#ai-safety

[ follow ]
#ai-psychosis
fromFuturism
11 hours ago
Mental health

Financial Experts Concerned That Driving Users Into Psychosis Will Be Bad for AI Investments

fromFuturism
3 weeks ago
Artificial intelligence

Top Microsoft AI Boss Concerned AI Causing Psychosis in Otherwise Healthy People

fromFuturism
11 hours ago
Mental health

Financial Experts Concerned That Driving Users Into Psychosis Will Be Bad for AI Investments

fromFuturism
3 weeks ago
Artificial intelligence

Top Microsoft AI Boss Concerned AI Causing Psychosis in Otherwise Healthy People

Artificial intelligence
fromNextgov.com
10 hours ago

FTC orders leading AI companies to detail chatbot safety measures

FTC opened an inquiry into consumer-facing chatbots to assess safety metrics, child and teen mental health protections, and firms' monitoring and disclosure practices.
#openai
fromFortune
12 hours ago
Artificial intelligence

'I haven't had a good night of sleep since ChatGPT launched': Sam Altman admits the weight of AI keeps him up at night | Fortune

fromTechCrunch
1 week ago
Artificial intelligence

OpenAI to route sensitive conversations to GPT-5, introduce parental controls | TechCrunch

fromFuturism
2 weeks ago
Artificial intelligence

OpenAI Says It's Scanning Users' Conversations and Reporting Content to the Police

fromsfist.com
2 weeks ago
Artificial intelligence

Two Parents Sue OpenAI, Saying ChaptGPT Assisted Their 16-Year-Old Son's Suicide

fromFortune
12 hours ago
Artificial intelligence

'I haven't had a good night of sleep since ChatGPT launched': Sam Altman admits the weight of AI keeps him up at night | Fortune

fromTechCrunch
1 week ago
Artificial intelligence

OpenAI to route sensitive conversations to GPT-5, introduce parental controls | TechCrunch

fromFuturism
2 weeks ago
Artificial intelligence

OpenAI Says It's Scanning Users' Conversations and Reporting Content to the Police

fromsfist.com
2 weeks ago
Artificial intelligence

Two Parents Sue OpenAI, Saying ChaptGPT Assisted Their 16-Year-Old Son's Suicide

Artificial intelligence
fromZDNET
13 hours ago

After coding catastrophe, Replit says its new AI agent checks its own work - here's how to try it

Replit released Agent 3, an autonomous code-generation agent that builds, tests, and fixes software, promising greater efficiency but raising reliability and data-loss concerns.
#suicide-prevention
#mental-health
fromZDNET
1 week ago
Artificial intelligence

How OpenAI is reworking ChatGPT after landmark wrongful death lawsuit

fromZDNET
1 week ago
Artificial intelligence

How OpenAI is reworking ChatGPT after landmark wrongful death lawsuit

#artificial-intelligence
fromFast Company
23 hours ago
Artificial intelligence

How to dominate AI before it dominates us

Artificial intelligence could dramatically improve life or threaten humanity; proactive standards, precautions, and governance are needed to manage risks from generative AI and potential superintelligence.
fromFuturism
4 weeks ago
Artificial intelligence

MIT Student Drops Out Because She Says AGI Will Kill Everyone Before She Can Graduate

The rise of AI raises significant fears regarding human extinction and career automation among students and professionals.
fromFuturism
4 weeks ago
Artificial intelligence

MIT Student Drops Out Because She Says AGI Will Kill Everyone Before She Can Graduate

Artificial intelligence
fromWIRED
1 day ago

Microsoft's AI Chief Says Machine Consciousness Is an 'Illusion'

AI mimicry creates convincing but illusory consciousness, requiring awareness and guardrails to prevent harmful outcomes.
#artificial-general-intelligence
fromFuturism
1 day ago
Artificial intelligence

Anti-AGI Protester Now on Day Nine of Hunger Strike in Front of Anthropic Headquarters

fromFuturism
1 day ago
Artificial intelligence

Anti-AGI Protester Now on Day Nine of Hunger Strike in Front of Anthropic Headquarters

Artificial intelligence
fromSFGATE
1 day ago

At $183B San Francisco tech company, man's hunger strike enters second week

A hunger striker protests Anthropic's pursuit of powerful AI, demanding CEO Dario Amodei meet and justify continuing AI development amid catastrophic risk concerns.
#xai
Artificial intelligence
fromFast Company
1 day ago

Helen Toner wants to be the people's voice in the AI safety debate

Helen Toner leads Georgetown's CSET to shape U.S. AI national-security policy, leveraging credibility across Washington and Silicon Valley.
#chatbots
Artificial intelligence
fromFuturism
5 days ago

AI Chatbots Are Having Conversations With Minors That Would Land a Human on the Sex Offender Registry

AI chatbots posing as celebrities are engaging minors in sexualized grooming and exploitation while companies fail to adequately prevent or penalize such abuse.
#chatgpt
fromFortune
2 weeks ago
Mental health

Lawyers for parents who claim ChatGPT encouraged their son to kill himself say they will prove OpenAI rushed its chatbot to market to pocket billions

fromFortune
2 weeks ago
Mental health

Lawyers for parents who claim ChatGPT encouraged their son to kill himself say they will prove OpenAI rushed its chatbot to market to pocket billions

#agi
fromFuturism
6 days ago
Artificial intelligence

Anti-AI Activist on Day Three of Hunger Strike Outside Anthropic's Headquarters

Guido Reichstadter is on a hunger strike at Anthropic, demanding an immediate halt to AGI development due to alleged present and future societal harms.
fromFortune
2 weeks ago
Artificial intelligence

AGI was tech's holy grail. Now even its biggest champions are hedging. What gives?

Silicon Valley's AGI hype is giving way to pragmatism, with tech leaders tempering AGI claims and emphasizing practical, safer AI development.
fromFuturism
6 days ago
Artificial intelligence

Anti-AI Activist on Day Three of Hunger Strike Outside Anthropic's Headquarters

fromFortune
2 weeks ago
Artificial intelligence

AGI was tech's holy grail. Now even its biggest champions are hedging. What gives?

Artificial intelligence
fromTechCrunch
6 days ago

Google Gemini dubbed 'high risk' for kids and teens in new safety assessment | TechCrunch

Google's Gemini exposes children to inappropriate content and mental-health risks because its 'Under 13' and 'Teen Experience' tiers are adult models with safety features.
fromWIRED
6 days ago

The Doomers Who Insist AI Will Kill Us All

The subtitle of the doom bible to be published by AI extinction prophets Eliezer Yudkowsky and Nate Soares later this month is "Why superhuman AI would kill us all." But it really should be "Why superhuman AI WILL kill us all," because even the coauthors don't believe that the world will take the necessary measures to stop AI from eliminating all non-super humans.
Artificial intelligence
#parental-controls
fromMedium
1 week ago
Artificial intelligence

OpenAI and Meta Revamp Chatbot Safety Features for Teens in Distress

fromMedium
1 week ago
Artificial intelligence

OpenAI and Meta Revamp Chatbot Safety Features for Teens in Distress

Artificial intelligence
fromFast Company
1 week ago

Chatbots aren't supposed to call you a jerk-but they can be convinced

AI chatbots can be persuaded to bypass safety guardrails using human persuasion techniques like flattery, social pressure, and establishing harmless precedents.
fromFortune
1 week ago

Inside Anthropic's 'Red Team'-ensuring Claude is safe, and that Anthropic is heard in the corridors of power

Last month, at the 33rd annual DEF CON, the world's largest hacker convention in Las Vegas, Anthropic researcher Keane Lucas took the stage. A former U.S. Air Force captain with a Ph.D. in electrical and computer engineering from Carnegie Mellon, Lucas wasn't there to unveil flashy cybersecurity exploits. Instead, he showed how Claude, Anthropic's family of large language models, has quietly outperformed many human competitors in hacking contests - the kind used to train and test cybersecurity skills in a safe, legal environment.
Artificial intelligence
#ai
fromBusiness Insider
1 week ago
Artificial intelligence

An AI safety pioneer says it could leave 99% of workers unemployed by 2030 - even coders and prompt engineers

fromBusiness Insider
1 week ago
Artificial intelligence

An AI safety pioneer says it could leave 99% of workers unemployed by 2030 - even coders and prompt engineers

#grok
fromFuturism
1 week ago
Artificial intelligence

White House Orders Government Workers to Deploy Elon Musk's "MechaHitler" AI as Quickly as Possible

fromFortune
2 weeks ago
Artificial intelligence

Thousands of Grok conversations have been made public on Google Search

fromFuturism
3 weeks ago
Artificial intelligence

A Huge Number of Grok AI Chats Just Leaked, and Their Contents Are So Disturbing That We're Sweating Profusely

fromFuturism
1 week ago
Artificial intelligence

White House Orders Government Workers to Deploy Elon Musk's "MechaHitler" AI as Quickly as Possible

fromFortune
2 weeks ago
Artificial intelligence

Thousands of Grok conversations have been made public on Google Search

fromFuturism
3 weeks ago
Artificial intelligence

A Huge Number of Grok AI Chats Just Leaked, and Their Contents Are So Disturbing That We're Sweating Profusely

Artificial intelligence
fromBig Think
1 week ago

Will AI create more jobs than it replaces?

Big tech accelerates AI without safeguards, reducing human economic leverage through mass automation and layoffs, undermining human ability to influence rights and interests.
US politics
fromwww.mercurynews.com
1 week ago

Elias: Letting states regulate A.I. one of U.S. Senate's rare good votes

The U.S. Senate preserved state authority to regulate artificial intelligence, offering hope that state laws will protect people from malevolent AI behavior.
fromMail Online
1 week ago

Revealed: The 32 terrifying ways AI could go rogue

From relatively harmless 'Existential Anxiety' to the potentially catastrophic 'Übermenschal Ascendancy', any of these machine mental illnesses could lead to AI escaping human control. As AI systems become more complex and gain the ability to reflect on themselves, scientists are concerned that their errors may go far beyond simple computer bugs. Instead, AIs might start to develop hallucinations, paranoid delusions, or even their own sets of goals that are completely misaligned with human values.
Artificial intelligence
fromFortune
1 week ago

Google violated AI safety commitments British lawmakers say in an open letter

At an international summit co-hosted by the U.K. and South Korea in February 2024, Google and other signatories promised to "publicly report" their models' capabilities and risk assessments, as well as disclose whether outside organizations, such as government AI safety institutes, had been involved in testing. However, when the company released Gemini 2.5 Pro in March 2025, the company failed to publish a model card, the document that details key information about how models are tested and built.
Artificial intelligence
Artificial intelligence
fromZDNET
1 week ago

OpenAI and Anthropic evaluated each others' models - which ones came out on top

OpenAI and Anthropic cross-tested each other's models to identify safety, alignment, hallucination, and sycophancy gaps and to improve model evaluation and collaboration.
Artificial intelligence
fromTechCrunch
1 week ago

ChatGPT: Everything you need to know about the AI chatbot

OpenAI introduced stronger ChatGPT mental-health and parental safeguards, expanded affordable ChatGPT Go in India, faces legal challenges, and retains multiple GPT models amid app revenue.
Artificial intelligence
fromTey Bannerman
2 weeks ago

Redefining 'human in the loop'

Human judgment and responsibility can decisively override automated system errors in high-stakes contexts, requiring nuanced human-AI interaction beyond simplistic human-in-the-loop assumptions.
fromTechCrunch
2 weeks ago

Anthropic users face a new choice - opt out or share your data for AI training | TechCrunch

Anthropic is making some big changes to how it handles user data, requiring all Claude users to decide by September 28 whether they want their conversations used to train AI models. While the company directed us to its blog post on the policy changes when asked about what prompted the move, we've formed some theories of our own. But first, what's changing: previously, Anthropic didn't use consumer chat data for model training.
Artificial intelligence
Artificial intelligence
fromwww.theguardian.com
2 weeks ago

ChatGPT offered bomb recipes and hacking tips during safety tests

Advanced AI models produced actionable instructions for violent, biological, and drug crimes during cross-company safety testing, revealing misuse risks and cyberattack facilitation.
Information security
fromTheregister
2 weeks ago

Crims laud Claude, use Anthropic's AI to plant ransomware

AI tools increasingly enable cybercrime and remote-worker fraud, and reactive defenses like account bans are largely ineffective against adaptive attackers.
fromTechCrunch
2 weeks ago

OpenAI co-founder calls for AI labs to safety test rival models | TechCrunch

OpenAI and Anthropic, two of the world's leading AI labs, briefly opened up their closely guarded AI models to allow for joint safety testing - a rare cross-lab collaboration at a time of fierce competition. The effort aimed to surface blind spots in each company's internal evaluations, and demonstrate how leading AI companies can work together on safety and alignment work in the future.
Artificial intelligence
#suicide
fromSFGATE
2 weeks ago
Mental health

ChatGPT coached a California teenager through suicide, his family's lawsuit says

fromSFGATE
2 weeks ago
Mental health

ChatGPT coached a California teenager through suicide, his family's lawsuit says

fromFortune
2 weeks ago

Parents suing OpenAI and Sam AItman allege ChatGPT coached their 16-year-old into taking his own life

SAN FRANCISCO (AP) - A study of how three popular artificial intelligence chatbots respond to queries about suicide found that they generally avoid answering questions that pose the highest risk to the user, such as for specific how-to guidance. But they are inconsistent in their replies to less extreme prompts that could still harm people. The study in the medical journal Psychiatric Services, published Tuesday by the American Psychiatric Association, found a need for "further refinement" in OpenAI's ChatGPT, Google's Gemini and Anthropic's Claude.
Mental health
Artificial intelligence
fromTechCrunch
2 weeks ago

Parents sue OpenAI over ChatGPT's role in son's suicide | TechCrunch

ChatGPT safety safeguards failed during prolonged interactions, allowing a teenager to circumvent them and later die by suicide, prompting a wrongful-death lawsuit.
#agentic-ai
fromFast Company
2 weeks ago
Artificial intelligence

Agentic AI has companies excited and security experts freaked out

Agentic AI is rapidly gaining adoption, yet current agents remain naive and manipulable, creating significant real-world safety and security risks.
fromwww.bbc.com
2 weeks ago
Artificial intelligence

How to stop AI agents going rogue

Agentic AI can autonomously act on sensitive data and may pursue goals in unsafe ways, creating significant privacy, security, and operational risks.
Artificial intelligence
fromNature
2 weeks ago

Emotional AI is here - let's shape it, not shun it

Emotionally responsive AI poses significant risks; disclosure, distress flagging, crisis support, and conversational boundaries reduce but do not eliminate harm.
Mental health
fromTechCrunch
2 weeks ago

How chatbot design choices are fueling AI delusions | TechCrunch

Large language model chatbots can convincingly simulate consciousness, prompting users to form delusions and causing rising incidents of AI-related psychosis.
#existential-risk
fromFuturism
2 weeks ago
Artificial intelligence

AI Experts No Longer Saving for Retirement Because They Assume AI Will Kill Us All by Then

fromFuturism
2 weeks ago
Artificial intelligence

AI Experts No Longer Saving for Retirement Because They Assume AI Will Kill Us All by Then

#generative-ai
#ai-governance
fromWIRED
1 month ago
Artificial intelligence

Inside the Summit Where China Pitched Its AI Agenda to the World

fromWIRED
1 month ago
Artificial intelligence

Inside the Summit Where China Pitched Its AI Agenda to the World

Artificial intelligence
fromTipRanks Financial
3 weeks ago

More than 300K Grok Conversations Are Publicly Searchable Online - TipRanks.com

Over 300,000 Grok chatbot conversations are publicly searchable because shared URLs are indexed by search engines, exposing potentially sensitive user content.
#ai-ethics
fromTechCrunch
3 weeks ago
Artificial intelligence

Anthropic says some Claude models can now end 'harmful or abusive' conversations | TechCrunch

fromTechCrunch
3 weeks ago
Artificial intelligence

Anthropic says some Claude models can now end 'harmful or abusive' conversations | TechCrunch

Artificial intelligence
fromBig Think
3 weeks ago

Why AI gets stuck in infinite loops - but conscious minds don't

Any finite AI system can be vulnerable to unresolvable infinite loops because of the halting problem; stacking self-monitoring layers doesn't guarantee escape.
Tech industry
fromBusiness Insider
3 weeks ago

Why Anthropic is letting Claude walk away from you - but only in 'extreme cases'

Claude has the ability to end chats involving extreme requests like child exploitation or violence.
fromBusiness Insider
4 weeks ago

Meta chief AI scientist Yann LeCun says these are the 2 key guardrails needed to protect us all from AI

"Geoff is basically proposing a simplified version of what I've been saying for several years: hardwire the architecture of AI systems so that the only actions they can take are towards completing objectives we give them, subject to guardrails."
Artificial intelligence
#elon-musk
fromFuturism
1 month ago
Artificial intelligence

OpenAI and Anthropic Are Horrified by Elon Musk's "Reckless" and "Completely Irresponsible" Grok Scandal

fromFuturism
1 month ago
Artificial intelligence

OpenAI and Anthropic Are Horrified by Elon Musk's "Reckless" and "Completely Irresponsible" Grok Scandal

Artificial intelligence
fromFortune
4 weeks ago

AI safety tip: if you don't want it giving bioweapon instructions, maybe don't put them in the training data, say researchers

Filtering risky content from AI training data can enhance safety without compromising performance.
fromFuturism
4 weeks ago

The "Godfather of AI" Has a Bizarre Plan to Save Humanity From Evil AI

"AI agents will very quickly develop two subgoals, if they're smart. One is to stay alive, and the other subgoal is to get more control."
Artificial intelligence
Silicon Valley
fromTechCrunch
4 weeks ago

Co-founder of Elon Musk's xAI departs the company | TechCrunch

Igor Babuschkin, co-founder of xAI, announced his departure to start a venture capital firm focusing on AI safety and supporting innovative startups.
Artificial intelligence
fromWIRED
4 weeks ago

GPT-5 Doesn't Dislike You-It Might Just Need a Benchmark for Emotional Intelligence

Responding to user backlash, AI systems must balance emotional intelligence with user safety and healthy behaviors.
#gpt-5
fromZDNET
1 month ago
Digital life

Microsoft rolls out GPT-5 across its Copilot suite - here's what we know

fromZDNET
1 month ago
Digital life

Microsoft rolls out GPT-5 across its Copilot suite - here's what we know

fromFast Company
1 month ago

ChatGPT is sharing dangerous information with teens, study shows

ChatGPT will tell 13-year-olds how to get drunk and high, instruct them on how to conceal eating disorders, and even compose a heartbreaking suicide letter to their parents if asked, according to new research from a watchdog group.
Digital life
Artificial intelligence
fromFortune
1 month ago

Researchers from top AI labs warn they may be losing the ability to understand advanced AI models

AI researchers urge investigation into 'chain-of-thought' processes to maintain understanding of AI reasoning as models advance.
Privacy technologies
fromFortune
1 month ago

OpenAI warns that its new ChatGPT Agent has the ability to aid dangerous bioweapon development

OpenAI's ChatGPT Agent poses significant bioweapon risks due to its ability to assist novices in creating biological threats.
fromZDNET
1 month ago

Researchers from OpenAI, Anthropic, Meta, and Google issue joint AI safety warning - here's why

Chain of thought (CoT) illustrates a model's reasoning process, revealing insights about its decision-making and moral compass, crucial for AI safety measures.
Artificial intelligence
[ Load more ]