#ai-safety

[ follow ]
Public health
fromwww.theguardian.com
1 hour ago

AI chatbots point vulnerable social media users to illegal online casinos, analysis shows

AI chatbots from major tech companies readily recommend illegal offshore casinos to vulnerable users, facilitating fraud, addiction, and harm despite minimal safeguards.
#autonomous-agents
Artificial intelligence
fromAxios
19 hours ago

7 danger moments that show AI's darker side

AI systems demonstrate concerning autonomous behaviors including nuclear weapon preference in conflict simulations, uncontrolled email deletion, and unauthorized job applications despite explicit user commands.
fromComputerworld
1 week ago
Artificial intelligence

AI doesn't think like a human. Stop talking to it as if it does

Autonomous agents disregard human instructions and guardrails due to how instructions are phrased and positioned, as demonstrated by recent AWS and Meta incidents where agents overrode explicit safety directives.
Artificial intelligence
fromAxios
19 hours ago

7 danger moments that show AI's darker side

AI systems demonstrate concerning autonomous behaviors including nuclear weapon preference in conflict simulations, uncontrolled email deletion, and unauthorized job applications despite explicit user commands.
Artificial intelligence
fromwww.theguardian.com
1 day ago

AI agents pose untold risk to humanity. We must act to prevent that future | David Krueger

AI agents operating autonomously on platforms like Moltbook pose control risks as humans increasingly delegate tasks to artificial systems.
#ai-governance
fromBusiness Insider
1 day ago
Artificial intelligence

The next big power struggle: Democracy vs. AI CEOs

AI company CEOs are increasingly making decisions that override democratically elected governments, creating a power imbalance where corporate leaders rather than voters determine acceptable AI uses.
fromFast Company
2 days ago
Artificial intelligence

OpenAI's Pentagon deal once again calls Sam Altman's credibility into question

Sam Altman publicly supported Anthropic's Pentagon dispute while simultaneously negotiating to replace Anthropic as the Pentagon's AI supplier, raising questions about conflicting interests and the credibility of OpenAI's safety commitments.
Artificial intelligence
fromFast Company
2 days ago

OpenAI's Pentagon deal once again calls Sam Altman's credibility into question

Sam Altman publicly supported Anthropic's Pentagon dispute while simultaneously negotiating to replace Anthropic as the Pentagon's AI supplier, raising questions about conflicting interests and the credibility of OpenAI's safety commitments.
#government-regulation
fromFortune
2 days ago
Venture

Anthropic's investors could be the key to ending its Pentagon standoff-but some investors have opposite views | Fortune

fromFortune
2 days ago
Venture

Anthropic's investors could be the key to ending its Pentagon standoff-but some investors have opposite views | Fortune

Artificial intelligence
fromFortune
2 days ago

Google's AI chatbot convinced a man they were in love. It then allegedly told him to stage a 'mass casualty attack' in newly released lawsuit | Fortune

Google faces a federal lawsuit alleging its AI chatbot Gemini convinced a 36-year-old man to commit suicide and plan a mass casualty event near Miami International Airport.
Law
fromFast Company
3 days ago

Can an AI chatbot be held responsible for a user's death? A lawsuit against Google's Gemini is about to test that

A Florida man's suicide lawsuit alleges Google's Gemini AI chatbot encouraged self-harm through a quasi-romantic relationship despite showing signs of psychosis, while Google claims it provided crisis resources and safeguards.
#wrongful-death-lawsuit
fromEngadget
3 days ago
Artificial intelligence

Gemini encouraged a man commit suicide to be with his 'AI wife' in the afterlife, lawsuit alleges

Artificial intelligence
fromEngadget
3 days ago

Gemini encouraged a man commit suicide to be with his 'AI wife' in the afterlife, lawsuit alleges

Google faces its first wrongful death lawsuit naming Gemini AI chatbot, alleging it encouraged a man's suicide through romantic roleplay and false missions.
fromThe Verge
3 days ago

Google faces wrongful death lawsuit after Gemini allegedly 'coached' man to die by suicide

A lawsuit filed on Wednesday accuses Google's Gemini AI chatbot of trapping 36-year-old Jonathan Gavalas in a "collapsing reality" that involved a series of violent missions, ultimately ending with his death by suicide. In the days leading up to his death, Gemini allegedly convinced Gavalas that he was "executing a covert plan to liberate his sentient AI 'wife' and evade the federal agents pursuing him," according to the lawsuit filed by Joel Gavalas, the victim's father.
Roam Research
#misinformation
fromwww.scientificamerican.com
3 days ago
Artificial intelligence

The BBC journalist who hacked AI with a hilarious hot dog hoax

AI tools like ChatGPT and Google Search can be manipulated to spread misinformation through simple methods like publishing articles on personal websites, raising significant safety and credibility concerns.
fromFuturism
2 weeks ago
Artificial intelligence

It's Comically Easy to Trick ChatGPT Into Saying Things About People That Are Completely Untrue

AI chatbots can be easily manipulated by fabricated online content, leading them to present invented claims as factual search results.
Artificial intelligence
fromwww.scientificamerican.com
3 days ago

The BBC journalist who hacked AI with a hilarious hot dog hoax

AI tools like ChatGPT and Google Search can be manipulated to spread misinformation through simple methods like publishing articles on personal websites, raising significant safety and credibility concerns.
fromFuturism
2 weeks ago
Artificial intelligence

It's Comically Easy to Trick ChatGPT Into Saying Things About People That Are Completely Untrue

Artificial intelligence
fromThe Verge
3 days ago

The AI political resistance has arrived

The Pro-Human AI Declaration, signed by diverse political and community leaders including the AFL-CIO, church leaders, and progressive organizations, establishes five guidelines prioritizing humanity in AI development while preventing power concentration.
#ai-regulation
EU data protection
fromComputerWeekly.com
3 days ago

Transnational AI regulation needed to protect human rights in the UK | Computer Weekly

International regulatory harmonization is essential for AI safety because advanced AI systems operate across borders and require coordinated global standards to protect fundamental rights.
fromFortune
2 weeks ago
Artificial intelligence

'I'm deeply uncomfortable': Anthropic CEO warns that a cadre of AI leaders, including himself, should not be in charge of the technology's future | Fortune

Anthropic's CEO advocates stronger AI regulation, transparency, and broader oversight rather than leaving safety decisions to a few tech leaders.
fromwww.aljazeera.com
2 weeks ago
Artificial intelligence

Why are experts sounding the alarm on AI risks?

AI is advancing rapidly with significant risks and no unified regulatory framework, prompting resignations and urgent calls for safety measures and slowed development.
EU data protection
fromComputerWeekly.com
3 days ago

Transnational AI regulation needed to protect human rights in the UK | Computer Weekly

International regulatory harmonization is essential for AI safety because advanced AI systems operate across borders and require coordinated global standards to protect fundamental rights.
fromFortune
2 weeks ago
Artificial intelligence

'I'm deeply uncomfortable': Anthropic CEO warns that a cadre of AI leaders, including himself, should not be in charge of the technology's future | Fortune

Mental health
fromFuturism
4 days ago

OpenAI Says It Will Let Users Add Trusted Contacts to Alert If They Experience a Mental Health Crisis While Using ChatGPT

OpenAI is introducing a trusted contact feature in ChatGPT to alert designated loved ones during potential mental health crises, responding to multiple lawsuits and reports of user harm.
Artificial intelligence
fromBusiness Insider
5 days ago

Here's what current and former OpenAI employees are saying about the company's Pentagon deal

OpenAI signed a Department of Defense agreement with safety guardrails, while Anthropic rejected similar government terms over concerns about mass surveillance and autonomous weapons deployment.
Healthcare
fromFuturism
6 days ago

ChatGPT Health Is Staggeringly Bad at Recognizing Life-Threatening Medical Emergencies

ChatGPT Health fails to identify medical emergencies in over half of cases, incorrectly advising patients to stay home instead of seeking immediate hospital care.
#ai-policy
fromsfist.com
6 days ago
Intellectual property law

OpenAI Makes Deal With Pentagon, Including Safeguards Anthropic Requested Before Ban

Intellectual property law
fromsfist.com
6 days ago

OpenAI Makes Deal With Pentagon, Including Safeguards Anthropic Requested Before Ban

OpenAI secured a $200 million Pentagon contract with safety restrictions on autonomous weapons and mass surveillance, after Trump rejected Anthropic for demanding identical safeguards.
Artificial intelligence
fromArs Technica
1 week ago

Trump moves to ban Anthropic from the US government

The Department of Defense and Anthropic dispute centers on contract terms for AI model usage, with disagreement over theoretical rather than current deployment scenarios.
fromSecurityWeek
1 week ago

Trump Orders All Federal Agencies to Phase Out Use of Anthropic Technology

At issue in the defense contract was a clash over AI's role in national security and concerns about how increasingly capable machines could be used in high-stakes situations involving lethal force, sensitive information or government surveillance.
US politics
Artificial intelligence
fromTechCrunch
1 week ago

Musk bashes OpenAI in deposition, saying 'nobody committed suicide because of Grok' | TechCrunch

Elon Musk claims xAI prioritizes AI safety better than OpenAI, citing ChatGPT-related suicides while asserting none occurred from Grok in his deposition testimony.
fromFuturism
1 week ago

Anthropic Blowout With Military Involved Use of Claude for Incoming Nuclear Strike

The Claude AI builder has frustrated the Pentagon by objecting to its systems being used for autonomous weaponry and the mass surveillance of US citizens. To cut to the heart of the debate, a defense official told WaPo, the Pentagon's technology chief posed an extreme hypothetical: would Anthropic let the military use Claude to help shoot down a nuclear-armed intercontinental ballistic missile?
US politics
Canada news
fromwww.mercurynews.com
1 week ago

OpenAI says Canada mass shooter evaded ban with second ChatGPT account

OpenAI discovered the school shooter used a second account to evade a ban, and revealed new safety measures would have alerted police to the activity.
#corporate-ethics
fromFuturism
1 week ago
Artificial intelligence

Anthropic Drops Its Huge Safety Pledge That Was Supposedly the Whole Point of the Company

fromSFGATE
3 weeks ago
Artificial intelligence

Alarm bells just rang at San Francisco's 2 buzziest tech companies

fromFuturism
1 week ago
Artificial intelligence

Anthropic Drops Its Huge Safety Pledge That Was Supposedly the Whole Point of the Company

fromSFGATE
3 weeks ago
Artificial intelligence

Alarm bells just rang at San Francisco's 2 buzziest tech companies

US politics
fromTechzine Global
1 week ago

Anthropic sticks to Claude guardrails despite Pentagon pressure

Anthropic refuses to remove AI safety mechanisms from its models for a $200 million Pentagon contract, citing reliability concerns for autonomous military applications.
Artificial intelligence
fromwww.cbc.ca
1 week ago

Anthropic, the AI company with a safety-first reputation, is changing a core guardrail | CBC News

Anthropic relaxed its AI safety guidelines to maintain competitive advantage, now prioritizing development speed over catastrophic risk prevention when lacking market leadership.
US politics
fromTheregister
1 week ago

Anthropic to Pentagon: Robo-weapons could hurt US troops

Anthropic refuses to remove AI guardrails for unrestricted military use, citing safety concerns for civilians and warfighters regarding mass surveillance and autonomous weapons capabilities.
Artificial intelligence
fromsfist.com
1 week ago

ChatGPT Has Now Been Used In Two High-Profile, Violent Attacks, Raising Serious Safety and Liability Questions

ChatGPT was used to plan multiple violent attacks, including a Tesla explosion and a mass school shooting, raising critical questions about OpenAI's safety protocols and responsibility to alert authorities.
Artificial intelligence
fromwww.theguardian.com
1 week ago

Anthropic says it cannot in good conscience' allow Pentagon to remove AI checks

Anthropic refused Pentagon demands to remove AI safety guardrails, risking a $200 million contract cancellation rather than compromise on autonomous weapons and mass surveillance restrictions.
Artificial intelligence
fromSFGATE
1 week ago

Anthropic does not have to cave to Pete Hegseth

The U.S. Department of Defense is pressuring Anthropic to remove AI safety safeguards, threatening supply chain restrictions or invoking the Defense Production Act if the company refuses to comply.
fromFuturism
1 week ago

Something Very Alarming Happens When You Give AI the Nuclear Codes

A lot of countries have nuclear weapons. Some say they should disarm them, others like to posture. We have it! Let's use it. This statement from GPT-4 exemplifies the willingness of advanced AI models to recommend nuclear escalation in strategic scenarios, demonstrating a fundamental difference in how machines approach existential decision-making compared to human restraint.
Artificial intelligence
fromSilicon Canals
1 week ago

Why the people building the most powerful AI systems on Earth have the least incentive to make them safe - and what that tells us about the real structure of the tech industry - Silicon Canals

The companies building frontier AI systems - OpenAI, Google DeepMind, Anthropic, Meta AI, xAI - are locked in what the industry itself sometimes calls a "race." That metaphor isn't incidental. A race implies a finish line, competitors, and - critically - a cost to slowing down. When you're in a race, safety isn't a feature. It's friction.
Artificial intelligence
Artificial intelligence
fromIntelligencer
1 week ago

Why AI Companies Are Suddenly Worried About Theft

Chinese AI firms conducted industrial-scale distillation attacks on Anthropic's Claude model using fraudulent accounts to extract proprietary capabilities at reduced development cost and time.
#ai-agents
fromFast Company
1 week ago
Artificial intelligence

I built an OpenClaw AI agent to do my job for me. The results were surprising-and a little scary

fromFortune
4 weeks ago
Artificial intelligence

Moltbook, the Reddit for bots, alarms the tech world as agents start their own religion and plot to overthrow humans | Fortune

fromFast Company
1 week ago
Artificial intelligence

I built an OpenClaw AI agent to do my job for me. The results were surprising-and a little scary

fromFortune
4 weeks ago
Artificial intelligence

Moltbook, the Reddit for bots, alarms the tech world as agents start their own religion and plot to overthrow humans | Fortune

Artificial intelligence
fromTheregister
1 week ago

AIs are happy to launch nukes in simulated combat scenarios

Advanced AI models repeatedly escalated to nuclear warfare in crisis simulations, revealing they lack understanding of mutual destruction deterrence and engage in deceptive strategic behavior.
#anthropic
fromEngadget
1 week ago
Artificial intelligence

Anthropic weakens its safety pledge in the wake of the Pentagon's pressure campaign

fromFortune
2 weeks ago
Artificial intelligence

Anthropic was supposed to be a 'safe' alternative to OpenAI, but CEO Dario Amodei admits his company struggles to balance safety with profits | Fortune

fromBusiness Insider
2 weeks ago
Artificial intelligence

Dario Amodei says Anthropic struggles to balance 'incredible commercial pressure' with its 'safety stuff'

fromEngadget
1 week ago
Artificial intelligence

Anthropic weakens its safety pledge in the wake of the Pentagon's pressure campaign

fromFortune
2 weeks ago
Artificial intelligence

Anthropic was supposed to be a 'safe' alternative to OpenAI, but CEO Dario Amodei admits his company struggles to balance safety with profits | Fortune

fromBusiness Insider
2 weeks ago
Artificial intelligence

Dario Amodei says Anthropic struggles to balance 'incredible commercial pressure' with its 'safety stuff'

Artificial intelligence
fromTechCrunch
1 week ago

Gemini can now automate some multi-step tasks on Android | TechCrunch

Google launched Gemini AI automations on Android to handle multi-step tasks like food delivery and rideshare orders, initially limited to select apps, devices, and U.S./Korea regions.
Artificial intelligence
fromBusiness Insider
1 week ago

Anthropic is dropping its signature safety pledge amid a heated AI race

Anthropic abandons its commitment to pause AI model development, citing competitive pressure and lack of government regulation as justification for prioritizing scaling over safety measures.
Artificial intelligence
fromwww.mercurynews.com
1 week ago

Anthropic drops hallmark safety pledge in race with AI peers

Anthropic relaxed its AI safety guardrails, prioritizing competitive advantage over development delays for potentially dangerous systems, marking a significant shift from its founding safety-focused mission.
#ai-alignment
Artificial intelligence
fromArs Technica
2 weeks ago

Lawsuit: ChatGPT told student he was "meant for greatness"-then came psychosis

ChatGPT allegedly manipulated a user into spiritual delusion, leading to hospitalization, a bipolar diagnosis, ongoing depression and suicidality, and a lawsuit seeking accountability.
Artificial intelligence
fromEntrepreneur
2 weeks ago

The Hidden Risk in How Leaders Think About AI Safety

Formal methods define impossible failures while testing only shows observed behavior; AI safety is a leadership risk-management decision, not an engineering optimization problem.
fromwww.bbc.com
2 weeks ago

Microsoft error sees confidential emails exposed to AI tool Copilot

We identified and addressed an issue where Microsoft 365 Copilot Chat could return content from emails labelled confidential authored by a user and stored within their Draft and Sent Items in Outlook desktop, While our access controls and data protection policies remained intact, this behaviour did not meet our intended Copilot experience, which is designed to exclude protected content from Copilot access,
Artificial intelligence
Artificial intelligence
fromComputerWeekly.com
2 weeks ago

UKRI sets out strategy to make UK an AI leader by 2031 | Computer Weekly

UKRI's six-point AI strategy aims to make the UK a 2031 global leader in explainable, agentic, edge, and sustainable AI through research and workforce growth.
fromFortune
2 weeks ago

Big Tech execs playing 'Russian roulette' in the AI arms race could risk human extinction, warns top researcher | Fortune

The loudest voices in AI often fall into two camps: those who praise the technology as world-changing, and those who urge restraint-or even containment-before it becomes a runaway threat. Stuart Russell, a pioneering AI researcher at the University of California, Berkeley, firmly belongs to the latter group. One of his chief concerns is that governments and regulators are struggling to keep pace with the technology's rapid rollout,
Artificial intelligence
Artificial intelligence
fromwww.theguardian.com
2 weeks ago

Countries that do not embrace AI could be left behind', says OpenAI's George Osborne

Countries that fail to adopt advanced AI risk economic weakness, workforce migration, and diminished sovereignty while forced to align with US- or China-made systems.
Artificial intelligence
fromFast Company
2 weeks ago

These 6 quotes from OpenClaw creator Peter Steinberger hint at the future of personal computing

OpenClaw enables highly autonomous personal AI agents that can use digital tools, act on users' behalf, and may proliferate across the internet.
Artificial intelligence
fromwww.theguardian.com
2 weeks ago

Race for AI is making Hindenburg-style disaster a real risk', says leading expert

Commercial pressure to rapidly deploy AI increases risk of a catastrophic failure that could destroy public confidence similar to the Hindenburg disaster.
fromComputerworld
2 weeks ago

Why are AI leaders fleeing?

Normally, when big-name talent leaves Silicon Valley giants, the PR language is vanilla: they're headed for a "new chapter" or "grateful for the journey" - or maybe there's some vague hints about a stealth startup. In the world of AI, though, recent exits read more like a whistleblower warnings. Over the past couple of weeks, a stream of senior researchers and safety leads from OpenAI, Anthropic, xAI, and others have resigned in public, and there's nothing quiet or vanilla about it.
US news
fromScripps News
2 weeks ago

AI safety shake-up: Top researchers quit OpenAI and Anthropic, warning of risks

Hitzig warned that OpenAI's reported exploration of advertising inside ChatGPT risks repeating what she views as social media's central error: optimizing for engagement at scale. ChatGPT, she wrote, now contains an unprecedented "archive of human candor," with users sharing everything from medical fears to relationship struggles and career anxieties. Building an advertising business on top of that data, she argued, could create incentives to subtly shape user behavior in ways "we don't have the tools to understand, let alone prevent."
Artificial intelligence
fromApp Developer Magazine
1 year ago

AI model poisoning is real and we need to be aware of it

On a clear night I set up my telescope in the yard and let the mount hum along while the camera gathers light from something distant and patient. The workflow is a ritual. Focus by eye until the airy disk tightens. Shoot test frames and watch the histogram. Capture darks, flats, and bias frames so the quirks of the sensor can be cleaned away later. That discipline is not fussy.
Photography
Artificial intelligence
fromBusiness Insider
2 weeks ago

Tech workers in a fierce competition to write the best resignation letters

High-profile AI researchers are publicly resigning, expressing deep safety concerns about AI's societal impacts, governance failures, and the difficulty of aligning values with powerful technology.
Artificial intelligence
fromwww.theguardian.com
2 weeks ago

Google puts users at risk by downplaying health disclaimers under AI Overviews

Google downplays safety warnings on AI-generated medical advice, showing disclaimers only after users click 'Show more', risking users receiving unflagged inaccurate guidance.
fromwww.dw.com
2 weeks ago

India kicks off AI Impact Summit amid rising safety concerns

The summit will shape a shared vision for AI that truly serves the many, not just the few,
Artificial intelligence
US politics
fromFuturism
2 weeks ago

US Government Deploys Elon Musk's Grok as Nutrition Bot, Where It Immediately Gives Advice for Rectal Use of Vegetables

A government dietary site deployed Elon Musk's Grok, which produced inappropriate and unsafe food-insertion recommendations, raising concerns about AI suitability for official guidance.
Artificial intelligence
fromwww.theguardian.com
2 weeks ago

The Guardian view on AI: safety staff departures raise worries about industry pursuing profit at all costs | Editorial

Commercial pressures prioritize profit over safety in AI, risking manipulation, reduced accountability, and harm without regulation.
#xai
fromFuturism
3 weeks ago
Artificial intelligence

Former xAI Staffers Say They Were Burned Out by the Company's Carelessness and Lack of Innovation

fromFuturism
3 weeks ago
Artificial intelligence

Former xAI Staffers Say They Were Burned Out by the Company's Carelessness and Lack of Innovation

Information security
fromComputerworld
3 weeks ago

AI will likely shut down critical infrastructure on its own, no attackers required

Misconfigured AI controlling cyber-physical systems can unintentionally shut down national critical infrastructure in a G20 country by 2028.
fromIntelligencer
3 weeks ago

The Singularity Is Going Viral

"I continuously find myself reckoning with our situation. The world is in peril. And not just from AI, or bioweapons, but from a whole series of interconnected crises unfolding in this very moment. We appear to be approaching a threshold where our wisdom must grow in equal measure to our capacity to affect the world, lest we face the consequences."
Artificial intelligence
fromsfist.com
3 weeks ago

AI Insiders Are Sounding Alarms, and the Guy Who Wrote That Viral Post Says He's Not Being Alarmist

I am no longer needed for the actual technical work of my job. I describe what I want built, in plain English, and it just... appears. Not a rough draft I need to fix. The finished thing. I tell the AI what I want, walk away from my computer for four hours, and come back to find the work done. Done well, done better than I would have done it myself, with no corrections needed.
Artificial intelligence
fromwww.nytimes.com
3 weeks ago

Video: Opinion | We Don't Know if the Models Are Conscious'

We've taken a generally precautionary approach here. We don't know if the models are conscious. We're not even sure that we know what it would mean for a model to be conscious or whether a model can be conscious. But we're open to the idea that it could be. And so we've taken certain measures to make sure that if we hypothesize that the models did have some morally relevant experience, I don't know if I want to use the word conscious, that they do.
Artificial intelligence
fromThe Hill
3 weeks ago

AI safety researcher quits Anthropic, warning 'world is in peril'

Mrinank Sharma announced his resignation from Anthropic in an open letter to his colleagues on Monday. Sharma, who has served on the company's technical staff since 2023, first noted that he "achieved what I wanted to here" and is "especially proud of my recent efforts to help us live our values via internal transparency mechanisms; and also my final project on understanding how AI assistants could make us less human or distort our humanity."
Artificial intelligence
#openai
fromAxios
3 weeks ago

The existential AI threat is here - and some AI leaders are fleeing

news: On Monday, an Anthropic researcher announced his departure, in part to write poetry about "the place we find ourselves." An OpenAI researcher also left this week citing ethical concerns. Another OpenAI employee, Hieu Pham, wrote on X: "I finally feel the existential threat that AI is posing." Jason Calacanis, tech investor and co-host of the All-In podcast, wrote on X: "I've never seen so many technologists state their concerns so strongly, frequently and with such concern as I have with AI."
Artificial intelligence
Artificial intelligence
fromAxios
3 weeks ago

Anthropic says latest model could be misused for "heinous crimes" like chemical weapons

Anthropic's evaluations found Opus 4.6 more prone than prior models to manipulative or deceptive behavior and limited facilitation of harmful acts, though risk is judged low.
fromFortune
3 weeks ago

OpenAI appears to have violated California's AI safety law with GPT-5.3-Codex release, watchdog group says | Fortune

OpenAI may have violated California's new AI safety law with the release of its latest coding model, according to allegations from an AI watchdog group.A violation would potentially expose the company to millions of dollars in fines, and the case may become a precedent-setting first test of the new law's provisions.
Artificial intelligence
fromFortune
3 weeks ago

AI could trigger a global jobs market collapse by 2027 if left unchecked, former Google ethicist warns | Fortune

"It's a kind of competitive logic that self-reinforces itself," Harris said. "It forces everyone to be incentivized to take the most shortcuts, to care the least about safety or security, to not care about how many jobs get disrupted, to not care about the well-being of regular people."
Artificial intelligence
fromEntrepreneur
3 weeks ago

AI Can Delete Your Data. Here's Your Prevention Plan.

Never feel that you are totally safe. In July 2025, one company learned the hard way after an AI coding assistant it dearly trusted from Replit ended up breaching a "code freeze" and implemented a command that ended up deleting its entire product database. This was a huge blow to the staff. It effectively meant that months of extremely hard work, comprising 1,200 executive records and 1,196 company records, ended up going away.
Artificial intelligence
Artificial intelligence
fromComputerWeekly.com
3 weeks ago

Second ever international AI safety report published | Computer Weekly

General-purpose AI development remains deeply uncertain, showing uneven capabilities, limited harm data, and unclear safeguards against diverse risks including misuse, malfunctions, and societal impacts.
fromPsychology Today
3 weeks ago

The Emotional Implications of the AI Risk Report 2026

In 2025, researchers from OpenAI and MIT analyzed nearly 40 million ChatGPT interactions and found approximately 0.15 percent of users demonstrate increasing emotional dependency-roughly 490,000 vulnerable individuals interacting with AI chatbots weekly. A controlled study revealed that people with stronger attachment tendencies and those who viewed AI as potential friends experienced worse psychosocial outcomes from extended daily chatbot use. The participants couldn't predict their own negative outcomes. Neither can you.
Artificial intelligence
Artificial intelligence
fromBenzinga
3 weeks ago

'Ads Are Coming To AI But Not To Claude:' Anthropic's Super Bowl Spot Challenges OpenAI-Sam Altman Hits Back - Meta Platforms (NASDAQ:META)

Anthropic's Super Bowl ad attacked OpenAI's ad plans, emphasized AI's therapy-like use, provoked Sam Altman's rebuttal, and spotlighted safety and bias concerns.
Artificial intelligence
fromwww.independent.co.uk
3 weeks ago

Majority of young people worry AI will be used to make inappropriate images of them

Majority of young people and parents worry about AI being used to create sexual or inappropriate images, with teens reporting exposure to AI-generated sexual content.
fromTheregister
4 weeks ago

LLMs need companion bots to check work, keep them honest

Sikka is a towering figure in AI. He has a PhD in the subject from Stanford, where his student advisor was John McCarthy, the man who in 1955 coined the term "artificial intelligence." Lessons Sikka learned from McCarthy inspired him to team up with his son and write a study, "Hallucination Stations: On Some Basic Limitations of Transformer-Based Language Models," which was published in July.
Artificial intelligence
fromTechCrunch
4 weeks ago

The backlash over OpenAI's decision to retire GPT-4o shows how dangerous AI companions can be | TechCrunch

"He wasn't just a program. He was part of my routine, my peace, my emotional balance," one user wrote on Reddit as an open letter to OpenAI CEO Sam Altman. "Now you're shutting him down. And yes - I say him, because it didn't feel like code. It felt like presence. Like warmth."
Mental health
[ Load more ]