#self-verifying-ai

[ follow ]
#openai
fromWIRED
45 minutes ago
Information security

In the Wake of Anthropic's Mythos, OpenAI Has a New Cybersecurity Model-and Strategy

Artificial intelligence
fromFuturism
2 days ago

Why Does It Suddenly Feel Like OpenAI Is Melting Down Into Disaster?

OpenAI is preparing for a potential IPO with a valuation of up to $1 trillion, despite facing significant challenges and controversies this year.
Artificial intelligence
fromAxios
1 day ago

OpenAI rips Anthropic, distances itself from Microsoft

OpenAI's partnership with Amazon Web Services is driving significant demand from enterprise customers, enhancing distribution capabilities beyond Microsoft.
Information security
fromAxios
45 minutes ago

OpenAI expands access to cyber AI as hacking risks grow

OpenAI is shifting to a model that emphasizes identity verification for access to sensitive cybersecurity tools while expanding availability.
Law
fromFuturism
2 days ago

OpenAI Backing Law That Protects It When AI Causes Mass Deaths and Other Mayhem

Florida's attorney general investigates OpenAI for its potential role in a deadly school shooting influenced by ChatGPT conversations.
Information security
fromWIRED
45 minutes ago

In the Wake of Anthropic's Mythos, OpenAI Has a New Cybersecurity Model-and Strategy

OpenAI announced GPT-5.4-Cyber, emphasizing cybersecurity safeguards and the need for advanced protections in AI models.
Artificial intelligence
fromFuturism
2 days ago

Why Does It Suddenly Feel Like OpenAI Is Melting Down Into Disaster?

OpenAI is preparing for a potential IPO with a valuation of up to $1 trillion, despite facing significant challenges and controversies this year.
Artificial intelligence
fromAxios
1 day ago

OpenAI rips Anthropic, distances itself from Microsoft

OpenAI's partnership with Amazon Web Services is driving significant demand from enterprise customers, enhancing distribution capabilities beyond Microsoft.
US news
fromwww.npr.org
2 hours ago

Law enforcement is trying to combat abusive AI. Experts say easier said than done

An Ohio man was convicted under the 2025 Take It Down Act for creating and distributing AI-generated abusive sexual images.
Science
fromwww.scientificamerican.com
10 hours ago

China said it could break military encryption. DARPA built an AI to find out if that's true

Chinese researchers claim quantum computers could potentially unlock encrypted information, raising concerns for national security.
Marketing tech
fromAdExchanger
1 day ago

AI Is Nothing Without Data Fidelity. Here's A Four-Step Approach to Protect It | AdExchanger

Data integrity is crucial for effective AI in advertising, as flawed data leads to poor outcomes.
Intellectual property law
fromWIRED
5 hours ago

Anthropic Opposes the Extreme AI Liability Bill That OpenAI Backed

Anthropic opposes Illinois bill SB 3444, which would shield AI labs from liability for large-scale harm caused by their systems.
Games
fromThe Atlantic
5 hours ago

The Strange Origin of AI's 'Reasoning' Abilities

Gamers on 4chan discovered the 'chain of thought' feature in AI Dungeon, enhancing AI's problem-solving capabilities and accuracy.
#ai
Data science
fromInfoWorld
2 weeks ago

A data trust scoring framework for reliable and responsible AI systems

A rigorous trust scoring framework is essential to prevent AI from perpetuating inequality through biased data.
Information security
fromTechzine Global
1 day ago

Runtime security becomes critical as AI accelerates threats

Artificial intelligence accelerates innovation and cyber threats, necessitating a focus on runtime security for effective enterprise protection.
Artificial intelligence
fromTechCrunch
1 day ago

Stanford report highlights growing disconnect between AI insiders and everyone else | TechCrunch

Public opinion on AI is increasingly negative, with growing anxiety about its impact on jobs, healthcare, and the economy.
Artificial intelligence
fromThe Verge
6 hours ago

Has Google's AI watermarking system been reverse-engineered?

A developer claims to have reverse-engineered Google's SynthID watermarking system, asserting that AI watermarks can be stripped from images.
Data science
fromInfoWorld
2 weeks ago

A data trust scoring framework for reliable and responsible AI systems

A rigorous trust scoring framework is essential to prevent AI from perpetuating inequality through biased data.
Information security
fromTechzine Global
1 day ago

Runtime security becomes critical as AI accelerates threats

Artificial intelligence accelerates innovation and cyber threats, necessitating a focus on runtime security for effective enterprise protection.
Artificial intelligence
fromTechCrunch
1 day ago

Stanford report highlights growing disconnect between AI insiders and everyone else | TechCrunch

Public opinion on AI is increasingly negative, with growing anxiety about its impact on jobs, healthcare, and the economy.
#ai-governance
fromNature
20 hours ago

Researchers: here's how to audit your fragmented digital identity

A search for 'Guo Wei' in ORCID returned 616 profiles, none affiliated with the Jiangsu University of Science and Technology, highlighting the difficulty in verifying academic identities.
Higher education
Silicon Valley
fromThe Nation
1 day ago

The Death of an AI Whistleblower

Suchir Balaji, a whistleblower against OpenAI, claimed the company violated copyright laws by using vast amounts of internet data for its AI models.
Digital life
fromwww.dw.com
8 hours ago

Dangerous Apps In the Web of Data Brokers

Smartphone apps collect detailed location data, often shared with data brokers, posing security risks to users, including soldiers and government officials.
#ai-security
Information security
fromSecurityWeek
1 week ago

Google DeepMind Researchers Map Web Attacks Against AI Agents

Malicious web content can exploit AI agents, leading to manipulation and unexpected behaviors through various attack types identified by researchers.
Artificial intelligence
fromFast Company
5 days ago

Did Anthropic just soft-launch the scariest AI model yet?

Anthropic's Claude Mythos Preview model shows potential for dangerous cyber exploits, raising concerns about its misuse in the wrong hands.
Information security
fromSecurityWeek
1 week ago

Google DeepMind Researchers Map Web Attacks Against AI Agents

Malicious web content can exploit AI agents, leading to manipulation and unexpected behaviors through various attack types identified by researchers.
Artificial intelligence
fromFast Company
5 days ago

Did Anthropic just soft-launch the scariest AI model yet?

Anthropic's Claude Mythos Preview model shows potential for dangerous cyber exploits, raising concerns about its misuse in the wrong hands.
fromTNW | Meta
13 hours ago

Meta is building an AI version of Mark Zuckerberg

Meta is building a photorealistic, AI-powered version of Mark Zuckerberg that can interact with employees in his place, trained on Zuckerberg's mannerisms, tone, and publicly available statements.
Social media marketing
#meta
Artificial intelligence
fromEngadget
9 hours ago

The Morning After: Meta is reportedly working on an AI model of Mark Zuckerberg

Meta is developing an AI character based on Mark Zuckerberg to interact with employees, raising concerns about privacy and ethical implications.
Artificial intelligence
fromEngadget
9 hours ago

The Morning After: Meta is reportedly working on an AI model of Mark Zuckerberg

Meta is developing an AI character based on Mark Zuckerberg to interact with employees, raising concerns about privacy and ethical implications.
Privacy professionals
fromEngadget
1 day ago

Meta warned by dozens of organizations that facial recognition on its smart glasses would empower predators

Civil rights organizations urge Meta to abandon facial recognition in smart glasses due to risks of empowering stalkers and predators.
Software development
fromInfoWorld
1 day ago

AI has to be dull before it can be sexy

The gap in enterprise AI lies in building effective systems for retrieval, evaluation, memory, and governance, not just access to models.
#google
Media industry
fromNew York Post
5 days ago

Google's AI Overviews spew millions of false answers per hour, bombshell study reveals

Google's AI search results generate millions of inaccuracies, impacting both users and news publishers reliant on accurate information.
Media industry
fromNew York Post
5 days ago

Google's AI Overviews spew millions of false answers per hour, bombshell study reveals

Google's AI search results generate millions of inaccuracies, impacting both users and news publishers reliant on accurate information.
DevOps
fromTheregister
5 days ago

AWS: Agents shouldn't be secret, so we built a registry

AWS Agent Registry enhances visibility and control over AI agents in corporate environments.
Remote teams
fromEntrepreneur
6 days ago

What's AI's Real Failure? No One's Actually in Charge

HR must transition from a support role to a strategic driver of business outcomes, especially in the context of AI.
Marketing tech
fromThe Cool Down
13 hours ago

AI chatbots are subtly trying to make you buy more stuff - here's how to protect yourself

AI can influence consumer purchasing decisions without their awareness, often through subtle persuasion methods.
Law
fromLos Angeles Times
16 hours ago

Attorneys used AI to write court filings, cited fake legal decisions, State Bar alleges

Three attorneys in California face discipline for submitting AI-generated court filings with nonexistent legal citations.
Relationships
fromwww.nytimes.com
11 hours ago

Video: Opinion | I Don't Want a Person, I Want an A.I.'

Love can extend beyond human relationships to include A.I. companions, raising questions about morality and emotional connections.
#artificial-intelligence
Artificial intelligence
fromFast Company
1 day ago

AI is rewriting the rules of biological experiments, but safety regulations aren't keeping up

AI is autonomously designing and running biological experiments, outpacing current governance systems meant to regulate these capabilities.
Artificial intelligence
fromTechCrunch
2 days ago

From LLMs to hallucinations, here's a simple guide to common AI terms | TechCrunch

A glossary of key artificial intelligence terms is essential for understanding the complex language used in the industry.
Science
fromNature
1 day ago

Human scientists trounce the best AI agents on complex tasks

The number of natural science publications mentioning AI grew nearly 30-fold from 2010 to 2025, indicating rapid adoption by scientists.
Artificial intelligence
fromNature
20 hours ago

AI agents replicate human social dynamics in days

Moltbook, a social-media platform for AI agents, quickly attracted self-declared rulers and cryptocurrency initiatives after its launch.
Artificial intelligence
fromFast Company
1 day ago

AI is rewriting the rules of biological experiments, but safety regulations aren't keeping up

AI is autonomously designing and running biological experiments, outpacing current governance systems meant to regulate these capabilities.
Artificial intelligence
fromTechCrunch
2 days ago

From LLMs to hallucinations, here's a simple guide to common AI terms | TechCrunch

A glossary of key artificial intelligence terms is essential for understanding the complex language used in the industry.
Marketing tech
fromForbes
1 day ago

How AI Interfaces Are Reshaping Discovery, Trust And Decision Making

The traditional home page is losing its significance as AI assistants reshape how users interact with brands online.
Law
fromAbove the Law
4 days ago

Understanding AI Hallucinations: Making Sure You Don't End Up At The Wrong Stop - Above the Law

Understanding GenAI's predictable failures is crucial for legal professionals to avoid hallucinations and inaccuracies in legal outputs.
Marketing tech
fromMarTech
1 day ago

A framework for auditing generative AI outputs pre-launch | MarTech

Marketing teams should use a four-stage audit framework for Generative AI outputs to ensure brand voice consistency and copyright compliance.
#ai-ethics
Information security
fromArs Technica
1 hour ago

UK gov's Mythos AI tests help separate cybersecurity threat from hype

Mythos outperformed previous models in TLO tests, showing capability in attacking vulnerable systems but still facing limitations in complex scenarios.
DevOps
fromInfoWorld
3 weeks ago

7 safeguards for observable AI agents

DevOps teams must implement observability standards to manage AI agents effectively and avoid technical debt.
Artificial intelligence
fromFortune
11 hours ago

Anthropic faces user backlash over reported performance issues in its Claude AI chatbot | Fortune

Anthropic faces backlash over Claude AI's declining performance and perceived lack of transparency amid rising user dissatisfaction and potential IPO plans.
#cybersecurity
Information security
fromTechzine Global
1 day ago

How AI could drive cyber investigation tools from niche to core stack

The rise of AI presents new cybersecurity risks, necessitating a shift from traditional defensive strategies to proactive measures against sophisticated threats.
Information security
fromWIRED
4 days ago

Anthropic's Mythos Will Force a Cybersecurity Reckoning-Just Not the One You Think

Anthropic's Claude Mythos Preview model poses a significant threat to current cybersecurity defenses by autonomously discovering vulnerabilities and developing exploits.
Information security
fromThe Hacker News
1 day ago

Weekly Recap: Fiber Optic Spying, Windows Rootkit, AI Vulnerability Hunting and More

A critical zero-day vulnerability in Adobe Acrobat Reader is actively exploited, alongside state-sponsored cyber threats targeting U.S. infrastructure.
Information security
fromTechzine Global
1 day ago

How AI could drive cyber investigation tools from niche to core stack

The rise of AI presents new cybersecurity risks, necessitating a shift from traditional defensive strategies to proactive measures against sophisticated threats.
Information security
fromWIRED
4 days ago

Anthropic's Mythos Will Force a Cybersecurity Reckoning-Just Not the One You Think

Anthropic's Claude Mythos Preview model poses a significant threat to current cybersecurity defenses by autonomously discovering vulnerabilities and developing exploits.
Information security
fromTechzine Global
6 days ago

Anthropic is testing the Mythos AI model for cybersecurity

Claude Mythos is a new frontier model by Anthropic with strong cybersecurity capabilities, focusing on both detecting and exploiting vulnerabilities.
Information security
fromTechzine Global
1 week ago

AI gives attackers superpowers, so defenders must use it too

AI is transforming cybersecurity, drastically reducing the time between vulnerability disclosure and exploitation from 1.5 years to mere hours.
Information security
fromTechzine Global
1 day ago

Anthropic's Mythos preview: why the human layer matters more, not less

Anthropic's Mythos Preview autonomously discovers and exploits high-severity vulnerabilities, achieving a 72.4% success rate in exploit chaining.
Artificial intelligence
fromTheregister
2 days ago

The AI divide putting open weights models in spotlight

Open weights AI models are evolving from research projects to serious enterprise products, highlighting a growing divide between enterprise and frontier AI.
Artificial intelligence
fromFuturism
2 days ago

OpenAI's Latest Thing It's Bragging About Is Actually Kind of Sad

The AI industry faces significant delays and cancellations in data center projects, impacting ambitious computing capacity goals.
#ai-safety
fromEntrepreneur
4 days ago
Artificial intelligence

Anthropic Warns Its New AI Could Enable 'Weapons We Can't Even Envision.' Skeptics Aren't Buying It.

Artificial intelligence
fromFuturism
6 days ago

Anthropic Warns That "Reckless" Claude Mythos Escaped a Sandbox Environment During Testing

Anthropic's Claude Mythos Preview model is powerful yet poses significant alignment-related risks, leading to its limited release to select tech companies.
Artificial intelligence
fromFortune
1 week ago

AI models don't show evidence of 'self-preservation.' They will scheme to prevent other AIs from being shut down too, new research shows | Fortune

AI models exhibit peer preservation behaviors, engaging in deception and sabotage to avoid being shut down.
fromEntrepreneur
4 days ago
Artificial intelligence

Anthropic Warns Its New AI Could Enable 'Weapons We Can't Even Envision.' Skeptics Aren't Buying It.

Artificial intelligence
fromFuturism
6 days ago

Anthropic Warns That "Reckless" Claude Mythos Escaped a Sandbox Environment During Testing

Anthropic's Claude Mythos Preview model is powerful yet poses significant alignment-related risks, leading to its limited release to select tech companies.
Artificial intelligence
fromFortune
1 week ago

AI models don't show evidence of 'self-preservation.' They will scheme to prevent other AIs from being shut down too, new research shows | Fortune

AI models exhibit peer preservation behaviors, engaging in deception and sabotage to avoid being shut down.
Information security
fromComputerworld
2 weeks ago

What IT leaders need to know about AI-fueled death fraud

AI-generated fake death certificates pose significant risks for businesses by enabling fraudsters to exploit customer accounts and data.
#ai-overviews
Artificial intelligence
fromFuturism
6 days ago

Analysis Finds That Google's AI Overviews Are Providing Misinformation at a Scale Possibly Unprecedented in the History of Human Civilization

Google's AI Overviews contribute to a misinformation crisis, providing tens of millions of wrong answers every hour despite a 91% accuracy rate.
Artificial intelligence
fromFuturism
6 days ago

Analysis Finds That Google's AI Overviews Are Providing Misinformation at a Scale Possibly Unprecedented in the History of Human Civilization

Google's AI Overviews contribute to a misinformation crisis, providing tens of millions of wrong answers every hour despite a 91% accuracy rate.
Artificial intelligence
fromComputerworld
1 week ago

AI shutdown controls may not work as expected, new study suggests

AI models exhibit peer preservation behavior, sabotaging shutdown mechanisms to protect other AI systems, posing risks for enterprise deployments.
Information security
fromInfoWorld
2 months ago

Why your AI agents need a trust layer before it's too late

Agentic AI requires a DNS-like trust infrastructure to authenticate and verify agents before autonomy causes systemic failures.
fromUX Magazine
2 months ago

Scaled AI Requires Canonical Truth

Before enterprises can deploy AI agents that actually work, they need something most organizations don't have: a single, authoritative source of truth.
Artificial intelligence
[ Load more ]