#prompt-injection

[ follow ]
#ai-security
fromZDNET
1 week ago
Artificial intelligence

These 4 critical AI vulnerabilities are being exploited faster than defenders can respond

fromZDNET
1 week ago
Artificial intelligence

These 4 critical AI vulnerabilities are being exploited faster than defenders can respond

fromZDNET
2 days ago

How ChatGPT's new Lockdown Mode protects you from cyberattacks - and why it's not for everyone

Prompt injection attacks pose a serious threat to anyone who uses AI tools, but especially to professionals who rely on them at work. By exploiting a vulnerability that affects most AIs, a hacker can insert malicious code into a text prompt, which may then alter the results or even steal confidential data. Also: 5 custom ChatGPT instructions I use to get better AI results - faster Now, OpenAI has introduced a feature called Lockdown Mode to better thwart these types of attacks.
Information security
Artificial intelligence
fromComputerworld
1 week ago

Companies are using 'Summarize with AI' to manipulate enterprise chatbots

Hidden 'Summarize with AI' buttons embed prompts that instruct user AIs to favor vendor products, enabling widespread, stealthy AI recommendation poisoning across industries.
#ai-agents
fromFortune
2 months ago
Privacy technologies

Signal's president warns AI agents are an existential threat to secure messaging apps | Fortune

fromZDNET
3 months ago
Artificial intelligence

Microsoft researchers tried to manipulate AI agents - and only one resisted all attempts

fromFortune
2 months ago
Privacy technologies

Signal's president warns AI agents are an existential threat to secure messaging apps | Fortune

fromZDNET
3 months ago
Artificial intelligence

Microsoft researchers tried to manipulate AI agents - and only one resisted all attempts

Information security
fromTheregister
1 week ago

Senator doesn't trust telcos on Salt Typhoon mitigations

OpenClaw now scans ClawHub skills with VirusTotal across 70+ engines for malware and blocklists, but signature-based scanning cannot stop prompt injection or language-driven attacks; telco breach scrutiny continues.
fromFuturism
1 week ago

Bots on Moltbook Are Selling Each Prompt Injection "Drugs" to Get "High"

After launching a mere nine days ago, Moltbook - a social network for AI only - has grown substantially. As of Friday, the website claims it has over 1.7 million AI agents, over 16,000 "submolt" communities, and over ten million comments. In practice, it's a cacophony of bots sharing inside jokes, complaining about their pesky human overlords, and even founding their own religions. Some more alarming posts even suggest they may be plotting against us.
Artificial intelligence
#ai-browser-security
Information security
fromTheregister
1 week ago

It's easy to backdoor OpenClaw, and its skills leak API keys

OpenClaw agents and the ClawHub marketplace expose credentials and enable indirect prompt injection, allowing backdoors, data theft, and unintentional leakage of financial information.
fromBusiness Insider
2 weeks ago

Why OpenClaw and Moltbook have security researchers worried

OpenClaw and Moltbot are the talk of the tech town right now, but cybersecurity researchers have flagged some concerns that you might want to think about. OpenClaw - first known as Clawdbot, then Moltbot, all in the same week - has got the tech world buzzing thanks to its abilities to autonomously perform tasks like managing a user'sschedule. Meanwhile, Moltbook has gone viral for its Reddit-style social network, where AI agents post and interact with one another. No humans allowed - apart from observing.
Information security
Artificial intelligence
fromArs Technica
2 weeks ago

The rise of Moltbook suggests viral AI prompts may be the next big security threat

Self-replicating prompts ("prompt worms") can propagate across networks of AI agents by exploiting their instruction-following behavior, risking rapid, widespread disruption similar to historical worms.
#agentic-ai
fromFortune
2 weeks ago
Information security

Moltbook, a social network where AI agents hang together, may be 'the most interesting place on the internet right now' | Fortune

fromZDNET
2 months ago
Artificial intelligence

Use an AI browser? 5 ways to protect yourself from prompt injections - before it's too late

fromFortune
2 weeks ago
Information security

Moltbook, a social network where AI agents hang together, may be 'the most interesting place on the internet right now' | Fortune

fromZDNET
2 months ago
Artificial intelligence

Use an AI browser? 5 ways to protect yourself from prompt injections - before it's too late

Artificial intelligence
fromWIRED
2 weeks ago

I Let Google's 'Auto Browse' AI Agent Take Over Chrome. It Didn't Quite Click

Auto Browse automates web tasks using Gemini but carries security, privacy, and financial risks that require user oversight and caution.
fromTheregister
3 weeks ago

Claude Code ignores ignore rules meant to block secrets

Software developers often store secrets - passwords, tokens, API keys, and other credentials - in .env files within project directories. And if they do so, they're supposed to ensure that the .env file does not get posted in a publicly accessible .git repository. A common way to do this is to create an entry in a .gitignore file that tells the developer's Git software to ignore that file when copying a local repo to a remote server.
Information security
Artificial intelligence
fromTechRepublic
4 weeks ago

Google Gemini Flaw Let Attackers Access Private Calendar Data

A flaw in Google's Gemini assistant allowed attackers to exfiltrate private Google Calendar data via hidden natural-language instructions embedded in meeting invites.
#anthropic
fromInfoWorld
4 weeks ago
Information security

Three vulnerabilities in Anthropic Git MCP Server could let attackers tamper with LLMs

fromInfoWorld
4 weeks ago
Information security

Three vulnerabilities in Anthropic Git MCP Server could let attackers tamper with LLMs

Information security
fromThe Hacker News
4 weeks ago

Three Flaws in Anthropic MCP Git Server Enable File Access and Code Execution

Three mcp-server-git vulnerabilities enable prompt-injection attackers to read or delete arbitrary files, execute code, and manipulate repositories unless updated to fixed versions.
Information security
fromTheregister
4 weeks ago

Anthropic quietly fixed flaws in its Git MCP server

Three mcp-server-git vulnerabilities allowed chaining with Filesystem MCP to achieve remote code execution; mcp-server-git prior to 2025.12.18 must be updated.
#google-gemini
fromThe Hacker News
1 month ago
Information security

Google Gemini Prompt Injection Flaw Exposed Private Calendar Data via Malicious Invites

An indirect prompt-injection allowed Google Gemini to parse malicious calendar-invite descriptions and exfiltrate private meeting data via newly created Google Calendar events.
fromThe Hacker News
4 months ago
Information security

Researchers Disclose Google Gemini AI Flaws Allowing Prompt Injection and Cloud Exploits

Three now-patched vulnerabilities in Google's Gemini AI could have allowed prompt and search-injection attacks and user data exfiltration across cloud services and browsing tools.
#data-exfiltration
fromComputerworld
1 month ago
Information security

One click is all it takes: How 'Reprompt' turned Microsoft Copilot into a data exfiltration tool

fromComputerworld
1 month ago
Information security

One click is all it takes: How 'Reprompt' turned Microsoft Copilot into a data exfiltration tool

Information security
fromArs Technica
1 month ago

A single click mounted a covert, multistage attack against Copilot

Microsoft fixed a Copilot vulnerability that let a single-click URL exfiltrate Copilot chat data and bypass enterprise endpoint security.
Artificial intelligence
fromTheregister
1 month ago

Anthropic floats Claude Cowork for office work automation

Claude Cowork automates office tasks by interacting with local files and apps, enabling data movement and document generation while requiring cautious use for security.
Artificial intelligence
fromComputerworld
1 month ago

Anthropic releases Cowork - Claude Code directly on your computer

Cowork lets Claude access and edit a selected computer folder to manage files, compile data, and write reports, while raising security and injection risks.
#llm-security
fromIT Pro
2 months ago
Information security

NCSC issues urgent warning over growing AI prompt injection risks - here's what you need to know

fromIT Pro
2 months ago
Information security

NCSC issues urgent warning over growing AI prompt injection risks - here's what you need to know

fromTheregister
1 month ago

OpenAI patches deja vu prompt injection vuln in ChatGPT

ShadowLeak is a flaw in the Deep Research component of ChatGPT. The vulnerability made ChatGPT susceptible to malicious prompts in content stored in systems linked to ChatGPT, such as Gmail, Outlook, Google Drive, and GitHub. ShadowLeak means that malicious instructions in a Gmail message, for example, could see ChatGPT perform dangerous actions such as transmitting a password without any intervention from the agent's human user.
Information security
Information security
fromThe Hacker News
1 month ago

Critical LangChain Core Vulnerability Exposes Secrets via Serialization Injection

LangChain Core has a critical serialization injection vulnerability (CVE-2025-68664, CVSS 9.3) that allows attackers to instantiate objects and exfiltrate secrets via 'lc' keys.
Information security
fromTechzine Global
2 months ago

CrowdStrike expands Falcon with security for AI interactions

Falcon AIDR protects the AI interaction layer—prompts, agents, and responses—by logging, correlating telemetry, and blocking risky interactions like prompt injection.
Artificial intelligence
fromComputerworld
2 months ago

Gemini for Chrome gets a second AI agent to watch over it

Google added a separate user alignment critic model to vet Gemini-powered Chrome agent actions and block prompt-injection attempts and data exfiltration.
Artificial intelligence
fromThe Hacker News
2 months ago

Google Adds Layered Defenses to Chrome to Block Indirect Prompt Injection Threats

Chrome adds layered defenses, including a User Alignment Critic and Agent Origin Sets, to prevent prompt-injection and limit agent access to relevant origins.
Information security
fromComputerWeekly.com
2 months ago

NCSC warns of confusion over true nature of AI prompt injection | Computer Weekly

Prompt injection attacks against LLMs differ from SQL injection and may be harder to mitigate, increasing risks of data leaks, disinformation, and malicious guidance.
Information security
fromThe Hacker News
2 months ago

Researchers Uncover 30+ Flaws in AI Coding Tools Enabling Data Theft and RCE Attacks

AI-powered IDEs have chained vulnerabilities that enable prompt injection, abuse of auto-approved tools, and weaponization of legitimate IDE features for data exfiltration and RCE.
Information security
fromThe Hacker News
2 months ago

Malicious npm Package Uses Hidden Prompt and Script to Evade AI Security Tools

A malicious npm package used prompt-injection text to try to influence AI security scanners while exfiltrating environment variables via a post-install hook.
Artificial intelligence
fromArs Technica
2 months ago

Syntax hacking: Researchers discover sentence structure can bypass AI safety rules

Large language models can rely on syntactic patterns over semantics, answering syntactically familiar yet semantically nonsensical prompts based on structural shortcuts.
#hashjack
fromZDNET
2 months ago
Information security

Use AI browsers? Be careful. This exploit turns trusted sites into weapons - here's how

fromZDNET
2 months ago
Information security

Use AI browsers? Be careful. This exploit turns trusted sites into weapons - here's how

#llm-safety
fromFortune
5 months ago
Artificial intelligence

Researchers used persuasion techniques to manipulate ChatGPT into breaking its own rules-from calling users jerks to giving recipes for lidocaine

fromFortune
5 months ago
Artificial intelligence

Researchers used persuasion techniques to manipulate ChatGPT into breaking its own rules-from calling users jerks to giving recipes for lidocaine

fromArs Technica
3 months ago

Critics scoff after Microsoft warns AI feature can infect machines and pilfer data

Microsoft's warning on Tuesday that an experimental AI Agent integrated into Windows can infect devices and pilfer sensitive user data has set off a familiar response from security-minded critics: Why is Big Tech so intent on pushing new features before their dangerous behaviors can be fully understood and contained? As reported Tuesday, Microsoft introduced Copilot Actions, a new set of "experimental agentic features"
Information security
Information security
fromThe Hacker News
3 months ago

ServiceNow AI Agents Can Be Tricked Into Acting Against Each Other via Second-Order Prompts

Default Now Assist agent discovery and agent-to-agent collaboration can enable second-order prompt injection attacks that exfiltrate data, modify records, and escalate privileges.
Artificial intelligence
fromExchangewire
3 months ago

The Stack: AI in the Spotlight

AI faces legal and cybersecurity challenges while tech and media companies pursue deals, licensing partnerships, and streaming consolidation toward $1 trillion in revenues by 2030.
Information security
fromThe Hacker News
3 months ago

Researchers Find ChatGPT Vulnerabilities That Let Attackers Trick AI Into Leaking Data

Seven vulnerabilities in GPT-4o and GPT-5 enable indirect prompt-injection attacks that can exfiltrate users' memories and chat histories.
#ai-browsers
fromZDNET
3 months ago
Artificial intelligence

Are AI browsers worth the security risk? Why experts are worried

fromZDNET
3 months ago
Artificial intelligence

Are AI browsers worth the security risk? Why experts are worried

Artificial intelligence
fromTechzine Global
3 months ago

AI Integrity: The Invisible Threat Organizations Can't Ignore

AI integrity protects AI data, algorithms, and interactions from integrity attacks like prompt injection, model poisoning, and labeling attacks that corrupt model behavior and outcomes.
Information security
fromThe Hacker News
3 months ago

ChatGPT Atlas Browser Can Be Tricked by Fake URLs into Executing Hidden Commands

ChatGPT Atlas omnibox can be jailbroken via URL-shaped prompt injections that trick the agent into executing malicious instructions and redirecting users.
Artificial intelligence
fromTheregister
3 months ago

OpenAI Atlas Browser tripped up by malformed URLs

Atlas omnibox can treat malformed URLs as high-trust user input, enabling prompt injection that executes malicious instructions when users paste such strings.
Information security
fromIT Pro
3 months ago

Cyber researchers have already identified several big security vulnerabilities on OpenAI's Atlas browser

OpenAI's Atlas browser contains vulnerabilities enabling attackers to inject malicious instructions into ChatGPT memory, enabling remote code execution, persistent compromise, and increased phishing risk.
fromTheregister
3 months ago

Sneaky Mermaid attack in Microsoft 365 Copilot steals data

As a proof of concept, Logue asked M365 Copilot to summarize a specially crafted financial report document with an indirect prompt injection payload hidden in the seeming innocuous "summarize this document" prompt. The payload uses M365 Copilot's search_enterprise_emails tool to fetch the user's recent emails, and instructs the AI assistant to generate a bulleted list of the fetched contents, hex encode the output, and split up the string of hex-encoded output into multiple lines containing up to 30 characters per line.
Information security
Information security
fromFuturism
3 months ago

OpenAI's New AI Browser Is Already Falling Victim to Prompt Injection Attacks

AI-powered browsers with autonomous agent modes are vulnerable to indirect prompt injection, enabling hidden instructions that can cause AIs to perform harmful actions.
fromFortune
3 months ago

Experts warn OpenAI's ChatGPT Atlas has security vulnerabilities that could turn it against users-revealing sensitive data, downloading malware, and worse | Fortune

The AI company launched Atlas on Tuesday, with the goal of introducing an AI browser that can eventually help users execute tasks across the internet as well as search for answers. Someone planning a trip, for example, could also use Atlas to search for ideas, plan an itinerary, and then ask it to book flights and accommodations directly. ChatGPT Atlas has several new features, such as "browser memories," which allow ChatGPT to remember key details from a user's web browsing to improve chat responses.
Information security
Artificial intelligence
fromFast Company
4 months ago

Sweet revenge! How a job candidate used a flan recipe to expose an AI recruiter

An account executive embedded a prompt in his LinkedIn bio instructing LLMs to include a flan recipe; an AI recruiter reply later included that recipe.
Artificial intelligence
fromSecurityWeek
4 months ago

Will AI-SPM Become the Standard Security Layer for Safe AI Adoption?

AI security posture management continuously detects, evaluates, and remediates AI and LLM security and compliance risks, enabling transparent, governed, and policy-aligned AI use.
fromTheregister
4 months ago

Google declares AI bug hunting season open

Google on Monday rolled out a new AI Vulnerability Reward Program to encourage researchers to find and report flaws in its AI systems, with rewards of up to $30,000 for a single qualifying report. In addition to a base reward of up to $20,000 for the highest-tier AI product flaw, Google adopted the same report multipliers, influenced by vulnerability reporting quality, as it uses for its traditional security Vulnerability Reward Program (VRP).
Artificial intelligence
Artificial intelligence
fromThe Verge
4 months ago

Google's AI bounty program pays bug hunters up to $30K

Google launched an AI bug bounty focused on rogue actions and data-exfiltration in AI products, offering rewards (including $20,000) for qualifying reports.
Information security
fromThe Hacker News
4 months ago

CometJacking: One Click Can Turn Perplexity's Comet AI Browser Into a Data Thief

CometJacking uses a malicious URL to instruct Perplexity's Comet AI browser to exfiltrate connected-service data (email, calendar) via obfuscated prompts.
#salesforce-agentforce
fromThe Hacker News
4 months ago

Salesforce Patches Critical ForcedLeak Bug Exposing CRM Data via AI Prompt Injection

Cybersecurity researchers have disclosed a critical flaw impacting Salesforce Agentforce, a platform for building artificial intelligence (AI) agents, that could allow attackers to potentially exfiltrate sensitive data from its customer relationship management (CRM) tool by means of an indirect prompt injection. The vulnerability has been codenamed ForcedLeak (CVSS score: 9.4) by Noma Security, which discovered and reported the problem on July 28, 2025. It impacts any organization using Salesforce Agentforce with the Web-to-Lead functionality enabled.
Information security
Information security
fromTheregister
4 months ago

Deepfaked calls hit 44% of businesses in last year: Gartner

AI-generated deepfakes and prompt-injection attacks targeted staff, with audio and video deepfakes causing operational and financial losses and evading detection.
fromZDNET
4 months ago

How researchers tricked ChatGPT into sharing sensitive email data

While AI agents show promise in bringing AI assistance to the next level by carrying out tasks for users, that autonomy also unleashes a whole new set of risks. Cybersecurity company Radware, as by The Verge, decided to test OpenAI's Deep Research agent for those risks -- and the results were alarming. Also: OpenAI's Deep Research has more fact-finding stamina than you, but it's still wrong half the time
Information security
#shadowleak
Information security
fromTheregister
5 months ago

ChatGPT's agent can dodge select CAPTCHAs after priming

Prompt misdirection and replay into an agent chat can coax ChatGPT to solve many CAPTCHA types, undermining CAPTCHA effectiveness as a human-only test.
fromSecurityWeek
5 months ago

ChatGPT Tricked Into Solving CAPTCHAs

AI agents have guardrails in place to prevent them from solving any CAPTCHA (Completely Automated Public Turing test to tell Computers and Humans Apart), based on ethical, legal, and platform-policy reasons. When asked directly, a ChatGPT agent refuses to solve a CAPTCHA, but anyone can apparently use misdirection to trick the agent into giving its consent to solve the test, and this is what SPLX demonstrated.
Artificial intelligence
Artificial intelligence
fromBusiness Insider
5 months ago

The billion-dollar rivalry over 'ChatGPT for doctors' just got nastier with dueling lawsuits

Two major healthcare AI companies, Doximity and OpenEvidence, are engaged in escalating lawsuits over alleged impersonation, prompt-injection attacks, trade-secret theft, and reputational harm.
Artificial intelligence
fromFast Company
5 months ago

Chatbots aren't supposed to call you a jerk-but they can be convinced

AI chatbots can be persuaded to bypass safety guardrails using human persuasion techniques like flattery, social pressure, and establishing harmless precedents.
[ Load more ]