#prompt-injection

[ follow ]
fromTheregister
1 day ago

Claude Code ignores ignore rules meant to block secrets

Software developers often store secrets - passwords, tokens, API keys, and other credentials - in .env files within project directories. And if they do so, they're supposed to ensure that the .env file does not get posted in a publicly accessible .git repository. A common way to do this is to create an entry in a .gitignore file that tells the developer's Git software to ignore that file when copying a local repo to a remote server.
Information security
Artificial intelligence
fromTechRepublic
1 week ago

Google Gemini Flaw Let Attackers Access Private Calendar Data

A flaw in Google's Gemini assistant allowed attackers to exfiltrate private Google Calendar data via hidden natural-language instructions embedded in meeting invites.
#anthropic
fromInfoWorld
1 week ago
Information security

Three vulnerabilities in Anthropic Git MCP Server could let attackers tamper with LLMs

fromInfoWorld
1 week ago
Information security

Three vulnerabilities in Anthropic Git MCP Server could let attackers tamper with LLMs

Information security
fromThe Hacker News
1 week ago

Three Flaws in Anthropic MCP Git Server Enable File Access and Code Execution

Three mcp-server-git vulnerabilities enable prompt-injection attackers to read or delete arbitrary files, execute code, and manipulate repositories unless updated to fixed versions.
Information security
fromTheregister
1 week ago

Anthropic quietly fixed flaws in its Git MCP server

Three mcp-server-git vulnerabilities allowed chaining with Filesystem MCP to achieve remote code execution; mcp-server-git prior to 2025.12.18 must be updated.
#google-gemini
fromThe Hacker News
1 week ago
Information security

Google Gemini Prompt Injection Flaw Exposed Private Calendar Data via Malicious Invites

An indirect prompt-injection allowed Google Gemini to parse malicious calendar-invite descriptions and exfiltrate private meeting data via newly created Google Calendar events.
fromThe Hacker News
4 months ago
Information security

Researchers Disclose Google Gemini AI Flaws Allowing Prompt Injection and Cloud Exploits

Three now-patched vulnerabilities in Google's Gemini AI could have allowed prompt and search-injection attacks and user data exfiltration across cloud services and browsing tools.
#data-exfiltration
fromComputerworld
2 weeks ago
Information security

One click is all it takes: How 'Reprompt' turned Microsoft Copilot into a data exfiltration tool

fromComputerworld
2 weeks ago
Information security

One click is all it takes: How 'Reprompt' turned Microsoft Copilot into a data exfiltration tool

#ai-security
fromZDNET
1 month ago
Artificial intelligence

How OpenAI is defending ChatGPT Atlas from attacks now - and why safety's not guaranteed

fromFortune
1 month ago
Information security

OpenAI says AI browsers like ChatGPT Atlas may never be fully secure from hackers-and experts say the risks are 'a feature not a bug' | Fortune

fromZDNET
1 month ago
Artificial intelligence

How OpenAI is defending ChatGPT Atlas from attacks now - and why safety's not guaranteed

fromFortune
1 month ago
Information security

OpenAI says AI browsers like ChatGPT Atlas may never be fully secure from hackers-and experts say the risks are 'a feature not a bug' | Fortune

Information security
fromArs Technica
2 weeks ago

A single click mounted a covert, multistage attack against Copilot

Microsoft fixed a Copilot vulnerability that let a single-click URL exfiltrate Copilot chat data and bypass enterprise endpoint security.
Artificial intelligence
fromTheregister
2 weeks ago

Anthropic floats Claude Cowork for office work automation

Claude Cowork automates office tasks by interacting with local files and apps, enabling data movement and document generation while requiring cautious use for security.
Artificial intelligence
fromComputerworld
2 weeks ago

Anthropic releases Cowork - Claude Code directly on your computer

Cowork lets Claude access and edit a selected computer folder to manage files, compile data, and write reports, while raising security and injection risks.
#llm-security
fromIT Pro
1 month ago
Information security

NCSC issues urgent warning over growing AI prompt injection risks - here's what you need to know

fromIT Pro
1 month ago
Information security

NCSC issues urgent warning over growing AI prompt injection risks - here's what you need to know

fromTheregister
3 weeks ago

OpenAI patches deja vu prompt injection vuln in ChatGPT

ShadowLeak is a flaw in the Deep Research component of ChatGPT. The vulnerability made ChatGPT susceptible to malicious prompts in content stored in systems linked to ChatGPT, such as Gmail, Outlook, Google Drive, and GitHub. ShadowLeak means that malicious instructions in a Gmail message, for example, could see ChatGPT perform dangerous actions such as transmitting a password without any intervention from the agent's human user.
Information security
Information security
fromThe Hacker News
1 month ago

Critical LangChain Core Vulnerability Exposes Secrets via Serialization Injection

LangChain Core has a critical serialization injection vulnerability (CVE-2025-68664, CVSS 9.3) that allows attackers to instantiate objects and exfiltrate secrets via 'lc' keys.
Information security
fromTechzine Global
1 month ago

CrowdStrike expands Falcon with security for AI interactions

Falcon AIDR protects the AI interaction layer—prompts, agents, and responses—by logging, correlating telemetry, and blocking risky interactions like prompt injection.
Artificial intelligence
fromComputerworld
1 month ago

Gemini for Chrome gets a second AI agent to watch over it

Google added a separate user alignment critic model to vet Gemini-powered Chrome agent actions and block prompt-injection attempts and data exfiltration.
Artificial intelligence
fromThe Hacker News
1 month ago

Google Adds Layered Defenses to Chrome to Block Indirect Prompt Injection Threats

Chrome adds layered defenses, including a User Alignment Critic and Agent Origin Sets, to prevent prompt-injection and limit agent access to relevant origins.
Information security
fromComputerWeekly.com
1 month ago

NCSC warns of confusion over true nature of AI prompt injection | Computer Weekly

Prompt injection attacks against LLMs differ from SQL injection and may be harder to mitigate, increasing risks of data leaks, disinformation, and malicious guidance.
Information security
fromThe Hacker News
1 month ago

Researchers Uncover 30+ Flaws in AI Coding Tools Enabling Data Theft and RCE Attacks

AI-powered IDEs have chained vulnerabilities that enable prompt injection, abuse of auto-approved tools, and weaponization of legitimate IDE features for data exfiltration and RCE.
Information security
fromThe Hacker News
1 month ago

Malicious npm Package Uses Hidden Prompt and Script to Evade AI Security Tools

A malicious npm package used prompt-injection text to try to influence AI security scanners while exfiltrating environment variables via a post-install hook.
Artificial intelligence
fromArs Technica
1 month ago

Syntax hacking: Researchers discover sentence structure can bypass AI safety rules

Large language models can rely on syntactic patterns over semantics, answering syntactically familiar yet semantically nonsensical prompts based on structural shortcuts.
#agentic-ai
fromZDNET
1 month ago
Artificial intelligence

Use an AI browser? 5 ways to protect yourself from prompt injections - before it's too late

fromZDNET
5 months ago
Information security

Perplexity's Comet AI browser could expose your data to attackers - here's how

fromZDNET
1 month ago
Artificial intelligence

Use an AI browser? 5 ways to protect yourself from prompt injections - before it's too late

fromZDNET
5 months ago
Information security

Perplexity's Comet AI browser could expose your data to attackers - here's how

#ai-agents
fromFortune
2 months ago
Privacy technologies

Signal's president warns AI agents are an existential threat to secure messaging apps | Fortune

fromZDNET
2 months ago
Artificial intelligence

Microsoft researchers tried to manipulate AI agents - and only one resisted all attempts

fromFortune
2 months ago
Privacy technologies

Signal's president warns AI agents are an existential threat to secure messaging apps | Fortune

fromZDNET
2 months ago
Artificial intelligence

Microsoft researchers tried to manipulate AI agents - and only one resisted all attempts

#hashjack
fromZDNET
2 months ago
Information security

Use AI browsers? Be careful. This exploit turns trusted sites into weapons - here's how

fromZDNET
2 months ago
Information security

Use AI browsers? Be careful. This exploit turns trusted sites into weapons - here's how

#llm-safety
fromFortune
4 months ago
Artificial intelligence

Researchers used persuasion techniques to manipulate ChatGPT into breaking its own rules-from calling users jerks to giving recipes for lidocaine

fromFortune
4 months ago
Artificial intelligence

Researchers used persuasion techniques to manipulate ChatGPT into breaking its own rules-from calling users jerks to giving recipes for lidocaine

fromArs Technica
2 months ago

Critics scoff after Microsoft warns AI feature can infect machines and pilfer data

Microsoft's warning on Tuesday that an experimental AI Agent integrated into Windows can infect devices and pilfer sensitive user data has set off a familiar response from security-minded critics: Why is Big Tech so intent on pushing new features before their dangerous behaviors can be fully understood and contained? As reported Tuesday, Microsoft introduced Copilot Actions, a new set of "experimental agentic features"
Information security
Information security
fromThe Hacker News
2 months ago

ServiceNow AI Agents Can Be Tricked Into Acting Against Each Other via Second-Order Prompts

Default Now Assist agent discovery and agent-to-agent collaboration can enable second-order prompt injection attacks that exfiltrate data, modify records, and escalate privileges.
Artificial intelligence
fromExchangewire
2 months ago

The Stack: AI in the Spotlight

AI faces legal and cybersecurity challenges while tech and media companies pursue deals, licensing partnerships, and streaming consolidation toward $1 trillion in revenues by 2030.
Information security
fromThe Hacker News
2 months ago

Researchers Find ChatGPT Vulnerabilities That Let Attackers Trick AI Into Leaking Data

Seven vulnerabilities in GPT-4o and GPT-5 enable indirect prompt-injection attacks that can exfiltrate users' memories and chat histories.
#ai-browsers
fromZDNET
3 months ago
Artificial intelligence

Are AI browsers worth the security risk? Why experts are worried

fromZDNET
3 months ago
Artificial intelligence

Are AI browsers worth the security risk? Why experts are worried

Artificial intelligence
fromTechzine Global
2 months ago

AI Integrity: The Invisible Threat Organizations Can't Ignore

AI integrity protects AI data, algorithms, and interactions from integrity attacks like prompt injection, model poisoning, and labeling attacks that corrupt model behavior and outcomes.
Information security
fromThe Hacker News
2 months ago

ChatGPT Atlas Browser Can Be Tricked by Fake URLs into Executing Hidden Commands

ChatGPT Atlas omnibox can be jailbroken via URL-shaped prompt injections that trick the agent into executing malicious instructions and redirecting users.
Artificial intelligence
fromTheregister
3 months ago

OpenAI Atlas Browser tripped up by malformed URLs

Atlas omnibox can treat malformed URLs as high-trust user input, enabling prompt injection that executes malicious instructions when users paste such strings.
Information security
fromIT Pro
3 months ago

Cyber researchers have already identified several big security vulnerabilities on OpenAI's Atlas browser

OpenAI's Atlas browser contains vulnerabilities enabling attackers to inject malicious instructions into ChatGPT memory, enabling remote code execution, persistent compromise, and increased phishing risk.
fromFuturism
3 months ago

Serious New Hack Discovered Against OpenAI's New AI Browser

The browser, which puts OpenAI's blockbuster ChatGPT front and center, features an "agent mode" - currently limited to paying subscribers - that allows it to complete entire tasks, such as booking a flight or purchasing groceries. However, that makes the browser vulnerable to "prompt injection" attacks, allowing hackers to embed hidden messages on the web that force it to carry out harmful instructions. For instance, one researcher tricked the browser into spitting out the words "Trust No AI" instead of generating a summary of a document in Google Docs, as prompted.
Artificial intelligence
fromTheregister
3 months ago

Sneaky Mermaid attack in Microsoft 365 Copilot steals data

As a proof of concept, Logue asked M365 Copilot to summarize a specially crafted financial report document with an indirect prompt injection payload hidden in the seeming innocuous "summarize this document" prompt. The payload uses M365 Copilot's search_enterprise_emails tool to fetch the user's recent emails, and instructs the AI assistant to generate a bulleted list of the fetched contents, hex encode the output, and split up the string of hex-encoded output into multiple lines containing up to 30 characters per line.
Information security
Information security
fromFuturism
3 months ago

OpenAI's New AI Browser Is Already Falling Victim to Prompt Injection Attacks

AI-powered browsers with autonomous agent modes are vulnerable to indirect prompt injection, enabling hidden instructions that can cause AIs to perform harmful actions.
#ai-browser
fromFortune
3 months ago
Information security

Experts warn OpenAI's ChatGPT Atlas has security vulnerabilities that could turn it against users-revealing sensitive data, downloading malware, and worse | Fortune

fromFuturism
5 months ago
Information security

Using an AI Browser Lets Hackers Drain Your Bank Account Just by Showing You a Public Reddit Post

fromFortune
3 months ago
Information security

Experts warn OpenAI's ChatGPT Atlas has security vulnerabilities that could turn it against users-revealing sensitive data, downloading malware, and worse | Fortune

fromFuturism
5 months ago
Information security

Using an AI Browser Lets Hackers Drain Your Bank Account Just by Showing You a Public Reddit Post

Artificial intelligence
fromFast Company
3 months ago

Sweet revenge! How a job candidate used a flan recipe to expose an AI recruiter

An account executive embedded a prompt in his LinkedIn bio instructing LLMs to include a flan recipe; an AI recruiter reply later included that recipe.
Artificial intelligence
fromSecurityWeek
3 months ago

Will AI-SPM Become the Standard Security Layer for Safe AI Adoption?

AI security posture management continuously detects, evaluates, and remediates AI and LLM security and compliance risks, enabling transparent, governed, and policy-aligned AI use.
fromTheregister
3 months ago

Google declares AI bug hunting season open

Google on Monday rolled out a new AI Vulnerability Reward Program to encourage researchers to find and report flaws in its AI systems, with rewards of up to $30,000 for a single qualifying report. In addition to a base reward of up to $20,000 for the highest-tier AI product flaw, Google adopted the same report multipliers, influenced by vulnerability reporting quality, as it uses for its traditional security Vulnerability Reward Program (VRP).
Artificial intelligence
Artificial intelligence
fromThe Verge
3 months ago

Google's AI bounty program pays bug hunters up to $30K

Google launched an AI bug bounty focused on rogue actions and data-exfiltration in AI products, offering rewards (including $20,000) for qualifying reports.
Information security
fromThe Hacker News
3 months ago

CometJacking: One Click Can Turn Perplexity's Comet AI Browser Into a Data Thief

CometJacking uses a malicious URL to instruct Perplexity's Comet AI browser to exfiltrate connected-service data (email, calendar) via obfuscated prompts.
#salesforce-agentforce
fromThe Hacker News
4 months ago

Salesforce Patches Critical ForcedLeak Bug Exposing CRM Data via AI Prompt Injection

Cybersecurity researchers have disclosed a critical flaw impacting Salesforce Agentforce, a platform for building artificial intelligence (AI) agents, that could allow attackers to potentially exfiltrate sensitive data from its customer relationship management (CRM) tool by means of an indirect prompt injection. The vulnerability has been codenamed ForcedLeak (CVSS score: 9.4) by Noma Security, which discovered and reported the problem on July 28, 2025. It impacts any organization using Salesforce Agentforce with the Web-to-Lead functionality enabled.
Information security
Information security
fromTheregister
4 months ago

Deepfaked calls hit 44% of businesses in last year: Gartner

AI-generated deepfakes and prompt-injection attacks targeted staff, with audio and video deepfakes causing operational and financial losses and evading detection.
fromZDNET
4 months ago

How researchers tricked ChatGPT into sharing sensitive email data

While AI agents show promise in bringing AI assistance to the next level by carrying out tasks for users, that autonomy also unleashes a whole new set of risks. Cybersecurity company Radware, as by The Verge, decided to test OpenAI's Deep Research agent for those risks -- and the results were alarming. Also: OpenAI's Deep Research has more fact-finding stamina than you, but it's still wrong half the time
Information security
#shadowleak
Information security
fromTheregister
4 months ago

ChatGPT's agent can dodge select CAPTCHAs after priming

Prompt misdirection and replay into an agent chat can coax ChatGPT to solve many CAPTCHA types, undermining CAPTCHA effectiveness as a human-only test.
fromSecurityWeek
4 months ago

ChatGPT Tricked Into Solving CAPTCHAs

AI agents have guardrails in place to prevent them from solving any CAPTCHA (Completely Automated Public Turing test to tell Computers and Humans Apart), based on ethical, legal, and platform-policy reasons. When asked directly, a ChatGPT agent refuses to solve a CAPTCHA, but anyone can apparently use misdirection to trick the agent into giving its consent to solve the test, and this is what SPLX demonstrated.
Artificial intelligence
Artificial intelligence
fromBusiness Insider
4 months ago

The billion-dollar rivalry over 'ChatGPT for doctors' just got nastier with dueling lawsuits

Two major healthcare AI companies, Doximity and OpenEvidence, are engaged in escalating lawsuits over alleged impersonation, prompt-injection attacks, trade-secret theft, and reputational harm.
Artificial intelligence
fromFast Company
4 months ago

Chatbots aren't supposed to call you a jerk-but they can be convinced

AI chatbots can be persuaded to bypass safety guardrails using human persuasion techniques like flattery, social pressure, and establishing harmless precedents.
Artificial intelligence
fromCSO Online
5 months ago

LLMs easily exploited using run-on sentences, bad grammar, image scaling

Large language models remain easily manipulated into revealing sensitive data via prompt formatting and hidden-image attacks due to alignment training gaps and brittle prompt security.
Artificial intelligence
fromTheregister
5 months ago

Anthropic teases Claude for Chrome with massive warnings

Claude for Chrome gives Max-tier users automated web browsing control while introducing significant browser-extension security, privacy, and prompt-injection risks.
Science
fromSecuritymagazine
5 months ago

Agentic AI Browsers Exploited by "PromptFix" Trick Technique

A new prompt injection technique uses fake CAPTCHA pages to trick generative AI agents into executing malicious actions and visiting lookalike storefronts.
Information security
fromTheregister
5 months ago

AWS patches Q Developer after prompt injection, RCE demo

Amazon fixed prompt-injection and RCE-capable vulnerabilities in the Amazon Q Developer VS Code extension by updating the language server and adding human-in-the-loop approval.
Information security
fromThe Hacker News
5 months ago

Experts Find AI Browsers Can Be Tricked by PromptFix Exploit to Run Malicious Hidden Prompts

PromptFix hides malicious instructions inside fake CAPTCHA checks to trick GenAI browsers and agentic AI into interacting with phishing sites and performing attacker actions.
#cybersecurity
fromHackernoon
6 months ago
Privacy professionals

The Prompt Protocol: Why Tomorrow's Security Nightmares Will Be Whispered, Not Coded | HackerNoon

fromHackernoon
6 months ago
Privacy professionals

The Prompt Protocol: Why Tomorrow's Security Nightmares Will Be Whispered, Not Coded | HackerNoon

[ Load more ]