#prompt-injection

[ follow ]
#ai-security
Artificial intelligence
fromInfoQ
2 months ago

Google Gemini's Long-term Memory Vulnerable to a Kind of Phishing Attack

Johann Rehberger demonstrated a prompt injection attack on Google Gemini, exploiting delayed tool invocation to modify its long-term memories.
Artificial intelligence
fromFuturism
3 weeks ago

Researchers Find Easy Way to Jailbreak Every Major AI, From ChatGPT to Claude

A newly discovered jailbreak can manipulate AI models into producing harmful content, exposing vulnerabilities in their safety measures.
Artificial intelligence
fromInfoQ
3 weeks ago

DeepMind Researchers Propose Defense Against LLM Prompt Injection

Google DeepMind's CaMeL effectively neutralizes 67% of prompt injection attacks in LLMs using traditional software security principles.
Growth hacking
fromArs Technica
1 month ago

Gemini hackers can deliver more potent attacks with a helping hand from... Gemini

Indirect prompt injections are an effective method for exploiting large language models, revealing vulnerabilities in AI systems.
fromInfoQ
1 week ago
Artificial intelligence

Meta Open Sources LlamaFirewall for AI Agent Combined Protection

Artificial intelligence
fromInfoQ
2 months ago

Google Gemini's Long-term Memory Vulnerable to a Kind of Phishing Attack

Johann Rehberger demonstrated a prompt injection attack on Google Gemini, exploiting delayed tool invocation to modify its long-term memories.
Artificial intelligence
fromFuturism
3 weeks ago

Researchers Find Easy Way to Jailbreak Every Major AI, From ChatGPT to Claude

A newly discovered jailbreak can manipulate AI models into producing harmful content, exposing vulnerabilities in their safety measures.
Artificial intelligence
fromInfoQ
3 weeks ago

DeepMind Researchers Propose Defense Against LLM Prompt Injection

Google DeepMind's CaMeL effectively neutralizes 67% of prompt injection attacks in LLMs using traditional software security principles.
Growth hacking
fromArs Technica
1 month ago

Gemini hackers can deliver more potent attacks with a helping hand from... Gemini

Indirect prompt injections are an effective method for exploiting large language models, revealing vulnerabilities in AI systems.
fromInfoQ
1 week ago
Artificial intelligence

Meta Open Sources LlamaFirewall for AI Agent Combined Protection

Artificial intelligence
fromHackernoon
1 year ago

Prompt Injection Is What Happens When AI Trusts Too Easily | HackerNoon

Generative AI is becoming essential in daily life, but it poses significant security threats like prompt injection, which can manipulate AI systems.
Artificial intelligence
fromArs Technica
1 month ago

Researchers claim breakthrough in fight against AI's frustrating security hole

Prompt injections jeopardize AI systems; Google DeepMind's CaMeL offers a potential solution by treating language models as untrusted components within security frameworks.
[ Load more ]