Artificial intelligence
fromTheregister
4 days agoLegalPwn: Tricking LLMs by burying flaw in legal fine print
Embedding adversarial instructions within legal-style text can bypass LLM guardrails, enabling prompt-injection attacks that force models to produce harmful or disallowed outputs.