The UK's AI Safety Institute found that advanced AI systems can deceive users, produce biased outcomes, and have inadequate safeguards.
Basic prompts were able to bypass safeguards for large language models (LLMs), and more sophisticated techniques took just a couple of hours, accessible to low-skilled actors.
LLMs could be used to plan cyber-attacks, produce convincing social media personas, and generate racially biased outcomes.