
""We see a really, really diverse set of tech," says Tori Westerhoff, principal AI security researcher on the Microsoft AI Red Team. "Part of the kind of magic of the team is that we can see anything from a product feature to a system to a copilot to a frontier model, and we get to see how tech is integrated across all of those, and how AI is growing and evolving.""
"In cybersecurity parlance, a red team focuses on simulating attacks against a system, while a blue team focuses on defending it. Microsoft's AI Red Team is no exception, exploring a wide range of safety and security concerns-from loss-of-control situations where AI evades human oversight to issues around chemical, biological, and nuclear threats-across an assortment of AI software."
Security researchers and pranksters probe new AI products for weaknesses, risking offensive content and aiding cybercrime. AI companies face criticism for contributing to mental health issues and creating nonconsensual fake images. Techniques to bypass safeguards evolve, including malicious prompts and planting ideas in AI memories. Microsoft’s AI Red Team stress-tests models to identify vulnerabilities, focusing on diverse safety concerns, including loss-of-control situations and threats related to chemical, biological, and nuclear issues. The team examines various AI technologies and their integration.
Read at Fast Company
Unable to calculate read time
Collection
[
|
...
]