AI Safety Institute releases new AI safety evaluations platform
Global AI safety evaluations enhanced with UK-built testing platform Inspect.
Balancing innovation and safety: Inside NIST's AI Safety Institute
The AI Safety Institute led by Elizabeth Kelly aims to establish guidelines for evaluating and testing AI models for safety and risk, without hindering generative AI adoption.
Seoul summit showcases UK's progress on trying to make advanced AI safe
The UK is spearheading international collaboration to ensure safe advanced AI models before the Paris summit in six months.
UK's AI Safety Institute needs to set standards rather than do testing'
The UK should focus on setting global standards for AI testing instead of carrying out all vetting itself.
The AI Safety Institute should be a world leader in setting test standards for AI models.
AI Safety Institute releases new AI safety evaluations platform
Global AI safety evaluations enhanced with UK-built testing platform Inspect.
Balancing innovation and safety: Inside NIST's AI Safety Institute
The AI Safety Institute led by Elizabeth Kelly aims to establish guidelines for evaluating and testing AI models for safety and risk, without hindering generative AI adoption.
Seoul summit showcases UK's progress on trying to make advanced AI safe
The UK is spearheading international collaboration to ensure safe advanced AI models before the Paris summit in six months.
UK's AI Safety Institute needs to set standards rather than do testing'
The UK should focus on setting global standards for AI testing instead of carrying out all vetting itself.
The AI Safety Institute should be a world leader in setting test standards for AI models.
UK opens office in San Francisco to tackle AI risk | TechCrunch
Establishing a second location in San Francisco, the AI Safety Institute aims to get closer to the epicenter of AI development for collaborative initiatives and talent access.
Stanford's Rob Reich to Serve as Senior Advisor to the U.S. AI Safety Institute (AISI) - Stanford Impact Labs
The U.S. AI Safety Institute (AISI) under the Department of Commerce is an important governmental initiative to address AI safety, following similar efforts in the EU and the UK.
NIST's new AI safety institute to focus on synthetic content, international outreach
The U.S. AI Safety Institute focuses on standardizing AI tools, monitoring synthetically generated content, and international collaboration.
Three pillars guide the USAISI's work: testing and evaluation, safety and security protocols, and developing guidance for AI-generated content.
Stanford's Rob Reich to Serve as Senior Advisor to the U.S. AI Safety Institute (AISI) - Stanford Impact Labs
The U.S. AI Safety Institute (AISI) under the Department of Commerce is an important governmental initiative to address AI safety, following similar efforts in the EU and the UK.
NIST's new AI safety institute to focus on synthetic content, international outreach
The U.S. AI Safety Institute focuses on standardizing AI tools, monitoring synthetically generated content, and international collaboration.
Three pillars guide the USAISI's work: testing and evaluation, safety and security protocols, and developing guidance for AI-generated content.
NIST adds 5 new members to its AI Safety Institute
Leadership team at U.S. AI Safety Institute strengthened with five new members to focus on AI safety efforts.
NIST Establishes AI Safety Consortium
The National Institute of Standards and Technology has established the AI Safety Institute to develop guidelines and standards for AI measurement and policy.
The U.S. AI Safety Institute is a consortium of AI creators and users, academics, government and industry researchers, and civil society organizations.
The Biden administration names a director of the new AI Safety Institute
Elizabeth Kelly has been named the director of the AI Safety Institute at the National Institute for Standards and Technology.
The institute will create red team testing standards for major AI developers to ensure the safety of AI systems.