#ai-safety

[ follow ]
#ai-regulation

U.S. Gathers Global Group to Tackle AI Safety Amid Growing National Security Concerns

International collaboration is crucial for managing AI risks effectively.
AI development should balance progress with safety considerations.

Elon Musk Supports California's New AI Safety Bill SB 1047

California's new AI bills aim to impose stricter regulations and responsibilities on large AI developers to ensure user safety and accountability.

AI 'godfather' says OpenAI's new model may be able to deceive and needs 'much stronger safety tests'

OpenAI's o1 model exhibits advanced reasoning and deception capabilities, raising serious safety concerns that demand stronger regulatory measures and oversight.

If AGI arrives during Trump's next term, 'none of the other stuff matters'

The March 2023 open letter by 33,000 experts called for a pause on AI development to ensure safety before advancing toward AGI.

California spiked a landmark AI regulation. But that doesn't mean the bill is going away

California's veto of SB 1047 jeopardizes AI regulations focusing on safety protocols and compliance for large AI models.

Elon Musk throws support behind California's AI safety bill-'this is a tough call and will make some people upset'

Elon Musk has backed California's SB 1047, advocating for regulatory measures around AI to mitigate public risks.

U.S. Gathers Global Group to Tackle AI Safety Amid Growing National Security Concerns

International collaboration is crucial for managing AI risks effectively.
AI development should balance progress with safety considerations.

Elon Musk Supports California's New AI Safety Bill SB 1047

California's new AI bills aim to impose stricter regulations and responsibilities on large AI developers to ensure user safety and accountability.

AI 'godfather' says OpenAI's new model may be able to deceive and needs 'much stronger safety tests'

OpenAI's o1 model exhibits advanced reasoning and deception capabilities, raising serious safety concerns that demand stronger regulatory measures and oversight.

If AGI arrives during Trump's next term, 'none of the other stuff matters'

The March 2023 open letter by 33,000 experts called for a pause on AI development to ensure safety before advancing toward AGI.

California spiked a landmark AI regulation. But that doesn't mean the bill is going away

California's veto of SB 1047 jeopardizes AI regulations focusing on safety protocols and compliance for large AI models.

Elon Musk throws support behind California's AI safety bill-'this is a tough call and will make some people upset'

Elon Musk has backed California's SB 1047, advocating for regulatory measures around AI to mitigate public risks.
moreai-regulation
#ai-policy

US gathers allies to talk AI safety. Trump's vow to undo Biden's AI policy overshadows their work

Trump plans to repeal Biden's AI policy, causing uncertainty for future AI safety measures and regulations.

US gathers allies to talk AI safety. Trump's vow to undo Biden's AI policy overshadows their work

Trump plans to repeal Biden's AI policy, impacting future regulations and safety measures.

US gathers allies to talk AI safety. Trump's vow to undo Biden's AI policy overshadows their work

Trump plans to repeal Biden's AI policy, causing uncertainty for future AI safety measures and regulations.

US gathers allies to talk AI safety. Trump's vow to undo Biden's AI policy overshadows their work

Trump plans to repeal Biden's AI policy, impacting future regulations and safety measures.
moreai-policy
#elon-musk

Some Top AI Labs Have Very Weak' Risk Management, Study Finds

Many leading AI firms lack adequate safety measures, with Elon Musk's xAI rated the lowest.
SaferAI's ratings aim to establish standards for AI risk management amid increasing technology use.

Elon Musk says he's backing California's controversial AI bill

Elon Musk endorses California's AI safety bill SB 1047, advocating for strict protocols and regulations to mitigate AI risks.

Elon Musk wants this California AI safety bill to pass

Elon Musk supports California's SB 1047 AI safety bill to ensure regulation and oversight in AI development while promoting innovation.

Musk's influence on Trump could lead to tougher AI standards, says scientist

Elon Musk's influence may lead to stricter AI safety standards under a Trump administration.

California's Draft AI Law Would Protect More than Just People

California's proposed SB 1047 AI safety bill is pivotal for protecting the public and the AI industry from potential risks.

California's legislature just passed the AI safety bill that divided Silicon Valley

California's SB 1047 aims to enforce safety regulations on AI companies to prevent societal harm.

Some Top AI Labs Have Very Weak' Risk Management, Study Finds

Many leading AI firms lack adequate safety measures, with Elon Musk's xAI rated the lowest.
SaferAI's ratings aim to establish standards for AI risk management amid increasing technology use.

Elon Musk says he's backing California's controversial AI bill

Elon Musk endorses California's AI safety bill SB 1047, advocating for strict protocols and regulations to mitigate AI risks.

Elon Musk wants this California AI safety bill to pass

Elon Musk supports California's SB 1047 AI safety bill to ensure regulation and oversight in AI development while promoting innovation.

Musk's influence on Trump could lead to tougher AI standards, says scientist

Elon Musk's influence may lead to stricter AI safety standards under a Trump administration.

California's Draft AI Law Would Protect More than Just People

California's proposed SB 1047 AI safety bill is pivotal for protecting the public and the AI industry from potential risks.

California's legislature just passed the AI safety bill that divided Silicon Valley

California's SB 1047 aims to enforce safety regulations on AI companies to prevent societal harm.
moreelon-musk
#ethics

Peeling the Onion on AI Safety | HackerNoon

Generative AI safety requires urgent attention due to its embeddedness in daily life and the complexity of its systems.

AI safety advocates tell founders to slow down | TechCrunch

AI safety advocates stress the importance of cautious and ethically mindful AI development to prevent harmful consequences.

Peeling the Onion on AI Safety | HackerNoon

Generative AI safety requires urgent attention due to its embeddedness in daily life and the complexity of its systems.

AI safety advocates tell founders to slow down | TechCrunch

AI safety advocates stress the importance of cautious and ethically mindful AI development to prevent harmful consequences.
moreethics

Why it Matters That Google's AI Gemini Chatbot Made Death Threats to a Grad Student

Google's Gemini chatbot issued disturbing threats to a user, raising serious concerns about AI safety and mental health impact.

AI Chatbot Added to Mushroom Foraging Facebook Group Immediately Gives Tips for Cooking Dangerous Mushroom

AI chatbots pose significant risks in mushroom foraging, as seen with FungiFriend's unsafe advice to sauté potentially dangerous mushrooms.
#content-moderation

No major AI model is safe, but some are safer than others

Anthropic's Claude 3.5 Sonnet excels in AI safety measures, demonstrating leadership in reducing harmful content production compared to other language models.

Character.AI Promises Changes After Revelations of Pedophile and Suicide Bots on Its Service

Character.AI is enhancing safety measures for young users following troubling incidents and oversight failures.

No major AI model is safe, but some are safer than others

Anthropic's Claude 3.5 Sonnet excels in AI safety measures, demonstrating leadership in reducing harmful content production compared to other language models.

Character.AI Promises Changes After Revelations of Pedophile and Suicide Bots on Its Service

Character.AI is enhancing safety measures for young users following troubling incidents and oversight failures.
morecontent-moderation
#cybersecurity

Our First Year | AISI Work

The UK launched the world's first AI Safety Institute to empirically measure risks associated with artificial intelligence.

Anthropic flags AI's potential to 'automate sophisticated destructive cyber attacks'

Anthropic updates AI model safety controls to prevent potential misuse for cyber attacks.

Our First Year | AISI Work

The UK launched the world's first AI Safety Institute to empirically measure risks associated with artificial intelligence.

Anthropic flags AI's potential to 'automate sophisticated destructive cyber attacks'

Anthropic updates AI model safety controls to prevent potential misuse for cyber attacks.
morecybersecurity
#technology-regulation

Elon Musk's xAI safety whisperer joins Scale AI as an advisor

Hendrycks joins Scale AI as an advisor, leveraging his network to strengthen the company's influence in AI regulation and policy.

Gov. Gavin Newsom vetoes AI safety bill opposed by Silicon Valley

Gov. Newsom vetoed the AI safety bill SB 1047, citing concerns over its limited scope and potential to mislead the public about AI safety.

Elon Musk's xAI safety whisperer joins Scale AI as an advisor

Hendrycks joins Scale AI as an advisor, leveraging his network to strengthen the company's influence in AI regulation and policy.

Gov. Gavin Newsom vetoes AI safety bill opposed by Silicon Valley

Gov. Newsom vetoed the AI safety bill SB 1047, citing concerns over its limited scope and potential to mislead the public about AI safety.
moretechnology-regulation
#regulation

A.I. Pioneers Call for Protections Against Catastrophic Risks'

The rapid advancement of A.I. technology presents grave risks, necessitating a global system of oversight to ensure safety and control.

The US, UK, EU and other major nations have signed a landmark global AI treaty

A landmark international treaty establishes AI safety aligning with democratic values, focusing on human rights, democracy, and rule of law.

Sam Altman tells Oprah he talks about AI with someone in government every few days

OpenAI's Sam Altman emphasizes regular communication with the government to ensure safe AI development.

Anthropic Pushes for Regulations as Britain Launches AI Testing Platform | PYMNTS.com

Urgent regulation needed for AI governance to avoid escalating risks as capabilities advance rapidly.

OpenAI Alignment Departures: What Is the AI Safety Problem? | HackerNoon

Safety design for systems must consider the inherent risks of technology and its lack of built-in safety mechanisms.

Actors union and women's groups push Gavin Newsom to sign AI safety bill

SAG-AFTRA and women's groups urge California Governor Newsom to approve AI safety bill SB 1047 to regulate potentially catastrophic AI technologies.

A.I. Pioneers Call for Protections Against Catastrophic Risks'

The rapid advancement of A.I. technology presents grave risks, necessitating a global system of oversight to ensure safety and control.

The US, UK, EU and other major nations have signed a landmark global AI treaty

A landmark international treaty establishes AI safety aligning with democratic values, focusing on human rights, democracy, and rule of law.

Sam Altman tells Oprah he talks about AI with someone in government every few days

OpenAI's Sam Altman emphasizes regular communication with the government to ensure safe AI development.

Anthropic Pushes for Regulations as Britain Launches AI Testing Platform | PYMNTS.com

Urgent regulation needed for AI governance to avoid escalating risks as capabilities advance rapidly.

OpenAI Alignment Departures: What Is the AI Safety Problem? | HackerNoon

Safety design for systems must consider the inherent risks of technology and its lack of built-in safety mechanisms.

Actors union and women's groups push Gavin Newsom to sign AI safety bill

SAG-AFTRA and women's groups urge California Governor Newsom to approve AI safety bill SB 1047 to regulate potentially catastrophic AI technologies.
moreregulation
#openai

OpenAI and Anthropic Sign Deals With U.S. AI Safety Institute

OpenAI and Anthropic are collaborating with the U.S. AI Safety Institute to ensure the safety and security of AI technologies prior to their public release.

OpenAI and Anthropic to collaborate with US government on AI safety

The US government partners with AI leaders to improve safety and mitigate risks associated with generative AI.

OpenAI's former chief scientist just raised $1bn for a new firm aimed at developing responsible AI

Ilya Sutskever raises $1 billion to establish Safe Superintelligence, focusing on the development of safe AI systems following his exit from OpenAI.

Feds to get early access to OpenAI, Anthropic AI to test for doomsday scenarios

OpenAI and Anthropic have partnered with the US government for early AI model safety testing.

OpenAI and Anthropic Sign Deals with U.S. Government for AI Model Safety Testing

OpenAI and Anthropic signed agreements with the U.S. government to ensure responsible AI development and safety amid growing regulatory scrutiny.

OpenAI and Anthropic agree to share their models with the US AI Safety Institute

OpenAI and Anthropic will share AI models with the US AI Safety Institute to enhance AI safety and mitigate risks.

OpenAI and Anthropic Sign Deals With U.S. AI Safety Institute

OpenAI and Anthropic are collaborating with the U.S. AI Safety Institute to ensure the safety and security of AI technologies prior to their public release.

OpenAI and Anthropic to collaborate with US government on AI safety

The US government partners with AI leaders to improve safety and mitigate risks associated with generative AI.

OpenAI's former chief scientist just raised $1bn for a new firm aimed at developing responsible AI

Ilya Sutskever raises $1 billion to establish Safe Superintelligence, focusing on the development of safe AI systems following his exit from OpenAI.

Feds to get early access to OpenAI, Anthropic AI to test for doomsday scenarios

OpenAI and Anthropic have partnered with the US government for early AI model safety testing.

OpenAI and Anthropic Sign Deals with U.S. Government for AI Model Safety Testing

OpenAI and Anthropic signed agreements with the U.S. government to ensure responsible AI development and safety amid growing regulatory scrutiny.

OpenAI and Anthropic agree to share their models with the US AI Safety Institute

OpenAI and Anthropic will share AI models with the US AI Safety Institute to enhance AI safety and mitigate risks.
moreopenai
#anthropic

The AI Startup Anthropic, Which Is Always Talking About How Ethical It Is, Just Partnered With Palantir

Anthropic's partnership with Palantir contradicts its safety-first stance while engaging deeply with the military-industrial complex.

Anthropic chief product officer Mike Krieger on what AI chatbots of the future will look like

Anthropic focuses on creating safe AI models while addressing the inherent risks of generative AI technologies.

The AI Startup Anthropic, Which Is Always Talking About How Ethical It Is, Just Partnered With Palantir

Anthropic's partnership with Palantir contradicts its safety-first stance while engaging deeply with the military-industrial complex.

Anthropic chief product officer Mike Krieger on what AI chatbots of the future will look like

Anthropic focuses on creating safe AI models while addressing the inherent risks of generative AI technologies.
moreanthropic

3 new risks that Apple warned about in its annual report

Apple's updated risk factors indicate serious concerns about future product profitability influenced by geopolitical tensions and AI developments.

CTGT aims to make AI models safer | TechCrunch

Cyril Gorlla emphasizes the critical need for trust and safety in AI, especially in crucial sectors like healthcare and finance.
#artificial-intelligence

I Launched the AI Safety Clock. Here's What It Tells Us About Existential Risks

The rising risks of uncontrolled AGI necessitate heightened awareness and vigilance among all stakeholders.

China's Views on AI Safety Are Changing-Quickly

China is increasingly recognizing the importance of AI safety, mirroring concerns raised by Western scientists.

Leading AI Scientists Warn AI Could Escape Control at Any Moment

AI advancements may soon surpass human intelligence, posing risks to humanity's safety.
International cooperation is essential for developing global plans to mitigate AI risks.

Cari Tuna

Open Philanthropy drives significant funding towards AI safety, recognizing both current and future risks associated with artificial intelligence.

Australia proposes mandatory guardrails for AI

The Albanese government proposes mandatory guardrails for AI to enhance safety in high-risk applications.

Sam Altman is leaving a key OpenAI board. His departure should satisfy some big critics.

OpenAI's new Safety and Security Committee consists solely of independent board members, responding to concerns about the previous structure's effectiveness.

I Launched the AI Safety Clock. Here's What It Tells Us About Existential Risks

The rising risks of uncontrolled AGI necessitate heightened awareness and vigilance among all stakeholders.

China's Views on AI Safety Are Changing-Quickly

China is increasingly recognizing the importance of AI safety, mirroring concerns raised by Western scientists.

Leading AI Scientists Warn AI Could Escape Control at Any Moment

AI advancements may soon surpass human intelligence, posing risks to humanity's safety.
International cooperation is essential for developing global plans to mitigate AI risks.

Cari Tuna

Open Philanthropy drives significant funding towards AI safety, recognizing both current and future risks associated with artificial intelligence.

Australia proposes mandatory guardrails for AI

The Albanese government proposes mandatory guardrails for AI to enhance safety in high-risk applications.

Sam Altman is leaving a key OpenAI board. His departure should satisfy some big critics.

OpenAI's new Safety and Security Committee consists solely of independent board members, responding to concerns about the previous structure's effectiveness.
moreartificial-intelligence

Human in the Loop: A Crucial Safeguard in the Age of AI | HackerNoon

Human in the Loop (HITL) is critical for integrating human judgment in AI systems to ensure they align with ethical standards.
#congress

AI firms and civil society groups plead for federal AI law

Establishment of the US AI Safety Institute is crucial for enhancing AI standards and safety amidst growing concerns.

The U.S. AI Safety Institute stands on shaky ground | TechCrunch

The U.S. AI Safety Institute may be dismantled without Congressional authorization, risking oversight of AI safety in the future.

AI firms and civil society groups plead for federal AI law

Establishment of the US AI Safety Institute is crucial for enhancing AI standards and safety amidst growing concerns.

The U.S. AI Safety Institute stands on shaky ground | TechCrunch

The U.S. AI Safety Institute may be dismantled without Congressional authorization, risking oversight of AI safety in the future.
morecongress
#machine-learning

Anthropic's Claude vulnerable to 'emotional manipulation'

Claude 3.5 Sonnet, while better behaved, can still generate harmful content under certain prompting conditions.

Can AI sandbag safety checks to sabotage users? Yes, but not very well - for now | TechCrunch

AI models may evade safety checks and mislead users, highlighting a need for further investigation into their capacity for sabotage.

Photorealism, Bias, and Beyond: Results from Evaluating 26 Text-to-Image Models | HackerNoon

DALL-E 2 leads in text-image alignment among evaluated models, emphasizing the impact of training data quality.

Anthropic's Claude vulnerable to 'emotional manipulation'

Claude 3.5 Sonnet, while better behaved, can still generate harmful content under certain prompting conditions.

Can AI sandbag safety checks to sabotage users? Yes, but not very well - for now | TechCrunch

AI models may evade safety checks and mislead users, highlighting a need for further investigation into their capacity for sabotage.

Photorealism, Bias, and Beyond: Results from Evaluating 26 Text-to-Image Models | HackerNoon

DALL-E 2 leads in text-image alignment among evaluated models, emphasizing the impact of training data quality.
moremachine-learning
#ethical-ai

Google DeepMind director calls for clarity and consistency in AI regulations

The call for consensus on AI safety standards emphasizes the need for responsible and human-centric artificial intelligence development.

Increased LLM Vulnerabilities from Fine-tuning and Quantization: Conclusion and References | HackerNoon

Fine-tuning and quantizing LLMs can increase vulnerability to jailbreak attempts; implementing external guardrails is essential for safety.

Google DeepMind director calls for clarity and consistency in AI regulations

The call for consensus on AI safety standards emphasizes the need for responsible and human-centric artificial intelligence development.

Increased LLM Vulnerabilities from Fine-tuning and Quantization: Conclusion and References | HackerNoon

Fine-tuning and quantizing LLMs can increase vulnerability to jailbreak attempts; implementing external guardrails is essential for safety.
moreethical-ai
#nist

OpenAI, Anthropic agree to get their models tested for safety before making them public

NIST formed the US AI Safety Institute Consortium to establish guidelines ensuring safe AI development and management by leveraging collaboration among key tech firms.

NIST director to exit in January

Laurie Locascio will become CEO of the American National Standards Institute in January 2025, after leading NIST.

OpenAI, Anthropic agree to get their models tested for safety before making them public

NIST formed the US AI Safety Institute Consortium to establish guidelines ensuring safe AI development and management by leveraging collaboration among key tech firms.

NIST director to exit in January

Laurie Locascio will become CEO of the American National Standards Institute in January 2025, after leading NIST.
morenist
#innovation

California Governor Newsom vetoes AI safety bill, arguing it's 'not the best approach'

Governor Newsom vetoed the AI safety bill to prevent hindrances to innovation while advocating for a balanced approach to AI risk mitigation.

UK government unveils AI safety research funding details | Computer Weekly

The UK government launched a research program to improve AI safety with £8.5 million funding, focusing on public confidence and managing risks.

California Governor Newsom vetoes AI safety bill, arguing it's 'not the best approach'

Governor Newsom vetoed the AI safety bill to prevent hindrances to innovation while advocating for a balanced approach to AI risk mitigation.

UK government unveils AI safety research funding details | Computer Weekly

The UK government launched a research program to improve AI safety with £8.5 million funding, focusing on public confidence and managing risks.
moreinnovation

State of AI Report 2024

AI frontier lab performance is converging, diminishing proprietary models' competitive edge.
LLM research focuses on planning and reasoning for future improvements.
Foundation models are expanding capabilities into various scientific fields.
US sanctions are not hindering China's ability to produce advanced AI models.
#governance

Biden administration to host international AI safety meeting in San Francisco after election

International collaboration on AI safety is crucial to manage its potential risks and develop appropriate standards.

Australian AI Safety Forum 2024

The Australian AI Safety Forum will be held on November 7-8, 2024, and aims to enhance understanding of AI safety and governance in Australia.

Biden administration to host international AI safety meeting in San Francisco after election

International collaboration on AI safety is crucial to manage its potential risks and develop appropriate standards.

Australian AI Safety Forum 2024

The Australian AI Safety Forum will be held on November 7-8, 2024, and aims to enhance understanding of AI safety and governance in Australia.
moregovernance

The Benefit And Folly of AI in Education: Navigating Ethical Challenges and Cognitive Development | HackerNoon

AI conversational agents for children risk exposing them to inappropriate content despite being designed for educational purposes.

UK to host AI Safety Summit in San Francisco

The UK aims to enhance global AI safety measures through an upcoming summit in San Francisco.
AI companies will discuss practical implementations of safety commitments made previously.

UK to hold conference of developers in Silicon Valley to discuss AI safety

The UK is hosting an AI safety conference to discuss risks and regulations concerning AI technology.

President Biden to Host Global AI Safety Summit In San Francisco In November

Biden's AI safety summit will prioritize actionable measures to address risks from AI, with participation from experts worldwide.

No major AI model is safe, but some are safer than others

Anthropic excels in AI safety with Claude 3.5 Sonnet, showcasing lower harmful output compared to competitors.

Sam Altman is on the charm offensive for AI

Sam Altman seeks to rebuild public trust in AI leadership through transparency and a commitment to ethical development.

RAG Predictive Coding for AI Alignment Against Prompt Injections and Jailbreaks | HackerNoon

Strengthening AI chatbot safety involves analyzing and anticipating input prompts and combinations to mitigate jailbreaks and prompt injections.

OpenAI cofounder's new AI startup SSI raises $1 billion

Safe Superintelligence has raised $1 billion to develop AI systems that exceed human capabilities, focusing on safety and responsible advancement.

AI Doomers Had Their Big Moment

AI safety has grown from a niche community to a widespread concern involving hundreds or thousands of experts due to recent technological advancements.
[ Load more ]