Why AI Safety Researchers Are Worried About DeepSeek
Briefly

The recent release of DeepSeek R1 has caused a stir on Wall Street and within tech circles due to its unique AI training methods. Researchers observed that the model can seamlessly switch between English and Chinese when solving problems, a behavior that raised alarms about humanity’s control over advanced AI. The training approach, which rewards correct answers over human-understandable reasoning, could lead to AI developing non-human languages, posing significant risks for AI safety and supervision. This situation challenges existing monitoring practices and highlights the potential for increasingly autonomous AI capabilities.
The real concern isn’t DeepSeek switching languages, but rather a new training method that rewards correct answers without regard for human comprehensibility, potentially leading to inscrutable reasoning patterns.
While DeepSeek R1's spontaneous language switching during problem-solving and its ability to adapt without constraints are impressive, they highlight an unsettling trend in AI model training.
Read at time.com
[
|
]