New AI models like ChatGPT pursue superintelligence', but can't be trusted when it comes to basic questions
Briefly

As Lexin Zhou points out, the performance of language models is a mixed bag: they show improvements in handling complex tasks, yet their reliability in basic queries diminishes, raising concerns about their practical utility in diverse job settings.
Jose Hernandez-Orallo highlights a troubling duality: these models excel at complex challenges but stumble on simple tasks, leading to potential misinformation and reliability issues that complicate their integration into roles requiring accuracy.
Despite advancements, Zhou notes that the situation appears to be deteriorating. Models like ChatGPT-4 may take excessively long to attempt answers rather than simply admitting lack of knowledge, leading to inefficiency.
The difference between human expectations and the models' performance is exacerbating the reliability issue, as users progressively pose harder questions, inadvertently setting the stage for greater discrepancies in perceived capability.
Read at english.elpais.com
[
]
[
|
]