Chatbots Are Academically Dishonest
Briefly

The AI landscape in 2025 has been characterized by a surge in new models from major players like OpenAI and Anthropic, who market their offerings as the most intelligent. However, true evaluation of these models' capabilities is complicated by benchmark manipulation; many leading AI systems have been trained on data that includes these tests, undermining the accuracy of their reported intelligence. Critics argue that this undermines the legitimacy of AI assessments, drawing a parallel with a student who cheats on exams to appear knowledgeable.
Think of it like a human student who steals and memorizes a math test, fooling his teacher into thinking he's learned how to do long division.
Determining how "intelligent" programs like GPT-4.5 or Claude 3.7 are is tricky, as vague metrics make for easy claims by companies.
Read at The Atlantic
[
|
]