"A new study from MIT suggests the biggest and most computationally intensive AI models may soon offer diminishing returns compared to smaller models. By mapping scaling laws against continued improvements in model efficiency, the researchers found that it could become harder to wring leaps in performance from giant models whereas efficiency gains could make models running on more modest hardware increasingly capable over the next decade."
"Leaps in efficiency, like those seen with DeepSeek's remarkably low-cost model in January, have already served as a reality check for the AI industry, which is accustomed to burning massive amounts of compute. As things stand, a frontier model from a company like OpenAI is currently much better than a model trained with a fraction of the compute from an academic lab."
Mapping scaling laws against continued efficiency improvements indicates that extremely large, compute-intensive AI models may produce smaller marginal performance gains compared with more efficient, modestly sized models. Efficiency breakthroughs already demonstrated by low-cost models have reduced the compute advantage held by frontier systems. The narrowing gap is expected to be most pronounced for reasoning models that rely on extra inference computation. Continued algorithmic refinement and investment in efficiency could yield large benefits for organizations without massive compute budgets. Novel training methods, such as reinforcement learning advances, could alter this trajectory but do not erase current efficiency trends.
Read at WIRED
Unable to calculate read time
Collection
[
|
...
]