Intel, Ampere show LLMs on CPUs isn't as crazy as it soundsSmaller AI models can now be efficiently run on CPUs with reduced latency, challenging the dominance of GPUs for AI tasks.