#mixture-of-experts-moe

[ follow ]
Artificial intelligence
fromTheregister
1 day ago

Baidu teases next-gen AI training, inference accelerators

Baidu unveiled M100 (inference) and M300 (training) chips and will deploy Tianchi256/512 clusters to lower inference costs and support larger MoE models.
Artificial intelligence
fromZDNET
1 week ago

A new Chinese AI model claims to outperform GPT-5 and Sonnet 4.5 - and it's free

Moonshot's open-source Kimi K2 Thinking claims superior reasoning performance to GPT-5 and Sonnet 4.5 on several benchmarks while offering tool use and MoE architecture.
fromTheregister
1 week ago

How Perplexity optimized 1T parameter AI models for AWS EFA

AI search provider Perplexity's research wing has developed a new set of software optimizations that allows for trillion parameter or large models to run efficiently across older, cheaper hardware using a variety of existing network technologies, including Amazon's proprietary Elastic Fabric Adapter. These innovations, detailed in a paper published this week and released on GitHub for further scrutiny, present a novel approach to addressing one of the biggest challenges in serving large-scale mixture of experts models (MoE) at scale: memory and network latency.
Artificial intelligence
Artificial intelligence
fromLogRocket Blog
2 months ago

Agentic AI for 5x less: Why Kimi K2 is a frontend game-changer - LogRocket Blog

Kimi K2 is an open-source, agentic mixture-of-experts model that activates 32B of 1T parameters to execute tasks efficiently and rapidly.
[ Load more ]