AI startup Cerebras debuts 'world's fastest inference' service - with a twist
Briefly

"We have never seen a technology market growing this fast," said Cerebras cofounder and CEO Andrew Feldman in a press conference in San Francisco. "We intend to take meaningful share."
Cerebras has set up its own inference data centers in multiple locations and will rent inference capacity for a fee on a per-query basis. It will also sell its CS-3 computers to companies that wish to perform inference on-premise, either managed by the customer or as a service managed by Cerebras.
Nvidia currently dominates the market both for training neural nets, including generative AI, and the sales of accelerator chips for performing inference.
The new service puts those machines behind the scenes, creating a revenue model based not on machine sales but on volume of transactions.
Read at ZDNET
[
|
]