#inference

[ follow ]
Tech industry
fromwww.businessinsider.com
2 days ago

Google's new chips are a shot at Nvidia and a big hint at where AI goes next

Google unveiled its latest AI chips, TPU 8t for training and TPU 8i for inference, responding to industry shifts towards inference computing.
Tech industry
fromTNW | Artificial-Intelligence
4 days ago

Google in talks with Marvell Technology to build new AI inference chips alongside Broadcom TPU programme

Google is collaborating with Marvell Technology to develop new AI chips, enhancing its custom silicon supply chain for inference processing.
#ai
fromTechCrunch
1 week ago
Startup companies

This startup is betting tokenmaxxing will create the next compute giant | TechCrunch

fromTechCrunch
1 month ago
Silicon Valley

Startup Gimlet Labs is solving the AI inference bottleneck in a surprisingly elegant way | TechCrunch

Startup companies
fromTechCrunch
1 week ago

This startup is betting tokenmaxxing will create the next compute giant | TechCrunch

Developers demand fast, cheap tokens for AI models, driving companies like Parasail to innovate in cloud processing for inference.
Silicon Valley
fromTechCrunch
1 month ago

Startup Gimlet Labs is solving the AI inference bottleneck in a surprisingly elegant way | TechCrunch

Gimlet Labs raised $80 million to enhance AI inference efficiency across diverse hardware types.
Artificial intelligence
fromTechCrunch
1 year ago

Ironwood is Google's newest AI accelerator chip | TechCrunch

Google unveiled its seventh-generation TPU chip, Ironwood, optimized for AI inference.
Ironwood will enhance AI model processing capabilities significantly.
fromCointelegraph
2 months ago

What Role Is Left for Decentralized GPU Networks in AI?

What we are beginning to see is that many open-source and other models are becoming compact enough and sufficiently optimized to run very efficiently on consumer GPUs,
Artificial intelligence
Artificial intelligence
fromTechzine Global
3 months ago

Inferencing startup Baseten valued at $5B after new funding round

Baseten raised $300 million at a $5 billion valuation to provide scalable inference infrastructure for deploying AI models, with Nvidia investing about $150 million.
Artificial intelligence
fromInfoWorld
3 months ago

Edge AI: The future of AI inference is smarter local compute

Edge AI shifts computation from cloud to devices, enabling low-latency, cost-efficient, and privacy-preserving AI inference while facing performance and ecosystem challenges.
Artificial intelligence
fromInfoWorld
5 months ago

AI is all about inference now

Enterprise AI success depends more on deploying models against governed business data with guardrails and scalable inference infrastructure than on creating new models.
Artificial intelligence
fromFortune
6 months ago

OpenAI is putting apps in ChatGPT. Why that's a bigger deal than you might think. | Fortune

OpenAI partnered with AMD for M4150 GPUs for inference, committed to six gigawatts and warrants, and launched ChatGPT apps integrating third-party services.
Artificial intelligence
fromComputerworld
10 months ago

Canalys: Companies limit genAI use due to unclear costs

Companies face challenges in predicting cloud costs as they move from testing to real-world use of generative AI due to the recurring operational costs of inference.
[ Load more ]