Google's Gemma 3 is an open source, single-GPU AI with a 128K context window
Briefly

Google's Gemma 3 is presented as a leading single-accelerator model, showcasing improved chat capabilities and performance metrics over peers. The range of model sizes caters to different hardware, with the smallest version requiring under a gigabyte of memory and the largest demanding 20-30GB. When tested with the Elo metric, Gemma 3's 27 billion parameter model outperformed many competitors, but needs a single Nvidia H100. Although Google claims superior math and coding abilities, they provide no statistical backing. The open-source nature allows flexibility while adhering to Google’s licensing restrictions.
Gemma 3, touted by Google as the 'world's best single-accelerator model,' marks a significant enhancement in chat and functionality over prior models.
Offering models ranging from 1 billion to 27 billion parameters, Gemma 3 delivers flexibility for various hardware capabilities, showcasing its efficiency in local processing.
Read at Ars Technica
[
|
]