Meta Llama: Everything you need to know about the open generative AI model | TechCrunch
Briefly

Meta Llama: Everything you need to know about the open generative AI model | TechCrunch
"Like every Big Tech company these days, Meta has its own flagship generative AI model, called Llama. Llama is somewhat unique among major models in that it's "open," meaning developers can download and use it however they please (with certain limitations). That's in contrast to models like Anthropic's Claude, Google's Gemini, xAI's Grok, and most of OpenAI's ChatGPT models, which can only be accessed via APIs."
"In the interest of giving developers choice, however, Meta has also partnered with vendors, including AWS, Google Cloud, and Microsoft Azure, to make cloud-hosted versions of Llama available. In addition, the company publishes tools, libraries, and recipes in its Llama cookbook to help developers fine-tune, evaluate, and adapt the models to their domain. With newer generations likeLlama 3 and Llama 4, these capabilities have expanded to include native multimodal support and broader cloud rollouts."
"What is Llama? Llama is a family of models - not just one. The latest version is Llama 4; it was released in April 2025 and includes three models: Scout: 17 billion active parameters, 109 billion total parameters, and a context window of 10 million tokens. Maverick: 17 billion active parameters, 400 billion total parameters, and a context window of 1 million tokens."
Llama is a family of open generative AI models from Meta, with Llama 4 released in April 2025. Llama 4 comprises Scout (17 billion active, 109 billion total, 10 million token context), Maverick (17 billion active, 400 billion total, 1 million token context), and the forthcoming Behemoth (expected 288 billion active, 2 trillion total). Meta provides both downloadable open models (with limitations) and cloud-hosted deployments through AWS, Google Cloud, and Microsoft Azure. Meta also publishes tools, libraries, and a Llama cookbook to enable fine-tuning, evaluation, and domain adaptation. Newer generations add native multimodal support and wider cloud rollouts.
Read at TechCrunch
Unable to calculate read time
[
|
]