Meta releases its first open AI model that can process images
Briefly

Meta’s release of Llama 3.2 marks a significant advancement in AI models, combining image and text processing, thus opening new avenues for developers to create innovative applications.
According to Ahmad Al-Dahle, the new multimodal model makes it simpler for developers, requiring merely the integration of 'new multimodality' for it to operate effectively.
With two vision models and two text-only models, Llama 3.2 caters to various hardware specifications, including mobile-friendly sizes to maximize broad usability across devices.
Meta's strategy shows a readiness to enhance real-time image understanding capabilities, positioning itself more competitively against existing multimodal AI applications from rivals like OpenAI and Google.
Read at The Verge
[
]
[
|
]