LLaVA-Phi: How We Rigorously Evaluated It Using an Extensive Array of Academic Benchmarks | HackerNoonLLaVA-Phi shows significant advancements in visual question-answering, surpassing existing large multimodal models.
Introducing LLaVA-Phi: A Compact Vision-Language Assistant Powered By a Small Language Model | HackerNoonLLaVA-Phi showcases the capabilities of smaller language models in multi-modal tasks with only 2.7B parameters.
Alibaba Releases Two Open-Weight Language Models for Math and Voice ChatAlibaba unveiled Qwen2-Math for mathematical solutions and Qwen2-Audio for multi-modal input, showcasing significant advancements in language models.