#multi-modal-models

[ follow ]
fromHackernoon
1 year ago

Introducing LLaVA-Phi: A Compact Vision-Language Assistant Powered By a Small Language Model | HackerNoon

LLaVA-Phi leverages the Phi-2 model to deliver effective multi-modal dialogues with only 2.7B parameters, demonstrating that smaller models can achieve high performance.
Data science
[ Load more ]