#multi-modal-models

[ follow ]
fromHackernoon
1 year ago

Introducing LLaVA-Phi: A Compact Vision-Language Assistant Powered By a Small Language Model | HackerNoon

LLaVA-Phi showcases the capabilities of smaller language models in multi-modal tasks with only 2.7B parameters.
[ Load more ]