LLaVA-Phi leverages the Phi-2 model to deliver effective multi-modal dialogues with only 2.7B parameters, demonstrating that smaller models can achieve high performance.
We train a bilingual Arabic-Hebrew language model using a transliterated version of Arabic texts in Hebrew, ensuring both languages are represented in the same script.