Training LLaVA-Phi follows a structured pipeline that includes pretraining and instruction tuning phases, enhancing the model's visual comprehension and language processing capabilities.
Supervised fine-tuning on Phi-2 demonstrates that even a small amount of high-quality data can greatly improve performance in tasks such as language reasoning and coding.
Collection
[
|
...
]