The bilingual English/Chinese Baichuan2-7B-Chat VLM trained via LORA for https://arxiv.org/abs/2406.11665.

The Chinese half of the training data used for multimodal alignment and visual instruction tuning is sampled from here.

Downloads last month
11
Safetensors
Model size
7.83B params
Tensor type
F32
·
FP16
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and HF Inference API does not yet support model repos that contain custom code.

Dataset used to train amitha/mllava-baichuan2-en-zh