The bilingual English/Chinese Baichuan2-7B-Chat VLM trained via LORA for https://arxiv.org/abs/2406.11665.

The Chinese half of the training data used for multimodal alignment and visual instruction tuning is sampled from here.

Downloads last month
12
Safetensors
Model size
7.83B params
Tensor type
F32
·
FP16
·
Inference Examples
Inference API (serverless) does not yet support model repos that contain custom code.

Dataset used to train amitha/mllava-baichuan2-en-zh