Model Card for LLaVA-1.6-Mistral-7B-Offensive-Meme-Singapore

This model is described in the paper Detecting Offensive Memes with Social Biases in Singapore Context Using Multimodal Large Language Models. It classifies memes as offensive or not offensive, specifically within the Singaporean context.

Model Details

This model is a fine-tuned Vision-Language Model (VLM) designed to detect offensive memes in the Singaporean context. It leverages the strengths of VLMs to handle the nuanced and culturally specific nature of meme interpretation, addressing the limitations of traditional content moderation systems. The model was fine-tuned on a dataset of 112K memes labeled by GPT-4V. The fine-tuning process involved a pipeline incorporating OCR, translation, and a 7-billion parameter VLM (LLaVA-v1.6-Mistral-7b-hf). The resulting model demonstrates strong performance in offensive meme detection, achieving high accuracy and AUROC scores on a held-out test set.

Uses

Direct Use

The model can be used directly for classifying memes as offensive or non-offensive. Input is expected to be a meme image. The model processes this using OCR and translation where necessary, then utilizes a VLM for classification.

Downstream Use

This model can be integrated into larger content moderation systems to enhance the detection of offensive memes, specifically targeting the Singaporean context.

Out-of-Scope Use

This model is specifically trained for the Singaporean context. Its performance may degrade significantly when applied to memes from other cultures or regions. It is also not suitable for general-purpose image classification tasks.

Bias, Risks, and Limitations

The model's performance is inherently tied to the quality and representativeness of the training data. Biases present in the training data may be reflected in the model's output, particularly regarding the interpretation of culturally specific humor or references. The model may misclassify memes due to ambiguities in language or visual representation. It is crucial to use this model responsibly and acknowledge its limitations.

Recommendations

Users should be aware of the potential biases and limitations of the model. Human review of the model's output is strongly recommended, especially in high-stakes scenarios. Further research into mitigating bias and enhancing robustness is needed.

How to Get Started with the Model

[More Information Needed]

Training Details

Training Data

[More Information Needed]

Training Procedure

[More Information Needed]

Evaluation

Testing Data, Factors & Metrics

Testing Data

[More Information Needed]

Factors

[More Information Needed]

Metrics

[More Information Needed]

Results

[More Information Needed]

Summary

[More Information Needed]

Model Examination

[More Information Needed]

Environmental Impact

[More Information Needed]

Technical Specifications

[More Information Needed]

Citation

@misc{yuxuan2025detectingoffensivememessocial,
      title={Detecting Offensive Memes with Social Biases in Singapore Context Using Multimodal Large Language Models},
      author={Cao Yuxuan and Wu Jiayang and Alistair Cheong Liang Chuen and Bryan Shan Guanrong and Theodore Lee Chong Jen and Sherman Chann Zhi Shen},
      year={2025},
      eprint={2502.18101},
      archivePrefix={arXiv},
      primaryClass={cs.CV},
      url={https://arxiv.org/abs/2502.18101},
}

Glossary

[More Information Needed]

More Information

[More Information Needed]

Model Card Authors

[More Information Needed]

Model Card Contact

[More Information Needed]

Downloads last month
36
Safetensors
Model size
7.57B params
Tensor type
FP16
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for aliencaocao/llava-1.6-mistral-7b-offensive-meme-singapore

Finetuned
(171)
this model

Dataset used to train aliencaocao/llava-1.6-mistral-7b-offensive-meme-singapore

Evaluation results