-
RLHF-V: Towards Trustworthy MLLMs via Behavior Alignment from Fine-grained Correctional Human Feedback
Paper • 2312.00849 • Published • 12 -
RLAIF-V: Aligning MLLMs through Open-Source AI Feedback for Super GPT-4V Trustworthiness
Paper • 2405.17220 • Published • 1 -
ImageReward: Learning and Evaluating Human Preferences for Text-to-Image Generation
Paper • 2304.05977 • Published • 2 -
Interpretable Preferences via Multi-Objective Reward Modeling and Mixture-of-Experts
Paper • 2406.12845 • Published • 1
Robert Wijaya
wijayarobert
AI & ML interests
Computer Vision
Recent Activity
authored
a paper
4 days ago
Crowdsource, Crawl, or Generate? Creating SEA-VL, a Multicultural
Vision-Language Dataset for Southeast Asia
Organizations
None yet
Collections
1
Papers
2
models
3
datasets
None public yet