SelectiveDPO
Collection
Released models trained by Selective DPO.
•
5 items
•
Updated
This model is fine-tuned from the HuggingFaceH4/mistral-7b-sft-beta model using the SelectiveDPO algorithm on the Ultrafeedback_binarized dataset.
For the recipe to reproduce this model, please visit our GitHub page.
Base model
mistralai/Mistral-7B-v0.1