Update README.md
Browse files
README.md
CHANGED
@@ -12,7 +12,7 @@ library_name: transformers
|
|
12 |
|
13 |
# OLMo-2-1124-7B-DPO
|
14 |
|
15 |
-
OLMo-2
|
16 |
Tülu 3 is designed for state-of-the-art performance on a diversity of tasks in addition to chat, such as MATH, GSM8K, and IFEval.
|
17 |
Check out [the OLMo-2 paper](https://TODO) or [Tülu 3 paper](https://arxiv.org/abs/2411.15124) for more details!
|
18 |
|
|
|
12 |
|
13 |
# OLMo-2-1124-7B-DPO
|
14 |
|
15 |
+
OLMo-2 7B DPO November 2024 is finetuned variant of the [OLMo-2 7B November 2024](https://huggingface.co/allenai/OLMo2-7B-1124) model, which has undergone supervised finetuning on the [Tülu 3 dataset](https://huggingface.co/datasets/allenai/tulu-3-sft-mixture) and further DPO training.
|
16 |
Tülu 3 is designed for state-of-the-art performance on a diversity of tasks in addition to chat, such as MATH, GSM8K, and IFEval.
|
17 |
Check out [the OLMo-2 paper](https://TODO) or [Tülu 3 paper](https://arxiv.org/abs/2411.15124) for more details!
|
18 |
|