|
--- |
|
license: apache-2.0 |
|
--- |
|
|
|
# OLMo 7B-Instruct-hf |
|
|
|
> For more details on OLMO-7B-Instruct, refer to [Allen AI's OLMo-7B-Instruct model card](https://huggingface.co/allenai/OLMo-7B-Instruct). |
|
|
|
OLMo is a series of **O**pen **L**anguage **Mo**dels designed to enable the science of language models. |
|
The OLMo base models are trained on the [Dolma](https://huggingface.co/datasets/allenai/dolma) dataset. |
|
The Instruct version is trained on the [cleaned version of the UltraFeedback dataset](https://huggingface.co/datasets/allenai/ultrafeedback_binarized_cleaned). |
|
|
|
OLMo 7B Instruct is trained for better question answering. They show the performance gain that OLMo base models can achieve with existing fine-tuning techniques. |
|
|
|
**This version is for direct use with HuggingFace Transformers** from v4.40 on. |
|
|
|
Run instructions are forthcoming. |
|
|
|
For faster inference with llama.cpp or similar software supporting the GGUF format, |
|
you can find this model as GGUF at [ssec-uw/OLMo-7B-Instruct-GGUF](https://huggingface.co/ssec-uw/OLMo-7B-Instruct-GGUF). |
|
|
|
## Contact |
|
|
|
For errors in this model card, contact Don or Anant, {landungs, anmittal} at uw dot edu. |
|
|
|
## Acknowledgement |
|
|
|
We would like to thank the hardworking folks at [Allen AI](https://huggingface.co/allenai) for providing the original model. |
|
|
|
Additionally, the work to convert the model to the new `hf` version was done by the |
|
[University of Washington Scientific Software Engineering Center (SSEC)](https://escience.washington.edu/software-engineering/ssec/), |
|
as part of the [Schmidt Futures Virtual Institute for Scientific Software (VISS)](https://www.schmidtsciences.org/viss/). |