Add citations
Browse files
README.md
CHANGED
@@ -286,3 +286,20 @@ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-le
|
|
286 |
|MuSR (0-shot) | 4.79|
|
287 |
|MMLU-PRO (5-shot) | 1.53|
|
288 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
286 |
|MuSR (0-shot) | 4.79|
|
287 |
|MMLU-PRO (5-shot) | 1.53|
|
288 |
|
289 |
+
## Citation
|
290 |
+
If you use MicroLlama in your research or work, please cite the project using the following reference:
|
291 |
+
APA:
|
292 |
+
```
|
293 |
+
Wang, Z. K. (2024). MicroLlama: A 300M-parameter language model trained from scratch. GitHub & Hugging Face. https://github.com/keeeeenw/MicroLlama, https://huggingface.co/keeeeenw/MicroLlama
|
294 |
+
```
|
295 |
+
BibTeX:
|
296 |
+
```
|
297 |
+
@misc{wang2024microllama,
|
298 |
+
author = {Zixiao Ken Wang},
|
299 |
+
title = {MicroLlama: A 300M-parameter language model trained from scratch},
|
300 |
+
year = {2024},
|
301 |
+
howpublished = {\url{https://github.com/keeeeenw/MicroLlama}, \url{https://huggingface.co/keeeeenw/MicroLlama}},
|
302 |
+
note = {GitHub and Hugging Face repositories}
|
303 |
+
}
|
304 |
+
```
|
305 |
+
🙏 Please cite this work if you find it useful.
|