File size: 656 Bytes
d57a808 ddcbd88 d57a808 954a6d6 d57a808 2b025ae d57a808 2b025ae d57a808 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 |
---
base_model: google/gemma-2-2b
datasets: mlabonne/TheTome
---
# Distil Gemma 2 2b
This model is a gemma 2 2b model distilled from google/gemma-2-9b-it and finetuned on the tome.
![image/webp](https://cdn-uploads.huggingface.co/production/uploads/6455cc8d679315e4ef16fbec/89XFihSa8o08wWw8w53uh.webp)
## Prompt Template
ChatML
```
<|im_start|>system
{system}<|im_end|>
<|im_start|>user
{user}<|im_end|>
<|im_start|>assistant
```
## Training Information
This model trained on 8x Nvidia H100 NVL for the equivalent of 120 GPU hours.
+ Loss Achieved: 0.27
+ Epochs: 3
Checkpoints are available in the repo to continue training
## Evals
IN PROGRESS |