MambaInLlama-distill
Collection
Directly distill from Llama without doing SFT and DPO
•
4 items
•
Updated
This model is a fine-tuned version of meta-llama/Llama-3.1-8B-Instruct on the JunxiongWang/sftdatasetv3 dataset. It achieves the following results on the evaluation set:
More information needed
More information needed
More information needed
The following hyperparameters were used during training:
Training Loss | Epoch | Step | Validation Loss |
---|---|---|---|
220.5766 | 1.0000 | 51995 | 242.9601 |
Base model
meta-llama/Llama-3.1-8B