Update README.md
Browse files
README.md
CHANGED
|
@@ -184,6 +184,20 @@ model.push_to_hub_gguf("Ramikan-BR/tinyllama-coder-py-4bit-v10", tokenizer, quan
|
|
| 184 |
model.save_pretrained_gguf("model", tokenizer, quantization_method = "q4_k_m")
|
| 185 |
model.push_to_hub_gguf("Ramikan-BR/tinyllama-coder-py-4bit-v10", tokenizer, quantization_method = "q4_k_m", token = "hf_...")
|
| 186 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 187 |
Parameters:
|
| 188 |
|
| 189 |
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
|
|
|
|
| 184 |
model.save_pretrained_gguf("model", tokenizer, quantization_method = "q4_k_m")
|
| 185 |
model.push_to_hub_gguf("Ramikan-BR/tinyllama-coder-py-4bit-v10", tokenizer, quantization_method = "q4_k_m", token = "hf_...")
|
| 186 |
|
| 187 |
+
Loss for 5 epochs in the last training session of the last part of the dataset:
|
| 188 |
+
==((====))== Unsloth - 2x faster free finetuning | Num GPUs = 1
|
| 189 |
+
\\ /| Num examples = 407 | Num Epochs = 5
|
| 190 |
+
O^O/ \_/ \ Batch size per device = 2 | Gradient Accumulation steps = 256
|
| 191 |
+
\ / Total batch size = 512 | Total steps = 5
|
| 192 |
+
"-____-" Number of trainable parameters = 201,850,880
|
| 193 |
+
[5/5 29:36, Epoch 3/5]
|
| 194 |
+
Step Training Loss
|
| 195 |
+
1 0.568000
|
| 196 |
+
2 0.145300
|
| 197 |
+
3 0.506100
|
| 198 |
+
4 0.331900
|
| 199 |
+
5 0.276100
|
| 200 |
+
|
| 201 |
Parameters:
|
| 202 |
|
| 203 |
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
|