afrideva commited on
Commit
4c03ea5
·
1 Parent(s): 5c951be

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +100 -0
README.md ADDED
@@ -0,0 +1,100 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: lpetreadg/trained-tinyllama-ultrachat
3
+ inference: false
4
+ license: apache-2.0
5
+ model-index:
6
+ - name: trained-tinyllama-ultrachat
7
+ results: []
8
+ model_creator: lpetreadg
9
+ model_name: trained-tinyllama-ultrachat
10
+ pipeline_tag: text-generation
11
+ quantized_by: afrideva
12
+ tags:
13
+ - generated_from_trainer
14
+ - gguf
15
+ - ggml
16
+ - quantized
17
+ - q2_k
18
+ - q3_k_m
19
+ - q4_k_m
20
+ - q5_k_m
21
+ - q6_k
22
+ - q8_0
23
+ ---
24
+ # lpetreadg/trained-tinyllama-ultrachat-GGUF
25
+
26
+ Quantized GGUF model files for [trained-tinyllama-ultrachat](https://huggingface.co/lpetreadg/trained-tinyllama-ultrachat) from [lpetreadg](https://huggingface.co/lpetreadg)
27
+
28
+
29
+ | Name | Quant method | Size |
30
+ | ---- | ---- | ---- |
31
+ | [trained-tinyllama-ultrachat.q2_k.gguf](https://huggingface.co/afrideva/trained-tinyllama-ultrachat-GGUF/resolve/main/trained-tinyllama-ultrachat.q2_k.gguf) | q2_k | None |
32
+ | [trained-tinyllama-ultrachat.q3_k_m.gguf](https://huggingface.co/afrideva/trained-tinyllama-ultrachat-GGUF/resolve/main/trained-tinyllama-ultrachat.q3_k_m.gguf) | q3_k_m | None |
33
+ | [trained-tinyllama-ultrachat.q4_k_m.gguf](https://huggingface.co/afrideva/trained-tinyllama-ultrachat-GGUF/resolve/main/trained-tinyllama-ultrachat.q4_k_m.gguf) | q4_k_m | None |
34
+ | [trained-tinyllama-ultrachat.q5_k_m.gguf](https://huggingface.co/afrideva/trained-tinyllama-ultrachat-GGUF/resolve/main/trained-tinyllama-ultrachat.q5_k_m.gguf) | q5_k_m | None |
35
+ | [trained-tinyllama-ultrachat.q6_k.gguf](https://huggingface.co/afrideva/trained-tinyllama-ultrachat-GGUF/resolve/main/trained-tinyllama-ultrachat.q6_k.gguf) | q6_k | None |
36
+ | [trained-tinyllama-ultrachat.q8_0.gguf](https://huggingface.co/afrideva/trained-tinyllama-ultrachat-GGUF/resolve/main/trained-tinyllama-ultrachat.q8_0.gguf) | q8_0 | None |
37
+
38
+
39
+
40
+ ## Original Model Card:
41
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
42
+ should probably proofread and complete it, then remove this comment. -->
43
+
44
+ # trained-tinyllama-ultrachat
45
+
46
+ This model is a fine-tuned version of [PY007/TinyLlama-1.1B-intermediate-step-715k-1.5T](https://huggingface.co/PY007/TinyLlama-1.1B-intermediate-step-715k-1.5T) on the None dataset.
47
+ It achieves the following results on the evaluation set:
48
+ - Loss: 1.3258
49
+
50
+ ## Model description
51
+
52
+ More information needed
53
+
54
+ ## Intended uses & limitations
55
+
56
+ More information needed
57
+
58
+ ## Training and evaluation data
59
+
60
+ More information needed
61
+
62
+ ## Training procedure
63
+
64
+ ### Training hyperparameters
65
+
66
+ The following hyperparameters were used during training:
67
+ - learning_rate: 2e-05
68
+ - train_batch_size: 64
69
+ - eval_batch_size: 64
70
+ - seed: 42
71
+ - gradient_accumulation_steps: 2
72
+ - total_train_batch_size: 128
73
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
74
+ - lr_scheduler_type: cosine
75
+ - lr_scheduler_warmup_steps: 1
76
+ - num_epochs: 1
77
+
78
+ ### Training results
79
+
80
+ | Training Loss | Epoch | Step | Validation Loss |
81
+ |:-------------:|:-----:|:----:|:---------------:|
82
+ | 1.3767 | 0.08 | 100 | 1.3685 |
83
+ | 1.3494 | 0.17 | 200 | 1.3490 |
84
+ | 1.3436 | 0.25 | 300 | 1.3389 |
85
+ | 1.3231 | 0.33 | 400 | 1.3331 |
86
+ | 1.3278 | 0.42 | 500 | 1.3296 |
87
+ | 1.3214 | 0.5 | 600 | 1.3276 |
88
+ | 1.3376 | 0.58 | 700 | 1.3266 |
89
+ | 1.3227 | 0.67 | 800 | 1.3261 |
90
+ | 1.3329 | 0.75 | 900 | 1.3259 |
91
+ | 1.3185 | 0.83 | 1000 | 1.3258 |
92
+ | 1.332 | 0.92 | 1100 | 1.3258 |
93
+
94
+
95
+ ### Framework versions
96
+
97
+ - Transformers 4.34.1
98
+ - Pytorch 2.0.1+cu118
99
+ - Datasets 2.14.5
100
+ - Tokenizers 0.14.1