aashish1904 commited on
Commit
a671d03
1 Parent(s): 52686d5

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +157 -0
README.md ADDED
@@ -0,0 +1,157 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+
4
+ library_name: transformers
5
+ base_model: nvidia/Llama-3.1-Minitron-4B-Width-Base
6
+ tags:
7
+ - axolotl
8
+ - generated_from_trainer
9
+ model-index:
10
+ - name: MagpieLM-4B-SFT-v0.1
11
+ results: []
12
+ datasets:
13
+ - Magpie-Align/MagpieLM-SFT-Data-v0.1
14
+ language:
15
+ - en
16
+
17
+ ---
18
+
19
+ [![QuantFactory Banner](https://lh7-rt.googleusercontent.com/docsz/AD_4nXeiuCm7c8lEwEJuRey9kiVZsRn2W-b4pWlu3-X534V3YmVuVc2ZL-NXg2RkzSOOS2JXGHutDuyyNAUtdJI65jGTo8jT9Y99tMi4H4MqL44Uc5QKG77B0d6-JfIkZHFaUA71-RtjyYZWVIhqsNZcx8-OMaA?key=xt3VSDoCbmTY7o-cwwOFwQ)](https://hf.co/QuantFactory)
20
+
21
+
22
+ # QuantFactory/MagpieLM-4B-SFT-v0.1-GGUF
23
+ This is quantized version of [Magpie-Align/MagpieLM-4B-SFT-v0.1](https://huggingface.co/Magpie-Align/MagpieLM-4B-SFT-v0.1) created using llama.cpp
24
+
25
+ # Original Model Card
26
+
27
+
28
+ ![Magpie](https://cdn-uploads.huggingface.co/production/uploads/653df1323479e9ebbe3eb6cc/FWWILXrAGNwWr52aghV0S.png)
29
+
30
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://api.wandb.ai/links/uw-nsl/7grozq8s)
31
+
32
+ # 🐦 MagpieLM-4B-SFT-v0.1
33
+
34
+ Project Web: [https://magpie-align.github.io/](https://magpie-align.github.io/)
35
+
36
+ Arxiv Technical Report: [https://arxiv.org/abs/2406.08464](https://arxiv.org/abs/2406.08464)
37
+
38
+ Codes: [https://github.com/magpie-align/magpie](https://github.com/magpie-align/magpie)
39
+
40
+ ## About This Model
41
+
42
+ *Model full name: Llama3.1-MagpieLM-4B-SFT-v0.1*
43
+
44
+ This model is a fine-tuned version of [nvidia/Llama-3.1-Minitron-4B-Width-Base](https://huggingface.co/nvidia/Llama-3.1-Minitron-4B-Width-Base) on [Magpie-Align/MagpieLM-SFT-Data-v0.1](https://huggingface.co/datasets/Magpie-Align/MagpieLM-SFT-Data-v0.1) dataset.
45
+
46
+ This is the intermediate checkpoint for fine-tuning [Magpie-Align/MagpieLM-4B-Chat-v0.1](https://huggingface.co/Magpie-Align/MagpieLM-4B-Chat-v0.1).
47
+
48
+ ## Training procedure
49
+
50
+ ### Training hyperparameters
51
+
52
+ The following hyperparameters were used during training:
53
+ - learning_rate: 2e-05
54
+ - train_batch_size: 1
55
+ - eval_batch_size: 1
56
+ - seed: 42
57
+ - distributed_type: multi-GPU
58
+ - num_devices: 4
59
+ - gradient_accumulation_steps: 32
60
+ - total_train_batch_size: 128
61
+ - total_eval_batch_size: 4
62
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
63
+ - lr_scheduler_type: cosine
64
+ - lr_scheduler_warmup_steps: 51
65
+ - num_epochs: 2
66
+
67
+ ### Training results
68
+
69
+ | Training Loss | Epoch | Step | Validation Loss |
70
+ |:-------------:|:------:|:----:|:---------------:|
71
+ | 1.1026 | 0.0038 | 1 | 1.1547 |
72
+ | 0.6994 | 0.2015 | 53 | 0.7142 |
73
+ | 0.6181 | 0.4030 | 106 | 0.6375 |
74
+ | 0.5967 | 0.6045 | 159 | 0.6134 |
75
+ | 0.5793 | 0.8060 | 212 | 0.6004 |
76
+ | 0.5736 | 1.0075 | 265 | 0.5914 |
77
+ | 0.5411 | 1.1938 | 318 | 0.5883 |
78
+ | 0.5402 | 1.3953 | 371 | 0.5864 |
79
+ | 0.5423 | 1.5968 | 424 | 0.5856 |
80
+ | 0.5408 | 1.7983 | 477 | 0.5854 |
81
+
82
+ ### Framework versions
83
+
84
+ - Transformers 4.45.0.dev0
85
+ - Pytorch 2.3.0+cu121
86
+ - Datasets 2.19.1
87
+ - Tokenizers 0.19.1
88
+
89
+ [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
90
+ <details><summary>See axolotl config</summary>
91
+
92
+ axolotl version: `0.4.1`
93
+ ```yaml
94
+ base_model: nvidia/Llama-3.1-Minitron-4B-Width-Base
95
+ model_type: AutoModelForCausalLM
96
+ tokenizer_type: AutoTokenizer
97
+ chat_template: llama3
98
+
99
+ load_in_8bit: false
100
+ load_in_4bit: false
101
+ strict: false
102
+
103
+ datasets:
104
+ - path: Magpie-Align/MagpieLM-SFT-Data-v0.1
105
+ type: sharegpt
106
+ conversation: llama3
107
+ dataset_prepared_path: last_run_prepared
108
+ val_set_size: 0.001
109
+ output_dir: axolotl_out/MagpieLM-4B-SFT-v0.1
110
+
111
+ sequence_len: 8192
112
+ sample_packing: true
113
+ eval_sample_packing: false
114
+ pad_to_sequence_len: true
115
+
116
+ wandb_project: SynDa
117
+ wandb_entity:
118
+ wandb_watch:
119
+ wandb_name: Llama3.1-MagpieLM-4B-SFT-v0.1
120
+ wandb_log_model:
121
+ hub_model_id: Magpie-Align/MagpieLM-4B-SFT-v0.1
122
+
123
+ gradient_accumulation_steps: 32
124
+ micro_batch_size: 1
125
+ num_epochs: 2
126
+ optimizer: paged_adamw_8bit
127
+ lr_scheduler: cosine
128
+ learning_rate: 2e-5
129
+
130
+ train_on_inputs: false
131
+ group_by_length: false
132
+ bf16: true
133
+ fp16:
134
+ tf32: false
135
+
136
+ gradient_checkpointing: true
137
+ gradient_checkpointing_kwargs:
138
+ use_reentrant: false
139
+ early_stopping_patience:
140
+ resume_from_checkpoint:
141
+ logging_steps: 1
142
+ xformers_attention:
143
+ flash_attention: true
144
+
145
+ warmup_ratio: 0.1
146
+ evals_per_epoch: 5
147
+ eval_table_size:
148
+ saves_per_epoch: 1
149
+ debug:
150
+ deepspeed:
151
+ weight_decay: 0.0
152
+ fsdp:
153
+ fsdp_config:
154
+ special_tokens:
155
+ pad_token: <|end_of_text|>
156
+ ```
157
+ </details><br>