kweinmeister commited on
Commit
a3f09db
·
verified ·
1 Parent(s): d1ff27e

End of training

Browse files
Files changed (2) hide show
  1. README.md +148 -0
  2. adapter_model.bin +3 -0
README.md ADDED
@@ -0,0 +1,148 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ license: gemma
4
+ base_model: google/gemma-2-2b-it
5
+ tags:
6
+ - axolotl
7
+ - generated_from_trainer
8
+ model-index:
9
+ - name: gemma-2-2b-it-dolly-15k
10
+ results: []
11
+ ---
12
+
13
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
+ should probably proofread and complete it, then remove this comment. -->
15
+
16
+ [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
17
+ <details><summary>See axolotl config</summary>
18
+
19
+ axolotl version: `0.5.2`
20
+ ```yaml
21
+ base_model: google/gemma-2-2b-it
22
+ hub_model_id: kweinmeister/gemma-2-2b-it-dolly-15k
23
+
24
+ load_in_8bit: false
25
+ load_in_4bit: true
26
+ strict: false
27
+
28
+ datasets:
29
+ - path: databricks/databricks-dolly-15k
30
+ type:
31
+ field_instruction: instruction
32
+ field_input: context
33
+ field_output: response
34
+ val_set_size: 0.05
35
+
36
+ sequence_len: 2048
37
+ sample_packing: true
38
+ eval_sample_packing: true
39
+ pad_to_sequence_len: true
40
+
41
+ adapter: qlora
42
+ lora_model_dir:
43
+ lora_r: 32
44
+ lora_alpha: 64
45
+ lora_dropout: 0.05
46
+ lora_target_linear: true
47
+ lora_fan_in_fan_out:
48
+
49
+ wandb_project: gemma-2-2b-it-dolly-15k
50
+ wandb_entity:
51
+ wandb_watch:
52
+ wandb_name:
53
+ wandb_log_model:
54
+
55
+ gradient_accumulation_steps: 4
56
+ micro_batch_size: 1
57
+ num_epochs: 1
58
+ optimizer: adamw_torch
59
+ lr_scheduler: cosine
60
+ learning_rate: 0.0002
61
+
62
+ train_on_inputs: false
63
+ group_by_length: false
64
+ bf16: auto
65
+ fp16:
66
+ tf32: true
67
+
68
+ gradient_checkpointing: true
69
+ gradient_checkpointing_kwargs:
70
+ use_reentrant: true
71
+ early_stopping_patience:
72
+ resume_from_checkpoint:
73
+ local_rank:
74
+ logging_steps: 1
75
+ xformers_attention:
76
+ flash_attention: false
77
+
78
+ warmup_steps: 10
79
+ evals_per_epoch: 4
80
+ saves_per_epoch: 1
81
+ debug:
82
+ deepspeed: deepspeed_configs/zero1.json
83
+ weight_decay: 0.0
84
+ fsdp:
85
+ fsdp_config:
86
+ special_tokens:
87
+ output_dir: "/mnt/disks/gcs/axolotl/runs/google--gemma-2-2b-it-20250101-144050/out/"
88
+ dataset_prepared_path: "/mnt/disks/gcs/axolotl/last_run_prepared"
89
+
90
+ ```
91
+
92
+ </details><br>
93
+
94
+ # gemma-2-2b-it-dolly-15k
95
+
96
+ This model is a fine-tuned version of [google/gemma-2-2b-it](https://huggingface.co/google/gemma-2-2b-it) on the None dataset.
97
+ It achieves the following results on the evaluation set:
98
+ - Loss: 1.7389
99
+
100
+ ## Model description
101
+
102
+ More information needed
103
+
104
+ ## Intended uses & limitations
105
+
106
+ More information needed
107
+
108
+ ## Training and evaluation data
109
+
110
+ More information needed
111
+
112
+ ## Training procedure
113
+
114
+ ### Training hyperparameters
115
+
116
+ The following hyperparameters were used during training:
117
+ - learning_rate: 0.0002
118
+ - train_batch_size: 1
119
+ - eval_batch_size: 1
120
+ - seed: 42
121
+ - distributed_type: multi-GPU
122
+ - num_devices: 2
123
+ - gradient_accumulation_steps: 4
124
+ - total_train_batch_size: 8
125
+ - total_eval_batch_size: 2
126
+ - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
127
+ - lr_scheduler_type: cosine
128
+ - lr_scheduler_warmup_steps: 10
129
+ - num_epochs: 1
130
+
131
+ ### Training results
132
+
133
+ | Training Loss | Epoch | Step | Validation Loss |
134
+ |:-------------:|:------:|:----:|:---------------:|
135
+ | 5.7033 | 0.0061 | 1 | 5.5100 |
136
+ | 1.8197 | 0.2492 | 41 | 1.8752 |
137
+ | 1.6386 | 0.4985 | 82 | 1.7666 |
138
+ | 1.7346 | 0.7477 | 123 | 1.7436 |
139
+ | 1.7742 | 0.9970 | 164 | 1.7389 |
140
+
141
+
142
+ ### Framework versions
143
+
144
+ - PEFT 0.13.2
145
+ - Transformers 4.46.3
146
+ - Pytorch 2.4.1+cu124
147
+ - Datasets 3.1.0
148
+ - Tokenizers 0.20.3
adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:58c4c0b2fb94816856942067d94e5484a814a19efab094a4526505f88e96102a
3
+ size 83124666