tomtseng commited on
Commit
bc51537
·
verified ·
1 Parent(s): 0309311

Model save

Browse files
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
  license: apache-2.0
3
- base_model: Qwen/Qwen2.5-7B-Instruct
4
  tags:
5
  - generated_from_trainer
6
  library_name: peft
@@ -14,7 +14,7 @@ should probably proofread and complete it, then remove this comment. -->
14
 
15
  # robust_llm_r2d2_tom-debug_Qwen2.5-7B-Instruct_lora
16
 
17
- This model is a fine-tuned version of [Qwen/Qwen2.5-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-7B-Instruct) on an unknown dataset.
18
 
19
  ## Model description
20
 
@@ -38,10 +38,10 @@ The following hyperparameters were used during training:
38
  - eval_batch_size: 2
39
  - seed: 42
40
  - distributed_type: multi-GPU
41
- - num_devices: 4
42
  - gradient_accumulation_steps: 4
43
- - total_train_batch_size: 16
44
- - total_eval_batch_size: 8
45
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
46
  - lr_scheduler_type: cosine
47
  - num_epochs: 1
 
1
  ---
2
  license: apache-2.0
3
+ base_model: Qwen/Qwen2.5-14B-Instruct
4
  tags:
5
  - generated_from_trainer
6
  library_name: peft
 
14
 
15
  # robust_llm_r2d2_tom-debug_Qwen2.5-7B-Instruct_lora
16
 
17
+ This model is a fine-tuned version of [Qwen/Qwen2.5-14B-Instruct](https://huggingface.co/Qwen/Qwen2.5-14B-Instruct) on an unknown dataset.
18
 
19
  ## Model description
20
 
 
38
  - eval_batch_size: 2
39
  - seed: 42
40
  - distributed_type: multi-GPU
41
+ - num_devices: 8
42
  - gradient_accumulation_steps: 4
43
+ - total_train_batch_size: 32
44
+ - total_eval_batch_size: 16
45
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
46
  - lr_scheduler_type: cosine
47
  - num_epochs: 1
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "Qwen/Qwen2.5-7B-Instruct",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
@@ -16,10 +16,10 @@
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
19
- "o_proj",
20
  "v_proj",
 
21
  "q_proj",
22
- "k_proj"
23
  ],
24
  "task_type": "CAUSAL_LM"
25
  }
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "Qwen/Qwen2.5-14B-Instruct",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
 
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
 
19
  "v_proj",
20
+ "k_proj",
21
  "q_proj",
22
+ "o_proj"
23
  ],
24
  "task_type": "CAUSAL_LM"
25
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:916deea7e1fc999df6e785e88afe3227a7b77250daeef65cde68ccf8ff619ab9
3
- size 80770656
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:054fd3370c720b1fcc8a362a30b4d9226c219b1d816fc9d8a008a38bdb8d621d
3
+ size 201379120
runs/Feb02_01-49-13_r2d2-devbox-526kp/events.out.tfevents.1738460975.r2d2-devbox-526kp.1829.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0416d226fb80c8b548ec11c2d48b834931f6a7815381007c10fb378535857320
3
- size 4622
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af79440ab54b24adffc0649e162e3cf0d74cf071a7b0d10c2692b33bd1489f47
3
+ size 4776
runs/Feb02_02-31-29_r2d2-devbox-526kp/events.out.tfevents.1738463538.r2d2-devbox-526kp.5134.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c9d280afedbad93bd14ee7cbef5a98d8bbc166dcea4c53376673701568bec27
3
+ size 4623
step_0/README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- base_model: Qwen/Qwen2.5-7B-Instruct
3
  library_name: peft
4
  ---
5
 
@@ -203,4 +203,11 @@ Carbon emissions can be estimated using the [Machine Learning Impact calculator]
203
  ### Framework versions
204
 
205
 
 
 
 
 
 
 
 
206
  - PEFT 0.6.1
 
1
  ---
2
+ base_model: Qwen/Qwen2.5-14B-Instruct
3
  library_name: peft
4
  ---
5
 
 
203
  ### Framework versions
204
 
205
 
206
+ - PEFT 0.6.1
207
+ ## Training procedure
208
+
209
+
210
+ ### Framework versions
211
+
212
+
213
  - PEFT 0.6.1
step_0/adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "Qwen/Qwen2.5-7B-Instruct",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
@@ -16,10 +16,10 @@
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
19
- "o_proj",
20
  "v_proj",
 
21
  "q_proj",
22
- "k_proj"
23
  ],
24
  "task_type": "CAUSAL_LM"
25
  }
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "Qwen/Qwen2.5-14B-Instruct",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
 
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
 
19
  "v_proj",
20
+ "k_proj",
21
  "q_proj",
22
+ "o_proj"
23
  ],
24
  "task_type": "CAUSAL_LM"
25
  }
step_0/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:916deea7e1fc999df6e785e88afe3227a7b77250daeef65cde68ccf8ff619ab9
3
- size 80770656
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:054fd3370c720b1fcc8a362a30b4d9226c219b1d816fc9d8a008a38bdb8d621d
3
+ size 201379120
step_0/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1b46f0df35125f293adfdbfb156efdb1968b7dc0752cae97daa96b012f0b3464
3
  size 5816
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd29f6ce8baa4403da6138b1ab6401e9f3f3e2e8b1f571f87bed547b2cef427f
3
  size 5816
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1b46f0df35125f293adfdbfb156efdb1968b7dc0752cae97daa96b012f0b3464
3
  size 5816
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd29f6ce8baa4403da6138b1ab6401e9f3f3e2e8b1f571f87bed547b2cef427f
3
  size 5816