msab97 commited on
Commit
fb950f6
·
verified ·
1 Parent(s): 1dee721

End of training

Browse files
Files changed (3) hide show
  1. README.md +10 -10
  2. adapter_config.json +6 -3
  3. adapter_model.safetensors +2 -2
README.md CHANGED
@@ -18,10 +18,10 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  This model is a fine-tuned version of [meta-llama/Llama-3.2-3B](https://huggingface.co/meta-llama/Llama-3.2-3B) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 0.4414
22
- - Model Preparation Time: 0.0045
23
  - Accuracy: 0.8313
24
- - F1 Macro: 0.8354
25
 
26
  ## Model description
27
 
@@ -56,17 +56,17 @@ The following hyperparameters were used during training:
56
 
57
  | Training Loss | Epoch | Step | Validation Loss | Model Preparation Time | Accuracy | F1 Macro |
58
  |:-------------:|:-----:|:----:|:---------------:|:----------------------:|:--------:|:--------:|
59
- | 0.9162 | 1.0 | 368 | 0.8379 | 0.0045 | 0.6361 | 0.6334 |
60
- | 0.5201 | 2.0 | 736 | 0.5242 | 0.0045 | 0.7782 | 0.7849 |
61
- | 0.3988 | 3.0 | 1104 | 0.4936 | 0.0045 | 0.7993 | 0.8024 |
62
- | 0.3288 | 4.0 | 1472 | 0.4774 | 0.0045 | 0.8007 | 0.8084 |
63
- | 0.3602 | 5.0 | 1840 | 0.4858 | 0.0045 | 0.8034 | 0.8103 |
64
 
65
 
66
  ### Framework versions
67
 
68
  - PEFT 0.14.0
69
- - Transformers 4.48.2
70
  - Pytorch 2.5.1+cu124
71
- - Datasets 3.2.0
72
  - Tokenizers 0.21.0
 
18
 
19
  This model is a fine-tuned version of [meta-llama/Llama-3.2-3B](https://huggingface.co/meta-llama/Llama-3.2-3B) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 0.4550
22
+ - Model Preparation Time: 0.0066
23
  - Accuracy: 0.8313
24
+ - F1 Macro: 0.8378
25
 
26
  ## Model description
27
 
 
56
 
57
  | Training Loss | Epoch | Step | Validation Loss | Model Preparation Time | Accuracy | F1 Macro |
58
  |:-------------:|:-----:|:----:|:---------------:|:----------------------:|:--------:|:--------:|
59
+ | 0.5588 | 1.0 | 368 | 0.5780 | 0.0066 | 0.7503 | 0.7548 |
60
+ | 0.4576 | 2.0 | 736 | 0.4660 | 0.0066 | 0.8048 | 0.8131 |
61
+ | 0.2993 | 3.0 | 1104 | 0.4513 | 0.0066 | 0.8177 | 0.8253 |
62
+ | 0.1717 | 4.0 | 1472 | 0.5759 | 0.0066 | 0.8020 | 0.8123 |
63
+ | 0.1154 | 5.0 | 1840 | 0.6625 | 0.0066 | 0.8136 | 0.8214 |
64
 
65
 
66
  ### Framework versions
67
 
68
  - PEFT 0.14.0
69
+ - Transformers 4.49.0
70
  - Pytorch 2.5.1+cu124
71
+ - Datasets 3.3.1
72
  - Tokenizers 0.21.0
adapter_config.json CHANGED
@@ -12,7 +12,7 @@
12
  "layers_pattern": null,
13
  "layers_to_transform": null,
14
  "loftq_config": {},
15
- "lora_alpha": 8,
16
  "lora_bias": false,
17
  "lora_dropout": 0.05,
18
  "megatron_config": null,
@@ -26,10 +26,13 @@
26
  "rank_pattern": {},
27
  "revision": null,
28
  "target_modules": [
 
 
 
29
  "k_proj",
30
  "v_proj",
31
- "q_proj",
32
- "o_proj"
33
  ],
34
  "task_type": "SEQ_CLS",
35
  "use_dora": false,
 
12
  "layers_pattern": null,
13
  "layers_to_transform": null,
14
  "loftq_config": {},
15
+ "lora_alpha": 32,
16
  "lora_bias": false,
17
  "lora_dropout": 0.05,
18
  "megatron_config": null,
 
26
  "rank_pattern": {},
27
  "revision": null,
28
  "target_modules": [
29
+ "q_proj",
30
+ "down_proj",
31
+ "o_proj",
32
  "k_proj",
33
  "v_proj",
34
+ "up_proj",
35
+ "gate_proj"
36
  ],
37
  "task_type": "SEQ_CLS",
38
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9b2aedb4a849cb0cdfe72064eef4e3fd9ac7cc5c4f2212eda363357d9a2d31e8
3
- size 36779480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:39d49fde721f4ce82ef56835c3a74559bc9cad244bb81df60863a3bf7506474d
3
+ size 97356792