lvcalucioli commited on
Commit
ca0e8d1
1 Parent(s): 71bd856

llamantino7b_question_answering_finetuining

Browse files
README.md CHANGED
@@ -18,7 +18,7 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  This model is a fine-tuned version of [swap-uniba/LLaMAntino-2-7b-hf-ITA](https://huggingface.co/swap-uniba/LLaMAntino-2-7b-hf-ITA) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 0.2277
22
 
23
  ## Model description
24
 
@@ -39,27 +39,19 @@ More information needed
39
  The following hyperparameters were used during training:
40
  - learning_rate: 0.0002
41
  - train_batch_size: 4
42
- - eval_batch_size: 4
43
  - seed: 42
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: linear
46
  - lr_scheduler_warmup_ratio: 0.03
47
- - num_epochs: 10
48
 
49
  ### Training results
50
 
51
  | Training Loss | Epoch | Step | Validation Loss |
52
  |:-------------:|:-----:|:----:|:---------------:|
53
- | 1.6498 | 1.0 | 90 | 0.0351 |
54
- | 1.1895 | 2.0 | 180 | 0.0237 |
55
- | 0.8519 | 3.0 | 270 | 0.0219 |
56
- | 0.5762 | 4.0 | 360 | 0.0155 |
57
- | 0.3852 | 5.0 | 450 | 0.0987 |
58
- | 0.2445 | 6.0 | 540 | 0.1373 |
59
- | 0.1554 | 7.0 | 630 | 0.2655 |
60
- | 0.0989 | 8.0 | 720 | 0.0678 |
61
- | 0.0607 | 9.0 | 810 | 0.1935 |
62
- | 0.0397 | 10.0 | 900 | 0.2277 |
63
 
64
 
65
  ### Framework versions
 
18
 
19
  This model is a fine-tuned version of [swap-uniba/LLaMAntino-2-7b-hf-ITA](https://huggingface.co/swap-uniba/LLaMAntino-2-7b-hf-ITA) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 1.4340
22
 
23
  ## Model description
24
 
 
39
  The following hyperparameters were used during training:
40
  - learning_rate: 0.0002
41
  - train_batch_size: 4
42
+ - eval_batch_size: 1
43
  - seed: 42
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: linear
46
  - lr_scheduler_warmup_ratio: 0.03
47
+ - num_epochs: 2
48
 
49
  ### Training results
50
 
51
  | Training Loss | Epoch | Step | Validation Loss |
52
  |:-------------:|:-----:|:----:|:---------------:|
53
+ | 1.4152 | 1.0 | 3 | 1.4624 |
54
+ | 1.3209 | 2.0 | 6 | 1.4340 |
 
 
 
 
 
 
 
 
55
 
56
 
57
  ### Framework versions
adapter_config.json CHANGED
@@ -19,8 +19,8 @@
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
- "q_proj",
23
- "v_proj"
24
  ],
25
  "task_type": "CAUSAL_LM",
26
  "use_rslora": false
 
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
+ "v_proj",
23
+ "q_proj"
24
  ],
25
  "task_type": "CAUSAL_LM",
26
  "use_rslora": false
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5cca94709411e6f43a71c53735a3796e0415d3643f0c62de07cddeba3c06cbea
3
  size 134235048
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eb8e6c929e9d22e8c7f5910f58212d2876b080e3e4b8ae99d32e569eaba24e59
3
  size 134235048
tokenizer.json CHANGED
@@ -1,6 +1,11 @@
1
  {
2
  "version": "1.0",
3
- "truncation": null,
 
 
 
 
 
4
  "padding": null,
5
  "added_tokens": [
6
  {
 
1
  {
2
  "version": "1.0",
3
+ "truncation": {
4
+ "direction": "Right",
5
+ "max_length": 1024,
6
+ "strategy": "LongestFirst",
7
+ "stride": 0
8
+ },
9
  "padding": null,
10
  "added_tokens": [
11
  {
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:de3f6ccc8a4fe4d9d7af9f98b6ced1b4b4aa7db3a8e15dde4f62a932939aaecf
3
  size 4411
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:88218a9037911cdc6f26e5076c993ebc2d7b2abe368fd7f6168160042fa72cb1
3
  size 4411