yasmineee commited on
Commit
692d088
·
verified ·
1 Parent(s): 2bdd93c

finetune-t5-base-on-opus100-Ar2En-with-lora

Browse files
Files changed (2) hide show
  1. README.md +4 -24
  2. adapter_model.safetensors +1 -1
README.md CHANGED
@@ -1,9 +1,6 @@
1
  ---
2
  base_model: UBC-NLP/AraT5v2-base-1024
3
  library_name: peft
4
- metrics:
5
- - bleu
6
- - rouge
7
  tags:
8
  - generated_from_trainer
9
  model-index:
@@ -14,16 +11,10 @@ model-index:
14
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
  should probably proofread and complete it, then remove this comment. -->
16
 
17
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/FinalProject_/T5/runs/vvxagyr8)
18
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/FinalProject_/T5/runs/vvxagyr8)
19
  # finetune-t5-base-on-opus100-Ar2En-with-lora
20
 
21
  This model is a fine-tuned version of [UBC-NLP/AraT5v2-base-1024](https://huggingface.co/UBC-NLP/AraT5v2-base-1024) on an unknown dataset.
22
- It achieves the following results on the evaluation set:
23
- - Loss: 3.7552
24
- - Bleu: 4.3018
25
- - Rouge: 0.2386
26
- - Gen Len: 10.572
27
 
28
  ## Model description
29
 
@@ -46,24 +37,13 @@ The following hyperparameters were used during training:
46
  - train_batch_size: 10
47
  - eval_batch_size: 10
48
  - seed: 42
 
 
49
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
50
  - lr_scheduler_type: linear
51
- - num_epochs: 7
52
  - mixed_precision_training: Native AMP
53
 
54
- ### Training results
55
-
56
- | Training Loss | Epoch | Step | Validation Loss | Bleu | Rouge | Gen Len |
57
- |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:-------:|
58
- | 6.6745 | 1.0 | 700 | 4.6813 | 3.2487 | 0.2249 | 10.726 |
59
- | 6.1243 | 2.0 | 1400 | 4.0666 | 3.3995 | 0.2273 | 10.0245 |
60
- | 5.3863 | 3.0 | 2100 | 3.9208 | 3.8728 | 0.2335 | 10.3965 |
61
- | 5.1275 | 4.0 | 2800 | 3.8485 | 3.9535 | 0.2331 | 10.5655 |
62
- | 4.975 | 5.0 | 3500 | 3.7971 | 3.9941 | 0.2318 | 10.572 |
63
- | 4.8991 | 6.0 | 4200 | 3.7639 | 4.0786 | 0.2349 | 10.6005 |
64
- | 4.857 | 7.0 | 4900 | 3.7552 | 4.3018 | 0.2386 | 10.572 |
65
-
66
-
67
  ### Framework versions
68
 
69
  - PEFT 0.12.0
 
1
  ---
2
  base_model: UBC-NLP/AraT5v2-base-1024
3
  library_name: peft
 
 
 
4
  tags:
5
  - generated_from_trainer
6
  model-index:
 
11
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
  should probably proofread and complete it, then remove this comment. -->
13
 
14
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/FinalProject_/T5/runs/fy2nx2eb)
 
15
  # finetune-t5-base-on-opus100-Ar2En-with-lora
16
 
17
  This model is a fine-tuned version of [UBC-NLP/AraT5v2-base-1024](https://huggingface.co/UBC-NLP/AraT5v2-base-1024) on an unknown dataset.
 
 
 
 
 
18
 
19
  ## Model description
20
 
 
37
  - train_batch_size: 10
38
  - eval_batch_size: 10
39
  - seed: 42
40
+ - gradient_accumulation_steps: 4
41
+ - total_train_batch_size: 40
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
  - lr_scheduler_type: linear
44
+ - num_epochs: 37
45
  - mixed_precision_training: Native AMP
46
 
 
 
 
 
 
 
 
 
 
 
 
 
 
47
  ### Framework versions
48
 
49
  - PEFT 0.12.0
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:635f7af0a43c505277d7a537a7ca262d7660d46f23d6959ba0399894db6d80d6
3
  size 4463560
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b2c34ad319211f06b61ac534e856be32916502097be79fdf8c973ddd610a303
3
  size 4463560