NourFakih commited on
Commit
cc5d1dd
·
verified ·
1 Parent(s): 3ae22ce

Model save

Browse files
README.md ADDED
@@ -0,0 +1,111 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: nlpconnect/vit-gpt2-image-captioning
4
+ tags:
5
+ - generated_from_trainer
6
+ metrics:
7
+ - rouge
8
+ model-index:
9
+ - name: image-captioning-Vit-GPT2-Flickr8k
10
+ results: []
11
+ ---
12
+
13
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
+ should probably proofread and complete it, then remove this comment. -->
15
+
16
+ # image-captioning-Vit-GPT2-Flickr8k
17
+
18
+ This model is a fine-tuned version of [nlpconnect/vit-gpt2-image-captioning](https://huggingface.co/nlpconnect/vit-gpt2-image-captioning) on an unknown dataset.
19
+ It achieves the following results on the evaluation set:
20
+ - Loss: 0.4624
21
+ - Rouge1: 38.4598
22
+ - Rouge2: 14.1356
23
+ - Rougel: 35.4001
24
+ - Rougelsum: 35.4044
25
+ - Gen Len: 12.1355
26
+
27
+ ## Model description
28
+
29
+ More information needed
30
+
31
+ ## Intended uses & limitations
32
+
33
+ More information needed
34
+
35
+ ## Training and evaluation data
36
+
37
+ More information needed
38
+
39
+ ## Training procedure
40
+
41
+ ### Training hyperparameters
42
+
43
+ The following hyperparameters were used during training:
44
+ - learning_rate: 5e-05
45
+ - train_batch_size: 4
46
+ - eval_batch_size: 4
47
+ - seed: 42
48
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
49
+ - lr_scheduler_type: linear
50
+ - num_epochs: 3.0
51
+
52
+ ### Training results
53
+
54
+ | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
55
+ |:-------------:|:-----:|:-----:|:---------------:|:-------:|:-------:|:-------:|:---------:|:-------:|
56
+ | 0.5495 | 0.06 | 500 | 0.4942 | 35.0813 | 11.7169 | 32.4184 | 32.4273 | 11.5738 |
57
+ | 0.4945 | 0.12 | 1000 | 0.4903 | 35.4868 | 12.037 | 32.835 | 32.8388 | 11.8682 |
58
+ | 0.4984 | 0.19 | 1500 | 0.4862 | 35.3878 | 11.996 | 32.8196 | 32.8268 | 12.0544 |
59
+ | 0.4783 | 0.25 | 2000 | 0.4808 | 36.1063 | 12.3478 | 33.4632 | 33.4783 | 11.3468 |
60
+ | 0.4736 | 0.31 | 2500 | 0.4772 | 35.9266 | 12.3362 | 33.5046 | 33.5103 | 11.1066 |
61
+ | 0.4685 | 0.37 | 3000 | 0.4708 | 36.9089 | 13.0915 | 34.2896 | 34.2995 | 11.4739 |
62
+ | 0.4687 | 0.43 | 3500 | 0.4704 | 36.1844 | 12.5731 | 33.4609 | 33.4733 | 11.9201 |
63
+ | 0.4709 | 0.49 | 4000 | 0.4696 | 36.1774 | 12.8262 | 33.3824 | 33.3814 | 12.1733 |
64
+ | 0.4575 | 0.56 | 4500 | 0.4675 | 37.4417 | 13.7581 | 34.5386 | 34.5523 | 12.6302 |
65
+ | 0.4484 | 0.62 | 5000 | 0.4662 | 36.6864 | 13.0727 | 33.9056 | 33.9339 | 12.6007 |
66
+ | 0.4507 | 0.68 | 5500 | 0.4656 | 36.5144 | 12.7924 | 34.0484 | 34.0759 | 11.4316 |
67
+ | 0.4445 | 0.74 | 6000 | 0.4628 | 37.0553 | 13.3404 | 34.4096 | 34.4153 | 12.3211 |
68
+ | 0.4557 | 0.8 | 6500 | 0.4594 | 37.3241 | 13.1468 | 34.45 | 34.4658 | 12.2522 |
69
+ | 0.4451 | 0.87 | 7000 | 0.4600 | 37.33 | 13.5726 | 34.6534 | 34.6635 | 12.0494 |
70
+ | 0.4381 | 0.93 | 7500 | 0.4588 | 37.6255 | 13.8048 | 34.817 | 34.8252 | 12.1347 |
71
+ | 0.4357 | 0.99 | 8000 | 0.4571 | 37.2088 | 13.4177 | 34.3316 | 34.3372 | 12.2670 |
72
+ | 0.3869 | 1.05 | 8500 | 0.4612 | 37.7054 | 13.683 | 34.9637 | 34.9821 | 11.3216 |
73
+ | 0.377 | 1.11 | 9000 | 0.4616 | 37.2701 | 13.2182 | 34.3249 | 34.3396 | 12.3221 |
74
+ | 0.3736 | 1.17 | 9500 | 0.4607 | 37.2101 | 13.1285 | 34.3812 | 34.3767 | 11.8274 |
75
+ | 0.3801 | 1.24 | 10000 | 0.4617 | 37.9963 | 13.7537 | 35.2402 | 35.2374 | 11.6079 |
76
+ | 0.3816 | 1.3 | 10500 | 0.4599 | 37.3247 | 13.619 | 34.6494 | 34.6538 | 12.2101 |
77
+ | 0.377 | 1.36 | 11000 | 0.4619 | 37.2827 | 13.4471 | 34.3588 | 34.3861 | 12.3911 |
78
+ | 0.3745 | 1.42 | 11500 | 0.4604 | 37.5469 | 13.3948 | 34.5403 | 34.5613 | 12.2747 |
79
+ | 0.3785 | 1.48 | 12000 | 0.4568 | 38.085 | 14.0087 | 35.0549 | 35.0564 | 12.3179 |
80
+ | 0.3675 | 1.54 | 12500 | 0.4587 | 37.6241 | 13.8529 | 34.7614 | 34.7853 | 11.8732 |
81
+ | 0.3731 | 1.61 | 13000 | 0.4554 | 38.4418 | 14.1464 | 35.6658 | 35.6502 | 11.4294 |
82
+ | 0.3731 | 1.67 | 13500 | 0.4548 | 37.9045 | 13.7524 | 34.9001 | 34.9092 | 12.1241 |
83
+ | 0.371 | 1.73 | 14000 | 0.4542 | 38.412 | 14.212 | 35.473 | 35.4781 | 12.1014 |
84
+ | 0.3615 | 1.79 | 14500 | 0.4551 | 38.0734 | 14.1066 | 35.1289 | 35.1552 | 12.1135 |
85
+ | 0.3687 | 1.85 | 15000 | 0.4550 | 38.1762 | 14.1402 | 35.288 | 35.2936 | 12.2255 |
86
+ | 0.3711 | 1.92 | 15500 | 0.4532 | 37.6439 | 13.611 | 34.7558 | 34.7601 | 12.1632 |
87
+ | 0.3685 | 1.98 | 16000 | 0.4515 | 38.5682 | 14.5305 | 35.552 | 35.5703 | 11.9162 |
88
+ | 0.3333 | 2.04 | 16500 | 0.4626 | 38.4527 | 14.4649 | 35.6252 | 35.6307 | 11.9506 |
89
+ | 0.3129 | 2.1 | 17000 | 0.4660 | 38.203 | 14.0699 | 35.1626 | 35.1595 | 12.3313 |
90
+ | 0.3155 | 2.16 | 17500 | 0.4674 | 37.8903 | 13.9159 | 34.9097 | 34.9101 | 12.4853 |
91
+ | 0.3134 | 2.22 | 18000 | 0.4644 | 38.1489 | 13.9448 | 35.0351 | 35.0351 | 11.9748 |
92
+ | 0.3167 | 2.29 | 18500 | 0.4653 | 37.8449 | 13.9106 | 34.7773 | 34.7854 | 12.5273 |
93
+ | 0.322 | 2.35 | 19000 | 0.4673 | 37.9832 | 14.0115 | 34.8505 | 34.8597 | 12.4680 |
94
+ | 0.312 | 2.41 | 19500 | 0.4641 | 38.4627 | 14.2528 | 35.4297 | 35.4377 | 11.9315 |
95
+ | 0.3173 | 2.47 | 20000 | 0.4654 | 38.1591 | 13.9126 | 35.1114 | 35.1042 | 12.4845 |
96
+ | 0.3081 | 2.53 | 20500 | 0.4640 | 38.6969 | 14.3244 | 35.6933 | 35.692 | 11.8932 |
97
+ | 0.3093 | 2.6 | 21000 | 0.4633 | 38.2944 | 14.103 | 35.2407 | 35.2629 | 11.8932 |
98
+ | 0.3154 | 2.66 | 21500 | 0.4637 | 38.0668 | 13.7427 | 35.0547 | 35.0585 | 12.1310 |
99
+ | 0.3096 | 2.72 | 22000 | 0.4630 | 38.3647 | 14.0445 | 35.2568 | 35.2511 | 12.2591 |
100
+ | 0.3101 | 2.78 | 22500 | 0.4627 | 38.6366 | 14.3013 | 35.4955 | 35.4956 | 12.2836 |
101
+ | 0.309 | 2.84 | 23000 | 0.4620 | 38.3486 | 14.0403 | 35.3173 | 35.3265 | 12.3281 |
102
+ | 0.312 | 2.9 | 23500 | 0.4623 | 38.423 | 14.0759 | 35.3766 | 35.3853 | 12.2208 |
103
+ | 0.3135 | 2.97 | 24000 | 0.4624 | 38.4598 | 14.1356 | 35.4001 | 35.4044 | 12.1355 |
104
+
105
+
106
+ ### Framework versions
107
+
108
+ - Transformers 4.39.3
109
+ - Pytorch 2.1.2
110
+ - Datasets 2.18.0
111
+ - Tokenizers 0.15.2
generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 50256,
3
+ "decoder_start_token_id": 50256,
4
+ "eos_token_id": 50256,
5
+ "pad_token_id": 50256,
6
+ "transformers_version": "4.39.3"
7
+ }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:940b21177b16a07f4df269b1c2d69f9be28b2024dbd0dae0f02ca0b68927773a
3
  size 956835520
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f7d99eda5e1c86aed8e773c732c4ed535f9fa5c7494dc2d817e4ab19dec725e3
3
  size 956835520
runs/Apr30_23-43-46_3e0cabee4dc0/events.out.tfevents.1714520677.3e0cabee4dc0.34.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9c78054d516e46e5db4f991d3bdffad0c94b18376f8a373af7359f90cda4b6de
3
- size 44624
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cde88082626623a0531a772b64b5d2af5832e5223dd15615b155e9812f29cf4e
3
+ size 44984