Model save
Browse files
README.md
ADDED
@@ -0,0 +1,111 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
base_model: nlpconnect/vit-gpt2-image-captioning
|
4 |
+
tags:
|
5 |
+
- generated_from_trainer
|
6 |
+
metrics:
|
7 |
+
- rouge
|
8 |
+
model-index:
|
9 |
+
- name: image-captioning-Vit-GPT2-Flickr8k
|
10 |
+
results: []
|
11 |
+
---
|
12 |
+
|
13 |
+
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
14 |
+
should probably proofread and complete it, then remove this comment. -->
|
15 |
+
|
16 |
+
# image-captioning-Vit-GPT2-Flickr8k
|
17 |
+
|
18 |
+
This model is a fine-tuned version of [nlpconnect/vit-gpt2-image-captioning](https://huggingface.co/nlpconnect/vit-gpt2-image-captioning) on an unknown dataset.
|
19 |
+
It achieves the following results on the evaluation set:
|
20 |
+
- Loss: 0.4624
|
21 |
+
- Rouge1: 38.4598
|
22 |
+
- Rouge2: 14.1356
|
23 |
+
- Rougel: 35.4001
|
24 |
+
- Rougelsum: 35.4044
|
25 |
+
- Gen Len: 12.1355
|
26 |
+
|
27 |
+
## Model description
|
28 |
+
|
29 |
+
More information needed
|
30 |
+
|
31 |
+
## Intended uses & limitations
|
32 |
+
|
33 |
+
More information needed
|
34 |
+
|
35 |
+
## Training and evaluation data
|
36 |
+
|
37 |
+
More information needed
|
38 |
+
|
39 |
+
## Training procedure
|
40 |
+
|
41 |
+
### Training hyperparameters
|
42 |
+
|
43 |
+
The following hyperparameters were used during training:
|
44 |
+
- learning_rate: 5e-05
|
45 |
+
- train_batch_size: 4
|
46 |
+
- eval_batch_size: 4
|
47 |
+
- seed: 42
|
48 |
+
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
49 |
+
- lr_scheduler_type: linear
|
50 |
+
- num_epochs: 3.0
|
51 |
+
|
52 |
+
### Training results
|
53 |
+
|
54 |
+
| Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
|
55 |
+
|:-------------:|:-----:|:-----:|:---------------:|:-------:|:-------:|:-------:|:---------:|:-------:|
|
56 |
+
| 0.5495 | 0.06 | 500 | 0.4942 | 35.0813 | 11.7169 | 32.4184 | 32.4273 | 11.5738 |
|
57 |
+
| 0.4945 | 0.12 | 1000 | 0.4903 | 35.4868 | 12.037 | 32.835 | 32.8388 | 11.8682 |
|
58 |
+
| 0.4984 | 0.19 | 1500 | 0.4862 | 35.3878 | 11.996 | 32.8196 | 32.8268 | 12.0544 |
|
59 |
+
| 0.4783 | 0.25 | 2000 | 0.4808 | 36.1063 | 12.3478 | 33.4632 | 33.4783 | 11.3468 |
|
60 |
+
| 0.4736 | 0.31 | 2500 | 0.4772 | 35.9266 | 12.3362 | 33.5046 | 33.5103 | 11.1066 |
|
61 |
+
| 0.4685 | 0.37 | 3000 | 0.4708 | 36.9089 | 13.0915 | 34.2896 | 34.2995 | 11.4739 |
|
62 |
+
| 0.4687 | 0.43 | 3500 | 0.4704 | 36.1844 | 12.5731 | 33.4609 | 33.4733 | 11.9201 |
|
63 |
+
| 0.4709 | 0.49 | 4000 | 0.4696 | 36.1774 | 12.8262 | 33.3824 | 33.3814 | 12.1733 |
|
64 |
+
| 0.4575 | 0.56 | 4500 | 0.4675 | 37.4417 | 13.7581 | 34.5386 | 34.5523 | 12.6302 |
|
65 |
+
| 0.4484 | 0.62 | 5000 | 0.4662 | 36.6864 | 13.0727 | 33.9056 | 33.9339 | 12.6007 |
|
66 |
+
| 0.4507 | 0.68 | 5500 | 0.4656 | 36.5144 | 12.7924 | 34.0484 | 34.0759 | 11.4316 |
|
67 |
+
| 0.4445 | 0.74 | 6000 | 0.4628 | 37.0553 | 13.3404 | 34.4096 | 34.4153 | 12.3211 |
|
68 |
+
| 0.4557 | 0.8 | 6500 | 0.4594 | 37.3241 | 13.1468 | 34.45 | 34.4658 | 12.2522 |
|
69 |
+
| 0.4451 | 0.87 | 7000 | 0.4600 | 37.33 | 13.5726 | 34.6534 | 34.6635 | 12.0494 |
|
70 |
+
| 0.4381 | 0.93 | 7500 | 0.4588 | 37.6255 | 13.8048 | 34.817 | 34.8252 | 12.1347 |
|
71 |
+
| 0.4357 | 0.99 | 8000 | 0.4571 | 37.2088 | 13.4177 | 34.3316 | 34.3372 | 12.2670 |
|
72 |
+
| 0.3869 | 1.05 | 8500 | 0.4612 | 37.7054 | 13.683 | 34.9637 | 34.9821 | 11.3216 |
|
73 |
+
| 0.377 | 1.11 | 9000 | 0.4616 | 37.2701 | 13.2182 | 34.3249 | 34.3396 | 12.3221 |
|
74 |
+
| 0.3736 | 1.17 | 9500 | 0.4607 | 37.2101 | 13.1285 | 34.3812 | 34.3767 | 11.8274 |
|
75 |
+
| 0.3801 | 1.24 | 10000 | 0.4617 | 37.9963 | 13.7537 | 35.2402 | 35.2374 | 11.6079 |
|
76 |
+
| 0.3816 | 1.3 | 10500 | 0.4599 | 37.3247 | 13.619 | 34.6494 | 34.6538 | 12.2101 |
|
77 |
+
| 0.377 | 1.36 | 11000 | 0.4619 | 37.2827 | 13.4471 | 34.3588 | 34.3861 | 12.3911 |
|
78 |
+
| 0.3745 | 1.42 | 11500 | 0.4604 | 37.5469 | 13.3948 | 34.5403 | 34.5613 | 12.2747 |
|
79 |
+
| 0.3785 | 1.48 | 12000 | 0.4568 | 38.085 | 14.0087 | 35.0549 | 35.0564 | 12.3179 |
|
80 |
+
| 0.3675 | 1.54 | 12500 | 0.4587 | 37.6241 | 13.8529 | 34.7614 | 34.7853 | 11.8732 |
|
81 |
+
| 0.3731 | 1.61 | 13000 | 0.4554 | 38.4418 | 14.1464 | 35.6658 | 35.6502 | 11.4294 |
|
82 |
+
| 0.3731 | 1.67 | 13500 | 0.4548 | 37.9045 | 13.7524 | 34.9001 | 34.9092 | 12.1241 |
|
83 |
+
| 0.371 | 1.73 | 14000 | 0.4542 | 38.412 | 14.212 | 35.473 | 35.4781 | 12.1014 |
|
84 |
+
| 0.3615 | 1.79 | 14500 | 0.4551 | 38.0734 | 14.1066 | 35.1289 | 35.1552 | 12.1135 |
|
85 |
+
| 0.3687 | 1.85 | 15000 | 0.4550 | 38.1762 | 14.1402 | 35.288 | 35.2936 | 12.2255 |
|
86 |
+
| 0.3711 | 1.92 | 15500 | 0.4532 | 37.6439 | 13.611 | 34.7558 | 34.7601 | 12.1632 |
|
87 |
+
| 0.3685 | 1.98 | 16000 | 0.4515 | 38.5682 | 14.5305 | 35.552 | 35.5703 | 11.9162 |
|
88 |
+
| 0.3333 | 2.04 | 16500 | 0.4626 | 38.4527 | 14.4649 | 35.6252 | 35.6307 | 11.9506 |
|
89 |
+
| 0.3129 | 2.1 | 17000 | 0.4660 | 38.203 | 14.0699 | 35.1626 | 35.1595 | 12.3313 |
|
90 |
+
| 0.3155 | 2.16 | 17500 | 0.4674 | 37.8903 | 13.9159 | 34.9097 | 34.9101 | 12.4853 |
|
91 |
+
| 0.3134 | 2.22 | 18000 | 0.4644 | 38.1489 | 13.9448 | 35.0351 | 35.0351 | 11.9748 |
|
92 |
+
| 0.3167 | 2.29 | 18500 | 0.4653 | 37.8449 | 13.9106 | 34.7773 | 34.7854 | 12.5273 |
|
93 |
+
| 0.322 | 2.35 | 19000 | 0.4673 | 37.9832 | 14.0115 | 34.8505 | 34.8597 | 12.4680 |
|
94 |
+
| 0.312 | 2.41 | 19500 | 0.4641 | 38.4627 | 14.2528 | 35.4297 | 35.4377 | 11.9315 |
|
95 |
+
| 0.3173 | 2.47 | 20000 | 0.4654 | 38.1591 | 13.9126 | 35.1114 | 35.1042 | 12.4845 |
|
96 |
+
| 0.3081 | 2.53 | 20500 | 0.4640 | 38.6969 | 14.3244 | 35.6933 | 35.692 | 11.8932 |
|
97 |
+
| 0.3093 | 2.6 | 21000 | 0.4633 | 38.2944 | 14.103 | 35.2407 | 35.2629 | 11.8932 |
|
98 |
+
| 0.3154 | 2.66 | 21500 | 0.4637 | 38.0668 | 13.7427 | 35.0547 | 35.0585 | 12.1310 |
|
99 |
+
| 0.3096 | 2.72 | 22000 | 0.4630 | 38.3647 | 14.0445 | 35.2568 | 35.2511 | 12.2591 |
|
100 |
+
| 0.3101 | 2.78 | 22500 | 0.4627 | 38.6366 | 14.3013 | 35.4955 | 35.4956 | 12.2836 |
|
101 |
+
| 0.309 | 2.84 | 23000 | 0.4620 | 38.3486 | 14.0403 | 35.3173 | 35.3265 | 12.3281 |
|
102 |
+
| 0.312 | 2.9 | 23500 | 0.4623 | 38.423 | 14.0759 | 35.3766 | 35.3853 | 12.2208 |
|
103 |
+
| 0.3135 | 2.97 | 24000 | 0.4624 | 38.4598 | 14.1356 | 35.4001 | 35.4044 | 12.1355 |
|
104 |
+
|
105 |
+
|
106 |
+
### Framework versions
|
107 |
+
|
108 |
+
- Transformers 4.39.3
|
109 |
+
- Pytorch 2.1.2
|
110 |
+
- Datasets 2.18.0
|
111 |
+
- Tokenizers 0.15.2
|
generation_config.json
ADDED
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token_id": 50256,
|
3 |
+
"decoder_start_token_id": 50256,
|
4 |
+
"eos_token_id": 50256,
|
5 |
+
"pad_token_id": 50256,
|
6 |
+
"transformers_version": "4.39.3"
|
7 |
+
}
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 956835520
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f7d99eda5e1c86aed8e773c732c4ed535f9fa5c7494dc2d817e4ab19dec725e3
|
3 |
size 956835520
|
runs/Apr30_23-43-46_3e0cabee4dc0/events.out.tfevents.1714520677.3e0cabee4dc0.34.0
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cde88082626623a0531a772b64b5d2af5832e5223dd15615b155e9812f29cf4e
|
3 |
+
size 44984
|