swaroopajit commited on
Commit
6106c0c
1 Parent(s): 43d3c2d

End of training

Browse files
Files changed (5) hide show
  1. README.md +8 -8
  2. config.json +1 -1
  3. generation_config.json +1 -1
  4. pytorch_model.bin +1 -1
  5. training_args.bin +1 -1
README.md CHANGED
@@ -13,7 +13,7 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  # git-base-fashion
15
 
16
- This model is a fine-tuned version of [microsoft/git-base](https://huggingface.co/microsoft/git-base) on [this fashion caption dataset].
17
 
18
  ## Model description
19
 
@@ -21,7 +21,7 @@ More information needed
21
 
22
  ## Intended uses & limitations
23
 
24
- This is intended to create captions for clothing.
25
 
26
  ## Training and evaluation data
27
 
@@ -32,15 +32,15 @@ More information needed
32
  ### Training hyperparameters
33
 
34
  The following hyperparameters were used during training:
35
- - learning_rate: 3e-05
36
- - train_batch_size: 2
37
- - eval_batch_size: 2
38
  - seed: 42
39
  - gradient_accumulation_steps: 8
40
- - total_train_batch_size: 16
41
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
42
  - lr_scheduler_type: linear
43
- - num_epochs: 5
44
 
45
  ### Training results
46
 
@@ -48,7 +48,7 @@ The following hyperparameters were used during training:
48
 
49
  ### Framework versions
50
 
51
- - Transformers 4.32.1
52
  - Pytorch 2.0.1+cu118
53
  - Datasets 2.14.4
54
  - Tokenizers 0.13.3
 
13
 
14
  # git-base-fashion
15
 
16
+ This model is a fine-tuned version of [microsoft/git-base](https://huggingface.co/microsoft/git-base) on an unknown dataset.
17
 
18
  ## Model description
19
 
 
21
 
22
  ## Intended uses & limitations
23
 
24
+ More information needed
25
 
26
  ## Training and evaluation data
27
 
 
32
  ### Training hyperparameters
33
 
34
  The following hyperparameters were used during training:
35
+ - learning_rate: 5e-05
36
+ - train_batch_size: 4
37
+ - eval_batch_size: 4
38
  - seed: 42
39
  - gradient_accumulation_steps: 8
40
+ - total_train_batch_size: 32
41
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
42
  - lr_scheduler_type: linear
43
+ - num_epochs: 30
44
 
45
  ### Training results
46
 
 
48
 
49
  ### Framework versions
50
 
51
+ - Transformers 4.33.0
52
  - Pytorch 2.0.1+cu118
53
  - Datasets 2.14.4
54
  - Tokenizers 0.13.3
config.json CHANGED
@@ -22,7 +22,7 @@
22
  "position_embedding_type": "absolute",
23
  "tie_word_embeddings": false,
24
  "torch_dtype": "float32",
25
- "transformers_version": "4.32.1",
26
  "use_cache": true,
27
  "vision_config": {
28
  "dropout": 0.0,
 
22
  "position_embedding_type": "absolute",
23
  "tie_word_embeddings": false,
24
  "torch_dtype": "float32",
25
+ "transformers_version": "4.33.0",
26
  "use_cache": true,
27
  "vision_config": {
28
  "dropout": 0.0,
generation_config.json CHANGED
@@ -3,5 +3,5 @@
3
  "bos_token_id": 101,
4
  "eos_token_id": 102,
5
  "pad_token_id": 0,
6
- "transformers_version": "4.32.1"
7
  }
 
3
  "bos_token_id": 101,
4
  "eos_token_id": 102,
5
  "pad_token_id": 0,
6
+ "transformers_version": "4.33.0"
7
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:14e0b4312d07709c285c55f786a1cbc0057afa1db45822673dc6ac779c6516c6
3
  size 706584273
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d8ec05432efbf811011fba1a16e73c6e9da932d24f62f30e94601f604c38c434
3
  size 706584273
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d0a8791c693f2464ff5e423ab08b9e9311396fc61d4886f5d92328fa8287c3e9
3
  size 4027
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9006efc26cd0b9c238c4367f198ce7c7163fb0b5f0f14998f56ff72f0da54a06
3
  size 4027