ghost613 commited on
Commit
90ed6eb
·
verified ·
1 Parent(s): e3eb082

Training in progress, step 700

Browse files
config.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "vocab_size": 51865,
3
- "num_mel_bins": 80,
4
  "d_model": 1280,
5
- "encoder_layers": 32,
6
  "encoder_attention_heads": 20,
7
- "decoder_layers": 6,
8
  "decoder_attention_heads": 20,
9
  "decoder_ffn_dim": 5120,
10
  "encoder_ffn_dim": 5120,
@@ -34,7 +34,7 @@
34
  "output_hidden_states": false,
35
  "output_attentions": false,
36
  "torchscript": false,
37
- "torch_dtype": null,
38
  "use_bfloat16": false,
39
  "tf_legacy_loss": false,
40
  "pruned_heads": {},
@@ -73,7 +73,9 @@
73
  220,
74
  50256
75
  ],
76
- "architectures": null,
 
 
77
  "finetuning_task": null,
78
  "id2label": {
79
  "0": "LABEL_0",
@@ -85,17 +87,17 @@
85
  },
86
  "tokenizer_class": null,
87
  "prefix": null,
88
- "bos_token_id": 50256,
89
- "pad_token_id": 50256,
90
- "eos_token_id": 50256,
91
  "sep_token_id": null,
92
- "decoder_start_token_id": 50257,
93
  "task_specific_params": null,
94
  "problem_type": null,
95
- "_name_or_path": "",
96
  "transformers_version": "4.42.3",
97
  "model_type": "syllable_whisper",
98
- "encoder_type": "embedding",
99
  "freeze_wav2vec": false,
100
  "phoneme_vocab_size": 55,
101
  "syllable_vocab_size": 1207,
@@ -111,13 +113,13 @@
111
  "num_layer_stacks": 1,
112
  "noise_std": 0.01,
113
  "apply_noise_to_original": false,
114
- "original_encoder_layers": 32,
115
  "unfreeze_top_k_layers": 0,
116
- "original_decoder_layers": 6,
117
- "reduce_encoder_layers": false,
118
  "reduce_decoder_layers": false,
119
- "encoder_layers_to_keep": 32,
120
- "decoder_layers_to_keep": 6,
121
  "decoder_dropout_rate": 0.0,
122
  "decoder_dropout_type": "standard",
123
  "apply_dropout_to_stacked_only": false
 
1
  {
2
+ "vocab_size": 51866,
3
+ "num_mel_bins": 128,
4
  "d_model": 1280,
5
+ "encoder_layers": 4,
6
  "encoder_attention_heads": 20,
7
+ "decoder_layers": 4,
8
  "decoder_attention_heads": 20,
9
  "decoder_ffn_dim": 5120,
10
  "encoder_ffn_dim": 5120,
 
34
  "output_hidden_states": false,
35
  "output_attentions": false,
36
  "torchscript": false,
37
+ "torch_dtype": "float16",
38
  "use_bfloat16": false,
39
  "tf_legacy_loss": false,
40
  "pruned_heads": {},
 
73
  220,
74
  50256
75
  ],
76
+ "architectures": [
77
+ "WhisperForConditionalGeneration"
78
+ ],
79
  "finetuning_task": null,
80
  "id2label": {
81
  "0": "LABEL_0",
 
87
  },
88
  "tokenizer_class": null,
89
  "prefix": null,
90
+ "bos_token_id": 50257,
91
+ "pad_token_id": 50257,
92
+ "eos_token_id": 50257,
93
  "sep_token_id": null,
94
+ "decoder_start_token_id": 0,
95
  "task_specific_params": null,
96
  "problem_type": null,
97
+ "_name_or_path": "openai/whisper-large-v3-turbo",
98
  "transformers_version": "4.42.3",
99
  "model_type": "syllable_whisper",
100
+ "encoder_type": "whisper",
101
  "freeze_wav2vec": false,
102
  "phoneme_vocab_size": 55,
103
  "syllable_vocab_size": 1207,
 
113
  "num_layer_stacks": 1,
114
  "noise_std": 0.01,
115
  "apply_noise_to_original": false,
116
+ "original_encoder_layers": 4,
117
  "unfreeze_top_k_layers": 0,
118
+ "original_decoder_layers": 4,
119
+ "reduce_encoder_layers": true,
120
  "reduce_decoder_layers": false,
121
+ "encoder_layers_to_keep": 4,
122
+ "decoder_layers_to_keep": 4,
123
  "decoder_dropout_rate": 0.0,
124
  "decoder_dropout_type": "standard",
125
  "apply_dropout_to_stacked_only": false
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9d44eaf9150b296406c717ac15317030c53f088307f8eb002be6de766ab54c33
3
- size 1305713496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:046677bd41d5fded7a1d310f00d8e6d817b32836d46d60e319645601a253d03e
3
+ size 1229405408
runs/Jul18_07-19-50_whisper-finetune.asia-northeast3-a.c.sabina-404403.internal/events.out.tfevents.1752823193.whisper-finetune.asia-northeast3-a.c.sabina-404403.internal.52480.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f342508fecd68b70602b6439e595e2212598ec3fbbfb6cc67f4c4d3d5d24437
3
+ size 6359
runs/Jul18_07-19-50_whisper-finetune.asia-northeast3-a.c.sabina-404403.internal/events.out.tfevents.1752825377.whisper-finetune.asia-northeast3-a.c.sabina-404403.internal.52480.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7bc932387e3c1ac1c6ca4a78c60060ac9f96ce538e97822234574189c708b25a
3
+ size 4624
runs/Jul18_07-19-50_whisper-finetune.asia-northeast3-a.c.sabina-404403.internal/events.out.tfevents.1752825377.whisper-finetune.asia-northeast3-a.c.sabina-404403.internal.52480.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:548034074c46f5b887ec8752a750a7339f465ba42a73c0a335afc2f7998075a6
3
+ size 88
runs/Jul18_07-19-50_whisper-finetune.asia-northeast3-a.c.sabina-404403.internal/events.out.tfevents.1752825409.whisper-finetune.asia-northeast3-a.c.sabina-404403.internal.52480.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8170bbd0d45af728c78a3d7f4530d8f66e4607704e34744ed1d66aed1dae504
3
+ size 4624
runs/Jul18_07-19-50_whisper-finetune.asia-northeast3-a.c.sabina-404403.internal/events.out.tfevents.1752825409.whisper-finetune.asia-northeast3-a.c.sabina-404403.internal.52480.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7fceb8c18317377459342f71ae1b5547ddc2cb1e782f6e0fc72731564b25b739
3
+ size 88
runs/Jul18_07-19-50_whisper-finetune.asia-northeast3-a.c.sabina-404403.internal/events.out.tfevents.1752825420.whisper-finetune.asia-northeast3-a.c.sabina-404403.internal.52480.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:96afe3c8c3feb17ef3fc8d96e55f9e0e52fe433729cf8c5825cb5c3742938998
3
+ size 4624
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3edea9c945b057438824151077ed6160f9ce4fd1768b15a3b0e24084ed19cdc8
3
  size 4923
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:177ce8ad5379038863796f3962a5fdb6b94aeb0f38a89d4ae680f491ddbef2f9
3
  size 4923