Upload WhisperForConditionalGeneration

#8
by bvkbharadwaj - opened
Files changed (3) hide show
  1. config.json +3 -2
  2. generation_config.json +37 -2
  3. model.safetensors +3 -0
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "openai/whisper-small",
3
  "activation_dropout": 0.0,
4
  "activation_function": "gelu",
5
  "apply_spec_augment": false,
@@ -37,6 +37,7 @@
37
  "max_length": 448,
38
  "max_source_positions": 1500,
39
  "max_target_positions": 448,
 
40
  "model_type": "whisper",
41
  "num_hidden_layers": 12,
42
  "num_mel_bins": 80,
@@ -44,7 +45,7 @@
44
  "scale_embedding": false,
45
  "suppress_tokens": [],
46
  "torch_dtype": "float32",
47
- "transformers_version": "4.30.2",
48
  "use_cache": true,
49
  "use_weighted_layer_sum": false,
50
  "vocab_size": 51865
 
1
  {
2
+ "_name_or_path": "bvkbharadwaj/whisper-small-sanskasr",
3
  "activation_dropout": 0.0,
4
  "activation_function": "gelu",
5
  "apply_spec_augment": false,
 
37
  "max_length": 448,
38
  "max_source_positions": 1500,
39
  "max_target_positions": 448,
40
+ "median_filter_width": 7,
41
  "model_type": "whisper",
42
  "num_hidden_layers": 12,
43
  "num_mel_bins": 80,
 
45
  "scale_embedding": false,
46
  "suppress_tokens": [],
47
  "torch_dtype": "float32",
48
+ "transformers_version": "4.37.2",
49
  "use_cache": true,
50
  "use_weighted_layer_sum": false,
51
  "vocab_size": 51865
generation_config.json CHANGED
@@ -1,4 +1,38 @@
1
  {
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  "begin_suppress_tokens": [
3
  220,
4
  50257
@@ -118,10 +152,11 @@
118
  "<|yo|>": 50325,
119
  "<|zh|>": 50260
120
  },
121
- "max_initial_timestamp_index": 1,
122
  "max_length": 448,
123
  "no_timestamps_token_id": 50363,
124
  "pad_token_id": 50257,
 
125
  "return_timestamps": false,
126
  "suppress_tokens": [
127
  1,
@@ -217,5 +252,5 @@
217
  "transcribe": 50359,
218
  "translate": 50358
219
  },
220
- "transformers_version": "4.30.2"
221
  }
 
1
  {
2
+ "alignment_heads": [
3
+ [
4
+ 3,
5
+ 1
6
+ ],
7
+ [
8
+ 4,
9
+ 2
10
+ ],
11
+ [
12
+ 4,
13
+ 3
14
+ ],
15
+ [
16
+ 4,
17
+ 7
18
+ ],
19
+ [
20
+ 5,
21
+ 1
22
+ ],
23
+ [
24
+ 5,
25
+ 2
26
+ ],
27
+ [
28
+ 5,
29
+ 4
30
+ ],
31
+ [
32
+ 5,
33
+ 6
34
+ ]
35
+ ],
36
  "begin_suppress_tokens": [
37
  220,
38
  50257
 
152
  "<|yo|>": 50325,
153
  "<|zh|>": 50260
154
  },
155
+ "max_initial_timestamp_index": 50,
156
  "max_length": 448,
157
  "no_timestamps_token_id": 50363,
158
  "pad_token_id": 50257,
159
+ "prev_sot_token_id": 50361,
160
  "return_timestamps": false,
161
  "suppress_tokens": [
162
  1,
 
252
  "transcribe": 50359,
253
  "translate": 50358
254
  },
255
+ "transformers_version": "4.37.2"
256
  }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8280fe5aefb0d278aabb2ff93c0bc93a9aac6d91ba4c1430566b37c775a62630
3
+ size 966995080