erfanzar commited on
Commit
7e35881
·
verified ·
1 Parent(s): dc92a50

Upload WhisperForConditionalGeneration

Browse files
Files changed (3) hide show
  1. config.json +6 -7
  2. generation_config.json +12 -28
  3. model.safetensors +3 -0
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "openai/whisper-large-v3",
3
  "activation_dropout": 0.0,
4
  "activation_function": "gelu",
5
  "apply_spec_augment": false,
@@ -9,7 +9,7 @@
9
  "attention_dropout": 0.0,
10
  "begin_suppress_tokens": [
11
  220,
12
- 50257
13
  ],
14
  "bos_token_id": 50257,
15
  "classifier_proj_size": 256,
@@ -17,7 +17,7 @@
17
  "decoder_attention_heads": 20,
18
  "decoder_ffn_dim": 5120,
19
  "decoder_layerdrop": 0.0,
20
- "decoder_layers": 32,
21
  "decoder_start_token_id": 50258,
22
  "dropout": 0.0,
23
  "encoder_attention_heads": 20,
@@ -33,17 +33,16 @@
33
  "mask_time_length": 10,
34
  "mask_time_min_masks": 2,
35
  "mask_time_prob": 0.05,
36
- "max_length": 448,
37
  "max_source_positions": 1500,
38
  "max_target_positions": 448,
39
  "median_filter_width": 7,
40
  "model_type": "whisper",
41
  "num_hidden_layers": 32,
42
  "num_mel_bins": 128,
43
- "pad_token_id": 50256,
44
  "scale_embedding": false,
45
- "torch_dtype": "float32",
46
- "transformers_version": "4.44.2",
47
  "use_cache": true,
48
  "use_weighted_layer_sum": false,
49
  "vocab_size": 51866
 
1
  {
2
+ "_name_or_path": "openai/whisper-large-v3-turbo",
3
  "activation_dropout": 0.0,
4
  "activation_function": "gelu",
5
  "apply_spec_augment": false,
 
9
  "attention_dropout": 0.0,
10
  "begin_suppress_tokens": [
11
  220,
12
+ 50256
13
  ],
14
  "bos_token_id": 50257,
15
  "classifier_proj_size": 256,
 
17
  "decoder_attention_heads": 20,
18
  "decoder_ffn_dim": 5120,
19
  "decoder_layerdrop": 0.0,
20
+ "decoder_layers": 4,
21
  "decoder_start_token_id": 50258,
22
  "dropout": 0.0,
23
  "encoder_attention_heads": 20,
 
33
  "mask_time_length": 10,
34
  "mask_time_min_masks": 2,
35
  "mask_time_prob": 0.05,
 
36
  "max_source_positions": 1500,
37
  "max_target_positions": 448,
38
  "median_filter_width": 7,
39
  "model_type": "whisper",
40
  "num_hidden_layers": 32,
41
  "num_mel_bins": 128,
42
+ "pad_token_id": 50257,
43
  "scale_embedding": false,
44
+ "torch_dtype": "float16",
45
+ "transformers_version": "4.44.0",
46
  "use_cache": true,
47
  "use_weighted_layer_sum": false,
48
  "vocab_size": 51866
generation_config.json CHANGED
@@ -1,44 +1,28 @@
1
  {
2
  "alignment_heads": [
3
  [
4
- 7,
5
- 0
6
- ],
7
- [
8
- 10,
9
- 17
10
- ],
11
- [
12
- 12,
13
- 18
14
  ],
15
  [
16
- 13,
17
- 12
18
  ],
19
  [
20
- 16,
21
- 1
22
  ],
23
  [
24
- 17,
25
- 14
26
  ],
27
  [
28
- 19,
29
  11
30
  ],
31
  [
32
- 21,
33
- 4
34
- ],
35
- [
36
- 24,
37
- 1
38
- ],
39
- [
40
- 25,
41
- 6
42
  ]
43
  ],
44
  "begin_suppress_tokens": [
@@ -261,5 +245,5 @@
261
  "transcribe": 50360,
262
  "translate": 50359
263
  },
264
- "transformers_version": "4.44.2"
265
  }
 
1
  {
2
  "alignment_heads": [
3
  [
4
+ 2,
5
+ 4
 
 
 
 
 
 
 
 
6
  ],
7
  [
8
+ 2,
9
+ 11
10
  ],
11
  [
12
+ 3,
13
+ 3
14
  ],
15
  [
16
+ 3,
17
+ 6
18
  ],
19
  [
20
+ 3,
21
  11
22
  ],
23
  [
24
+ 3,
25
+ 14
 
 
 
 
 
 
 
 
26
  ]
27
  ],
28
  "begin_suppress_tokens": [
 
245
  "transcribe": 50360,
246
  "translate": 50359
247
  },
248
+ "transformers_version": "4.44.0"
249
  }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:542566a422ae4f3fd23f1ba11add198fca01bbf82e66e6a2857b3f608b1eb9d1
3
+ size 1617824864