From d13d0e4bec77ec497c0fbde187bdc71a6f913388 Mon Sep 17 00:00:00 2001 From: Arthur Zucker Date: Tue, 4 Oct 2022 14:58:21 +0000 Subject: [PATCH] Upload config --- config.json | 144 ++++++++++++++++++++++++++-------------------------- 1 file changed, 73 insertions(+), 71 deletions(-) diff --git a/config.json b/config.json index 3efc49f..718cdee 100644 --- a/config.json +++ b/config.json @@ -1,11 +1,11 @@ { - "_name_or_path": "openai/whisper-tiny.en", "activation_dropout": 0.0, "activation_function": "gelu", - "architectures": [ - "WhisperForConditionalGeneration" - ], "attention_dropout": 0.0, + "begin_suppress_tokens": [ + 220, + 50256 + ], "bos_token_id": 50257, "d_model": 384, "decoder_attention_heads": 6, @@ -19,14 +19,16 @@ "encoder_layerdrop": 0.0, "encoder_layers": 4, "eos_token_id": 50256, - "feature_size": 1, "init_std": 0.02, - "input_channels": 1, "is_encoder_decoder": true, "max_source_positions": 1500, "max_target_positions": 448, "model_type": "whisper", - "non_speech_tokens": [ + "num_hidden_layers": 4, + "num_mel_bins": 80, + "pad_token_id": 0, + "scale_embedding": false, + "suppress_tokens": [ 1, 2, 6, @@ -52,71 +54,71 @@ 91, 92, 93, - 359, - 503, - 522, - 542, - 873, - 893, - 902, - 918, - 922, - 931, - 1350, - 1853, - 1982, - 2460, - 2627, - 3246, - 3253, - 3268, - 3536, - 3846, - 3961, - 4183, - 4667, - 6585, - 6647, - 7273, - 9061, - 9383, - 10428, - 10929, - 11938, - 12033, - 12331, - 12562, - 13793, - 14157, - 14635, - 15265, - 15618, - 16553, - 16604, - 18362, - 18956, - 20075, - 21675, - 22520, - 26130, - 26161, - 26435, - 28279, - 29464, - 31650, - 32302, - 32470, - 36865, - 42863, - 47425, - 49870, - 50254 + 357, + 366, + 438, + 532, + 685, + 705, + 796, + 930, + 1058, + 1220, + 1267, + 1279, + 1303, + 1343, + 1377, + 1391, + 1635, + 1782, + 1875, + 2162, + 2361, + 2488, + 3467, + 4008, + 4211, + 4600, + 4808, + 5299, + 5855, + 6329, + 7203, + 9609, + 9959, + 10563, + 10786, + 11420, + 11709, + 11907, + 13163, + 13697, + 13700, + 14808, + 15306, + 16410, + 16791, + 17992, + 19203, + 19510, + 20724, + 22305, + 22935, + 27007, + 30109, + 30420, + 33409, + 34949, + 40283, + 40493, + 40549, + 47282, + 49146, + 50257, + 50359, + 50360 ], - "num_hidden_layers": 4, - "num_mel_bins": 80, - "pad_token_id": 0, - "scale_embedding": false, - "torch_dtype": "float32", "transformers_version": "4.23.0.dev0", "use_cache": true, "vocab_size": 51864