fairseq tuned
This commit is contained in:
parent
27503d5d38
commit
3d0bc0ac66
19
config.json
19
config.json
|
@ -1,12 +1,11 @@
|
|||
{
|
||||
"_name_or_path": "./pretrained/checkpoint_0.60",
|
||||
"activation_dropout": 0.1,
|
||||
"apply_spec_augment": true,
|
||||
"architectures": [
|
||||
"Wav2Vec2ForCTC"
|
||||
],
|
||||
"attention_dropout": 0.1,
|
||||
"bos_token_id": 1,
|
||||
"bos_token_id": 0,
|
||||
"conv_bias": false,
|
||||
"conv_dim": [
|
||||
512,
|
||||
|
@ -35,17 +34,17 @@
|
|||
2,
|
||||
2
|
||||
],
|
||||
"ctc_loss_reduction": "mean",
|
||||
"ctc_zero_infinity": true,
|
||||
"ctc_loss_reduction": "sum",
|
||||
"ctc_zero_infinity": false,
|
||||
"do_stable_layer_norm": false,
|
||||
"eos_token_id": 2,
|
||||
"feat_extract_activation": "gelu",
|
||||
"feat_extract_norm": "group",
|
||||
"feat_proj_dropout": 0.1,
|
||||
"final_dropout": 0.1,
|
||||
"gradient_checkpointing": true,
|
||||
"gradient_checkpointing": false,
|
||||
"hidden_act": "gelu",
|
||||
"hidden_dropout": 0.05,
|
||||
"hidden_dropout": 0.1,
|
||||
"hidden_size": 768,
|
||||
"initializer_range": 0.02,
|
||||
"intermediate_size": 3072,
|
||||
|
@ -54,14 +53,14 @@
|
|||
"mask_feature_length": 10,
|
||||
"mask_feature_prob": 0.0,
|
||||
"mask_time_length": 10,
|
||||
"mask_time_prob": 0.5,
|
||||
"mask_time_prob": 0.05,
|
||||
"model_type": "wav2vec2",
|
||||
"num_attention_heads": 12,
|
||||
"num_conv_pos_embedding_groups": 16,
|
||||
"num_conv_pos_embeddings": 128,
|
||||
"num_feat_extract_layers": 7,
|
||||
"num_hidden_layers": 12,
|
||||
"pad_token_id": 36,
|
||||
"transformers_version": "4.5.0.dev0",
|
||||
"vocab_size": 37
|
||||
"pad_token_id": 1,
|
||||
"transformers_version": "4.4.2",
|
||||
"vocab_size": 39
|
||||
}
|
||||
|
|
|
@ -1,3 +1,3 @@
|
|||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:ecb5a521af5105ca67abd19a841c08e97b071a8b69e186b06b25633c6fc36804
|
||||
size 377690860
|
||||
oid sha256:7ac7eaad990b15315d1772928ea15b9c77d2e259311b5189f9772b04da157294
|
||||
size 377691502
|
||||
|
|
|
@ -1 +1 @@
|
|||
{"bos_token": "<s>", "eos_token": "</s>", "unk_token": "[UNK]", "pad_token": "[PAD]"}
|
||||
{"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>"}
|
|
@ -1 +1 @@
|
|||
{"unk_token": "[UNK]", "bos_token": "<s>", "eos_token": "</s>", "pad_token": "[PAD]", "do_lower_case": false, "word_delimiter_token": "|"}
|
||||
{"unk_token": "<unk>", "bos_token": "<s>", "eos_token": "</s>", "pad_token": "<pad>", "do_lower_case": false, "word_delimiter_token": "|","special_tokens_map_file": "/home/ceyda/workspace/libs/fairseq/hf_finetuned_output/special_tokens_map.json", "tokenizer_file": null}
|
40
vocab.json
40
vocab.json
|
@ -1 +1,39 @@
|
|||
{"e": 0, "x": 1, "î": 2, "l": 3, "â": 4, "j": 5, "ç": 6, "ş": 7, "g": 8, "ı": 9, "v": 10, "d": 11, "t": 12, "n": 13, "a": 14, "c": 15, "h": 16, "p": 17, "r": 18, "w": 19, "z": 20, "k": 21, "u": 22, "b": 23, "ü": 24, "y": 26, "o": 27, "q": 28, "m": 29, "f": 30, "s": 31, "ö": 32, "ğ": 33, "i": 34, "|": 25, "[UNK]": 35, "[PAD]": 36}
|
||||
{"|": 4,
|
||||
"p": 5,
|
||||
"i": 6,
|
||||
"r": 7,
|
||||
"n": 8,
|
||||
"s": 9,
|
||||
"ö": 10,
|
||||
"z": 11,
|
||||
"l": 12,
|
||||
"e": 13,
|
||||
"h": 14,
|
||||
"â": 15,
|
||||
"y": 16,
|
||||
"a": 17,
|
||||
"k": 18,
|
||||
"ı": 19,
|
||||
"o": 20,
|
||||
"m": 21,
|
||||
"ü": 22,
|
||||
"g": 23,
|
||||
"c": 24,
|
||||
"b": 25,
|
||||
"ş": 26,
|
||||
"d": 27,
|
||||
"u": 28,
|
||||
"t": 29,
|
||||
"ç": 30,
|
||||
"ğ": 31,
|
||||
"v": 32,
|
||||
"f": 33,
|
||||
"j": 34,
|
||||
"x": 35,
|
||||
"w": 36,
|
||||
"q": 37,
|
||||
"î": 38,
|
||||
"<s>": 0,
|
||||
"<pad>": 1,
|
||||
"</s>": 2,
|
||||
"<unk>": 3}
|
||||
|
|
Loading…
Reference in New Issue