fairseq tuned

This commit is contained in:
ceyda 2021-04-02 18:49:51 +00:00
parent 27503d5d38
commit 3d0bc0ac66
5 changed files with 52 additions and 15 deletions

View File

@ -1,12 +1,11 @@
{
"_name_or_path": "./pretrained/checkpoint_0.60",
"activation_dropout": 0.1,
"apply_spec_augment": true,
"architectures": [
"Wav2Vec2ForCTC"
],
"attention_dropout": 0.1,
"bos_token_id": 1,
"bos_token_id": 0,
"conv_bias": false,
"conv_dim": [
512,
@ -35,17 +34,17 @@
2,
2
],
"ctc_loss_reduction": "mean",
"ctc_zero_infinity": true,
"ctc_loss_reduction": "sum",
"ctc_zero_infinity": false,
"do_stable_layer_norm": false,
"eos_token_id": 2,
"feat_extract_activation": "gelu",
"feat_extract_norm": "group",
"feat_proj_dropout": 0.1,
"final_dropout": 0.1,
"gradient_checkpointing": true,
"gradient_checkpointing": false,
"hidden_act": "gelu",
"hidden_dropout": 0.05,
"hidden_dropout": 0.1,
"hidden_size": 768,
"initializer_range": 0.02,
"intermediate_size": 3072,
@ -54,14 +53,14 @@
"mask_feature_length": 10,
"mask_feature_prob": 0.0,
"mask_time_length": 10,
"mask_time_prob": 0.5,
"mask_time_prob": 0.05,
"model_type": "wav2vec2",
"num_attention_heads": 12,
"num_conv_pos_embedding_groups": 16,
"num_conv_pos_embeddings": 128,
"num_feat_extract_layers": 7,
"num_hidden_layers": 12,
"pad_token_id": 36,
"transformers_version": "4.5.0.dev0",
"vocab_size": 37
"pad_token_id": 1,
"transformers_version": "4.4.2",
"vocab_size": 39
}

View File

@ -1,3 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:ecb5a521af5105ca67abd19a841c08e97b071a8b69e186b06b25633c6fc36804
size 377690860
oid sha256:7ac7eaad990b15315d1772928ea15b9c77d2e259311b5189f9772b04da157294
size 377691502

View File

@ -1 +1 @@
{"bos_token": "<s>", "eos_token": "</s>", "unk_token": "[UNK]", "pad_token": "[PAD]"}
{"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>"}

View File

@ -1 +1 @@
{"unk_token": "[UNK]", "bos_token": "<s>", "eos_token": "</s>", "pad_token": "[PAD]", "do_lower_case": false, "word_delimiter_token": "|"}
{"unk_token": "<unk>", "bos_token": "<s>", "eos_token": "</s>", "pad_token": "<pad>", "do_lower_case": false, "word_delimiter_token": "|","special_tokens_map_file": "/home/ceyda/workspace/libs/fairseq/hf_finetuned_output/special_tokens_map.json", "tokenizer_file": null}

View File

@ -1 +1,39 @@
{"e": 0, "x": 1, "î": 2, "l": 3, "â": 4, "j": 5, "ç": 6, "ş": 7, "g": 8, "ı": 9, "v": 10, "d": 11, "t": 12, "n": 13, "a": 14, "c": 15, "h": 16, "p": 17, "r": 18, "w": 19, "z": 20, "k": 21, "u": 22, "b": 23, "ü": 24, "y": 26, "o": 27, "q": 28, "m": 29, "f": 30, "s": 31, "ö": 32, "ğ": 33, "i": 34, "|": 25, "[UNK]": 35, "[PAD]": 36}
{"|": 4,
"p": 5,
"i": 6,
"r": 7,
"n": 8,
"s": 9,
"ö": 10,
"z": 11,
"l": 12,
"e": 13,
"h": 14,
"â": 15,
"y": 16,
"a": 17,
"k": 18,
"ı": 19,
"o": 20,
"m": 21,
"ü": 22,
"g": 23,
"c": 24,
"b": 25,
"ş": 26,
"d": 27,
"u": 28,
"t": 29,
"ç": 30,
"ğ": 31,
"v": 32,
"f": 33,
"j": 34,
"x": 35,
"w": 36,
"q": 37,
"î": 38,
"<s>": 0,
"<pad>": 1,
"</s>": 2,
"<unk>": 3}