diff --git a/config.json b/config.json
index cf53f72..7275e8e 100644
--- a/config.json
+++ b/config.json
@@ -1,12 +1,11 @@
{
- "_name_or_path": "./pretrained/checkpoint_0.60",
"activation_dropout": 0.1,
"apply_spec_augment": true,
"architectures": [
"Wav2Vec2ForCTC"
],
"attention_dropout": 0.1,
- "bos_token_id": 1,
+ "bos_token_id": 0,
"conv_bias": false,
"conv_dim": [
512,
@@ -35,17 +34,17 @@
2,
2
],
- "ctc_loss_reduction": "mean",
- "ctc_zero_infinity": true,
+ "ctc_loss_reduction": "sum",
+ "ctc_zero_infinity": false,
"do_stable_layer_norm": false,
"eos_token_id": 2,
"feat_extract_activation": "gelu",
"feat_extract_norm": "group",
"feat_proj_dropout": 0.1,
"final_dropout": 0.1,
- "gradient_checkpointing": true,
+ "gradient_checkpointing": false,
"hidden_act": "gelu",
- "hidden_dropout": 0.05,
+ "hidden_dropout": 0.1,
"hidden_size": 768,
"initializer_range": 0.02,
"intermediate_size": 3072,
@@ -54,14 +53,14 @@
"mask_feature_length": 10,
"mask_feature_prob": 0.0,
"mask_time_length": 10,
- "mask_time_prob": 0.5,
+ "mask_time_prob": 0.05,
"model_type": "wav2vec2",
"num_attention_heads": 12,
"num_conv_pos_embedding_groups": 16,
"num_conv_pos_embeddings": 128,
"num_feat_extract_layers": 7,
"num_hidden_layers": 12,
- "pad_token_id": 36,
- "transformers_version": "4.5.0.dev0",
- "vocab_size": 37
+ "pad_token_id": 1,
+ "transformers_version": "4.4.2",
+ "vocab_size": 39
}
diff --git a/pytorch_model.bin b/pytorch_model.bin
index 6b20cde..72adfa1 100644
--- a/pytorch_model.bin
+++ b/pytorch_model.bin
@@ -1,3 +1,3 @@
version https://git-lfs.github.com/spec/v1
-oid sha256:ecb5a521af5105ca67abd19a841c08e97b071a8b69e186b06b25633c6fc36804
-size 377690860
+oid sha256:7ac7eaad990b15315d1772928ea15b9c77d2e259311b5189f9772b04da157294
+size 377691502
diff --git a/special_tokens_map.json b/special_tokens_map.json
index 9abf719..25bc396 100644
--- a/special_tokens_map.json
+++ b/special_tokens_map.json
@@ -1 +1 @@
-{"bos_token": "", "eos_token": "", "unk_token": "[UNK]", "pad_token": "[PAD]"}
\ No newline at end of file
+{"bos_token": "", "eos_token": "", "unk_token": "", "pad_token": ""}
\ No newline at end of file
diff --git a/tokenizer_config.json b/tokenizer_config.json
index a2a8340..42e4a14 100644
--- a/tokenizer_config.json
+++ b/tokenizer_config.json
@@ -1 +1 @@
-{"unk_token": "[UNK]", "bos_token": "", "eos_token": "", "pad_token": "[PAD]", "do_lower_case": false, "word_delimiter_token": "|"}
\ No newline at end of file
+{"unk_token": "", "bos_token": "", "eos_token": "", "pad_token": "", "do_lower_case": false, "word_delimiter_token": "|","special_tokens_map_file": "/home/ceyda/workspace/libs/fairseq/hf_finetuned_output/special_tokens_map.json", "tokenizer_file": null}
\ No newline at end of file
diff --git a/vocab.json b/vocab.json
index 7f974b7..2edf7b2 100644
--- a/vocab.json
+++ b/vocab.json
@@ -1 +1,39 @@
-{"e": 0, "x": 1, "î": 2, "l": 3, "â": 4, "j": 5, "ç": 6, "ş": 7, "g": 8, "ı": 9, "v": 10, "d": 11, "t": 12, "n": 13, "a": 14, "c": 15, "h": 16, "p": 17, "r": 18, "w": 19, "z": 20, "k": 21, "u": 22, "b": 23, "ü": 24, "y": 26, "o": 27, "q": 28, "m": 29, "f": 30, "s": 31, "ö": 32, "ğ": 33, "i": 34, "|": 25, "[UNK]": 35, "[PAD]": 36}
\ No newline at end of file
+{"|": 4,
+ "p": 5,
+ "i": 6,
+ "r": 7,
+ "n": 8,
+ "s": 9,
+ "ö": 10,
+ "z": 11,
+ "l": 12,
+ "e": 13,
+ "h": 14,
+ "â": 15,
+ "y": 16,
+ "a": 17,
+ "k": 18,
+ "ı": 19,
+ "o": 20,
+ "m": 21,
+ "ü": 22,
+ "g": 23,
+ "c": 24,
+ "b": 25,
+ "ş": 26,
+ "d": 27,
+ "u": 28,
+ "t": 29,
+ "ç": 30,
+ "ğ": 31,
+ "v": 32,
+ "f": 33,
+ "j": 34,
+ "x": 35,
+ "w": 36,
+ "q": 37,
+ "î": 38,
+ "": 0,
+ "": 1,
+ "": 2,
+ "": 3}