From 842ba042d27e06f65c2e7b5ec66a58ba64835b59 Mon Sep 17 00:00:00 2001 From: system Date: Fri, 24 Apr 2020 15:54:21 +0000 Subject: [PATCH] Update config.json --- config.json | 26 ++++++++++++++++---------- 1 file changed, 16 insertions(+), 10 deletions(-) diff --git a/config.json b/config.json index 7f13b8c..c2add12 100644 --- a/config.json +++ b/config.json @@ -3,21 +3,27 @@ "AlbertForMaskedLM" ], "attention_probs_dropout_prob": 0, + "bos_token_id": 2, + "classifier_dropout_prob": 0.1, + "down_scale_factor": 1, + "embedding_size": 128, + "eos_token_id": 3, + "gap_size": 0, "hidden_act": "gelu_new", "hidden_dropout_prob": 0, - "embedding_size": 128, "hidden_size": 768, "initializer_range": 0.02, - "intermediate_size": 3072, - "max_position_embeddings": 512, - "num_attention_heads": 12, - "num_hidden_layers": 12, - "num_hidden_groups": 1, - "net_structure_type": 0, - "gap_size": 0, - "num_memory_blocks": 0, "inner_group_num": 1, - "down_scale_factor": 1, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "albert", + "net_structure_type": 0, + "num_attention_heads": 12, + "num_hidden_groups": 1, + "num_hidden_layers": 12, + "num_memory_blocks": 0, + "pad_token_id": 0, "type_vocab_size": 2, "vocab_size": 30000 }