From 220857605d793c642827eceb57479bc77a015631 Mon Sep 17 00:00:00 2001 From: Stella Biderman Date: Thu, 5 Aug 2021 20:05:38 +0000 Subject: [PATCH] Update to GPT2 style --- config.json | 20 ++++---------------- 1 file changed, 4 insertions(+), 16 deletions(-) diff --git a/config.json b/config.json index 7b25bc7..2494b6c 100644 --- a/config.json +++ b/config.json @@ -4,30 +4,18 @@ "GPTJForCausalLM" ], "attention_dropout": 0, - "attention_layers": ["global", "global", "global", "global", "global", "global", "global", "global", "global", "global", "global", "global", "global", "global", "global", "global", "global", "global", "global", "global", "global", "global", "global", "global", "global", "global", "global", "global"], - "attention_types": [ - [ - [ - "global" - ], - 28 - ] - ], "bos_token_id": 50256, "embed_dropout": 0, "eos_token_id": 50256, "gradient_checkpointing": false, - "hidden_size": 4096, + "n_embd": 4096, "initializer_range": 0.02, "intermediate_size": null, - "jax": true, "layer_norm_epsilon": 1e-05, - "max_position_embeddings": 2048, + "n_positions": 2048, "model_type": "gptj", - "num_heads": 16, - "num_layers": 28, - "resid_dropout": 0, - "rotary": true, + "n_head": 16, + "n_layer": 28, "rotary_dim": 64, "summary_activation": null, "summary_first_dropout": 0.1,