From 262613e53384a4830c0abb437976cc285d56e692 Mon Sep 17 00:00:00 2001 From: Niels Rogge Date: Mon, 2 Jan 2023 10:52:35 +0000 Subject: [PATCH] Upload GitForCausalLM --- config.json | 106 ++++++++++++++++++++++++++++++++++++++++++++++ pytorch_model.bin | 3 ++ 2 files changed, 109 insertions(+) create mode 100644 config.json create mode 100644 pytorch_model.bin diff --git a/config.json b/config.json new file mode 100644 index 0000000..678f1d1 --- /dev/null +++ b/config.json @@ -0,0 +1,106 @@ +{ + "_commit_hash": null, + "architectures": [ + "GitForCausalLM" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 101, + "classifier_dropout": null, + "eos_token_id": 102, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 1024, + "model_type": "git", + "num_attention_heads": 12, + "num_hidden_layers": 6, + "num_image_with_embedding": null, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "tie_word_embeddings": false, + "torch_dtype": "float32", + "transformers_version": null, + "use_cache": true, + "vision_config": { + "_name_or_path": "", + "add_cross_attention": false, + "architectures": null, + "attention_dropout": 0.0, + "bad_words_ids": null, + "begin_suppress_tokens": null, + "bos_token_id": null, + "chunk_size_feed_forward": 0, + "cross_attention_hidden_size": null, + "decoder_start_token_id": null, + "diversity_penalty": 0.0, + "do_sample": false, + "dropout": 0.0, + "early_stopping": false, + "encoder_no_repeat_ngram_size": 0, + "eos_token_id": null, + "exponential_decay_length_penalty": null, + "finetuning_task": null, + "forced_bos_token_id": null, + "forced_eos_token_id": null, + "hidden_act": "quick_gelu", + "hidden_size": 1024, + "id2label": { + "0": "LABEL_0", + "1": "LABEL_1" + }, + "image_size": 224, + "initializer_factor": 1.0, + "initializer_range": 0.02, + "intermediate_size": 4096, + "is_decoder": false, + "is_encoder_decoder": false, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1 + }, + "layer_norm_eps": 1e-05, + "length_penalty": 1.0, + "max_length": 20, + "min_length": 0, + "model_type": "git_vision_model", + "no_repeat_ngram_size": 0, + "num_attention_heads": 16, + "num_beam_groups": 1, + "num_beams": 1, + "num_channels": 3, + "num_hidden_layers": 24, + "num_return_sequences": 1, + "output_attentions": false, + "output_hidden_states": false, + "output_scores": false, + "pad_token_id": null, + "patch_size": 14, + "prefix": null, + "problem_type": null, + "projection_dim": 512, + "pruned_heads": {}, + "remove_invalid_values": false, + "repetition_penalty": 1.0, + "return_dict": true, + "return_dict_in_generate": false, + "sep_token_id": null, + "suppress_tokens": null, + "task_specific_params": null, + "temperature": 1.0, + "tf_legacy_loss": false, + "tie_encoder_decoder": false, + "tie_word_embeddings": true, + "tokenizer_class": null, + "top_k": 50, + "top_p": 1.0, + "torch_dtype": null, + "torchscript": false, + "transformers_version": "4.26.0.dev0", + "typical_p": 1.0, + "use_bfloat16": false + }, + "vocab_size": 30522 +} diff --git a/pytorch_model.bin b/pytorch_model.bin new file mode 100644 index 0000000..33a284b --- /dev/null +++ b/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cdfc6b68ba3499a63652eb2dc0f1b46aa2f4e4a01d0e18c2866057cb764c4792 +size 1576966105