Upload TF model

This commit is contained in:
Niels Rogge 2021-11-29 11:02:34 +00:00
parent 4a7bc62e6f
commit 70ef0dc38f
2 changed files with 72 additions and 67 deletions

View File

@ -1,67 +1,69 @@
{ {
"aggregation_labels": { "_name_or_path": "google/tapas-large-finetuned-wtq",
"0": "NONE", "aggregation_labels": {
"1": "SUM", "0": "NONE",
"2": "AVERAGE", "1": "SUM",
"3": "COUNT" "2": "AVERAGE",
}, "3": "COUNT"
"aggregation_loss_weight": 1.0, },
"aggregation_temperature": 1.0, "aggregation_loss_weight": 1.0,
"allow_empty_column_selection": false, "aggregation_temperature": 1.0,
"answer_loss_cutoff": 0.664694, "allow_empty_column_selection": false,
"answer_loss_importance": 1.0, "answer_loss_cutoff": 0.664694,
"architectures": [ "answer_loss_importance": 1.0,
"TapasForQuestionAnswering" "architectures": [
], "TapasForQuestionAnswering"
"attention_probs_dropout_prob": 0.1, ],
"average_approximation_function": "ratio", "attention_probs_dropout_prob": 0.1,
"average_logits_per_cell": false, "average_approximation_function": "ratio",
"cell_selection_preference": 0.207951, "average_logits_per_cell": false,
"disable_per_token_loss": false, "cell_selection_preference": 0.207951,
"gradient_checkpointing": false, "disable_per_token_loss": false,
"hidden_act": "gelu", "gradient_checkpointing": false,
"hidden_dropout_prob": 0.1, "hidden_act": "gelu",
"hidden_size": 1024, "hidden_dropout_prob": 0.1,
"huber_loss_delta": 0.121194, "hidden_size": 1024,
"init_cell_selection_weights_to_zero": true, "huber_loss_delta": 0.121194,
"initializer_range": 0.02, "init_cell_selection_weights_to_zero": true,
"intermediate_size": 4096, "initializer_range": 0.02,
"layer_norm_eps": 1e-12, "intermediate_size": 4096,
"max_num_columns": 32, "layer_norm_eps": 1e-12,
"max_num_rows": 64, "max_num_columns": 32,
"max_position_embeddings": 1024, "max_num_rows": 64,
"model_type": "tapas", "max_position_embeddings": 1024,
"no_aggregation_label_index": 0, "model_type": "tapas",
"num_aggregation_labels": 4, "no_aggregation_label_index": 0,
"num_attention_heads": 16, "num_aggregation_labels": 4,
"num_hidden_layers": 24, "num_attention_heads": 16,
"pad_token_id": 0, "num_hidden_layers": 24,
"positive_label_weight": 10.0, "pad_token_id": 0,
"reset_position_index_per_cell": true, "positive_label_weight": 10.0,
"select_one_column": true, "reset_position_index_per_cell": true,
"softmax_temperature": 1.0, "select_one_column": true,
"temperature": 0.0352513, "softmax_temperature": 1.0,
"type_vocab_size": [ "temperature": 0.0352513,
3, "transformers_version": "4.13.0.dev0",
256, "type_vocab_size": [
256, 3,
2, 256,
256, 256,
256, 2,
10 256,
], 256,
"type_vocab_sizes": [ 10
3, ],
256, "type_vocab_sizes": [
256, 3,
2, 256,
256, 256,
256, 2,
10 256,
], 256,
"use_answer_as_supervision": true, 10
"use_gumbel_for_aggregation": false, ],
"use_gumbel_for_cells": false, "use_answer_as_supervision": true,
"use_normalized_answer_loss": false, "use_gumbel_for_aggregation": false,
"vocab_size": 30522 "use_gumbel_for_cells": false,
} "use_normalized_answer_loss": false,
"vocab_size": 30522
}

BIN
tf_model.h5 (Stored with Git LFS) Normal file

Binary file not shown.