|
{ |
|
"_name_or_path": "AutoTrain", |
|
"_num_labels": 57, |
|
"architectures": [ |
|
"BertForSequenceClassification" |
|
], |
|
"attention_probs_dropout_prob": 0.1, |
|
"classifier_dropout": null, |
|
"gradient_checkpointing": false, |
|
"hidden_act": "gelu", |
|
"hidden_dropout_prob": 0.1, |
|
"hidden_size": 768, |
|
"id2label": { |
|
"0": "129.0", |
|
"1": "131.0", |
|
"2": "142.0", |
|
"3": "149.0", |
|
"4": "151.0", |
|
"5": "159.0", |
|
"6": "191.0", |
|
"7": "192.0", |
|
"8": "193.0", |
|
"9": "199.0", |
|
"10": "212.0", |
|
"11": "222.0", |
|
"12": "231.0", |
|
"13": "232.0", |
|
"14": "234.0", |
|
"15": "234.1", |
|
"16": "234.2", |
|
"17": "234.3", |
|
"18": "234.4", |
|
"19": "235.0", |
|
"20": "236.0", |
|
"21": "239.0", |
|
"22": "240.0", |
|
"23": "251.0", |
|
"24": "252.0", |
|
"25": "262.0", |
|
"26": "313.0", |
|
"27": "314.0", |
|
"28": "319.0", |
|
"29": "321.0", |
|
"30": "325.0", |
|
"31": "330.0", |
|
"32": "342.0", |
|
"33": "350.0", |
|
"34": "361.0", |
|
"35": "362.0", |
|
"36": "370.0", |
|
"37": "380.0", |
|
"38": "390.0", |
|
"39": "410.0", |
|
"40": "422.0", |
|
"41": "423.0", |
|
"42": "424.0", |
|
"43": "429.0", |
|
"44": "449.0", |
|
"45": "490.0", |
|
"46": "511.0", |
|
"47": "512.0", |
|
"48": "513.0", |
|
"49": "519.0", |
|
"50": "521.0", |
|
"51": "523.0", |
|
"52": "526.0", |
|
"53": "529.0", |
|
"54": "539.0", |
|
"55": "611.0", |
|
"56": "690.0" |
|
}, |
|
"initializer_range": 0.02, |
|
"intermediate_size": 3072, |
|
"label2id": { |
|
"129.0": 0, |
|
"131.0": 1, |
|
"142.0": 2, |
|
"149.0": 3, |
|
"151.0": 4, |
|
"159.0": 5, |
|
"191.0": 6, |
|
"192.0": 7, |
|
"193.0": 8, |
|
"199.0": 9, |
|
"212.0": 10, |
|
"222.0": 11, |
|
"231.0": 12, |
|
"232.0": 13, |
|
"234.0": 14, |
|
"234.1": 15, |
|
"234.2": 16, |
|
"234.3": 17, |
|
"234.4": 18, |
|
"235.0": 19, |
|
"236.0": 20, |
|
"239.0": 21, |
|
"240.0": 22, |
|
"251.0": 23, |
|
"252.0": 24, |
|
"262.0": 25, |
|
"313.0": 26, |
|
"314.0": 27, |
|
"319.0": 28, |
|
"321.0": 29, |
|
"325.0": 30, |
|
"330.0": 31, |
|
"342.0": 32, |
|
"350.0": 33, |
|
"361.0": 34, |
|
"362.0": 35, |
|
"370.0": 36, |
|
"380.0": 37, |
|
"390.0": 38, |
|
"410.0": 39, |
|
"422.0": 40, |
|
"423.0": 41, |
|
"424.0": 42, |
|
"429.0": 43, |
|
"449.0": 44, |
|
"490.0": 45, |
|
"511.0": 46, |
|
"512.0": 47, |
|
"513.0": 48, |
|
"519.0": 49, |
|
"521.0": 50, |
|
"523.0": 51, |
|
"526.0": 52, |
|
"529.0": 53, |
|
"539.0": 54, |
|
"611.0": 55, |
|
"690.0": 56 |
|
}, |
|
"layer_norm_eps": 1e-12, |
|
"max_length": 192, |
|
"max_position_embeddings": 512, |
|
"model_type": "bert", |
|
"num_attention_heads": 12, |
|
"num_hidden_layers": 12, |
|
"output_past": true, |
|
"pad_token_id": 1, |
|
"padding": "max_length", |
|
"position_embedding_type": "absolute", |
|
"problem_type": "single_label_classification", |
|
"torch_dtype": "float32", |
|
"transformers_version": "4.15.0", |
|
"type_vocab_size": 2, |
|
"use_cache": true, |
|
"vocab_size": 31002 |
|
} |
|
|