60emotions / config.json
system's picture
system HF staff
Commit From AutoTrain
9fc1aed
{
"_name_or_path": "AutoTrain",
"_num_labels": 58,
"architectures": [
"RobertaForSequenceClassification"
],
"attention_probs_dropout_prob": 0.1,
"bos_token_id": 0,
"classifier_dropout": null,
"eos_token_id": 2,
"gradient_checkpointing": false,
"hidden_act": "gelu",
"hidden_dropout_prob": 0.1,
"hidden_size": 1024,
"id2label": {
"0": "\uac00\ub09c\ud55c, \ubd88\uc6b0\ud55c",
"1": "\uac10\uc0ac\ud558\ub294",
"2": "\uac71\uc815\uc2a4\ub7ec\uc6b4",
"3": "\uace0\ub9bd\ub41c",
"4": "\uad34\ub85c\uc6cc\ud558\ub294",
"5": "\uad6c\uc5ed\uc9c8 \ub098\ub294",
"6": "\uae30\uc068",
"7": "\ub099\ub2f4\ud55c",
"8": "\ub0a8\uc758 \uc2dc\uc120\uc744 \uc758\uc2dd\ud558\ub294",
"9": "\ub178\uc5ec\uc6cc\ud558\ub294",
"10": "\ub208\ubb3c\uc774 \ub098\ub294",
"11": "\ub290\uae0b",
"12": "\ub2f9\ud639\uc2a4\ub7ec\uc6b4",
"13": "\ub2f9\ud669",
"14": "\ub450\ub824\uc6b4",
"15": "\ub9c8\ube44\ub41c",
"16": "\ub9cc\uc871\uc2a4\ub7ec\uc6b4",
"17": "\ubc29\uc5b4\uc801\uc778",
"18": "\ubc30\uc2e0\ub2f9\ud55c",
"19": "\ubc84\ub824\uc9c4",
"20": "\ubd80\ub044\ub7ec\uc6b4",
"21": "\ubd84\ub178",
"22": "\ubd88\uc548",
"23": "\ube44\ud1b5\ud55c",
"24": "\uc0c1\ucc98",
"25": "\uc131\uac00\uc2e0",
"26": "\uc2a4\ud2b8\ub808\uc2a4 \ubc1b\ub294",
"27": "\uc2ac\ud514",
"28": "\uc2e0\ub8b0\ud558\ub294",
"29": "\uc2e0\uc774 \ub09c",
"30": "\uc2e4\ub9dd\ud55c",
"31": "\uc545\uc758\uc801\uc778",
"32": "\uc548\ub2ec\ud558\ub294",
"33": "\uc548\ub3c4",
"34": "\uc5b5\uc6b8\ud55c",
"35": "\uc5f4\ub4f1\uac10",
"36": "\uc5fc\uc138\uc801\uc778",
"37": "\uc678\ub85c\uc6b4",
"38": "\uc6b0\uc6b8\ud55c",
"39": "\uc790\uc2e0\ud558\ub294",
"40": "\uc870\uc2ec\uc2a4\ub7ec\uc6b4",
"41": "\uc88c\uc808\ud55c",
"42": "\uc8c4\ucc45\uac10\uc758",
"43": "\uc9c8\ud22c\ud558\ub294",
"44": "\uc9dc\uc99d\ub0b4\ub294",
"45": "\ucd08\uc870\ud55c",
"46": "\ucda9\uaca9 \ubc1b\uc740",
"47": "\ucde8\uc57d\ud55c",
"48": "\ud234\ud234\ub300\ub294",
"49": "\ud3b8\uc548\ud55c",
"50": "\ud55c\uc2ec\ud55c",
"51": "\ud610\uc624\uc2a4\ub7ec\uc6b4",
"52": "\ud63c\ub780\uc2a4\ub7ec\uc6b4",
"53": "\ud658\uba78\uc744 \ub290\ub07c\ub294",
"54": "\ud68c\uc758\uc801\uc778",
"55": "\ud6c4\ud68c\ub418\ub294",
"56": "\ud765\ubd84",
"57": "\ud76c\uc0dd\ub41c"
},
"initializer_range": 0.02,
"intermediate_size": 4096,
"label2id": {
"\uac00\ub09c\ud55c, \ubd88\uc6b0\ud55c": 0,
"\uac10\uc0ac\ud558\ub294": 1,
"\uac71\uc815\uc2a4\ub7ec\uc6b4": 2,
"\uace0\ub9bd\ub41c": 3,
"\uad34\ub85c\uc6cc\ud558\ub294": 4,
"\uad6c\uc5ed\uc9c8 \ub098\ub294": 5,
"\uae30\uc068": 6,
"\ub099\ub2f4\ud55c": 7,
"\ub0a8\uc758 \uc2dc\uc120\uc744 \uc758\uc2dd\ud558\ub294": 8,
"\ub178\uc5ec\uc6cc\ud558\ub294": 9,
"\ub208\ubb3c\uc774 \ub098\ub294": 10,
"\ub290\uae0b": 11,
"\ub2f9\ud639\uc2a4\ub7ec\uc6b4": 12,
"\ub2f9\ud669": 13,
"\ub450\ub824\uc6b4": 14,
"\ub9c8\ube44\ub41c": 15,
"\ub9cc\uc871\uc2a4\ub7ec\uc6b4": 16,
"\ubc29\uc5b4\uc801\uc778": 17,
"\ubc30\uc2e0\ub2f9\ud55c": 18,
"\ubc84\ub824\uc9c4": 19,
"\ubd80\ub044\ub7ec\uc6b4": 20,
"\ubd84\ub178": 21,
"\ubd88\uc548": 22,
"\ube44\ud1b5\ud55c": 23,
"\uc0c1\ucc98": 24,
"\uc131\uac00\uc2e0": 25,
"\uc2a4\ud2b8\ub808\uc2a4 \ubc1b\ub294": 26,
"\uc2ac\ud514": 27,
"\uc2e0\ub8b0\ud558\ub294": 28,
"\uc2e0\uc774 \ub09c": 29,
"\uc2e4\ub9dd\ud55c": 30,
"\uc545\uc758\uc801\uc778": 31,
"\uc548\ub2ec\ud558\ub294": 32,
"\uc548\ub3c4": 33,
"\uc5b5\uc6b8\ud55c": 34,
"\uc5f4\ub4f1\uac10": 35,
"\uc5fc\uc138\uc801\uc778": 36,
"\uc678\ub85c\uc6b4": 37,
"\uc6b0\uc6b8\ud55c": 38,
"\uc790\uc2e0\ud558\ub294": 39,
"\uc870\uc2ec\uc2a4\ub7ec\uc6b4": 40,
"\uc88c\uc808\ud55c": 41,
"\uc8c4\ucc45\uac10\uc758": 42,
"\uc9c8\ud22c\ud558\ub294": 43,
"\uc9dc\uc99d\ub0b4\ub294": 44,
"\ucd08\uc870\ud55c": 45,
"\ucda9\uaca9 \ubc1b\uc740": 46,
"\ucde8\uc57d\ud55c": 47,
"\ud234\ud234\ub300\ub294": 48,
"\ud3b8\uc548\ud55c": 49,
"\ud55c\uc2ec\ud55c": 50,
"\ud610\uc624\uc2a4\ub7ec\uc6b4": 51,
"\ud63c\ub780\uc2a4\ub7ec\uc6b4": 52,
"\ud658\uba78\uc744 \ub290\ub07c\ub294": 53,
"\ud68c\uc758\uc801\uc778": 54,
"\ud6c4\ud68c\ub418\ub294": 55,
"\ud765\ubd84": 56,
"\ud76c\uc0dd\ub41c": 57
},
"layer_norm_eps": 1e-05,
"max_length": 192,
"max_position_embeddings": 514,
"model_type": "roberta",
"num_attention_heads": 16,
"num_hidden_layers": 24,
"pad_token_id": 1,
"padding": "max_length",
"position_embedding_type": "absolute",
"problem_type": "single_label_classification",
"tokenizer_class": "BertTokenizer",
"torch_dtype": "float32",
"transformers_version": "4.29.2",
"type_vocab_size": 1,
"use_cache": true,
"vocab_size": 39255
}