ner-jp-gliner / gliner_config.json
vumichien's picture
Training in progress, step 3000
fbf7086 verified
raw
history blame
1.14 kB
{
"class_token_index": 96872,
"dropout": 0.4,
"encoder_config": null,
"ent_token": "<<ENT>>",
"eval_every": 500,
"fine_tune": true,
"freeze_token_rep": false,
"has_rnn": true,
"hidden_size": 768,
"label_smoothing": 0,
"log_dir": "ner-jp-gliner",
"loss_alpha": -1,
"loss_gamma": 0,
"loss_reduction": "sum",
"lr_encoder": "1e-5",
"lr_others": "5e-5",
"max_grad_norm": 1.0,
"max_len": 384,
"max_neg_type_ratio": 1,
"max_types": 25,
"max_width": 12,
"model_name": "ku-nlp/deberta-v3-base-japanese",
"model_type": "gliner",
"name": "span level gliner",
"num_steps": 1000,
"prev_path": "none",
"random_drop": true,
"root_dir": "gliner_logs",
"save_total_limit": 10,
"scheduler_type": "cosine",
"sep_token": "<<SEP>>",
"shuffle_types": true,
"size_sup": -1,
"span_mode": "markerV0",
"subtoken_pooling": "first",
"train_batch_size": 8,
"train_data": "ner_jp.json",
"transformers_version": "4.41.2",
"val_data_dir": "none",
"vocab_size": 96874,
"warmup_ratio": 0.1,
"weight_decay_encoder": 0.01,
"weight_decay_other": 0.01,
"words_splitter_type": "whitespace"
}