friend-chatbot / config.json
lagoma's picture
creando config.json
4cdf027 verified
{
"model_name": "huggingchat/friend-chatbot",
"model_type": "transformer",
"num_layers": 6,
"hidden_size": 512,
"num_attention_heads": 8,
"intermediate_size": 2048,
"hidden_act": "gelu",
"num_labels": 1,
"max_position_embeddings": 512,
"type_vocab_size": 2,
"initializer_range": 0.02,
"layer_norm_eps": 1e-12,
"pad_token_id": 0,
"vocab_size": 30522,
"max_length": 128,
"train_batch_size": 8,
"eval_batch_size": 8,
"learning_rate": 5e-5,
"num_train_epochs": 3,
"warmup_ratio": 0.1,
"weight_decay": 0.01,
"adam_epsilon": 1e-8,
"max_grad_norm": 1.0,
"seed": 42,
"output_dir": "./models/friend-chatbot",
"fp16": true,
"gradient_accumulation_steps": 1,
"save_steps": 1000,
"logging_steps": 100,
"evaluate_during_training": true,
"evaluate_during_training_steps": 1000,
"overwrite_output_dir": true,
"report_to": "tensorboard",
"dataloader_drop_last": true,
"use_multiprocessing": false,
"cache_dir": "./cache",
"model_revision": "main",
"tokenizer_name": "huggingchat/friend-chatbot",
"do_train": true,
"do_eval": true,
"evaluation_strategy": "steps",
"per_device_train_batch_size": 8,
"per_device_eval_batch_size": 8,
"save_strategy": "steps",
"load_best_model_at_end": true,
"metric_for_best_model": "accuracy",
"greater_is_better": true,
"push_to_hub": true,
"hub_model_id": "huggingchat/friend-chatbot",
"hub_token": "your-huggingface-token"
}