kbert_base / config.json
lapp0's picture
Upload KBERTForMaskedLM
37212c8 verified
raw
history blame contribute delete
309 Bytes
{
"architectures": [
"KBERTForMaskedLM"
],
"head_dropout": 0.0,
"intermediate_dim": 2048,
"logit_softcap": 15,
"model_dim": 768,
"num_attention_heads": 6,
"num_layers": 20,
"tokenizer_uri": "answerdotai/ModernBERT-base",
"torch_dtype": "bfloat16",
"transformers_version": "4.48.0"
}