transformer_base / config.json
huyenquinn282's picture
Upload model
959f027 verified
raw
history blame contribute delete
295 Bytes
{
"architectures": [
"TransformerModel"
],
"embedding_dim": 128,
"fully_connected_dim": 128,
"max_position_embeddings": 512,
"model_type": "transformer",
"num_heads": 2,
"num_layers": 2,
"torch_dtype": "float32",
"transformers_version": "4.41.2",
"vocab_size": 30522
}