ltoniazzi commited on
Commit
f76605c
·
verified ·
1 Parent(s): bf8abbe

Upload Gemma2ForCausalLM/hidden_size=64/base/config.json with huggingface_hub

Browse files
Gemma2ForCausalLM/hidden_size=64/base/config.json CHANGED
@@ -10,7 +10,7 @@
10
  "cache_implementation": "hybrid",
11
  "eos_token_id": 1,
12
  "final_logit_softcapping": 30.0,
13
- "head_dim": 256,
14
  "hidden_act": "gelu_pytorch_tanh",
15
  "hidden_activation": "gelu_pytorch_tanh",
16
  "hidden_size": 64,
 
10
  "cache_implementation": "hybrid",
11
  "eos_token_id": 1,
12
  "final_logit_softcapping": 30.0,
13
+ "head_dim": 64,
14
  "hidden_act": "gelu_pytorch_tanh",
15
  "hidden_activation": "gelu_pytorch_tanh",
16
  "hidden_size": 64,