Correct maximum positional embeddings

#17
Files changed (1) hide show
  1. config.json +1 -1
config.json CHANGED
@@ -28,7 +28,7 @@
28
  "initializer_range": 0.02,
29
  "intermediate_size": null,
30
  "layer_norm_epsilon": 1e-05,
31
- "max_position_embeddings": 2048,
32
  "model_type": "gpt_neo",
33
  "num_heads": 16,
34
  "num_layers": 4,
 
28
  "initializer_range": 0.02,
29
  "intermediate_size": null,
30
  "layer_norm_epsilon": 1e-05,
31
+ "max_position_embeddings": 512,
32
  "model_type": "gpt_neo",
33
  "num_heads": 16,
34
  "num_layers": 4,