m-nagoudi commited on
Commit
ad4188b
·
1 Parent(s): d590f0b

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +3 -3
config.json CHANGED
@@ -32,9 +32,9 @@
32
  "intermediate_size": null,
33
  "layer_norm_epsilon": 1e-05,
34
  "max_position_embeddings": 2048,
35
- "eos_token_id" = 0,
36
- "bos_token_id" = 0,
37
- "pad_token_id" = 1,
38
  "model_type": "gpt_neo",
39
  "num_heads": 12,
40
  "num_layers": 12,
 
32
  "intermediate_size": null,
33
  "layer_norm_epsilon": 1e-05,
34
  "max_position_embeddings": 2048,
35
+ "eos_token_id" : 0,
36
+ "bos_token_id" : 0,
37
+ "pad_token_id" : 1,
38
  "model_type": "gpt_neo",
39
  "num_heads": 12,
40
  "num_layers": 12,