m-nagoudi commited on
Commit
d590f0b
·
1 Parent(s): 5ec18f3

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +3 -0
config.json CHANGED
@@ -32,6 +32,9 @@
32
  "intermediate_size": null,
33
  "layer_norm_epsilon": 1e-05,
34
  "max_position_embeddings": 2048,
 
 
 
35
  "model_type": "gpt_neo",
36
  "num_heads": 12,
37
  "num_layers": 12,
 
32
  "intermediate_size": null,
33
  "layer_norm_epsilon": 1e-05,
34
  "max_position_embeddings": 2048,
35
+ "eos_token_id" = 0,
36
+ "bos_token_id" = 0,
37
+ "pad_token_id" = 1,
38
  "model_type": "gpt_neo",
39
  "num_heads": 12,
40
  "num_layers": 12,