Muennighoff
commited on
Commit
•
fb716d3
1
Parent(s):
59facdf
Correct pretraining tp
Browse files- config.json +1 -1
config.json
CHANGED
@@ -20,7 +20,7 @@
|
|
20 |
"n_layer": 30,
|
21 |
"num_attention_heads": 32,
|
22 |
"offset_alibi": 100,
|
23 |
-
"pretraining_tp":
|
24 |
"seq_length": 2048,
|
25 |
"skip_bias_add": true,
|
26 |
"skip_bias_add_qkv": false,
|
|
|
20 |
"n_layer": 30,
|
21 |
"num_attention_heads": 32,
|
22 |
"offset_alibi": 100,
|
23 |
+
"pretraining_tp": 1,
|
24 |
"seq_length": 2048,
|
25 |
"skip_bias_add": true,
|
26 |
"skip_bias_add_qkv": false,
|