svakhreev commited on
Commit
34f8d9a
1 Parent(s): da95b3f

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +4 -8
config.json CHANGED
@@ -2,15 +2,13 @@
2
  "architectures": [
3
  "GPTRefactForCausalLM"
4
  ],
5
- "attention_softmax_in_fp32": false,
6
- "attn_pdrop": 0.1,
7
  "auto_map": {
8
  "AutoConfig": "configuration_gpt_refact.GPTRefactConfig",
9
  "AutoModelForCausalLM": "modeling_gpt_refact.GPTRefactForCausalLM"
10
  },
11
- "bos_token_id": -1,
12
  "do_sample": true,
13
- "embd_pdrop": 0.1,
14
  "eos_token_id": 0,
15
  "initializer_range": 0.02,
16
  "layer_norm_epsilon": 1e-05,
@@ -21,10 +19,8 @@
21
  "n_inner": null,
22
  "n_layer": 32,
23
  "n_positions": 4096,
24
- "resid_pdrop": 0.1,
25
- "scale_attention_softmax_in_fp32": false,
26
- "scale_attn_weights": true,
27
- "torch_dtype": "float32",
28
  "transformers_version": "4.31.0",
29
  "use_cache": true,
30
  "vocab_size": 49216
 
2
  "architectures": [
3
  "GPTRefactForCausalLM"
4
  ],
5
+ "attention_bias_in_fp32": true,
6
+ "attention_softmax_in_fp32": true,
7
  "auto_map": {
8
  "AutoConfig": "configuration_gpt_refact.GPTRefactConfig",
9
  "AutoModelForCausalLM": "modeling_gpt_refact.GPTRefactForCausalLM"
10
  },
 
11
  "do_sample": true,
 
12
  "eos_token_id": 0,
13
  "initializer_range": 0.02,
14
  "layer_norm_epsilon": 1e-05,
 
19
  "n_inner": null,
20
  "n_layer": 32,
21
  "n_positions": 4096,
22
+ "scale_attention_softmax_in_fp32": true,
23
+ "torch_dtype": "bfloat16",
 
 
24
  "transformers_version": "4.31.0",
25
  "use_cache": true,
26
  "vocab_size": 49216