bjoernp commited on
Commit
768e55b
·
verified ·
1 Parent(s): 5da55ee

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +5 -1
config.json CHANGED
@@ -2,6 +2,10 @@
2
  "architectures": [
3
  "BitLlamaForCausalLM"
4
  ],
 
 
 
 
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
7
  "bos_token_id": 1,
@@ -18,7 +22,7 @@
18
  "pretraining_tp": 1,
19
  "rms_norm_eps": 1e-05,
20
  "rope_scaling": null,
21
- "rope_theta": 10000.0,
22
  "tie_word_embeddings": true,
23
  "torch_dtype": "float32",
24
  "transformers_version": "4.36.0.dev0",
 
2
  "architectures": [
3
  "BitLlamaForCausalLM"
4
  ],
5
+ "auto_map": {
6
+ "AutoConfig": "configuration_bitllama.LlamaConfig",
7
+ "AutoModelForCausalLM": "modeling_bitllama.BitLlamaForCausalLM"
8
+ },
9
  "attention_bias": false,
10
  "attention_dropout": 0.0,
11
  "bos_token_id": 1,
 
22
  "pretraining_tp": 1,
23
  "rms_norm_eps": 1e-05,
24
  "rope_scaling": null,
25
+ "rope_theta": 100000.0,
26
  "tie_word_embeddings": true,
27
  "torch_dtype": "float32",
28
  "transformers_version": "4.36.0.dev0",