danielhanchen commited on
Commit
d0f81d6
·
verified ·
1 Parent(s): 08a5c65

Add files using upload-large-folder tool

Browse files
Files changed (2) hide show
  1. config.json +2 -2
  2. generation_config.json +1 -1
config.json CHANGED
@@ -4,7 +4,7 @@
4
  "Qwen2ForCausalLM"
5
  ],
6
  "attention_dropout": 0.0,
7
- "bos_token_id": 151643,
8
  "eos_token_id": 151643,
9
  "hidden_act": "silu",
10
  "hidden_size": 1536,
@@ -43,7 +43,7 @@
43
  "sliding_window": null,
44
  "tie_word_embeddings": false,
45
  "torch_dtype": "bfloat16",
46
- "transformers_version": "4.48.1",
47
  "unsloth_fixed": true,
48
  "use_cache": true,
49
  "use_mrope": false,
 
4
  "Qwen2ForCausalLM"
5
  ],
6
  "attention_dropout": 0.0,
7
+ "bos_token_id": 151646,
8
  "eos_token_id": 151643,
9
  "hidden_act": "silu",
10
  "hidden_size": 1536,
 
43
  "sliding_window": null,
44
  "tie_word_embeddings": false,
45
  "torch_dtype": "bfloat16",
46
+ "transformers_version": "4.49.0.dev0",
47
  "unsloth_fixed": true,
48
  "use_cache": true,
49
  "use_mrope": false,
generation_config.json CHANGED
@@ -7,5 +7,5 @@
7
  "pad_token_id": 151654,
8
  "temperature": 0.6,
9
  "top_p": 0.95,
10
- "transformers_version": "4.48.1"
11
  }
 
7
  "pad_token_id": 151654,
8
  "temperature": 0.6,
9
  "top_p": 0.95,
10
+ "transformers_version": "4.49.0.dev0"
11
  }