danielhanchen commited on
Commit
5a77b42
·
verified ·
1 Parent(s): 75b37d2

Add files using upload-large-folder tool

Browse files
Files changed (2) hide show
  1. config.json +2 -2
  2. generation_config.json +1 -1
config.json CHANGED
@@ -4,7 +4,7 @@
4
  "Qwen2ForCausalLM"
5
  ],
6
  "attention_dropout": 0.0,
7
- "bos_token_id": 151643,
8
  "eos_token_id": 151643,
9
  "hidden_act": "silu",
10
  "hidden_size": 3584,
@@ -57,7 +57,7 @@
57
  "sliding_window": null,
58
  "tie_word_embeddings": false,
59
  "torch_dtype": "bfloat16",
60
- "transformers_version": "4.48.1",
61
  "unsloth_fixed": true,
62
  "use_cache": true,
63
  "use_mrope": false,
 
4
  "Qwen2ForCausalLM"
5
  ],
6
  "attention_dropout": 0.0,
7
+ "bos_token_id": 151646,
8
  "eos_token_id": 151643,
9
  "hidden_act": "silu",
10
  "hidden_size": 3584,
 
57
  "sliding_window": null,
58
  "tie_word_embeddings": false,
59
  "torch_dtype": "bfloat16",
60
+ "transformers_version": "4.49.0.dev0",
61
  "unsloth_fixed": true,
62
  "use_cache": true,
63
  "use_mrope": false,
generation_config.json CHANGED
@@ -7,5 +7,5 @@
7
  "pad_token_id": 151654,
8
  "temperature": 0.6,
9
  "top_p": 0.95,
10
- "transformers_version": "4.48.1"
11
  }
 
7
  "pad_token_id": 151654,
8
  "temperature": 0.6,
9
  "top_p": 0.95,
10
+ "transformers_version": "4.49.0.dev0"
11
  }