Upload folder using huggingface_hub
Browse files- mlc-chat-config.json +76 -0
- ndarray-cache.json +0 -0
- params_shard_0.bin +3 -0
- params_shard_1.bin +3 -0
- params_shard_10.bin +3 -0
- params_shard_11.bin +3 -0
- params_shard_12.bin +3 -0
- params_shard_13.bin +3 -0
- params_shard_14.bin +3 -0
- params_shard_15.bin +3 -0
- params_shard_16.bin +3 -0
- params_shard_17.bin +3 -0
- params_shard_18.bin +3 -0
- params_shard_19.bin +3 -0
- params_shard_2.bin +3 -0
- params_shard_20.bin +3 -0
- params_shard_21.bin +3 -0
- params_shard_22.bin +3 -0
- params_shard_23.bin +3 -0
- params_shard_24.bin +3 -0
- params_shard_25.bin +3 -0
- params_shard_26.bin +3 -0
- params_shard_27.bin +3 -0
- params_shard_28.bin +3 -0
- params_shard_29.bin +3 -0
- params_shard_3.bin +3 -0
- params_shard_30.bin +3 -0
- params_shard_4.bin +3 -0
- params_shard_5.bin +3 -0
- params_shard_6.bin +3 -0
- params_shard_7.bin +3 -0
- params_shard_8.bin +3 -0
- params_shard_9.bin +3 -0
- tokenizer.json +0 -0
- tokenizer_config.json +35 -0
mlc-chat-config.json
ADDED
@@ -0,0 +1,76 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"version": "0.1.0",
|
3 |
+
"model_type": "qwen2",
|
4 |
+
"quantization": "q4f16_1",
|
5 |
+
"model_config": {
|
6 |
+
"hidden_act": "silu",
|
7 |
+
"hidden_size": 1536,
|
8 |
+
"intermediate_size": 8960,
|
9 |
+
"num_attention_heads": 12,
|
10 |
+
"num_hidden_layers": 28,
|
11 |
+
"num_key_value_heads": 2,
|
12 |
+
"rms_norm_eps": 1e-06,
|
13 |
+
"rope_theta": 10000,
|
14 |
+
"vocab_size": 151936,
|
15 |
+
"tie_word_embeddings": false,
|
16 |
+
"context_window_size": 131072,
|
17 |
+
"prefill_chunk_size": 8192,
|
18 |
+
"tensor_parallel_shards": 1,
|
19 |
+
"head_dim": 128,
|
20 |
+
"dtype": "float32",
|
21 |
+
"max_batch_size": 128
|
22 |
+
},
|
23 |
+
"vocab_size": 151936,
|
24 |
+
"context_window_size": 131072,
|
25 |
+
"sliding_window_size": -1,
|
26 |
+
"prefill_chunk_size": 8192,
|
27 |
+
"attention_sink_size": -1,
|
28 |
+
"tensor_parallel_shards": 1,
|
29 |
+
"pipeline_parallel_stages": 1,
|
30 |
+
"temperature": 1.0,
|
31 |
+
"presence_penalty": 0.0,
|
32 |
+
"frequency_penalty": 0.0,
|
33 |
+
"repetition_penalty": 1.0,
|
34 |
+
"top_p": 1.0,
|
35 |
+
"tokenizer_files": [
|
36 |
+
"tokenizer.json",
|
37 |
+
"tokenizer_config.json"
|
38 |
+
],
|
39 |
+
"tokenizer_info": {
|
40 |
+
"token_postproc_method": "byte_level",
|
41 |
+
"prepend_space_in_encode": false,
|
42 |
+
"strip_space_in_decode": false
|
43 |
+
},
|
44 |
+
"conv_template": {
|
45 |
+
"name": "deepseek_v3",
|
46 |
+
"system_template": "<\uff5cbegin\u2581of\u2581sentence\uff5c>{system_message}",
|
47 |
+
"system_message": "You are a helpful assistant.",
|
48 |
+
"system_prefix_token_ids": null,
|
49 |
+
"add_role_after_system_message": true,
|
50 |
+
"roles": {
|
51 |
+
"user": "<\uff5cUser\uff5c>",
|
52 |
+
"assistant": "<\uff5cAssistant\uff5c>"
|
53 |
+
},
|
54 |
+
"role_templates": {
|
55 |
+
"user": "{user_message}",
|
56 |
+
"assistant": "{assistant_message}",
|
57 |
+
"tool": "{tool_message}"
|
58 |
+
},
|
59 |
+
"messages": [],
|
60 |
+
"seps": [
|
61 |
+
"",
|
62 |
+
"<\uff5cend\u2581of\u2581sentence\uff5c>"
|
63 |
+
],
|
64 |
+
"role_content_sep": "",
|
65 |
+
"role_empty_sep": "",
|
66 |
+
"stop_str": [],
|
67 |
+
"stop_token_ids": [
|
68 |
+
151643
|
69 |
+
],
|
70 |
+
"function_string": "",
|
71 |
+
"use_function_calling": false
|
72 |
+
},
|
73 |
+
"pad_token_id": 0,
|
74 |
+
"bos_token_id": 151643,
|
75 |
+
"eos_token_id": 151643
|
76 |
+
}
|
ndarray-cache.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
params_shard_0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ec4813c1c77b8d0d3d362848966ed3f6bec760c329a9b4024278758d97c1c36d
|
3 |
+
size 116686848
|
params_shard_1.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d27b68ff3b61089610d2431d3097a288ac8abbf6027012755013b280e4fb438f
|
3 |
+
size 116686848
|
params_shard_10.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2c434bf8c1a12c70f82cbfd1a1fd032c36062e0d0624ee854869a2918b95adcc
|
3 |
+
size 26331136
|
params_shard_11.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:658ee406da70e5f872ea5f93d749c2f23e68c68a6cc7ad63f855dc3f6e0639d1
|
3 |
+
size 26331136
|
params_shard_12.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2970d7dd049a487bb1415cfd7ccc15bec06cea60a38a42c68e5886cfaa02643c
|
3 |
+
size 26331136
|
params_shard_13.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a014ab75a88523f552997270404aad87a746887ed9a3092c83b250594b7badc6
|
3 |
+
size 26331136
|
params_shard_14.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5cff56a70ee401359ff0f2cd9169d3abc8fff240ee42f713954cd959195fc786
|
3 |
+
size 26331136
|
params_shard_15.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0df414862e8be0cf7ba70ebefc89eafb80108a492c5d2190ff16b523b651fbfa
|
3 |
+
size 26331136
|
params_shard_16.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:868cc2f94a7fa22126a155cc5e06b14fed5f95b83af00b736755b846c7b3ba13
|
3 |
+
size 26331136
|
params_shard_17.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f73f777b8611905d18bae711f8f2a97a88f92f45aa21d8b6f05c3ebea3453135
|
3 |
+
size 26331136
|
params_shard_18.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dd03d9b6d8d2dd9fbdcb490c798c544957bb9569e3e58bcb00a932795e2cd7f3
|
3 |
+
size 26331136
|
params_shard_19.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5eb39dac820c506c9c37ba3da690a88da7a2a48c5920deb096d4c7ee88371678
|
3 |
+
size 26331136
|
params_shard_2.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f2c9ac194aa5ed68623826af040327a6ee43a8923d71eeced5a9554666dce213
|
3 |
+
size 29174784
|
params_shard_20.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:61fd3429478f1c85e5d9cdf34d37a3306fea8be8acdced8e2e407c16d910389b
|
3 |
+
size 26331136
|
params_shard_21.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:268247137646af01cad889dc6a12e229cfcf29d3ed9b96d0fe1850821e9362d1
|
3 |
+
size 26331136
|
params_shard_22.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:31b1a3a86c5198b3e0b537f9f05d92ca3b22b5d6747ff4450e4413171289f549
|
3 |
+
size 26331136
|
params_shard_23.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c54e39305bea2b781d6e5c518ea81d9dfe05ccb1a19306d02250fb2411bb41ae
|
3 |
+
size 26331136
|
params_shard_24.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:879358a57f9e2e8c15645afe76147f6f668769e2b02d33193f480de0acd80b47
|
3 |
+
size 26331136
|
params_shard_25.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3b9d831dacea12adda34c79bcfc061574093e64290fd8fada3726f5c1efd8b77
|
3 |
+
size 26331136
|
params_shard_26.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8607e8b3fe2ce80d0825e2dd4035e26318166bf487c2e062c792f5e2001192e4
|
3 |
+
size 26331136
|
params_shard_27.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:01e21acdb9f3a722227431e258fd7b3aae74eea80ba10e9ebe78c9ca6b37bdd5
|
3 |
+
size 26331136
|
params_shard_28.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:75b01f6c5daf9e775d5c50e75d4cd99d10ddf9e0a5458e0f602630161bea7085
|
3 |
+
size 26331136
|
params_shard_29.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:018f365ae1584512127502951877098287ac39f5ad7856e2eea263ad4f4262c6
|
3 |
+
size 26331136
|
params_shard_3.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:807993187284a5a883cd20f80a113c25c5fde13d69606d526dcd0448cba20f82
|
3 |
+
size 33212416
|
params_shard_30.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bd225141d3e031ab77600c180cedc8c5ae3854b8fc584236a5317f12c0be380c
|
3 |
+
size 18589696
|
params_shard_4.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:eeda98d75827843b164b71597c17f33c833e859914017510728e9834505aa273
|
3 |
+
size 27191296
|
params_shard_5.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f9abf5d9053c11f6ad700d6f7572dad2832c9f579d6f217e0998a9a6e4aed4e2
|
3 |
+
size 26331136
|
params_shard_6.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e4cd6910039baad6fce4467956a5179ca074e609dbcbebfe7261701ecf4a377a
|
3 |
+
size 26331136
|
params_shard_7.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:904b9b0c57e91b88a96c4aa7f8b47d80ec7f1ba2d0313743b85268b2c7a747ec
|
3 |
+
size 26331136
|
params_shard_8.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:141a631925d702b7e681421c972908c707e0a544095a412547eacb68eb4c2de9
|
3 |
+
size 26331136
|
params_shard_9.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ab485e831ccc43b7e10b2d0fc6880327ec14c9350c9d9157e8d54664c5871b3f
|
3 |
+
size 26331136
|
tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
ADDED
@@ -0,0 +1,35 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"add_bos_token": true,
|
3 |
+
"add_eos_token": false,
|
4 |
+
"bos_token": {
|
5 |
+
"__type": "AddedToken",
|
6 |
+
"content": "<|begin▁of▁sentence|>",
|
7 |
+
"lstrip": false,
|
8 |
+
"normalized": true,
|
9 |
+
"rstrip": false,
|
10 |
+
"single_word": false
|
11 |
+
},
|
12 |
+
"clean_up_tokenization_spaces": false,
|
13 |
+
"eos_token": {
|
14 |
+
"__type": "AddedToken",
|
15 |
+
"content": "<|end▁of▁sentence|>",
|
16 |
+
"lstrip": false,
|
17 |
+
"normalized": true,
|
18 |
+
"rstrip": false,
|
19 |
+
"single_word": false
|
20 |
+
},
|
21 |
+
"legacy": true,
|
22 |
+
"model_max_length": 16384,
|
23 |
+
"pad_token": {
|
24 |
+
"__type": "AddedToken",
|
25 |
+
"content": "<|end▁of▁sentence|>",
|
26 |
+
"lstrip": false,
|
27 |
+
"normalized": true,
|
28 |
+
"rstrip": false,
|
29 |
+
"single_word": false
|
30 |
+
},
|
31 |
+
"sp_model_kwargs": {},
|
32 |
+
"unk_token": null,
|
33 |
+
"tokenizer_class": "LlamaTokenizerFast",
|
34 |
+
"chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% set ns = namespace(is_first=false, is_tool=false, is_output_first=true, system_prompt='') %}{%- for message in messages %}{%- if message['role'] == 'system' %}{% set ns.system_prompt = message['content'] %}{%- endif %}{%- endfor %}{{bos_token}}{{ns.system_prompt}}{%- for message in messages %}{%- if message['role'] == 'user' %}{%- set ns.is_tool = false -%}{{'<|User|>' + message['content']}}{%- endif %}{%- if message['role'] == 'assistant' and message['content'] is none %}{%- set ns.is_tool = false -%}{%- for tool in message['tool_calls']%}{%- if not ns.is_first %}{{'<|Assistant|><|tool▁calls▁begin|><|tool▁call▁begin|>' + tool['type'] + '<|tool▁sep|>' + tool['function']['name'] + '\\n' + '```json' + '\\n' + tool['function']['arguments'] + '\\n' + '```' + '<|tool▁call▁end|>'}}{%- set ns.is_first = true -%}{%- else %}{{'\\n' + '<|tool▁call▁begin|>' + tool['type'] + '<|tool▁sep|>' + tool['function']['name'] + '\\n' + '```json' + '\\n' + tool['function']['arguments'] + '\\n' + '```' + '<|tool▁call▁end|>'}}{{'<|tool▁calls▁end|><|end▁of▁sentence|>'}}{%- endif %}{%- endfor %}{%- endif %}{%- if message['role'] == 'assistant' and message['content'] is not none %}{%- if ns.is_tool %}{{'<|tool▁outputs▁end|>' + message['content'] + '<|end▁of▁sentence|>'}}{%- set ns.is_tool = false -%}{%- else %}{% set content = message['content'] %}{% if '</think>' in content %}{% set content = content.split('</think>')[-1] %}{% endif %}{{'<|Assistant|>' + content + '<|end▁of▁sentence|>'}}{%- endif %}{%- endif %}{%- if message['role'] == 'tool' %}{%- set ns.is_tool = true -%}{%- if ns.is_output_first %}{{'<|tool▁outputs▁begin|><|tool▁output▁begin|>' + message['content'] + '<|tool▁output▁end|>'}}{%- set ns.is_output_first = false %}{%- else %}{{'\\n<|tool▁output▁begin|>' + message['content'] + '<|tool▁output▁end|>'}}{%- endif %}{%- endif %}{%- endfor -%}{% if ns.is_tool %}{{'<|tool▁outputs▁end|>'}}{% endif %}{% if add_generation_prompt and not ns.is_tool %}{{'<|Assistant|>'}}{% endif %}"
|
35 |
+
}
|