Upload folder using huggingface_hub
Browse files- README.md +6 -0
- added_tokens.json +5 -0
- merges.txt +0 -0
- mlc-chat-config.json +80 -0
- model.wasm +3 -0
- ndarray-cache.json +0 -0
- params_shard_0.bin +3 -0
- params_shard_1.bin +3 -0
- params_shard_10.bin +3 -0
- params_shard_11.bin +3 -0
- params_shard_12.bin +3 -0
- params_shard_13.bin +3 -0
- params_shard_14.bin +3 -0
- params_shard_15.bin +3 -0
- params_shard_16.bin +3 -0
- params_shard_17.bin +3 -0
- params_shard_18.bin +3 -0
- params_shard_19.bin +3 -0
- params_shard_2.bin +3 -0
- params_shard_20.bin +3 -0
- params_shard_21.bin +3 -0
- params_shard_22.bin +3 -0
- params_shard_23.bin +3 -0
- params_shard_24.bin +3 -0
- params_shard_25.bin +3 -0
- params_shard_26.bin +3 -0
- params_shard_27.bin +3 -0
- params_shard_28.bin +3 -0
- params_shard_29.bin +3 -0
- params_shard_3.bin +3 -0
- params_shard_4.bin +3 -0
- params_shard_5.bin +3 -0
- params_shard_6.bin +3 -0
- params_shard_7.bin +3 -0
- params_shard_8.bin +3 -0
- params_shard_9.bin +3 -0
- tokenizer.json +0 -0
- tokenizer_config.json +43 -0
- vocab.json +0 -0
README.md
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
base_model: arcee-ai/arcee-lite
|
3 |
+
---
|
4 |
+
|
5 |
+
[MLC](https://llm.mlc.ai/) version of [arcee-ai/arcee-lite](https://huggingface.co/arcee-ai/arcee-lite), using `q4f16_1` quantization.
|
6 |
+
|
added_tokens.json
ADDED
@@ -0,0 +1,5 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"<|endoftext|>": 151643,
|
3 |
+
"<|im_end|>": 151645,
|
4 |
+
"<|im_start|>": 151644
|
5 |
+
}
|
merges.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
mlc-chat-config.json
ADDED
@@ -0,0 +1,80 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"version": "0.1.0",
|
3 |
+
"model_type": "qwen2",
|
4 |
+
"quantization": "q4f16_1",
|
5 |
+
"model_config": {
|
6 |
+
"hidden_act": "silu",
|
7 |
+
"hidden_size": 1536,
|
8 |
+
"intermediate_size": 8960,
|
9 |
+
"num_attention_heads": 12,
|
10 |
+
"num_hidden_layers": 28,
|
11 |
+
"num_key_value_heads": 2,
|
12 |
+
"rms_norm_eps": 1e-06,
|
13 |
+
"rope_theta": 1000000.0,
|
14 |
+
"vocab_size": 151936,
|
15 |
+
"tie_word_embeddings": true,
|
16 |
+
"context_window_size": 32768,
|
17 |
+
"prefill_chunk_size": 2048,
|
18 |
+
"tensor_parallel_shards": 1,
|
19 |
+
"head_dim": 128,
|
20 |
+
"dtype": "float32",
|
21 |
+
"max_batch_size": 80
|
22 |
+
},
|
23 |
+
"vocab_size": 151936,
|
24 |
+
"context_window_size": 32768,
|
25 |
+
"sliding_window_size": -1,
|
26 |
+
"prefill_chunk_size": 2048,
|
27 |
+
"attention_sink_size": -1,
|
28 |
+
"tensor_parallel_shards": 1,
|
29 |
+
"pipeline_parallel_stages": 1,
|
30 |
+
"temperature": 1.0,
|
31 |
+
"presence_penalty": 0.0,
|
32 |
+
"frequency_penalty": 0.0,
|
33 |
+
"repetition_penalty": 1.0,
|
34 |
+
"top_p": 1.0,
|
35 |
+
"tokenizer_files": [
|
36 |
+
"tokenizer.json",
|
37 |
+
"vocab.json",
|
38 |
+
"merges.txt",
|
39 |
+
"added_tokens.json",
|
40 |
+
"tokenizer_config.json"
|
41 |
+
],
|
42 |
+
"tokenizer_info": {
|
43 |
+
"token_postproc_method": "byte_level",
|
44 |
+
"prepend_space_in_encode": false,
|
45 |
+
"strip_space_in_decode": false
|
46 |
+
},
|
47 |
+
"conv_template": {
|
48 |
+
"name": "chatml",
|
49 |
+
"system_template": "<|im_start|>system\n{system_message}<|im_end|>\n",
|
50 |
+
"system_message": "A conversation between a user and an LLM-based AI assistant. The assistant gives helpful and honest answers.",
|
51 |
+
"system_prefix_token_ids": null,
|
52 |
+
"add_role_after_system_message": true,
|
53 |
+
"roles": {
|
54 |
+
"user": "<|im_start|>user",
|
55 |
+
"assistant": "<|im_start|>assistant"
|
56 |
+
},
|
57 |
+
"role_templates": {
|
58 |
+
"user": "{user_message}",
|
59 |
+
"assistant": "{assistant_message}",
|
60 |
+
"tool": "{tool_message}"
|
61 |
+
},
|
62 |
+
"messages": [],
|
63 |
+
"seps": [
|
64 |
+
"<|im_end|>\n"
|
65 |
+
],
|
66 |
+
"role_content_sep": "\n",
|
67 |
+
"role_empty_sep": "\n",
|
68 |
+
"stop_str": [
|
69 |
+
"<|im_end|>"
|
70 |
+
],
|
71 |
+
"stop_token_ids": [
|
72 |
+
2
|
73 |
+
],
|
74 |
+
"function_string": "",
|
75 |
+
"use_function_calling": false
|
76 |
+
},
|
77 |
+
"pad_token_id": 0,
|
78 |
+
"bos_token_id": 151643,
|
79 |
+
"eos_token_id": 151645
|
80 |
+
}
|
model.wasm
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7f8f86862754427ffe6c8662c1a7baf07ab9b7eea07b0d244f1bc2f5afd5275c
|
3 |
+
size 4252774
|
ndarray-cache.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
params_shard_0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2c5dfb575a434baf7503d3e448454c3c80749d9a744cc2e0e84155d38915f0be
|
3 |
+
size 116686848
|
params_shard_1.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b9286de77eea236058d4fd432db9802e53bc41273655fc62e01417dff94495f5
|
3 |
+
size 22330368
|
params_shard_10.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1d447aba0019d3e8e1dbd7a8343643d445a42548e8d0582952b0cb1c35c91f6a
|
3 |
+
size 26331136
|
params_shard_11.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3ce05bf20751799957c031d486ae7aa1a6c79a90fae18f8e8a0f650b85519385
|
3 |
+
size 26331136
|
params_shard_12.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d1f12c454f8eaffdc443d454495f6f61aa541084a66b915634f58c5fa096b227
|
3 |
+
size 26331136
|
params_shard_13.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:01d13b5f466862d613aff9ff87796cd7503fd1b04b79498aab451f490c524983
|
3 |
+
size 26331136
|
params_shard_14.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:306bf3132ce4d7e3aea3cd9c9412a4158368982485f3e6309d09058fb46837a7
|
3 |
+
size 26331136
|
params_shard_15.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2888d56cee11ffc000101927637044d8c387a71526eec1a87e2b8d1051834fe5
|
3 |
+
size 26331136
|
params_shard_16.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fbe4160a7d1f4ec68e7be5274dc18e29d532ca4c4b465dc98fe2351a79e313c0
|
3 |
+
size 26331136
|
params_shard_17.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:83198c94051fb8fab6131bdec3ea53d3ae260dd0e1b3a91e03b452cb3ac59114
|
3 |
+
size 26331136
|
params_shard_18.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dbadf5a7317d01753f68f13fb24c014852c75a00945a64f7cef6ddfa2328916a
|
3 |
+
size 26331136
|
params_shard_19.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7e340bcd1b5a743c4a815dcc06c69069fd20b99bc348e3c6eab818a494496c2e
|
3 |
+
size 26331136
|
params_shard_2.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c8cdda97d07d3ddb2098dfec71a2b284c07b84abb07e8b180d3f7f71f22a6c63
|
3 |
+
size 26331136
|
params_shard_20.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fb4cb7ba5746fc2585a60468aac9506bf0e00f1513b5e2fa12e596f431d00f19
|
3 |
+
size 26331136
|
params_shard_21.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:aa09ab5baf75024cd05de2995ada093881f0f4613d83c343b695af7fc2fe74be
|
3 |
+
size 26331136
|
params_shard_22.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a297e27aaf22e317592eaaf532a721f7bfd945cdc76aebd51e137e0bda33301b
|
3 |
+
size 26331136
|
params_shard_23.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9520a3f6e3585f781fa06515f3630626dac3c8b4d89b35d33e6447f22564cdc6
|
3 |
+
size 26331136
|
params_shard_24.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:73f7498cb96bf18d51d7f5d502115d7304c7dfd8d3da2d2800651eeb4341b263
|
3 |
+
size 26331136
|
params_shard_25.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a5d65936857974b872a3542c05e8d326c9da0d85dee5dc60f662f3f9bb9f2506
|
3 |
+
size 26331136
|
params_shard_26.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fcc227323563bf658f96cbe2ba23105266972e0fb2d86aaf2e84f37f92cfc5b2
|
3 |
+
size 26331136
|
params_shard_27.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c30bb7b3e787ed2b7ed0b48256e1a8b191f60a952afb261c2293e600fa4ce6b9
|
3 |
+
size 26331136
|
params_shard_28.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:13ac5590ea5beb04fce2687da1e4bdb823b63995811ed70312c8a2bb997186e9
|
3 |
+
size 26331136
|
params_shard_29.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7f73db55c55a4f5823e3208a56d635653f190173cf5865d4a8c2371683c2821e
|
3 |
+
size 18589696
|
params_shard_3.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:076324acd41822e1cf69b1848bf67cf35e6314d7953455f2400acebe658b91e2
|
3 |
+
size 26331136
|
params_shard_4.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:87022d378ab8e1d4aa2bbf73f666ed5f4a469fef14c98993d512f2e861fc0680
|
3 |
+
size 26331136
|
params_shard_5.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3e76781ba6d018f33db6bdd750b28b9c2c1c9ace8c671ecbcc0bc71b89ed4d18
|
3 |
+
size 26331136
|
params_shard_6.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dd9da31a9a7e89550c4da2e01a1135a138fa82bf7d9ad9068f4abb4558488216
|
3 |
+
size 26331136
|
params_shard_7.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8025033084dbaf0e4281865666a04ee667bc62293af47ff3c06ce9a8f0e7b811
|
3 |
+
size 26331136
|
params_shard_8.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c93c3eb1d26facab213c53100ca8fda9c9c51122dbad711928786bb45b5b8078
|
3 |
+
size 26331136
|
params_shard_9.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a82e72a29340cc44782dc138e4db796cb8095e026f053cab6db8e7fbccd6f02f
|
3 |
+
size 26331136
|
tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
ADDED
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"add_prefix_space": false,
|
3 |
+
"added_tokens_decoder": {
|
4 |
+
"151643": {
|
5 |
+
"content": "<|endoftext|>",
|
6 |
+
"lstrip": false,
|
7 |
+
"normalized": false,
|
8 |
+
"rstrip": false,
|
9 |
+
"single_word": false,
|
10 |
+
"special": true
|
11 |
+
},
|
12 |
+
"151644": {
|
13 |
+
"content": "<|im_start|>",
|
14 |
+
"lstrip": false,
|
15 |
+
"normalized": false,
|
16 |
+
"rstrip": false,
|
17 |
+
"single_word": false,
|
18 |
+
"special": true
|
19 |
+
},
|
20 |
+
"151645": {
|
21 |
+
"content": "<|im_end|>",
|
22 |
+
"lstrip": false,
|
23 |
+
"normalized": false,
|
24 |
+
"rstrip": false,
|
25 |
+
"single_word": false,
|
26 |
+
"special": true
|
27 |
+
}
|
28 |
+
},
|
29 |
+
"additional_special_tokens": [
|
30 |
+
"<|im_start|>",
|
31 |
+
"<|im_end|>"
|
32 |
+
],
|
33 |
+
"bos_token": null,
|
34 |
+
"chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
|
35 |
+
"clean_up_tokenization_spaces": false,
|
36 |
+
"eos_token": "<|im_end|>",
|
37 |
+
"errors": "replace",
|
38 |
+
"model_max_length": 32768,
|
39 |
+
"pad_token": "<|endoftext|>",
|
40 |
+
"split_special_tokens": false,
|
41 |
+
"tokenizer_class": "Qwen2Tokenizer",
|
42 |
+
"unk_token": null
|
43 |
+
}
|
vocab.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|