Update models
Browse files- .gitattributes +13 -0
- Mistral-Small-24B-Instruct-2501-Q2_K.gguf +3 -0
- Mistral-Small-24B-Instruct-2501-Q3_K_L.gguf +3 -0
- Mistral-Small-24B-Instruct-2501-Q3_K_M.gguf +3 -0
- Mistral-Small-24B-Instruct-2501-Q3_K_S.gguf +3 -0
- Mistral-Small-24B-Instruct-2501-Q4_0.gguf +3 -0
- Mistral-Small-24B-Instruct-2501-Q4_K_M.gguf +3 -0
- Mistral-Small-24B-Instruct-2501-Q4_K_S.gguf +3 -0
- Mistral-Small-24B-Instruct-2501-Q5_0.gguf +3 -0
- Mistral-Small-24B-Instruct-2501-Q5_K_M.gguf +3 -0
- Mistral-Small-24B-Instruct-2501-Q5_K_S.gguf +3 -0
- Mistral-Small-24B-Instruct-2501-Q6_K.gguf +3 -0
- Mistral-Small-24B-Instruct-2501-Q8_0.gguf +3 -0
- Mistral-Small-24B-Instruct-2501-f16.gguf +3 -0
- config.json +26 -0
.gitattributes
CHANGED
@@ -33,3 +33,16 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
Mistral-Small-24B-Instruct-2501-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
|
37 |
+
Mistral-Small-24B-Instruct-2501-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
|
38 |
+
Mistral-Small-24B-Instruct-2501-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
39 |
+
Mistral-Small-24B-Instruct-2501-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
40 |
+
Mistral-Small-24B-Instruct-2501-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
|
41 |
+
Mistral-Small-24B-Instruct-2501-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
42 |
+
Mistral-Small-24B-Instruct-2501-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
43 |
+
Mistral-Small-24B-Instruct-2501-Q5_0.gguf filter=lfs diff=lfs merge=lfs -text
|
44 |
+
Mistral-Small-24B-Instruct-2501-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
45 |
+
Mistral-Small-24B-Instruct-2501-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
46 |
+
Mistral-Small-24B-Instruct-2501-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
|
47 |
+
Mistral-Small-24B-Instruct-2501-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
|
48 |
+
Mistral-Small-24B-Instruct-2501-f16.gguf filter=lfs diff=lfs merge=lfs -text
|
Mistral-Small-24B-Instruct-2501-Q2_K.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:857fdf52140fcc91db4b1d3ddc0332d65a7f69a15288c706f592b858ce4e62a1
|
3 |
+
size 8890324384
|
Mistral-Small-24B-Instruct-2501-Q3_K_L.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:19846a991a1eb10103c456bd63c490ef0dc7dd913ed3b37d1eb5965a8ad757e0
|
3 |
+
size 12400760224
|
Mistral-Small-24B-Instruct-2501-Q3_K_M.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9489cf4eee82e0819524f0e28778d3a6484e64e2e5f88b1cf3369effc5012ee4
|
3 |
+
size 11474081184
|
Mistral-Small-24B-Instruct-2501-Q3_K_S.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:73477e921a77cde3e73bdacb74e13d451b3ea56feef90bd2c716c40464e7c71b
|
3 |
+
size 10400273824
|
Mistral-Small-24B-Instruct-2501-Q4_0.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c991b3e82deae411f3d3182363f04ae499dd5059eece0d5f60101f4da3c81d59
|
3 |
+
size 13441799584
|
Mistral-Small-24B-Instruct-2501-Q4_K_M.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:41ad86031ea06210ba7790d6a4742f8ad2ab39984d83254c24869822a72e64f2
|
3 |
+
size 14333908384
|
Mistral-Small-24B-Instruct-2501-Q4_K_S.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1e9b4f8d42a2c28ae7fd4bf5a8c7d872399b0b48b4b81480d4a37000eb00be81
|
3 |
+
size 13549278624
|
Mistral-Small-24B-Instruct-2501-Q5_0.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a7f3088616569a3ac9a575e7db30ba7479afa249e9b365b0c2e591f6651bad65
|
3 |
+
size 16304412064
|
Mistral-Small-24B-Instruct-2501-Q5_K_M.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c62374efe76ede58f43367f5f273c8260ddb50b0496917c119cf8e36dbf45d76
|
3 |
+
size 16763983264
|
Mistral-Small-24B-Instruct-2501-Q5_K_S.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:73b03f98a78d5f21abef01ae6f378f2908d8699674ff43b29823f9e65d5024a4
|
3 |
+
size 16304412064
|
Mistral-Small-24B-Instruct-2501-Q6_K.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b91dfb3e9264e2ba61e9890b4f266d92fd78bb85ec0aa01318b0108c4c60b3b6
|
3 |
+
size 19345937824
|
Mistral-Small-24B-Instruct-2501-Q8_0.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1b5f926b6e87e443949ed30327a93a3792492963e570457228366d86705433eb
|
3 |
+
size 25054778784
|
Mistral-Small-24B-Instruct-2501-f16.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ab5720f26eb3e8f22346ee7af5834a72019e77cd7806fc07c757e75e7b44c4b8
|
3 |
+
size 47153517984
|
config.json
ADDED
@@ -0,0 +1,26 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"architectures": [
|
3 |
+
"MistralForCausalLM"
|
4 |
+
],
|
5 |
+
"attention_dropout": 0.0,
|
6 |
+
"bos_token_id": 1,
|
7 |
+
"eos_token_id": 2,
|
8 |
+
"head_dim": 128,
|
9 |
+
"hidden_act": "silu",
|
10 |
+
"hidden_size": 5120,
|
11 |
+
"initializer_range": 0.02,
|
12 |
+
"intermediate_size": 32768,
|
13 |
+
"max_position_embeddings": 32768,
|
14 |
+
"model_type": "mistral",
|
15 |
+
"num_attention_heads": 32,
|
16 |
+
"num_hidden_layers": 40,
|
17 |
+
"num_key_value_heads": 8,
|
18 |
+
"rms_norm_eps": 1e-05,
|
19 |
+
"rope_theta": 100000000.0,
|
20 |
+
"sliding_window": null,
|
21 |
+
"tie_word_embeddings": false,
|
22 |
+
"torch_dtype": "bfloat16",
|
23 |
+
"transformers_version": "4.49.0.dev0",
|
24 |
+
"use_cache": true,
|
25 |
+
"vocab_size": 131072
|
26 |
+
}
|