Update README.md
Browse files
README.md
CHANGED
@@ -27,4 +27,6 @@ Using llama.cpp fork: [https://github.com/fairydreaming/llama.cpp/tree/deepseek-
|
|
27 |
- ~~q2_k (after q4_k_m) [estimated size: ~65gb]~~
|
28 |
- ~~q3_k_s (low priority) [estimated size: 96.05gb]~~
|
29 |
|
30 |
-
If quantize.exe supports it I will make RTN quants (edit: it doesn't).
|
|
|
|
|
|
27 |
- ~~q2_k (after q4_k_m) [estimated size: ~65gb]~~
|
28 |
- ~~q3_k_s (low priority) [estimated size: 96.05gb]~~
|
29 |
|
30 |
+
If quantize.exe supports it I will make RTN quants (edit: it doesn't, will try building from fork).
|
31 |
+
|
32 |
+
Note: the bf16 GGUF does not have some DeepSeek v2 specific parameters, will look into adding them
|