Update README.md
Browse files
README.md
CHANGED
@@ -53,14 +53,20 @@ git clone https://github.com/ggerganov/llama.cpp && cd llama.cpp && make -j
|
|
53 |
```verilog
|
54 |
//the 4bit iq4xm gets better perplexity than bf16 lol but it's likely just a rounding error
|
55 |
|
56 |
-
deepseek-0628-bf16-00001-of-00011.gguf
|
57 |
Model size: 440 Gib
|
58 |
perplexity: 735.50 seconds per pass - ETA 36.77 minutes
|
59 |
[1]2.4827,[2]3.3887,[3]2.9470,[4]3.4768,[5]3.9012,[6]4.5128,[7]4.7533,[8]4.9550,[9]5.2863,[10]5.6824,[11]5.7541,[12]5.8734,
|
60 |
Final estimate: PPL = 5.8734 +/- 0.26967
|
61 |
|
62 |
-
|
63 |
-
size
|
|
|
|
|
|
|
|
|
|
|
|
|
64 |
perplexity: 59.49 seconds per pass - ETA 2.97 minutes
|
65 |
[1]2.4954,[2]3.3941,[3]2.9607,[4]3.4755,[5]3.8889,[6]4.5036,[7]4.7364,[8]4.9401,[9]5.2737,[10]5.6651,[11]5.7354,[12]5.8620,
|
66 |
Final estimate: PPL = 5.8620 +/- 0.26853
|
|
|
53 |
```verilog
|
54 |
//the 4bit iq4xm gets better perplexity than bf16 lol but it's likely just a rounding error
|
55 |
|
56 |
+
deepseek-0628-bf16-00001-of-00011.gguf //16B
|
57 |
Model size: 440 Gib
|
58 |
perplexity: 735.50 seconds per pass - ETA 36.77 minutes
|
59 |
[1]2.4827,[2]3.3887,[3]2.9470,[4]3.4768,[5]3.9012,[6]4.5128,[7]4.7533,[8]4.9550,[9]5.2863,[10]5.6824,[11]5.7541,[12]5.8734,
|
60 |
Final estimate: PPL = 5.8734 +/- 0.26967
|
61 |
|
62 |
+
deepseek-0628-q8_0-00001-of-00006.gguf //8Bit
|
63 |
+
model size = 233.41 GiB (8.50 BPW)
|
64 |
+
perplexity: 49.96 seconds per pass - ETA 2.48 minutes
|
65 |
+
[1]2.5022,[2]3.3930,[3]2.9422,[4]3.4757,[5]3.8977,[6]4.5114,[7]4.7577,[8]4.9631,[9]5.2926,[10]5.6878,[11]5.7580,[12]5.8782,
|
66 |
+
Final estimate: PPL = 5.8782 +/- 0.27021
|
67 |
+
|
68 |
+
deepseek_0628_cpu_optimized_iq4xm-00001-of-00004.gguf //4Bit
|
69 |
+
model size: 132.1 GiB
|
70 |
perplexity: 59.49 seconds per pass - ETA 2.97 minutes
|
71 |
[1]2.4954,[2]3.3941,[3]2.9607,[4]3.4755,[5]3.8889,[6]4.5036,[7]4.7364,[8]4.9401,[9]5.2737,[10]5.6651,[11]5.7354,[12]5.8620,
|
72 |
Final estimate: PPL = 5.8620 +/- 0.26853
|