nisten commited on
Commit
434fbe3
·
verified ·
1 Parent(s): c7ec7b4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +19 -2
README.md CHANGED
@@ -2,8 +2,11 @@
2
  base_model: HuggingFaceH4/zephyr-orpo-141b-A35b-v0.1
3
  license: apache-2.0
4
  ---
 
5
 
6
- # Importance-Matrix quantizations of HuggingFaceH4/zephyr-orpo-141b-A35b-v0.1 first mixtral8x22b finetune 💫
 
 
7
 
8
  the imatrix.dat file was calcuated over 1000 chunks with wikitext.train.raw( included )
9
 
@@ -29,4 +32,18 @@ git clone https://github.com/ggerganov/llama.cpp && cd llama.cpp/ && make -j
29
 
30
  Command I used to run these on 48 core CPU only machine, you can add -ngl 16 to offload 16 layers or more to gpu on your own.
31
 
32
- ```./perplexity -m ~/orpo4ns.gguf -f wiki.test.raw --chunks 12 -t 48 ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  base_model: HuggingFaceH4/zephyr-orpo-141b-A35b-v0.1
3
  license: apache-2.0
4
  ---
5
+ ## orpo4ns.gguf is good to go, 2bit also done but not recommended, other quants STILL UPLOADING.
6
 
7
+
8
+ # Importance-Matrix quantizations of HuggingFaceH4/zephyr-orpo-141b-A35b-v0.1
9
+ # first mixtral8x22b finetune 💫
10
 
11
  the imatrix.dat file was calcuated over 1000 chunks with wikitext.train.raw( included )
12
 
 
32
 
33
  Command I used to run these on 48 core CPU only machine, you can add -ngl 16 to offload 16 layers or more to gpu on your own.
34
 
35
+ ```./perplexity -m ~/orpo4ns.gguf -f wiki.test.raw --chunks 12 -t 48 ```
36
+
37
+ ```bash
38
+ orpor4ns.gguf 71260MB
39
+ [1]2.6970,[2]3.1781,[3]3.7390,[4]3.4159,[5]2.8977,[6]2.7126,[7]2.5597,[8]2.5013,[9]2.5279,[10]2.5175,[11]2.5315,[12]2.5455,
40
+ Final estimate: PPL = 2.5455 +/- 0.07697
41
+
42
+ orpo2ns.gguf 44026MB
43
+ [1]3.0077,[2]3.5575,[3]4.1028,[4]4.4088,[5]4.2206,[6]4.1056,[7]4.1029,[8]4.1305,[9]4.1791,[10]4.3247,[11]4.4759,[12]4.4659,
44
+ Final estimate: PPL = 4.4659 +/- 0.16582
45
+
46
+ orpo2n.gguf 49420MB
47
+ [1]3.0082,[2]3.5829,[3]4.1414,[4]4.1671,[5]3.8567,[6]3.7209,[7]3.7150,[8]3.7210,[9]3.8445,[10]3.9332,[11]4.0879,[12]4.0884,
48
+ Final estimate: PPL = 4.0884 +/- 0.1499
49
+ ```