Update README.md
Browse filesUpdated links to quants
README.md
CHANGED
@@ -7,6 +7,7 @@ tags:
|
|
7 |
- mergekit
|
8 |
- merge
|
9 |
license: cc-by-nc-4.0
|
|
|
10 |
---
|
11 |
# kukulemon-7B-8.0bpw-h8_exl2
|
12 |
|
@@ -15,9 +16,9 @@ This is an 8.0bpw h8 exl2 quant of a merger of two similar models with strong re
|
|
15 |
I've tested with ChatML prompts with temperature=1.1 and minP=0.03. The model itself supports Alpaca format prompts. The model claims a context length of 32K, but I found it lost coherence after 8K in informal testing. I prefer to stick with 8.0bpw h8 exl2 or Q8_0 GGUF for maximum coherence.
|
16 |
|
17 |
Alternative downloads:
|
18 |
-
[GGUF quants courtesy of Lewdiculous](https://huggingface.co/Lewdiculous/kukulemon-7B-GGUF-IQ-Imatrix)
|
19 |
-
[fp16 safetensors](https://huggingface.co/grimjim/kukulemon-7B)
|
20 |
-
[
|
21 |
|
22 |
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
|
23 |
|
|
|
7 |
- mergekit
|
8 |
- merge
|
9 |
license: cc-by-nc-4.0
|
10 |
+
pipeline_tag: text-generation
|
11 |
---
|
12 |
# kukulemon-7B-8.0bpw-h8_exl2
|
13 |
|
|
|
16 |
I've tested with ChatML prompts with temperature=1.1 and minP=0.03. The model itself supports Alpaca format prompts. The model claims a context length of 32K, but I found it lost coherence after 8K in informal testing. I prefer to stick with 8.0bpw h8 exl2 or Q8_0 GGUF for maximum coherence.
|
17 |
|
18 |
Alternative downloads:
|
19 |
+
- [iMatrix GGUF quants courtesy of Lewdiculous](https://huggingface.co/Lewdiculous/kukulemon-7B-GGUF-IQ-Imatrix)
|
20 |
+
- [fp16 safetensors](https://huggingface.co/grimjim/kukulemon-7B)
|
21 |
+
- [GGUF quants](https://huggingface.co/grimjim/kukulemon-7B-GGUF)
|
22 |
|
23 |
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
|
24 |
|