Update README.md
Browse files
README.md
CHANGED
@@ -29,14 +29,14 @@ These files are pytorch format fp16 model files for [Mikael10's Llama2 70b Guana
|
|
29 |
|
30 |
It is the result of merging and/or converting the source repository to float16.
|
31 |
|
32 |
-
|
33 |
## Repositories available
|
34 |
|
35 |
* [GPTQ models for GPU inference, with multiple quantisation parameter options.](https://huggingface.co/TheBloke/llama-2-70b-Guanaco-QLoRA-GPTQ)
|
36 |
-
* [
|
37 |
-
* [
|
|
|
38 |
|
39 |
-
## Prompt template:
|
40 |
|
41 |
```
|
42 |
### Human: {prompt}
|
|
|
29 |
|
30 |
It is the result of merging and/or converting the source repository to float16.
|
31 |
|
|
|
32 |
## Repositories available
|
33 |
|
34 |
* [GPTQ models for GPU inference, with multiple quantisation parameter options.](https://huggingface.co/TheBloke/llama-2-70b-Guanaco-QLoRA-GPTQ)
|
35 |
+
* [GGML experimental 4, 5, 6 and 8-bit models for CPU only inference](https://huggingface.co/TheBloke/llama-2-70b-Guanaco-QLoRA-GGML)
|
36 |
+
* [Merged fp16 model in pytorch model format for GPU inference and further conversions](https://huggingface.co/TheBloke/llama-2-70b-Guanaco-QLoRA-fp16)
|
37 |
+
* [Original QLoRA model](https://huggingface.co/Mikael110/llama-2-70b-guanaco-qlora)
|
38 |
|
39 |
+
## Prompt template: Guanaco
|
40 |
|
41 |
```
|
42 |
### Human: {prompt}
|