Update README.md
Browse files
README.md
CHANGED
@@ -23,6 +23,10 @@ tags:
|
|
23 |
</div>
|
24 |
<!-- header end -->
|
25 |
|
|
|
|
|
|
|
|
|
26 |
# Mikael110's Llama2 70b Guanaco QLoRA fp16
|
27 |
|
28 |
These files are pytorch format fp16 model files for [Mikael110's Llama2 70b Guanaco QLoRA](https://huggingface.co/Mikael110/llama-2-70b-guanaco-qlora).
|
@@ -32,8 +36,9 @@ It is the result of merging and/or converting the source repository to float16.
|
|
32 |
## Repositories available
|
33 |
|
34 |
* [GPTQ models for GPU inference, with multiple quantisation parameter options.](https://huggingface.co/TheBloke/llama-2-70b-Guanaco-QLoRA-GPTQ)
|
35 |
-
* [
|
36 |
-
* [
|
|
|
37 |
|
38 |
## Prompt template: Guanaco
|
39 |
|
|
|
23 |
</div>
|
24 |
<!-- header end -->
|
25 |
|
26 |
+
# Llama2 70b Guanaco QLoRA - fp16
|
27 |
+
- Model creator: [Mikael110](https://huggingface.co/Mikael110)
|
28 |
+
- Original model: [Llama2 70b Guanaco QLoRA](https://huggingface.co/Mikael110/llama-2-70b-guanaco-qlora)
|
29 |
+
|
30 |
# Mikael110's Llama2 70b Guanaco QLoRA fp16
|
31 |
|
32 |
These files are pytorch format fp16 model files for [Mikael110's Llama2 70b Guanaco QLoRA](https://huggingface.co/Mikael110/llama-2-70b-guanaco-qlora).
|
|
|
36 |
## Repositories available
|
37 |
|
38 |
* [GPTQ models for GPU inference, with multiple quantisation parameter options.](https://huggingface.co/TheBloke/llama-2-70b-Guanaco-QLoRA-GPTQ)
|
39 |
+
* [2, 3, 4, 5, 6 and 8-bit GGML models for CPU+GPU inference](https://huggingface.co/TheBloke/llama-2-70b-Guanaco-QLoRA-GGML)
|
40 |
+
* [Merged fp16 model, for GPU inference and further conversions](https://huggingface.co/TheBloke/llama-2-70b-Guanaco-QLoRA-fp16)
|
41 |
+
* [Mikael110's original QLoRA adapter](https://huggingface.co/Mikael110/llama-2-70b-guanaco-qlora)
|
42 |
|
43 |
## Prompt template: Guanaco
|
44 |
|