|
--- |
|
license: apache-2.0 |
|
tags: |
|
- gguf |
|
--- |
|
GGUF [llama.cpp](https://github.com/ggerganov/llama.cpp) quantized version of: |
|
- Original model: [Mistral-7B-Instruct-v0.3](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.3) |
|
- Model creator: [Mistral AI](https://huggingface.co/mistralai) |
|
- [License](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/apache-2.0.md) |
|
|
|
## Recommended Prompt Format (Mistral) |
|
``` |
|
<s>[INST]Provide some context and/or instructions to the model.[/INST] |
|
AI message goes here</s> |
|
[INST] The user’s message goes here [/INST] |
|
``` |
|
|
|
Quant Version: [b3437](https://github.com/ggerganov/llama.cpp/releases/tag/b3437) with [imatrix](https://github.com/ggerganov/llama.cpp/discussions/5263#discussioncomment-8395384) |