metadata
tags:
- not-for-all-audiences
- nsfw
license: other
language:
- en
EXL2 Quantization of Undi95's's MXLewd-L2-20B.
Model details
First attempt to quantize a 20B model so it can run on 16GB VRAM with the highest quality possible. Quantized at 3.23bpw with hb 6
Perplexity:
Base = 6.4744
3.23 h6 = 6.5369
Dataset = wikitext
Prompt Format
Below is an instruction that describes a task. Write a response that appropriately completes the request.
### Instruction:
{prompt}
### Response: