Mid-range models exl2 quants
Collection
Exl2 quants of mid-range (20-40B) LLM models. Usually around 4-5 BPW
•
8 items
•
Updated
This is a 5.6bpw EXL2 quant of TheDrummer/Big-Tiger-Gemma-27B-v1
This quant was made using exllamav2-0.1.8 with default dataset.
I tested this quant shortly in some random RPs and a few assistant-type tasks and it seems to work fine.
This quant fits nicely in 24GB VRAM on Windows with no KV cache quantization (using standard 8k context length)
Seems to use Gemma 2 format.
Decensored Gemma 27B. No refusals so far (other than some rare instances from 9B). No apparent brain damage.
In memory of Tiger (the happy street cat on the right)