m33393 commited on
Commit
73b449d
1 Parent(s): 074723c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -12,5 +12,5 @@ Converted to HF with `transformers 4.30.0.dev0`, then quantized to 4 bit with GP
12
  PPL should be marginally better than group size 128 at the cost of more VRAM. An A6000 should still be able to fit it all at full 2048 context.
13
 
14
  ---
15
- Note that this model is quantized under GPTQ's `cuda` branch. Which means this model should work with 0cc4m's KoboldAI fork:
16
  https://github.com/0cc4m/KoboldAI
 
12
  PPL should be marginally better than group size 128 at the cost of more VRAM. An A6000 should still be able to fit it all at full 2048 context.
13
 
14
  ---
15
+ Note that this model was quantized under GPTQ's `cuda` branch. Which means it should work with 0cc4m's KoboldAI fork:
16
  https://github.com/0cc4m/KoboldAI