Based on https://huggingface.co/minhngh/CodeCapybara
Model by https://github.com/AI4Code-Research/CodeCapybara
Using https://github.com/qwopqwop200/GPTQ-for-LLaMa triton branch
python llama.py CodeCapybara/ c4 --wbits 4 --true-sequential --act-order --groupsize 128 --save_safetensors codecapybara-4bit-128g-gptq.safetensors