miqu-1-70b-AQLM-2Bit-1x16-hf / configuration_llama_aqlm.py
AlexWortega's picture
Upload LlamaForCausalLM
3a86f34 verified
raw
history blame
421 Bytes
from transformers import LlamaConfig as OrigLlamaConfig
class LlamaConfig(OrigLlamaConfig):
model_type = "llama_aqlm"
def __init__(
self,
aqlm: dict[str, int] = {
"nbits_per_codebook": 16,
"num_codebooks": 1,
"out_group_size": 8,
"in_group_size": 1,
},
**kwargs,
):
super().__init__(**kwargs)
self.aqlm = aqlm