OOM on 12 GB loading the model

#1
by mapa17 - opened

Hello,

i am running out of memory when loading the model with 12GB VRAM. What is the minimum memory needed for an AWQ quantized 7B model with group size of 128?

thx,
Manuel

Sign up or log in to comment