Change max_position_embeddings to original value
#18 opened 4 months ago
by
AshtonIsNotHere
![](https://cdn-avatars.huggingface.co/v1/production/uploads/6372d4b1112fb535baf166bb/hKuTfj5i8MpBhshN1xaK4.jpeg)
Can you provide one model using `group_size=1024` to make the model smaller?
#15 opened 6 months ago
by
shuyuej
![](https://cdn-avatars.huggingface.co/v1/production/uploads/63874ebe20244d72a740548f/pmREOHPGNwpG1_Eeif6Sv.jpeg)
optimum version cannot support llama3.1 405b
#14 opened 6 months ago
by
Atomheart-Father
Source codes to quantize the LLaMA 3.1 405B model
3
#10 opened 7 months ago
by
shuyuej
![](https://cdn-avatars.huggingface.co/v1/production/uploads/63874ebe20244d72a740548f/pmREOHPGNwpG1_Eeif6Sv.jpeg)
quantization gptq_marlin (not found gptq_marlin) not work. , remove it. work.
8
#7 opened 7 months ago
by
linpan
Accuracy tradeoff
#6 opened 7 months ago
by
shaamil101
Value Error when trying to run
2
#4 opened 7 months ago
by
itaytricks