Adding `safetensors` variant of this model
#29 opened 10 months ago
by
SFconvertbot
![](https://cdn-avatars.huggingface.co/v1/production/uploads/635fd4cc14657fb8cff2a081/GDkyDwAcuqDBpaOvQgJuq.png)
How to use all gpu?
#19 opened over 1 year ago
by
fuliu2023
How to only get the model answer as output
#18 opened over 1 year ago
by
ibrim
Why inference is very slow?
#17 opened over 1 year ago
by
hanswang73
mosaicml/mpt-30b-chat on sagemaker ml.p3.8xlarge
1
#16 opened over 1 year ago
by
markdoucette
How to Finetune mpt-30b-chat?
#15 opened over 1 year ago
by
nandakishorej8
Decreased performance with recent updated model?
4
#14 opened over 1 year ago
by
Roy-Shih
The model keeps generating up to the maximum length but no EOS token.
1
#13 opened over 1 year ago
by
xianf
could yo provide a prompt template
#11 opened over 1 year ago
by
enzii
DIfferent results here in the chat and locally
2
#10 opened over 1 year ago
by
KarBik
how to load the model with multiple GPUs
#9 opened over 1 year ago
by
Sven00
How many GPUs for this model to run this fast?
3
#8 opened over 1 year ago
by
edureisMD
The model is extremelly slow in 4bit, is my code for loading ok?
#7 opened over 1 year ago
by
zokica
How to run on Colab's CPU?
7
#4 opened over 1 year ago
by
deepakkaura26
![](https://cdn-avatars.huggingface.co/v1/production/uploads/63c8ef6c00104ea998d92645/8zVt_tzR2fPgk7s6dA03k.jpeg)
Sagemaker endpoint inference script help?
1
#3 opened over 1 year ago
by
varuntejay