Molmo-7B-D BnB 4bit quant 30GB -> 7GB
approx. 12GB VRAM required
base model for more information:
https://huggingface.co/allenai/Molmo-7B-D-0924
example code:
https://github.com/cyan2k/molmo-7b-bnb-4bit
performance metrics & benchmarks to compare with base will follow over the next week
- Downloads last month
- 7,944
Inference API (serverless) does not yet support model repos that contain custom code.