Tired of waiting in queue. How to eval locally?

#1103
by Blazgo - opened

How to run same eval scripts locallly?? I don't see anything as "open and reproducible" without there being any documentation. All three README's are useless. I do see the Standalone Installation (without Docker) section but I can't seem to find out if it's just the voting/submitting/viewing app or it's the entire evaluator.

Open LLM Leaderboard org

Thank you @icefog72 !

Yes, to evaluate a model locally you can use either lm-evaluation-harness or to reproduce the Leaderboard scores as we do it, please check "Reproducibility" section in our FAQ

Feel free to ping me in case of any questions here!

alozowski changed discussion status to closed

@alozowski Will I need a CUDA-compat GPU, or can I run it on my CPU?

Sign up or log in to comment