
TIGER-Lab/AceCodeRM-7B
Updated
•
248
•
3
Note The state-of-the-art 7B reward model for code generation
Note The state-of-the-art 32B reward model for code generation
Note The first large-scale coding dataset with an average of 16 test cases per prompt, synthesized by GPT-4o-mini
Note DeepSeek-R1 style RL-tuned model with binary pass rate as the verifiable reward