Thank you deepseek
#8
by
teknium
- opened
that is all
The comment above is understated. I just glanced their DeepSeek R1 paper. They have:
- Proved that LLMs can be post-trained to reason with CoT without any human supervised data, with R1 zero.
- Released R1 and R1 Zero with MIT license. Explicitly allows for distillation of CoT and model responses to other models.
- Actually, they distilled R1 to 6 Qwen and Llama 3 based models with sizes between 1.5B - 70B and will opensource these as well, so that researchers with smaller compute budget can once again have a level playing field.
A great day for the global opensource and research community. These guys deserve not 1, but 10 Million GB200 GPUs. And of course, thanks again, and all the best.