Thank you deepseek

#8
by teknium - opened

that is all

Same 🤝
Screenshot 2025-01-20 154703.jpg

The comment above is understated. I just glanced their DeepSeek R1 paper. They have:

  • Proved that LLMs can be post-trained to reason with CoT without any human supervised data, with R1 zero.
  • Released R1 and R1 Zero with MIT license. Explicitly allows for distillation of CoT and model responses to other models.
  • Actually, they distilled R1 to 6 Qwen and Llama 3 based models with sizes between 1.5B - 70B and will opensource these as well, so that researchers with smaller compute budget can once again have a level playing field.

A great day for the global opensource and research community. These guys deserve not 1, but 10 Million GB200 GPUs. And of course, thanks again, and all the best.

Sign up or log in to comment