Yi Cui

onekq

AI & ML interests

Benchmark, Code Generation Model

Recent Activity

Articles

Organizations

MLX Community's profile picture ONEKQ AI's profile picture

onekq's activity

posted an update 1 day ago
view post
Post
1272
So πŸ‹DeepSeekπŸ‹ hits the mainstream media. But it has been a star in our little cult for at least 6 months. Its meteoric success is not overnight, but two years in the making.

To learn their history, just look at their πŸ€— repo https://huggingface.co/deepseek-ai

* End of 2023, they launched the first model (pretrained by themselves) following Llama 2 architecture
* June 2024, v2 (MoE architecture) surpassed Gemini 1.5, but behind Mistral
* September, v2.5 surpassed GPT 4o mini
* December, v3 surpassed GPT 4o
* Now R1 surpassed o1

Most importantly, if you think DeepSeek success is singular and unrivaled, that's WRONG. The following models are also near or equal the o1 bar.

* Minimax-01
* Kimi k1.5
* Doubao 1.5 pro
reacted to clem's post with πŸ”₯ 1 day ago
replied to their post 2 days ago
view reply

My conclusion is the same. The R1 paper already reported lower success rate of the distilled models. This is not surprising since we cannot expect the same outcomes out of a much smaller model.

Here is the problem. The small models released by frontier labs are always generic, i.e. decent but lower performance than the flagship model on every benchmark. But we GPU deplorables often want a specialized model which is excellent on only one thing, hence the disappointment.

I guess we will have to help ourselves on this one. Distill an opinionated dataset from the flagship model to a small model of your choice, then hill climb the benchmark you care about.

replied to their post 3 days ago
view reply

1000% agree.

Also reasoning models sure spit out lots of tokens. The same benchmark cost 4x or 5x the money and time to run than regular LLMs. Exciting time for inference players.

Have you tried the distilled models of R1(Qwen and Llama)?

replied to their post 4 days ago
view reply

+1

Also the velocity of progress. I have wanted to learn Monte Carlo Tree Search and process rewards etc. and haven't got the time. I guess now I can skip them πŸ€—

posted an update 5 days ago
view post
Post
2560
This is historical. πŸŽ‰

DeepSeek πŸ‹R1πŸ‹ surpassed OpenAI πŸ“o1πŸ“ on the dual leaderboard. What a year for the open source!

onekq-ai/WebApp1K-models-leaderboard
posted an update 6 days ago
view post
Post
4556
πŸ‹DeepSeek πŸ‹ is the real OpenAI 😯
Β·
posted an update 12 days ago