feihu.hf commited on
Commit
7c0a8dc
·
1 Parent(s): 2ebdfe8

update readme

Browse files
Files changed (2) hide show
  1. README.md +8 -1
  2. figures/benchmark.jpg +0 -0
README.md CHANGED
@@ -19,6 +19,11 @@ tags:
19
 
20
  QwQ is the reasoning model of the Qwen series. Compared with conventional instruction-tuned models, QwQ, which is capable of thinking and reasoning, can achieve significantly enhanced performance in downstream tasks, especially hard problems. QwQ-32B is the medium-sized reasoning model, which is capable of achieving competitive performance against state-of-the-art reasoning models, e.g., DeepSeek-R1, o1-mini.
21
 
 
 
 
 
 
22
  **This repo contains the QwQ 32B model**, which has the following features:
23
  - Type: Causal Language Models
24
  - Training Stage: Pretraining & Post-training (Supervised Finetuning and Reinforcement Learning)
@@ -31,6 +36,8 @@ QwQ is the reasoning model of the Qwen series. Compared with conventional instru
31
 
32
  **Note:** For the best experience, please review the [usage guidelines](#usage-guidelines) before deploying QwQ models.
33
 
 
 
34
  For more details, please refer to our [blog](https://qwenlm.github.io/blog/qwq-32b/), [GitHub](https://github.com/QwenLM/Qwen2.5), and [Documentation](https://qwen.readthedocs.io/en/latest/).
35
 
36
  ## Requirements
@@ -126,7 +133,7 @@ If you find our work helpful, feel free to give us a cite.
126
 
127
  ```
128
  @misc{qwq32b,
129
- title = {Qwen2.5: A Party of Foundation Models},
130
  url = {https://qwenlm.github.io/blog/qwq-32b/},
131
  author = {Qwen Team},
132
  month = {March},
 
19
 
20
  QwQ is the reasoning model of the Qwen series. Compared with conventional instruction-tuned models, QwQ, which is capable of thinking and reasoning, can achieve significantly enhanced performance in downstream tasks, especially hard problems. QwQ-32B is the medium-sized reasoning model, which is capable of achieving competitive performance against state-of-the-art reasoning models, e.g., DeepSeek-R1, o1-mini.
21
 
22
+ <p align="center">
23
+ <img width="100%" src="figures/benchmark.jpg">
24
+ </p>
25
+
26
+
27
  **This repo contains the QwQ 32B model**, which has the following features:
28
  - Type: Causal Language Models
29
  - Training Stage: Pretraining & Post-training (Supervised Finetuning and Reinforcement Learning)
 
36
 
37
  **Note:** For the best experience, please review the [usage guidelines](#usage-guidelines) before deploying QwQ models.
38
 
39
+ You can try our [demo](https://huggingface.co/spaces/Qwen/QwQ-32B-Demo) or access QwQ models via [QwenChat](https://chat.qwen.ai).
40
+
41
  For more details, please refer to our [blog](https://qwenlm.github.io/blog/qwq-32b/), [GitHub](https://github.com/QwenLM/Qwen2.5), and [Documentation](https://qwen.readthedocs.io/en/latest/).
42
 
43
  ## Requirements
 
133
 
134
  ```
135
  @misc{qwq32b,
136
+ title = {QwQ-32B: The Power of Scaling RL},
137
  url = {https://qwenlm.github.io/blog/qwq-32b/},
138
  author = {Qwen Team},
139
  month = {March},
figures/benchmark.jpg ADDED