feihu.hf
commited on
Commit
·
7c0a8dc
1
Parent(s):
2ebdfe8
update readme
Browse files- README.md +8 -1
- figures/benchmark.jpg +0 -0
README.md
CHANGED
@@ -19,6 +19,11 @@ tags:
|
|
19 |
|
20 |
QwQ is the reasoning model of the Qwen series. Compared with conventional instruction-tuned models, QwQ, which is capable of thinking and reasoning, can achieve significantly enhanced performance in downstream tasks, especially hard problems. QwQ-32B is the medium-sized reasoning model, which is capable of achieving competitive performance against state-of-the-art reasoning models, e.g., DeepSeek-R1, o1-mini.
|
21 |
|
|
|
|
|
|
|
|
|
|
|
22 |
**This repo contains the QwQ 32B model**, which has the following features:
|
23 |
- Type: Causal Language Models
|
24 |
- Training Stage: Pretraining & Post-training (Supervised Finetuning and Reinforcement Learning)
|
@@ -31,6 +36,8 @@ QwQ is the reasoning model of the Qwen series. Compared with conventional instru
|
|
31 |
|
32 |
**Note:** For the best experience, please review the [usage guidelines](#usage-guidelines) before deploying QwQ models.
|
33 |
|
|
|
|
|
34 |
For more details, please refer to our [blog](https://qwenlm.github.io/blog/qwq-32b/), [GitHub](https://github.com/QwenLM/Qwen2.5), and [Documentation](https://qwen.readthedocs.io/en/latest/).
|
35 |
|
36 |
## Requirements
|
@@ -126,7 +133,7 @@ If you find our work helpful, feel free to give us a cite.
|
|
126 |
|
127 |
```
|
128 |
@misc{qwq32b,
|
129 |
-
title = {
|
130 |
url = {https://qwenlm.github.io/blog/qwq-32b/},
|
131 |
author = {Qwen Team},
|
132 |
month = {March},
|
|
|
19 |
|
20 |
QwQ is the reasoning model of the Qwen series. Compared with conventional instruction-tuned models, QwQ, which is capable of thinking and reasoning, can achieve significantly enhanced performance in downstream tasks, especially hard problems. QwQ-32B is the medium-sized reasoning model, which is capable of achieving competitive performance against state-of-the-art reasoning models, e.g., DeepSeek-R1, o1-mini.
|
21 |
|
22 |
+
<p align="center">
|
23 |
+
<img width="100%" src="figures/benchmark.jpg">
|
24 |
+
</p>
|
25 |
+
|
26 |
+
|
27 |
**This repo contains the QwQ 32B model**, which has the following features:
|
28 |
- Type: Causal Language Models
|
29 |
- Training Stage: Pretraining & Post-training (Supervised Finetuning and Reinforcement Learning)
|
|
|
36 |
|
37 |
**Note:** For the best experience, please review the [usage guidelines](#usage-guidelines) before deploying QwQ models.
|
38 |
|
39 |
+
You can try our [demo](https://huggingface.co/spaces/Qwen/QwQ-32B-Demo) or access QwQ models via [QwenChat](https://chat.qwen.ai).
|
40 |
+
|
41 |
For more details, please refer to our [blog](https://qwenlm.github.io/blog/qwq-32b/), [GitHub](https://github.com/QwenLM/Qwen2.5), and [Documentation](https://qwen.readthedocs.io/en/latest/).
|
42 |
|
43 |
## Requirements
|
|
|
133 |
|
134 |
```
|
135 |
@misc{qwq32b,
|
136 |
+
title = {QwQ-32B: The Power of Scaling RL},
|
137 |
url = {https://qwenlm.github.io/blog/qwq-32b/},
|
138 |
author = {Qwen Team},
|
139 |
month = {March},
|
figures/benchmark.jpg
ADDED
![]() |