Daemontatox
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -7,25 +7,27 @@ tags:
|
|
7 |
- unsloth
|
8 |
- qwen2
|
9 |
- trl
|
|
|
|
|
10 |
license: apache-2.0
|
11 |
language:
|
12 |
- en
|
13 |
-
metrics:
|
14 |
-
- accuracy
|
15 |
new_version: Daemontatox/CogitoZ
|
16 |
library_name: transformers
|
|
|
|
|
17 |
---
|
18 |
![image](./image.webp)
|
19 |
-
# CogitoZ -
|
20 |
|
21 |
## Model Overview
|
22 |
|
23 |
-
CogitoZ -
|
24 |
|
25 |
- **Developed by**: Daemontatox
|
26 |
- **License**: Apache 2.0
|
27 |
- **Base Model**: [Qwen/QwQ-32B-Preview](https://huggingface.co/Qwen/QwQ-32B-Preview)
|
28 |
-
- **Finetuned
|
29 |
|
30 |
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
|
31 |
|
@@ -95,10 +97,4 @@ The fine-tuning process utilized reasoning-specific datasets, including:
|
|
95 |
**Safe Deployment** **->** Not recommended for generating harmful or unethical content.
|
96 |
|
97 |
## Acknowledgments
|
98 |
-
This model was developed with contributions from Daemontatox and the Unsloth team, utilizing state-of-the-art techniques in fine-tuning and optimization.
|
99 |
-
|
100 |
-
For more information or collaboration inquiries, please contact:
|
101 |
-
|
102 |
-
Author: Daemontatox
|
103 |
-
GitHub: Daemontatox GitHub Profile
|
104 |
-
Unsloth: Unsloth GitHub
|
|
|
7 |
- unsloth
|
8 |
- qwen2
|
9 |
- trl
|
10 |
+
- Chain-of-thought
|
11 |
+
- Reasoning
|
12 |
license: apache-2.0
|
13 |
language:
|
14 |
- en
|
|
|
|
|
15 |
new_version: Daemontatox/CogitoZ
|
16 |
library_name: transformers
|
17 |
+
datasets:
|
18 |
+
- PJMixers/Math-Multiturn-100K-ShareGPT
|
19 |
---
|
20 |
![image](./image.webp)
|
21 |
+
# CogitoZ - 32B
|
22 |
|
23 |
## Model Overview
|
24 |
|
25 |
+
CogitoZ - 32B is a state-of-the-art large language model fine-tuned to excel in advanced reasoning and real-time decision-making tasks. This enhanced version was trained using [Unsloth](https://github.com/unslothai/unsloth), achieving a 2x faster training process. Leveraging Hugging Face's TRL (Transformers Reinforcement Learning) library, CogitoZ combines efficiency with exceptional reasoning performance.
|
26 |
|
27 |
- **Developed by**: Daemontatox
|
28 |
- **License**: Apache 2.0
|
29 |
- **Base Model**: [Qwen/QwQ-32B-Preview](https://huggingface.co/Qwen/QwQ-32B-Preview)
|
30 |
+
- **Finetuned To**: [Daemontatox/CogitoZ](https://huggingface.co/Daemontatox/CogitoZ)
|
31 |
|
32 |
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
|
33 |
|
|
|
97 |
**Safe Deployment** **->** Not recommended for generating harmful or unethical content.
|
98 |
|
99 |
## Acknowledgments
|
100 |
+
This model was developed with contributions from Daemontatox and the Unsloth team, utilizing state-of-the-art techniques in fine-tuning and optimization.
|
|
|
|
|
|
|
|
|
|
|
|