Daemontatox commited on
Commit
cd6c033
·
verified ·
1 Parent(s): 116fdd9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -12
README.md CHANGED
@@ -7,25 +7,27 @@ tags:
7
  - unsloth
8
  - qwen2
9
  - trl
 
 
10
  license: apache-2.0
11
  language:
12
  - en
13
- metrics:
14
- - accuracy
15
  new_version: Daemontatox/CogitoZ
16
  library_name: transformers
 
 
17
  ---
18
  ![image](./image.webp)
19
- # CogitoZ - Qwen2
20
 
21
  ## Model Overview
22
 
23
- CogitoZ - Qwen2 is a state-of-the-art large language model fine-tuned to excel in advanced reasoning and real-time decision-making tasks. This enhanced version was trained using [Unsloth](https://github.com/unslothai/unsloth), achieving a 2x faster training process. Leveraging Hugging Face's TRL (Transformers Reinforcement Learning) library, CogitoZ combines efficiency with exceptional reasoning performance.
24
 
25
  - **Developed by**: Daemontatox
26
  - **License**: Apache 2.0
27
  - **Base Model**: [Qwen/QwQ-32B-Preview](https://huggingface.co/Qwen/QwQ-32B-Preview)
28
- - **Finetuned from**: [Daemontatox/CogitoZ](https://huggingface.co/Daemontatox/CogitoZ)
29
 
30
  [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
31
 
@@ -95,10 +97,4 @@ The fine-tuning process utilized reasoning-specific datasets, including:
95
  **Safe Deployment** **->** Not recommended for generating harmful or unethical content.
96
 
97
  ## Acknowledgments
98
- This model was developed with contributions from Daemontatox and the Unsloth team, utilizing state-of-the-art techniques in fine-tuning and optimization.
99
-
100
- For more information or collaboration inquiries, please contact:
101
-
102
- Author: Daemontatox
103
- GitHub: Daemontatox GitHub Profile
104
- Unsloth: Unsloth GitHub
 
7
  - unsloth
8
  - qwen2
9
  - trl
10
+ - Chain-of-thought
11
+ - Reasoning
12
  license: apache-2.0
13
  language:
14
  - en
 
 
15
  new_version: Daemontatox/CogitoZ
16
  library_name: transformers
17
+ datasets:
18
+ - PJMixers/Math-Multiturn-100K-ShareGPT
19
  ---
20
  ![image](./image.webp)
21
+ # CogitoZ - 32B
22
 
23
  ## Model Overview
24
 
25
+ CogitoZ - 32B is a state-of-the-art large language model fine-tuned to excel in advanced reasoning and real-time decision-making tasks. This enhanced version was trained using [Unsloth](https://github.com/unslothai/unsloth), achieving a 2x faster training process. Leveraging Hugging Face's TRL (Transformers Reinforcement Learning) library, CogitoZ combines efficiency with exceptional reasoning performance.
26
 
27
  - **Developed by**: Daemontatox
28
  - **License**: Apache 2.0
29
  - **Base Model**: [Qwen/QwQ-32B-Preview](https://huggingface.co/Qwen/QwQ-32B-Preview)
30
+ - **Finetuned To**: [Daemontatox/CogitoZ](https://huggingface.co/Daemontatox/CogitoZ)
31
 
32
  [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
33
 
 
97
  **Safe Deployment** **->** Not recommended for generating harmful or unethical content.
98
 
99
  ## Acknowledgments
100
+ This model was developed with contributions from Daemontatox and the Unsloth team, utilizing state-of-the-art techniques in fine-tuning and optimization.