prithivMLmods commited on
Commit
e5d2392
·
verified ·
1 Parent(s): 6b99588

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +105 -0
README.md CHANGED
@@ -10,3 +10,108 @@ tags:
10
  - text-generation-inference
11
  ---
12
  ![omni.png](https://cdn-uploads.huggingface.co/production/uploads/65bb837dbfb878f46c77de4c/Zz_Uc6M06tyh3Euhm93fn.png)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
10
  - text-generation-inference
11
  ---
12
  ![omni.png](https://cdn-uploads.huggingface.co/production/uploads/65bb837dbfb878f46c77de4c/Zz_Uc6M06tyh3Euhm93fn.png)
13
+
14
+ # **Omni-Reasoner-o1: Overview**
15
+
16
+ *Omni-Reasoner-o1* is a specialized AI model built upon the Sky T1 32B architecture, combined with **Qwen 2.5 32B**, and fine-tuned using synthetic data from OpenAI pipeline-generated records. It is optimized for mathematical reasoning and complex problem-solving.
17
+
18
+ # **Quickstart with Transformers**
19
+
20
+ Here provides a code snippet with `apply_chat_template` to show you how to load the tokenizer and model and how to generate contents.
21
+
22
+ ```python
23
+ from transformers import AutoModelForCausalLM, AutoTokenizer
24
+
25
+ model_name = "prithivMLmods/Omni-Reasoner-o1"
26
+
27
+ model = AutoModelForCausalLM.from_pretrained(
28
+ model_name,
29
+ torch_dtype="auto",
30
+ device_map="auto"
31
+ )
32
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
33
+
34
+ prompt = "How many r in strawberry."
35
+ messages = [
36
+ {"role": "system", "content": "You are a helpful and harmless assistant. You are Qwen developed by Alibaba. You should think step-by-step."},
37
+ {"role": "user", "content": prompt}
38
+ ]
39
+ text = tokenizer.apply_chat_template(
40
+ messages,
41
+ tokenize=False,
42
+ add_generation_prompt=True
43
+ )
44
+ model_inputs = tokenizer([text], return_tensors="pt").to(model.device)
45
+
46
+ generated_ids = model.generate(
47
+ **model_inputs,
48
+ max_new_tokens=512
49
+ )
50
+ generated_ids = [
51
+ output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
52
+ ]
53
+
54
+ response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
55
+ ```
56
+
57
+ # **Key Features**
58
+ 1. **Hybrid Architecture:**
59
+ - Combines **Sky T1 32B** and **Qwen 2.5 32B** to leverage strengths in both natural language understanding and mathematical reasoning.
60
+ - Enables robust problem-solving across diverse domains.
61
+
62
+ 2. **Mathematical Expertise:**
63
+ - Trained specifically as a **mathematical reasoner and problem solver**.
64
+ - Excels in numerical computations, symbolic mathematics, proofs, and equation-solving.
65
+
66
+ 3. **Synthetic Data Fine-Tuning:**
67
+ - Leveraged high-quality synthetic data generated by OpenAI pipelines.
68
+ - Ensures enhanced generalization across a wide range of problem-solving scenarios.
69
+
70
+ 4. **Natural Language Processing (NLP):**
71
+ - Capable of understanding and interpreting complex language inputs related to mathematical queries.
72
+ - Provides step-by-step explanations for solutions, fostering user understanding.
73
+
74
+ 5. **Multi-Task Capability:**
75
+ - Handles a variety of mathematical tasks including algebra, calculus, combinatorics, and statistics.
76
+ - Suitable for word problems and domain-specific queries requiring logic and reasoning.
77
+
78
+ 6. **Scalability:**
79
+ - Designed for seamless integration into **educational platforms**, **scientific research tools**, and **automated reasoning systems**.
80
+
81
+ # **Intended Use**
82
+ 1. **Educational Applications:**
83
+ - Acts as a tutor for students in mathematics and related fields.
84
+ - Provides explanations, step-by-step solutions, and practice problem generation.
85
+
86
+ 2. **Scientific Research:**
87
+ - Aids researchers in automating repetitive mathematical calculations or exploring new problem-solving methodologies.
88
+
89
+ 3. **Professional Use Cases:**
90
+ - Supports professionals in domains like engineering, data science, and finance by solving domain-specific mathematical problems.
91
+
92
+ 4. **AI-Assisted Development:**
93
+ - Assists in coding environments for algorithm development and debugging by identifying mathematical bottlenecks or issues.
94
+
95
+ 5. **Automated Systems:**
96
+ - Integrates into automated reasoning and decision-making systems for operations requiring quantitative analysis.
97
+
98
+ # **Limitations**
99
+ 1. **Reliance on Synthetic Data:**
100
+ - Despite its extensive training, reliance on synthetic data might lead to **biases** or **overfitting** in specific scenarios.
101
+ - May struggle with real-world edge cases not reflected in its training data.
102
+
103
+ 2. **Domain-Specific Gaps:**
104
+ - While excelling in mathematics, it may not perform as well in non-mathematical or interdisciplinary problem-solving tasks.
105
+
106
+ 3. **Resource Intensive:**
107
+ - Due to its hybrid 32B architecture, deploying the model requires **significant computational resources**.
108
+
109
+ 4. **Interpretation Errors:**
110
+ - Misinterprets poorly structured or ambiguous natural language queries.
111
+ - May provide overly verbose explanations that aren't always user-friendly.
112
+
113
+ 5. **Limitations in Creativity:**
114
+ - Not designed for creative or abstract tasks outside mathematical reasoning, such as writing, art, or subjective decision-making.
115
+
116
+ 6. **Dependency on Prompt Quality:**
117
+ - Performance can degrade with unclear, poorly framed, or overly complex prompts