Triangle104 commited on
Commit
f0ea82a
·
verified ·
1 Parent(s): 11971d5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +87 -0
README.md CHANGED
@@ -117,6 +117,93 @@ model-index:
117
  This model was converted to GGUF format from [`prithivMLmods/Primal-Opus-14B-Optimus-v2`](https://huggingface.co/prithivMLmods/Primal-Opus-14B-Optimus-v2) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
118
  Refer to the [original model card](https://huggingface.co/prithivMLmods/Primal-Opus-14B-Optimus-v2) for more details on the model.
119
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
120
  ## Use with llama.cpp
121
  Install llama.cpp through brew (works on Mac and Linux)
122
 
 
117
  This model was converted to GGUF format from [`prithivMLmods/Primal-Opus-14B-Optimus-v2`](https://huggingface.co/prithivMLmods/Primal-Opus-14B-Optimus-v2) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
118
  Refer to the [original model card](https://huggingface.co/prithivMLmods/Primal-Opus-14B-Optimus-v2) for more details on the model.
119
 
120
+ ---
121
+
122
+ Multilingual Proficiency: Supports over 29 languages, including Chinese, English, French, Spanish, Portuguese, German, and more.
123
+
124
+
125
+
126
+
127
+
128
+
129
+
130
+ Quickstart with Transformers
131
+ -
132
+
133
+
134
+
135
+ from transformers import AutoModelForCausalLM, AutoTokenizer
136
+
137
+ model_name = "prithivMLmods/Primal-Opus-14B-Optimus-v2"
138
+
139
+ model = AutoModelForCausalLM.from_pretrained(
140
+ model_name,
141
+ torch_dtype="auto",
142
+ device_map="auto",
143
+ trust_remote_code=True
144
+ )
145
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
146
+
147
+ prompt = "Give me a short introduction to large language models."
148
+ messages = [
149
+ {"role": "system", "content": "You are an advanced AI assistant with expert-level reasoning and knowledge."},
150
+ {"role": "user", "content": prompt}
151
+ ]
152
+ text = tokenizer.apply_chat_template(
153
+ messages,
154
+ tokenize=False,
155
+ add_generation_prompt=True
156
+ )
157
+ model_inputs = tokenizer([text], return_tensors="pt").to(model.device)
158
+
159
+ generated_ids = model.generate(
160
+ **model_inputs,
161
+ max_new_tokens=512
162
+ )
163
+ generated_ids = [
164
+ output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
165
+ ]
166
+
167
+ response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
168
+ print(response)
169
+
170
+
171
+
172
+
173
+
174
+
175
+
176
+
177
+ Intended Use
178
+ -
179
+
180
+
181
+
182
+ Advanced Logical Reasoning: Designed for logical deduction, multi-step problem-solving, and knowledge-based tasks.
183
+ Mathematical & Scientific Problem-Solving: Enhanced capabilities for calculations, theorem proving, and scientific queries.
184
+ Code Generation & Debugging: Generates and optimizes code across multiple programming languages.
185
+ Structured Data Analysis: Processes tables, JSON, and structured outputs, making it ideal for data-centric tasks.
186
+ Multilingual Applications: High proficiency in over 29 languages, enabling global-scale applications.
187
+ Extended Content Generation: Supports detailed document writing, research reports, and instructional guides.
188
+
189
+
190
+
191
+
192
+
193
+
194
+
195
+ Limitations
196
+ -
197
+
198
+
199
+
200
+ High Computational Requirements: Due to its 14B parameters and 128K context support, it requires powerful GPUs or TPUs for efficient inference.
201
+ Language-Specific Variability: Performance may vary across supported languages, especially for low-resource languages.
202
+ Potential Error Accumulation: Long-text generation can sometimes introduce inconsistencies over extended outputs.
203
+ Limited Real-World Awareness: Knowledge is restricted to training data and may not reflect recent world events.
204
+ Prompt Sensitivity: Outputs can depend on the specificity and clarity of the input prompt.
205
+
206
+ ---
207
  ## Use with llama.cpp
208
  Install llama.cpp through brew (works on Mac and Linux)
209