File size: 963 Bytes
b64647b
fadeac0
b64647b
 
 
 
 
 
fa3b309
 
 
 
f8f36f3
8b59af1
0fc0c0d
b64647b
 
0fc0c0d
b64647b
 
8b59af1
 
 
f8f36f3
 
 
 
 
8b59af1
 
f8f36f3
b64647b
 
 
d094778
 
b64647b
0fc0c0d
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
---
base_model: Pinkstack/PARM-V1-phi-4-4k-CoT-pytorch
tags:
- text-generation-inference
- transformers
- unsloth
- llama
- gguf
- code
- phi3
- cot
- o1
- reasoning
- cot
license: mit
language:
- en
pipeline_tag: text-generation
---


this is out flagship model, with top-tier reasoning, rivaling gemini-flash-exp-2.0-thinking and o1 mini. results are overall similar to both of them, we are not comparing to qwq as it has much longer results which waste tokens.
the model uses this prompt: (modified phi-4 prompt)
```
{{ if .System }}<|system|>
{{ .System }}<|im_end|>
{{ end }}{{ if .Prompt }}<|user|>
{{ .Prompt }}<|im_end|>
{{ end }}<|assistant|>{{ .CoT }}<|CoT|>
{{ .Response }}<|FinalAnswer|><|im_end|>
```
# Uploaded  model

- **Developed by:** Pinkstack
- **License:** MIT
- **Finetuned from model :** Pinkstack/PARM-V1-phi-4-4k-CoT-pytorch

This Phi-4 model was trained with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.