File size: 963 Bytes
b64647b fadeac0 b64647b fa3b309 f8f36f3 8b59af1 0fc0c0d b64647b 0fc0c0d b64647b 8b59af1 f8f36f3 8b59af1 f8f36f3 b64647b d094778 b64647b 0fc0c0d |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 |
---
base_model: Pinkstack/PARM-V1-phi-4-4k-CoT-pytorch
tags:
- text-generation-inference
- transformers
- unsloth
- llama
- gguf
- code
- phi3
- cot
- o1
- reasoning
- cot
license: mit
language:
- en
pipeline_tag: text-generation
---
this is out flagship model, with top-tier reasoning, rivaling gemini-flash-exp-2.0-thinking and o1 mini. results are overall similar to both of them, we are not comparing to qwq as it has much longer results which waste tokens.
the model uses this prompt: (modified phi-4 prompt)
```
{{ if .System }}<|system|>
{{ .System }}<|im_end|>
{{ end }}{{ if .Prompt }}<|user|>
{{ .Prompt }}<|im_end|>
{{ end }}<|assistant|>{{ .CoT }}<|CoT|>
{{ .Response }}<|FinalAnswer|><|im_end|>
```
# Uploaded model
- **Developed by:** Pinkstack
- **License:** MIT
- **Finetuned from model :** Pinkstack/PARM-V1-phi-4-4k-CoT-pytorch
This Phi-4 model was trained with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. |