--- base_model: unsloth/phi-4-unsloth-bnb-4bit tags: - text-generation-inference - transformers - unsloth - llama - trl license: apache-2.0 language: - en datasets: - bespokelabs/Bespoke-Stratos-17k --- # Phi4 Turn R1Distill LoRA Adapters ## Overview Hey! These LoRA adapters are trained using different reasoning datasets that utilize **Thought** and **Solution** for reasoning responses. I hope these help jumpstart your project! These adapters have been trained on an **A800 GPU** and should provide a solid base for fine-tuning or merging. Everything on my page is left **public** for Open Source use. ## Available LoRA Adapters Here are the links to the available adapters as of **January 30, 2025**: - [Phi4.Turn.R1Distill-Lora1](https://huggingface.co/Quazim0t0/Phi4.Turn.R1Distill-Lora1) - [Phi4.Turn.R1Distill-Lora2](https://huggingface.co/Quazim0t0/Phi4.Turn.R1Distill-Lora2) - [Phi4.Turn.R1Distill-Lora3](https://huggingface.co/Quazim0t0/Phi4.Turn.R1Distill-Lora3) - [Phi4.Turn.R1Distill-Lora4](https://huggingface.co/Quazim0t0/Phi4.Turn.R1Distill-Lora4) - [Phi4.Turn.R1Distill-Lora5](https://huggingface.co/Quazim0t0/Phi4.Turn.R1Distill-Lora5) - [Phi4.Turn.R1Distill-Lora6](https://huggingface.co/Quazim0t0/Phi4.Turn.R1Distill-Lora6) - [Phi4.Turn.R1Distill-Lora7](https://huggingface.co/Quazim0t0/Phi4.Turn.R1Distill-Lora7) - [Phi4.Turn.R1Distill-Lora8](https://huggingface.co/Quazim0t0/Phi4.Turn.R1Distill-Lora8) ## Usage These adapters can be loaded and used with `peft` and `transformers`. Here’s a quick example: ```python from transformers import AutoModelForCausalLM, AutoTokenizer from peft import PeftModel base_model = "microsoft/Phi-4" lora_adapter = "Quazim0t0/Phi4.Turn.R1Distill-Lora1" tokenizer = AutoTokenizer.from_pretrained(base_model) model = AutoModelForCausalLM.from_pretrained(base_model) model = PeftModel.from_pretrained(model, lora_adapter) model.eval()