Text Generation
Transformers
Safetensors
English
qwen2
conversational
text-generation-inference
Inference Endpoints

You need to agree to share your contact information to access this model

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this model content.

Model Details:

  • Base Model: Qwen/Qwen2-0.5B-Instruct
  • Teacher Model: Qwen/QwQ-32B-Preview
  • Distillation Framework: Instruction Tuning
  • Task Type: Conversational AI / Causal Language Modeling
  • Parameters: 0.5B
  • Special Features:
    • Integrated gradient checkpointing for efficient training
    • Step-by-step reasoning capabilities for better problem-solving

Training:

QwQ-0.5B-Distilled was trained using the amphora/QwQ-LongCoT-130K-2, PowerInfer/QWQ-LONGCOT-500K, and PowerInfer/LONGCOT-Refine-500K with supervised finetuning. This model can be used as a competitive reasoning model on edge devices as well as a draft model for Qwen/QwQ-32B-Preview.

Training Progress:

[â–“â–“â–“â–“â–“â–“â–“â–“â–“â–“] 100%

Example Usage:

import torch
from transformers import AutoModelForCausalLM, AutoTokenizer
# Model name
model_name = "kz919/QwQ-0.5B-Distilled-SFT"
# Load the model
print(f"Starting to load the model {model_name} into memory")
model = AutoModelForCausalLM.from_pretrained(
    model_name,
    torch_dtype=torch.bfloat16,
    device_map={"": 0}
)
# Load the tokenizer
tokenizer = AutoTokenizer.from_pretrained(model_name)
# Define the prompt
prompt = "How many r in strawberry."
messages = [
    {"role": "system", "content": "You are a helpful and harmless assistant. You are Qwen developed by Alibaba. You should think step-by-step."},
    {"role": "user", "content": prompt}
]
# Tokenize the input
text = tokenizer.apply_chat_template(
    messages,
    tokenize=False,
    add_generation_prompt=True
)
model_inputs = tokenizer([text], return_tensors="pt").to(model.device)
# Generate a response
generated_ids = model.generate(
    **model_inputs,
    max_new_tokens=4096
)
generated_ids = [
    output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
]
# Decode the response
response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
print(response)

Applications:

  1. Conversational Assistants:
    Suitable for AI chatbots that require reasoning and long-context understanding.

  2. Educational Tools:
    Provides step-by-step explanations, making it ideal for learning environments.

  3. Creative Writing:
    Assists in generating coherent, contextually aware long-form content.

  4. Technical Support:
    Handles complex customer queries with precision and clarity.


Draft model for Qwen/QwQ-32B-Preview:

This model can be used as a draft model for Qwen/QwQ-32B-Preview in sepculative decoding. We observe out of 5 tokens it generates, on average 3 tokens are accepted for math queries and 2.3 tokens are accepted for general reasoning queries.


Limitations:

  • While distilled for efficiency, performance on highly complex reasoning tasks may slightly trail the teacher model.
  • This model could still be under trained, merely a proof of concept. Don't yell at me if it's outputing nonesense.

Citation:

If you use this model in your research or applications, please cite it as:

@model{qwq_0.5B_distilled,
  author = {Kaizhao Liang},
  title = {Mini-QwQ: A Reasoning Model for Edge Devices},
  year = {2024},
  publisher = {Hugging Face},
  version = {1.0}
}

This model is an example of how efficient fine-tuning and distillation methods can deliver robust conversational AI capabilities in a smaller, more manageable footprint.

Downloads last month
14
Safetensors
Model size
494M params
Tensor type
BF16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for sambanovasystems/QwQ-0.5B-SFT-Draft

Base model

Qwen/Qwen2.5-0.5B
Finetuned
(150)
this model

Datasets used to train sambanovasystems/QwQ-0.5B-SFT-Draft