Gemma2-2B Instruction Tuned Model (Transferred to Qwen Tokenizer) Model Card

Gemma2-2B-IT transferred to the Qwen2 Tokenizer. The model approximately preserves performance of the original on most benchmarks, except for some slight degradations.

Model Details

  • Base Model: Gemma2-2B
  • Tokenization: Transferred to the Qwen Tokenizer
  • Training Methodology: Instruction-tuned Gemma2-2B-IT transferred to the Qwen Tokenizer
Benchmark Gemma2-2B w/ Qwen Tokenizer Original Gemma2-2B-IT
PiQA 76.9 79.6
HS 70.7 72.5
ARC-C 46.8 50.4
BoolQ 82.8 83.8
MMLU 53.8 56.9
Arith. 83.9 84.8
IFEval 62.5 62.5

Model Details

Details on the training methodology are forthcoming.

Use

import torch
from transformers import pipeline

pipe = pipeline(
    "text-generation",
    model="benjamin/Gemma2-2B-IT-with-Qwen2-Tokenizer",
    model_kwargs={"torch_dtype": torch.bfloat16},
    device="cuda",  # replace with "mps" to run on a Mac device
)

messages = [
    {"role": "user", "content": "Who are you? Please, answer in pirate-speak."},
]

outputs = pipe(messages, max_new_tokens=256)
assistant_response = outputs[0]["generated_text"][-1]["content"].strip()
print(assistant_response)
Downloads last month
24
Safetensors
Model size
2.37B params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for benjamin/Gemma2-2B-IT-with-Qwen2-Tokenizer

Base model

google/gemma-2-2b
Finetuned
(512)
this model