Edit model card

monsterapi/gemma-2-2b-norobots

Base Model for Fine-tuning: google/gemma-2-2b-it
Service Used: MonsterAPI
License: Apache-2.0

Overview

monsterapi/gemma-2-2b-norobots is a fine-tuned language model designed to improve instruction-following capabilities. The model was trained using the "No Robots" dataset, a high-quality set of 10,000 instructions and demonstrations curated by expert human annotators. This fine-tuning process enhances the base model's performance in understanding and executing single-turn instructions, similar to the goals outlined in OpenAI's InstructGPT.

Dataset Details

Dataset Summary:
The "No Robots" dataset is a collection of 10,000 high-quality instructions and demonstrations created by skilled human annotators. The dataset is modeled after the instruction dataset described in OpenAI's InstructGPT paper. It mainly includes single-turn instructions across various categories, aiming to improve the instruction-following capabilities of language models during supervised fine-tuning (SFT).

Fine-tuning Details

Fine-tuned Model Name: monsterapi/gemma-2-2b-norobots
Training Time: 31 minutes
Cost: $1.10
Epochs: 1
Gradient Accumulation Steps: 32

The model was fine-tuned using MonsterAPI's finetuning service, optimizing the base model google/gemma-2-2b-it to perform better on instruction-following tasks.

Hyperparameters & Additional Details

  • Base Model: google/gemma-2-2b-it
  • Dataset: No Robots (10,000 instructions and demonstrations)
  • Training Duration: 31 minutes
  • Cost per Epoch: $1.10
  • Total Finetuning Cost: $1.10
  • Gradient Accumulation Steps: 32

Use Cases

This model is well-suited for tasks that require improved instruction-following capabilities, such as:

  • Chatbots and virtual assistants
  • Content creation tools
  • Automated customer support systems
  • Task automation in various industries

How to Use

You can load the model directly using the Hugging Face Transformers library:

from transformers import AutoModelForCausalLM, AutoTokenizer

model_name = "monsterapi/gemma-2-2b-norobots"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)

# Example usage
input_text = "Explain the concept of supervised fine-tuning in simple terms."
inputs = tokenizer(input_text, return_tensors="pt")
outputs = model.generate(**inputs)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))

Acknowledgements

The fine-tuning process was carried out using MonsterAPI's finetuning service, which offers a seamless experience for optimizing large language models.

Contact

For further details or queries, please contact MonsterAPI or visit the official documentation.

Downloads last month
25
Inference API
Unable to determine this model’s pipeline type. Check the docs .

Model tree for monsterapi/gemma-2-2b-norobots

Base model

google/gemma-2-2b
Adapter
this model

Dataset used to train monsterapi/gemma-2-2b-norobots