BluebrainAILabs's picture
Upload folder using huggingface_hub
195a7b1 verified
|
raw
history blame
1.62 kB
# Llama-3.2-3B-Instruct-function-calling-gorilla-style-5epochs
## Model Description
This is a fine-tuned version of unsloth/Llama-3.2-3B-Instruct for function calling capabilities, trained in a Gorilla-style format on a custom dataset. The model is trained to understand function calling instructions and generate appropriate responses.
## Training Parameters
- Base Model: unsloth/Llama-3.2-3B-Instruct
- Dataset: Custom Function Calling Dataset (Gorilla-style format)
- Training Type: Supervised Fine-tuning with LoRA
- Epochs: 5
## Dataset Format
The model was trained on a custom dataset with the following structure:
```json
{
"Instruction": "User instruction/query",
"Functions": ["Available function definitions"],
"Output": ["Model's response with function calls"]
}
```
## Usage
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("BluebrainAI/Llama-3.2-3B-Instruct-function-calling-gorilla-style-5epochs")
tokenizer = AutoTokenizer.from_pretrained("BluebrainAI/Llama-3.2-3B-Instruct-function-calling-gorilla-style-5epochs")
# Example usage
instruction = "Your instruction here"
chat = [
{"role": "user", "content": instruction}
]
input_text = tokenizer.apply_chat_template(chat, tokenize=False, add_generation_prompt=False)
inputs = tokenizer(input_text, return_tensors="pt", truncation=True)
outputs = model.generate(**inputs, max_new_tokens=512, temperature=0.7)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
```
## License
This model inherits the license of the base model unsloth/Llama-3.2-3B-Instruct.