# Llama-3.2-3B-Instruct-function-calling-gorilla-style-5epochs ## Model Description This is a fine-tuned version of unsloth/Llama-3.2-3B-Instruct for function calling capabilities, trained in a Gorilla-style format on a custom dataset. The model is trained to understand function calling instructions and generate appropriate responses. ## Training Parameters - Base Model: unsloth/Llama-3.2-3B-Instruct - Dataset: Custom Function Calling Dataset (Gorilla-style format) - Training Type: Supervised Fine-tuning with LoRA - Epochs: 5 ## Dataset Format The model was trained on a custom dataset with the following structure: ```json { "Instruction": "User instruction/query", "Functions": ["Available function definitions"], "Output": ["Model's response with function calls"] } ``` ## Usage ```python from transformers import AutoModelForCausalLM, AutoTokenizer model = AutoModelForCausalLM.from_pretrained("BluebrainAI/Llama-3.2-3B-Instruct-function-calling-gorilla-style-5epochs") tokenizer = AutoTokenizer.from_pretrained("BluebrainAI/Llama-3.2-3B-Instruct-function-calling-gorilla-style-5epochs") # Example usage instruction = "Your instruction here" chat = [ {"role": "user", "content": instruction} ] input_text = tokenizer.apply_chat_template(chat, tokenize=False, add_generation_prompt=False) inputs = tokenizer(input_text, return_tensors="pt", truncation=True) outputs = model.generate(**inputs, max_new_tokens=512, temperature=0.7) response = tokenizer.decode(outputs[0], skip_special_tokens=True) ``` ## License This model inherits the license of the base model unsloth/Llama-3.2-3B-Instruct.