--- library_name: transformers license: mit --- # Chocolatine-Fusion-14B **FINGU-AI/Chocolatine-Fusion-14B** is a merged model combining **jpacifico/Chocolatine-2-14B-Instruct-v2.0b3** and **jpacifico/Chocolatine-2-14B-Instruct-v2.0b2**. This model maintains the strengths of Chocolatine while benefiting from an optimized fusion for improved reasoning and multi-turn conversation capabilities. ## **Training & Fine-Tuning** Chocolatine-Fusion-14B is based on **DPO fine-tuning** from the Chocolatine-2 series, which originated as a fine-tuned version of **sometimesanotion/Lamarck-14B-v0.7**. - The model has been trained using **French and English RLHF datasets** (including jpacifico/french-orca-dpo-pairs-revised) for enhanced bilingual capabilities. - Long-context support has been extended up to **128K tokens** with the ability to generate up to **8K tokens**. ## **OpenLLM Leaderboard** Coming soon. ## **MT-Bench** Coming soon. ## **Usage** You can run this model using the following code: ```python import transformers from transformers import AutoTokenizer # Format prompt message = [ {"role": "system", "content": "You are a helpful assistant chatbot."}, {"role": "user", "content": "What is a Large Language Model?"} ] tokenizer = AutoTokenizer.from_pretrained("FINGU-AI/Chocolatine-Fusion-14B") prompt = tokenizer.apply_chat_template(message, add_generation_prompt=True, tokenize=False) # Create pipeline pipeline = transformers.pipeline( "text-generation", model="FINGU-AI/Chocolatine-Fusion-14B", tokenizer=tokenizer ) # Generate text sequences = pipeline( prompt, do_sample=True, temperature=0.7, top_p=0.9, num_return_sequences=1, max_length=200, ) print(sequences[0]['generated_text']) ``` ## **Limitations** Chocolatine-Fusion-14B is a **demonstration of model merging techniques** rather than a standalone fine-tuned model. - It does **not** have any built-in moderation mechanisms. - Responses may vary based on the interaction and prompt style. - Performance on **highly technical or domain-specific queries** may require further fine-tuning. ## **Developed by** - **Author:** FINGU-AI, 2025 - **Base Models:** jpacifico/Chocolatine-2-14B-Instruct-v2.0b3, jpacifico/Chocolatine-2-14B-Instruct-v2.0b2 - **Language(s):** French, English - **Model Type:** Merged LLM - **License:** Apache-2.0