Chocolatine-Fusion-14B

FINGU-AI/Chocolatine-Fusion-14B is a merged model combining jpacifico/Chocolatine-2-14B-Instruct-v2.0b3 and jpacifico/Chocolatine-2-14B-Instruct-v2.0b2. This model maintains the strengths of Chocolatine while benefiting from an optimized fusion for improved reasoning and multi-turn conversation capabilities.

Training & Fine-Tuning

Chocolatine-Fusion-14B is based on DPO fine-tuning from the Chocolatine-2 series, which originated as a fine-tuned version of sometimesanotion/Lamarck-14B-v0.7.

  • The model has been trained using French and English RLHF datasets (including jpacifico/french-orca-dpo-pairs-revised) for enhanced bilingual capabilities.
  • Long-context support has been extended up to 128K tokens with the ability to generate up to 8K tokens.

OpenLLM Leaderboard

Coming soon.

MT-Bench

Coming soon.

Usage

You can run this model using the following code:

import transformers
from transformers import AutoTokenizer

# Format prompt
message = [
    {"role": "system", "content": "You are a helpful assistant chatbot."},
    {"role": "user", "content": "What is a Large Language Model?"}
]
tokenizer = AutoTokenizer.from_pretrained("FINGU-AI/Chocolatine-Fusion-14B")
prompt = tokenizer.apply_chat_template(message, add_generation_prompt=True, tokenize=False)

# Create pipeline
pipeline = transformers.pipeline(
    "text-generation",
    model="FINGU-AI/Chocolatine-Fusion-14B",
    tokenizer=tokenizer
)

# Generate text
sequences = pipeline(
    prompt,
    do_sample=True,
    temperature=0.7,
    top_p=0.9,
    num_return_sequences=1,
    max_length=200,
)
print(sequences[0]['generated_text'])

Limitations

Chocolatine-Fusion-14B is a demonstration of model merging techniques rather than a standalone fine-tuned model.

  • It does not have any built-in moderation mechanisms.
  • Responses may vary based on the interaction and prompt style.
  • Performance on highly technical or domain-specific queries may require further fine-tuning.

Developed by

  • Author: FINGU-AI, 2025
  • Base Models: jpacifico/Chocolatine-2-14B-Instruct-v2.0b3, jpacifico/Chocolatine-2-14B-Instruct-v2.0b2
  • Language(s): French, English
  • Model Type: Merged LLM
  • License: Apache-2.0
Downloads last month
394
Safetensors
Model size
8.37B params
Tensor type
F32
FP16
U8
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Space using FINGU-AI/Chocolatine-Fusion-14B 1