1.png

Gacrux-R1-Qwen3-1.7B-MoD

Gacrux-R1-Qwen3-1.7B-MoD is a high-efficiency, multi-domain model fine-tuned on Qwen3-1.7B with traces of Mixture of Domains (MoD). It leverages the prithivMLmods/Gargantua-R1-Wee dataset, designed for rigorous mathematical problem-solving and enriched with multi-domain coverage across mathematics, coding, and science. This model blends symbolic precision, scientific logic, and structured output fluency—making it an ideal tool for developers, educators, and researchers seeking advanced reasoning under constrained compute.

GGUF: https://huggingface.co/prithivMLmods/Gacrux-R1-Qwen3-1.7B-MoD-GGUF


Key Features

  1. Unified Reasoning Across Math, Code & Science Fine-tuned on the Gargantua-R1-Wee dataset covering rigorous mathematics, coding, and scientific logic, enabling robust symbolic and multi-domain reasoning.

  2. Advanced Code Reasoning & Generation Supports multi-language coding with explanations, optimization hints, and error detection—ideal for full-stack prototyping, algorithm synthesis, and debugging workflows.

  3. Scientific & Mathematical Problem Solving Performs analytical reasoning in physics, biology, chemistry, and mathematics—explaining concepts, solving equations, and handling symbolic derivations step-by-step.

  4. Hybrid Symbolic-AI Thinking Combines structured logic, chain-of-thought reasoning, and open-ended inference, delivering robust performance on STEM tasks and complex prompt decomposition.

  5. Structured Output Mastery Seamlessly generates output in LaTeX, Markdown, JSON, CSV, and YAML, suited for research reports, technical documentation, and data formats.

  6. Optimized Lightweight Footprint for Versatile Deployment Balances performance and efficiency, making it deployable on mid-range GPUs, offline clusters, and advanced edge AI systems.


Quickstart with Transformers

from transformers import AutoModelForCausalLM, AutoTokenizer

model_name = "prithivMLmods/Gacrux-R1-Qwen3-1.7B-MoD"

model = AutoModelForCausalLM.from_pretrained(
    model_name,
    torch_dtype="auto",
    device_map="auto"
)
tokenizer = AutoTokenizer.from_pretrained(model_name)

prompt = "Explain the difference between Newtonian mechanics and quantum mechanics with examples."

messages = [
    {"role": "system", "content": "You are a scientific tutor skilled in code, math, and reasoning."},
    {"role": "user", "content": prompt}
]

text = tokenizer.apply_chat_template(
    messages,
    tokenize=False,
    add_generation_prompt=True
)

model_inputs = tokenizer([text], return_tensors="pt").to(model.device)

generated_ids = model.generate(
    **model_inputs,
    max_new_tokens=512
)
generated_ids = [
    output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
]

response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
print(response)

Intended Use

  • Scientific tutoring, computational logic, and mathematical education
  • Advanced coding assistant for algorithm design, code reviews, and documentation
  • Structured technical data generation across formats and fields
  • STEM-focused chatbot or API for research and education tools
  • Mid-resource deployment requiring high symbolic fidelity

Limitations

  • Not tuned for general-purpose or long-form creative writing
  • Context limitations may hinder multi-document or full codebase analysis
  • Specialized in technical and symbolic tasks—general chat may underperform
  • Prioritizes structured reasoning over emotional or casual tone generation
Downloads last month
5
Safetensors
Model size
1.72B params
Tensor type
BF16
·
F16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for prithivMLmods/Gacrux-R1-Qwen3-1.7B-MoD

Finetuned
Qwen/Qwen3-1.7B
Finetuned
(208)
this model
Quantizations
1 model

Dataset used to train prithivMLmods/Gacrux-R1-Qwen3-1.7B-MoD

Collection including prithivMLmods/Gacrux-R1-Qwen3-1.7B-MoD