LLAMA-3.2-3B-MathInstruct_LORA_SFT

This model is a fine-tuned version of meta-llama/Llama-3.2-3B-Instruct on the MathInstruct dataset. The fine-tuning process was designed to enhance the model's performance for mathematical instruction-following tasks, ensuring improved accuracy and precision when solving math-related problems.

It achieves the following results on the evaluation set:

  • Loss: 0.6895

Model Description

This model is specifically fine-tuned for mathematical reasoning, problem-solving, and instruction-following tasks. Leveraging the LLaMA-3.2-3B-Instruct base model, it has been optimized to handle mathematical queries and tasks with improved efficiency and context understanding.


Training and Evaluation Data

The model was fine-tuned on the MathInstruct dataset.

  • Dataset Source: TIGER-Lab.
  • Dataset Focus: Mathematical instruction-following and reasoning tasks.
  • Scope: A wide range of math topics, including arithmetic, algebra, calculus, and problem-solving.

The dataset was carefully curated to align with instructional objectives for solving mathematical problems and understanding step-by-step reasoning.


Training Procedure

Hyperparameters

  • Learning rate: 0.0001
  • Train batch size: 1
  • Eval batch size: 1
  • Gradient accumulation steps: 8
  • Total effective batch size: 8
  • Optimizer: AdamW (torch)
    • Betas: (0.9, 0.999)
    • Epsilon: 1e-08
  • Learning rate scheduler: Cosine schedule with 10% warmup.
  • Number of epochs: 3.0

Framework Versions

  • PEFT: 0.12.0
  • Transformers: 4.46.1
  • PyTorch: 2.5.1+cu124
  • Datasets: 3.1.0
  • Tokenizers: 0.20.3

Training Results

  • Loss: 0.6895
  • Evaluation indicates strong performance on math instruction-following tasks. Further testing on specific use cases is recommended to assess the model’s generalizability.

Additional Information

  • Author: Sri Santh M
  • Purpose: Fine-tuned for educational and development purposes, particularly for math-related tasks.
  • Dataset Link: MathInstruct Dataset

This model represents a focused effort to adapt the LLaMA-3.2-3B-Instruct model for specialized mathematical use cases. It can be further fine-tuned or extended for more specific mathematical domains or applications.

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.

Model tree for SriSanth2345/LLAMA-3.2-3B-MathInstruct_LORA_SFT

Adapter
(198)
this model

Dataset used to train SriSanth2345/LLAMA-3.2-3B-MathInstruct_LORA_SFT