sumink commited on
Commit
e6e1507
·
verified ·
1 Parent(s): 569e731

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +18 -0
README.md ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Model Overview
2
+ This model is a fine-tuned version of LLaMA-3.2B, specifically trained on a dataset processed using Facility Location (FL) and Facility Location Mutual Information (FLMI) techniques. These data selection methods were employed to reduce the dataset size while retaining high-quality and representative samples, ensuring the model is trained on the most informative and diverse data points.
3
+
4
+ Dataset Details
5
+ Original Dataset: A filtered subset of a conversational dataset, containing examples of chosen and rejected responses.
6
+ Data Preprocessing:
7
+ The dataset underwent an initial Facility Location (FL) process to select 1,000 samples from the original dataset.
8
+ Further refinement using Facility Location Mutual Information (FLMI) reduced the dataset to 500 highly informative samples.
9
+ These methods ensured that the final dataset preserved critical information and diversity, optimizing the training efficiency and model performance.
10
+ Training Configuration
11
+ Base Model: LLaMA-3.2B
12
+ Fine-Tuning Dataset: The final dataset of 500 samples refined through FL and FLMI techniques.
13
+ Objective: Enhance the model's ability to generate high-quality, contextually accurate responses in conversational settings.
14
+ Training Framework: Hugging Face Transformers library with PyTorch backend.
15
+ Training Hardware: Multi-GPU setup (e.g., NVIDIA A100 GPUs).
16
+ Batch Size: 16
17
+ Learning Rate: 5e-5 with linear decay.
18
+ Optimizer: AdamW