Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,18 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
Model Overview
|
2 |
+
This model is a fine-tuned version of LLaMA-3.2B, specifically trained on a dataset processed using Facility Location (FL) and Facility Location Mutual Information (FLMI) techniques. These data selection methods were employed to reduce the dataset size while retaining high-quality and representative samples, ensuring the model is trained on the most informative and diverse data points.
|
3 |
+
|
4 |
+
Dataset Details
|
5 |
+
Original Dataset: A filtered subset of a conversational dataset, containing examples of chosen and rejected responses.
|
6 |
+
Data Preprocessing:
|
7 |
+
The dataset underwent an initial Facility Location (FL) process to select 1,000 samples from the original dataset.
|
8 |
+
Further refinement using Facility Location Mutual Information (FLMI) reduced the dataset to 500 highly informative samples.
|
9 |
+
These methods ensured that the final dataset preserved critical information and diversity, optimizing the training efficiency and model performance.
|
10 |
+
Training Configuration
|
11 |
+
Base Model: LLaMA-3.2B
|
12 |
+
Fine-Tuning Dataset: The final dataset of 500 samples refined through FL and FLMI techniques.
|
13 |
+
Objective: Enhance the model's ability to generate high-quality, contextually accurate responses in conversational settings.
|
14 |
+
Training Framework: Hugging Face Transformers library with PyTorch backend.
|
15 |
+
Training Hardware: Multi-GPU setup (e.g., NVIDIA A100 GPUs).
|
16 |
+
Batch Size: 16
|
17 |
+
Learning Rate: 5e-5 with linear decay.
|
18 |
+
Optimizer: AdamW
|