InferenceLab commited on
Commit
a256681
·
verified ·
1 Parent(s): e2bda88

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -29
README.md CHANGED
@@ -32,12 +32,6 @@ This model is a domain-adapted version of LLaMA 3.2 3B Instruct. It has been fin
32
  - **License:** Apache 2.0
33
  - **Finetuned from model:** meta-llama/Llama-3.2-3B-Instruct
34
 
35
- ### Model Sources
36
-
37
- - **Repository:** [More Information Needed]
38
- - **Paper [optional]:** [More Information Needed]
39
- - **Demo [optional]:** [More Information Needed]
40
-
41
  ## Uses
42
 
43
  ### Direct Use
@@ -96,14 +90,11 @@ Tokenization using LLaMA tokenizer with special medical instruction formatting.
96
  #### Training Hyperparameters
97
 
98
  * **Training regime:** bf16 mixed precision
99
- * **Epochs:** 3
100
- * **Batch size:** 64
101
- * **Learning rate:** 2e-5
102
 
103
  #### Speeds, Sizes, Times
104
 
105
  * **Training time:** \~12 hours on 4×A100 GPUs
106
- * **Final model size:** \~3.1B parameters
107
 
108
  ## Evaluation
109
 
@@ -165,23 +156,6 @@ Explainability tools like LLaMA-MedLens (if available) are suggested to interpre
165
  * Unsloth
166
  * PyTorch 2.1
167
 
168
- ## Citation
169
-
170
- **BibTeX:**
171
-
172
- ```bibtex
173
- @misc{medillama_2025,
174
- author = {InferenceLab},
175
- title = {MediLlama-3.2: A Medical Chatbot Fine-Tuned from LLaMA 3.2},
176
- year = {2025},
177
- publisher = {HuggingFace},
178
- howpublished = {\url{https://huggingface.co/InferenceLab/MediLlama-3.2}},
179
- }
180
- ```
181
-
182
- **APA:**
183
-
184
- InferenceLab. (2025). *MediLlama-3.2: A Medical Chatbot Fine-Tuned from LLaMA 3.2*. Hugging Face. [https://huggingface.co/InferenceLab/MediLlama-3.2](https://huggingface.co/InferenceLab/MediLlama-3.2)
185
 
186
  ## Glossary
187
 
@@ -197,8 +171,6 @@ For collaborations, deployment help, or fine-tuning extensions, please contact t
197
 
198
  * InferenceLab Team
199
 
200
- ## Model Card Contact
201
 
202
203
 
204
 
 
32
  - **License:** Apache 2.0
33
  - **Finetuned from model:** meta-llama/Llama-3.2-3B-Instruct
34
 
 
 
 
 
 
 
35
  ## Uses
36
 
37
  ### Direct Use
 
90
  #### Training Hyperparameters
91
 
92
  * **Training regime:** bf16 mixed precision
93
+ * **Learning rate:** 1e-5
 
 
94
 
95
  #### Speeds, Sizes, Times
96
 
97
  * **Training time:** \~12 hours on 4×A100 GPUs
 
98
 
99
  ## Evaluation
100
 
 
156
  * Unsloth
157
  * PyTorch 2.1
158
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
159
 
160
  ## Glossary
161
 
 
171
 
172
  * InferenceLab Team
173
 
 
174
 
 
175
 
176