Update README.md
Browse files
README.md
CHANGED
@@ -1,17 +1,39 @@
|
|
1 |
-
|
|
|
|
|
|
|
2 |
|
3 |
-
Developed by
|
|
|
|
|
4 |
|
5 |
-
|
6 |
|
7 |
-
|
8 |
|
9 |
-
|
10 |
-
Using the Model
|
11 |
|
12 |
-
|
13 |
|
14 |
-
|
15 |
-
|
16 |
-
|
17 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
model-index:
|
2 |
+
- name: xmanii/llama-3-8b-instruct-bnb-4bit-persian
|
3 |
+
description: |
|
4 |
+
**Model Information**
|
5 |
|
6 |
+
**Developed by:** xmanii
|
7 |
+
**License:** Apache-2.0
|
8 |
+
**Finetuned from model:** unsloth/llama-3-8b-instruct-bnb-4bit
|
9 |
|
10 |
+
**Model Description**
|
11 |
|
12 |
+
This LLaMA model was fine-tuned on a unique Persian dataset of Alpaca chat conversations, consisting of approximately 8,000 rows. Our training process utilized two H100 GPUs, completing in just under 1 hour. We leveraged the power of Unsloth and Hugging Face's TRL library to accelerate our training process by 2x.
|
13 |
|
14 |
+
**Open-Source Contribution**
|
|
|
15 |
|
16 |
+
This model is open-source, and we invite the community to use and build upon our work. The fine-tuned LLaMA model is designed to improve Persian conversation capabilities, and we hope it will contribute to the advancement of natural language processing in the Persian language.
|
17 |
|
18 |
+
**Using the Model**
|
19 |
+
|
20 |
+
To use this model, you can utilize the Hugging Face Transformers library. **Note:** The default usage code provided by Hugging Face is not applicable for this model. Instead, follow the example below:
|
21 |
+
|
22 |
+
```python
|
23 |
+
messages = [{"from": "human", "value": prompt},]
|
24 |
+
```
|
25 |
+
|
26 |
+
Finally, use the pipeline to generate responses:
|
27 |
+
|
28 |
+
```python
|
29 |
+
pipe = pipeline("text-generation", model="xmanii/Llama3-8b-simorgh-16bit")
|
30 |
+
pipe(messages)
|
31 |
+
```
|
32 |
+
|
33 |
+
**Full 16-bit Merged Model**
|
34 |
+
|
35 |
+
For a full 16-bit merged model, please check out xmanii/Llama3-8b-simorgh-16bit.
|
36 |
+
|
37 |
+
**Future Work**
|
38 |
+
|
39 |
+
We are working on quantizing the models and bringing them to ollama.
|