xmanii commited on
Commit
514b017
1 Parent(s): 7d15c39

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +33 -11
README.md CHANGED
@@ -1,17 +1,39 @@
1
- Model Information
 
 
 
2
 
3
- Developed by: xmanii License: Apache-2.0 Finetuned from model: unsloth/llama-3-8b-instruct-bnb-4bit
 
 
4
 
5
- This LLaMA model was fine-tuned on a unique Persian dataset of Alpaca chat conversations, consisting of approximately 8,000 rows. Our training process utilized two H100 GPUs, completing in just under 1 hour. We leveraged the power of Unsloth and Hugging Face's TRL library to accelerate our training process by 2x.
6
 
7
- <img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>
8
 
9
- This model is open-source, and we invite the community to use and build upon our work. The fine-tuned LLaMA model is designed to improve Persian conversation capabilities, and we hope it will contribute to the advancement of natural language processing in the Persian language.
10
- Using the Model
11
 
12
- To use this model, you can utilize the Hugging Face Transformers library. Note: The default usage code provided by Hugging Face is not applicable for this model. Instead, follow the example below:
13
 
14
- messages = [ {"from": "human", "value": prompt},]
15
- Finally, use the pipeline to generate responses:
16
- pipe = pipeline("text-generation", model="xmanii/Llama3-8b-simorgh-16bit")
17
- pipe(messages)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model-index:
2
+ - name: xmanii/llama-3-8b-instruct-bnb-4bit-persian
3
+ description: |
4
+ **Model Information**
5
 
6
+ **Developed by:** xmanii
7
+ **License:** Apache-2.0
8
+ **Finetuned from model:** unsloth/llama-3-8b-instruct-bnb-4bit
9
 
10
+ **Model Description**
11
 
12
+ This LLaMA model was fine-tuned on a unique Persian dataset of Alpaca chat conversations, consisting of approximately 8,000 rows. Our training process utilized two H100 GPUs, completing in just under 1 hour. We leveraged the power of Unsloth and Hugging Face's TRL library to accelerate our training process by 2x.
13
 
14
+ **Open-Source Contribution**
 
15
 
16
+ This model is open-source, and we invite the community to use and build upon our work. The fine-tuned LLaMA model is designed to improve Persian conversation capabilities, and we hope it will contribute to the advancement of natural language processing in the Persian language.
17
 
18
+ **Using the Model**
19
+
20
+ To use this model, you can utilize the Hugging Face Transformers library. **Note:** The default usage code provided by Hugging Face is not applicable for this model. Instead, follow the example below:
21
+
22
+ ```python
23
+ messages = [{"from": "human", "value": prompt},]
24
+ ```
25
+
26
+ Finally, use the pipeline to generate responses:
27
+
28
+ ```python
29
+ pipe = pipeline("text-generation", model="xmanii/Llama3-8b-simorgh-16bit")
30
+ pipe(messages)
31
+ ```
32
+
33
+ **Full 16-bit Merged Model**
34
+
35
+ For a full 16-bit merged model, please check out xmanii/Llama3-8b-simorgh-16bit.
36
+
37
+ **Future Work**
38
+
39
+ We are working on quantizing the models and bringing them to ollama.