manojpreveen
commited on
Commit
•
0595a70
1
Parent(s):
149d292
Update README.md
Browse files
README.md
CHANGED
@@ -36,7 +36,7 @@ The prompt template for the multi-turn conversation looks like this:
|
|
36 |
|
37 |
This model follows the official Meta's chat model Prompt format. Please refer here : https://huggingface.co/blog/llama2#how-to-prompt-llama-2 on how to prompt the model for single/multi-turn conversations.
|
38 |
|
39 |
-
**Base model :** meta-llama/Llama-2-
|
40 |
|
41 |
**Data :**
|
42 |
1. 1M Orca dara (Gpt-4 Orca data - OpenOrca)
|
@@ -46,7 +46,7 @@ This model follows the official Meta's chat model Prompt format. Please refer he
|
|
46 |
**Training Params :**
|
47 |
```
|
48 |
Number of Epochs : 1
|
49 |
-
Batch Size :
|
50 |
Sequence Length : 4096
|
51 |
Learning Rate : 2e-5 (Cosine)
|
52 |
Weight Decay : 0.1
|
|
|
36 |
|
37 |
This model follows the official Meta's chat model Prompt format. Please refer here : https://huggingface.co/blog/llama2#how-to-prompt-llama-2 on how to prompt the model for single/multi-turn conversations.
|
38 |
|
39 |
+
**Base model :** meta-llama/Llama-2-70b-hf
|
40 |
|
41 |
**Data :**
|
42 |
1. 1M Orca dara (Gpt-4 Orca data - OpenOrca)
|
|
|
46 |
**Training Params :**
|
47 |
```
|
48 |
Number of Epochs : 1
|
49 |
+
Batch Size : 64
|
50 |
Sequence Length : 4096
|
51 |
Learning Rate : 2e-5 (Cosine)
|
52 |
Weight Decay : 0.1
|