ajibawa-2023 commited on
Commit
9755981
1 Parent(s): 0982a20

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -13,7 +13,7 @@ Besides this it is trained on following datasets:
13
 
14
  [orca-math-word-problems-200k](https://huggingface.co/datasets/microsoft/orca-math-word-problems-200k)
15
 
16
- [Openhermes](https://huggingface.co/datasets/teknium/openhermes)
17
 
18
  The idea was to check how this Model will perform with both Code & Maths datasets. This model is very good with Coding.
19
  Maths outputs are also very good. You can test out this model.
@@ -34,7 +34,7 @@ Exllama v2: TBA
34
 
35
  **Training:**
36
 
37
- Entire dataset was trained on 4 x A100 80GB. For 2 epoch, training took more than 160 Hours. Axolotl codebase was used for training purpose.
38
  Entire data is trained on Llama-3-8B by Meta.
39
 
40
  **Example Prompt:**
 
13
 
14
  [orca-math-word-problems-200k](https://huggingface.co/datasets/microsoft/orca-math-word-problems-200k)
15
 
16
+ [CodeFeedback-Filtered-Instruction](https://huggingface.co/datasets/m-a-p/CodeFeedback-Filtered-Instruction)
17
 
18
  The idea was to check how this Model will perform with both Code & Maths datasets. This model is very good with Coding.
19
  Maths outputs are also very good. You can test out this model.
 
34
 
35
  **Training:**
36
 
37
+ Entire dataset was trained on 4 x A100 80GB. For 2 epoch, training took more than 160 Hours. Axolotl & Deepspeed codebase was used for training purpose.
38
  Entire data is trained on Llama-3-8B by Meta.
39
 
40
  **Example Prompt:**