Text Generation
Transformers
PyTorch
English
llama
Inference Endpoints
text-generation-inference
arielnlee commited on
Commit
812717c
1 Parent(s): 342e41d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +14 -7
README.md CHANGED
@@ -10,26 +10,29 @@ license: cc-by-nc-4.0
10
 
11
  OpenOrca-Platypus2-13B is a merge of [`garage-bAInd/Platypus2-13B`](https://huggingface.co/garage-bAInd/Platypus2-13B) and [`Open-Orca/OpenOrcaxOpenChat-Preview2-13B`](https://huggingface.co/Open-Orca/OpenOrcaxOpenChat-Preview2-13B).
12
 
 
 
13
  ![Platty](./Best_Platty_small.jpeg)
14
 
15
  ### Benchmark Metrics
16
 
17
  | Metric | Value |
18
  |-----------------------|-------|
19
- | MMLU (5-shot) | - |
20
- | ARC (25-shot) | - |
21
- | HellaSwag (10-shot) | - |
22
- | TruthfulQA (0-shot) | - |
23
- | Avg. | - |
24
 
25
  We use state-of-the-art [Language Model Evaluation Harness](https://github.com/EleutherAI/lm-evaluation-harness) to run the benchmark tests above, using the same version as the HuggingFace LLM Leaderboard. Please see below for detailed instructions on reproducing benchmark results.
26
 
27
  ### Model Details
28
 
29
- * **Trained by**: **Platypus2-13B** trained by Cole Hunter & Ariel Lee; **OpenOrcaxOpenChat-Preview2-13B** trained by OpenOrca
30
  * **Model type:** **OpenOrca-Platypus2-13B** is an auto-regressive language model based on the LLaMA 2 transformer architecture.
31
  * **Language(s)**: English
32
  * **License for Platypus2-13B base weights**: Non-Commercial Creative Commons license ([CC BY-NC-4.0](https://creativecommons.org/licenses/by-nc/4.0/))
 
33
 
34
  ### Prompt Template for base Platypus2-13B
35
  ```
@@ -39,13 +42,17 @@ We use state-of-the-art [Language Model Evaluation Harness](https://github.com/E
39
 
40
  ### Response:
41
  ```
 
 
42
 
43
- ### Training Dataset
44
 
45
  `garage-bAInd/Platypus2-13B` trained using STEM and logic based dataset [`garage-bAInd/Open-Platypus`](https://huggingface.co/datasets/garage-bAInd/Open-Platypus).
46
 
47
  Please see our [paper](https://platypus-llm.github.io/Platypus.pdf) and [project webpage](https://platypus-llm.github.io) for additional information.
48
 
 
 
49
  ### Training Procedure
50
 
51
  `garage-bAInd/Platypus2-13B` was instruction fine-tuned using LoRA on 1 A100 80GB. For training details and inference instructions please see the [Platypus](https://github.com/arielnlee/Platypus) GitHub repo.
 
10
 
11
  OpenOrca-Platypus2-13B is a merge of [`garage-bAInd/Platypus2-13B`](https://huggingface.co/garage-bAInd/Platypus2-13B) and [`Open-Orca/OpenOrcaxOpenChat-Preview2-13B`](https://huggingface.co/Open-Orca/OpenOrcaxOpenChat-Preview2-13B).
12
 
13
+ Thank you Open-Orca for putting out a beast of a model and dataset—can't wait for the 70B version (and beyond)!
14
+
15
  ![Platty](./Best_Platty_small.jpeg)
16
 
17
  ### Benchmark Metrics
18
 
19
  | Metric | Value |
20
  |-----------------------|-------|
21
+ | MMLU (5-shot) | 59.5 |
22
+ | ARC (25-shot) | 62.88 |
23
+ | HellaSwag (10-shot) | 83.19 |
24
+ | TruthfulQA (0-shot) | 52.69 |
25
+ | Avg. | 64.56 |
26
 
27
  We use state-of-the-art [Language Model Evaluation Harness](https://github.com/EleutherAI/lm-evaluation-harness) to run the benchmark tests above, using the same version as the HuggingFace LLM Leaderboard. Please see below for detailed instructions on reproducing benchmark results.
28
 
29
  ### Model Details
30
 
31
+ * **Trained by**: **Platypus2-13B** trained by Cole Hunter & Ariel Lee; **OpenOrcaxOpenChat-Preview2-13B** trained by Open-Orca
32
  * **Model type:** **OpenOrca-Platypus2-13B** is an auto-regressive language model based on the LLaMA 2 transformer architecture.
33
  * **Language(s)**: English
34
  * **License for Platypus2-13B base weights**: Non-Commercial Creative Commons license ([CC BY-NC-4.0](https://creativecommons.org/licenses/by-nc/4.0/))
35
+ * **License for OpenOrcaxOpenChat-Preview2-13B base weights**: LLaMa-2 commercial
36
 
37
  ### Prompt Template for base Platypus2-13B
38
  ```
 
42
 
43
  ### Response:
44
  ```
45
+ ### Prompt Template for base OpenOrcaxOpenChat-Preview2-13B
46
+ OpenChat Llama2 V1: see [Open-Orca's page](https://huggingface.co/Open-Orca/OpenOrcaxOpenChat-Preview2-13B) for additional information.
47
 
48
+ ### Training Datasets
49
 
50
  `garage-bAInd/Platypus2-13B` trained using STEM and logic based dataset [`garage-bAInd/Open-Platypus`](https://huggingface.co/datasets/garage-bAInd/Open-Platypus).
51
 
52
  Please see our [paper](https://platypus-llm.github.io/Platypus.pdf) and [project webpage](https://platypus-llm.github.io) for additional information.
53
 
54
+ [`Open-Orca/OpenOrcaxOpenChat-Preview2-13B`] trained using a refined, 220k subset of the [OpenOrca dataset](https://huggingface.co/datasets/Open-Orca/OpenOrca).
55
+
56
  ### Training Procedure
57
 
58
  `garage-bAInd/Platypus2-13B` was instruction fine-tuned using LoRA on 1 A100 80GB. For training details and inference instructions please see the [Platypus](https://github.com/arielnlee/Platypus) GitHub repo.