Text Generation
Transformers
PyTorch
Safetensors
English
llama
Eval Results
text-generation-inference
Inference Endpoints
Pankaj Mathur commited on
Commit
7c09fc9
1 Parent(s): 17b2ccf

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +12 -5
README.md CHANGED
@@ -3,20 +3,27 @@ license: mit
3
  language:
4
  - en
5
  library_name: transformers
 
 
 
 
 
6
  ---
7
- # Wizardlm Alpaca Dolly Orca Open_LLaMa_3b
8
- An Open_LLaMA-3B model trained on custom explain tuned datasets, created using Instructions and Input from WizardLM, Alpaca & Dolly-V2 datasets and applying Orca Research Paper dataset construction approaches.
 
9
 
10
 
11
  # Dataset
12
 
13
- We trained [OpenLLaMa-3B model](https://github.com/openlm-research/open_llama) on custom explain tuned [WizardLM dataset ~70K](https://github.com/nlpxucan/WizardLM), [Alpaca dataset ~52K](https://crfm.stanford.edu/2023/03/13/alpaca.html) & [Dolly-V2 dataset ~15K](https://github.com/databrickslabs/dolly) created using approaches from [Orca Research Paper](https://arxiv.org/abs/2306.02707).
14
 
15
  We leverage all of the 15 system instructions provided in Orca Research Paper. to generate custom datasets, in contrast to vanilla instruction tuning approaches used by original datasets.
16
 
17
- This helps student model aka [wizardlm_alpaca_dolly_orca_open_llama_3b](https://huggingface.co/psmathur/wizardlm_alpaca_dolly_orca_open_llama_3b) to learn ***thought*** process from teacher model, which is ChatGPT (gpt-3.5-turbo-0301 version).
 
 
18
 
19
- Please see below example usage how the **System** prompt is added before each *instruction*.
20
 
21
  # Training
22
 
 
3
  language:
4
  - en
5
  library_name: transformers
6
+
7
+ datasets:
8
+ - psmathur/alpaca_orca
9
+ - psmathur/dolly-v2_orca
10
+ - psmathur/WizardLM_Orca
11
  ---
12
+ # orca_mini_3b
13
+
14
+ An [OpenLLaMa-3B model](https://github.com/openlm-research/open_llama) model trained on explain tuned datasets, created using Instructions and Input from WizardLM, Alpaca & Dolly-V2 datasets and applying Orca Research Paper dataset construction approaches.
15
 
16
 
17
  # Dataset
18
 
19
+ We build explain tuned [WizardLM dataset ~70K](https://github.com/nlpxucan/WizardLM), [Alpaca dataset ~52K](https://crfm.stanford.edu/2023/03/13/alpaca.html) & [Dolly-V2 dataset ~15K](https://github.com/databrickslabs/dolly) created using approaches from [Orca Research Paper](https://arxiv.org/abs/2306.02707).
20
 
21
  We leverage all of the 15 system instructions provided in Orca Research Paper. to generate custom datasets, in contrast to vanilla instruction tuning approaches used by original datasets.
22
 
23
+ This helps student model aka [wizardlm_alpaca_dolly_orca_open_llama_13b](https://huggingface.co/psmathur/wizardlm_alpaca_dolly_orca_open_llama_13b) to learn ***thought*** process from teacher model, which is ChatGPT (gpt-3.5-turbo-0301 version).
24
+
25
+ Please see below example usage how the **System** prompt is added before each **instruction**.
26
 
 
27
 
28
  # Training
29