Pankaj Mathur commited on
Commit
ad636d6
1 Parent(s): 8889e72

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -5
README.md CHANGED
@@ -4,14 +4,20 @@ language:
4
  - en
5
  library_name: adapter-transformers
6
  ---
7
- # alpaca_orca_open_llama: An Open_LLaMA-3B model trained on Alpaca dataset using Orca Research paper approaches
8
 
9
 
10
- # Dataset and Training
11
 
12
- We train OpenLLaMa-3B model to become more steerable by training it on the custom Alpaca dataset created using approaches from [Orca Research Paper](https://arxiv.org/abs/2306.02707).
13
 
14
- Please pay attention how the **System** prompt is added before each *instruction*.
 
 
 
 
 
 
15
 
16
  The training configurations are provided in the table below.
17
 
@@ -84,4 +90,4 @@ Next Goals:
84
  2) Try bigger OpenLLaMA models 7B and 13B
85
  3) Try better GPU for training, couldn't get 8xA100 (40GB), I guess they are in hot demand now.
86
  4) Provide more options for Text generation UI. (may be https://github.com/oobabooga/text-generation-webui)
87
- 6) Provide 4bit quantized model (may be TheBloke can help here)
 
4
  - en
5
  library_name: adapter-transformers
6
  ---
7
+ # alpaca_orca_open_llama: An Open_LLaMA-3B model trained on custom Alpaca dataset using Orca Research paper approaches
8
 
9
 
10
+ # Dataset
11
 
12
+ We train OpenLLaMa-3B model on custom explained tuned Alpaca dataset (~52K) created using approaches from [Orca Research Paper](https://arxiv.org/abs/2306.02707).
13
 
14
+ We leverage all of the 15 system instructions provided in [Orca Research Paper](https://arxiv.org/abs/2306.02707) to generate custom Alpaca dataset, in contrast to vanilla instruction tuning approaches used by original [Alpaca research paper](https://crfm.stanford.edu/2023/03/13/alpaca.html).
15
+
16
+ This helps student model aka [alpaca_orca_open_llama_3b](psmathur/alpaca_orca_open_llama_3b) to learn ***thought*** process from teacher model, which is ChatGPT (gpt-3.5-turbo-0301 version).
17
+
18
+ Please pay attention how the **System** prompt is added before each *instruction* in below example usage.
19
+
20
+ # Training
21
 
22
  The training configurations are provided in the table below.
23
 
 
90
  2) Try bigger OpenLLaMA models 7B and 13B
91
  3) Try better GPU for training, couldn't get 8xA100 (40GB), I guess they are in hot demand now.
92
  4) Provide more options for Text generation UI. (may be https://github.com/oobabooga/text-generation-webui)
93
+ 6) Provide 4bit GGML/GPTQ quantized model (may be [TheBloke](https://huggingface.co/TheBloke) can help here)