andreaskoepf commited on
Commit
7f844fd
1 Parent(s): 776288c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +23 -6
README.md CHANGED
@@ -7,7 +7,9 @@ datasets:
7
  - atom-in-the-universe/fanfics-10k-50k
8
  ---
9
 
10
- - **At least Huggingface Transformers [4.31.0](https://pypi.org/project/transformers/4.31.0/) is required to load this model!**
 
 
11
  - base model: [meta-llama/Llama-2-7b](https://huggingface.co/meta-llama/Llama-2-7b)
12
  - License: [Llama 2 Community License Agreement](https://ai.meta.com/resources/models-and-libraries/llama-downloads/)
13
  - wandb: [public-sft/runs/2jfazjt9](https://wandb.ai/open-assistant/public-sft/runs/2jfazjt9)
@@ -23,17 +25,19 @@ HF transformers >=4.31.0 is installed (`pip install transformers>=4.31.0`).
23
 
24
  ## Conversation Template
25
 
 
 
26
  ```
27
  <|system|>system message</s><|prompter|>user prompt</s><|assistant|>
28
  ```
29
 
30
- For multi-turn conversations:
31
 
32
  ```
33
  <|system|>system message</s><|prompter|>Q1</s><|assistant|>A1</s><|prompter|>Q2</s><|assistant|>
34
  ```
35
 
36
- The model was trained with the following 16 system messages that were used to generate the training examples (see [ORCA paper](https://arxiv.org/abs/2306.02707)):
37
 
38
  1. \<empty system message\>
39
  2. You are an AI assistant. Provide a detailed answer so user don’t need to search outside to understand the answer.
@@ -53,10 +57,19 @@ The model was trained with the following 16 system messages that were used to ge
53
  16. You are an AI assistant that helps people find information.
54
 
55
 
56
- ## Orca-Chat/Dolphin Datasets
 
 
57
 
58
- This model is trained on []()
59
- https://huggingface.co/datasets/ehartford/dolphin
 
 
 
 
 
 
 
60
 
61
 
62
  ## Model Configuration
@@ -110,6 +123,10 @@ llama2_13b_orca_8k:
110
  peft_model: false
111
  ```
112
 
 
 
 
 
113
  # License
114
 
115
  - Llama 2 is licensed under the LLAMA 2 Community License, Copyright © Meta Platforms, Inc. All Rights Reserved.
 
7
  - atom-in-the-universe/fanfics-10k-50k
8
  ---
9
 
10
+ Note: **At least Huggingface Transformers [4.31.0](https://pypi.org/project/transformers/4.31.0/) is required to load this model!**
11
+
12
+
13
  - base model: [meta-llama/Llama-2-7b](https://huggingface.co/meta-llama/Llama-2-7b)
14
  - License: [Llama 2 Community License Agreement](https://ai.meta.com/resources/models-and-libraries/llama-downloads/)
15
  - wandb: [public-sft/runs/2jfazjt9](https://wandb.ai/open-assistant/public-sft/runs/2jfazjt9)
 
25
 
26
  ## Conversation Template
27
 
28
+ For the initial response use (the system message is optional):
29
+
30
  ```
31
  <|system|>system message</s><|prompter|>user prompt</s><|assistant|>
32
  ```
33
 
34
+ For multi-turn conversations use:
35
 
36
  ```
37
  <|system|>system message</s><|prompter|>Q1</s><|assistant|>A1</s><|prompter|>Q2</s><|assistant|>
38
  ```
39
 
40
+ The model was trained with the following 16 system messages used to generate the training examples (see [ORCA paper](https://arxiv.org/abs/2306.02707)):
41
 
42
  1. \<empty system message\>
43
  2. You are an AI assistant. Provide a detailed answer so user don’t need to search outside to understand the answer.
 
57
  16. You are an AI assistant that helps people find information.
58
 
59
 
60
+ ## Datasets: Orca-Chat/Dolphin, RedPajama1T & FanFics
61
+
62
+ This model was trained on:
63
 
64
+ - [shahules786/orca-chat](https://huggingface.co/datasets/shahules786/orca-chat)
65
+ - [togethercomputer/RedPajama-Data-1T-Sample](https://huggingface.co/datasets/togethercomputer/RedPajama-Data-1T)
66
+ - [atom-in-the-universe/fanfics-10k-50k](https://huggingface.co/datasets/atom-in-the-universe/fanfics-10k-50k)
67
+
68
+ The dataset [shahules786/orca-chat](https://huggingface.co/datasets/shahules786/orca-chat) combines similar
69
+ examples of the GPT-4 subset of [ehartford/dolphin](https://huggingface.co/datasets/ehartford/dolphin) to form longer conversations
70
+ to improve long-context trainig.
71
+
72
+ RedPajama and FanFics were additionally used for classic language modelling to fine-tune the RoPE scaling for 8k context size.
73
 
74
 
75
  ## Model Configuration
 
123
  peft_model: false
124
  ```
125
 
126
+ # Special Thanks
127
+
128
+ We want to especially thank Eric Hardford for replicating ORCA and making it publicly available at [ehartford/dolphin](https://huggingface.co/datasets/ehartford/dolphin)!
129
+
130
  # License
131
 
132
  - Llama 2 is licensed under the LLAMA 2 Community License, Copyright © Meta Platforms, Inc. All Rights Reserved.