SebastianSchramm commited on
Commit
a83e129
·
1 Parent(s): 5b6fd23

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +44 -0
README.md ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ pipeline_tag: text-generation
5
+ library_name: transformers
6
+ tags:
7
+ - cerebras
8
+ - LLM
9
+ ---
10
+
11
+ # Instruction-tuned Cerebras GPT 111M
12
+
13
+ The smallest of [cerebras GPT models](https://huggingface.co/cerebras) with only 111M parameters instruction fine-tuned.
14
+
15
+ ## Model Description
16
+
17
+ Instruction fine-tuned [cerebras-GPT-111M](https://huggingface.co/cerebras/Cerebras-GPT-111M)
18
+
19
+ ## Training data
20
+
21
+ The model was fine-tuned with the following data: [alpaca_gpt4_data](https://github.com/Instruction-Tuning-with-GPT-4/GPT-4-LLM/blob/main/data/alpaca_gpt4_data.json) (data generated by GPT-4 using Alpaca prompts for fine-tuning LLMs) and [alpaca_data_cleaned](https://github.com/tloen/alpaca-lora/blob/a3027fea37c2087b8b0131b21a4cd948bbdcd9e0/alpaca_data_cleaned.json).
22
+
23
+ ## Prompt template
24
+
25
+ Fine-tuning was performed with the promp template from [stanford alpaca](https://github.com/tatsu-lab/stanford_alpaca):
26
+
27
+ ```python
28
+ PROMPT_DICT = {
29
+ "prompt_input": (
30
+ "Below is an instruction that describes a task, paired with an input that provides further context. "
31
+ "Write a response that appropriately completes the request.\n\n"
32
+ "### Instruction:\n{instruction}\n\n### Input:\n{input}\n\n### Response:"
33
+ ),
34
+ "prompt_no_input": (
35
+ "Below is an instruction that describes a task. "
36
+ "Write a response that appropriately completes the request.\n\n"
37
+ "### Instruction:\n{instruction}\n\n### Response:"
38
+ ),
39
+ }
40
+ ```
41
+
42
+ ## Usage
43
+
44
+ It is recommended to format input according to the prompt template mentioned above during inference for best results.