mike-conover-db commited on
Commit
f87ac98
1 Parent(s): 5447775

Updating README.

Browse files
Files changed (1) hide show
  1. README.md +86 -21
README.md CHANGED
@@ -1,32 +1,97 @@
 
 
 
1
  # Summary
2
- `databricks-dolly-15k` is an open source dataset of instruction-following records generated by thousnads of Databricks empoyees in several
3
- of the behavioral categories outlined in the InstructGPT paper, including brainstorming, classification, closed QA, generation, information
4
- extraction, open QA, and summarization.
 
 
 
 
 
 
 
 
 
 
 
 
 
5
 
6
- [LEGAL]
7
- This dataset is licensed for commercial use ...
8
 
9
  # Dataset Overview
10
- Databricks employees were invited to create prompt / response pairs using Google Forms in each of eight different instruction categories, including the seven
11
- outlined above as well as an open-ended free-form category. Employees were instructed to avoid using information from any source on the web with
12
- the exception of Wikipedia, and explicitly instructed to avoid using generative AI in formulating instructions or responses. Examples of each behavior were
13
- provided to motivate the types of questions and instructions appropriate to each category.
 
 
 
 
 
 
14
 
15
- Halfway through the data generation process, participants were given the option of answering questions posed by other employees. They were
16
- asked to rephrase the original question and only select questions they could be reasonably expected to answer correctly.
17
 
18
- For questions in the XYZ categories employees were asked to provide a reference passage copied from Wikipedia. This text may contain bracketed
19
- citations which we recommend users remove for downstream applications.
20
 
21
  # Intended Uses
22
  While immediately valuable for instruction fine tuning large language models, as a corpus of human-generated instruction prompts,
23
- this dataset also presents a valuable opportunity for synthetic data generation in the methods outlined in the [Self-Instruct](https://arxiv.org/abs/2212.10560) paper.
24
- For example, employee-generated prompts could be submitted as few-shot examples to a large open language model to generate a corpus of millions of examples of
25
- instructions in each of the respective InstructGPT categories.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
26
 
27
- Likewise, both the instructions and responses present fertile ground for data augmentation. A paraphrasing model might be used to restate each prompt or short responses, with the
28
- resulting text associated to the respective ground-truth sample. Such an approach might provide a form of regularization on the dataset that could allow for
29
- more robust instruction-following behavior in models derived from these synthetic datasets.
30
 
31
- # Dataset Limitations
32
- [LEGAL]
 
1
+ ---
2
+ license: cc-by-sa-3.0
3
+ ---
4
  # Summary
5
+ `databricks-dolly-15k` is an open source dataset of instruction-following records generated by thousands of Databricks employees in several
6
+ of the behavioral categories outlined in the [InstructGPT](https://arxiv.org/abs/2203.02155) paper, including brainstorming, classification,
7
+ closed QA, generation, information extraction, open QA, and summarization.
8
+
9
+ This dataset can be used for any purpose, whether academic or commercial, under the terms of the
10
+ [Creative Commons Attribution-ShareAlike 3.0 Unported License](https://creativecommons.org/licenses/by-sa/3.0/legalcode).
11
+
12
+ Supported Tasks:
13
+ - Training LLMs
14
+ - Synthetic Data Generation
15
+ - Data Augmentation
16
+
17
+ Languages: English
18
+ Version: 1.0
19
+
20
+ **Owner: Databricks, Inc.**
21
 
 
 
22
 
23
  # Dataset Overview
24
+ `databricks-dolly-15k` is a corpus of more than 15,000 records generated by thousands of Databricks employees to enable large language
25
+ models to exhibit the magical interactivity of ChatGPT.
26
+ Databricks employees were invited to create prompt / response pairs in each of eight different instruction categories, including
27
+ the seven outlined in the InstructGPT paper, as well as an open-ended free-form category. The contributors were instructed to avoid using
28
+ information from any source on the web with the exception of Wikipedia (for particular subsets of instruction categories), and explicitly
29
+ instructed to avoid using generative AI in formulating instructions or responses. Examples of each behavior were provided to motivate the
30
+ types of questions and instructions appropriate to each category.
31
+
32
+ Halfway through the data generation process, contributors were given the option of answering questions posed by other contributors.
33
+ They were asked to rephrase the original question and only select questions they could be reasonably expected to answer correctly.
34
 
35
+ For certain categories contributors were asked to provide reference texts copied from Wikipedia. Reference text (indicated by the `context`
36
+ field in the actual dataset) may contain bracketed Wikipedia citation numbers (e.g. `[42]`) which we recommend users remove for downstream applications.
37
 
 
 
38
 
39
  # Intended Uses
40
  While immediately valuable for instruction fine tuning large language models, as a corpus of human-generated instruction prompts,
41
+ this dataset also presents a valuable opportunity for synthetic data generation in the methods outlined in the Self-Instruct paper.
42
+ For example, contributor--generated prompts could be submitted as few-shot examples to a large open language model to generate a
43
+ corpus of millions of examples of instructions in each of the respective InstructGPT categories.
44
+
45
+ Likewise, both the instructions and responses present fertile ground for data augmentation. A paraphrasing model might be used to
46
+ restate each prompt or short responses, with the resulting text associated to the respective ground-truth sample. Such an approach might
47
+ provide a form of regularization on the dataset that could allow for more robust instruction-following behavior in models derived from
48
+ these synthetic datasets.
49
+
50
+
51
+ # Dataset
52
+ ## Purpose of Collection
53
+ As part of our continuing commitment to open source, Databricks developed what is, to the best of our knowledge, the first open source,
54
+ human-generated instruction corpus specifically designed to enable large language models to exhibit the magical interactivity of ChatGPT.
55
+ Unlike other datasets that are limited to non-commercial use, this dataset can be used, modified, and extended for any purpose, including
56
+ academic or commercial applications.
57
+
58
+ ## Sources
59
+ - **Human-generated data**: Databricks employees were invited to create prompt / response pairs in each of eight different instruction categories.
60
+ - **Wikipedia**: For instruction categories that require an annotator to consult a reference text (information extraction, closed QA, summarization)
61
+ contributors selected passages from Wikipedia for particular subsets of instruction categories. No guidance was given to annotators as to how to select the
62
+ target passages.
63
+
64
+ ## Annotator Guidelines
65
+ To create a record, employees were given a brief description of the annotation task as well as examples of the types of prompts typical
66
+ of each annotation task. Guidelines were succinct by design so as to encourage a high task completion rate, possibly at the cost of
67
+ rigorous compliance to an annotation rubric that concretely and reliably operationalizes the specific task. Caveat emptor.
68
+
69
+ The annotation guidelines for each of the categories are as follows:
70
+
71
+ - **Creative Writing**: Write a question or instruction that requires a creative, open-ended written response. The instruction should be reasonable to ask of a person with general world knowledge and should not require searching. In this task, your prompt should give very specific instructions to follow. Constraints, instructions, guidelines, or requirements all work, and the more of them the better.
72
+ - **Closed QA**: Write a question or instruction that requires factually correct response based on a passage of text from Wikipedia. The question can be complex and can involve human-level reasoning capabilities, but should not require special knowledge. To create a question for this task include both the text of the question as well as the reference text in the form.
73
+ - **Open QA**: Write a question that can be answered using general world knowledge or at most a single search. This task asks for opinions and facts about the world at large and does not provide any reference text for consultation.
74
+ - **Summarization**: Give a summary of a paragraph from Wikipedia. Please don't ask questions that will require more than 3-5 minutes to answer. To create a question for this task include both the text of the question as well as the reference text in the form.
75
+ - **Information Extraction**: These questions involve reading a paragraph from Wikipedia and extracting information from the passage. Everything required to produce an answer (e.g. a list, keywords etc) should be included in the passages. To create a question for this task include both the text of the question as well as the reference text in the form.
76
+ - **Classification**: These prompts contain lists or examples of entities to be classified, e.g. movie reviews, products, etc. In this task the text or list of entities under consideration is contained in the prompt (e.g. there is no reference text.). You can choose any categories for classification you like, the more diverse the better.
77
+ - **Brainstorming**: Think up lots of examples in response to a question asking to brainstorm ideas.
78
+
79
+ ## Personal or Sensitive Data
80
+ This dataset contains public information (e.g., some information from Wikipedia). To our knowledge, there are no private person’s personal identifiers or sensitive information.
81
+
82
+ ## Language
83
+ American English
84
+
85
+ # Known Limitations
86
+ - Wikipedia is a crowdsourced corpus and the contents of this dataset may reflect the bias, factual errors and topical focus found in Wikipedia
87
+ - Some annotators may not be native English speakers
88
+ - Annotator demographics and subject matter may reflect the makeup of Databricks employees
89
+
90
+ # License/Attribution
91
+ **Copyright (2023) Databricks, Inc.**
92
+ This dataset was developed at Databricks (https://www.databricks.com) and its use is subject to the CC BY-SA 3.0 license.
93
 
94
+ Certain categories of material in the dataset include materials from the following sources, licensed under the CC BY-SA 3.0 license:
 
 
95
 
96
+ Wikipedia (various pages) - https://www.wikipedia.org/
97
+ Copyright © Wikipedia editors and contributors.