VictorSanh
commited on
Commit
•
f52f478
1
Parent(s):
bbad117
update script and data card
Browse files
P3.py
CHANGED
@@ -12,7 +12,7 @@
|
|
12 |
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
13 |
# See the License for the specific language governing permissions and
|
14 |
# limitations under the License.
|
15 |
-
"""P3"""
|
16 |
|
17 |
|
18 |
import datasets
|
@@ -27,7 +27,7 @@ _CITATION = """\
|
|
27 |
TODO"""
|
28 |
|
29 |
_DESCRIPTION = """\
|
30 |
-
P3 is a collection of prompted English datasets covering a diverse set of NLP tasks. A prompt is the combination of an input template and a target template. The templates are functions mapping a data example into natural language for the input and target sequences. For example, in the case of an NLI dataset, the data example would include fields for *Premise, Hypothesis, Label*. An input template would be *If {Premise} is true, is it also true that {Hypothesis}?*, whereas a target template can be defined with the label choices *Choices[label]*. Here *Choices* is prompt-specific metadata that consists of the options *yes, maybe, no* corresponding to *label* being entailment (0), neutral (1) or contradiction (2).
|
31 |
|
32 |
Prompts are collected using [Promptsource](https://github.com/bigscience-workshop/promptsource), an interface to interactively write prompts on datasets, and collect prompt-specific metadata such as evaluation metrics. As of October 13th, there are 2'000 prompts collected for 270+ data(sub)sets. The collection of prompts is publicly available on [Promptsource](https://github.com/bigscience-workshop/promptsource).
|
33 |
|
@@ -83,8 +83,6 @@ def find_task_splits_and_features():
|
|
83 |
"""Find the available tasks under ./data and their available splits and features."""
|
84 |
task_and_their_splits = defaultdict(dict)
|
85 |
for stats in glob.glob(f"{_DATA_PATH}/*/stats.*.json"):
|
86 |
-
if "anli" not in stats:
|
87 |
-
continue
|
88 |
folder_path = os.path.dirname(stats)
|
89 |
task_name = folder_path.split("/")[-1]
|
90 |
split_name = os.path.basename(stats).split(".")[1]
|
|
|
12 |
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
13 |
# See the License for the specific language governing permissions and
|
14 |
# limitations under the License.
|
15 |
+
"""P3 (Public Pool of Prompts)"""
|
16 |
|
17 |
|
18 |
import datasets
|
|
|
27 |
TODO"""
|
28 |
|
29 |
_DESCRIPTION = """\
|
30 |
+
P3 (Pubic Pool of Prompts)is a collection of prompted English datasets covering a diverse set of NLP tasks. A prompt is the combination of an input template and a target template. The templates are functions mapping a data example into natural language for the input and target sequences. For example, in the case of an NLI dataset, the data example would include fields for *Premise, Hypothesis, Label*. An input template would be *If {Premise} is true, is it also true that {Hypothesis}?*, whereas a target template can be defined with the label choices *Choices[label]*. Here *Choices* is prompt-specific metadata that consists of the options *yes, maybe, no* corresponding to *label* being entailment (0), neutral (1) or contradiction (2).
|
31 |
|
32 |
Prompts are collected using [Promptsource](https://github.com/bigscience-workshop/promptsource), an interface to interactively write prompts on datasets, and collect prompt-specific metadata such as evaluation metrics. As of October 13th, there are 2'000 prompts collected for 270+ data(sub)sets. The collection of prompts is publicly available on [Promptsource](https://github.com/bigscience-workshop/promptsource).
|
33 |
|
|
|
83 |
"""Find the available tasks under ./data and their available splits and features."""
|
84 |
task_and_their_splits = defaultdict(dict)
|
85 |
for stats in glob.glob(f"{_DATA_PATH}/*/stats.*.json"):
|
|
|
|
|
86 |
folder_path = os.path.dirname(stats)
|
87 |
task_name = folder_path.split("/")[-1]
|
88 |
split_name = os.path.basename(stats).split(".")[1]
|
README.md
CHANGED
@@ -32,13 +32,7 @@ task_categories:
|
|
32 |
- [Curation Rationale](#curation-rationale)
|
33 |
- [Source Data](#source-data)
|
34 |
- [Annotations](#annotations)
|
35 |
-
- [Personal and Sensitive Information](#personal-and-sensitive-information)
|
36 |
-
- [Considerations for Using the Data](#considerations-for-using-the-data)
|
37 |
-
- [Social Impact of Dataset](#social-impact-of-dataset)
|
38 |
-
- [Discussion of Biases](#discussion-of-biases)
|
39 |
-
- [Other Known Limitations](#other-known-limitations)
|
40 |
- [Additional Information](#additional-information)
|
41 |
-
- [Dataset Curators](#dataset-curators)
|
42 |
- [Licensing Information](#licensing-information)
|
43 |
- [Citation Information](#citation-information)
|
44 |
- [Contributions](#contributions)
|
@@ -183,6 +177,8 @@ The main annotation guideline was that prompts needed to be grammatical and unde
|
|
183 |
|
184 |
The full annotation given to the contributors can be found [here](https://github.com/bigscience-workshop/promptsource/blob/main/CONTRIBUTING.md). *Note to self: the link is currently being updated with the)
|
185 |
|
|
|
|
|
186 |
### Licensing Information
|
187 |
|
188 |
The dataset is released under Apache 2.0.
|
|
|
32 |
- [Curation Rationale](#curation-rationale)
|
33 |
- [Source Data](#source-data)
|
34 |
- [Annotations](#annotations)
|
|
|
|
|
|
|
|
|
|
|
35 |
- [Additional Information](#additional-information)
|
|
|
36 |
- [Licensing Information](#licensing-information)
|
37 |
- [Citation Information](#citation-information)
|
38 |
- [Contributions](#contributions)
|
|
|
177 |
|
178 |
The full annotation given to the contributors can be found [here](https://github.com/bigscience-workshop/promptsource/blob/main/CONTRIBUTING.md). *Note to self: the link is currently being updated with the)
|
179 |
|
180 |
+
## Additional Information
|
181 |
+
|
182 |
### Licensing Information
|
183 |
|
184 |
The dataset is released under Apache 2.0.
|