File size: 3,332 Bytes
71754df
bd97593
71754df
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
bd97593
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
---
license: other
dataset_info:
  features:
  - name: inputs
    dtype: string
  - name: targets
    dtype: string
  - name: task_source
    dtype: string
  - name: task_name
    dtype: string
  - name: template_type
    dtype: string
  splits:
  - name: train
    num_bytes: 195544293.50492442
    num_examples: 116664
  download_size: 26571254
  dataset_size: 195544293.50492442
configs:
- config_name: default
  data_files:
  - split: train
    path: data/train-*
---

# Flan 2021 Coreference Tasks

- Project: https://github.com/google-research/FLAN/tree/main/flan/v2
- Data source: [DataProvenanceInitiative/flan2021_submix_original](https://huggingface.co/datasets/DataProvenanceInitiative/flan2021_submix_original)

## Details

This dataset contains all coreference examples that were included in the [Flan 2022 collection](https://github.com/google-research/FLAN/tree/main/flan/v2) which were orignally included in Flan 2021.

The data is copied from the preprocessed Flan2021 dataset at [DataProvenanceInitiative/flan2021_submix_original](https://huggingface.co/datasets/DataProvenanceInitiative/flan2021_submix_original).

```python
COREFERENCE_TASK_NAMES = {
    'definite_pronoun_resolution:1.1.0',
    'glue/wnli:2.0.0',
    'super_glue/wsc.fixed:1.0.2',
    'winogrande:1.1.0',
}
```

This does not include tasks that are tangentially coreference, e.g. "quoref" tasks in "DataProvenanceInitiative/t0_submix_original" and "qrecc" tasks in "DataProvenanceInitiative/dialog_submix_original".

### Fields

- `inputs`: a `string` feature.
- `targets`: a `string` feature.
- `task_source`: a `string` feature.
- `task_name`: a `string` feature.
- `template_type`: a `string` feature.

## Citation
```
@inproceedings{flan_2022_collection,
    author = {Longpre, Shayne and Hou, Le and Vu, Tu and Webson, Albert and Chung, Hyung Won and Tay, Yi and Zhou, Denny and Le, Quoc V. and Zoph, Barret and Wei, Jason and Roberts, Adam},
    title = {The flan collection: designing data and methods for effective instruction tuning},
    year = {2023},
    publisher = {JMLR.org},
    abstract = {We study the design decisions of publicly available instruction tuning methods, by reproducing and breaking down the development of Flan 2022 (Chung et al., 2022). Through careful ablation studies on the Flan Collection of tasks and methods, we tease apart the effect of design decisions which enable Flan-T5 to outperform prior work by 3-17\%+ across evaluation settings. We find task balancing and enrichment techniques are overlooked but critical to effective instruction tuning, and in particular, training with mixed prompt settings (zero-shot, few-shot, chain-of-thought) actually yields equivalent or stronger (2\%+) performance in all settings. In further experiments, we show Flan-T5 requires less finetuning to converge higher and faster than T5 on single downstream tasks--motivating instruction-tuned models as more computationally-efficient starting checkpoints for new tasks. Finally, to accelerate research on instruction tuning, we make the Flan 2022 collection of datasets, templates, and methods publicly available.},
    booktitle = {Proceedings of the 40th International Conference on Machine Learning},
    articleno = {941},
    numpages = {18},
    location = {Honolulu, Hawaii, USA},
    series = {ICML'23}
}
```