File size: 7,359 Bytes
ae73f56
 
38c0057
ae73f56
587ac79
afd53e3
 
 
 
 
 
 
 
186e3f8
587ac79
 
 
 
 
 
 
 
 
 
 
f8b0c8b
6720af7
f8b0c8b
6720af7
 
587ac79
c1550f9
 
 
 
afc7fbf
a73fe7b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
c1550f9
 
 
 
afc7fbf
a73fe7b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
c1550f9
 
 
 
afc7fbf
a73fe7b
 
 
 
 
 
 
 
 
 
 
 
 
 
c1550f9
 
afd53e3
587ac79
 
 
 
 
 
 
afd53e3
587ac79
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
f8b0c8b
6720af7
f8b0c8b
6720af7
f8b0c8b
587ac79
 
 
 
 
6720af7
587ac79
 
 
6720af7
587ac79
afd53e3
587ac79
afd53e3
587ac79
afd53e3
587ac79
afd53e3
f8b0c8b
9556685
6720af7
 
 
 
 
f8b0c8b
6720af7
afd53e3
f8b0c8b
 
6720af7
 
 
 
 
 
12cc1c1
f8b0c8b
 
 
 
 
6720af7
afd53e3
9556685
12cc1c1
 
 
 
 
 
9556685
 
 
f9a2ff5
9556685
 
 
 
 
 
 
 
 
 
 
6720af7
9556685
587ac79
12cc1c1
afd53e3
587ac79
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
afd53e3
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
---
license: unknown
viewer: false
---

**==========================================**

**_IN PROGRESS - NOT READY FOR LOADING OR USE_**

**==========================================**

---

# Dataset Card for climate-fever-nli-stsb

## Dataset Description

- **Homepage:** 
- **Repository:** 
- **Paper:** 
- **Leaderboard:** 
- **Point of Contact:** 

### Dataset Summary

The CLIMATE-FEVER dataset modified to supply NLI-style (**cf-nli**) features or STSb-style (**cf-stsb**) features that SentenceBERT training scripts can use as drop-in replacements for AllNLI and/or STSb datasets.

There are two **cf-nli** datasets: one derived from only SUPPORTS and REFUTES evidence (**cf-nli**), and one that also derived data from NOT_ENOUGH_INFO evidence based on the annotator votes (**cf-nli-nei**).

The feature style is specified as a named configuration when loading the dataset: cf-nli, cf-nli-nei, or cf-stsb. See usage notes below for `load_dataset` examples.

### Usage

Load the **cf-nli** dataset

```python
# if datasets not already in your environment
!pip install datasets

from datasets import load_dataset

# all splits...
dd = load_dataset('climate-fever-nli-stsb', 'cf-nli')

# ... or specific split (only 'train' is available)
ds_train = load_dataset('climate-fever-nli-stsb', 'cf-nli', split='train')

## ds_train can now be injected into SentenceBERT training scripts at the point
## where individual sentence pairs are aggregated into
## {'claim': {'entailment': set(), 'contradiction': set(), 'neutral': set()}} dicts
## for further processing into training samples
```

Load the **cf-nli-nei** dataset

```python
# if datasets not already in your environment
!pip install datasets

from datasets import load_dataset

# all splits...
dd = load_dataset('climate-fever-nli-stsb', 'cf-nli-nei')

# ... or specific split (only 'train' is available)
ds_train = load_dataset('climate-fever-nli-stsb', 'cf-nli-nei', split='train')

## ds_train can now be injected into SentenceBERT training scripts at the point
## where individual sentence pairs are aggregated into
## {'claim': {'entailment': set(), 'contradiction': set(), 'neutral': set()}} dicts
## for further processing into training samples
```

Load the **cf-stsb** dataset

```python
# if datasets not already in your environment
!pip install datasets

from datasets import load_dataset

# all splits...
dd = load_dataset('climate-fever-nli-stsb', 'cf-stsb')

# ... or specific split ('train', 'dev', 'test' available)
ds_dev = load_dataset('climate-fever-nli-stsb', 'cf-stsb', split='dev')

## ds_dev (or test) can now be injected into SentenceBERT training scripts at the point
## where individual sentence pairs are aggregated into
## a list of dev (or test) samples
```

<!--
### Supported Tasks and Leaderboards

[More Information Needed]

### Languages

[More Information Needed]
-->

## Dataset Structure

### Data Instances

[More Information Needed]

### Data Fields

[More Information Needed]

### Data Splits

[More Information Needed]

## Dataset Creation

### Curation Rationale

SentenceBERT models are designed for 'Domain Adaptation' and/or 'Fine-tuning' using labeled data in the downstream task domain. As a bi-encoder, the primary objective function is real-valued similarity scoring. Typical training datasets use NLI-style features as input, and STSb-style features as model evaluation during training, and to measure post-hoc, _intrinsic_ STSb performance. Classification tasks typically use a classifier network that accepts SentenceBERT encodings as input, and is trained on class-labeled datasets.

So, to fine-tune a SentenceBERT model in a climate-change domain, a labeled climate change dataset would be ideal. Much like the authors of the CLIMATE-FEVER dataset, we know of no other _labeled_ datasets specific to climate change. And while CLIMATE-FEVER is suitably labeled for classification tasks, it is not ready for similarity tuning in the style of SentenceBERT.

This modified CLIMATE-FEVER dataset attempts to fill that gap by deriving NLI-style features typically used in pre-training and fine-tuning a SentenceBERT model. SentenceBERT also uses STSb-style features to evaluate model performance both during training and after training to gauge _intrinsic_ model performance on STSb.

### Source Data

#### Initial Data Collection and Normalization

see CLIMATE-FEVER

#### Who are the source language producers?

see CLIMATE-FEVER

<!--
### Annotations
-->

### Annotation process

#### **cf-nli**

For each Claim that has both SUPPORTS evidence and REFUTES evidence, create labeled pairs in the style of NLI dataset:

| split | dataset | sentence1 | sentence2 | label |
|---|---|---|---|---|
| {'train', 'test'} | 'climate-fever' | claim | evidence | evidence_label SUPPORTS -> 'entailment', REFUTES -> 'contradiction' |

> Note that by defintion, only claims classified as DISPUTED include both SUPPORTS and REFUTES evidence, so this dataset is limited to a small subset of CLIMATE-FEVER.

### **cf-nli-nei**

This dataset uses the list of annotator 'votes' to cast a NOT_ENOUGH_INFO (NEI) evidence to a SUPPORTS or REFUTES evidence. By doing so, Claims in the SUPPORTS, REFUTES, and NEI classes can be used to generate additional sentence pairs.

| votes | effective evidence_label |
|---|---|
| SUPPORTS > REFUTES | _SUPPORTS_ |
| SUPPORTS < REFUTES | _REFUTES_ |

In addition to all the claims in **cf-nli**, any Claims that have,

* **_at least one_** SUPPORTS or REFUTES evidence, AND
* NEI evidences that can be cast to effective _SUPPORTS_ or _REFUTES_

are included in the datasset.

### **cf-stsb**

For each Claim <-> Evidence pair, create labeled pairs in the style of STSb dataset:

| split | dataset | score | sentence1 | sentence2 |
|---|---|---|---|---|
| {'train', 'dev', 'test'} | 'climate-fever' | cos_sim score | claim | evidence |

This dataset uses 'evidence_label', vote 'entropy', and the list of annotator 'votes' to derive a similarity score for each claim <-> evidence pairing. Similarity score conversion:

> `mean(entropy)` refers to the average entropy within the defined group of evidence

| evidence_label | votes | similarity score |
|---|---|---|
| SUPPORTS | SUPPORTS > 0, REFUTES == 0, NOT_ENOUGH_INFO (NEI) == 0 | 1 |
| | SUPPORTS > 0, REFUTES == 0 | mean(entropy) |
| | SUPPORTS > 0, REFUTES > 0 | 1 - mean(entropy) |
| NEI | SUPPORTS > REFUTES | (1 - mean(entropy)) / 2|
| | SUPPORTS == REFUTES | 0 |
| | SUPPORTS < REFUTES | -(1 - mean(entropy)) / 2 |
| REFUTES | SUPPORTS > 0, REFUTES > 0 | -(1 - mean(entropy)) |
| | SUPPORTS == 0, REFUTES > 0 | -mean(entropy) |
| | SUPPORTS == 0, REFUTES > 0, NEI == 0 | -1 |

The above derivation roughly maps the strength of evidence annotation (REFUTES..NEI..SUPPORTS) to cosine similarity (-1..0..1).


<!--
#### Who are the annotators?

[More Information Needed]

### Personal and Sensitive Information

[More Information Needed]

## Considerations for Using the Data

### Social Impact of Dataset

[More Information Needed]

### Discussion of Biases

[More Information Needed]

### Other Known Limitations

[More Information Needed]

## Additional Information

### Dataset Curators

[More Information Needed]

### Licensing Information

[More Information Needed]

### Citation Information

[More Information Needed]

### Contributions

[More Information Needed]
-->