climate_fever_fixed / README.md
dcisek93's picture
updated dataset card
5e7d270
|
raw
history blame
No virus
4.23 kB
---
dataset_info:
features:
- name: claim_id
dtype: int64
- name: claim
dtype: string
- name: evidence
dtype: string
- name: label
dtype:
class_label:
names:
'0': SUPPORTS
'1': REFUTES
'2': NOT_ENOUGH_INFO
- name: category
dtype: string
splits:
- name: train
num_bytes: 1467456
num_examples: 4298
- name: test
num_bytes: 526276
num_examples: 1535
- name: valid
num_bytes: 635174
num_examples: 1842
download_size: 1372892
dataset_size: 2628906
license: mit
task_categories:
- text-classification
language:
- en
tags:
- climate
pretty_name: climate_fever dataset with one-to-one claim-evidence pair
size_categories:
- 1K<n<10K
---
# Dataset Card for "climate_fever_fixed"
### Dataset Summary
This dataset was created to aid our team in developing a model to more accurately perform climate change-related fact checking. We approach this task from a perspective heavily impacted
by the work of the [ClimateBERT](https://climatebert.ai/about) team. With that in mind, our team likewise leveraged a BERT Language model to solve this task. This dataset presents an
edited version of the [Climate_Fever](https://huggingface.co/datasets/climate_fever) dataset, hosted by HuggingFace. Climate_Fever is composed of climate-related documents
that have been annotated with labels related to fact-checking and misinformation. However, in the climate-plus project, we decided to modify the dataset to remove redundancy
and keep only the essentials of a text-entailment problem: claim as the premise and evidence as the hypothesis.
### Data Fields
This dataset contains 7675 records, each of which is composed of several attributes:
- `claim_id`: a `integer` feature, which serves as a unique identifier for each record/row.
- `claim`: a `string` feature, containes the raw text of a given climate-related claim.
- `evidence`: a `string` feature, which provides free text evidence that relates to the previously established claim.
- `label`: a `class label` feature representing an assigned class, where values can either be 0: "supports", 1: "refutes" and 2: "not enough info".
- `category`: a `string` feature, which provides additional detail about the particular focus of a given claim.
<br>
This dataset was then broken into train, test and validation sets to enable proper evaluation of our model. These splits contain the following amount of data:
- `Train`: 4300 Records
- `Test`: 1540 Records
- `Val`: 1840 Records
### Source Data
This dataset represents an evolved version of the original [Climate_Fever](https://huggingface.co/datasets/climate_fever) dataset, hosted by HuggingFace. It was adapted to meet
the needs of our team, as we attempted to solve a specific climate change-related task. The original dataset adopted the FEVER methodology, discussed in more detail [here](https://www.amazon.science/blog/the-fever-data-set-what-doesnt-kill-it-will-make-it-stronger).
Their original dataset consists of 1,535 real-world claims regarding climate-change collected on the internet. Each claim is accompanied by five manually annotated evidence
sentences retrieved from the English Wikipedia that support, refute or do not give enough information to validate the claim totalling in 7,675 claim-evidence pairs.
### Methodology
This dataset was curated by our team to reduce redundancy and keep only the essentials of a text-entailment problem: claim as the premise and evidence as the hypothesis.
For each given claim, there are multiple sentences of evidence. We decided to expand the one-to-many relation to one-to-one.
This resulted in a modified version of the climate_fever dataset that includes only one evidence sentence per claim.
### Languages
The text contained in the dataset is entirely in English, as found in the real-world financial disclosures identified by the TCFD. The associated BCP-47 code is [`en`](https://www.techonthenet.com/js/language_tags.php), to ensure clear labeling of language usage for downstream tasks and other future applications.
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)