Quantitative101 / README.txt
NLPFin's picture
Upload 20 files
ed1c082
raw
history blame
No virus
2.41 kB
Quantitative 101 dataset, which is the combination of one generated dataset (CND) and three benchmark datasets: Numeracy-600K [1], EQUATE [2], and NumGLUE Task 3 [3]. The tasks in Quantitative 101 include Comparing Numbers (ComNum), Quantitative Prediction (QP), Quantitative Natural Language Inference (QNLI), and Quantitative Question Answering (QQA).
The details of how to separate the dataset are shown in this document.
(1) Task: ComNum
There are two JSON files in the ComNum folder. "CND-OOR.json" is used for testing the phenomenon. "CND-IR.json" can be separated into training and test sets with the following code:
import pandas as pd
train_size = 0.8
train_dataset=new_df.sample(frac=train_size,random_state=200)
test_dataset=new_df.drop(train_dataset.index).reset_index(drop=True)
train_dataset = train_dataset.reset_index(drop=True)
(2) Task: QP
In the QP folder, we already separated Numeracy-600K into training, development, and test sets. Note that, the original Numeracy-600K [1] did not provide such information.
(3) Task: QNLI
EQUATE has five subsets collected from different sources, including RTE-QUANT, AWP-NLI, NEWSNLI, REDDITNLI, and Stress Test.
For Stress Test, which contains 7,500 instances, we follow the splitting method in NumGLUE Task 7 to separate it into training, development, and test sets. All sets are in the "QNLI-Stress Test" folder.
Because other subsets are less than 1,000 instances, we perform the 10-fold cross-validation in the experiments. Please use the following code to separate folds:
from sklearn.model_selection import KFold
kf = KFold(n_splits=10,random_state=200)
(4) Task: QQA
We follow [3] to separate the dataset into training, development, and test sets.
Reference:
[1] Chen, Chung-Chi, et al. "Numeracy-600K: Learning numeracy for detecting exaggerated information in market comments." Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. 2019.
[2] Ravichander, Abhilasha, et al. "EQUATE: A Benchmark Evaluation Framework for Quantitative Reasoning in Natural Language Inference." Proceedings of the 23rd Conference on Computational Natural Language Learning (CoNLL). 2019.
[3] Mishra, Swaroop, et al. "NumGLUE: A Suite of Fundamental yet Challenging Mathematical Reasoning Tasks." Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics. 2022