Datasets:
Tasks:
Text Classification
Modalities:
Text
Sub-tasks:
sentiment-classification
Size:
10K - 100K
License:
metadata
pretty_name: IMDB
annotations_creators:
- expert-generated
language_creators:
- expert-generated
language:
- nl
- en
license:
- other
multilinguality:
- multilingual
size_categories:
- 10K<n<100K
source_datasets:
- original
task_categories:
- text-classification
task_ids:
- sentiment-classification
paperswithcode_id: imdb-movie-reviews
train-eval-index:
- config: plain_text
task: text-classification
task_id: binary_classification
splits:
train_split: train
eval_split: test
col_mapping:
text: text
label: target
metrics:
- type: accuracy
- name: Accuracy
- type: f1
name: F1 macro
args:
average: macro
- type: f1
name: F1 micro
args:
average: micro
- type: f1
name: F1 weighted
args:
average: weighted
- type: precision
name: Precision macro
args:
average: macro
- type: precision
name: Precision micro
args:
average: micro
- type: precision
name: Precision weighted
args:
average: weighted
- type: recall
name: Recall macro
args:
average: macro
- type: recall
name: Recall micro
args:
average: micro
- type: recall
name: Recall weighted
args:
average: weighted
dataset_info:
features:
- name: text
dtype: string
- name: text_en
dtype: string
- name: label
dtype:
class_label:
names:
'0': neg
'1': pos
config_name: plain_text
splits:
- name: train
num_bytes: 69589646
num_examples: 24992
- name: test
num_bytes: 67958995
num_examples: 24992
- name: unsupervised
num_examples: 49984
Dataset Card for "imdb"
Table of Contents
- Dataset Description
- Dataset Structure
- Dataset Creation
- Considerations for Using the Data
- Additional Information
Dataset Description
- Homepage: http://ai.stanford.edu/~amaas/data/sentiment/
- Repository: More Information Needed
- Paper: More Information Needed
- Point of Contact: More Information Needed
Dataset Summary
Large Movie Review Dataset translated to Dutch. This is a dataset for binary sentiment classification containing substantially more data than previous benchmark datasets. We provide a set of 25,000 highly polar movie reviews for training, and 25,000 for testing. There is additional unlabeled data for use as well.
Supported Tasks and Leaderboards
Languages
Dataset Structure
Data Instances
plain_text
- Size of downloaded dataset files: 80.23 MB
- Size of the generated dataset: 127.06 MB
- Total amount of disk used: 207.28 MB
An example of 'train' looks as follows.
{
"label": 0,
"text": "Holy shit. Dit was de slechtste film die ik in lange tijd heb gezien."
"text_en": "Holy crap. This was the worst film I have seen in a long time."
}
Data Fields
The data fields are the same among all splits.
plain_text
text
: astring
feature.text_en
: astring
feature.label
: a classification label, with possible values includingneg
(0),pos
(1).
Data Splits
name | train | unsupervised | test |
---|---|---|---|
plain_text | 25000 | 50000 | 25000 |
Dataset Creation
Curation Rationale
Source Data
Initial Data Collection and Normalization
Who are the source language producers?
Annotations
Annotation process
Who are the annotators?
Personal and Sensitive Information
Considerations for Using the Data
Social Impact of Dataset
Discussion of Biases
Other Known Limitations
Additional Information
Dataset Curators
Licensing Information
Citation Information
@InProceedings{maas-EtAl:2011:ACL-HLT2011,
author = {Maas, Andrew L. and Daly, Raymond E. and Pham, Peter T. and Huang, Dan and Ng, Andrew Y. and Potts, Christopher},
title = {Learning Word Vectors for Sentiment Analysis},
booktitle = {Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies},
month = {June},
year = {2011},
address = {Portland, Oregon, USA},
publisher = {Association for Computational Linguistics},
pages = {142--150},
url = {http://www.aclweb.org/anthology/P11-1015}
}
Contributions
Thanks to @ghazi-f, @patrickvonplaten, @lhoestq, @thomwolf for adding
the English imdb
dataset.
This project would not have been possible without compute generously provided by Google through the
TPU Research Cloud.
Created by Yeb Havinga