dataset_name
stringlengths
2
128
description
stringlengths
1
9.7k
prompt
stringlengths
59
185
VFITex
To test interpolation performance on various texture types, we developed a new test set, VFITex, which contains twenty 100-frame UHD or HD videos at 24, 30 or 50 FPS, collected from the Xiph, Mitch Martinez Free 4K Stock Footage, UVG database and pexels.com. This dataset covers diverse textured scenes, including crowds...
Provide a detailed description of the following dataset: VFITex
HONEST
The HONEST dataset is a template-based corpus for testing the hurtfulness of sentence completions in language models (e.g., BERT) in six different languages (English, Italian, French, Portuguese, Romanian, and Spanish). HONEST is composed of 420 instances for each language, which are generated from 28 identity terms (...
Provide a detailed description of the following dataset: HONEST
SID
The See-in-the-Dark (SID) dataset contains 5094 raw short-exposure images, each with a corresponding long-exposure reference image. Images were captured using two cameras: Sony α7SII and Fujifilm X-T2.
Provide a detailed description of the following dataset: SID
ELD
Extreme low-light denoising (ELD) dataset that covers 10 indoor scenes and 4 camera devices from multiple brands (SonyA7S2, NikonD850, CanonEOS70D, CanonEOS700D). It has three levels (800, 1600, 3200) and two low light factors(100, 200) for noisy images, resulting in 240 (3×2×10×4) raw image pairs in total.
Provide a detailed description of the following dataset: ELD
DoodleUINet
Doodle to UI Dataset contains 11 thousand drawings from 16 categories.
Provide a detailed description of the following dataset: DoodleUINet
CLEVR-X
**CLEVR-X** is a dataset that extends the [CLEVR](/dataset/clevr) dataset with natural language explanations in the context of VQA. It consists of 3.6 million natural language explanations for 850k question-image pairs. For each image-question pair in the CLEVR dataset, CLEVR-X contains multiple structured textual ...
Provide a detailed description of the following dataset: CLEVR-X
BirdClef 2020 (Pruned)
Due to the highly variable sample size of the original BirdClef2020 dataset and the issues that it presents with reproducibility, we propose a pruned version of the set, where samples longer than 180s are removed along with classes with fewer than 50 samples. This processing puts it further in line with other complex a...
Provide a detailed description of the following dataset: BirdClef 2020 (Pruned)
Cellcycle Funcat
Hierarchical multi-label classification dataset for functional genomics
Provide a detailed description of the following dataset: Cellcycle Funcat
Derisi Funcat
Hierarchical-multilabel classification dataset for functional genomics
Provide a detailed description of the following dataset: Derisi Funcat
Eisen Funcat
Hierarchical-multilabel classification dataset for functional genomics
Provide a detailed description of the following dataset: Eisen Funcat
Expr Funcat
Hierarchical-multilabel classification dataset for functional genomics
Provide a detailed description of the following dataset: Expr Funcat
Gasch1 Funcat
Hierarchical-multilabel classification dataset for functional genomics
Provide a detailed description of the following dataset: Gasch1 Funcat
Gasch2 Funcat
Hierarchical-multilabel classification dataset for functional genomics
Provide a detailed description of the following dataset: Gasch2 Funcat
Seq Funcat
Hierarchical-multilabel classification dataset for functional genomics
Provide a detailed description of the following dataset: Seq Funcat
Spo Funcat
Hierarchical-multilabel classification dataset for functional genomics
Provide a detailed description of the following dataset: Spo Funcat
IJB-S
[Paper Abstract](http://biometrics.cse.msu.edu/Publications/Face/Kalkaetal_IJBSIARPPAJanusSurveillanceVideoBenchmark_BTAS2018.pdf) We present IJB–S dataset, an open-source IARPA Janus Surveillance Video Benchmark and associated protocols. The dataset consists of images and surveillance video collected from 202 subje...
Provide a detailed description of the following dataset: IJB-S
EVICAN
Deep learning use for quantitative image analysis is exponentially increasing. However, training accurate, widely deployable deep learning algorithms requires a plethora of annotated (ground truth) data. Image collections must contain not only thousands of images to provide sufficient example objects (i.e. cells), but ...
Provide a detailed description of the following dataset: EVICAN
SR-RAW
Raw sensor dataset where each sequence captures 7 (few contain 6) images (RAW and JPG) taken by different focal lengths.
Provide a detailed description of the following dataset: SR-RAW
bladderbatch
Microarray gene expression data on 57 bladder samples from 5 batches.
Provide a detailed description of the following dataset: bladderbatch
HRSOD
There exist several datasets for saliency detection, but none of them is specifically designed for high-resolution salient object detection. High-Resolution Salient Object Detection (HRSOD) dataset, containing 1610 training images and 400 test images. The total 2010 images are collected from the website of Flickr with ...
Provide a detailed description of the following dataset: HRSOD
RASFF
In the actual globalized world, the transportation of goods between any country is something normal. Considering that the protocols in quality and security vary from one country to another, there is a risk with the products that do not comply with the legislation of a country cross the border. In the case of edible pro...
Provide a detailed description of the following dataset: RASFF
DAVIS-585
A dataset for interactive segmentation with simulated initial masks.
Provide a detailed description of the following dataset: DAVIS-585
CNN Filter DB-Robust
Dataset for the Paper "Adversarial Robustness through the Lens of Convolutional Filters".
Provide a detailed description of the following dataset: CNN Filter DB-Robust
CVL-DataBase
The CVL Database is a public database for writer retrieval, writer identification and word spotting. The database consists of 7 different handwritten texts (1 German and 6 Englisch Texts). In total 310 writers participated in the dataset. 27 of which wrote 7 texts and 283 writers had to write 5 texts. For each text a r...
Provide a detailed description of the following dataset: CVL-DataBase
Gait3D
Gait3D is a large-scale 3D representation-based gait recognition dataset. It contains 4,000 subjects and over 25,000 sequences extracted from 39 cameras in an unconstrained indoor scene.
Provide a detailed description of the following dataset: Gait3D
Korean UnSmile Dataset
1.9K Korean Online Hate Speech Comments for Multilabel Classification (Annotated by Three Independent Labelers per Data)
Provide a detailed description of the following dataset: Korean UnSmile Dataset
HateScore
2.2K neutral sentences from Wikipedia 1.7K additionally labeled sentences generated by the Human-in-the-Loop procedure (based on Korean Unsmile Dataset Base Model) 7.1K rule-generated neutral sentences
Provide a detailed description of the following dataset: HateScore
The Little Prince
This corpus is an annotation of the novel The Little Prince by Antoine de Saint-Exupéry, published in 1943. We were inspired by the UNL project to include this novel, so that different groups could compare representations on the same text.
Provide a detailed description of the following dataset: The Little Prince
Bio
This corpus includes annotations of cancer-related PubMed articles, covering 3 full papers (PMID:24651010, PMID:11777939, PMID:15630473) as well as the result sections of 46 additional PubMed papers. The corpus also includes about 1000 sentences each from the BEL BioCreative training corpus and the Chicago Corpus.
Provide a detailed description of the following dataset: Bio
New3
New3, a set of 527 instances from AMR 3.0, whose original source was the LORELEI DARPA project – not included in the AMR 2.0 training set – consisting of excerpts from newswires and online forum.
Provide a detailed description of the following dataset: New3
RepCount
Counting repetitive actions are widely seen in human activities such as physical exercise. Existing methods focus on performing repetitive action counting in short videos, which is tough for dealing with longer videos in more realistic scenarios. In the data-driven era, the degradation of such generalization capability...
Provide a detailed description of the following dataset: RepCount
Multispectral Image Database
We present a database of multispectral images that were used to emulate the GAP camera. The images are of a wide variety of real-world materials and objects. We are making this database available to the research community. Details of the database can be found in the following publication: "Generalized Assorted Pixel...
Provide a detailed description of the following dataset: Multispectral Image Database
RS-Haze
A large-scale non-homogeneous remote sensing image dehazing dataset
Provide a detailed description of the following dataset: RS-Haze
CP2A dataset
We present a new simulated dataset for pedestrian action anticipation collected using the CARLA simulator. To generate this dataset, we place a camera sensor on the ego-vehicle in the Carla environment and set the parameters to those of the camera used to record the PIE dataset (i.e., 1920x1080, 110° FOV). Then, we co...
Provide a detailed description of the following dataset: CP2A dataset
Digits-Five
Digits-Five is a collection of five most popular digit datasets, MNIST (mt) (55000 samples), MNIST-M (mm) (55000 samples), Synthetic Digits (syn) (25000 samples), SVHN (sv)(73257 samples), and USPS (up) (7438 samples). Each digit dataset includes a different style of 0-9 digit images.
Provide a detailed description of the following dataset: Digits-Five
Amazon Review
Amazon Review is a dataset to tackle the task of identifying whether the sentiment of a product review is positive or negative. This dataset includes reviews from four different merchandise categories: Books (B) (2834 samples), DVDs (D) (1199 samples), Electronics (E) (1883 samples), and Kitchen and housewares (K) (175...
Provide a detailed description of the following dataset: Amazon Review
DIBCO 2019
DIBCO 2019 is the international Competition on Document Image Binarization organized in conjunction with the ICDAR 2019 conference. The general objective of the contest is to identify current advances in document image binarization of machine-printed and handwritten document images using performance evaluation measures...
Provide a detailed description of the following dataset: DIBCO 2019
OSS for Social Good Project List
# Leaving My Fingerprints: Motivations and Challenges of Contributing to OSS for Social Good -> ICSE 2021 <- ### Authors + Yu Huang + Denae Ford + Thomas Zimmermann ### Abstract When inspiring software developers to contribute to open source software, the act is often referenced as an opportunity to b...
Provide a detailed description of the following dataset: OSS for Social Good Project List
Chart-to-text
**Chart-to-text** is a large-scale benchmark with two datasets and a total of 44,096 charts covering a wide range of topics and chart types.
Provide a detailed description of the following dataset: Chart-to-text
RELiC
**RELiC** is a large-scale dataset of 79k excerpts of literary scholarship, each containing a quotation from a primary source and the surrounding critical analysis. 79 public domain primary sources and over 8,836 secondary sources are represented in RELiC.
Provide a detailed description of the following dataset: RELiC
PhysioNet Challenge 2021
# Data Description The training data contains twelve-lead ECGs. The validation and test data contains twelve-lead, six-lead, four-lead, three-lead, and two-lead ECGs: 1. Twelve leads: I, II, III, aVR, aVL, aVF, V1, V2, V3, V4, V5, V6 2. Six leads: I, II, III, aVR, aVL, aVF 3. Four leads: I, II, III, V2 4. Three ...
Provide a detailed description of the following dataset: PhysioNet Challenge 2021
LUDB
# Abstract Lobachevsky University Electrocardiography Database (LUDB) is an ECG signal database with marked boundaries and peaks of P, T waves and QRS complexes. The database consists of 200 10-second 12-lead ECG signal records representing different morphologies of the ECG signal. The ECGs were collected from healthy...
Provide a detailed description of the following dataset: LUDB
CPSC2019
# Introduction The China Physiological Signal Challenge 2019 (CPSC 2019) aims to encourage the development of algorithms for challenging QRS detection and heart rate (HR) estimation from short-term single-lead ECG recordings usually with low signal quality and/or abnormal rhythm waveforms. ECG signal provides an ...
Provide a detailed description of the following dataset: CPSC2019
CPSC2020
# Introduction Abnormality of cardiac conduction system can induce arrhythmia. Abnormal heart rhythm can lead to other cardiac diseases and complications, and can be life-threatening [1]. There are various types of arrhythmias and each type is associated with a pattern, and as such, it is possible to be identified. ...
Provide a detailed description of the following dataset: CPSC2020
CPSC2021
# Introduction The 4th China Physiological Signal Challenge 2021 (CPSC 2021) aims to encourage the development of algorithms for searching the paroxysmal atrial fibrillation (PAF) events from dynamic ECG recordings. ECG signal provides an important role in non-invasively monitoring and clinical diagnosis for cardio...
Provide a detailed description of the following dataset: CPSC2021
AnthroProtect
For a detailed description, we refer to Section 3 in our research article.
Provide a detailed description of the following dataset: AnthroProtect
SSD_PHONE
SSD (Sub-slot Dialog) dataset: This is the dataset for the ACL 2022 paper "A Slot Is Not Built in One Utterance: Spoken Language Dialogs with Sub-Slots".
Provide a detailed description of the following dataset: SSD_PHONE
SSD_ID
SSD (Sub-slot Dialog) dataset: This is the dataset for the ACL 2022 paper "A Slot Is Not Built in One Utterance: Spoken Language Dialogs with Sub-Slots".
Provide a detailed description of the following dataset: SSD_ID
SSD_NAME
SSD (Sub-slot Dialog) dataset: This is the dataset for the ACL 2022 paper "A Slot Is Not Built in One Utterance: Spoken Language Dialogs with Sub-Slots".
Provide a detailed description of the following dataset: SSD_NAME
SSD_PLATE
SSD (Sub-slot Dialog) dataset: This is the dataset for the ACL 2022 paper "A Slot Is Not Built in One Utterance: Spoken Language Dialogs with Sub-Slots".
Provide a detailed description of the following dataset: SSD_PLATE
TUSC
Tweets from US and Canada (TUSC) is a large dataset of more than 45 million geo-located tweets posted between 2015 and 2021 from US and Canada (TUSC), especially curated for natural language analysis
Provide a detailed description of the following dataset: TUSC
SF-XL
Large scale dataset for visual geo-localization / visual place recognition. It provides images from the city of San Francisco, labeled with GPS coordinates and heading.
Provide a detailed description of the following dataset: SF-XL
SF-XL test v1
Test set version 1 for the San Francisco eXtra Large dataset
Provide a detailed description of the following dataset: SF-XL test v1
SF-XL test v2
Test set version 2 for the San Francisco eXtra Large dataset
Provide a detailed description of the following dataset: SF-XL test v2
AmsterTime
**AmsterTime** dataset offers a collection of 2,500 well-curated images matching the same scene from a street view matched to historical archival image data from Amsterdam city. The image pairs capture the same place with different cameras, viewpoints, and appearances. Unlike existing benchmark datasets, AmsterTime is ...
Provide a detailed description of the following dataset: AmsterTime
ChAII - Hindi and Tamil Question Answering
The dataset covers Hindi and Tamil, collected without the use of translation. It provides a realistic information-seeking task with questions written by native-speaking expert data annotators.
Provide a detailed description of the following dataset: ChAII - Hindi and Tamil Question Answering
MC_GRID
Here we release the dataset (Multi_Channel_Grid, abbreviated as **MC_Grid**) used in our paper [LIMUSE: LIGHTWEIGHT MULTI-MODAL SPEAKER EXTRACTION]([[2111.04063\] LiMuSE: Lightweight Multi-modal Speaker Extraction (arxiv.org)](https://arxiv.org/abs/2111.04063)). MC_Grid, which is based on [GRID](http://spandh.dcs.sh...
Provide a detailed description of the following dataset: MC_GRID
PDNC
A annotated dataset of quotations and within-quotation-mentions in 22 full-length English novels.
Provide a detailed description of the following dataset: PDNC
Korean Hate Speech Evaluation Datasets
APEACH is the first crowd-generated Korean evaluation dataset for hate speech detection. Sentences of the dataset are created by anonymous participants using an online crowdsourcing platform DeepNatural AI.
Provide a detailed description of the following dataset: Korean Hate Speech Evaluation Datasets
Forest CoverType
Predicting forest cover type from cartographic variables only (no remotely sensed data). The actual forest cover type for a given observation (30 x 30 meter cell) was determined from US Forest Service (USFS) Region 2 Resource Information System (RIS) data. Independent variables were derived from data originally obtaine...
Provide a detailed description of the following dataset: Forest CoverType
Casino Reviews
This dataset contain online reviews gathered from google reviews written by north american casino users. explain motivations and summary of its content. Can be used to study user experience and relative research directions such as cultural impacts on latency of aspects, domain importance, sentiment analysis, opinio...
Provide a detailed description of the following dataset: Casino Reviews
WHAMR_ext
WHAMR_ext is an extension to the WHAMR corpus with larger RT60 values (between 1s and 3s)
Provide a detailed description of the following dataset: WHAMR_ext
60k Stack Overflow Questions
The dataset contains 60,000 Stack Overflow questions from 2016-2020, classified into three categories: 1. HQ: High-quality posts without a single edit. 2. LQ_EDIT: Low-quality posts with a negative score, and multiple community edits. However, they still remain open after those changes. 3. LQ_CLOSE: Low-quality p...
Provide a detailed description of the following dataset: 60k Stack Overflow Questions
Wiki-ZSL
The Wiki-ZSL (Wiki Zero-Shot Learning) dataset contains 113 relations and 94,383 instances from Wikipedia. The dataset is divided into three subsets: training set (98 relations), validation set (5 relations) and test set (10 relations).
Provide a detailed description of the following dataset: Wiki-ZSL
Tweet IDs - Academic API Experiments
The lists of Tweet IDs for the experiments of the article: This Sample seems to be good enough! Assessing Coverage and Temporal Reliability of Twitter's Academic API by Juergen Pfeffer, Angelina Mooseder, Luca Hammer, Oliver Stritzel, David Garcia.
Provide a detailed description of the following dataset: Tweet IDs - Academic API Experiments
Anshita
Potential cases
Provide a detailed description of the following dataset: Anshita
KITTI-360
KITTI-360 is a large-scale dataset that contains rich sensory information and full annotations. It is the successor of the popular KITTI dataset, providing more comprehensive semantic/instance labels in 2D and 3D, richer 360 degree sensory information (fisheye images and pushbroom laser scans), very accurate and geo-l...
Provide a detailed description of the following dataset: KITTI-360
BDD100K-Subsets
Subsets of BDD100K Dataset that are used in Object Detection Under Rainy Conditions for Autonomous Vehicles: A Review of State-of-the-Art and Emerging Techniques
Provide a detailed description of the following dataset: BDD100K-Subsets
CholecT45
CholecT45 is a subset of CholecT50 consisting of 45 videos from the Cholec80 dataset. It is the first public release of part of CholecT50 dataset. CholecT50 is a dataset of 50 endoscopic videos of laparoscopic cholecystectomy surgery introduced to enable research on fine-grained action recognition in laparoscopic sur...
Provide a detailed description of the following dataset: CholecT45
tdcommons
Therapeutics Data Commons is an open-science initiative with AI/ML-ready datasets and AI/ML tasks for therapeutics, spanning the discovery and development of safe and effective medicines. TDC provides an ecosystem of tools, libraries, leaderboards, and community resources, including data functions, strategies for syste...
Provide a detailed description of the following dataset: tdcommons
Amazon Men
This datasets is a subset of the Amazon reviews dataset which contain Men related products
Provide a detailed description of the following dataset: Amazon Men
Amazon Fashion
This datasets is a subset of the Amazon reviews dataset which contain Fashion related products
Provide a detailed description of the following dataset: Amazon Fashion
TCR
A dataset of Joint Reasoning for Temporal and Causal Relations
Provide a detailed description of the following dataset: TCR
Four Shapes
This dataset contains 16,000 images of four shapes; square, star, circle, and triangle. Each image is 200x200 pixels.
Provide a detailed description of the following dataset: Four Shapes
RainCityscapes
A dataset for rain removal with scene depth information. Compared with previous datasets, this dataset are all outdoor photos, each with a depth map, and the rain images exhibit different degrees of rain and fog.
Provide a detailed description of the following dataset: RainCityscapes
OUMVLP-Pose
The OU-ISIR Gait Database, Multi-View Large Population Database with Pose Sequence (OUMVLP-Pose) is meant to aid research efforts in the general area of developing, testing and evaluating algorithms for model-based gait recognition. This data set was built upon OU-MVLP. It contains 10,307 subjects of round-trip wal...
Provide a detailed description of the following dataset: OUMVLP-Pose
PhC-C2DH-U373
Glioblastoma-astrocytoma U373 cells on a polyacrylamide substrate Dr. S. Kumar. Department of Bioengineering, University of California at Berkeley, Berkeley CA (USA)
Provide a detailed description of the following dataset: PhC-C2DH-U373
DIC-C2DH-HeLa
HeLa cells on a flat glass Dr. G. van Cappellen. Erasmus Medical Center, Rotterdam, The Netherlands
Provide a detailed description of the following dataset: DIC-C2DH-HeLa
Fluo-N2DH-SIM+
Simulated nuclei of HL60 cells stained with Hoescht Dr. V. Ulman and Dr. D. Svoboda. Centre for Biomedical Image Analysis (CBIA), Masaryk University, Brno, Czech Republic (Created using MitoGen, part of Cytopacq)
Provide a detailed description of the following dataset: Fluo-N2DH-SIM+
Fluo-N2DH-GOWT1
GFP-GOWT1 mouse stem cells Dr. E. Bártová. Institute of Biophysics, Academy of Sciences of the Czech Republic, Brno, Czech Republic
Provide a detailed description of the following dataset: Fluo-N2DH-GOWT1
Fluo-N2DL-HeLa
HeLa cells stably expressing H2b-GFP Mitocheck Consortium
Provide a detailed description of the following dataset: Fluo-N2DL-HeLa
Fluo-N3DL-TRIC
Developing Tribolium Castaneum embryo (3D cartographic projection) Dr. A. Jain. Max Planck Institute of Molecular Cell Biology and Genetics, Dresden, Germany
Provide a detailed description of the following dataset: Fluo-N3DL-TRIC
Fluo-C3DH-A549-SIM
Simulated GFP-actin-stained A549 Lung Cancer cells embedded in a Matrigel matrix Dr. M. Maška and Dr. D. V. Sorokin. Centre for Biomedical Image Analysis (CBIA), Masaryk University, Brno, Czech Republic (Created using FiloGen, part of Cytopacq)
Provide a detailed description of the following dataset: Fluo-C3DH-A549-SIM
Fluo-C3DL-MDA231
MDA231 human breast carcinoma cells infected with a pMSCV vector including the GFP sequence, embedded in a collagen matrix Dr. R. Kamm. Dept. of Biological Engineering, Massachusetts Institute of Technology, Cambridge MA (USA)
Provide a detailed description of the following dataset: Fluo-C3DL-MDA231
Dataset: Relationship extraction for knowledge graph creation from biomedical literature (Gene-Disease relationships)
This is the dataset used for classifying Gene-Disease relationship types from sentences. The dataset consists of 3 files: * manually_annotated_set.xlsx - set of 2000 manualy annotated sentences with entities * Unbalanced_dataset.xlsx - set of 12000 sentences, out of which 2000 are from the first set, manually annot...
Provide a detailed description of the following dataset: Dataset: Relationship extraction for knowledge graph creation from biomedical literature (Gene-Disease relationships)
VISUELLE2.0
Visuelle 2.0 is a dataset containing real data for 5355 clothing products of the retail fast-fashion Italian company, Nuna Lie. Specifically, Visuelle 2.0 provides data from 6 fashion seasons (partitioned in Autumn-Winter and Spring-Summer) from 2017-2019, right before the Covid-19 pandemic. Each product is accompanied...
Provide a detailed description of the following dataset: VISUELLE2.0
CB-ToF
# Cornell-Box Dataset ## Download The CornellBox Dataset can be downloaded from this URL >https://viscom.datasets.uni-ulm.de/radu/dataset.zip ## Dataset The dataset contains correlation measurements, ToF depth images and ground truth depth images in `.hdr` format. The script `simulate_noise_on_corre...
Provide a detailed description of the following dataset: CB-ToF
FLAT
FLAT, a synthetic dataset of 2000 ToF measurements that capture all of these nonidealities, and can be used to simulate different hardware
Provide a detailed description of the following dataset: FLAT
DeePore
DeePore is a deep learning workflow for rapid estimation of a wide range of porous material properties based on the binarized micro–tomography images. By combining naturally occurring porous textures we generated 17,700 semi–real 3–D micro–structures of porous geo–materials with the size of $256^3$ voxels and 30 physic...
Provide a detailed description of the following dataset: DeePore
MTHS
the MTHS dataset contains 30Hz PPG signals obtained from 62 patients, including 35 men and 27 women. The ground truth data includes heart rate and oxygen saturation levels sampled at 1Hz. The HR and SPo2 measurement is obtained using a pulse oximeter (M70). An iPhone 5s was used to obtain the ppg recordings at 30 f...
Provide a detailed description of the following dataset: MTHS
MUGEN
**MUGEN** is a large-scale video-audio-text dataset MUGEN, collected using the open-sourced platform game CoinRun. MUGEN can help progress research in many tasks in multimodal understanding and generation.
Provide a detailed description of the following dataset: MUGEN
Distinctions-646
Dinstinctions-646 are composed of 646 foreground images with manually annotated alpha mattes
Provide a detailed description of the following dataset: Distinctions-646
SMC Text Corpus
Contents (As on March 4, 2019) -------- The text corpus contains running text from various free licensed sources. - The whole content of Malayalam Wikipedia extracted on January 1, 2019 - News/Article from various sources, source mentioned in respective files: - 251 Mb - 8,60,159 lines - 98,15,533 words - 10,1...
Provide a detailed description of the following dataset: SMC Text Corpus
SILVR
We present _SILVR_, a dataset of light field images for six-degrees-of-freedom navigation in large fully-immersive volumes. The _SILVR_ dataset is short for _"**S**ynthetic **I**mmersive **L**arge-**V**olume **R**ay"_ dataset. ## Properties Our dataset exhibits the following properties: - **synthetic**: Rende...
Provide a detailed description of the following dataset: SILVR
MSU Super-Resolution for Video Compression
This is a dataset for a super-resolution task. The dataset contains 480x270 videos that were decoded with 6 different bitrates (100 - 4000 kbps) using 5 different codecs (H.264, H.265, H.266, AV1, and AVS3 standards). The dataset contains indoor and outdoor videos as well as animation. All videos have low SI/TI values ...
Provide a detailed description of the following dataset: MSU Super-Resolution for Video Compression
MASSIVE
MASSIVE is a parallel dataset of > 1M utterances across 51 languages with annotations for the Natural Language Understanding tasks of intent prediction and slot annotation. Utterances span 60 intents and include 55 slot types. MASSIVE was created by localizing the SLURP dataset, composed of general Intelligent Voice As...
Provide a detailed description of the following dataset: MASSIVE
MICCAI 2015 Head and Neck Challenge
This database is provided and maintained by Dr. Gregory C Sharp (Harvard Medical School – MGH, Boston) and his group. The data here provided have been used for the “Head and Neck Auto Segmentation MICCAI Challenge (2015)”. To cite the challenge or the data, please refer to: Raudaschl, P. F., Zaffino, P., Sharp, G....
Provide a detailed description of the following dataset: MICCAI 2015 Head and Neck Challenge
NR2R
To form the collection of nighttime RAW samples, we first selected a total of 150 images with the spatial resolution at 3464×5202 from the training and validation sets provided by the night image challenge. And then these RAW images are pre-processed to best produce noise-free samples using a notable CNN based denoiser...
Provide a detailed description of the following dataset: NR2R
160_subset
the 160x160 subset of the GasHisSDB dataset.
Provide a detailed description of the following dataset: 160_subset
Basketball Ballistic raw sequences
Ballistic trajectories
Provide a detailed description of the following dataset: Basketball Ballistic raw sequences