text
stringlengths
493
2.1k
inputs
dict
prediction
null
prediction_agent
null
annotation
stringclasses
2 values
annotation_agent
stringclasses
1 value
vectors
null
multi_label
bool
1 class
explanation
null
id
stringlengths
36
36
metadata
null
status
stringclasses
2 values
event_timestamp
stringlengths
26
26
metrics
dict
label
class label
2 classes
TITLE: Amplifying Pathological Detection in EEG Signaling Pathways through Cross-Dataset Transfer Learning ABSTRACT: Pathology diagnosis based on EEG signals and decoding brain activity holds immense importance in understanding neurological disorders. With the advancement of artificial intelligence methods and machine learning techniques, the potential for accurate data-driven diagnoses and effective treatments has grown significantly. However, applying machine learning algorithms to real-world datasets presents diverse challenges at multiple levels. The scarcity of labelled data, especially in low regime scenarios with limited availability of real patient cohorts due to high costs of recruitment, underscores the vital deployment of scaling and transfer learning techniques. In this study, we explore a real-world pathology classification task to highlight the effectiveness of data and model scaling and cross-dataset knowledge transfer. As such, we observe varying performance improvements through data scaling, indicating the need for careful evaluation and labelling. Additionally, we identify the challenges of possible negative transfer and emphasize the significance of some key components to overcome distribution shifts and potential spurious correlations and achieve positive transfer. We see improvement in the performance of the target model on the target (NMT) datasets by using the knowledge from the source dataset (TUAB) when a low amount of labelled data was available. Our findings indicate a small and generic model (e.g. ShallowNet) performs well on a single dataset, however, a larger model (e.g. TCN) performs better on transfer and learning from a larger and diverse dataset.
{ "abstract": "Pathology diagnosis based on EEG signals and decoding brain activity holds\nimmense importance in understanding neurological disorders. With the\nadvancement of artificial intelligence methods and machine learning techniques,\nthe potential for accurate data-driven diagnoses and effective treatments has\ngrown significantly. However, applying machine learning algorithms to\nreal-world datasets presents diverse challenges at multiple levels. The\nscarcity of labelled data, especially in low regime scenarios with limited\navailability of real patient cohorts due to high costs of recruitment,\nunderscores the vital deployment of scaling and transfer learning techniques.\nIn this study, we explore a real-world pathology classification task to\nhighlight the effectiveness of data and model scaling and cross-dataset\nknowledge transfer. As such, we observe varying performance improvements\nthrough data scaling, indicating the need for careful evaluation and labelling.\nAdditionally, we identify the challenges of possible negative transfer and\nemphasize the significance of some key components to overcome distribution\nshifts and potential spurious correlations and achieve positive transfer. We\nsee improvement in the performance of the target model on the target (NMT)\ndatasets by using the knowledge from the source dataset (TUAB) when a low\namount of labelled data was available. Our findings indicate a small and\ngeneric model (e.g. ShallowNet) performs well on a single dataset, however, a\nlarger model (e.g. TCN) performs better on transfer and learning from a larger\nand diverse dataset.", "title": "Amplifying Pathological Detection in EEG Signaling Pathways through Cross-Dataset Transfer Learning", "url": "http://arxiv.org/abs/2309.10910v1" }
null
null
no_new_dataset
admin
null
false
null
42dfbf23-60ae-415f-a73f-e934204e0d77
null
Validated
2023-10-04 15:19:51.863444
{ "text_length": 1725 }
1no_new_dataset
TITLE: NEREL-BIO: A Dataset of Biomedical Abstracts Annotated with Nested Named Entities ABSTRACT: This paper describes NEREL-BIO -- an annotation scheme and corpus of PubMed abstracts in Russian and smaller number of abstracts in English. NEREL-BIO extends the general domain dataset NEREL by introducing domain-specific entity types. NEREL-BIO annotation scheme covers both general and biomedical domains making it suitable for domain transfer experiments. NEREL-BIO provides annotation for nested named entities as an extension of the scheme employed for NEREL. Nested named entities may cross entity boundaries to connect to shorter entities nested within longer entities, making them harder to detect. NEREL-BIO contains annotations for 700+ Russian and 100+ English abstracts. All English PubMed annotations have corresponding Russian counterparts. Thus, NEREL-BIO comprises the following specific features: annotation of nested named entities, it can be used as a benchmark for cross-domain (NEREL -> NEREL-BIO) and cross-language (English -> Russian) transfer. We experiment with both transformer-based sequence models and machine reading comprehension (MRC) models and report their results. The dataset is freely available at https://github.com/nerel-ds/NEREL-BIO.
{ "abstract": "This paper describes NEREL-BIO -- an annotation scheme and corpus of PubMed\nabstracts in Russian and smaller number of abstracts in English. NEREL-BIO\nextends the general domain dataset NEREL by introducing domain-specific entity\ntypes. NEREL-BIO annotation scheme covers both general and biomedical domains\nmaking it suitable for domain transfer experiments. NEREL-BIO provides\nannotation for nested named entities as an extension of the scheme employed for\nNEREL. Nested named entities may cross entity boundaries to connect to shorter\nentities nested within longer entities, making them harder to detect.\n NEREL-BIO contains annotations for 700+ Russian and 100+ English abstracts.\nAll English PubMed annotations have corresponding Russian counterparts. Thus,\nNEREL-BIO comprises the following specific features: annotation of nested named\nentities, it can be used as a benchmark for cross-domain (NEREL -> NEREL-BIO)\nand cross-language (English -> Russian) transfer. We experiment with both\ntransformer-based sequence models and machine reading comprehension (MRC)\nmodels and report their results.\n The dataset is freely available at https://github.com/nerel-ds/NEREL-BIO.", "title": "NEREL-BIO: A Dataset of Biomedical Abstracts Annotated with Nested Named Entities", "url": "http://arxiv.org/abs/2210.11913v1" }
null
null
new_dataset
admin
null
false
null
17db3f79-0176-40de-8448-c76da8578952
null
Validated
2023-10-04 15:19:51.883373
{ "text_length": 1294 }
0new_dataset
TITLE: Crime Prediction using Machine Learning with a Novel Crime Dataset ABSTRACT: Crime is an unlawful act that carries legal repercussions. Bangladesh has a high crime rate due to poverty, population growth, and many other socio-economic issues. For law enforcement agencies, understanding crime patterns is essential for preventing future criminal activity. For this purpose, these agencies need structured crime database. This paper introduces a novel crime dataset that contains temporal, geographic, weather, and demographic data about 6574 crime incidents of Bangladesh. We manually gather crime news articles of a seven year time span from a daily newspaper archive. We extract basic features from these raw text. Using these basic features, we then consult standard service-providers of geo-location and weather data in order to garner these information related to the collected crime incidents. Furthermore, we collect demographic information from Bangladesh National Census data. All these information are combined that results in a standard machine learning dataset. Together, 36 features are engineered for the crime prediction task. Five supervised machine learning classification algorithms are then evaluated on this newly built dataset and satisfactory results are achieved. We also conduct exploratory analysis on various aspects the dataset. This dataset is expected to serve as the foundation for crime incidence prediction systems for Bangladesh and other countries. The findings of this study will help law enforcement agencies to forecast and contain crime as well as to ensure optimal resource allocation for crime patrol and prevention.
{ "abstract": "Crime is an unlawful act that carries legal repercussions. Bangladesh has a\nhigh crime rate due to poverty, population growth, and many other\nsocio-economic issues. For law enforcement agencies, understanding crime\npatterns is essential for preventing future criminal activity. For this\npurpose, these agencies need structured crime database. This paper introduces a\nnovel crime dataset that contains temporal, geographic, weather, and\ndemographic data about 6574 crime incidents of Bangladesh. We manually gather\ncrime news articles of a seven year time span from a daily newspaper archive.\nWe extract basic features from these raw text. Using these basic features, we\nthen consult standard service-providers of geo-location and weather data in\norder to garner these information related to the collected crime incidents.\nFurthermore, we collect demographic information from Bangladesh National Census\ndata. All these information are combined that results in a standard machine\nlearning dataset. Together, 36 features are engineered for the crime prediction\ntask. Five supervised machine learning classification algorithms are then\nevaluated on this newly built dataset and satisfactory results are achieved. We\nalso conduct exploratory analysis on various aspects the dataset. This dataset\nis expected to serve as the foundation for crime incidence prediction systems\nfor Bangladesh and other countries. The findings of this study will help law\nenforcement agencies to forecast and contain crime as well as to ensure optimal\nresource allocation for crime patrol and prevention.", "title": "Crime Prediction using Machine Learning with a Novel Crime Dataset", "url": "http://arxiv.org/abs/2211.01551v1" }
null
null
new_dataset
admin
null
false
null
10f466ee-7a83-42d0-911f-27a639abf0ee
null
Validated
2023-10-04 15:19:51.883184
{ "text_length": 1679 }
0new_dataset
TITLE: SIB-200: A Simple, Inclusive, and Big Evaluation Dataset for Topic Classification in 200+ Languages and Dialects ABSTRACT: Despite the progress we have recorded in the last few years in multilingual natural language processing, evaluation is typically limited to a small set of languages with available datasets which excludes a large number of low-resource languages. In this paper, we created SIB-200 -- a large-scale open-sourced benchmark dataset for topic classification in 200 languages and dialects to address the lack of evaluation dataset for Natural Language Understanding (NLU). For many of the languages covered in SIB-200, this is the first publicly available evaluation dataset for NLU. The dataset is based on Flores-200 machine translation corpus. We annotated the English portion of the dataset and extended the sentence-level annotation to the remaining 203 languages covered in the corpus. Despite the simplicity of this task, our evaluation in full-supervised setting, cross-lingual transfer setting and prompting of large language model setting show that there is still a large gap between the performance of high-resource and low-resource languages when multilingual evaluation is scaled to numerous world languages. We found that languages unseen during the pre-training of multilingual language models, under-represented language families (like Nilotic and Altantic-Congo), and languages from the regions of Africa, Americas, Oceania and South East Asia, often have the lowest performance on our topic classification dataset. We hope our dataset will encourage a more inclusive evaluation of multilingual language models on a more diverse set of languages. https://github.com/dadelani/sib-200
{ "abstract": "Despite the progress we have recorded in the last few years in multilingual\nnatural language processing, evaluation is typically limited to a small set of\nlanguages with available datasets which excludes a large number of low-resource\nlanguages. In this paper, we created SIB-200 -- a large-scale open-sourced\nbenchmark dataset for topic classification in 200 languages and dialects to\naddress the lack of evaluation dataset for Natural Language Understanding\n(NLU). For many of the languages covered in SIB-200, this is the first publicly\navailable evaluation dataset for NLU. The dataset is based on Flores-200\nmachine translation corpus. We annotated the English portion of the dataset and\nextended the sentence-level annotation to the remaining 203 languages covered\nin the corpus. Despite the simplicity of this task, our evaluation in\nfull-supervised setting, cross-lingual transfer setting and prompting of large\nlanguage model setting show that there is still a large gap between the\nperformance of high-resource and low-resource languages when multilingual\nevaluation is scaled to numerous world languages. We found that languages\nunseen during the pre-training of multilingual language models,\nunder-represented language families (like Nilotic and Altantic-Congo), and\nlanguages from the regions of Africa, Americas, Oceania and South East Asia,\noften have the lowest performance on our topic classification dataset. We hope\nour dataset will encourage a more inclusive evaluation of multilingual language\nmodels on a more diverse set of languages. https://github.com/dadelani/sib-200", "title": "SIB-200: A Simple, Inclusive, and Big Evaluation Dataset for Topic Classification in 200+ Languages and Dialects", "url": "http://arxiv.org/abs/2309.07445v1" }
null
null
new_dataset
admin
null
false
null
784adf59-a3bb-4781-8112-38379a7d642d
null
Validated
2023-10-04 15:19:51.863592
{ "text_length": 1740 }
0new_dataset
TITLE: Artificial Empathy Classification: A Survey of Deep Learning Techniques, Datasets, and Evaluation Scales ABSTRACT: From the last decade, researchers in the field of machine learning (ML) and assistive developmental robotics (ADR) have taken an interest in artificial empathy (AE) as a possible future paradigm for human-robot interaction (HRI). Humans learn empathy since birth, therefore, it is challenging to instill this sense in robots and intelligent machines. Nevertheless, by training over a vast amount of data and time, imitating empathy, to a certain extent, can be possible for robots. Training techniques for AE, along with findings from the field of empathetic AI research, are ever-evolving. The standard workflow for artificial empathy consists of three stages: 1) Emotion Recognition (ER) using the retrieved features from video or textual data, 2) analyzing the perceived emotion or degree of empathy to choose the best course of action, and 3) carrying out a response action. Recent studies that show AE being used with virtual agents or robots often include Deep Learning (DL) techniques. For instance, models like VGGFace are used to conduct ER. Semi-supervised models like Autoencoders generate the corresponding emotional states and behavioral responses. However, there has not been any study that presents an independent approach for evaluating AE, or the degree to which a reaction was empathetic. This paper aims to investigate and evaluate existing works for measuring and evaluating empathy, as well as the datasets that have been collected and used so far. Our goal is to highlight and facilitate the use of state-of-the-art methods in the area of AE by comparing their performance. This will aid researchers in the area of AE in selecting their approaches with precision.
{ "abstract": "From the last decade, researchers in the field of machine learning (ML) and\nassistive developmental robotics (ADR) have taken an interest in artificial\nempathy (AE) as a possible future paradigm for human-robot interaction (HRI).\nHumans learn empathy since birth, therefore, it is challenging to instill this\nsense in robots and intelligent machines. Nevertheless, by training over a vast\namount of data and time, imitating empathy, to a certain extent, can be\npossible for robots. Training techniques for AE, along with findings from the\nfield of empathetic AI research, are ever-evolving. The standard workflow for\nartificial empathy consists of three stages: 1) Emotion Recognition (ER) using\nthe retrieved features from video or textual data, 2) analyzing the perceived\nemotion or degree of empathy to choose the best course of action, and 3)\ncarrying out a response action. Recent studies that show AE being used with\nvirtual agents or robots often include Deep Learning (DL) techniques. For\ninstance, models like VGGFace are used to conduct ER. Semi-supervised models\nlike Autoencoders generate the corresponding emotional states and behavioral\nresponses. However, there has not been any study that presents an independent\napproach for evaluating AE, or the degree to which a reaction was empathetic.\nThis paper aims to investigate and evaluate existing works for measuring and\nevaluating empathy, as well as the datasets that have been collected and used\nso far. Our goal is to highlight and facilitate the use of state-of-the-art\nmethods in the area of AE by comparing their performance. This will aid\nresearchers in the area of AE in selecting their approaches with precision.", "title": "Artificial Empathy Classification: A Survey of Deep Learning Techniques, Datasets, and Evaluation Scales", "url": "http://arxiv.org/abs/2310.00010v1" }
null
null
no_new_dataset
admin
null
false
null
9d5c5ecc-2d4b-4911-915c-e668da3f6d16
null
Validated
2023-10-04 15:19:51.863813
{ "text_length": 1824 }
1no_new_dataset
TITLE: Defectors: A Large, Diverse Python Dataset for Defect Prediction ABSTRACT: Defect prediction has been a popular research topic where machine learning (ML) and deep learning (DL) have found numerous applications. However, these ML/DL-based defect prediction models are often limited by the quality and size of their datasets. In this paper, we present Defectors, a large dataset for just-in-time and line-level defect prediction. Defectors consists of $\approx$ 213K source code files ($\approx$ 93K defective and $\approx$ 120K defect-free) that span across 24 popular Python projects. These projects come from 18 different domains, including machine learning, automation, and internet-of-things. Such a scale and diversity make Defectors a suitable dataset for training ML/DL models, especially transformer models that require large and diverse datasets. We also foresee several application areas of our dataset including defect prediction and defect explanation. Dataset link: https://doi.org/10.5281/zenodo.7708984
{ "abstract": "Defect prediction has been a popular research topic where machine learning\n(ML) and deep learning (DL) have found numerous applications. However, these\nML/DL-based defect prediction models are often limited by the quality and size\nof their datasets. In this paper, we present Defectors, a large dataset for\njust-in-time and line-level defect prediction. Defectors consists of $\\approx$\n213K source code files ($\\approx$ 93K defective and $\\approx$ 120K defect-free)\nthat span across 24 popular Python projects. These projects come from 18\ndifferent domains, including machine learning, automation, and\ninternet-of-things. Such a scale and diversity make Defectors a suitable\ndataset for training ML/DL models, especially transformer models that require\nlarge and diverse datasets. We also foresee several application areas of our\ndataset including defect prediction and defect explanation.\n Dataset link: https://doi.org/10.5281/zenodo.7708984", "title": "Defectors: A Large, Diverse Python Dataset for Defect Prediction", "url": "http://arxiv.org/abs/2303.04738v4" }
null
null
new_dataset
admin
null
false
null
22fdc178-4a7f-4574-b86b-5e3b7bd01a7c
null
Validated
2023-10-04 15:19:51.880607
{ "text_length": 1043 }
0new_dataset
TITLE: ESTER: A Machine Reading Comprehension Dataset for Event Semantic Relation Reasoning ABSTRACT: Understanding how events are semantically related to each other is the essence of reading comprehension. Recent event-centric reading comprehension datasets focus mostly on event arguments or temporal relations. While these tasks partially evaluate machines' ability of narrative understanding, human-like reading comprehension requires the capability to process event-based information beyond arguments and temporal reasoning. For example, to understand causality between events, we need to infer motivation or purpose; to establish event hierarchy, we need to understand the composition of events. To facilitate these tasks, we introduce ESTER, a comprehensive machine reading comprehension (MRC) dataset for Event Semantic Relation Reasoning. The dataset leverages natural language queries to reason about the five most common event semantic relations, provides more than 6K questions and captures 10.1K event relation pairs. Experimental results show that the current SOTA systems achieve 22.1%, 63.3%, and 83.5% for token-based exact-match, F1, and event-based HIT@1 scores, which are all significantly below human performances (36.0%, 79.6%, 100% respectively), highlighting our dataset as a challenging benchmark.
{ "abstract": "Understanding how events are semantically related to each other is the\nessence of reading comprehension. Recent event-centric reading comprehension\ndatasets focus mostly on event arguments or temporal relations. While these\ntasks partially evaluate machines' ability of narrative understanding,\nhuman-like reading comprehension requires the capability to process event-based\ninformation beyond arguments and temporal reasoning. For example, to understand\ncausality between events, we need to infer motivation or purpose; to establish\nevent hierarchy, we need to understand the composition of events. To facilitate\nthese tasks, we introduce ESTER, a comprehensive machine reading comprehension\n(MRC) dataset for Event Semantic Relation Reasoning. The dataset leverages\nnatural language queries to reason about the five most common event semantic\nrelations, provides more than 6K questions and captures 10.1K event relation\npairs. Experimental results show that the current SOTA systems achieve 22.1%,\n63.3%, and 83.5% for token-based exact-match, F1, and event-based HIT@1 scores,\nwhich are all significantly below human performances (36.0%, 79.6%, 100%\nrespectively), highlighting our dataset as a challenging benchmark.", "title": "ESTER: A Machine Reading Comprehension Dataset for Event Semantic Relation Reasoning", "url": "http://arxiv.org/abs/2104.08350v2" }
null
null
new_dataset
admin
null
false
null
1db2c1e2-fa8e-442e-a345-cec8be294fd5
null
Validated
2023-10-04 15:19:51.895003
{ "text_length": 1339 }
0new_dataset
TITLE: Potential sources of dataset bias complicate investigation of underdiagnosis by machine learning algorithms ABSTRACT: An increasing number of reports raise concerns about the risk that machine learning algorithms could amplify health disparities due to biases embedded in the training data. Seyyed-Kalantari et al. find that models trained on three chest X-ray datasets yield disparities in false-positive rates (FPR) across subgroups on the 'no-finding' label (indicating the absence of disease). The models consistently yield higher FPR on subgroups known to be historically underserved, and the study concludes that the models exhibit and potentially even amplify systematic underdiagnosis. We argue that the experimental setup in the study is insufficient to study algorithmic underdiagnosis. In the absence of specific knowledge (or assumptions) about the extent and nature of the dataset bias, it is difficult to investigate model bias. Importantly, their use of test data exhibiting the same bias as the training data (due to random splitting) severely complicates the interpretation of the reported disparities.
{ "abstract": "An increasing number of reports raise concerns about the risk that machine\nlearning algorithms could amplify health disparities due to biases embedded in\nthe training data. Seyyed-Kalantari et al. find that models trained on three\nchest X-ray datasets yield disparities in false-positive rates (FPR) across\nsubgroups on the 'no-finding' label (indicating the absence of disease). The\nmodels consistently yield higher FPR on subgroups known to be historically\nunderserved, and the study concludes that the models exhibit and potentially\neven amplify systematic underdiagnosis. We argue that the experimental setup in\nthe study is insufficient to study algorithmic underdiagnosis. In the absence\nof specific knowledge (or assumptions) about the extent and nature of the\ndataset bias, it is difficult to investigate model bias. Importantly, their use\nof test data exhibiting the same bias as the training data (due to random\nsplitting) severely complicates the interpretation of the reported disparities.", "title": "Potential sources of dataset bias complicate investigation of underdiagnosis by machine learning algorithms", "url": "http://arxiv.org/abs/2201.07856v2" }
null
null
no_new_dataset
admin
null
false
null
a5c8f03d-e3e3-4775-8878-03fe44c8eaf0
null
Validated
2023-10-04 15:19:51.888773
{ "text_length": 1143 }
1no_new_dataset
TITLE: TTSWING: a Dataset for Table Tennis Swing Analysis ABSTRACT: We introduce TTSWING, a novel dataset designed for table tennis swing analysis. This dataset comprises comprehensive swing information obtained through 9-axis sensors integrated into custom-made racket grips, accompanied by anonymized demographic data of the players. We detail the data collection and annotation procedures. Furthermore, we conduct pilot studies utilizing diverse machine learning models for swing analysis. TTSWING holds tremendous potential to facilitate innovative research in table tennis analysis and is a valuable resource for the scientific community. We release the dataset and experimental codes at https://github.com/DEPhantom/TTSWING.
{ "abstract": "We introduce TTSWING, a novel dataset designed for table tennis swing\nanalysis. This dataset comprises comprehensive swing information obtained\nthrough 9-axis sensors integrated into custom-made racket grips, accompanied by\nanonymized demographic data of the players. We detail the data collection and\nannotation procedures. Furthermore, we conduct pilot studies utilizing diverse\nmachine learning models for swing analysis. TTSWING holds tremendous potential\nto facilitate innovative research in table tennis analysis and is a valuable\nresource for the scientific community. We release the dataset and experimental\ncodes at https://github.com/DEPhantom/TTSWING.", "title": "TTSWING: a Dataset for Table Tennis Swing Analysis", "url": "http://arxiv.org/abs/2306.17550v1" }
null
null
new_dataset
admin
null
false
null
1279b7a2-2a6a-437b-ad7e-bc5a87c03f53
null
Validated
2023-10-04 15:19:51.868615
{ "text_length": 747 }
0new_dataset
TITLE: The Second DiCOVA Challenge: Dataset and performance analysis for COVID-19 diagnosis using acoustics ABSTRACT: The Second Diagnosis of COVID-19 using Acoustics (DiCOVA) Challenge aimed at accelerating the research in acoustics based detection of COVID-19, a topic at the intersection of acoustics, signal processing, machine learning, and healthcare. This paper presents the details of the challenge, which was an open call for researchers to analyze a dataset of audio recordings consisting of breathing, cough and speech signals. This data was collected from individuals with and without COVID-19 infection, and the task in the challenge was a two-class classification. The development set audio recordings were collected from 965 (172 COVID-19 positive) individuals, while the evaluation set contained data from 471 individuals (71 COVID-19 positive). The challenge featured four tracks, one associated with each sound category of cough, speech and breathing, and a fourth fusion track. A baseline system was also released to benchmark the participants. In this paper, we present an overview of the challenge, the rationale for the data collection and the baseline system. Further, a performance analysis for the systems submitted by the $16$ participating teams in the leaderboard is also presented.
{ "abstract": "The Second Diagnosis of COVID-19 using Acoustics (DiCOVA) Challenge aimed at\naccelerating the research in acoustics based detection of COVID-19, a topic at\nthe intersection of acoustics, signal processing, machine learning, and\nhealthcare. This paper presents the details of the challenge, which was an open\ncall for researchers to analyze a dataset of audio recordings consisting of\nbreathing, cough and speech signals. This data was collected from individuals\nwith and without COVID-19 infection, and the task in the challenge was a\ntwo-class classification. The development set audio recordings were collected\nfrom 965 (172 COVID-19 positive) individuals, while the evaluation set\ncontained data from 471 individuals (71 COVID-19 positive). The challenge\nfeatured four tracks, one associated with each sound category of cough, speech\nand breathing, and a fourth fusion track. A baseline system was also released\nto benchmark the participants. In this paper, we present an overview of the\nchallenge, the rationale for the data collection and the baseline system.\nFurther, a performance analysis for the systems submitted by the $16$\nparticipating teams in the leaderboard is also presented.", "title": "The Second DiCOVA Challenge: Dataset and performance analysis for COVID-19 diagnosis using acoustics", "url": "http://arxiv.org/abs/2110.01177v3" }
null
null
new_dataset
admin
null
false
null
51ac2033-7d25-4d0c-9c3d-406cccf579b3
null
Validated
2023-10-04 15:19:51.891717
{ "text_length": 1327 }
0new_dataset
TITLE: DiCOVA Challenge: Dataset, task, and baseline system for COVID-19 diagnosis using acoustics ABSTRACT: The DiCOVA challenge aims at accelerating research in diagnosing COVID-19 using acoustics (DiCOVA), a topic at the intersection of speech and audio processing, respiratory health diagnosis, and machine learning. This challenge is an open call for researchers to analyze a dataset of sound recordings collected from COVID-19 infected and non-COVID-19 individuals for a two-class classification. These recordings were collected via crowdsourcing from multiple countries, through a website application. The challenge features two tracks, one focusing on cough sounds, and the other on using a collection of breath, sustained vowel phonation, and number counting speech recordings. In this paper, we introduce the challenge and provide a detailed description of the task, and present a baseline system for the task.
{ "abstract": "The DiCOVA challenge aims at accelerating research in diagnosing COVID-19\nusing acoustics (DiCOVA), a topic at the intersection of speech and audio\nprocessing, respiratory health diagnosis, and machine learning. This challenge\nis an open call for researchers to analyze a dataset of sound recordings\ncollected from COVID-19 infected and non-COVID-19 individuals for a two-class\nclassification. These recordings were collected via crowdsourcing from multiple\ncountries, through a website application. The challenge features two tracks,\none focusing on cough sounds, and the other on using a collection of breath,\nsustained vowel phonation, and number counting speech recordings. In this\npaper, we introduce the challenge and provide a detailed description of the\ntask, and present a baseline system for the task.", "title": "DiCOVA Challenge: Dataset, task, and baseline system for COVID-19 diagnosis using acoustics", "url": "http://arxiv.org/abs/2103.09148v3" }
null
null
no_new_dataset
admin
null
false
null
1f2c8daa-431f-4361-962a-f8337331510c
null
Validated
2023-10-04 15:19:51.895407
{ "text_length": 937 }
1no_new_dataset
TITLE: GrowliFlower: An image time series dataset for GROWth analysis of cauLIFLOWER ABSTRACT: This article presents GrowliFlower, a georeferenced, image-based UAV time series dataset of two monitored cauliflower fields of size 0.39 and 0.60 ha acquired in 2020 and 2021. The dataset contains RGB and multispectral orthophotos from which about 14,000 individual plant coordinates are derived and provided. The coordinates enable the dataset users the extraction of complete and incomplete time series of image patches showing individual plants. The dataset contains collected phenotypic traits of 740 plants, including the developmental stage as well as plant and cauliflower size. As the harvestable product is completely covered by leaves, plant IDs and coordinates are provided to extract image pairs of plants pre and post defoliation, to facilitate estimations of cauliflower head size. Moreover, the dataset contains pixel-accurate leaf and plant instance segmentations, as well as stem annotations to address tasks like classification, detection, segmentation, instance segmentation, and similar computer vision tasks. The dataset aims to foster the development and evaluation of machine learning approaches. It specifically focuses on the analysis of growth and development of cauliflower and the derivation of phenotypic traits to foster the development of automation in agriculture. Two baseline results of instance segmentation at plant and leaf level based on the labeled instance segmentation data are presented. The entire data set is publicly available.
{ "abstract": "This article presents GrowliFlower, a georeferenced, image-based UAV time\nseries dataset of two monitored cauliflower fields of size 0.39 and 0.60 ha\nacquired in 2020 and 2021. The dataset contains RGB and multispectral\northophotos from which about 14,000 individual plant coordinates are derived\nand provided. The coordinates enable the dataset users the extraction of\ncomplete and incomplete time series of image patches showing individual plants.\nThe dataset contains collected phenotypic traits of 740 plants, including the\ndevelopmental stage as well as plant and cauliflower size. As the harvestable\nproduct is completely covered by leaves, plant IDs and coordinates are provided\nto extract image pairs of plants pre and post defoliation, to facilitate\nestimations of cauliflower head size. Moreover, the dataset contains\npixel-accurate leaf and plant instance segmentations, as well as stem\nannotations to address tasks like classification, detection, segmentation,\ninstance segmentation, and similar computer vision tasks. The dataset aims to\nfoster the development and evaluation of machine learning approaches. It\nspecifically focuses on the analysis of growth and development of cauliflower\nand the derivation of phenotypic traits to foster the development of automation\nin agriculture. Two baseline results of instance segmentation at plant and leaf\nlevel based on the labeled instance segmentation data are presented. The entire\ndata set is publicly available.", "title": "GrowliFlower: An image time series dataset for GROWth analysis of cauLIFLOWER", "url": "http://arxiv.org/abs/2204.00294v1" }
null
null
new_dataset
admin
null
false
null
32436224-4867-448d-a4f8-ce01f80d97ef
null
Validated
2023-10-04 15:19:51.887287
{ "text_length": 1585 }
0new_dataset
TITLE: ERS: a novel comprehensive endoscopy image dataset for machine learning, compliant with the MST 3.0 specification ABSTRACT: The article presents a new multi-label comprehensive image dataset from flexible endoscopy, colonoscopy and capsule endoscopy, named ERS. The collection has been labeled according to the full medical specification of 'Minimum Standard Terminology 3.0' (MST 3.0), describing all possible findings in the gastrointestinal tract (104 possible labels), extended with an additional 19 labels useful in common machine learning applications. The dataset contains around 6000 precisely and 115,000 approximately labeled frames from endoscopy videos, 3600 precise and 22,600 approximate segmentation masks, and 1.23 million unlabeled frames from flexible and capsule endoscopy videos. The labeled data cover almost entirely the MST 3.0 standard. The data came from 1520 videos of 1135 patients. Additionally, this paper proposes and describes four exemplary experiments in gastrointestinal image classification task performed using the created dataset. The obtained results indicate the high usefulness and flexibility of the dataset in training and testing machine learning algorithms in the field of endoscopic data analysis.
{ "abstract": "The article presents a new multi-label comprehensive image dataset from\nflexible endoscopy, colonoscopy and capsule endoscopy, named ERS. The\ncollection has been labeled according to the full medical specification of\n'Minimum Standard Terminology 3.0' (MST 3.0), describing all possible findings\nin the gastrointestinal tract (104 possible labels), extended with an\nadditional 19 labels useful in common machine learning applications.\n The dataset contains around 6000 precisely and 115,000 approximately labeled\nframes from endoscopy videos, 3600 precise and 22,600 approximate segmentation\nmasks, and 1.23 million unlabeled frames from flexible and capsule endoscopy\nvideos. The labeled data cover almost entirely the MST 3.0 standard. The data\ncame from 1520 videos of 1135 patients.\n Additionally, this paper proposes and describes four exemplary experiments in\ngastrointestinal image classification task performed using the created dataset.\nThe obtained results indicate the high usefulness and flexibility of the\ndataset in training and testing machine learning algorithms in the field of\nendoscopic data analysis.", "title": "ERS: a novel comprehensive endoscopy image dataset for machine learning, compliant with the MST 3.0 specification", "url": "http://arxiv.org/abs/2201.08746v1" }
null
null
new_dataset
admin
null
false
null
5076f321-8f7e-43c2-a2ed-291ee575abe3
null
Validated
2023-10-04 15:19:51.888701
{ "text_length": 1270 }
0new_dataset
TITLE: Hierarchical Optimal Transport for Comparing Histopathology Datasets ABSTRACT: Scarcity of labeled histopathology data limits the applicability of deep learning methods to under-profiled cancer types and labels. Transfer learning allows researchers to overcome the limitations of small datasets by pre-training machine learning models on larger datasets similar to the small target dataset. However, similarity between datasets is often determined heuristically. In this paper, we propose a principled notion of distance between histopathology datasets based on a hierarchical generalization of optimal transport distances. Our method does not require any training, is agnostic to model type, and preserves much of the hierarchical structure in histopathology datasets imposed by tiling. We apply our method to H&E stained slides from The Cancer Genome Atlas from six different cancer types. We show that our method outperforms a baseline distance in a cancer-type prediction task. Our results also show that our optimal transport distance predicts difficulty of transferability in a tumor vs.normal prediction setting.
{ "abstract": "Scarcity of labeled histopathology data limits the applicability of deep\nlearning methods to under-profiled cancer types and labels. Transfer learning\nallows researchers to overcome the limitations of small datasets by\npre-training machine learning models on larger datasets similar to the small\ntarget dataset. However, similarity between datasets is often determined\nheuristically. In this paper, we propose a principled notion of distance\nbetween histopathology datasets based on a hierarchical generalization of\noptimal transport distances. Our method does not require any training, is\nagnostic to model type, and preserves much of the hierarchical structure in\nhistopathology datasets imposed by tiling. We apply our method to H&E stained\nslides from The Cancer Genome Atlas from six different cancer types. We show\nthat our method outperforms a baseline distance in a cancer-type prediction\ntask. Our results also show that our optimal transport distance predicts\ndifficulty of transferability in a tumor vs.normal prediction setting.", "title": "Hierarchical Optimal Transport for Comparing Histopathology Datasets", "url": "http://arxiv.org/abs/2204.08324v2" }
null
null
no_new_dataset
admin
null
false
null
aff2f5ac-c3fe-40d3-b604-617879609885
null
Validated
2023-10-04 15:19:51.886986
{ "text_length": 1143 }
1no_new_dataset
TITLE: Online learning techniques for prediction of temporal tabular datasets with regime changes ABSTRACT: The application of deep learning to non-stationary temporal datasets can lead to overfitted models that underperform under regime changes. In this work, we propose a modular machine learning pipeline for ranking predictions on temporal panel datasets which is robust under regime changes. The modularity of the pipeline allows the use of different models, including Gradient Boosting Decision Trees (GBDTs) and Neural Networks, with and without feature engineering. We evaluate our framework on financial data for stock portfolio prediction, and find that GBDT models with dropout display high performance, robustness and generalisability with reduced complexity and computational cost. We then demonstrate how online learning techniques, which require no retraining of models, can be used post-prediction to enhance the results. First, we show that dynamic feature projection improves robustness by reducing drawdown in regime changes. Second, we demonstrate that dynamical model ensembling based on selection of models with good recent performance leads to improved Sharpe and Calmar ratios of out-of-sample predictions. We also evaluate the robustness of our pipeline across different data splits and random seeds with good reproducibility.
{ "abstract": "The application of deep learning to non-stationary temporal datasets can lead\nto overfitted models that underperform under regime changes. In this work, we\npropose a modular machine learning pipeline for ranking predictions on temporal\npanel datasets which is robust under regime changes. The modularity of the\npipeline allows the use of different models, including Gradient Boosting\nDecision Trees (GBDTs) and Neural Networks, with and without feature\nengineering. We evaluate our framework on financial data for stock portfolio\nprediction, and find that GBDT models with dropout display high performance,\nrobustness and generalisability with reduced complexity and computational cost.\nWe then demonstrate how online learning techniques, which require no retraining\nof models, can be used post-prediction to enhance the results. First, we show\nthat dynamic feature projection improves robustness by reducing drawdown in\nregime changes. Second, we demonstrate that dynamical model ensembling based on\nselection of models with good recent performance leads to improved Sharpe and\nCalmar ratios of out-of-sample predictions. We also evaluate the robustness of\nour pipeline across different data splits and random seeds with good\nreproducibility.", "title": "Online learning techniques for prediction of temporal tabular datasets with regime changes", "url": "http://arxiv.org/abs/2301.00790v4" }
null
null
no_new_dataset
admin
null
false
null
688ad769-5a09-425e-a2d0-24d7feb41d7a
null
Validated
2023-10-04 15:19:51.881727
{ "text_length": 1368 }
1no_new_dataset
TITLE: Gender and Racial Bias in Visual Question Answering Datasets ABSTRACT: Vision-and-language tasks have increasingly drawn more attention as a means to evaluate human-like reasoning in machine learning models. A popular task in the field is visual question answering (VQA), which aims to answer questions about images. However, VQA models have been shown to exploit language bias by learning the statistical correlations between questions and answers without looking into the image content: e.g., questions about the color of a banana are answered with yellow, even if the banana in the image is green. If societal bias (e.g., sexism, racism, ableism, etc.) is present in the training data, this problem may be causing VQA models to learn harmful stereotypes. For this reason, we investigate gender and racial bias in five VQA datasets. In our analysis, we find that the distribution of answers is highly different between questions about women and men, as well as the existence of detrimental gender-stereotypical samples. Likewise, we identify that specific race-related attributes are underrepresented, whereas potentially discriminatory samples appear in the analyzed datasets. Our findings suggest that there are dangers associated to using VQA datasets without considering and dealing with the potentially harmful stereotypes. We conclude the paper by proposing solutions to alleviate the problem before, during, and after the dataset collection process.
{ "abstract": "Vision-and-language tasks have increasingly drawn more attention as a means\nto evaluate human-like reasoning in machine learning models. A popular task in\nthe field is visual question answering (VQA), which aims to answer questions\nabout images. However, VQA models have been shown to exploit language bias by\nlearning the statistical correlations between questions and answers without\nlooking into the image content: e.g., questions about the color of a banana are\nanswered with yellow, even if the banana in the image is green. If societal\nbias (e.g., sexism, racism, ableism, etc.) is present in the training data,\nthis problem may be causing VQA models to learn harmful stereotypes. For this\nreason, we investigate gender and racial bias in five VQA datasets. In our\nanalysis, we find that the distribution of answers is highly different between\nquestions about women and men, as well as the existence of detrimental\ngender-stereotypical samples. Likewise, we identify that specific race-related\nattributes are underrepresented, whereas potentially discriminatory samples\nappear in the analyzed datasets. Our findings suggest that there are dangers\nassociated to using VQA datasets without considering and dealing with the\npotentially harmful stereotypes. We conclude the paper by proposing solutions\nto alleviate the problem before, during, and after the dataset collection\nprocess.", "title": "Gender and Racial Bias in Visual Question Answering Datasets", "url": "http://arxiv.org/abs/2205.08148v3" }
null
null
no_new_dataset
admin
null
false
null
a2dc7836-2bbd-4b76-b785-1e06276846c0
null
Validated
2023-10-04 15:19:51.886444
{ "text_length": 1482 }
1no_new_dataset
TITLE: Classification of datasets with imputed missing values: does imputation quality matter? ABSTRACT: Classifying samples in incomplete datasets is a common aim for machine learning practitioners, but is non-trivial. Missing data is found in most real-world datasets and these missing values are typically imputed using established methods, followed by classification of the now complete, imputed, samples. The focus of the machine learning researcher is then to optimise the downstream classification performance. In this study, we highlight that it is imperative to consider the quality of the imputation. We demonstrate how the commonly used measures for assessing quality are flawed and propose a new class of discrepancy scores which focus on how well the method recreates the overall distribution of the data. To conclude, we highlight the compromised interpretability of classifier models trained using poorly imputed data.
{ "abstract": "Classifying samples in incomplete datasets is a common aim for machine\nlearning practitioners, but is non-trivial. Missing data is found in most\nreal-world datasets and these missing values are typically imputed using\nestablished methods, followed by classification of the now complete, imputed,\nsamples. The focus of the machine learning researcher is then to optimise the\ndownstream classification performance. In this study, we highlight that it is\nimperative to consider the quality of the imputation. We demonstrate how the\ncommonly used measures for assessing quality are flawed and propose a new class\nof discrepancy scores which focus on how well the method recreates the overall\ndistribution of the data. To conclude, we highlight the compromised\ninterpretability of classifier models trained using poorly imputed data.", "title": "Classification of datasets with imputed missing values: does imputation quality matter?", "url": "http://arxiv.org/abs/2206.08478v1" }
null
null
no_new_dataset
admin
null
false
null
2647bd44-498b-42f9-9ef8-6cb20b36eed9
null
Validated
2023-10-04 15:19:51.885725
{ "text_length": 950 }
1no_new_dataset
TITLE: ST-MNIST -- The Spiking Tactile MNIST Neuromorphic Dataset ABSTRACT: Tactile sensing is an essential modality for smart robots as it enables them to interact flexibly with physical objects in their environment. Recent advancements in electronic skins have led to the development of data-driven machine learning methods that exploit this important sensory modality. However, current datasets used to train such algorithms are limited to standard synchronous tactile sensors. There is a dearth of neuromorphic event-based tactile datasets, principally due to the scarcity of large-scale event-based tactile sensors. Having such datasets is crucial for the development and evaluation of new algorithms that process spatio-temporal event-based data. For example, evaluating spiking neural networks on conventional frame-based datasets is considered sub-optimal. Here, we debut a novel neuromorphic Spiking Tactile MNIST (ST-MNIST) dataset, which comprises handwritten digits obtained by human participants writing on a neuromorphic tactile sensor array. We also describe an initial effort to evaluate our ST-MNIST dataset using existing artificial and spiking neural network models. The classification accuracies provided herein can serve as performance benchmarks for future work. We anticipate that our ST-MNIST dataset will be of interest and useful to the neuromorphic and robotics research communities.
{ "abstract": "Tactile sensing is an essential modality for smart robots as it enables them\nto interact flexibly with physical objects in their environment. Recent\nadvancements in electronic skins have led to the development of data-driven\nmachine learning methods that exploit this important sensory modality. However,\ncurrent datasets used to train such algorithms are limited to standard\nsynchronous tactile sensors. There is a dearth of neuromorphic event-based\ntactile datasets, principally due to the scarcity of large-scale event-based\ntactile sensors. Having such datasets is crucial for the development and\nevaluation of new algorithms that process spatio-temporal event-based data. For\nexample, evaluating spiking neural networks on conventional frame-based\ndatasets is considered sub-optimal. Here, we debut a novel neuromorphic Spiking\nTactile MNIST (ST-MNIST) dataset, which comprises handwritten digits obtained\nby human participants writing on a neuromorphic tactile sensor array. We also\ndescribe an initial effort to evaluate our ST-MNIST dataset using existing\nartificial and spiking neural network models. The classification accuracies\nprovided herein can serve as performance benchmarks for future work. We\nanticipate that our ST-MNIST dataset will be of interest and useful to the\nneuromorphic and robotics research communities.", "title": "ST-MNIST -- The Spiking Tactile MNIST Neuromorphic Dataset", "url": "http://arxiv.org/abs/2005.04319v1" }
null
null
new_dataset
admin
null
false
null
1e5201eb-c28c-49a8-bb4f-9553c88d774d
null
Validated
2023-10-04 15:19:51.900070
{ "text_length": 1427 }
0new_dataset
TITLE: X-RiSAWOZ: High-Quality End-to-End Multilingual Dialogue Datasets and Few-shot Agents ABSTRACT: Task-oriented dialogue research has mainly focused on a few popular languages like English and Chinese, due to the high dataset creation cost for a new language. To reduce the cost, we apply manual editing to automatically translated data. We create a new multilingual benchmark, X-RiSAWOZ, by translating the Chinese RiSAWOZ to 4 languages: English, French, Hindi, Korean; and a code-mixed English-Hindi language. X-RiSAWOZ has more than 18,000 human-verified dialogue utterances for each language, and unlike most multilingual prior work, is an end-to-end dataset for building fully-functioning agents. The many difficulties we encountered in creating X-RiSAWOZ led us to develop a toolset to accelerate the post-editing of a new language dataset after translation. This toolset improves machine translation with a hybrid entity alignment technique that combines neural with dictionary-based methods, along with many automated and semi-automated validation checks. We establish strong baselines for X-RiSAWOZ by training dialogue agents in the zero- and few-shot settings where limited gold data is available in the target language. Our results suggest that our translation and post-editing methodology and toolset can be used to create new high-quality multilingual dialogue agents cost-effectively. Our dataset, code, and toolkit are released open-source.
{ "abstract": "Task-oriented dialogue research has mainly focused on a few popular languages\nlike English and Chinese, due to the high dataset creation cost for a new\nlanguage. To reduce the cost, we apply manual editing to automatically\ntranslated data. We create a new multilingual benchmark, X-RiSAWOZ, by\ntranslating the Chinese RiSAWOZ to 4 languages: English, French, Hindi, Korean;\nand a code-mixed English-Hindi language. X-RiSAWOZ has more than 18,000\nhuman-verified dialogue utterances for each language, and unlike most\nmultilingual prior work, is an end-to-end dataset for building\nfully-functioning agents.\n The many difficulties we encountered in creating X-RiSAWOZ led us to develop\na toolset to accelerate the post-editing of a new language dataset after\ntranslation. This toolset improves machine translation with a hybrid entity\nalignment technique that combines neural with dictionary-based methods, along\nwith many automated and semi-automated validation checks.\n We establish strong baselines for X-RiSAWOZ by training dialogue agents in\nthe zero- and few-shot settings where limited gold data is available in the\ntarget language. Our results suggest that our translation and post-editing\nmethodology and toolset can be used to create new high-quality multilingual\ndialogue agents cost-effectively. Our dataset, code, and toolkit are released\nopen-source.", "title": "X-RiSAWOZ: High-Quality End-to-End Multilingual Dialogue Datasets and Few-shot Agents", "url": "http://arxiv.org/abs/2306.17674v1" }
null
null
new_dataset
admin
null
false
null
d762d175-1fc5-43e4-8427-a5049dfff4e1
null
Validated
2023-10-04 15:19:51.868362
{ "text_length": 1483 }
0new_dataset
TITLE: EarthNet2021: A novel large-scale dataset and challenge for forecasting localized climate impacts ABSTRACT: Climate change is global, yet its concrete impacts can strongly vary between different locations in the same region. Seasonal weather forecasts currently operate at the mesoscale (> 1 km). For more targeted mitigation and adaptation, modelling impacts to < 100 m is needed. Yet, the relationship between driving variables and Earth's surface at such local scales remains unresolved by current physical models. Large Earth observation datasets now enable us to create machine learning models capable of translating coarse weather information into high-resolution Earth surface forecasts. Here, we define high-resolution Earth surface forecasting as video prediction of satellite imagery conditional on mesoscale weather forecasts. Video prediction has been tackled with deep learning models. Developing such models requires analysis-ready datasets. We introduce EarthNet2021, a new, curated dataset containing target spatio-temporal Sentinel 2 satellite imagery at 20 m resolution, matched with high-resolution topography and mesoscale (1.28 km) weather variables. With over 32000 samples it is suitable for training deep neural networks. Comparing multiple Earth surface forecasts is not trivial. Hence, we define the EarthNetScore, a novel ranking criterion for models forecasting Earth surface reflectance. For model intercomparison we frame EarthNet2021 as a challenge with four tracks based on different test sets. These allow evaluation of model validity and robustness as well as model applicability to extreme events and the complete annual vegetation cycle. In addition to forecasting directly observable weather impacts through satellite-derived vegetation indices, capable Earth surface models will enable downstream applications such as crop yield prediction, forest health assessments, coastline management, or biodiversity monitoring. Find data, code, and how to participate at www.earthnet.tech .
{ "abstract": "Climate change is global, yet its concrete impacts can strongly vary between\ndifferent locations in the same region. Seasonal weather forecasts currently\noperate at the mesoscale (> 1 km). For more targeted mitigation and adaptation,\nmodelling impacts to < 100 m is needed. Yet, the relationship between driving\nvariables and Earth's surface at such local scales remains unresolved by\ncurrent physical models. Large Earth observation datasets now enable us to\ncreate machine learning models capable of translating coarse weather\ninformation into high-resolution Earth surface forecasts. Here, we define\nhigh-resolution Earth surface forecasting as video prediction of satellite\nimagery conditional on mesoscale weather forecasts. Video prediction has been\ntackled with deep learning models. Developing such models requires\nanalysis-ready datasets. We introduce EarthNet2021, a new, curated dataset\ncontaining target spatio-temporal Sentinel 2 satellite imagery at 20 m\nresolution, matched with high-resolution topography and mesoscale (1.28 km)\nweather variables. With over 32000 samples it is suitable for training deep\nneural networks. Comparing multiple Earth surface forecasts is not trivial.\nHence, we define the EarthNetScore, a novel ranking criterion for models\nforecasting Earth surface reflectance. For model intercomparison we frame\nEarthNet2021 as a challenge with four tracks based on different test sets.\nThese allow evaluation of model validity and robustness as well as model\napplicability to extreme events and the complete annual vegetation cycle. In\naddition to forecasting directly observable weather impacts through\nsatellite-derived vegetation indices, capable Earth surface models will enable\ndownstream applications such as crop yield prediction, forest health\nassessments, coastline management, or biodiversity monitoring. Find data, code,\nand how to participate at www.earthnet.tech .", "title": "EarthNet2021: A novel large-scale dataset and challenge for forecasting localized climate impacts", "url": "http://arxiv.org/abs/2012.06246v1" }
null
null
new_dataset
admin
null
false
null
456fcbdf-9699-419b-9065-cd916c678795
null
Validated
2023-10-04 15:19:51.896759
{ "text_length": 2042 }
0new_dataset
TITLE: LAHM : Large Annotated Dataset for Multi-Domain and Multilingual Hate Speech Identification ABSTRACT: Current research on hate speech analysis is typically oriented towards monolingual and single classification tasks. In this paper, we present a new multilingual hate speech analysis dataset for English, Hindi, Arabic, French, German and Spanish languages for multiple domains across hate speech - Abuse, Racism, Sexism, Religious Hate and Extremism. To the best of our knowledge, this paper is the first to address the problem of identifying various types of hate speech in these five wide domains in these six languages. In this work, we describe how we created the dataset, created annotations at high level and low level for different domains and how we use it to test the current state-of-the-art multilingual and multitask learning approaches. We evaluate our dataset in various monolingual, cross-lingual and machine translation classification settings and compare it against open source English datasets that we aggregated and merged for this task. Then we discuss how this approach can be used to create large scale hate-speech datasets and how to leverage our annotations in order to improve hate speech detection and classification in general.
{ "abstract": "Current research on hate speech analysis is typically oriented towards\nmonolingual and single classification tasks. In this paper, we present a new\nmultilingual hate speech analysis dataset for English, Hindi, Arabic, French,\nGerman and Spanish languages for multiple domains across hate speech - Abuse,\nRacism, Sexism, Religious Hate and Extremism. To the best of our knowledge,\nthis paper is the first to address the problem of identifying various types of\nhate speech in these five wide domains in these six languages. In this work, we\ndescribe how we created the dataset, created annotations at high level and low\nlevel for different domains and how we use it to test the current\nstate-of-the-art multilingual and multitask learning approaches. We evaluate\nour dataset in various monolingual, cross-lingual and machine translation\nclassification settings and compare it against open source English datasets\nthat we aggregated and merged for this task. Then we discuss how this approach\ncan be used to create large scale hate-speech datasets and how to leverage our\nannotations in order to improve hate speech detection and classification in\ngeneral.", "title": "LAHM : Large Annotated Dataset for Multi-Domain and Multilingual Hate Speech Identification", "url": "http://arxiv.org/abs/2304.00913v1" }
null
null
new_dataset
admin
null
false
null
2ecb2c7b-c994-4cdc-b5a4-b50bb80b99c0
null
Validated
2023-10-04 15:19:51.880098
{ "text_length": 1279 }
0new_dataset
TITLE: Domain adaptation using optimal transport for invariant learning using histopathology datasets ABSTRACT: Histopathology is critical for the diagnosis of many diseases, including cancer. These protocols typically require pathologists to manually evaluate slides under a microscope, which is time-consuming and subjective, leading to interest in machine learning to automate analysis. However, computational techniques are limited by batch effects, where technical factors like differences in preparation protocol or scanners can alter the appearance of slides, causing models trained on one institution to fail when generalizing to others. Here, we propose a domain adaptation method that improves the generalization of histopathological models to data from unseen institutions, without the need for labels or retraining in these new settings. Our approach introduces an optimal transport (OT) loss, that extends adversarial methods that penalize models if images from different institutions can be distinguished in their representation space. Unlike previous methods, which operate on single samples, our loss accounts for distributional differences between batches of images. We show that on the Camelyon17 dataset, while both methods can adapt to global differences in color distribution, only our OT loss can reliably classify a cancer phenotype unseen during training. Together, our results suggest that OT improves generalization on rare but critical phenotypes that may only make up a small fraction of the total tiles and variation in a slide.
{ "abstract": "Histopathology is critical for the diagnosis of many diseases, including\ncancer. These protocols typically require pathologists to manually evaluate\nslides under a microscope, which is time-consuming and subjective, leading to\ninterest in machine learning to automate analysis. However, computational\ntechniques are limited by batch effects, where technical factors like\ndifferences in preparation protocol or scanners can alter the appearance of\nslides, causing models trained on one institution to fail when generalizing to\nothers. Here, we propose a domain adaptation method that improves the\ngeneralization of histopathological models to data from unseen institutions,\nwithout the need for labels or retraining in these new settings. Our approach\nintroduces an optimal transport (OT) loss, that extends adversarial methods\nthat penalize models if images from different institutions can be distinguished\nin their representation space. Unlike previous methods, which operate on single\nsamples, our loss accounts for distributional differences between batches of\nimages. We show that on the Camelyon17 dataset, while both methods can adapt to\nglobal differences in color distribution, only our OT loss can reliably\nclassify a cancer phenotype unseen during training. Together, our results\nsuggest that OT improves generalization on rare but critical phenotypes that\nmay only make up a small fraction of the total tiles and variation in a slide.", "title": "Domain adaptation using optimal transport for invariant learning using histopathology datasets", "url": "http://arxiv.org/abs/2303.02241v1" }
null
null
no_new_dataset
admin
null
false
null
c3175827-96a5-4e7e-afc5-d5e9267be533
null
Validated
2023-10-04 15:19:51.880752
{ "text_length": 1574 }
1no_new_dataset
TITLE: Acoustic scene classification using auditory datasets ABSTRACT: The approach used not only challenges some of the fundamental mathematical techniques used so far in early experiments of the same trend but also introduces new scopes and new horizons for interesting results. The physics governing spectrograms have been optimized in the project along with exploring how it handles the intense requirements of the problem at hand. Major contributions and developments brought under the light, through this project involve using better mathematical techniques and problem-specific machine learning methods. Improvised data analysis and data augmentation for audio datasets like frequency masking and random frequency-time stretching are used in the project and hence are explained in this paper. In the used methodology, the audio transforms principle were also tried and explored, and indeed the insights gained were used constructively in the later stages of the project. Using a deep learning principle is surely one of them. Also, in this paper, the potential scopes and upcoming research openings in both short and long term tunnel of time has been presented. Although much of the results gained are domain-specific as of now, they are surely potent enough to produce novel solutions in various different domains of diverse backgrounds.
{ "abstract": "The approach used not only challenges some of the fundamental mathematical\ntechniques used so far in early experiments of the same trend but also\nintroduces new scopes and new horizons for interesting results. The physics\ngoverning spectrograms have been optimized in the project along with exploring\nhow it handles the intense requirements of the problem at hand. Major\ncontributions and developments brought under the light, through this project\ninvolve using better mathematical techniques and problem-specific machine\nlearning methods. Improvised data analysis and data augmentation for audio\ndatasets like frequency masking and random frequency-time stretching are used\nin the project and hence are explained in this paper. In the used methodology,\nthe audio transforms principle were also tried and explored, and indeed the\ninsights gained were used constructively in the later stages of the project.\nUsing a deep learning principle is surely one of them. Also, in this paper, the\npotential scopes and upcoming research openings in both short and long term\ntunnel of time has been presented. Although much of the results gained are\ndomain-specific as of now, they are surely potent enough to produce novel\nsolutions in various different domains of diverse backgrounds.", "title": "Acoustic scene classification using auditory datasets", "url": "http://arxiv.org/abs/2112.13450v2" }
null
null
no_new_dataset
admin
null
false
null
22e36565-c3e6-4636-95d0-0d72ba44ee6d
null
Validated
2023-10-04 15:19:51.889118
{ "text_length": 1362 }
1no_new_dataset
TITLE: Multiplexed gradient descent: Fast online training of modern datasets on hardware neural networks without backpropagation ABSTRACT: We present multiplexed gradient descent (MGD), a gradient descent framework designed to easily train analog or digital neural networks in hardware. MGD utilizes zero-order optimization techniques for online training of hardware neural networks. We demonstrate its ability to train neural networks on modern machine learning datasets, including CIFAR-10 and Fashion-MNIST, and compare its performance to backpropagation. Assuming realistic timescales and hardware parameters, our results indicate that these optimization techniques can train a network on emerging hardware platforms orders of magnitude faster than the wall-clock time of training via backpropagation on a standard GPU, even in the presence of imperfect weight updates or device-to-device variations in the hardware. We additionally describe how it can be applied to existing hardware as part of chip-in-the-loop training, or integrated directly at the hardware level. Crucially, the MGD framework is highly flexible, and its gradient descent process can be optimized to compensate for specific hardware limitations such as slow parameter-update speeds or limited input bandwidth.
{ "abstract": "We present multiplexed gradient descent (MGD), a gradient descent framework\ndesigned to easily train analog or digital neural networks in hardware. MGD\nutilizes zero-order optimization techniques for online training of hardware\nneural networks. We demonstrate its ability to train neural networks on modern\nmachine learning datasets, including CIFAR-10 and Fashion-MNIST, and compare\nits performance to backpropagation. Assuming realistic timescales and hardware\nparameters, our results indicate that these optimization techniques can train a\nnetwork on emerging hardware platforms orders of magnitude faster than the\nwall-clock time of training via backpropagation on a standard GPU, even in the\npresence of imperfect weight updates or device-to-device variations in the\nhardware. We additionally describe how it can be applied to existing hardware\nas part of chip-in-the-loop training, or integrated directly at the hardware\nlevel. Crucially, the MGD framework is highly flexible, and its gradient\ndescent process can be optimized to compensate for specific hardware\nlimitations such as slow parameter-update speeds or limited input bandwidth.", "title": "Multiplexed gradient descent: Fast online training of modern datasets on hardware neural networks without backpropagation", "url": "http://arxiv.org/abs/2303.03986v1" }
null
null
no_new_dataset
admin
null
false
null
ec525273-b7d8-418f-9434-79fb74949007
null
Validated
2023-10-04 15:19:51.880679
{ "text_length": 1301 }
1no_new_dataset
TITLE: A Vietnamese Dataset for Evaluating Machine Reading Comprehension ABSTRACT: Over 97 million people speak Vietnamese as their native language in the world. However, there are few research studies on machine reading comprehension (MRC) for Vietnamese, the task of understanding a text and answering questions related to it. Due to the lack of benchmark datasets for Vietnamese, we present the Vietnamese Question Answering Dataset (UIT-ViQuAD), a new dataset for the low-resource language as Vietnamese to evaluate MRC models. This dataset comprises over 23,000 human-generated question-answer pairs based on 5,109 passages of 174 Vietnamese articles from Wikipedia. In particular, we propose a new process of dataset creation for Vietnamese MRC. Our in-depth analyses illustrate that our dataset requires abilities beyond simple reasoning like word matching and demands single-sentence and multiple-sentence inferences. Besides, we conduct experiments on state-of-the-art MRC methods for English and Chinese as the first experimental models on UIT-ViQuAD. We also estimate human performance on the dataset and compare it to the experimental results of powerful machine learning models. As a result, the substantial differences between human performance and the best model performance on the dataset indicate that improvements can be made on UIT-ViQuAD in future research. Our dataset is freely available on our website to encourage the research community to overcome challenges in Vietnamese MRC.
{ "abstract": "Over 97 million people speak Vietnamese as their native language in the\nworld. However, there are few research studies on machine reading comprehension\n(MRC) for Vietnamese, the task of understanding a text and answering questions\nrelated to it. Due to the lack of benchmark datasets for Vietnamese, we present\nthe Vietnamese Question Answering Dataset (UIT-ViQuAD), a new dataset for the\nlow-resource language as Vietnamese to evaluate MRC models. This dataset\ncomprises over 23,000 human-generated question-answer pairs based on 5,109\npassages of 174 Vietnamese articles from Wikipedia. In particular, we propose a\nnew process of dataset creation for Vietnamese MRC. Our in-depth analyses\nillustrate that our dataset requires abilities beyond simple reasoning like\nword matching and demands single-sentence and multiple-sentence inferences.\nBesides, we conduct experiments on state-of-the-art MRC methods for English and\nChinese as the first experimental models on UIT-ViQuAD. We also estimate human\nperformance on the dataset and compare it to the experimental results of\npowerful machine learning models. As a result, the substantial differences\nbetween human performance and the best model performance on the dataset\nindicate that improvements can be made on UIT-ViQuAD in future research. Our\ndataset is freely available on our website to encourage the research community\nto overcome challenges in Vietnamese MRC.", "title": "A Vietnamese Dataset for Evaluating Machine Reading Comprehension", "url": "http://arxiv.org/abs/2009.14725v3" }
null
null
new_dataset
admin
null
false
null
1732f1e0-19f3-4699-a9af-8dfab99b291d
null
Validated
2023-10-04 15:19:51.897958
{ "text_length": 1519 }
0new_dataset
TITLE: An Energy Activity Dataset for Smart Homes ABSTRACT: A smart home energy dataset that records miscellaneous energy consumption data is publicly offered. The proposed energy activity dataset (EAD) has a high data type diversity in contrast to existing load monitoring datasets. In EAD, a simple data point is labeled with the appliance, brand, and event information, whereas a complex data point has an extra application label. Several discoveries have been made on the energy consumption patterns of many appliances. Load curves of the appliances are measured when different events and applications are triggered and utilized. A revised longest-common-subsequence (LCS) similarity measurement algorithm is proposed to calculate energy dataset similarities. Thus, the data quality prior information becomes available before training machine learning models. In addition, a subsample convolutional neural network (SCNN) is put forward. It serves as a non-intrusive optical character recognition (OCR) approach to obtain energy data directly from monitors of power meters. The link for the EAD dataset is: https://drive.google.com/drive/folders/1zn0V6Q8eXXSKxKgcs8ZRValL5VEn3anD
{ "abstract": "A smart home energy dataset that records miscellaneous energy consumption\ndata is publicly offered. The proposed energy activity dataset (EAD) has a high\ndata type diversity in contrast to existing load monitoring datasets. In EAD, a\nsimple data point is labeled with the appliance, brand, and event information,\nwhereas a complex data point has an extra application label. Several\ndiscoveries have been made on the energy consumption patterns of many\nappliances. Load curves of the appliances are measured when different events\nand applications are triggered and utilized. A revised\nlongest-common-subsequence (LCS) similarity measurement algorithm is proposed\nto calculate energy dataset similarities. Thus, the data quality prior\ninformation becomes available before training machine learning models. In\naddition, a subsample convolutional neural network (SCNN) is put forward. It\nserves as a non-intrusive optical character recognition (OCR) approach to\nobtain energy data directly from monitors of power meters. The link for the EAD\ndataset is:\nhttps://drive.google.com/drive/folders/1zn0V6Q8eXXSKxKgcs8ZRValL5VEn3anD", "title": "An Energy Activity Dataset for Smart Homes", "url": "http://arxiv.org/abs/2208.13416v2" }
null
null
new_dataset
admin
null
false
null
fbc700f8-ea1f-4563-94d0-3781d522ad20
null
Validated
2023-10-04 15:19:51.884441
{ "text_length": 1199 }
0new_dataset
TITLE: Robustness-preserving Lifelong Learning via Dataset Condensation ABSTRACT: Lifelong learning (LL) aims to improve a predictive model as the data source evolves continuously. Most work in this learning paradigm has focused on resolving the problem of 'catastrophic forgetting,' which refers to a notorious dilemma between improving model accuracy over new data and retaining accuracy over previous data. Yet, it is also known that machine learning (ML) models can be vulnerable in the sense that tiny, adversarial input perturbations can deceive the models into producing erroneous predictions. This motivates the research objective of this paper - specification of a new LL framework that can salvage model robustness (against adversarial attacks) from catastrophic forgetting. Specifically, we propose a new memory-replay LL strategy that leverages modern bi-level optimization techniques to determine the 'coreset' of the current data (i.e., a small amount of data to be memorized) for ease of preserving adversarial robustness over time. We term the resulting LL framework 'Data-Efficient Robustness-Preserving LL' (DERPLL). The effectiveness of DERPLL is evaluated for class-incremental image classification using ResNet-18 over the CIFAR-10 dataset. Experimental results show that DERPLL outperforms the conventional coreset-guided LL baseline and achieves a substantial improvement in both standard accuracy and robust accuracy.
{ "abstract": "Lifelong learning (LL) aims to improve a predictive model as the data source\nevolves continuously. Most work in this learning paradigm has focused on\nresolving the problem of 'catastrophic forgetting,' which refers to a notorious\ndilemma between improving model accuracy over new data and retaining accuracy\nover previous data. Yet, it is also known that machine learning (ML) models can\nbe vulnerable in the sense that tiny, adversarial input perturbations can\ndeceive the models into producing erroneous predictions. This motivates the\nresearch objective of this paper - specification of a new LL framework that can\nsalvage model robustness (against adversarial attacks) from catastrophic\nforgetting. Specifically, we propose a new memory-replay LL strategy that\nleverages modern bi-level optimization techniques to determine the 'coreset' of\nthe current data (i.e., a small amount of data to be memorized) for ease of\npreserving adversarial robustness over time. We term the resulting LL framework\n'Data-Efficient Robustness-Preserving LL' (DERPLL). The effectiveness of DERPLL\nis evaluated for class-incremental image classification using ResNet-18 over\nthe CIFAR-10 dataset. Experimental results show that DERPLL outperforms the\nconventional coreset-guided LL baseline and achieves a substantial improvement\nin both standard accuracy and robust accuracy.", "title": "Robustness-preserving Lifelong Learning via Dataset Condensation", "url": "http://arxiv.org/abs/2303.04183v1" }
null
null
no_new_dataset
admin
null
false
null
ce2dc330-1032-4edf-9368-543cc1dbcdfa
null
Validated
2023-10-04 15:19:51.880631
{ "text_length": 1458 }
1no_new_dataset
TITLE: mMARCO: A Multilingual Version of the MS MARCO Passage Ranking Dataset ABSTRACT: The MS MARCO ranking dataset has been widely used for training deep learning models for IR tasks, achieving considerable effectiveness on diverse zero-shot scenarios. However, this type of resource is scarce in languages other than English. In this work, we present mMARCO, a multilingual version of the MS MARCO passage ranking dataset comprising 13 languages that was created using machine translation. We evaluated mMARCO by finetuning monolingual and multilingual reranking models, as well as a multilingual dense retrieval model on this dataset. We also evaluated models finetuned using the mMARCO dataset in a zero-shot scenario on Mr. TyDi dataset, demonstrating that multilingual models finetuned on our translated dataset achieve superior effectiveness to models finetuned on the original English version alone. Our experiments also show that a distilled multilingual reranker is competitive with non-distilled models while having 5.4 times fewer parameters. Lastly, we show a positive correlation between translation quality and retrieval effectiveness, providing evidence that improvements in translation methods might lead to improvements in multilingual information retrieval. The translated datasets and finetuned models are available at https://github.com/unicamp-dl/mMARCO.
{ "abstract": "The MS MARCO ranking dataset has been widely used for training deep learning\nmodels for IR tasks, achieving considerable effectiveness on diverse zero-shot\nscenarios. However, this type of resource is scarce in languages other than\nEnglish. In this work, we present mMARCO, a multilingual version of the MS\nMARCO passage ranking dataset comprising 13 languages that was created using\nmachine translation. We evaluated mMARCO by finetuning monolingual and\nmultilingual reranking models, as well as a multilingual dense retrieval model\non this dataset. We also evaluated models finetuned using the mMARCO dataset in\na zero-shot scenario on Mr. TyDi dataset, demonstrating that multilingual\nmodels finetuned on our translated dataset achieve superior effectiveness to\nmodels finetuned on the original English version alone. Our experiments also\nshow that a distilled multilingual reranker is competitive with non-distilled\nmodels while having 5.4 times fewer parameters. Lastly, we show a positive\ncorrelation between translation quality and retrieval effectiveness, providing\nevidence that improvements in translation methods might lead to improvements in\nmultilingual information retrieval. The translated datasets and finetuned\nmodels are available at https://github.com/unicamp-dl/mMARCO.", "title": "mMARCO: A Multilingual Version of the MS MARCO Passage Ranking Dataset", "url": "http://arxiv.org/abs/2108.13897v5" }
null
null
new_dataset
admin
null
false
null
00fffa41-4e56-4436-b7c8-fccaf32deaeb
null
Validated
2023-10-04 15:19:51.892387
{ "text_length": 1394 }
0new_dataset
TITLE: Generating Higher-Fidelity Synthetic Datasets with Privacy Guarantees ABSTRACT: This paper considers the problem of enhancing user privacy in common machine learning development tasks, such as data annotation and inspection, by substituting the real data with samples form a generative adversarial network. We propose employing Bayesian differential privacy as the means to achieve a rigorous theoretical guarantee while providing a better privacy-utility trade-off. We demonstrate experimentally that our approach produces higher-fidelity samples, compared to prior work, allowing to (1) detect more subtle data errors and biases, and (2) reduce the need for real data labelling by achieving high accuracy when training directly on artificial samples.
{ "abstract": "This paper considers the problem of enhancing user privacy in common machine\nlearning development tasks, such as data annotation and inspection, by\nsubstituting the real data with samples form a generative adversarial network.\nWe propose employing Bayesian differential privacy as the means to achieve a\nrigorous theoretical guarantee while providing a better privacy-utility\ntrade-off. We demonstrate experimentally that our approach produces\nhigher-fidelity samples, compared to prior work, allowing to (1) detect more\nsubtle data errors and biases, and (2) reduce the need for real data labelling\nby achieving high accuracy when training directly on artificial samples.", "title": "Generating Higher-Fidelity Synthetic Datasets with Privacy Guarantees", "url": "http://arxiv.org/abs/2003.00997v1" }
null
null
no_new_dataset
admin
null
false
null
73a60d1d-cc44-4f4e-86b3-7b38a7ea6785
null
Validated
2023-10-04 15:19:51.901455
{ "text_length": 776 }
1no_new_dataset
TITLE: Common Phone: A Multilingual Dataset for Robust Acoustic Modelling ABSTRACT: Current state of the art acoustic models can easily comprise more than 100 million parameters. This growing complexity demands larger training datasets to maintain a decent generalization of the final decision function. An ideal dataset is not necessarily large in size, but large with respect to the amount of unique speakers, utilized hardware and varying recording conditions. This enables a machine learning model to explore as much of the domain-specific input space as possible during parameter estimation. This work introduces Common Phone, a gender-balanced, multilingual corpus recorded from more than 11.000 contributors via Mozilla's Common Voice project. It comprises around 116 hours of speech enriched with automatically generated phonetic segmentation. A Wav2Vec 2.0 acoustic model was trained with the Common Phone to perform phonetic symbol recognition and validate the quality of the generated phonetic annotation. The architecture achieved a PER of 18.1 % on the entire test set, computed with all 101 unique phonetic symbols, showing slight differences between the individual languages. We conclude that Common Phone provides sufficient variability and reliable phonetic annotation to help bridging the gap between research and application of acoustic models.
{ "abstract": "Current state of the art acoustic models can easily comprise more than 100\nmillion parameters. This growing complexity demands larger training datasets to\nmaintain a decent generalization of the final decision function. An ideal\ndataset is not necessarily large in size, but large with respect to the amount\nof unique speakers, utilized hardware and varying recording conditions. This\nenables a machine learning model to explore as much of the domain-specific\ninput space as possible during parameter estimation. This work introduces\nCommon Phone, a gender-balanced, multilingual corpus recorded from more than\n11.000 contributors via Mozilla's Common Voice project. It comprises around 116\nhours of speech enriched with automatically generated phonetic segmentation. A\nWav2Vec 2.0 acoustic model was trained with the Common Phone to perform\nphonetic symbol recognition and validate the quality of the generated phonetic\nannotation. The architecture achieved a PER of 18.1 % on the entire test set,\ncomputed with all 101 unique phonetic symbols, showing slight differences\nbetween the individual languages. We conclude that Common Phone provides\nsufficient variability and reliable phonetic annotation to help bridging the\ngap between research and application of acoustic models.", "title": "Common Phone: A Multilingual Dataset for Robust Acoustic Modelling", "url": "http://arxiv.org/abs/2201.05912v2" }
null
null
new_dataset
admin
null
false
null
5df4810f-28b7-4f4a-8875-8f2b1c3c1df5
null
Validated
2023-10-04 15:19:51.888820
{ "text_length": 1380 }
0new_dataset
TITLE: EDEN: Multimodal Synthetic Dataset of Enclosed GarDEN Scenes ABSTRACT: Multimodal large-scale datasets for outdoor scenes are mostly designed for urban driving problems. The scenes are highly structured and semantically different from scenarios seen in nature-centered scenes such as gardens or parks. To promote machine learning methods for nature-oriented applications, such as agriculture and gardening, we propose the multimodal synthetic dataset for Enclosed garDEN scenes (EDEN). The dataset features more than 300K images captured from more than 100 garden models. Each image is annotated with various low/high-level vision modalities, including semantic segmentation, depth, surface normals, intrinsic colors, and optical flow. Experimental results on the state-of-the-art methods for semantic segmentation and monocular depth prediction, two important tasks in computer vision, show positive impact of pre-training deep networks on our dataset for unstructured natural scenes. The dataset and related materials will be available at https://lhoangan.github.io/eden.
{ "abstract": "Multimodal large-scale datasets for outdoor scenes are mostly designed for\nurban driving problems. The scenes are highly structured and semantically\ndifferent from scenarios seen in nature-centered scenes such as gardens or\nparks. To promote machine learning methods for nature-oriented applications,\nsuch as agriculture and gardening, we propose the multimodal synthetic dataset\nfor Enclosed garDEN scenes (EDEN). The dataset features more than 300K images\ncaptured from more than 100 garden models. Each image is annotated with various\nlow/high-level vision modalities, including semantic segmentation, depth,\nsurface normals, intrinsic colors, and optical flow. Experimental results on\nthe state-of-the-art methods for semantic segmentation and monocular depth\nprediction, two important tasks in computer vision, show positive impact of\npre-training deep networks on our dataset for unstructured natural scenes. The\ndataset and related materials will be available at\nhttps://lhoangan.github.io/eden.", "title": "EDEN: Multimodal Synthetic Dataset of Enclosed GarDEN Scenes", "url": "http://arxiv.org/abs/2011.04389v2" }
null
null
new_dataset
admin
null
false
null
688404c5-7c71-456d-b3d0-371c712d64e5
null
Validated
2023-10-04 15:19:51.897092
{ "text_length": 1097 }
0new_dataset
TITLE: SPICE, A Dataset of Drug-like Molecules and Peptides for Training Machine Learning Potentials ABSTRACT: Machine learning potentials are an important tool for molecular simulation, but their development is held back by a shortage of high quality datasets to train them on. We describe the SPICE dataset, a new quantum chemistry dataset for training potentials relevant to simulating drug-like small molecules interacting with proteins. It contains over 1.1 million conformations for a diverse set of small molecules, dimers, dipeptides, and solvated amino acids. It includes 15 elements, charged and uncharged molecules, and a wide range of covalent and non-covalent interactions. It provides both forces and energies calculated at the {\omega}B97M-D3(BJ)/def2-TZVPPD level of theory, along with other useful quantities such as multipole moments and bond orders. We train a set of machine learning potentials on it and demonstrate that they can achieve chemical accuracy across a broad region of chemical space. It can serve as a valuable resource for the creation of transferable, ready to use potential functions for use in molecular simulations.
{ "abstract": "Machine learning potentials are an important tool for molecular simulation,\nbut their development is held back by a shortage of high quality datasets to\ntrain them on. We describe the SPICE dataset, a new quantum chemistry dataset\nfor training potentials relevant to simulating drug-like small molecules\ninteracting with proteins. It contains over 1.1 million conformations for a\ndiverse set of small molecules, dimers, dipeptides, and solvated amino acids.\nIt includes 15 elements, charged and uncharged molecules, and a wide range of\ncovalent and non-covalent interactions. It provides both forces and energies\ncalculated at the {\\omega}B97M-D3(BJ)/def2-TZVPPD level of theory, along with\nother useful quantities such as multipole moments and bond orders. We train a\nset of machine learning potentials on it and demonstrate that they can achieve\nchemical accuracy across a broad region of chemical space. It can serve as a\nvaluable resource for the creation of transferable, ready to use potential\nfunctions for use in molecular simulations.", "title": "SPICE, A Dataset of Drug-like Molecules and Peptides for Training Machine Learning Potentials", "url": "http://arxiv.org/abs/2209.10702v2" }
null
null
new_dataset
admin
null
false
null
1218fc36-d914-47a0-b45d-25c4d61b317d
null
Validated
2023-10-04 15:19:51.883853
{ "text_length": 1171 }
0new_dataset
TITLE: S2abEL: A Dataset for Entity Linking from Scientific Tables ABSTRACT: Entity linking (EL) is the task of linking a textual mention to its corresponding entry in a knowledge base, and is critical for many knowledge-intensive NLP applications. When applied to tables in scientific papers, EL is a step toward large-scale scientific knowledge bases that could enable advanced scientific question answering and analytics. We present the first dataset for EL in scientific tables. EL for scientific tables is especially challenging because scientific knowledge bases can be very incomplete, and disambiguating table mentions typically requires understanding the papers's tet in addition to the table. Our dataset, S2abEL, focuses on EL in machine learning results tables and includes hand-labeled cell types, attributed sources, and entity links from the PaperswithCode taxonomy for 8,429 cells from 732 tables. We introduce a neural baseline method designed for EL on scientific tables containing many out-of-knowledge-base mentions, and show that it significantly outperforms a state-of-the-art generic table EL method. The best baselines fall below human performance, and our analysis highlights avenues for improvement.
{ "abstract": "Entity linking (EL) is the task of linking a textual mention to its\ncorresponding entry in a knowledge base, and is critical for many\nknowledge-intensive NLP applications. When applied to tables in scientific\npapers, EL is a step toward large-scale scientific knowledge bases that could\nenable advanced scientific question answering and analytics. We present the\nfirst dataset for EL in scientific tables. EL for scientific tables is\nespecially challenging because scientific knowledge bases can be very\nincomplete, and disambiguating table mentions typically requires understanding\nthe papers's tet in addition to the table. Our dataset, S2abEL, focuses on EL\nin machine learning results tables and includes hand-labeled cell types,\nattributed sources, and entity links from the PaperswithCode taxonomy for 8,429\ncells from 732 tables. We introduce a neural baseline method designed for EL on\nscientific tables containing many out-of-knowledge-base mentions, and show that\nit significantly outperforms a state-of-the-art generic table EL method. The\nbest baselines fall below human performance, and our analysis highlights\navenues for improvement.", "title": "S2abEL: A Dataset for Entity Linking from Scientific Tables", "url": "http://arxiv.org/abs/2305.00366v1" }
null
null
no_new_dataset
admin
null
false
null
091d8728-b3a4-4de9-a5e1-a2419e5729c4
null
Validated
2023-10-04 15:19:51.878972
{ "text_length": 1242 }
1no_new_dataset
TITLE: ReforesTree: A Dataset for Estimating Tropical Forest Carbon Stock with Deep Learning and Aerial Imagery ABSTRACT: Forest biomass is a key influence for future climate, and the world urgently needs highly scalable financing schemes, such as carbon offsetting certifications, to protect and restore forests. Current manual forest carbon stock inventory methods of measuring single trees by hand are time, labour, and cost-intensive and have been shown to be subjective. They can lead to substantial overestimation of the carbon stock and ultimately distrust in forest financing. The potential for impact and scale of leveraging advancements in machine learning and remote sensing technologies is promising but needs to be of high quality in order to replace the current forest stock protocols for certifications. In this paper, we present ReforesTree, a benchmark dataset of forest carbon stock in six agro-forestry carbon offsetting sites in Ecuador. Furthermore, we show that a deep learning-based end-to-end model using individual tree detection from low cost RGB-only drone imagery is accurately estimating forest carbon stock within official carbon offsetting certification standards. Additionally, our baseline CNN model outperforms state-of-the-art satellite-based forest biomass and carbon stock estimates for this type of small-scale, tropical agro-forestry sites. We present this dataset to encourage machine learning research in this area to increase accountability and transparency of monitoring, verification and reporting (MVR) in carbon offsetting projects, as well as scaling global reforestation financing through accurate remote sensing.
{ "abstract": "Forest biomass is a key influence for future climate, and the world urgently\nneeds highly scalable financing schemes, such as carbon offsetting\ncertifications, to protect and restore forests. Current manual forest carbon\nstock inventory methods of measuring single trees by hand are time, labour, and\ncost-intensive and have been shown to be subjective. They can lead to\nsubstantial overestimation of the carbon stock and ultimately distrust in\nforest financing. The potential for impact and scale of leveraging advancements\nin machine learning and remote sensing technologies is promising but needs to\nbe of high quality in order to replace the current forest stock protocols for\ncertifications.\n In this paper, we present ReforesTree, a benchmark dataset of forest carbon\nstock in six agro-forestry carbon offsetting sites in Ecuador. Furthermore, we\nshow that a deep learning-based end-to-end model using individual tree\ndetection from low cost RGB-only drone imagery is accurately estimating forest\ncarbon stock within official carbon offsetting certification standards.\nAdditionally, our baseline CNN model outperforms state-of-the-art\nsatellite-based forest biomass and carbon stock estimates for this type of\nsmall-scale, tropical agro-forestry sites. We present this dataset to encourage\nmachine learning research in this area to increase accountability and\ntransparency of monitoring, verification and reporting (MVR) in carbon\noffsetting projects, as well as scaling global reforestation financing through\naccurate remote sensing.", "title": "ReforesTree: A Dataset for Estimating Tropical Forest Carbon Stock with Deep Learning and Aerial Imagery", "url": "http://arxiv.org/abs/2201.11192v1" }
null
null
new_dataset
admin
null
false
null
55d43471-737a-40d3-b966-0060a4b01cb3
null
Validated
2023-10-04 15:19:51.888629
{ "text_length": 1680 }
0new_dataset
TITLE: Analyzing Wearables Dataset to Predict ADLs and Falls: A Pilot Study ABSTRACT: Healthcare is an important aspect of human life. Use of technologies in healthcare has increased manifolds after the pandemic. Internet of Things based systems and devices proposed in literature can help elders, children and adults facing/experiencing health problems. This paper exhaustively reviews thirty-nine wearable based datasets which can be used for evaluating the system to recognize Activities of Daily Living and Falls. A comparative analysis on the SisFall dataset using five machine learning methods i.e., Logistic Regression, Linear Discriminant Analysis, K-Nearest Neighbor, Decision Tree and Naive Bayes is performed in python. The dataset is modified in two ways, in first all the attributes present in dataset are used as it is and labelled in binary form. In second, magnitude of three axes(x,y,z) for three sensors value are computed and then used in experiment with label attribute. The experiments are performed on one subject, ten subjects and all the subjects and compared in terms of accuracy, precision and recall. The results obtained from this study proves that KNN outperforms other machine learning methods in terms of accuracy, precision and recall. It is also concluded that personalization of data improves accuracy.
{ "abstract": "Healthcare is an important aspect of human life. Use of technologies in\nhealthcare has increased manifolds after the pandemic. Internet of Things based\nsystems and devices proposed in literature can help elders, children and adults\nfacing/experiencing health problems. This paper exhaustively reviews\nthirty-nine wearable based datasets which can be used for evaluating the system\nto recognize Activities of Daily Living and Falls. A comparative analysis on\nthe SisFall dataset using five machine learning methods i.e., Logistic\nRegression, Linear Discriminant Analysis, K-Nearest Neighbor, Decision Tree and\nNaive Bayes is performed in python. The dataset is modified in two ways, in\nfirst all the attributes present in dataset are used as it is and labelled in\nbinary form. In second, magnitude of three axes(x,y,z) for three sensors value\nare computed and then used in experiment with label attribute. The experiments\nare performed on one subject, ten subjects and all the subjects and compared in\nterms of accuracy, precision and recall. The results obtained from this study\nproves that KNN outperforms other machine learning methods in terms of\naccuracy, precision and recall. It is also concluded that personalization of\ndata improves accuracy.", "title": "Analyzing Wearables Dataset to Predict ADLs and Falls: A Pilot Study", "url": "http://arxiv.org/abs/2209.04785v1" }
null
null
no_new_dataset
admin
null
false
null
1796464e-b557-4e4c-ba9e-6ab34a6c16e8
null
Validated
2023-10-04 15:19:51.884113
{ "text_length": 1353 }
1no_new_dataset
TITLE: Hausa Visual Genome: A Dataset for Multi-Modal English to Hausa Machine Translation ABSTRACT: Multi-modal Machine Translation (MMT) enables the use of visual information to enhance the quality of translations. The visual information can serve as a valuable piece of context information to decrease the ambiguity of input sentences. Despite the increasing popularity of such a technique, good and sizeable datasets are scarce, limiting the full extent of their potential. Hausa, a Chadic language, is a member of the Afro-Asiatic language family. It is estimated that about 100 to 150 million people speak the language, with more than 80 million indigenous speakers. This is more than any of the other Chadic languages. Despite a large number of speakers, the Hausa language is considered low-resource in natural language processing (NLP). This is due to the absence of sufficient resources to implement most NLP tasks. While some datasets exist, they are either scarce, machine-generated, or in the religious domain. Therefore, there is a need to create training and evaluation data for implementing machine learning tasks and bridging the research gap in the language. This work presents the Hausa Visual Genome (HaVG), a dataset that contains the description of an image or a section within the image in Hausa and its equivalent in English. To prepare the dataset, we started by translating the English description of the images in the Hindi Visual Genome (HVG) into Hausa automatically. Afterward, the synthetic Hausa data was carefully post-edited considering the respective images. The dataset comprises 32,923 images and their descriptions that are divided into training, development, test, and challenge test set. The Hausa Visual Genome is the first dataset of its kind and can be used for Hausa-English machine translation, multi-modal research, and image description, among various other natural language processing and generation tasks.
{ "abstract": "Multi-modal Machine Translation (MMT) enables the use of visual information\nto enhance the quality of translations. The visual information can serve as a\nvaluable piece of context information to decrease the ambiguity of input\nsentences. Despite the increasing popularity of such a technique, good and\nsizeable datasets are scarce, limiting the full extent of their potential.\nHausa, a Chadic language, is a member of the Afro-Asiatic language family. It\nis estimated that about 100 to 150 million people speak the language, with more\nthan 80 million indigenous speakers. This is more than any of the other Chadic\nlanguages. Despite a large number of speakers, the Hausa language is considered\nlow-resource in natural language processing (NLP). This is due to the absence\nof sufficient resources to implement most NLP tasks. While some datasets exist,\nthey are either scarce, machine-generated, or in the religious domain.\nTherefore, there is a need to create training and evaluation data for\nimplementing machine learning tasks and bridging the research gap in the\nlanguage. This work presents the Hausa Visual Genome (HaVG), a dataset that\ncontains the description of an image or a section within the image in Hausa and\nits equivalent in English. To prepare the dataset, we started by translating\nthe English description of the images in the Hindi Visual Genome (HVG) into\nHausa automatically. Afterward, the synthetic Hausa data was carefully\npost-edited considering the respective images. The dataset comprises 32,923\nimages and their descriptions that are divided into training, development,\ntest, and challenge test set. The Hausa Visual Genome is the first dataset of\nits kind and can be used for Hausa-English machine translation, multi-modal\nresearch, and image description, among various other natural language\nprocessing and generation tasks.", "title": "Hausa Visual Genome: A Dataset for Multi-Modal English to Hausa Machine Translation", "url": "http://arxiv.org/abs/2205.01133v2" }
null
null
new_dataset
admin
null
false
null
c9bad79b-ac52-41cd-8fab-1470e9828ded
null
Validated
2023-10-04 15:19:51.886681
{ "text_length": 1971 }
0new_dataset
TITLE: COVIDx CXR-3: A Large-Scale, Open-Source Benchmark Dataset of Chest X-ray Images for Computer-Aided COVID-19 Diagnostics ABSTRACT: After more than two years since the beginning of the COVID-19 pandemic, the pressure of this crisis continues to devastate globally. The use of chest X-ray (CXR) imaging as a complementary screening strategy to RT-PCR testing is not only prevailing but has greatly increased due to its routine clinical use for respiratory complaints. Thus far, many visual perception models have been proposed for COVID-19 screening based on CXR imaging. Nevertheless, the accuracy and the generalization capacity of these models are very much dependent on the diversity and the size of the dataset they were trained on. Motivated by this, we introduce COVIDx CXR-3, a large-scale benchmark dataset of CXR images for supporting COVID-19 computer vision research. COVIDx CXR-3 is composed of 30,386 CXR images from a multinational cohort of 17,026 patients from at least 51 countries, making it, to the best of our knowledge, the most extensive, most diverse COVID-19 CXR dataset in open access form. Here, we provide comprehensive details on the various aspects of the proposed dataset including patient demographics, imaging views, and infection types. The hope is that COVIDx CXR-3 can assist scientists in advancing machine learning research against both the COVID-19 pandemic and related diseases.
{ "abstract": "After more than two years since the beginning of the COVID-19 pandemic, the\npressure of this crisis continues to devastate globally. The use of chest X-ray\n(CXR) imaging as a complementary screening strategy to RT-PCR testing is not\nonly prevailing but has greatly increased due to its routine clinical use for\nrespiratory complaints. Thus far, many visual perception models have been\nproposed for COVID-19 screening based on CXR imaging. Nevertheless, the\naccuracy and the generalization capacity of these models are very much\ndependent on the diversity and the size of the dataset they were trained on.\nMotivated by this, we introduce COVIDx CXR-3, a large-scale benchmark dataset\nof CXR images for supporting COVID-19 computer vision research. COVIDx CXR-3 is\ncomposed of 30,386 CXR images from a multinational cohort of 17,026 patients\nfrom at least 51 countries, making it, to the best of our knowledge, the most\nextensive, most diverse COVID-19 CXR dataset in open access form. Here, we\nprovide comprehensive details on the various aspects of the proposed dataset\nincluding patient demographics, imaging views, and infection types. The hope is\nthat COVIDx CXR-3 can assist scientists in advancing machine learning research\nagainst both the COVID-19 pandemic and related diseases.", "title": "COVIDx CXR-3: A Large-Scale, Open-Source Benchmark Dataset of Chest X-ray Images for Computer-Aided COVID-19 Diagnostics", "url": "http://arxiv.org/abs/2206.03671v3" }
null
null
new_dataset
admin
null
false
null
92963ad3-350f-49c8-b149-cbe55e288a61
null
Validated
2023-10-04 15:19:51.885890
{ "text_length": 1440 }
0new_dataset
TITLE: A Case Study on the Classification of Lost Circulation Events During Drilling using Machine Learning Techniques on an Imbalanced Large Dataset ABSTRACT: This study presents machine learning models that forecast and categorize lost circulation severity preemptively using a large class imbalanced drilling dataset. We demonstrate reproducible core techniques involved in tackling a large drilling engineering challenge utilizing easily interpretable machine learning approaches. We utilized a 65,000+ records data with class imbalance problem from Azadegan oilfield formations in Iran. Eleven of the dataset's seventeen parameters are chosen to be used in the classification of five lost circulation events. To generate classification models, we used six basic machine learning algorithms and four ensemble learning methods. Linear Discriminant Analysis (LDA), Logistic Regression (LR), Support Vector Machines (SVM), Classification and Regression Trees (CART), k-Nearest Neighbors (KNN), and Gaussian Naive Bayes (GNB) are the six fundamental techniques. We also used bagging and boosting ensemble learning techniques in the investigation of solutions for improved predicting performance. The performance of these algorithms is measured using four metrics: accuracy, precision, recall, and F1-score. The F1-score weighted to represent the data imbalance is chosen as the preferred evaluation criterion. The CART model was found to be the best in class for identifying drilling fluid circulation loss events with an average weighted F1-score of 0.9904 and standard deviation of 0.0015. Upon application of ensemble learning techniques, a Random Forest ensemble of decision trees showed the best predictive performance. It identified and classified lost circulation events with a perfect weighted F1-score of 1.0. Using Permutation Feature Importance (PFI), the measured depth was found to be the most influential factor in accurately recognizing lost circulation events while drilling.
{ "abstract": "This study presents machine learning models that forecast and categorize lost\ncirculation severity preemptively using a large class imbalanced drilling\ndataset. We demonstrate reproducible core techniques involved in tackling a\nlarge drilling engineering challenge utilizing easily interpretable machine\nlearning approaches.\n We utilized a 65,000+ records data with class imbalance problem from Azadegan\noilfield formations in Iran. Eleven of the dataset's seventeen parameters are\nchosen to be used in the classification of five lost circulation events. To\ngenerate classification models, we used six basic machine learning algorithms\nand four ensemble learning methods. Linear Discriminant Analysis (LDA),\nLogistic Regression (LR), Support Vector Machines (SVM), Classification and\nRegression Trees (CART), k-Nearest Neighbors (KNN), and Gaussian Naive Bayes\n(GNB) are the six fundamental techniques. We also used bagging and boosting\nensemble learning techniques in the investigation of solutions for improved\npredicting performance. The performance of these algorithms is measured using\nfour metrics: accuracy, precision, recall, and F1-score. The F1-score weighted\nto represent the data imbalance is chosen as the preferred evaluation\ncriterion.\n The CART model was found to be the best in class for identifying drilling\nfluid circulation loss events with an average weighted F1-score of 0.9904 and\nstandard deviation of 0.0015. Upon application of ensemble learning techniques,\na Random Forest ensemble of decision trees showed the best predictive\nperformance. It identified and classified lost circulation events with a\nperfect weighted F1-score of 1.0. Using Permutation Feature Importance (PFI),\nthe measured depth was found to be the most influential factor in accurately\nrecognizing lost circulation events while drilling.", "title": "A Case Study on the Classification of Lost Circulation Events During Drilling using Machine Learning Techniques on an Imbalanced Large Dataset", "url": "http://arxiv.org/abs/2209.01607v2" }
null
null
no_new_dataset
admin
null
false
null
aa7fbe27-3a54-4e61-9172-64185ae18634
null
Validated
2023-10-04 15:19:51.884322
{ "text_length": 2012 }
1no_new_dataset
TITLE: Making Machine Learning Datasets and Models FAIR for HPC: A Methodology and Case Study ABSTRACT: The FAIR Guiding Principles aim to improve the findability, accessibility, interoperability, and reusability of digital content by making them both human and machine actionable. However, these principles have not yet been broadly adopted in the domain of machine learning-based program analyses and optimizations for High-Performance Computing (HPC). In this paper, we design a methodology to make HPC datasets and machine learning models FAIR after investigating existing FAIRness assessment and improvement techniques. Our methodology includes a comprehensive, quantitative assessment for elected data, followed by concrete, actionable suggestions to improve FAIRness with respect to common issues related to persistent identifiers, rich metadata descriptions, license and provenance information. Moreover, we select a representative training dataset to evaluate our methodology. The experiment shows the methodology can effectively improve the dataset and model's FAIRness from an initial score of 19.1% to the final score of 83.0%.
{ "abstract": "The FAIR Guiding Principles aim to improve the findability, accessibility,\ninteroperability, and reusability of digital content by making them both human\nand machine actionable. However, these principles have not yet been broadly\nadopted in the domain of machine learning-based program analyses and\noptimizations for High-Performance Computing (HPC). In this paper, we design a\nmethodology to make HPC datasets and machine learning models FAIR after\ninvestigating existing FAIRness assessment and improvement techniques. Our\nmethodology includes a comprehensive, quantitative assessment for elected data,\nfollowed by concrete, actionable suggestions to improve FAIRness with respect\nto common issues related to persistent identifiers, rich metadata descriptions,\nlicense and provenance information. Moreover, we select a representative\ntraining dataset to evaluate our methodology. The experiment shows the\nmethodology can effectively improve the dataset and model's FAIRness from an\ninitial score of 19.1% to the final score of 83.0%.", "title": "Making Machine Learning Datasets and Models FAIR for HPC: A Methodology and Case Study", "url": "http://arxiv.org/abs/2211.02092v1" }
null
null
no_new_dataset
admin
null
false
null
0e687961-6e1d-4894-976e-690d0f572f05
null
Validated
2023-10-04 15:19:51.883114
{ "text_length": 1156 }
1no_new_dataset
TITLE: BDSL 49: A Comprehensive Dataset of Bangla Sign Language ABSTRACT: Language is a method by which individuals express their thoughts. Each language has its own set of alphabetic and numeric characters. People can communicate with one another through either oral or written communication. However, each language has a sign language counterpart. Individuals who are deaf and/or mute communicate through sign language. The Bangla language also has a sign language, which is called BDSL. The dataset is about Bangla hand sign images. The collection contains 49 individual Bangla alphabet images in sign language. BDSL49 is a dataset that consists of 29,490 images with 49 labels. Images of 14 different adult individuals, each with a distinct background and appearance, have been recorded during data collection. Several strategies have been used to eliminate noise from datasets during preparation. This dataset is available to researchers for free. They can develop automated systems using machine learning, computer vision, and deep learning techniques. In addition, two models were used in this dataset. The first is for detection, while the second is for recognition.
{ "abstract": "Language is a method by which individuals express their thoughts. Each\nlanguage has its own set of alphabetic and numeric characters. People can\ncommunicate with one another through either oral or written communication.\nHowever, each language has a sign language counterpart. Individuals who are\ndeaf and/or mute communicate through sign language. The Bangla language also\nhas a sign language, which is called BDSL. The dataset is about Bangla hand\nsign images. The collection contains 49 individual Bangla alphabet images in\nsign language. BDSL49 is a dataset that consists of 29,490 images with 49\nlabels. Images of 14 different adult individuals, each with a distinct\nbackground and appearance, have been recorded during data collection. Several\nstrategies have been used to eliminate noise from datasets during preparation.\nThis dataset is available to researchers for free. They can develop automated\nsystems using machine learning, computer vision, and deep learning techniques.\nIn addition, two models were used in this dataset. The first is for detection,\nwhile the second is for recognition.", "title": "BDSL 49: A Comprehensive Dataset of Bangla Sign Language", "url": "http://arxiv.org/abs/2208.06827v1" }
null
null
new_dataset
admin
null
false
null
bf1872e3-f3b0-4029-bbea-57dfd9131db4
null
Validated
2023-10-04 15:19:51.884816
{ "text_length": 1191 }
0new_dataset
TITLE: Uncovering Political Hate Speech During Indian Election Campaign: A New Low-Resource Dataset and Baselines ABSTRACT: The detection of hate speech in political discourse is a critical issue, and this becomes even more challenging in low-resource languages. To address this issue, we introduce a new dataset named IEHate, which contains 11,457 manually annotated Hindi tweets related to the Indian Assembly Election Campaign from November 1, 2021, to March 9, 2022. We performed a detailed analysis of the dataset, focusing on the prevalence of hate speech in political communication and the different forms of hateful language used. Additionally, we benchmark the dataset using a range of machine learning, deep learning, and transformer-based algorithms. Our experiments reveal that the performance of these models can be further improved, highlighting the need for more advanced techniques for hate speech detection in low-resource languages. In particular, the relatively higher score of human evaluation over algorithms emphasizes the importance of utilizing both human and automated approaches for effective hate speech moderation. Our IEHate dataset can serve as a valuable resource for researchers and practitioners working on developing and evaluating hate speech detection techniques in low-resource languages. Overall, our work underscores the importance of addressing the challenges of identifying and mitigating hate speech in political discourse, particularly in the context of low-resource languages. The dataset and resources for this work are made available at https://github.com/Farhan-jafri/Indian-Election.
{ "abstract": "The detection of hate speech in political discourse is a critical issue, and\nthis becomes even more challenging in low-resource languages. To address this\nissue, we introduce a new dataset named IEHate, which contains 11,457 manually\nannotated Hindi tweets related to the Indian Assembly Election Campaign from\nNovember 1, 2021, to March 9, 2022. We performed a detailed analysis of the\ndataset, focusing on the prevalence of hate speech in political communication\nand the different forms of hateful language used. Additionally, we benchmark\nthe dataset using a range of machine learning, deep learning, and\ntransformer-based algorithms. Our experiments reveal that the performance of\nthese models can be further improved, highlighting the need for more advanced\ntechniques for hate speech detection in low-resource languages. In particular,\nthe relatively higher score of human evaluation over algorithms emphasizes the\nimportance of utilizing both human and automated approaches for effective hate\nspeech moderation. Our IEHate dataset can serve as a valuable resource for\nresearchers and practitioners working on developing and evaluating hate speech\ndetection techniques in low-resource languages. Overall, our work underscores\nthe importance of addressing the challenges of identifying and mitigating hate\nspeech in political discourse, particularly in the context of low-resource\nlanguages. The dataset and resources for this work are made available at\nhttps://github.com/Farhan-jafri/Indian-Election.", "title": "Uncovering Political Hate Speech During Indian Election Campaign: A New Low-Resource Dataset and Baselines", "url": "http://arxiv.org/abs/2306.14764v2" }
null
null
new_dataset
admin
null
false
null
de814005-b946-4aef-b269-5f1760a84ba6
null
Validated
2023-10-04 15:19:51.869961
{ "text_length": 1648 }
0new_dataset
TITLE: Hyperparameter Importance of Quantum Neural Networks Across Small Datasets ABSTRACT: As restricted quantum computers are slowly becoming a reality, the search for meaningful first applications intensifies. In this domain, one of the more investigated approaches is the use of a special type of quantum circuit - a so-called quantum neural network -- to serve as a basis for a machine learning model. Roughly speaking, as the name suggests, a quantum neural network can play a similar role to a neural network. However, specifically for applications in machine learning contexts, very little is known about suitable circuit architectures, or model hyperparameters one should use to achieve good learning performance. In this work, we apply the functional ANOVA framework to quantum neural networks to analyze which of the hyperparameters were most influential for their predictive performance. We analyze one of the most typically used quantum neural network architectures. We then apply this to $7$ open-source datasets from the OpenML-CC18 classification benchmark whose number of features is small enough to fit on quantum hardware with less than $20$ qubits. Three main levels of importance were detected from the ranking of hyperparameters obtained with functional ANOVA. Our experiment both confirmed expected patterns and revealed new insights. For instance, setting well the learning rate is deemed the most critical hyperparameter in terms of marginal contribution on all datasets, whereas the particular choice of entangling gates used is considered the least important except on one dataset. This work introduces new methodologies to study quantum machine learning models and provides new insights toward quantum model selection.
{ "abstract": "As restricted quantum computers are slowly becoming a reality, the search for\nmeaningful first applications intensifies. In this domain, one of the more\ninvestigated approaches is the use of a special type of quantum circuit - a\nso-called quantum neural network -- to serve as a basis for a machine learning\nmodel. Roughly speaking, as the name suggests, a quantum neural network can\nplay a similar role to a neural network. However, specifically for applications\nin machine learning contexts, very little is known about suitable circuit\narchitectures, or model hyperparameters one should use to achieve good learning\nperformance. In this work, we apply the functional ANOVA framework to quantum\nneural networks to analyze which of the hyperparameters were most influential\nfor their predictive performance. We analyze one of the most typically used\nquantum neural network architectures. We then apply this to $7$ open-source\ndatasets from the OpenML-CC18 classification benchmark whose number of features\nis small enough to fit on quantum hardware with less than $20$ qubits. Three\nmain levels of importance were detected from the ranking of hyperparameters\nobtained with functional ANOVA. Our experiment both confirmed expected patterns\nand revealed new insights. For instance, setting well the learning rate is\ndeemed the most critical hyperparameter in terms of marginal contribution on\nall datasets, whereas the particular choice of entangling gates used is\nconsidered the least important except on one dataset. This work introduces new\nmethodologies to study quantum machine learning models and provides new\ninsights toward quantum model selection.", "title": "Hyperparameter Importance of Quantum Neural Networks Across Small Datasets", "url": "http://arxiv.org/abs/2206.09992v1" }
null
null
no_new_dataset
admin
null
false
null
bfc8fcf3-486b-43bb-97d2-204ccbba07b4
null
Validated
2023-10-04 15:19:51.885630
{ "text_length": 1763 }
1no_new_dataset
TITLE: Analyzing the Use of Character-Level Translation with Sparse and Noisy Datasets ABSTRACT: This paper provides an analysis of character-level machine translation models used in pivot-based translation when applied to sparse and noisy datasets, such as crowdsourced movie subtitles. In our experiments, we find that such character-level models cut the number of untranslated words by over 40% and are especially competitive (improvements of 2-3 BLEU points) in the case of limited training data. We explore the impact of character alignment, phrase table filtering, bitext size and the choice of pivot language on translation quality. We further compare cascaded translation models to the use of synthetic training data via multiple pivots, and we find that the latter works significantly better. Finally, we demonstrate that neither word-nor character-BLEU correlate perfectly with human judgments, due to BLEU's sensitivity to length.
{ "abstract": "This paper provides an analysis of character-level machine translation models\nused in pivot-based translation when applied to sparse and noisy datasets, such\nas crowdsourced movie subtitles. In our experiments, we find that such\ncharacter-level models cut the number of untranslated words by over 40% and are\nespecially competitive (improvements of 2-3 BLEU points) in the case of limited\ntraining data. We explore the impact of character alignment, phrase table\nfiltering, bitext size and the choice of pivot language on translation quality.\nWe further compare cascaded translation models to the use of synthetic training\ndata via multiple pivots, and we find that the latter works significantly\nbetter. Finally, we demonstrate that neither word-nor character-BLEU correlate\nperfectly with human judgments, due to BLEU's sensitivity to length.", "title": "Analyzing the Use of Character-Level Translation with Sparse and Noisy Datasets", "url": "http://arxiv.org/abs/2109.13723v1" }
null
null
no_new_dataset
admin
null
false
null
75afec9f-e139-4424-a813-01ed35f385ad
null
Validated
2023-10-04 15:19:51.891924
{ "text_length": 958 }
1no_new_dataset
TITLE: Computer Vision based inspection on post-earthquake with UAV synthetic dataset ABSTRACT: The area affected by the earthquake is vast and often difficult to entirely cover, and the earthquake itself is a sudden event that causes multiple defects simultaneously, that cannot be effectively traced using traditional, manual methods. This article presents an innovative approach to the problem of detecting damage after sudden events by using an interconnected set of deep machine learning models organized in a single pipeline and allowing for easy modification and swapping models seamlessly. Models in the pipeline were trained with a synthetic dataset and were adapted to be further evaluated and used with unmanned aerial vehicles (UAVs) in real-world conditions. Thanks to the methods presented in the article, it is possible to obtain high accuracy in detecting buildings defects, segmenting constructions into their components and estimating their technical condition based on a single drone flight.
{ "abstract": "The area affected by the earthquake is vast and often difficult to entirely\ncover, and the earthquake itself is a sudden event that causes multiple defects\nsimultaneously, that cannot be effectively traced using traditional, manual\nmethods. This article presents an innovative approach to the problem of\ndetecting damage after sudden events by using an interconnected set of deep\nmachine learning models organized in a single pipeline and allowing for easy\nmodification and swapping models seamlessly. Models in the pipeline were\ntrained with a synthetic dataset and were adapted to be further evaluated and\nused with unmanned aerial vehicles (UAVs) in real-world conditions. Thanks to\nthe methods presented in the article, it is possible to obtain high accuracy in\ndetecting buildings defects, segmenting constructions into their components and\nestimating their technical condition based on a single drone flight.", "title": "Computer Vision based inspection on post-earthquake with UAV synthetic dataset", "url": "http://arxiv.org/abs/2210.05282v1" }
null
null
no_new_dataset
admin
null
false
null
436923c6-aa42-45c0-a8a4-7e6e9dbb60e4
null
Validated
2023-10-04 15:19:51.883569
{ "text_length": 1027 }
1no_new_dataset
TITLE: Deep Learning Hyperspectral Pansharpening on large scale PRISMA dataset ABSTRACT: In this work, we assess several deep learning strategies for hyperspectral pansharpening. First, we present a new dataset with a greater extent than any other in the state of the art. This dataset, collected using the ASI PRISMA satellite, covers about 262200 km2, and its heterogeneity is granted by randomly sampling the Earth's soil. Second, we adapted several state of the art approaches based on deep learning to fit PRISMA hyperspectral data and then assessed, quantitatively and qualitatively, the performance in this new scenario. The investigation has included two settings: Reduced Resolution (RR) to evaluate the techniques in a supervised environment and Full Resolution (FR) for a real-world evaluation. The main purpose is the evaluation of the reconstruction fidelity of the considered methods. In both scenarios, for the sake of completeness, we also included machine-learning-free approaches. From this extensive analysis has emerged that data-driven neural network methods outperform machine-learning-free approaches and adapt better to the task of hyperspectral pansharpening, both in RR and FR protocols.
{ "abstract": "In this work, we assess several deep learning strategies for hyperspectral\npansharpening. First, we present a new dataset with a greater extent than any\nother in the state of the art. This dataset, collected using the ASI PRISMA\nsatellite, covers about 262200 km2, and its heterogeneity is granted by\nrandomly sampling the Earth's soil. Second, we adapted several state of the art\napproaches based on deep learning to fit PRISMA hyperspectral data and then\nassessed, quantitatively and qualitatively, the performance in this new\nscenario. The investigation has included two settings: Reduced Resolution (RR)\nto evaluate the techniques in a supervised environment and Full Resolution (FR)\nfor a real-world evaluation. The main purpose is the evaluation of the\nreconstruction fidelity of the considered methods. In both scenarios, for the\nsake of completeness, we also included machine-learning-free approaches. From\nthis extensive analysis has emerged that data-driven neural network methods\noutperform machine-learning-free approaches and adapt better to the task of\nhyperspectral pansharpening, both in RR and FR protocols.", "title": "Deep Learning Hyperspectral Pansharpening on large scale PRISMA dataset", "url": "http://arxiv.org/abs/2307.11666v2" }
null
null
new_dataset
admin
null
false
null
0bab9564-1361-40ed-a693-9e6f87deeda9
null
Validated
2023-10-04 15:19:51.865000
{ "text_length": 1230 }
0new_dataset
TITLE: On the Role of Dataset Quality and Heterogeneity in Model Confidence ABSTRACT: Safety-critical applications require machine learning models that output accurate and calibrated probabilities. While uncalibrated deep networks are known to make over-confident predictions, it is unclear how model confidence is impacted by the variations in the data, such as label noise or class size. In this paper, we investigate the role of the dataset quality by studying the impact of dataset size and the label noise on the model confidence. We theoretically explain and experimentally demonstrate that, surprisingly, label noise in the training data leads to under-confident networks, while reduced dataset size leads to over-confident models. We then study the impact of dataset heterogeneity, where data quality varies across classes, on model confidence. We demonstrate that this leads to heterogenous confidence/accuracy behavior in the test data and is poorly handled by the standard calibration algorithms. To overcome this, we propose an intuitive heterogenous calibration technique and show that the proposed approach leads to improved calibration metrics (both average and worst-case errors) on the CIFAR datasets.
{ "abstract": "Safety-critical applications require machine learning models that output\naccurate and calibrated probabilities. While uncalibrated deep networks are\nknown to make over-confident predictions, it is unclear how model confidence is\nimpacted by the variations in the data, such as label noise or class size. In\nthis paper, we investigate the role of the dataset quality by studying the\nimpact of dataset size and the label noise on the model confidence. We\ntheoretically explain and experimentally demonstrate that, surprisingly, label\nnoise in the training data leads to under-confident networks, while reduced\ndataset size leads to over-confident models. We then study the impact of\ndataset heterogeneity, where data quality varies across classes, on model\nconfidence. We demonstrate that this leads to heterogenous confidence/accuracy\nbehavior in the test data and is poorly handled by the standard calibration\nalgorithms. To overcome this, we propose an intuitive heterogenous calibration\ntechnique and show that the proposed approach leads to improved calibration\nmetrics (both average and worst-case errors) on the CIFAR datasets.", "title": "On the Role of Dataset Quality and Heterogeneity in Model Confidence", "url": "http://arxiv.org/abs/2002.09831v1" }
null
null
no_new_dataset
admin
null
false
null
6556831b-211d-4fa6-a255-53d603bba65a
null
Validated
2023-10-04 15:19:51.901554
{ "text_length": 1235 }
1no_new_dataset
TITLE: Dataset of Solution-based Inorganic Materials Synthesis Recipes Extracted from the Scientific Literature ABSTRACT: The development of a materials synthesis route is usually based on heuristics and experience. A possible new approach would be to apply data-driven approaches to learn the patterns of synthesis from past experience and use them to predict the syntheses of novel materials. However, this route is impeded by the lack of a large-scale database of synthesis formulations. In this work, we applied advanced machine learning and natural language processing techniques to construct a dataset of 35,675 solution-based synthesis "recipes" extracted from the scientific literature. Each recipe contains essential synthesis information including the precursors and target materials, their quantities, and the synthesis actions and corresponding attributes. Every recipe is also augmented with the reaction formula. Through this work, we are making freely available the first large dataset of solution-based inorganic materials synthesis recipes.
{ "abstract": "The development of a materials synthesis route is usually based on heuristics\nand experience. A possible new approach would be to apply data-driven\napproaches to learn the patterns of synthesis from past experience and use them\nto predict the syntheses of novel materials. However, this route is impeded by\nthe lack of a large-scale database of synthesis formulations. In this work, we\napplied advanced machine learning and natural language processing techniques to\nconstruct a dataset of 35,675 solution-based synthesis \"recipes\" extracted from\nthe scientific literature. Each recipe contains essential synthesis information\nincluding the precursors and target materials, their quantities, and the\nsynthesis actions and corresponding attributes. Every recipe is also augmented\nwith the reaction formula. Through this work, we are making freely available\nthe first large dataset of solution-based inorganic materials synthesis\nrecipes.", "title": "Dataset of Solution-based Inorganic Materials Synthesis Recipes Extracted from the Scientific Literature", "url": "http://arxiv.org/abs/2111.10874v1" }
null
null
new_dataset
admin
null
false
null
7ce010f4-a197-4d8f-840b-0d1150e371d9
null
Validated
2023-10-04 15:19:51.889692
{ "text_length": 1074 }
0new_dataset
TITLE: Interpretable Models Capable of Handling Systematic Missingness in Imbalanced Classes and Heterogeneous Datasets ABSTRACT: Application of interpretable machine learning techniques on medical datasets facilitate early and fast diagnoses, along with getting deeper insight into the data. Furthermore, the transparency of these models increase trust among application domain experts. Medical datasets face common issues such as heterogeneous measurements, imbalanced classes with limited sample size, and missing data, which hinder the straightforward application of machine learning techniques. In this paper we present a family of prototype-based (PB) interpretable models which are capable of handling these issues. The models introduced in this contribution show comparable or superior performance to alternative techniques applicable in such situations. However, unlike ensemble based models, which have to compromise on easy interpretation, the PB models here do not. Moreover we propose a strategy of harnessing the power of ensembles while maintaining the intrinsic interpretability of the PB models, by averaging the model parameter manifolds. All the models were evaluated on a synthetic (publicly available dataset) in addition to detailed analyses of two real-world medical datasets (one publicly available). Results indicated that the models and strategies we introduced addressed the challenges of real-world medical data, while remaining computationally inexpensive and transparent, as well as similar or superior in performance compared to their alternatives.
{ "abstract": "Application of interpretable machine learning techniques on medical datasets\nfacilitate early and fast diagnoses, along with getting deeper insight into the\ndata. Furthermore, the transparency of these models increase trust among\napplication domain experts. Medical datasets face common issues such as\nheterogeneous measurements, imbalanced classes with limited sample size, and\nmissing data, which hinder the straightforward application of machine learning\ntechniques. In this paper we present a family of prototype-based (PB)\ninterpretable models which are capable of handling these issues. The models\nintroduced in this contribution show comparable or superior performance to\nalternative techniques applicable in such situations. However, unlike ensemble\nbased models, which have to compromise on easy interpretation, the PB models\nhere do not. Moreover we propose a strategy of harnessing the power of\nensembles while maintaining the intrinsic interpretability of the PB models, by\naveraging the model parameter manifolds. All the models were evaluated on a\nsynthetic (publicly available dataset) in addition to detailed analyses of two\nreal-world medical datasets (one publicly available). Results indicated that\nthe models and strategies we introduced addressed the challenges of real-world\nmedical data, while remaining computationally inexpensive and transparent, as\nwell as similar or superior in performance compared to their alternatives.", "title": "Interpretable Models Capable of Handling Systematic Missingness in Imbalanced Classes and Heterogeneous Datasets", "url": "http://arxiv.org/abs/2206.02056v1" }
null
null
no_new_dataset
admin
null
false
null
28ba8e6a-d02a-4700-a7eb-46a3face3530
null
Validated
2023-10-04 15:19:51.886014
{ "text_length": 1596 }
1no_new_dataset
TITLE: Towards Accelerated Localization Performance Across Indoor Positioning Datasets ABSTRACT: The localization speed and accuracy in the indoor scenario can greatly impact the Quality of Experience of the user. While many individual machine learning models can achieve comparable positioning performance, their prediction mechanisms offer different complexity to the system. In this work, we propose a fingerprinting positioning method for multi-building and multi-floor deployments, composed of a cascade of three models for building classification, floor classification, and 2D localization regression. We conduct an exhaustive search for the optimally performing one in each step of the cascade while validating on 14 different openly available datasets. As a result, we bring forward the best-performing combination of models in terms of overall positioning accuracy and processing speed and evaluate on independent sets of samples. We reduce the mean prediction time by 71% while achieving comparable positioning performance across all considered datasets. Moreover, in case of voluminous training dataset, the prediction time is reduced down to 1% of the benchmark's.
{ "abstract": "The localization speed and accuracy in the indoor scenario can greatly impact\nthe Quality of Experience of the user. While many individual machine learning\nmodels can achieve comparable positioning performance, their prediction\nmechanisms offer different complexity to the system. In this work, we propose a\nfingerprinting positioning method for multi-building and multi-floor\ndeployments, composed of a cascade of three models for building classification,\nfloor classification, and 2D localization regression. We conduct an exhaustive\nsearch for the optimally performing one in each step of the cascade while\nvalidating on 14 different openly available datasets. As a result, we bring\nforward the best-performing combination of models in terms of overall\npositioning accuracy and processing speed and evaluate on independent sets of\nsamples. We reduce the mean prediction time by 71% while achieving comparable\npositioning performance across all considered datasets. Moreover, in case of\nvoluminous training dataset, the prediction time is reduced down to 1% of the\nbenchmark's.", "title": "Towards Accelerated Localization Performance Across Indoor Positioning Datasets", "url": "http://arxiv.org/abs/2204.10788v1" }
null
null
no_new_dataset
admin
null
false
null
868b2c6c-2bb8-4a26-abd1-7206bc592d8c
null
Validated
2023-10-04 15:19:51.886915
{ "text_length": 1193 }
1no_new_dataset
TITLE: Synthetic Dataset Generation of Driver Telematics ABSTRACT: This article describes techniques employed in the production of a synthetic dataset of driver telematics emulated from a similar real insurance dataset. The synthetic dataset generated has 100,000 policies that included observations about driver's claims experience together with associated classical risk variables and telematics-related variables. This work is aimed to produce a resource that can be used to advance models to assess risks for usage-based insurance. It follows a three-stage process using machine learning algorithms. The first stage is simulating values for the number of claims as multiple binary classifications applying feedforward neural networks. The second stage is simulating values for aggregated amount of claims as regression using feedforward neural networks, with number of claims included in the set of feature variables. In the final stage, a synthetic portfolio of the space of feature variables is generated applying an extended $\texttt{SMOTE}$ algorithm. The resulting dataset is evaluated by comparing the synthetic and real datasets when Poisson and gamma regression models are fitted to the respective data. Other visualization and data summarization produce remarkable similar statistics between the two datasets. We hope that researchers interested in obtaining telematics datasets to calibrate models or learning algorithms will find our work valuable.
{ "abstract": "This article describes techniques employed in the production of a synthetic\ndataset of driver telematics emulated from a similar real insurance dataset.\nThe synthetic dataset generated has 100,000 policies that included observations\nabout driver's claims experience together with associated classical risk\nvariables and telematics-related variables. This work is aimed to produce a\nresource that can be used to advance models to assess risks for usage-based\ninsurance. It follows a three-stage process using machine learning algorithms.\nThe first stage is simulating values for the number of claims as multiple\nbinary classifications applying feedforward neural networks. The second stage\nis simulating values for aggregated amount of claims as regression using\nfeedforward neural networks, with number of claims included in the set of\nfeature variables. In the final stage, a synthetic portfolio of the space of\nfeature variables is generated applying an extended $\\texttt{SMOTE}$ algorithm.\nThe resulting dataset is evaluated by comparing the synthetic and real datasets\nwhen Poisson and gamma regression models are fitted to the respective data.\nOther visualization and data summarization produce remarkable similar\nstatistics between the two datasets. We hope that researchers interested in\nobtaining telematics datasets to calibrate models or learning algorithms will\nfind our work valuable.", "title": "Synthetic Dataset Generation of Driver Telematics", "url": "http://arxiv.org/abs/2102.00252v1" }
null
null
no_new_dataset
admin
null
false
null
68e1c39d-6d31-4a6c-abfa-8099ad6d7fc0
null
Default
2023-10-04 15:19:51.896107
{ "text_length": 1480 }
1no_new_dataset
TITLE: XLCoST: A Benchmark Dataset for Cross-lingual Code Intelligence ABSTRACT: Recent advances in machine learning have significantly improved the understanding of source code data and achieved good performance on a number of downstream tasks. Open source repositories like GitHub enable this process with rich unlabeled code data. However, the lack of high quality labeled data has largely hindered the progress of several code related tasks, such as program translation, summarization, synthesis, and code search. This paper introduces XLCoST, Cross-Lingual Code SnippeT dataset, a new benchmark dataset for cross-lingual code intelligence. Our dataset contains fine-grained parallel data from 8 languages (7 commonly used programming languages and English), and supports 10 cross-lingual code tasks. To the best of our knowledge, it is the largest parallel dataset for source code both in terms of size and the number of languages. We also provide the performance of several state-of-the-art baseline models for each task. We believe this new dataset can be a valuable asset for the research community and facilitate the development and validation of new methods for cross-lingual code intelligence.
{ "abstract": "Recent advances in machine learning have significantly improved the\nunderstanding of source code data and achieved good performance on a number of\ndownstream tasks. Open source repositories like GitHub enable this process with\nrich unlabeled code data. However, the lack of high quality labeled data has\nlargely hindered the progress of several code related tasks, such as program\ntranslation, summarization, synthesis, and code search. This paper introduces\nXLCoST, Cross-Lingual Code SnippeT dataset, a new benchmark dataset for\ncross-lingual code intelligence. Our dataset contains fine-grained parallel\ndata from 8 languages (7 commonly used programming languages and English), and\nsupports 10 cross-lingual code tasks. To the best of our knowledge, it is the\nlargest parallel dataset for source code both in terms of size and the number\nof languages. We also provide the performance of several state-of-the-art\nbaseline models for each task. We believe this new dataset can be a valuable\nasset for the research community and facilitate the development and validation\nof new methods for cross-lingual code intelligence.", "title": "XLCoST: A Benchmark Dataset for Cross-lingual Code Intelligence", "url": "http://arxiv.org/abs/2206.08474v1" }
null
null
new_dataset
admin
null
false
null
a105b929-87c4-4981-99c6-4d6687c61b36
null
Validated
2023-10-04 15:19:51.885749
{ "text_length": 1221 }
0new_dataset
TITLE: A Puzzle-Based Dataset for Natural Language Inference ABSTRACT: We provide here a dataset for tasks related to natural language understanding and natural language inference. The dataset contains logical puzzles in natural language from three domains: comparing puzzles, knighs and knaves, and zebra puzzles. Each puzzle is associated with the entire set of atomic questions that can be generated based on the relations and individuals occurring in the text. For each question we provide the correct answer: entailment, contradiction or ambiguity. The answer's correctness is verified against theorem provers. Good puzzles have two properties: (i) each piece of information is necessary and (ii) no unnecessary information is provided. These properties make puzzles interesting candidates for machine comprehension tasks.
{ "abstract": "We provide here a dataset for tasks related to natural language understanding\nand natural language inference. The dataset contains logical puzzles in natural\nlanguage from three domains: comparing puzzles, knighs and knaves, and zebra\npuzzles. Each puzzle is associated with the entire set of atomic questions that\ncan be generated based on the relations and individuals occurring in the text.\nFor each question we provide the correct answer: entailment, contradiction or\nambiguity. The answer's correctness is verified against theorem provers. Good\npuzzles have two properties: (i) each piece of information is necessary and\n(ii) no unnecessary information is provided. These properties make puzzles\ninteresting candidates for machine comprehension tasks.", "title": "A Puzzle-Based Dataset for Natural Language Inference", "url": "http://arxiv.org/abs/2112.05742v1" }
null
null
new_dataset
admin
null
false
null
522fd1f6-1359-45dd-a8f2-da7e01993850
null
Validated
2023-10-04 15:19:51.889322
{ "text_length": 844 }
0new_dataset
TITLE: A Benchmark dataset for predictive maintenance ABSTRACT: The paper describes the MetroPT data set, an outcome of a eXplainable Predictive Maintenance (XPM) project with an urban metro public transportation service in Porto, Portugal. The data was collected in 2022 that aimed to evaluate machine learning methods for online anomaly detection and failure prediction. By capturing several analogic sensor signals (pressure, temperature, current consumption), digital signals (control signals, discrete signals), and GPS information (latitude, longitude, and speed), we provide a dataset that can be easily used to evaluate online machine learning methods. This dataset contains some interesting characteristics and can be a good benchmark for predictive maintenance models.
{ "abstract": "The paper describes the MetroPT data set, an outcome of a eXplainable\nPredictive Maintenance (XPM) project with an urban metro public transportation\nservice in Porto, Portugal. The data was collected in 2022 that aimed to\nevaluate machine learning methods for online anomaly detection and failure\nprediction. By capturing several analogic sensor signals (pressure,\ntemperature, current consumption), digital signals (control signals, discrete\nsignals), and GPS information (latitude, longitude, and speed), we provide a\ndataset that can be easily used to evaluate online machine learning methods.\nThis dataset contains some interesting characteristics and can be a good\nbenchmark for predictive maintenance models.", "title": "A Benchmark dataset for predictive maintenance", "url": "http://arxiv.org/abs/2207.05466v3" }
null
null
new_dataset
admin
null
false
null
35d2af1e-7208-481b-b3eb-491c6f840920
null
Validated
2023-10-04 15:19:51.885344
{ "text_length": 795 }
0new_dataset
TITLE: CLEVR-X: A Visual Reasoning Dataset for Natural Language Explanations ABSTRACT: Providing explanations in the context of Visual Question Answering (VQA) presents a fundamental problem in machine learning. To obtain detailed insights into the process of generating natural language explanations for VQA, we introduce the large-scale CLEVR-X dataset that extends the CLEVR dataset with natural language explanations. For each image-question pair in the CLEVR dataset, CLEVR-X contains multiple structured textual explanations which are derived from the original scene graphs. By construction, the CLEVR-X explanations are correct and describe the reasoning and visual information that is necessary to answer a given question. We conducted a user study to confirm that the ground-truth explanations in our proposed dataset are indeed complete and relevant. We present baseline results for generating natural language explanations in the context of VQA using two state-of-the-art frameworks on the CLEVR-X dataset. Furthermore, we provide a detailed analysis of the explanation generation quality for different question and answer types. Additionally, we study the influence of using different numbers of ground-truth explanations on the convergence of natural language generation (NLG) metrics. The CLEVR-X dataset is publicly available at \url{https://explainableml.github.io/CLEVR-X/}.
{ "abstract": "Providing explanations in the context of Visual Question Answering (VQA)\npresents a fundamental problem in machine learning. To obtain detailed insights\ninto the process of generating natural language explanations for VQA, we\nintroduce the large-scale CLEVR-X dataset that extends the CLEVR dataset with\nnatural language explanations. For each image-question pair in the CLEVR\ndataset, CLEVR-X contains multiple structured textual explanations which are\nderived from the original scene graphs. By construction, the CLEVR-X\nexplanations are correct and describe the reasoning and visual information that\nis necessary to answer a given question. We conducted a user study to confirm\nthat the ground-truth explanations in our proposed dataset are indeed complete\nand relevant. We present baseline results for generating natural language\nexplanations in the context of VQA using two state-of-the-art frameworks on the\nCLEVR-X dataset. Furthermore, we provide a detailed analysis of the explanation\ngeneration quality for different question and answer types. Additionally, we\nstudy the influence of using different numbers of ground-truth explanations on\nthe convergence of natural language generation (NLG) metrics. The CLEVR-X\ndataset is publicly available at\n\\url{https://explainableml.github.io/CLEVR-X/}.", "title": "CLEVR-X: A Visual Reasoning Dataset for Natural Language Explanations", "url": "http://arxiv.org/abs/2204.02380v1" }
null
null
new_dataset
admin
null
false
null
4379ec49-a7ee-420f-b803-9ab7a27bc0a4
null
Validated
2023-10-04 15:19:51.887215
{ "text_length": 1408 }
0new_dataset
TITLE: dMelodies: A Music Dataset for Disentanglement Learning ABSTRACT: Representation learning focused on disentangling the underlying factors of variation in given data has become an important area of research in machine learning. However, most of the studies in this area have relied on datasets from the computer vision domain and thus, have not been readily extended to music. In this paper, we present a new symbolic music dataset that will help researchers working on disentanglement problems demonstrate the efficacy of their algorithms on diverse domains. This will also provide a means for evaluating algorithms specifically designed for music. To this end, we create a dataset comprising of 2-bar monophonic melodies where each melody is the result of a unique combination of nine latent factors that span ordinal, categorical, and binary types. The dataset is large enough (approx. 1.3 million data points) to train and test deep networks for disentanglement learning. In addition, we present benchmarking experiments using popular unsupervised disentanglement algorithms on this dataset and compare the results with those obtained on an image-based dataset.
{ "abstract": "Representation learning focused on disentangling the underlying factors of\nvariation in given data has become an important area of research in machine\nlearning. However, most of the studies in this area have relied on datasets\nfrom the computer vision domain and thus, have not been readily extended to\nmusic. In this paper, we present a new symbolic music dataset that will help\nresearchers working on disentanglement problems demonstrate the efficacy of\ntheir algorithms on diverse domains. This will also provide a means for\nevaluating algorithms specifically designed for music. To this end, we create a\ndataset comprising of 2-bar monophonic melodies where each melody is the result\nof a unique combination of nine latent factors that span ordinal, categorical,\nand binary types. The dataset is large enough (approx. 1.3 million data points)\nto train and test deep networks for disentanglement learning. In addition, we\npresent benchmarking experiments using popular unsupervised disentanglement\nalgorithms on this dataset and compare the results with those obtained on an\nimage-based dataset.", "title": "dMelodies: A Music Dataset for Disentanglement Learning", "url": "http://arxiv.org/abs/2007.15067v1" }
null
null
new_dataset
admin
null
false
null
65fa59c1-f125-48d0-be83-120ea19693b5
null
Validated
2023-10-04 15:19:51.899021
{ "text_length": 1188 }
0new_dataset
TITLE: Machine Learning Models Evaluation and Feature Importance Analysis on NPL Dataset ABSTRACT: Predicting the probability of non-performing loans for individuals has a vital and beneficial role for banks to decrease credit risk and make the right decisions before giving the loan. The trend to make these decisions are based on credit study and in accordance with generally accepted standards, loan payment history, and demographic data of the clients. In this work, we evaluate how different Machine learning models such as Random Forest, Decision tree, KNN, SVM, and XGBoost perform on the dataset provided by a private bank in Ethiopia. Further, motivated by this evaluation we explore different feature selection methods to state the important features for the bank. Our findings show that XGBoost achieves the highest F1 score on the KMeans SMOTE over-sampled data. We also found that the most important features are the age of the applicant, years of employment, and total income of the applicant rather than collateral-related features in evaluating credit risk.
{ "abstract": "Predicting the probability of non-performing loans for individuals has a\nvital and beneficial role for banks to decrease credit risk and make the right\ndecisions before giving the loan. The trend to make these decisions are based\non credit study and in accordance with generally accepted standards, loan\npayment history, and demographic data of the clients. In this work, we evaluate\nhow different Machine learning models such as Random Forest, Decision tree,\nKNN, SVM, and XGBoost perform on the dataset provided by a private bank in\nEthiopia. Further, motivated by this evaluation we explore different feature\nselection methods to state the important features for the bank. Our findings\nshow that XGBoost achieves the highest F1 score on the KMeans SMOTE\nover-sampled data. We also found that the most important features are the age\nof the applicant, years of employment, and total income of the applicant rather\nthan collateral-related features in evaluating credit risk.", "title": "Machine Learning Models Evaluation and Feature Importance Analysis on NPL Dataset", "url": "http://arxiv.org/abs/2209.09638v1" }
null
null
no_new_dataset
admin
null
false
null
f7f340d8-ba7b-4189-9251-379eb2f0a28b
null
Validated
2023-10-04 15:19:51.884489
{ "text_length": 1090 }
1no_new_dataset
TITLE: Domain Adaptation in Highly Imbalanced and Overlapping Datasets ABSTRACT: In many machine learning domains, datasets are characterized by highly imbalanced and overlapping classes. Particularly in the medical domain, a specific list of symptoms can be labeled as one of various different conditions. Some of these conditions may be more prevalent than others by several orders of magnitude. Here we present a novel unsupervised domain adaptation scheme for such datasets. The scheme, based on a specific type of Quantification, is designed to work under both label and conditional shifts. It is demonstrated on datasets generated from electronic health records and provides high quality results for both Quantification and Domain Adaptation in very challenging scenarios. Potential benefits of using this scheme in the current COVID-19 outbreak, for estimation of prevalence and probability of infection are discussed.
{ "abstract": "In many machine learning domains, datasets are characterized by highly\nimbalanced and overlapping classes. Particularly in the medical domain, a\nspecific list of symptoms can be labeled as one of various different\nconditions. Some of these conditions may be more prevalent than others by\nseveral orders of magnitude. Here we present a novel unsupervised domain\nadaptation scheme for such datasets. The scheme, based on a specific type of\nQuantification, is designed to work under both label and conditional shifts. It\nis demonstrated on datasets generated from electronic health records and\nprovides high quality results for both Quantification and Domain Adaptation in\nvery challenging scenarios. Potential benefits of using this scheme in the\ncurrent COVID-19 outbreak, for estimation of prevalence and probability of\ninfection are discussed.", "title": "Domain Adaptation in Highly Imbalanced and Overlapping Datasets", "url": "http://arxiv.org/abs/2005.03585v2" }
null
null
no_new_dataset
admin
null
false
null
0c76ed2a-f4d2-4669-acf2-b34fbe962854
null
Validated
2023-10-04 15:19:51.900142
{ "text_length": 942 }
1no_new_dataset
TITLE: KazNERD: Kazakh Named Entity Recognition Dataset ABSTRACT: We present the development of a dataset for Kazakh named entity recognition. The dataset was built as there is a clear need for publicly available annotated corpora in Kazakh, as well as annotation guidelines containing straightforward--but rigorous--rules and examples. The dataset annotation, based on the IOB2 scheme, was carried out on television news text by two native Kazakh speakers under the supervision of the first author. The resulting dataset contains 112,702 sentences and 136,333 annotations for 25 entity classes. State-of-the-art machine learning models to automatise Kazakh named entity recognition were also built, with the best-performing model achieving an exact match F1-score of 97.22% on the test set. The annotated dataset, guidelines, and codes used to train the models are freely available for download under the CC BY 4.0 licence from https://github.com/IS2AI/KazNERD.
{ "abstract": "We present the development of a dataset for Kazakh named entity recognition.\nThe dataset was built as there is a clear need for publicly available annotated\ncorpora in Kazakh, as well as annotation guidelines containing\nstraightforward--but rigorous--rules and examples. The dataset annotation,\nbased on the IOB2 scheme, was carried out on television news text by two native\nKazakh speakers under the supervision of the first author. The resulting\ndataset contains 112,702 sentences and 136,333 annotations for 25 entity\nclasses. State-of-the-art machine learning models to automatise Kazakh named\nentity recognition were also built, with the best-performing model achieving an\nexact match F1-score of 97.22% on the test set. The annotated dataset,\nguidelines, and codes used to train the models are freely available for\ndownload under the CC BY 4.0 licence from https://github.com/IS2AI/KazNERD.", "title": "KazNERD: Kazakh Named Entity Recognition Dataset", "url": "http://arxiv.org/abs/2111.13419v2" }
null
null
new_dataset
admin
null
false
null
fe537a85-42f2-4ce7-ad75-3a796caa8343
null
Validated
2023-10-04 15:19:51.889620
{ "text_length": 979 }
0new_dataset
TITLE: Benchmarks for Pirá 2.0, a Reading Comprehension Dataset about the Ocean, the Brazilian Coast, and Climate Change ABSTRACT: Pir\'a is a reading comprehension dataset focused on the ocean, the Brazilian coast, and climate change, built from a collection of scientific abstracts and reports on these topics. This dataset represents a versatile language resource, particularly useful for testing the ability of current machine learning models to acquire expert scientific knowledge. Despite its potential, a detailed set of baselines has not yet been developed for Pir\'a. By creating these baselines, researchers can more easily utilize Pir\'a as a resource for testing machine learning models across a wide range of question answering tasks. In this paper, we define six benchmarks over the Pir\'a dataset, covering closed generative question answering, machine reading comprehension, information retrieval, open question answering, answer triggering, and multiple choice question answering. As part of this effort, we have also produced a curated version of the original dataset, where we fixed a number of grammar issues, repetitions, and other shortcomings. Furthermore, the dataset has been extended in several new directions, so as to face the aforementioned benchmarks: translation of supporting texts from English into Portuguese, classification labels for answerability, automatic paraphrases of questions and answers, and multiple choice candidates. The results described in this paper provide several points of reference for researchers interested in exploring the challenges provided by the Pir\'a dataset.
{ "abstract": "Pir\\'a is a reading comprehension dataset focused on the ocean, the Brazilian\ncoast, and climate change, built from a collection of scientific abstracts and\nreports on these topics. This dataset represents a versatile language resource,\nparticularly useful for testing the ability of current machine learning models\nto acquire expert scientific knowledge. Despite its potential, a detailed set\nof baselines has not yet been developed for Pir\\'a. By creating these\nbaselines, researchers can more easily utilize Pir\\'a as a resource for testing\nmachine learning models across a wide range of question answering tasks. In\nthis paper, we define six benchmarks over the Pir\\'a dataset, covering closed\ngenerative question answering, machine reading comprehension, information\nretrieval, open question answering, answer triggering, and multiple choice\nquestion answering. As part of this effort, we have also produced a curated\nversion of the original dataset, where we fixed a number of grammar issues,\nrepetitions, and other shortcomings. Furthermore, the dataset has been extended\nin several new directions, so as to face the aforementioned benchmarks:\ntranslation of supporting texts from English into Portuguese, classification\nlabels for answerability, automatic paraphrases of questions and answers, and\nmultiple choice candidates. The results described in this paper provide several\npoints of reference for researchers interested in exploring the challenges\nprovided by the Pir\\'a dataset.", "title": "Benchmarks for Pirá 2.0, a Reading Comprehension Dataset about the Ocean, the Brazilian Coast, and Climate Change", "url": "http://arxiv.org/abs/2309.10945v1" }
null
null
new_dataset
admin
null
false
null
bb978cb7-7a24-49c1-8d32-95589ab33e54
null
Validated
2023-10-04 15:19:51.863420
{ "text_length": 1640 }
0new_dataset
TITLE: SUTD-PRCM Dataset and Neural Architecture Search Approach for Complex Metasurface Design ABSTRACT: Metasurfaces have received a lot of attentions recently due to their versatile capability in manipulating electromagnetic wave. Advanced designs to satisfy multiple objectives with non-linear constraints have motivated researchers in using machine learning (ML) techniques like deep learning (DL) for accelerated design of metasurfaces. For metasurfaces, it is difficult to make quantitative comparisons between different ML models without having a common and yet complex dataset used in many disciplines like image classification. Many studies were directed to a relatively constrained datasets that are limited to specified patterns or shapes in metasurfaces. In this paper, we present our SUTD polarized reflection of complex metasurfaces (SUTD-PRCM) dataset, which contains approximately 260,000 samples of complex metasurfaces created from electromagnetic simulation, and it has been used to benchmark our DL models. The metasurface patterns are divided into different classes to facilitate different degree of complexity, which involves identifying and exploiting the relationship between the patterns and the electromagnetic responses that can be compared in using different DL models. With the release of this SUTD-PRCM dataset, we hope that it will be useful for benchmarking existing or future DL models developed in the ML community. We also propose a classification problem that is less encountered and apply neural architecture search to have a preliminary understanding of potential modification to the neural architecture that will improve the prediction by DL models. Our finding shows that convolution stacking is not the dominant element of the neural architecture anymore, which implies that low-level features are preferred over the traditional deep hierarchical high-level features thus explains why deep convolutional neural network based models are not performing well in our dataset.
{ "abstract": "Metasurfaces have received a lot of attentions recently due to their\nversatile capability in manipulating electromagnetic wave. Advanced designs to\nsatisfy multiple objectives with non-linear constraints have motivated\nresearchers in using machine learning (ML) techniques like deep learning (DL)\nfor accelerated design of metasurfaces. For metasurfaces, it is difficult to\nmake quantitative comparisons between different ML models without having a\ncommon and yet complex dataset used in many disciplines like image\nclassification. Many studies were directed to a relatively constrained datasets\nthat are limited to specified patterns or shapes in metasurfaces. In this\npaper, we present our SUTD polarized reflection of complex metasurfaces\n(SUTD-PRCM) dataset, which contains approximately 260,000 samples of complex\nmetasurfaces created from electromagnetic simulation, and it has been used to\nbenchmark our DL models. The metasurface patterns are divided into different\nclasses to facilitate different degree of complexity, which involves\nidentifying and exploiting the relationship between the patterns and the\nelectromagnetic responses that can be compared in using different DL models.\nWith the release of this SUTD-PRCM dataset, we hope that it will be useful for\nbenchmarking existing or future DL models developed in the ML community. We\nalso propose a classification problem that is less encountered and apply neural\narchitecture search to have a preliminary understanding of potential\nmodification to the neural architecture that will improve the prediction by DL\nmodels. Our finding shows that convolution stacking is not the dominant element\nof the neural architecture anymore, which implies that low-level features are\npreferred over the traditional deep hierarchical high-level features thus\nexplains why deep convolutional neural network based models are not performing\nwell in our dataset.", "title": "SUTD-PRCM Dataset and Neural Architecture Search Approach for Complex Metasurface Design", "url": "http://arxiv.org/abs/2203.00002v1" }
null
null
new_dataset
admin
null
false
null
a8552922-7488-4ac2-bc46-9b2eddfa898e
null
Validated
2023-10-04 15:19:51.888123
{ "text_length": 2030 }
0new_dataset
TITLE: MetaGraspNet_v0: A Large-Scale Benchmark Dataset for Vision-driven Robotic Grasping via Physics-based Metaverse Synthesis ABSTRACT: There has been increasing interest in smart factories powered by robotics systems to tackle repetitive, laborious tasks. One impactful yet challenging task in robotics-powered smart factory applications is robotic grasping: using robotic arms to grasp objects autonomously in different settings. Robotic grasping requires a variety of computer vision tasks such as object detection, segmentation, grasp prediction, pick planning, etc. While significant progress has been made in leveraging of machine learning for robotic grasping, particularly with deep learning, a big challenge remains in the need for large-scale, high-quality RGBD datasets that cover a wide diversity of scenarios and permutations. To tackle this big, diverse data problem, we are inspired by the recent rise in the concept of metaverse, which has greatly closed the gap between virtual worlds and the physical world. Metaverses allow us to create digital twins of real-world manufacturing scenarios and to virtually create different scenarios from which large volumes of data can be generated for training models. In this paper, we present MetaGraspNet: a large-scale benchmark dataset for vision-driven robotic grasping via physics-based metaverse synthesis. The proposed dataset contains 100,000 images and 25 different object types and is split into 5 difficulties to evaluate object detection and segmentation model performance in different grasping scenarios. We also propose a new layout-weighted performance metric alongside the dataset for evaluating object detection and segmentation performance in a manner that is more appropriate for robotic grasp applications compared to existing general-purpose performance metrics. Our benchmark dataset is available open-source on Kaggle, with the first phase consisting of detailed object detection, segmentation, layout annotations, and a layout-weighted performance metric script.
{ "abstract": "There has been increasing interest in smart factories powered by robotics\nsystems to tackle repetitive, laborious tasks. One impactful yet challenging\ntask in robotics-powered smart factory applications is robotic grasping: using\nrobotic arms to grasp objects autonomously in different settings. Robotic\ngrasping requires a variety of computer vision tasks such as object detection,\nsegmentation, grasp prediction, pick planning, etc. While significant progress\nhas been made in leveraging of machine learning for robotic grasping,\nparticularly with deep learning, a big challenge remains in the need for\nlarge-scale, high-quality RGBD datasets that cover a wide diversity of\nscenarios and permutations. To tackle this big, diverse data problem, we are\ninspired by the recent rise in the concept of metaverse, which has greatly\nclosed the gap between virtual worlds and the physical world. Metaverses allow\nus to create digital twins of real-world manufacturing scenarios and to\nvirtually create different scenarios from which large volumes of data can be\ngenerated for training models. In this paper, we present MetaGraspNet: a\nlarge-scale benchmark dataset for vision-driven robotic grasping via\nphysics-based metaverse synthesis. The proposed dataset contains 100,000 images\nand 25 different object types and is split into 5 difficulties to evaluate\nobject detection and segmentation model performance in different grasping\nscenarios. We also propose a new layout-weighted performance metric alongside\nthe dataset for evaluating object detection and segmentation performance in a\nmanner that is more appropriate for robotic grasp applications compared to\nexisting general-purpose performance metrics. Our benchmark dataset is\navailable open-source on Kaggle, with the first phase consisting of detailed\nobject detection, segmentation, layout annotations, and a layout-weighted\nperformance metric script.", "title": "MetaGraspNet_v0: A Large-Scale Benchmark Dataset for Vision-driven Robotic Grasping via Physics-based Metaverse Synthesis", "url": "http://arxiv.org/abs/2112.14663v3" }
null
null
new_dataset
admin
null
false
null
771e2acc-6f67-4aba-87fe-53f097b1db2b
null
Validated
2023-10-04 15:19:51.889047
{ "text_length": 2062 }
0new_dataset
TITLE: A Suite of Fairness Datasets for Tabular Classification ABSTRACT: There have been many papers with algorithms for improving fairness of machine-learning classifiers for tabular data. Unfortunately, most use only very few datasets for their experimental evaluation. We introduce a suite of functions for fetching 20 fairness datasets and providing associated fairness metadata. Hopefully, these will lead to more rigorous experimental evaluations in future fairness-aware machine learning research.
{ "abstract": "There have been many papers with algorithms for improving fairness of\nmachine-learning classifiers for tabular data. Unfortunately, most use only\nvery few datasets for their experimental evaluation. We introduce a suite of\nfunctions for fetching 20 fairness datasets and providing associated fairness\nmetadata. Hopefully, these will lead to more rigorous experimental evaluations\nin future fairness-aware machine learning research.", "title": "A Suite of Fairness Datasets for Tabular Classification", "url": "http://arxiv.org/abs/2308.00133v1" }
null
null
no_new_dataset
admin
null
false
null
b7b3c206-e26f-46b8-8932-7838812ba1a1
null
Validated
2023-10-04 15:19:51.864569
{ "text_length": 521 }
1no_new_dataset
TITLE: MADLAD-400: A Multilingual And Document-Level Large Audited Dataset ABSTRACT: We introduce MADLAD-400, a manually audited, general domain 3T token monolingual dataset based on CommonCrawl, spanning 419 languages. We discuss the limitations revealed by self-auditing MADLAD-400, and the role data auditing had in the dataset creation process. We then train and release a 10.7B-parameter multilingual machine translation model on 250 billion tokens covering over 450 languages using publicly available data, and find that it is competitive with models that are significantly larger, and report the results on different domains. In addition, we train a 8B-parameter language model, and assess the results on few-shot translation. We make the baseline models available to the research community.
{ "abstract": "We introduce MADLAD-400, a manually audited, general domain 3T token\nmonolingual dataset based on CommonCrawl, spanning 419 languages. We discuss\nthe limitations revealed by self-auditing MADLAD-400, and the role data\nauditing had in the dataset creation process. We then train and release a\n10.7B-parameter multilingual machine translation model on 250 billion tokens\ncovering over 450 languages using publicly available data, and find that it is\ncompetitive with models that are significantly larger, and report the results\non different domains. In addition, we train a 8B-parameter language model, and\nassess the results on few-shot translation. We make the baseline models\navailable to the research community.", "title": "MADLAD-400: A Multilingual And Document-Level Large Audited Dataset", "url": "http://arxiv.org/abs/2309.04662v1" }
null
null
new_dataset
admin
null
false
null
921cff9c-70e1-4bfa-afe1-ab167527ddfc
null
Validated
2023-10-04 15:19:51.863691
{ "text_length": 815 }
0new_dataset
TITLE: SubseasonalClimateUSA: A Dataset for Subseasonal Forecasting and Benchmarking ABSTRACT: Subseasonal forecasting of the weather two to six weeks in advance is critical for resource allocation and climate adaptation but poses many challenges for the forecasting community. At this forecast horizon, physics-based dynamical models have limited skill, and the targets for prediction depend in a complex manner on both local weather and global climate variables. Recently, machine learning methods have shown promise in advancing the state of the art but only at the cost of complex data curation, integrating expert knowledge with aggregation across multiple relevant data sources, file formats, and temporal and spatial resolutions. To streamline this process and accelerate future development, we introduce SubseasonalClimateUSA, a curated dataset for training and benchmarking subseasonal forecasting models in the United States. We use this dataset to benchmark a diverse suite of subseasonal models, including operational dynamical models, classical meteorological baselines, and ten state-of-the-art machine learning and deep learning-based methods from the literature. Overall, our benchmarks suggest simple and effective ways to extend the accuracy of current operational models. SubseasonalClimateUSA is regularly updated and accessible via the https://github.com/microsoft/subseasonal_data/ Python package.
{ "abstract": "Subseasonal forecasting of the weather two to six weeks in advance is\ncritical for resource allocation and climate adaptation but poses many\nchallenges for the forecasting community. At this forecast horizon,\nphysics-based dynamical models have limited skill, and the targets for\nprediction depend in a complex manner on both local weather and global climate\nvariables. Recently, machine learning methods have shown promise in advancing\nthe state of the art but only at the cost of complex data curation, integrating\nexpert knowledge with aggregation across multiple relevant data sources, file\nformats, and temporal and spatial resolutions. To streamline this process and\naccelerate future development, we introduce SubseasonalClimateUSA, a curated\ndataset for training and benchmarking subseasonal forecasting models in the\nUnited States. We use this dataset to benchmark a diverse suite of subseasonal\nmodels, including operational dynamical models, classical meteorological\nbaselines, and ten state-of-the-art machine learning and deep learning-based\nmethods from the literature. Overall, our benchmarks suggest simple and\neffective ways to extend the accuracy of current operational models.\nSubseasonalClimateUSA is regularly updated and accessible via the\nhttps://github.com/microsoft/subseasonal_data/ Python package.", "title": "SubseasonalClimateUSA: A Dataset for Subseasonal Forecasting and Benchmarking", "url": "http://arxiv.org/abs/2109.10399v3" }
null
null
new_dataset
admin
null
false
null
0bc818ba-e944-48fa-b660-12e49dde2661
null
Validated
2023-10-04 15:19:51.892044
{ "text_length": 1436 }
0new_dataset
TITLE: RenderIH: A Large-scale Synthetic Dataset for 3D Interacting Hand Pose Estimation ABSTRACT: The current interacting hand (IH) datasets are relatively simplistic in terms of background and texture, with hand joints being annotated by a machine annotator, which may result in inaccuracies, and the diversity of pose distribution is limited. However, the variability of background, pose distribution, and texture can greatly influence the generalization ability. Therefore, we present a large-scale synthetic dataset RenderIH for interacting hands with accurate and diverse pose annotations. The dataset contains 1M photo-realistic images with varied backgrounds, perspectives, and hand textures. To generate natural and diverse interacting poses, we propose a new pose optimization algorithm. Additionally, for better pose estimation accuracy, we introduce a transformer-based pose estimation network, TransHand, to leverage the correlation between interacting hands and verify the effectiveness of RenderIH in improving results. Our dataset is model-agnostic and can improve more accuracy of any hand pose estimation method in comparison to other real or synthetic datasets. Experiments have shown that pretraining on our synthetic data can significantly decrease the error from 6.76mm to 5.79mm, and our Transhand surpasses contemporary methods. Our dataset and code are available at https://github.com/adwardlee/RenderIH.
{ "abstract": "The current interacting hand (IH) datasets are relatively simplistic in terms\nof background and texture, with hand joints being annotated by a machine\nannotator, which may result in inaccuracies, and the diversity of pose\ndistribution is limited. However, the variability of background, pose\ndistribution, and texture can greatly influence the generalization ability.\nTherefore, we present a large-scale synthetic dataset RenderIH for interacting\nhands with accurate and diverse pose annotations. The dataset contains 1M\nphoto-realistic images with varied backgrounds, perspectives, and hand\ntextures. To generate natural and diverse interacting poses, we propose a new\npose optimization algorithm. Additionally, for better pose estimation accuracy,\nwe introduce a transformer-based pose estimation network, TransHand, to\nleverage the correlation between interacting hands and verify the effectiveness\nof RenderIH in improving results. Our dataset is model-agnostic and can improve\nmore accuracy of any hand pose estimation method in comparison to other real or\nsynthetic datasets. Experiments have shown that pretraining on our synthetic\ndata can significantly decrease the error from 6.76mm to 5.79mm, and our\nTranshand surpasses contemporary methods. Our dataset and code are available at\nhttps://github.com/adwardlee/RenderIH.", "title": "RenderIH: A Large-scale Synthetic Dataset for 3D Interacting Hand Pose Estimation", "url": "http://arxiv.org/abs/2309.09301v3" }
null
null
new_dataset
admin
null
false
null
bd459f9b-5366-48f6-8f1c-648b42ed945d
null
Validated
2023-10-04 15:19:51.863493
{ "text_length": 1446 }
0new_dataset
TITLE: A Dataset for Speech Emotion Recognition in Greek Theatrical Plays ABSTRACT: Machine learning methodologies can be adopted in cultural applications and propose new ways to distribute or even present the cultural content to the public. For instance, speech analytics can be adopted to automatically generate subtitles in theatrical plays, in order to (among other purposes) help people with hearing loss. Apart from a typical speech-to-text transcription with Automatic Speech Recognition (ASR), Speech Emotion Recognition (SER) can be used to automatically predict the underlying emotional content of speech dialogues in theatrical plays, and thus to provide a deeper understanding how the actors utter their lines. However, real-world datasets from theatrical plays are not available in the literature. In this work we present GreThE, the Greek Theatrical Emotion dataset, a new publicly available data collection for speech emotion recognition in Greek theatrical plays. The dataset contains utterances from various actors and plays, along with respective valence and arousal annotations. Towards this end, multiple annotators have been asked to provide their input for each speech recording and inter-annotator agreement is taken into account in the final ground truth generation. In addition, we discuss the results of some indicative experiments that have been conducted with machine and deep learning frameworks, using the dataset, along with some widely used databases in the field of speech emotion recognition.
{ "abstract": "Machine learning methodologies can be adopted in cultural applications and\npropose new ways to distribute or even present the cultural content to the\npublic. For instance, speech analytics can be adopted to automatically generate\nsubtitles in theatrical plays, in order to (among other purposes) help people\nwith hearing loss. Apart from a typical speech-to-text transcription with\nAutomatic Speech Recognition (ASR), Speech Emotion Recognition (SER) can be\nused to automatically predict the underlying emotional content of speech\ndialogues in theatrical plays, and thus to provide a deeper understanding how\nthe actors utter their lines. However, real-world datasets from theatrical\nplays are not available in the literature. In this work we present GreThE, the\nGreek Theatrical Emotion dataset, a new publicly available data collection for\nspeech emotion recognition in Greek theatrical plays. The dataset contains\nutterances from various actors and plays, along with respective valence and\narousal annotations. Towards this end, multiple annotators have been asked to\nprovide their input for each speech recording and inter-annotator agreement is\ntaken into account in the final ground truth generation. In addition, we\ndiscuss the results of some indicative experiments that have been conducted\nwith machine and deep learning frameworks, using the dataset, along with some\nwidely used databases in the field of speech emotion recognition.", "title": "A Dataset for Speech Emotion Recognition in Greek Theatrical Plays", "url": "http://arxiv.org/abs/2203.15568v1" }
null
null
new_dataset
admin
null
false
null
ae95482e-4071-488e-876e-dc04b5d599e5
null
Validated
2023-10-04 15:19:51.887382
{ "text_length": 1543 }
0new_dataset
TITLE: Confidence-Nets: A Step Towards better Prediction Intervals for regression Neural Networks on small datasets ABSTRACT: The recent decade has seen an enormous rise in the popularity of deep learning and neural networks. These algorithms have broken many previous records and achieved remarkable results. Their outstanding performance has significantly sped up the progress of AI, and so far various milestones have been achieved earlier than expected. However, in the case of relatively small datasets, the performance of Deep Neural Networks (DNN) may suffer from reduced accuracy compared to other Machine Learning models. Furthermore, it is difficult to construct prediction intervals or evaluate the uncertainty of predictions when dealing with regression tasks. In this paper, we propose an ensemble method that attempts to estimate the uncertainty of predictions, increase their accuracy and provide an interval for the expected variation. Compared with traditional DNNs that only provide a prediction, our proposed method can output a prediction interval by combining DNNs, extreme gradient boosting (XGBoost) and dissimilarity computation techniques. Albeit the simple design, this approach significantly increases accuracy on small datasets and does not introduce much complexity to the architecture of the neural network. The proposed method is tested on various datasets, and a significant improvement in the performance of the neural network model is seen. The model's prediction interval can include the ground truth value at an average rate of 71% and 78% across training sizes of 90% and 55%, respectively. Finally, we highlight other aspects and applications of the approach in experimental error estimation, and the application of transfer learning.
{ "abstract": "The recent decade has seen an enormous rise in the popularity of deep\nlearning and neural networks. These algorithms have broken many previous\nrecords and achieved remarkable results. Their outstanding performance has\nsignificantly sped up the progress of AI, and so far various milestones have\nbeen achieved earlier than expected. However, in the case of relatively small\ndatasets, the performance of Deep Neural Networks (DNN) may suffer from reduced\naccuracy compared to other Machine Learning models. Furthermore, it is\ndifficult to construct prediction intervals or evaluate the uncertainty of\npredictions when dealing with regression tasks. In this paper, we propose an\nensemble method that attempts to estimate the uncertainty of predictions,\nincrease their accuracy and provide an interval for the expected variation.\nCompared with traditional DNNs that only provide a prediction, our proposed\nmethod can output a prediction interval by combining DNNs, extreme gradient\nboosting (XGBoost) and dissimilarity computation techniques. Albeit the simple\ndesign, this approach significantly increases accuracy on small datasets and\ndoes not introduce much complexity to the architecture of the neural network.\nThe proposed method is tested on various datasets, and a significant\nimprovement in the performance of the neural network model is seen. The model's\nprediction interval can include the ground truth value at an average rate of\n71% and 78% across training sizes of 90% and 55%, respectively. Finally, we\nhighlight other aspects and applications of the approach in experimental error\nestimation, and the application of transfer learning.", "title": "Confidence-Nets: A Step Towards better Prediction Intervals for regression Neural Networks on small datasets", "url": "http://arxiv.org/abs/2210.17092v1" }
null
null
no_new_dataset
admin
null
false
null
94ca527f-ee32-4c5e-a72f-b59b6610c380
null
Validated
2023-10-04 15:19:51.883232
{ "text_length": 1789 }
1no_new_dataset
TITLE: SuperBench: A Super-Resolution Benchmark Dataset for Scientific Machine Learning ABSTRACT: Super-Resolution (SR) techniques aim to enhance data resolution, enabling the retrieval of finer details, and improving the overall quality and fidelity of the data representation. There is growing interest in applying SR methods to complex spatiotemporal systems within the Scientific Machine Learning (SciML) community, with the hope of accelerating numerical simulations and/or improving forecasts in weather, climate, and related areas. However, the lack of standardized benchmark datasets for comparing and validating SR methods hinders progress and adoption in SciML. To address this, we introduce SuperBench, the first benchmark dataset featuring high-resolution datasets (up to $2048\times2048$ dimensions), including data from fluid flows, cosmology, and weather. Here, we focus on validating spatial SR performance from data-centric and physics-preserved perspectives, as well as assessing robustness to data degradation tasks. While deep learning-based SR methods (developed in the computer vision community) excel on certain tasks, despite relatively limited prior physics information, we identify limitations of these methods in accurately capturing intricate fine-scale features and preserving fundamental physical properties and constraints in scientific data. These shortcomings highlight the importance and subtlety of incorporating domain knowledge into ML models. We anticipate that SuperBench will significantly advance SR methods for scientific tasks.
{ "abstract": "Super-Resolution (SR) techniques aim to enhance data resolution, enabling the\nretrieval of finer details, and improving the overall quality and fidelity of\nthe data representation. There is growing interest in applying SR methods to\ncomplex spatiotemporal systems within the Scientific Machine Learning (SciML)\ncommunity, with the hope of accelerating numerical simulations and/or improving\nforecasts in weather, climate, and related areas. However, the lack of\nstandardized benchmark datasets for comparing and validating SR methods hinders\nprogress and adoption in SciML. To address this, we introduce SuperBench, the\nfirst benchmark dataset featuring high-resolution datasets (up to\n$2048\\times2048$ dimensions), including data from fluid flows, cosmology, and\nweather. Here, we focus on validating spatial SR performance from data-centric\nand physics-preserved perspectives, as well as assessing robustness to data\ndegradation tasks. While deep learning-based SR methods (developed in the\ncomputer vision community) excel on certain tasks, despite relatively limited\nprior physics information, we identify limitations of these methods in\naccurately capturing intricate fine-scale features and preserving fundamental\nphysical properties and constraints in scientific data. These shortcomings\nhighlight the importance and subtlety of incorporating domain knowledge into ML\nmodels. We anticipate that SuperBench will significantly advance SR methods for\nscientific tasks.", "title": "SuperBench: A Super-Resolution Benchmark Dataset for Scientific Machine Learning", "url": "http://arxiv.org/abs/2306.14070v1" }
null
null
new_dataset
admin
null
false
null
5f5e7dd6-18a1-4d0d-a7cb-3b50a8096791
null
Validated
2023-10-04 15:19:51.870028
{ "text_length": 1587 }
0new_dataset
TITLE: Data Collaboration Analysis applied to Compound Datasets and the Introduction of Projection data to Non-IID settings ABSTRACT: Given the time and expense associated with bringing a drug to market, numerous studies have been conducted to predict the properties of compounds based on their structure using machine learning. Federated learning has been applied to compound datasets to increase their prediction accuracy while safeguarding potentially proprietary information. However, federated learning is encumbered by low accuracy in not identically and independently distributed (non-IID) settings, i.e., data partitioning has a large label bias, and is considered unsuitable for compound datasets, which tend to have large label bias. To address this limitation, we utilized an alternative method of distributed machine learning to chemical compound data from open sources, called data collaboration analysis (DC). We also proposed data collaboration analysis using projection data (DCPd), which is an improved method that utilizes auxiliary PubChem data. This improves the quality of individual user-side data transformations for the projection data for the creation of intermediate representations. The classification accuracy, i.e., area under the curve in the receiver operating characteristic curve (ROC-AUC) and AUC in the precision-recall curve (PR-AUC), of federated averaging (FedAvg), DC, and DCPd was compared for five compound datasets. We determined that the machine learning performance for non-IID settings was in the order of DCPd, DC, and FedAvg, although they were almost the same in identically and independently distributed (IID) settings. Moreover, the results showed that compared to other methods, DCPd exhibited a negligible decline in classification accuracy in experiments with different degrees of label bias. Thus, DCPd can address the low performance in non-IID settings, which is one of the challenges of federated learning.
{ "abstract": "Given the time and expense associated with bringing a drug to market,\nnumerous studies have been conducted to predict the properties of compounds\nbased on their structure using machine learning. Federated learning has been\napplied to compound datasets to increase their prediction accuracy while\nsafeguarding potentially proprietary information. However, federated learning\nis encumbered by low accuracy in not identically and independently distributed\n(non-IID) settings, i.e., data partitioning has a large label bias, and is\nconsidered unsuitable for compound datasets, which tend to have large label\nbias. To address this limitation, we utilized an alternative method of\ndistributed machine learning to chemical compound data from open sources,\ncalled data collaboration analysis (DC). We also proposed data collaboration\nanalysis using projection data (DCPd), which is an improved method that\nutilizes auxiliary PubChem data. This improves the quality of individual\nuser-side data transformations for the projection data for the creation of\nintermediate representations. The classification accuracy, i.e., area under the\ncurve in the receiver operating characteristic curve (ROC-AUC) and AUC in the\nprecision-recall curve (PR-AUC), of federated averaging (FedAvg), DC, and DCPd\nwas compared for five compound datasets. We determined that the machine\nlearning performance for non-IID settings was in the order of DCPd, DC, and\nFedAvg, although they were almost the same in identically and independently\ndistributed (IID) settings. Moreover, the results showed that compared to other\nmethods, DCPd exhibited a negligible decline in classification accuracy in\nexperiments with different degrees of label bias. Thus, DCPd can address the\nlow performance in non-IID settings, which is one of the challenges of\nfederated learning.", "title": "Data Collaboration Analysis applied to Compound Datasets and the Introduction of Projection data to Non-IID settings", "url": "http://arxiv.org/abs/2308.00280v1" }
null
null
no_new_dataset
admin
null
false
null
857aac25-cf99-4db4-82ea-3d0718e5b1bc
null
Validated
2023-10-04 15:19:51.864545
{ "text_length": 1980 }
1no_new_dataset
TITLE: On Intrinsic Dataset Properties for Adversarial Machine Learning ABSTRACT: Deep neural networks (DNNs) have played a key role in a wide range of machine learning applications. However, DNN classifiers are vulnerable to human-imperceptible adversarial perturbations, which can cause them to misclassify inputs with high confidence. Thus, creating robust DNNs which can defend against malicious examples is critical in applications where security plays a major role. In this paper, we study the effect of intrinsic dataset properties on the performance of adversarial attack and defense methods, testing on five popular image classification datasets - MNIST, Fashion-MNIST, CIFAR10/CIFAR100, and ImageNet. We find that input size and image contrast play key roles in attack and defense success. Our discoveries highlight that dataset design and data preprocessing steps are important to boost the adversarial robustness of DNNs. To our best knowledge, this is the first comprehensive work that studies the effect of intrinsic dataset properties on adversarial machine learning.
{ "abstract": "Deep neural networks (DNNs) have played a key role in a wide range of machine\nlearning applications. However, DNN classifiers are vulnerable to\nhuman-imperceptible adversarial perturbations, which can cause them to\nmisclassify inputs with high confidence. Thus, creating robust DNNs which can\ndefend against malicious examples is critical in applications where security\nplays a major role. In this paper, we study the effect of intrinsic dataset\nproperties on the performance of adversarial attack and defense methods,\ntesting on five popular image classification datasets - MNIST, Fashion-MNIST,\nCIFAR10/CIFAR100, and ImageNet. We find that input size and image contrast play\nkey roles in attack and defense success. Our discoveries highlight that dataset\ndesign and data preprocessing steps are important to boost the adversarial\nrobustness of DNNs. To our best knowledge, this is the first comprehensive work\nthat studies the effect of intrinsic dataset properties on adversarial machine\nlearning.", "title": "On Intrinsic Dataset Properties for Adversarial Machine Learning", "url": "http://arxiv.org/abs/2005.09170v1" }
null
null
no_new_dataset
admin
null
false
null
4862c26e-143c-4cd4-94ca-81c5387b92df
null
Validated
2023-10-04 15:19:51.899927
{ "text_length": 1099 }
1no_new_dataset
TITLE: A High-Resolution Chest CT-Scan Image Dataset for COVID-19 Diagnosis and Differentiation ABSTRACT: During the COVID-19 pandemic, computed tomography (CT) is a good way to diagnose COVID-19 patients. HRCT (High-Resolution Computed Tomography) is a form of computed tomography that uses advanced methods to improve image resolution. Publicly accessible COVID-19 CT image datasets are very difficult to come by due to privacy concerns, which impedes the study and development of AI-powered COVID-19 diagnostic algorithms based on CT images. To address this problem, we have introduced HRCTv1-COVID-19, a new COVID-19 high resolution chest CT Scan image dataset that includes not only COVID-19 cases of Ground Glass Opacity (GGO), Crazy Paving, and Air Space Consolidation, but also CT images of cases with negative COVID-19. The HRCTv1-COVID-19 dataset, which includes slice-level, and patient-level labels, has the potential to aid COVID-19 research, especially for diagnosis and differentiation using artificial intelligence algorithms, machine learning and deep learning methods. This dataset is accessible through web at: http://databiox.com and includes 181,106 chest HRCT images from 395 patients with four labels: GGO, Crazy Paving, Air Space Consolidation and Negative. Keywords- Dataset, COVID-19, CT-Scan, Computed Tomography, Medical Imaging, Chest Image.
{ "abstract": "During the COVID-19 pandemic, computed tomography (CT) is a good way to\ndiagnose COVID-19 patients. HRCT (High-Resolution Computed Tomography) is a\nform of computed tomography that uses advanced methods to improve image\nresolution. Publicly accessible COVID-19 CT image datasets are very difficult\nto come by due to privacy concerns, which impedes the study and development of\nAI-powered COVID-19 diagnostic algorithms based on CT images. To address this\nproblem, we have introduced HRCTv1-COVID-19, a new COVID-19 high resolution\nchest CT Scan image dataset that includes not only COVID-19 cases of Ground\nGlass Opacity (GGO), Crazy Paving, and Air Space Consolidation, but also CT\nimages of cases with negative COVID-19. The HRCTv1-COVID-19 dataset, which\nincludes slice-level, and patient-level labels, has the potential to aid\nCOVID-19 research, especially for diagnosis and differentiation using\nartificial intelligence algorithms, machine learning and deep learning methods.\nThis dataset is accessible through web at: http://databiox.com and includes\n181,106 chest HRCT images from 395 patients with four labels: GGO, Crazy\nPaving, Air Space Consolidation and Negative.\n Keywords- Dataset, COVID-19, CT-Scan, Computed Tomography, Medical Imaging,\nChest Image.", "title": "A High-Resolution Chest CT-Scan Image Dataset for COVID-19 Diagnosis and Differentiation", "url": "http://arxiv.org/abs/2205.03408v1" }
null
null
new_dataset
admin
null
false
null
a3fd3207-7327-4e20-a8b5-e4516ccf4144
null
Validated
2023-10-04 15:19:51.886609
{ "text_length": 1389 }
0new_dataset
TITLE: Occams Razor for Big Data? On Detecting Quality in Large Unstructured Datasets ABSTRACT: Detecting quality in large unstructured datasets requires capacities far beyond the limits of human perception and communicability and, as a result, there is an emerging trend towards increasingly complex analytic solutions in data science to cope with this problem. This new trend towards analytic complexity represents a severe challenge for the principle of parsimony or Occams Razor in science. This review article combines insight from various domains such as physics, computational science, data engineering, and cognitive science to review the specific properties of big data. Problems for detecting data quality without losing the principle of parsimony are then highlighted on the basis of specific examples. Computational building block approaches for data clustering can help to deal with large unstructured datasets in minimized computation time, and meaning can be extracted rapidly from large sets of unstructured image or video data parsimoniously through relatively simple unsupervised machine learning algorithms. Why we still massively lack in expertise for exploiting big data wisely to extract relevant information for specific tasks, recognize patterns, generate new information, or store and further process large amounts of sensor data is then reviewed; examples illustrating why we need subjective views and pragmatic methods to analyze big data contents are brought forward. The review concludes on how cultural differences between East and West are likely to affect the course of big data analytics, and the development of increasingly autonomous artificial intelligence aimed at coping with the big data deluge in the near future.
{ "abstract": "Detecting quality in large unstructured datasets requires capacities far\nbeyond the limits of human perception and communicability and, as a result,\nthere is an emerging trend towards increasingly complex analytic solutions in\ndata science to cope with this problem. This new trend towards analytic\ncomplexity represents a severe challenge for the principle of parsimony or\nOccams Razor in science. This review article combines insight from various\ndomains such as physics, computational science, data engineering, and cognitive\nscience to review the specific properties of big data. Problems for detecting\ndata quality without losing the principle of parsimony are then highlighted on\nthe basis of specific examples. Computational building block approaches for\ndata clustering can help to deal with large unstructured datasets in minimized\ncomputation time, and meaning can be extracted rapidly from large sets of\nunstructured image or video data parsimoniously through relatively simple\nunsupervised machine learning algorithms. Why we still massively lack in\nexpertise for exploiting big data wisely to extract relevant information for\nspecific tasks, recognize patterns, generate new information, or store and\nfurther process large amounts of sensor data is then reviewed; examples\nillustrating why we need subjective views and pragmatic methods to analyze big\ndata contents are brought forward. The review concludes on how cultural\ndifferences between East and West are likely to affect the course of big data\nanalytics, and the development of increasingly autonomous artificial\nintelligence aimed at coping with the big data deluge in the near future.", "title": "Occams Razor for Big Data? On Detecting Quality in Large Unstructured Datasets", "url": "http://arxiv.org/abs/2011.08663v1" }
null
null
no_new_dataset
admin
null
false
null
1189fce5-fcac-458b-920e-6e91827dab85
null
Validated
2023-10-04 15:19:51.897044
{ "text_length": 1770 }
1no_new_dataset
TITLE: Surgical Phase and Instrument Recognition: How to identify appropriate Dataset Splits ABSTRACT: Purpose: The development of machine learning models for surgical workflow and instrument recognition from temporal data represents a challenging task due to the complex nature of surgical workflows. In particular, the imbalanced distribution of data is one of the major challenges in the domain of surgical workflow recognition. In order to obtain meaningful results, careful partitioning of data into training, validation, and test sets, as well as the selection of suitable evaluation metrics are crucial. Methods: In this work, we present an openly available web-based application that enables interactive exploration of dataset partitions. The proposed visual framework facilitates the assessment of dataset splits for surgical workflow recognition, especially with regard to identifying sub-optimal dataset splits. Currently, it supports visualization of surgical phase and instrument annotations. Results: In order to validate the dedicated interactive visualizations, we use a dataset split of the Cholec80 dataset. This dataset split was specifically selected to reflect a case of strong data imbalance. Using our software, we were able to identify phases, phase transitions, and combinations of surgical instruments that were not represented in one of the sets. Conclusion: In order to obtain meaningful results in highly unbalanced class distributions, special care should be taken with respect to the selection of an appropriate split. Interactive data visualization represents a promising approach for the assessment of machine learning datasets. The source code is available at https://github.com/Cardio-AI/endovis-ml
{ "abstract": "Purpose: The development of machine learning models for surgical workflow and\ninstrument recognition from temporal data represents a challenging task due to\nthe complex nature of surgical workflows. In particular, the imbalanced\ndistribution of data is one of the major challenges in the domain of surgical\nworkflow recognition. In order to obtain meaningful results, careful\npartitioning of data into training, validation, and test sets, as well as the\nselection of suitable evaluation metrics are crucial. Methods: In this work, we\npresent an openly available web-based application that enables interactive\nexploration of dataset partitions. The proposed visual framework facilitates\nthe assessment of dataset splits for surgical workflow recognition, especially\nwith regard to identifying sub-optimal dataset splits. Currently, it supports\nvisualization of surgical phase and instrument annotations. Results: In order\nto validate the dedicated interactive visualizations, we use a dataset split of\nthe Cholec80 dataset. This dataset split was specifically selected to reflect a\ncase of strong data imbalance. Using our software, we were able to identify\nphases, phase transitions, and combinations of surgical instruments that were\nnot represented in one of the sets. Conclusion: In order to obtain meaningful\nresults in highly unbalanced class distributions, special care should be taken\nwith respect to the selection of an appropriate split. Interactive data\nvisualization represents a promising approach for the assessment of machine\nlearning datasets. The source code is available at\nhttps://github.com/Cardio-AI/endovis-ml", "title": "Surgical Phase and Instrument Recognition: How to identify appropriate Dataset Splits", "url": "http://arxiv.org/abs/2306.16879v1" }
null
null
no_new_dataset
admin
null
false
null
c3bd27fc-b37a-48b3-99cc-5d65b5af68a1
null
Validated
2023-10-04 15:19:51.868697
{ "text_length": 1750 }
1no_new_dataset
TITLE: Presenting a Larger Up-to-date Movie Dataset and Investigating the Effects of Pre-released Attributes on Gross Revenue ABSTRACT: Movie-making has become one of the most costly and risky endeavors in the entertainment industry. Continuous change in the preference of the audience makes it harder to predict what kind of movie will be financially successful at the box office. So, it is no wonder that cautious, intelligent stakeholders and large production houses will always want to know the probable revenue that will be generated by a movie before making an investment. Researchers have been working on finding an optimal strategy to help investors in making the right decisions. But the lack of a large, up-to-date dataset makes their work harder. In this work, we introduce an up-to-date, richer, and larger dataset that we have prepared by scraping IMDb for researchers and data analysts to work with. The compiled dataset contains the summery data of 7.5 million titles and detail information of more than 200K movies. Additionally, we perform different statistical analysis approaches on our dataset to find out how a movie's revenue is affected by different pre-released attributes such as budget, runtime, release month, content rating, genre etc. In our analysis, we have found that having a star cast/director has a positive impact on generated revenue. We introduce a novel approach for calculating the star power of a movie. Based on our analysis we select a set of attributes as features and train different machine learning algorithms to predict a movie's expected revenue. Based on generated revenue, we classified the movies in 10 categories and achieved a one-class-away accuracy rate of almost 60% (bingo accuracy of 30%). All the generated datasets and analysis codes are available online. We also made the source codes of our scraper bots public, so that researchers interested in extending this work can easily modify these bots as they need and prepare their own up-to-date datasets.
{ "abstract": "Movie-making has become one of the most costly and risky endeavors in the\nentertainment industry. Continuous change in the preference of the audience\nmakes it harder to predict what kind of movie will be financially successful at\nthe box office. So, it is no wonder that cautious, intelligent stakeholders and\nlarge production houses will always want to know the probable revenue that will\nbe generated by a movie before making an investment. Researchers have been\nworking on finding an optimal strategy to help investors in making the right\ndecisions. But the lack of a large, up-to-date dataset makes their work harder.\nIn this work, we introduce an up-to-date, richer, and larger dataset that we\nhave prepared by scraping IMDb for researchers and data analysts to work with.\nThe compiled dataset contains the summery data of 7.5 million titles and detail\ninformation of more than 200K movies. Additionally, we perform different\nstatistical analysis approaches on our dataset to find out how a movie's\nrevenue is affected by different pre-released attributes such as budget,\nruntime, release month, content rating, genre etc. In our analysis, we have\nfound that having a star cast/director has a positive impact on generated\nrevenue. We introduce a novel approach for calculating the star power of a\nmovie. Based on our analysis we select a set of attributes as features and\ntrain different machine learning algorithms to predict a movie's expected\nrevenue. Based on generated revenue, we classified the movies in 10 categories\nand achieved a one-class-away accuracy rate of almost 60% (bingo accuracy of\n30%). All the generated datasets and analysis codes are available online. We\nalso made the source codes of our scraper bots public, so that researchers\ninterested in extending this work can easily modify these bots as they need and\nprepare their own up-to-date datasets.", "title": "Presenting a Larger Up-to-date Movie Dataset and Investigating the Effects of Pre-released Attributes on Gross Revenue", "url": "http://arxiv.org/abs/2110.07039v2" }
null
null
new_dataset
admin
null
false
null
14366fed-8b51-40fd-9cc3-c8ff049dd855
null
Validated
2023-10-04 15:19:51.891029
{ "text_length": 2030 }
0new_dataset
TITLE: Dataset Factory: A Toolchain For Generative Computer Vision Datasets ABSTRACT: Generative AI workflows heavily rely on data-centric tasks - such as filtering samples by annotation fields, vector distances, or scores produced by custom classifiers. At the same time, computer vision datasets are quickly approaching petabyte volumes, rendering data wrangling difficult. In addition, the iterative nature of data preparation necessitates robust dataset sharing and versioning mechanisms, both of which are hard to implement ad-hoc. To solve these challenges, we propose a "dataset factory" approach that separates the storage and processing of samples from metadata and enables data-centric operations at scale for machine learning teams and individual researchers.
{ "abstract": "Generative AI workflows heavily rely on data-centric tasks - such as\nfiltering samples by annotation fields, vector distances, or scores produced by\ncustom classifiers. At the same time, computer vision datasets are quickly\napproaching petabyte volumes, rendering data wrangling difficult. In addition,\nthe iterative nature of data preparation necessitates robust dataset sharing\nand versioning mechanisms, both of which are hard to implement ad-hoc. To solve\nthese challenges, we propose a \"dataset factory\" approach that separates the\nstorage and processing of samples from metadata and enables data-centric\noperations at scale for machine learning teams and individual researchers.", "title": "Dataset Factory: A Toolchain For Generative Computer Vision Datasets", "url": "http://arxiv.org/abs/2309.11608v1" }
null
null
no_new_dataset
admin
null
false
null
fa011680-5d73-4e07-aaa1-11e087eb022d
null
Validated
2023-10-04 15:19:51.863345
{ "text_length": 787 }
1no_new_dataset
TITLE: A Synthetic Dataset for 5G UAV Attacks Based on Observable Network Parameters ABSTRACT: Synthetic datasets are beneficial for machine learning researchers due to the possibility of experimenting with new strategies and algorithms in the training and testing phases. These datasets can easily include more scenarios that might be costly to research with real data or can complement and, in some cases, replace real data measurements, depending on the quality of the synthetic data. They can also solve the unbalanced data problem, avoid overfitting, and can be used in training while testing can be done with real data. In this paper, we present, to the best of our knowledge, the first synthetic dataset for Unmanned Aerial Vehicle (UAV) attacks in 5G and beyond networks based on the following key observable network parameters that indicate power levels: the Received Signal Strength Indicator (RSSI) and the Signal to Interference-plus-Noise Ratio (SINR). The main objective of this data is to enable deep network development for UAV communication security. Especially, for algorithm development or the analysis of time-series data applied to UAV attack recognition. Our proposed dataset provides insights into network functionality when static or moving UAV attackers target authenticated UAVs in an urban environment. The dataset also considers the presence and absence of authenticated terrestrial users in the network, which may decrease the deep networks ability to identify attacks. Furthermore, the data provides deeper comprehension of the metrics available in the 5G physical and MAC layers for machine learning and statistics research. The dataset will available at link archive-beta.ics.uci.edu
{ "abstract": "Synthetic datasets are beneficial for machine learning researchers due to the\npossibility of experimenting with new strategies and algorithms in the training\nand testing phases. These datasets can easily include more scenarios that might\nbe costly to research with real data or can complement and, in some cases,\nreplace real data measurements, depending on the quality of the synthetic data.\nThey can also solve the unbalanced data problem, avoid overfitting, and can be\nused in training while testing can be done with real data. In this paper, we\npresent, to the best of our knowledge, the first synthetic dataset for Unmanned\nAerial Vehicle (UAV) attacks in 5G and beyond networks based on the following\nkey observable network parameters that indicate power levels: the Received\nSignal Strength Indicator (RSSI) and the Signal to Interference-plus-Noise\nRatio (SINR). The main objective of this data is to enable deep network\ndevelopment for UAV communication security. Especially, for algorithm\ndevelopment or the analysis of time-series data applied to UAV attack\nrecognition. Our proposed dataset provides insights into network functionality\nwhen static or moving UAV attackers target authenticated UAVs in an urban\nenvironment. The dataset also considers the presence and absence of\nauthenticated terrestrial users in the network, which may decrease the deep\nnetworks ability to identify attacks. Furthermore, the data provides deeper\ncomprehension of the metrics available in the 5G physical and MAC layers for\nmachine learning and statistics research. The dataset will available at link\narchive-beta.ics.uci.edu", "title": "A Synthetic Dataset for 5G UAV Attacks Based on Observable Network Parameters", "url": "http://arxiv.org/abs/2211.09706v1" }
null
null
new_dataset
admin
null
false
null
a407ec96-3a6e-432c-88ce-b3caf3cd1e90
null
Validated
2023-10-04 15:19:51.883043
{ "text_length": 1732 }
0new_dataset
TITLE: Automatic Dataset Builder for Machine Learning Applications to Satellite Imagery ABSTRACT: Nowadays the use of Machine Learning (ML) algorithms is spreading in the field of Remote Sensing, with applications ranging from detection and classification of land use and monitoring to the prediction of many natural or anthropic phenomena of interest. One main limit of their employment is related to the need for a huge amount of data for training the neural network, chosen for the specific application, and the resulting computational weight and time required to collect the necessary data. In this letter the architecture of an innovative tool, enabling researchers to create in an automatic way suitable datasets for AI (Artificial Intelligence) applications in the EO (Earth Observation) context, is presented. Two versions of the architecture have been implemented and made available on Git-Hub, with a specific Graphical User Interface (GUI) for non-expert users.
{ "abstract": "Nowadays the use of Machine Learning (ML) algorithms is spreading in the\nfield of Remote Sensing, with applications ranging from detection and\nclassification of land use and monitoring to the prediction of many natural or\nanthropic phenomena of interest. One main limit of their employment is related\nto the need for a huge amount of data for training the neural network, chosen\nfor the specific application, and the resulting computational weight and time\nrequired to collect the necessary data. In this letter the architecture of an\ninnovative tool, enabling researchers to create in an automatic way suitable\ndatasets for AI (Artificial Intelligence) applications in the EO (Earth\nObservation) context, is presented. Two versions of the architecture have been\nimplemented and made available on Git-Hub, with a specific Graphical User\nInterface (GUI) for non-expert users.", "title": "Automatic Dataset Builder for Machine Learning Applications to Satellite Imagery", "url": "http://arxiv.org/abs/2008.01578v1" }
null
null
no_new_dataset
admin
null
false
null
2f4a9cc8-cb31-49c8-9b11-1521acb50919
null
Validated
2023-10-04 15:19:51.898922
{ "text_length": 989 }
1no_new_dataset
TITLE: IFCNet: A Benchmark Dataset for IFC Entity Classification ABSTRACT: Enhancing interoperability and information exchange between domain-specific software products for BIM is an important aspect in the Architecture, Engineering, Construction and Operations industry. Recent research started investigating methods from the areas of machine and deep learning for semantic enrichment of BIM models. However, training and evaluation of these machine learning algorithms requires sufficiently large and comprehensive datasets. This work presents IFCNet, a dataset of single-entity IFC files spanning a broad range of IFC classes containing both geometric and semantic information. Using only the geometric information of objects, the experiments show that three different deep learning models are able to achieve good classification performance.
{ "abstract": "Enhancing interoperability and information exchange between domain-specific\nsoftware products for BIM is an important aspect in the Architecture,\nEngineering, Construction and Operations industry. Recent research started\ninvestigating methods from the areas of machine and deep learning for semantic\nenrichment of BIM models. However, training and evaluation of these machine\nlearning algorithms requires sufficiently large and comprehensive datasets.\nThis work presents IFCNet, a dataset of single-entity IFC files spanning a\nbroad range of IFC classes containing both geometric and semantic information.\nUsing only the geometric information of objects, the experiments show that\nthree different deep learning models are able to achieve good classification\nperformance.", "title": "IFCNet: A Benchmark Dataset for IFC Entity Classification", "url": "http://arxiv.org/abs/2106.09712v1" }
null
null
new_dataset
admin
null
false
null
e6fb60d6-17c4-4d32-a8b2-8666742d1ee0
null
Validated
2023-10-04 15:19:51.894074
{ "text_length": 862 }
0new_dataset
TITLE: TextileNet: A Material Taxonomy-based Fashion Textile Dataset ABSTRACT: The rise of Machine Learning (ML) is gradually digitalizing and reshaping the fashion industry. Recent years have witnessed a number of fashion AI applications, for example, virtual try-ons. Textile material identification and categorization play a crucial role in the fashion textile sector, including fashion design, retails, and recycling. At the same time, Net Zero is a global goal and the fashion industry is undergoing a significant change so that textile materials can be reused, repaired and recycled in a sustainable manner. There is still a challenge in identifying textile materials automatically for garments, as we lack a low-cost and effective technique for identifying them. In light of this, we build the first fashion textile dataset, TextileNet, based on textile material taxonomies - a fibre taxonomy and a fabric taxonomy generated in collaboration with material scientists. TextileNet can be used to train and evaluate the state-of-the-art Deep Learning models for textile materials. We hope to standardize textile related datasets through the use of taxonomies. TextileNet contains 33 fibres labels and 27 fabrics labels, and has in total 760,949 images. We use standard Convolutional Neural Networks (CNNs) and Vision Transformers (ViTs) to establish baselines for this dataset. Future applications for this dataset range from textile classification to optimization of the textile supply chain and interactive design for consumers. We envision that this can contribute to the development of a new AI-based fashion platform.
{ "abstract": "The rise of Machine Learning (ML) is gradually digitalizing and reshaping the\nfashion industry. Recent years have witnessed a number of fashion AI\napplications, for example, virtual try-ons. Textile material identification and\ncategorization play a crucial role in the fashion textile sector, including\nfashion design, retails, and recycling. At the same time, Net Zero is a global\ngoal and the fashion industry is undergoing a significant change so that\ntextile materials can be reused, repaired and recycled in a sustainable manner.\nThere is still a challenge in identifying textile materials automatically for\ngarments, as we lack a low-cost and effective technique for identifying them.\nIn light of this, we build the first fashion textile dataset, TextileNet, based\non textile material taxonomies - a fibre taxonomy and a fabric taxonomy\ngenerated in collaboration with material scientists. TextileNet can be used to\ntrain and evaluate the state-of-the-art Deep Learning models for textile\nmaterials. We hope to standardize textile related datasets through the use of\ntaxonomies. TextileNet contains 33 fibres labels and 27 fabrics labels, and has\nin total 760,949 images. We use standard Convolutional Neural Networks (CNNs)\nand Vision Transformers (ViTs) to establish baselines for this dataset. Future\napplications for this dataset range from textile classification to optimization\nof the textile supply chain and interactive design for consumers. We envision\nthat this can contribute to the development of a new AI-based fashion platform.", "title": "TextileNet: A Material Taxonomy-based Fashion Textile Dataset", "url": "http://arxiv.org/abs/2301.06160v1" }
null
null
new_dataset
admin
null
false
null
0f5da146-7ae1-4ac8-a996-f59a73535f9c
null
Validated
2023-10-04 15:19:51.881509
{ "text_length": 1643 }
0new_dataset
TITLE: IGRF-RFE: A Hybrid Feature Selection Method for MLP-based Network Intrusion Detection on UNSW-NB15 Dataset ABSTRACT: The effectiveness of machine learning models is significantly affected by the size of the dataset and the quality of features as redundant and irrelevant features can radically degrade the performance. This paper proposes IGRF-RFE: a hybrid feature selection method tasked for multi-class network anomalies using a Multilayer perceptron (MLP) network. IGRF-RFE can be considered as a feature reduction technique based on both the filter feature selection method and the wrapper feature selection method. In our proposed method, we use the filter feature selection method, which is the combination of Information Gain and Random Forest Importance, to reduce the feature subset search space. Then, we apply recursive feature elimination(RFE) as a wrapper feature selection method to further eliminate redundant features recursively on the reduced feature subsets. Our experimental results obtained based on the UNSW-NB15 dataset confirm that our proposed method can improve the accuracy of anomaly detection while reducing the feature dimension. The results show that the feature dimension is reduced from 42 to 23 while the multi-classification accuracy of MLP is improved from 82.25% to 84.24%.
{ "abstract": "The effectiveness of machine learning models is significantly affected by the\nsize of the dataset and the quality of features as redundant and irrelevant\nfeatures can radically degrade the performance. This paper proposes IGRF-RFE: a\nhybrid feature selection method tasked for multi-class network anomalies using\na Multilayer perceptron (MLP) network. IGRF-RFE can be considered as a feature\nreduction technique based on both the filter feature selection method and the\nwrapper feature selection method. In our proposed method, we use the filter\nfeature selection method, which is the combination of Information Gain and\nRandom Forest Importance, to reduce the feature subset search space. Then, we\napply recursive feature elimination(RFE) as a wrapper feature selection method\nto further eliminate redundant features recursively on the reduced feature\nsubsets. Our experimental results obtained based on the UNSW-NB15 dataset\nconfirm that our proposed method can improve the accuracy of anomaly detection\nwhile reducing the feature dimension. The results show that the feature\ndimension is reduced from 42 to 23 while the multi-classification accuracy of\nMLP is improved from 82.25% to 84.24%.", "title": "IGRF-RFE: A Hybrid Feature Selection Method for MLP-based Network Intrusion Detection on UNSW-NB15 Dataset", "url": "http://arxiv.org/abs/2203.16365v2" }
null
null
no_new_dataset
admin
null
false
null
1e7dc758-d310-4b1f-a822-9a1e142e609d
null
Validated
2023-10-04 15:19:51.887310
{ "text_length": 1335 }
1no_new_dataset
TITLE: MVA2023 Small Object Detection Challenge for Spotting Birds: Dataset, Methods, and Results ABSTRACT: Small Object Detection (SOD) is an important machine vision topic because (i) a variety of real-world applications require object detection for distant objects and (ii) SOD is a challenging task due to the noisy, blurred, and less-informative image appearances of small objects. This paper proposes a new SOD dataset consisting of 39,070 images including 137,121 bird instances, which is called the Small Object Detection for Spotting Birds (SOD4SB) dataset. The detail of the challenge with the SOD4SB dataset is introduced in this paper. In total, 223 participants joined this challenge. This paper briefly introduces the award-winning methods. The dataset, the baseline code, and the website for evaluation on the public testset are publicly available.
{ "abstract": "Small Object Detection (SOD) is an important machine vision topic because (i)\na variety of real-world applications require object detection for distant\nobjects and (ii) SOD is a challenging task due to the noisy, blurred, and\nless-informative image appearances of small objects. This paper proposes a new\nSOD dataset consisting of 39,070 images including 137,121 bird instances, which\nis called the Small Object Detection for Spotting Birds (SOD4SB) dataset. The\ndetail of the challenge with the SOD4SB dataset is introduced in this paper. In\ntotal, 223 participants joined this challenge. This paper briefly introduces\nthe award-winning methods. The dataset, the baseline code, and the website for\nevaluation on the public testset are publicly available.", "title": "MVA2023 Small Object Detection Challenge for Spotting Birds: Dataset, Methods, and Results", "url": "http://arxiv.org/abs/2307.09143v1" }
null
null
new_dataset
admin
null
false
null
5f019e01-32e6-4692-8c02-324e6a1bb7a0
null
Validated
2023-10-04 15:19:51.866808
{ "text_length": 880 }
0new_dataset
TITLE: Data Models for Dataset Drift Controls in Machine Learning With Optical Images ABSTRACT: Camera images are ubiquitous in machine learning research. They also play a central role in the delivery of important services spanning medicine and environmental surveying. However, the application of machine learning models in these domains has been limited because of robustness concerns. A primary failure mode are performance drops due to differences between the training and deployment data. While there are methods to prospectively validate the robustness of machine learning models to such dataset drifts, existing approaches do not account for explicit models of the primary object of interest: the data. This limits our ability to study and understand the relationship between data generation and downstream machine learning model performance in a physically accurate manner. In this study, we demonstrate how to overcome this limitation by pairing traditional machine learning with physical optics to obtain explicit and differentiable data models. We demonstrate how such data models can be constructed for image data and used to control downstream machine learning model performance related to dataset drift. The findings are distilled into three applications. First, drift synthesis enables the controlled generation of physically faithful drift test cases to power model selection and targeted generalization. Second, the gradient connection between machine learning task model and data model allows advanced, precise tolerancing of task model sensitivity to changes in the data generation. These drift forensics can be used to precisely specify the acceptable data environments in which a task model may be run. Third, drift optimization opens up the possibility to create drifts that can help the task model learn better faster, effectively optimizing the data generating process itself. A guide to access the open code and datasets is available at https://github.com/aiaudit-org/raw2logit.
{ "abstract": "Camera images are ubiquitous in machine learning research. They also play a\ncentral role in the delivery of important services spanning medicine and\nenvironmental surveying. However, the application of machine learning models in\nthese domains has been limited because of robustness concerns. A primary\nfailure mode are performance drops due to differences between the training and\ndeployment data. While there are methods to prospectively validate the\nrobustness of machine learning models to such dataset drifts, existing\napproaches do not account for explicit models of the primary object of\ninterest: the data. This limits our ability to study and understand the\nrelationship between data generation and downstream machine learning model\nperformance in a physically accurate manner. In this study, we demonstrate how\nto overcome this limitation by pairing traditional machine learning with\nphysical optics to obtain explicit and differentiable data models. We\ndemonstrate how such data models can be constructed for image data and used to\ncontrol downstream machine learning model performance related to dataset drift.\nThe findings are distilled into three applications. First, drift synthesis\nenables the controlled generation of physically faithful drift test cases to\npower model selection and targeted generalization. Second, the gradient\nconnection between machine learning task model and data model allows advanced,\nprecise tolerancing of task model sensitivity to changes in the data\ngeneration. These drift forensics can be used to precisely specify the\nacceptable data environments in which a task model may be run. Third, drift\noptimization opens up the possibility to create drifts that can help the task\nmodel learn better faster, effectively optimizing the data generating process\nitself. A guide to access the open code and datasets is available at\nhttps://github.com/aiaudit-org/raw2logit.", "title": "Data Models for Dataset Drift Controls in Machine Learning With Optical Images", "url": "http://arxiv.org/abs/2211.02578v3" }
null
null
no_new_dataset
admin
null
false
null
75bfa54b-db90-45f0-89a9-3102ff3b3df3
null
Validated
2023-10-04 15:19:51.883067
{ "text_length": 2020 }
1no_new_dataset
TITLE: The Influence of Dataset Partitioning on Dysfluency Detection Systems ABSTRACT: This paper empirically investigates the influence of different data splits and splitting strategies on the performance of dysfluency detection systems. For this, we perform experiments using wav2vec 2.0 models with a classification head as well as support vector machines (SVM) in conjunction with the features extracted from the wav2vec 2.0 model to detect dysfluencies. We train and evaluate the systems with different non-speaker-exclusive and speaker-exclusive splits of the Stuttering Events in Podcasts (SEP-28k) dataset to shed some light on the variability of results w.r.t. to the partition method used. Furthermore, we show that the SEP-28k dataset is dominated by only a few speakers, making it difficult to evaluate. To remedy this problem, we created SEP-28k-Extended (SEP-28k-E), containing semi-automatically generated speaker and gender information for the SEP-28k corpus, and suggest different data splits, each useful for evaluating other aspects of methods for dysfluency detection.
{ "abstract": "This paper empirically investigates the influence of different data splits\nand splitting strategies on the performance of dysfluency detection systems.\nFor this, we perform experiments using wav2vec 2.0 models with a classification\nhead as well as support vector machines (SVM) in conjunction with the features\nextracted from the wav2vec 2.0 model to detect dysfluencies. We train and\nevaluate the systems with different non-speaker-exclusive and speaker-exclusive\nsplits of the Stuttering Events in Podcasts (SEP-28k) dataset to shed some\nlight on the variability of results w.r.t. to the partition method used.\nFurthermore, we show that the SEP-28k dataset is dominated by only a few\nspeakers, making it difficult to evaluate. To remedy this problem, we created\nSEP-28k-Extended (SEP-28k-E), containing semi-automatically generated speaker\nand gender information for the SEP-28k corpus, and suggest different data\nsplits, each useful for evaluating other aspects of methods for dysfluency\ndetection.", "title": "The Influence of Dataset Partitioning on Dysfluency Detection Systems", "url": "http://arxiv.org/abs/2206.03400v1" }
null
null
no_new_dataset
admin
null
false
null
199c1346-2b99-47f9-8524-a305f8e79dee
null
Validated
2023-10-04 15:19:51.885937
{ "text_length": 1105 }
1no_new_dataset
TITLE: Potrika: Raw and Balanced Newspaper Datasets in the Bangla Language with Eight Topics and Five Attributes ABSTRACT: Knowledge is central to human and scientific developments. Natural Language Processing (NLP) allows automated analysis and creation of knowledge. Data is a crucial NLP and machine learning ingredient. The scarcity of open datasets is a well-known problem in machine and deep learning research. This is very much the case for textual NLP datasets in English and other major world languages. For the Bangla language, the situation is even more challenging and the number of large datasets for NLP research is practically nil. We hereby present Potrika, a large single-label Bangla news article textual dataset curated for NLP research from six popular online news portals in Bangladesh (Jugantor, Jaijaidin, Ittefaq, Kaler Kontho, Inqilab, and Somoyer Alo) for the period 2014-2020. The articles are classified into eight distinct categories (National, Sports, International, Entertainment, Economy, Education, Politics, and Science \& Technology) providing five attributes (News Article, Category, Headline, Publication Date, and Newspaper Source). The raw dataset contains 185.51 million words and 12.57 million sentences contained in 664,880 news articles. Moreover, using NLP augmentation techniques, we create from the raw (unbalanced) dataset another (balanced) dataset comprising 320,000 news articles with 40,000 articles in each of the eight news categories. Potrika contains both the datasets (raw and balanced) to suit a wide range of NLP research. By far, to the best of our knowledge, Potrika is the largest and the most extensive dataset for news classification.
{ "abstract": "Knowledge is central to human and scientific developments. Natural Language\nProcessing (NLP) allows automated analysis and creation of knowledge. Data is a\ncrucial NLP and machine learning ingredient. The scarcity of open datasets is a\nwell-known problem in machine and deep learning research. This is very much the\ncase for textual NLP datasets in English and other major world languages. For\nthe Bangla language, the situation is even more challenging and the number of\nlarge datasets for NLP research is practically nil. We hereby present Potrika,\na large single-label Bangla news article textual dataset curated for NLP\nresearch from six popular online news portals in Bangladesh (Jugantor,\nJaijaidin, Ittefaq, Kaler Kontho, Inqilab, and Somoyer Alo) for the period\n2014-2020. The articles are classified into eight distinct categories\n(National, Sports, International, Entertainment, Economy, Education, Politics,\nand Science \\& Technology) providing five attributes (News Article, Category,\nHeadline, Publication Date, and Newspaper Source). The raw dataset contains\n185.51 million words and 12.57 million sentences contained in 664,880 news\narticles. Moreover, using NLP augmentation techniques, we create from the raw\n(unbalanced) dataset another (balanced) dataset comprising 320,000 news\narticles with 40,000 articles in each of the eight news categories. Potrika\ncontains both the datasets (raw and balanced) to suit a wide range of NLP\nresearch. By far, to the best of our knowledge, Potrika is the largest and the\nmost extensive dataset for news classification.", "title": "Potrika: Raw and Balanced Newspaper Datasets in the Bangla Language with Eight Topics and Five Attributes", "url": "http://arxiv.org/abs/2210.09389v1" }
null
null
new_dataset
admin
null
false
null
7fe08fc6-104f-4522-b868-4114f14aa33c
null
Validated
2023-10-04 15:19:51.883427
{ "text_length": 1714 }
0new_dataset
TITLE: Creating Synthetic Datasets for Collaborative Filtering Recommender Systems using Generative Adversarial Networks ABSTRACT: Research and education in machine learning needs diverse, representative, and open datasets that contain sufficient samples to handle the necessary training, validation, and testing tasks. Currently, the Recommender Systems area includes a large number of subfields in which accuracy and beyond accuracy quality measures are continuously improved. To feed this research variety, it is necessary and convenient to reinforce the existing datasets with synthetic ones. This paper proposes a Generative Adversarial Network (GAN)-based method to generate collaborative filtering datasets in a parameterized way, by selecting their preferred number of users, items, samples, and stochastic variability. This parameterization cannot be made using regular GANs. Our GAN model is fed with dense, short, and continuous embedding representations of items and users, instead of sparse, large, and discrete vectors, to make an accurate and quick learning, compared to the traditional approach based on large and sparse input vectors. The proposed architecture includes a DeepMF model to extract the dense user and item embeddings, as well as a clustering process to convert from the dense GAN generated samples to the discrete and sparse ones, necessary to create each required synthetic dataset. The results of three different source datasets show adequate distributions and expected quality values and evolutions on the generated datasets compared to the source ones. Synthetic datasets and source codes are available to researchers.
{ "abstract": "Research and education in machine learning needs diverse, representative, and\nopen datasets that contain sufficient samples to handle the necessary training,\nvalidation, and testing tasks. Currently, the Recommender Systems area includes\na large number of subfields in which accuracy and beyond accuracy quality\nmeasures are continuously improved. To feed this research variety, it is\nnecessary and convenient to reinforce the existing datasets with synthetic\nones. This paper proposes a Generative Adversarial Network (GAN)-based method\nto generate collaborative filtering datasets in a parameterized way, by\nselecting their preferred number of users, items, samples, and stochastic\nvariability. This parameterization cannot be made using regular GANs. Our GAN\nmodel is fed with dense, short, and continuous embedding representations of\nitems and users, instead of sparse, large, and discrete vectors, to make an\naccurate and quick learning, compared to the traditional approach based on\nlarge and sparse input vectors. The proposed architecture includes a DeepMF\nmodel to extract the dense user and item embeddings, as well as a clustering\nprocess to convert from the dense GAN generated samples to the discrete and\nsparse ones, necessary to create each required synthetic dataset. The results\nof three different source datasets show adequate distributions and expected\nquality values and evolutions on the generated datasets compared to the source\nones. Synthetic datasets and source codes are available to researchers.", "title": "Creating Synthetic Datasets for Collaborative Filtering Recommender Systems using Generative Adversarial Networks", "url": "http://arxiv.org/abs/2303.01297v1" }
null
null
no_new_dataset
admin
null
false
null
575ace44-3cdd-420d-910d-cbae628cbf19
null
Validated
2023-10-04 15:19:51.880800
{ "text_length": 1670 }
1no_new_dataset
TITLE: An investigation of licensing of datasets for machine learning based on the GQM model ABSTRACT: Dataset licensing is currently an issue in the development of machine learning systems. And in the development of machine learning systems, the most widely used are publicly available datasets. However, since the images in the publicly available dataset are mainly obtained from the Internet, some images are not commercially available. Furthermore, developers of machine learning systems do not often care about the license of the dataset when training machine learning models with it. In summary, the licensing of datasets for machine learning systems is in a state of incompleteness in all aspects at this stage. Our investigation of two collection datasets revealed that most of the current datasets lacked licenses, and the lack of licenses made it impossible to determine the commercial availability of the datasets. Therefore, we decided to take a more scientific and systematic approach to investigate the licensing of datasets and the licensing of machine learning systems that use the dataset to make it easier and more compliant for future developers of machine learning systems.
{ "abstract": "Dataset licensing is currently an issue in the development of machine\nlearning systems. And in the development of machine learning systems, the most\nwidely used are publicly available datasets. However, since the images in the\npublicly available dataset are mainly obtained from the Internet, some images\nare not commercially available. Furthermore, developers of machine learning\nsystems do not often care about the license of the dataset when training\nmachine learning models with it. In summary, the licensing of datasets for\nmachine learning systems is in a state of incompleteness in all aspects at this\nstage.\n Our investigation of two collection datasets revealed that most of the\ncurrent datasets lacked licenses, and the lack of licenses made it impossible\nto determine the commercial availability of the datasets. Therefore, we decided\nto take a more scientific and systematic approach to investigate the licensing\nof datasets and the licensing of machine learning systems that use the dataset\nto make it easier and more compliant for future developers of machine learning\nsystems.", "title": "An investigation of licensing of datasets for machine learning based on the GQM model", "url": "http://arxiv.org/abs/2303.13735v1" }
null
null
no_new_dataset
admin
null
false
null
0859a6fa-09bc-4167-9272-370e9ffd4a82
null
Validated
2023-10-04 15:19:51.880245
{ "text_length": 1212 }
1no_new_dataset
TITLE: SignBank+: Multilingual Sign Language Translation Dataset ABSTRACT: This work advances the field of sign language machine translation by focusing on dataset quality and simplification of the translation system. We introduce SignBank+, a clean version of the SignBank dataset, optimized for machine translation. Contrary to previous works that employ complex factorization techniques for translation, we advocate for a simplified text-to-text translation approach. Our evaluation shows that models trained on SignBank+ surpass those on the original dataset, establishing a new benchmark and providing an open resource for future research.
{ "abstract": "This work advances the field of sign language machine translation by focusing\non dataset quality and simplification of the translation system. We introduce\nSignBank+, a clean version of the SignBank dataset, optimized for machine\ntranslation. Contrary to previous works that employ complex factorization\ntechniques for translation, we advocate for a simplified text-to-text\ntranslation approach. Our evaluation shows that models trained on SignBank+\nsurpass those on the original dataset, establishing a new benchmark and\nproviding an open resource for future research.", "title": "SignBank+: Multilingual Sign Language Translation Dataset", "url": "http://arxiv.org/abs/2309.11566v1" }
null
null
new_dataset
admin
null
false
null
c1f9ba01-67c4-4b08-88f7-4d5c4caf890a
null
Validated
2023-10-04 15:19:51.863370
{ "text_length": 661 }
0new_dataset
TITLE: Harvard Glaucoma Fairness: A Retinal Nerve Disease Dataset for Fairness Learning and Fair Identity Normalization ABSTRACT: Fairness (also known as equity interchangeably) in machine learning is important for societal well-being, but limited public datasets hinder its progress. Currently, no dedicated public medical datasets with imaging data for fairness learning are available, though minority groups suffer from more health issues. To address this gap, we introduce Harvard Glaucoma Fairness (Harvard-GF), a retinal nerve disease dataset with both 2D and 3D imaging data and balanced racial groups for glaucoma detection. Glaucoma is the leading cause of irreversible blindness globally with Blacks having doubled glaucoma prevalence than other races. We also propose a fair identity normalization (FIN) approach to equalize the feature importance between different identity groups. Our FIN approach is compared with various the-state-of-the-art fairness learning methods with superior performance in the racial, gender, and ethnicity fairness tasks with 2D and 3D imaging data, which demonstrate the utilities of our dataset Harvard-GF for fairness learning. To facilitate fairness comparisons between different models, we propose an equity-scaled performance measure, which can be flexibly used to compare all kinds of performance metrics in the context of fairness. The dataset and code are publicly accessible via \url{https://ophai.hms.harvard.edu/datasets/harvard-glaucoma-fairness-3300-samples/}.
{ "abstract": "Fairness (also known as equity interchangeably) in machine learning is\nimportant for societal well-being, but limited public datasets hinder its\nprogress. Currently, no dedicated public medical datasets with imaging data for\nfairness learning are available, though minority groups suffer from more health\nissues. To address this gap, we introduce Harvard Glaucoma Fairness\n(Harvard-GF), a retinal nerve disease dataset with both 2D and 3D imaging data\nand balanced racial groups for glaucoma detection. Glaucoma is the leading\ncause of irreversible blindness globally with Blacks having doubled glaucoma\nprevalence than other races. We also propose a fair identity normalization\n(FIN) approach to equalize the feature importance between different identity\ngroups. Our FIN approach is compared with various the-state-of-the-art fairness\nlearning methods with superior performance in the racial, gender, and ethnicity\nfairness tasks with 2D and 3D imaging data, which demonstrate the utilities of\nour dataset Harvard-GF for fairness learning. To facilitate fairness\ncomparisons between different models, we propose an equity-scaled performance\nmeasure, which can be flexibly used to compare all kinds of performance metrics\nin the context of fairness. The dataset and code are publicly accessible via\n\\url{https://ophai.hms.harvard.edu/datasets/harvard-glaucoma-fairness-3300-samples/}.", "title": "Harvard Glaucoma Fairness: A Retinal Nerve Disease Dataset for Fairness Learning and Fair Identity Normalization", "url": "http://arxiv.org/abs/2306.09264v2" }
null
null
new_dataset
admin
null
false
null
646c74b0-2d86-4afa-821e-831c8c9baadb
null
Validated
2023-10-04 15:19:51.870770
{ "text_length": 1531 }
0new_dataset
TITLE: DynaBench: A benchmark dataset for learning dynamical systems from low-resolution data ABSTRACT: Previous work on learning physical systems from data has focused on high-resolution grid-structured measurements. However, real-world knowledge of such systems (e.g. weather data) relies on sparsely scattered measuring stations. In this paper, we introduce a novel simulated benchmark dataset, DynaBench, for learning dynamical systems directly from sparsely scattered data without prior knowledge of the equations. The dataset focuses on predicting the evolution of a dynamical system from low-resolution, unstructured measurements. We simulate six different partial differential equations covering a variety of physical systems commonly used in the literature and evaluate several machine learning models, including traditional graph neural networks and point cloud processing models, with the task of predicting the evolution of the system. The proposed benchmark dataset is expected to advance the state of art as an out-of-the-box easy-to-use tool for evaluating models in a setting where only unstructured low-resolution observations are available. The benchmark is available at https://anonymous.4open.science/r/code-2022-dynabench/.
{ "abstract": "Previous work on learning physical systems from data has focused on\nhigh-resolution grid-structured measurements. However, real-world knowledge of\nsuch systems (e.g. weather data) relies on sparsely scattered measuring\nstations. In this paper, we introduce a novel simulated benchmark dataset,\nDynaBench, for learning dynamical systems directly from sparsely scattered data\nwithout prior knowledge of the equations. The dataset focuses on predicting the\nevolution of a dynamical system from low-resolution, unstructured measurements.\nWe simulate six different partial differential equations covering a variety of\nphysical systems commonly used in the literature and evaluate several machine\nlearning models, including traditional graph neural networks and point cloud\nprocessing models, with the task of predicting the evolution of the system. The\nproposed benchmark dataset is expected to advance the state of art as an\nout-of-the-box easy-to-use tool for evaluating models in a setting where only\nunstructured low-resolution observations are available. The benchmark is\navailable at https://anonymous.4open.science/r/code-2022-dynabench/.", "title": "DynaBench: A benchmark dataset for learning dynamical systems from low-resolution data", "url": "http://arxiv.org/abs/2306.05805v2" }
null
null
new_dataset
admin
null
false
null
ea320f3e-d049-455e-adcc-bfbdd1fa8a2b
null
Validated
2023-10-04 15:19:51.874159
{ "text_length": 1261 }
0new_dataset
TITLE: Continuous User Authentication Using Machine Learning and Multi-Finger Mobile Touch Dynamics with a Novel Dataset ABSTRACT: As technology grows and evolves rapidly, it is increasingly clear that mobile devices are more commonly used for sensitive matters than ever before. A need to authenticate users continuously is sought after as a single-factor or multi factor authentication may only initially validate a user, which does not help if an impostor can bypass this initial validation. The field of touch dynamics emerges as a clear way to non intrusively collect data about a user and their behaviors in order to develop and make imperative security related decisions in real time. In this paper we present a novel dataset consisting of tracking 25 users playing two mobile games Snake.io and Minecraft each for 10 minutes, along with their relevant gesture data. From this data, we ran machine learning binary classifiers namely Random Forest and K Nearest Neighbor to attempt to authenticate whether a sample of a particular users actions were genuine. Our strongest model returned an average accuracy of roughly 93% for both games, showing touch dynamics can differentiate users effectively and is a feasible consideration for authentication schemes. Our dataset can be observed at https://github.com/zderidder/MC-Snake-Results
{ "abstract": "As technology grows and evolves rapidly, it is increasingly clear that mobile\ndevices are more commonly used for sensitive matters than ever before. A need\nto authenticate users continuously is sought after as a single-factor or multi\nfactor authentication may only initially validate a user, which does not help\nif an impostor can bypass this initial validation. The field of touch dynamics\nemerges as a clear way to non intrusively collect data about a user and their\nbehaviors in order to develop and make imperative security related decisions in\nreal time. In this paper we present a novel dataset consisting of tracking 25\nusers playing two mobile games Snake.io and Minecraft each for 10 minutes,\nalong with their relevant gesture data. From this data, we ran machine learning\nbinary classifiers namely Random Forest and K Nearest Neighbor to attempt to\nauthenticate whether a sample of a particular users actions were genuine. Our\nstrongest model returned an average accuracy of roughly 93% for both games,\nshowing touch dynamics can differentiate users effectively and is a feasible\nconsideration for authentication schemes. Our dataset can be observed at\nhttps://github.com/zderidder/MC-Snake-Results", "title": "Continuous User Authentication Using Machine Learning and Multi-Finger Mobile Touch Dynamics with a Novel Dataset", "url": "http://arxiv.org/abs/2207.13648v1" }
null
null
new_dataset
admin
null
false
null
e601727a-1174-4c9c-84e8-b1d2abb35093
null
Validated
2023-10-04 15:19:51.885077
{ "text_length": 1357 }
0new_dataset
TITLE: Model Evaluation in Medical Datasets Over Time ABSTRACT: Machine learning models deployed in healthcare systems face data drawn from continually evolving environments. However, researchers proposing such models typically evaluate them in a time-agnostic manner, with train and test splits sampling patients throughout the entire study period. We introduce the Evaluation on Medical Datasets Over Time (EMDOT) framework and Python package, which evaluates the performance of a model class over time. Across five medical datasets and a variety of models, we compare two training strategies: (1) using all historical data, and (2) using a window of the most recent data. We note changes in performance over time, and identify possible explanations for these shocks.
{ "abstract": "Machine learning models deployed in healthcare systems face data drawn from\ncontinually evolving environments. However, researchers proposing such models\ntypically evaluate them in a time-agnostic manner, with train and test splits\nsampling patients throughout the entire study period. We introduce the\nEvaluation on Medical Datasets Over Time (EMDOT) framework and Python package,\nwhich evaluates the performance of a model class over time. Across five medical\ndatasets and a variety of models, we compare two training strategies: (1) using\nall historical data, and (2) using a window of the most recent data. We note\nchanges in performance over time, and identify possible explanations for these\nshocks.", "title": "Model Evaluation in Medical Datasets Over Time", "url": "http://arxiv.org/abs/2211.07165v1" }
null
null
no_new_dataset
admin
null
false
null
14c4fda3-5210-42ac-b939-bbe5e881a6bc
null
Validated
2023-10-04 15:19:51.882842
{ "text_length": 786 }
1no_new_dataset
TITLE: LEMMA: A Multi-view Dataset for Learning Multi-agent Multi-task Activities ABSTRACT: Understanding and interpreting human actions is a long-standing challenge and a critical indicator of perception in artificial intelligence. However, a few imperative components of daily human activities are largely missed in prior literature, including the goal-directed actions, concurrent multi-tasks, and collaborations among multi-agents. We introduce the LEMMA dataset to provide a single home to address these missing dimensions with meticulously designed settings, wherein the number of tasks and agents varies to highlight different learning objectives. We densely annotate the atomic-actions with human-object interactions to provide ground-truths of the compositionality, scheduling, and assignment of daily activities. We further devise challenging compositional action recognition and action/task anticipation benchmarks with baseline models to measure the capability of compositional action understanding and temporal reasoning. We hope this effort would drive the machine vision community to examine goal-directed human activities and further study the task scheduling and assignment in the real world.
{ "abstract": "Understanding and interpreting human actions is a long-standing challenge and\na critical indicator of perception in artificial intelligence. However, a few\nimperative components of daily human activities are largely missed in prior\nliterature, including the goal-directed actions, concurrent multi-tasks, and\ncollaborations among multi-agents. We introduce the LEMMA dataset to provide a\nsingle home to address these missing dimensions with meticulously designed\nsettings, wherein the number of tasks and agents varies to highlight different\nlearning objectives. We densely annotate the atomic-actions with human-object\ninteractions to provide ground-truths of the compositionality, scheduling, and\nassignment of daily activities. We further devise challenging compositional\naction recognition and action/task anticipation benchmarks with baseline models\nto measure the capability of compositional action understanding and temporal\nreasoning. We hope this effort would drive the machine vision community to\nexamine goal-directed human activities and further study the task scheduling\nand assignment in the real world.", "title": "LEMMA: A Multi-view Dataset for Learning Multi-agent Multi-task Activities", "url": "http://arxiv.org/abs/2007.15781v1" }
null
null
new_dataset
admin
null
false
null
238beaef-fde6-4639-afd6-26f57f4322dd
null
Validated
2023-10-04 15:19:51.898971
{ "text_length": 1226 }
0new_dataset
TITLE: WikiLingua: A New Benchmark Dataset for Cross-Lingual Abstractive Summarization ABSTRACT: We introduce WikiLingua, a large-scale, multilingual dataset for the evaluation of crosslingual abstractive summarization systems. We extract article and summary pairs in 18 languages from WikiHow, a high quality, collaborative resource of how-to guides on a diverse set of topics written by human authors. We create gold-standard article-summary alignments across languages by aligning the images that are used to describe each how-to step in an article. As a set of baselines for further studies, we evaluate the performance of existing cross-lingual abstractive summarization methods on our dataset. We further propose a method for direct crosslingual summarization (i.e., without requiring translation at inference time) by leveraging synthetic data and Neural Machine Translation as a pre-training step. Our method significantly outperforms the baseline approaches, while being more cost efficient during inference.
{ "abstract": "We introduce WikiLingua, a large-scale, multilingual dataset for the\nevaluation of crosslingual abstractive summarization systems. We extract\narticle and summary pairs in 18 languages from WikiHow, a high quality,\ncollaborative resource of how-to guides on a diverse set of topics written by\nhuman authors. We create gold-standard article-summary alignments across\nlanguages by aligning the images that are used to describe each how-to step in\nan article. As a set of baselines for further studies, we evaluate the\nperformance of existing cross-lingual abstractive summarization methods on our\ndataset. We further propose a method for direct crosslingual summarization\n(i.e., without requiring translation at inference time) by leveraging synthetic\ndata and Neural Machine Translation as a pre-training step. Our method\nsignificantly outperforms the baseline approaches, while being more cost\nefficient during inference.", "title": "WikiLingua: A New Benchmark Dataset for Cross-Lingual Abstractive Summarization", "url": "http://arxiv.org/abs/2010.03093v1" }
null
null
new_dataset
admin
null
false
null
2b06bf83-4565-40c7-bd94-55d327c90489
null
Validated
2023-10-04 15:19:51.897786
{ "text_length": 1034 }
0new_dataset
TITLE: Feature Selection on Thermal-stress Dataset ABSTRACT: Physical symptoms caused by high stress commonly happen in our daily lives, leading to the importance of stress recognition systems. This study aims to improve stress classification by selecting appropriate features from Thermal-stress data, ANUstressDB. We explored three different feature selection techniques: correlation analysis, magnitude measure, and genetic algorithm. Support Vector Machine (SVM) and Artificial Neural Network (ANN) models were involved in measuring these three algorithms. Our result indicates that the genetic algorithm combined with ANNs can improve the prediction accuracy by 19.1% compared to the baseline. Moreover, the magnitude measure performed best among the three feature selection algorithms regarding the balance of computation time and performance. These findings are likely to improve the accuracy of current stress recognition systems.
{ "abstract": "Physical symptoms caused by high stress commonly happen in our daily lives,\nleading to the importance of stress recognition systems. This study aims to\nimprove stress classification by selecting appropriate features from\nThermal-stress data, ANUstressDB. We explored three different feature selection\ntechniques: correlation analysis, magnitude measure, and genetic algorithm.\nSupport Vector Machine (SVM) and Artificial Neural Network (ANN) models were\ninvolved in measuring these three algorithms. Our result indicates that the\ngenetic algorithm combined with ANNs can improve the prediction accuracy by\n19.1% compared to the baseline. Moreover, the magnitude measure performed best\namong the three feature selection algorithms regarding the balance of\ncomputation time and performance. These findings are likely to improve the\naccuracy of current stress recognition systems.", "title": "Feature Selection on Thermal-stress Dataset", "url": "http://arxiv.org/abs/2109.03755v1" }
null
null
no_new_dataset
admin
null
false
null
6f0cc4c4-15fe-4882-8d9b-8522e3d34298
null
Validated
2023-10-04 15:19:51.892243
{ "text_length": 955 }
1no_new_dataset
TITLE: Binary classifiers for noisy datasets: a comparative study of existing quantum machine learning frameworks and some new approaches ABSTRACT: One of the most promising areas of research to obtain practical advantage is Quantum Machine Learning which was born as a result of cross-fertilisation of ideas between Quantum Computing and Classical Machine Learning. In this paper, we apply Quantum Machine Learning (QML) frameworks to improve binary classification models for noisy datasets which are prevalent in financial datasets. The metric we use for assessing the performance of our quantum classifiers is the area under the receiver operating characteristic curve (ROC/AUC). By combining such approaches as hybrid-neural networks, parametric circuits, and data re-uploading we create QML inspired architectures and utilise them for the classification of non-convex 2 and 3-dimensional figures. An extensive benchmarking of our new FULL HYBRID classifiers against existing quantum and classical classifier models, reveals that our novel models exhibit better learning characteristics to asymmetrical Gaussian noise in the dataset compared to known quantum classifiers and performs equally well for existing classical classifiers, with a slight improvement over classical results in the region of the high noise.
{ "abstract": "One of the most promising areas of research to obtain practical advantage is\nQuantum Machine Learning which was born as a result of cross-fertilisation of\nideas between Quantum Computing and Classical Machine Learning. In this paper,\nwe apply Quantum Machine Learning (QML) frameworks to improve binary\nclassification models for noisy datasets which are prevalent in financial\ndatasets. The metric we use for assessing the performance of our quantum\nclassifiers is the area under the receiver operating characteristic curve\n(ROC/AUC). By combining such approaches as hybrid-neural networks, parametric\ncircuits, and data re-uploading we create QML inspired architectures and\nutilise them for the classification of non-convex 2 and 3-dimensional figures.\nAn extensive benchmarking of our new FULL HYBRID classifiers against existing\nquantum and classical classifier models, reveals that our novel models exhibit\nbetter learning characteristics to asymmetrical Gaussian noise in the dataset\ncompared to known quantum classifiers and performs equally well for existing\nclassical classifiers, with a slight improvement over classical results in the\nregion of the high noise.", "title": "Binary classifiers for noisy datasets: a comparative study of existing quantum machine learning frameworks and some new approaches", "url": "http://arxiv.org/abs/2111.03372v1" }
null
null
no_new_dataset
admin
null
false
null
b301c76c-d50e-461a-8b0a-6503883f321f
null
Validated
2023-10-04 15:19:51.890085
{ "text_length": 1335 }
1no_new_dataset
TITLE: Active label cleaning for improved dataset quality under resource constraints ABSTRACT: Imperfections in data annotation, known as label noise, are detrimental to the training of machine learning models and have an often-overlooked confounding effect on the assessment of model performance. Nevertheless, employing experts to remove label noise by fully re-annotating large datasets is infeasible in resource-constrained settings, such as healthcare. This work advocates for a data-driven approach to prioritising samples for re-annotation - which we term "active label cleaning". We propose to rank instances according to estimated label correctness and labelling difficulty of each sample, and introduce a simulation framework to evaluate relabelling efficacy. Our experiments on natural images and on a new medical imaging benchmark show that cleaning noisy labels mitigates their negative impact on model training, evaluation, and selection. Crucially, the proposed active label cleaning enables correcting labels up to 4 times more effectively than typical random selection in realistic conditions, making better use of experts' valuable time for improving dataset quality.
{ "abstract": "Imperfections in data annotation, known as label noise, are detrimental to\nthe training of machine learning models and have an often-overlooked\nconfounding effect on the assessment of model performance. Nevertheless,\nemploying experts to remove label noise by fully re-annotating large datasets\nis infeasible in resource-constrained settings, such as healthcare. This work\nadvocates for a data-driven approach to prioritising samples for re-annotation\n- which we term \"active label cleaning\". We propose to rank instances according\nto estimated label correctness and labelling difficulty of each sample, and\nintroduce a simulation framework to evaluate relabelling efficacy. Our\nexperiments on natural images and on a new medical imaging benchmark show that\ncleaning noisy labels mitigates their negative impact on model training,\nevaluation, and selection. Crucially, the proposed active label cleaning\nenables correcting labels up to 4 times more effectively than typical random\nselection in realistic conditions, making better use of experts' valuable time\nfor improving dataset quality.", "title": "Active label cleaning for improved dataset quality under resource constraints", "url": "http://arxiv.org/abs/2109.00574v2" }
null
null
no_new_dataset
admin
null
false
null
2664980c-6dd3-4e96-9645-ed72da54a84b
null
Validated
2023-10-04 15:19:51.892338
{ "text_length": 1202 }
1no_new_dataset
TITLE: Dataset Inference: Ownership Resolution in Machine Learning ABSTRACT: With increasingly more data and computation involved in their training, machine learning models constitute valuable intellectual property. This has spurred interest in model stealing, which is made more practical by advances in learning with partial, little, or no supervision. Existing defenses focus on inserting unique watermarks in a model's decision surface, but this is insufficient: the watermarks are not sampled from the training distribution and thus are not always preserved during model stealing. In this paper, we make the key observation that knowledge contained in the stolen model's training set is what is common to all stolen copies. The adversary's goal, irrespective of the attack employed, is always to extract this knowledge or its by-products. This gives the original model's owner a strong advantage over the adversary: model owners have access to the original training data. We thus introduce $dataset$ $inference$, the process of identifying whether a suspected model copy has private knowledge from the original model's dataset, as a defense against model stealing. We develop an approach for dataset inference that combines statistical testing with the ability to estimate the distance of multiple data points to the decision boundary. Our experiments on CIFAR10, SVHN, CIFAR100 and ImageNet show that model owners can claim with confidence greater than 99% that their model (or dataset as a matter of fact) was stolen, despite only exposing 50 of the stolen model's training points. Dataset inference defends against state-of-the-art attacks even when the adversary is adaptive. Unlike prior work, it does not require retraining or overfitting the defended model.
{ "abstract": "With increasingly more data and computation involved in their training,\nmachine learning models constitute valuable intellectual property. This has\nspurred interest in model stealing, which is made more practical by advances in\nlearning with partial, little, or no supervision. Existing defenses focus on\ninserting unique watermarks in a model's decision surface, but this is\ninsufficient: the watermarks are not sampled from the training distribution and\nthus are not always preserved during model stealing. In this paper, we make the\nkey observation that knowledge contained in the stolen model's training set is\nwhat is common to all stolen copies. The adversary's goal, irrespective of the\nattack employed, is always to extract this knowledge or its by-products. This\ngives the original model's owner a strong advantage over the adversary: model\nowners have access to the original training data. We thus introduce $dataset$\n$inference$, the process of identifying whether a suspected model copy has\nprivate knowledge from the original model's dataset, as a defense against model\nstealing. We develop an approach for dataset inference that combines\nstatistical testing with the ability to estimate the distance of multiple data\npoints to the decision boundary. Our experiments on CIFAR10, SVHN, CIFAR100 and\nImageNet show that model owners can claim with confidence greater than 99% that\ntheir model (or dataset as a matter of fact) was stolen, despite only exposing\n50 of the stolen model's training points. Dataset inference defends against\nstate-of-the-art attacks even when the adversary is adaptive. Unlike prior\nwork, it does not require retraining or overfitting the defended model.", "title": "Dataset Inference: Ownership Resolution in Machine Learning", "url": "http://arxiv.org/abs/2104.10706v1" }
null
null
no_new_dataset
admin
null
false
null
34e24349-1528-48dd-a7cc-b25e36470f4d
null
Validated
2023-10-04 15:19:51.894907
{ "text_length": 1786 }
1no_new_dataset
TITLE: CLIP: A Dataset for Extracting Action Items for Physicians from Hospital Discharge Notes ABSTRACT: Continuity of care is crucial to ensuring positive health outcomes for patients discharged from an inpatient hospital setting, and improved information sharing can help. To share information, caregivers write discharge notes containing action items to share with patients and their future caregivers, but these action items are easily lost due to the lengthiness of the documents. In this work, we describe our creation of a dataset of clinical action items annotated over MIMIC-III, the largest publicly available dataset of real clinical notes. This dataset, which we call CLIP, is annotated by physicians and covers 718 documents representing 100K sentences. We describe the task of extracting the action items from these documents as multi-aspect extractive summarization, with each aspect representing a type of action to be taken. We evaluate several machine learning models on this task, and show that the best models exploit in-domain language model pre-training on 59K unannotated documents, and incorporate context from neighboring sentences. We also propose an approach to pre-training data selection that allows us to explore the trade-off between size and domain-specificity of pre-training datasets for this task.
{ "abstract": "Continuity of care is crucial to ensuring positive health outcomes for\npatients discharged from an inpatient hospital setting, and improved\ninformation sharing can help. To share information, caregivers write discharge\nnotes containing action items to share with patients and their future\ncaregivers, but these action items are easily lost due to the lengthiness of\nthe documents. In this work, we describe our creation of a dataset of clinical\naction items annotated over MIMIC-III, the largest publicly available dataset\nof real clinical notes. This dataset, which we call CLIP, is annotated by\nphysicians and covers 718 documents representing 100K sentences. We describe\nthe task of extracting the action items from these documents as multi-aspect\nextractive summarization, with each aspect representing a type of action to be\ntaken. We evaluate several machine learning models on this task, and show that\nthe best models exploit in-domain language model pre-training on 59K\nunannotated documents, and incorporate context from neighboring sentences. We\nalso propose an approach to pre-training data selection that allows us to\nexplore the trade-off between size and domain-specificity of pre-training\ndatasets for this task.", "title": "CLIP: A Dataset for Extracting Action Items for Physicians from Hospital Discharge Notes", "url": "http://arxiv.org/abs/2106.02524v1" }
null
null
new_dataset
admin
null
false
null
00207d9e-f241-43fd-81d6-65b657045f7d
null
Validated
2023-10-04 15:19:51.894350
{ "text_length": 1350 }
0new_dataset
TITLE: SANGEET: A XML based Open Dataset for Research in Hindustani Sangeet ABSTRACT: It is very important to access a rich music dataset that is useful in a wide variety of applications. Currently, available datasets are mostly focused on storing vocal or instrumental recording data and ignoring the requirement of its visual representation and retrieval. This paper attempts to build an XML-based public dataset, called SANGEET, that stores comprehensive information of Hindustani Sangeet (North Indian Classical Music) compositions written by famous musicologist Pt. Vishnu Narayan Bhatkhande. SANGEET preserves all the required information of any given composition including metadata, structural, notational, rhythmic, and melodic information in a standardized way for easy and efficient storage and extraction of musical information. The dataset is intended to provide the ground truth information for music information research tasks, thereby supporting several data-driven analysis from a machine learning perspective. We present the usefulness of the dataset by demonstrating its application on music information retrieval using XQuery, visualization through Omenad rendering system. Finally, we propose approaches to transform the dataset for performing statistical and machine learning tasks for a better understanding of Hindustani Sangeet. The dataset can be found at https://github.com/cmisra/Sangeet.
{ "abstract": "It is very important to access a rich music dataset that is useful in a wide\nvariety of applications. Currently, available datasets are mostly focused on\nstoring vocal or instrumental recording data and ignoring the requirement of\nits visual representation and retrieval. This paper attempts to build an\nXML-based public dataset, called SANGEET, that stores comprehensive information\nof Hindustani Sangeet (North Indian Classical Music) compositions written by\nfamous musicologist Pt. Vishnu Narayan Bhatkhande. SANGEET preserves all the\nrequired information of any given composition including metadata, structural,\nnotational, rhythmic, and melodic information in a standardized way for easy\nand efficient storage and extraction of musical information. The dataset is\nintended to provide the ground truth information for music information research\ntasks, thereby supporting several data-driven analysis from a machine learning\nperspective. We present the usefulness of the dataset by demonstrating its\napplication on music information retrieval using XQuery, visualization through\nOmenad rendering system. Finally, we propose approaches to transform the\ndataset for performing statistical and machine learning tasks for a better\nunderstanding of Hindustani Sangeet. The dataset can be found at\nhttps://github.com/cmisra/Sangeet.", "title": "SANGEET: A XML based Open Dataset for Research in Hindustani Sangeet", "url": "http://arxiv.org/abs/2306.04148v1" }
null
null
new_dataset
admin
null
false
null
48fd9e3f-2663-4a46-8703-9a85fc73987b
null
Validated
2023-10-04 15:19:51.874454
{ "text_length": 1432 }
0new_dataset
TITLE: RDD2022: A multi-national image dataset for automatic Road Damage Detection ABSTRACT: The data article describes the Road Damage Dataset, RDD2022, which comprises 47,420 road images from six countries, Japan, India, the Czech Republic, Norway, the United States, and China. The images have been annotated with more than 55,000 instances of road damage. Four types of road damage, namely longitudinal cracks, transverse cracks, alligator cracks, and potholes, are captured in the dataset. The annotated dataset is envisioned for developing deep learning-based methods to detect and classify road damage automatically. The dataset has been released as a part of the Crowd sensing-based Road Damage Detection Challenge (CRDDC2022). The challenge CRDDC2022 invites researchers from across the globe to propose solutions for automatic road damage detection in multiple countries. The municipalities and road agencies may utilize the RDD2022 dataset, and the models trained using RDD2022 for low-cost automatic monitoring of road conditions. Further, computer vision and machine learning researchers may use the dataset to benchmark the performance of different algorithms for other image-based applications of the same type (classification, object detection, etc.).
{ "abstract": "The data article describes the Road Damage Dataset, RDD2022, which comprises\n47,420 road images from six countries, Japan, India, the Czech Republic,\nNorway, the United States, and China. The images have been annotated with more\nthan 55,000 instances of road damage. Four types of road damage, namely\nlongitudinal cracks, transverse cracks, alligator cracks, and potholes, are\ncaptured in the dataset. The annotated dataset is envisioned for developing\ndeep learning-based methods to detect and classify road damage automatically.\nThe dataset has been released as a part of the Crowd sensing-based Road Damage\nDetection Challenge (CRDDC2022). The challenge CRDDC2022 invites researchers\nfrom across the globe to propose solutions for automatic road damage detection\nin multiple countries. The municipalities and road agencies may utilize the\nRDD2022 dataset, and the models trained using RDD2022 for low-cost automatic\nmonitoring of road conditions. Further, computer vision and machine learning\nresearchers may use the dataset to benchmark the performance of different\nalgorithms for other image-based applications of the same type (classification,\nobject detection, etc.).", "title": "RDD2022: A multi-national image dataset for automatic Road Damage Detection", "url": "http://arxiv.org/abs/2209.08538v1" }
null
null
new_dataset
admin
null
false
null
c0bcb794-f281-44ee-819a-6bb0253283a0
null
Validated
2023-10-04 15:19:51.883925
{ "text_length": 1284 }
0new_dataset