paper_id
stringlengths 10
10
| paper_url
stringlengths 37
80
| title
stringlengths 4
518
| abstract
stringlengths 3
7.27k
| arxiv_id
stringlengths 9
16
⌀ | url_abs
stringlengths 18
601
| url_pdf
stringlengths 21
601
| aspect_tasks
sequence | aspect_methods
sequence | aspect_datasets
sequence |
---|---|---|---|---|---|---|---|---|---|
u5kx-s23o9 | https://paperswithcode.com/paper/high-resolution-medical-image-analysis-with | High Resolution Medical Image Analysis with Spatial Partitioning | Medical images such as 3D computerized tomography (CT) scans and pathology images, have hundreds of millions or billions of voxels/pixels. It is infeasible to train CNN models directly on such high resolution images, because neural activations of a single image do not fit in the memory of a single GPU/TPU, and naive data and model parallelism approaches do not work. Existing image analysis approaches alleviate this problem by cropping or down-sampling input images, which leads to complicated implementation and sub-optimal performance due to information loss. In this paper, we implement spatial partitioning, which internally distributes the input and output of convolutional layers across GPUs/TPUs. Our implementation is based on the Mesh-TensorFlow framework and the computation distribution is transparent to end users. With this technique, we train a 3D Unet on up to 512 by 512 by 512 resolution data. To the best of our knowledge, this is the first work for handling such high resolution images end-to-end. | 1909.03108 | https://arxiv.org/abs/1909.03108v3 | https://arxiv.org/pdf/1909.03108v3.pdf | [] | [] | [] |
sn-XcX5t6G | https://paperswithcode.com/paper/on-the-origin-of-deep-learning | On the Origin of Deep Learning | This paper is a review of the evolutionary history of deep learning models.
It covers from the genesis of neural networks when associationism modeling of
the brain is studied, to the models that dominate the last decade of research
in deep learning like convolutional neural networks, deep belief networks, and
recurrent neural networks. In addition to a review of these models, this paper
primarily focuses on the precedents of the models above, examining how the
initial ideas are assembled to construct the early models and how these
preliminary models are developed into their current forms. Many of these
evolutionary paths last more than half a century and have a diversity of
directions. For example, CNN is built on prior knowledge of biological vision
system; DBN is evolved from a trade-off of modeling power and computation
complexity of graphical models and many nowadays models are neural counterparts
of ancient linear models. This paper reviews these evolutionary paths and
offers a concise thought flow of how these models are developed, and aims to
provide a thorough background for deep learning. More importantly, along with
the path, this paper summarizes the gist behind these milestones and proposes
many directions to guide the future research of deep learning. | 1702.07800 | http://arxiv.org/abs/1702.07800v4 | http://arxiv.org/pdf/1702.07800v4.pdf | [] | [] | [] |
ozdxIKOlTH | https://paperswithcode.com/paper/deep-learning-predicts-total-knee-replacement | Deep learning predicts total knee replacement from magnetic resonance images | Knee Osteoarthritis (OA) is a common musculoskeletal disorder in the United States. When diagnosed at early stages, lifestyle interventions such as exercise and weight loss can slow OA progression, but at later stages, only an invasive option is available: total knee replacement (TKR). Though a generally successful procedure, only 2/3 of patients who undergo the procedure report their knees feeling ''normal'' post-operation, and complications can arise that require revision. This necessitates a model to identify a population at higher risk of TKR, particularly at less advanced stages of OA, such that appropriate treatments can be implemented that slow OA progression and delay TKR. Here, we present a deep learning pipeline that leverages MRI images and clinical and demographic information to predict TKR with AUC $0.834 \pm 0.036$ (p < 0.05). Most notably, the pipeline predicts TKR with AUC $0.943 \pm 0.057$ (p < 0.05) for patients without OA. Furthermore, we develop occlusion maps for case-control pairs in test data and compare regions used by the model in both, thereby identifying TKR imaging biomarkers. As such, this work takes strides towards a pipeline with clinical utility, and the biomarkers identified further our understanding of OA progression and eventual TKR onset. | 2002.10591 | https://arxiv.org/abs/2002.10591v1 | https://arxiv.org/pdf/2002.10591v1.pdf | [] | [] | [] |
f6D2xZQLNv | https://paperswithcode.com/paper/building-statistical-shape-spaces-for-3d | Building Statistical Shape Spaces for 3D Human Modeling | Statistical models of 3D human shape and pose learned from scan databases
have developed into valuable tools to solve a variety of vision and graphics
problems. Unfortunately, most publicly available models are of limited
expressiveness as they were learned on very small databases that hardly reflect
the true variety in human body shapes. In this paper, we contribute by
rebuilding a widely used statistical body representation from the largest
commercially available scan database, and making the resulting model available
to the community (visit http://humanshape.mpi-inf.mpg.de). As preprocessing
several thousand scans for learning the model is a challenge in itself, we
contribute by developing robust best practice solutions for scan alignment that
quantitatively lead to the best learned models. We make implementations of
these preprocessing steps also publicly available. We extensively evaluate the
improved accuracy and generality of our new model, and show its improved
performance for human body reconstruction from sparse input data. | 1503.05860 | http://arxiv.org/abs/1503.05860v2 | http://arxiv.org/pdf/1503.05860v2.pdf | [] | [] | [] |
fq_p8PvKw0 | https://paperswithcode.com/paper/a-deep-learning-approach-to-digitally-stain | A Deep Learning Approach to Digitally Stain Optical Coherence Tomography Images of the Optic Nerve Head | Purpose: To develop a deep learning approach to digitally-stain optical
coherence tomography (OCT) images of the optic nerve head (ONH).
Methods: A horizontal B-scan was acquired through the center of the ONH using
OCT (Spectralis) for 1 eye of each of 100 subjects (40 normal & 60 glaucoma).
All images were enhanced using adaptive compensation. A custom deep learning
network was then designed and trained with the compensated images to digitally
stain (i.e. highlight) 6 tissue layers of the ONH. The accuracy of our
algorithm was assessed (against manual segmentations) using the Dice
coefficient, sensitivity, and specificity. We further studied how compensation
and the number of training images affected the performance of our algorithm.
Results: For images it had not yet assessed, our algorithm was able to
digitally stain the retinal nerve fiber layer + prelamina, the retinal pigment
epithelium, all other retinal layers, the choroid, and the peripapillary sclera
and lamina cribrosa. For all tissues, the mean dice coefficient was $0.84 \pm
0.03$, the mean sensitivity $0.92 \pm 0.03$, and the mean specificity $0.99 \pm
0.00$. Our algorithm performed significantly better when compensated images
were used for training. Increasing the number of images (from 10 to 40) to
train our algorithm did not significantly improve performance, except for the
RPE.
Conclusion. Our deep learning algorithm can simultaneously stain neural and
connective tissues in ONH images. Our approach offers a framework to
automatically measure multiple key structural parameters of the ONH that may be
critical to improve glaucoma management. | 1707.07609 | http://arxiv.org/abs/1707.07609v1 | http://arxiv.org/pdf/1707.07609v1.pdf | [] | [] | [] |
868-l1zhB7 | https://paperswithcode.com/paper/faster-unsupervised-semantic-inpainting-a-gan | Faster Unsupervised Semantic Inpainting: A GAN Based Approach | In this paper, we propose to improve the inference speed and visual quality of contemporary baseline of Generative Adversarial Networks (GAN) based unsupervised semantic inpainting. This is made possible with better initialization of the core iterative optimization involved in the framework. To our best knowledge, this is also the first attempt of GAN based video inpainting with consideration to temporal cues. On single image inpainting, we achieve about 4.5-5$\times$ speedup and 80$\times$ on videos compared to baseline. Simultaneously, our method has better spatial and temporal reconstruction qualities as found on three image and one video dataset. | 1908.04968 | https://arxiv.org/abs/1908.04968v1 | https://arxiv.org/pdf/1908.04968v1.pdf | [
"Image Inpainting",
"Video Inpainting"
] | [
"Convolution",
"GAN"
] | [] |
Nc2msNDWxJ | https://paperswithcode.com/paper/specifying-treebanks-outsourcing-parsebanks | Specifying Treebanks, Outsourcing Parsebanks: FinnTreeBank 3 | Corpus-based treebank annotation is known to result in incomplete coverage of mid- and low-frequency linguistic constructions: the linguistic representation and corpus annotation quality are sometimes suboptimal. Large descriptive grammars cover also many mid- and low-frequency constructions. We argue for use of large descriptive grammars and their sample sentences as a basis for specifying higher-coverage grammatical representations. We present an sample case from an ongoing project (FIN-CLARIN FinnTreeBank) where an grammatical representation is documented as an annotator's manual alongside manual annotation of sample sentences extracted from a large descriptive grammar of Finnish. We outline the linguistic representation (morphology and dependency syntax) for Finnish, and show how the resulting `Grammar Definition Corpus' and the documentation is used as a task specification for an external subcontractor for building a parser engine for use in morphological and dependency syntactic analysis of large volumes of Finnish for parsebanking purposes. The resulting corpus, FinnTreeBank 3, is due for release in June 2012, and will contain tens of millions of words from publicly available corpora of Finnish with automatic morphological and dependency syntactic analysis, for use in research on the corpus linguistics and language engineering. | null | https://www.aclweb.org/anthology/L12-1448/ | http://www.lrec-conf.org/proceedings/lrec2012/pdf/766_Paper.pdf | [] | [] | [] |
m9rRay5h1Z | https://paperswithcode.com/paper/checking-chase-termination-over-ontologies-of | Checking Chase Termination over Ontologies of Existential Rules with Equality | The chase is a sound and complete algorithm for conjunctive query answering over ontologies of existential rules with equality. To enable its effective use, we can apply acyclicity notions; that is, sufficient conditions that guarantee chase termination. Unfortunately, most of these notions have only been defined for existential rule sets without equality. A proposed solution to circumvent this issue is to treat equality as an ordinary predicate with an explicit axiomatisation. We empirically show that this solution is not efficient in practice and propose an alternative approach. More precisely, we show that, if the chase terminates for any equality axiomatisation of an ontology, then it terminates for the original ontology (which may contain equality). Therefore, one can apply existing acyclicity notions to check chase termination over an axiomatisation of an ontology and then use the original ontology for reasoning. We show that, in practice, doing so results in a more efficient reasoning procedure. Furthermore, we present equality model-faithful acyclicity, a general acyclicity notion that can be directly applied to ontologies with equality. | 1911.10981 | https://arxiv.org/abs/1911.10981v1 | https://arxiv.org/pdf/1911.10981v1.pdf | [] | [] | [] |
KZUq5PSkhR | https://paperswithcode.com/paper/ijcnlp-2017-task-1-chinese-grammatical-error | IJCNLP-2017 Task 1: Chinese Grammatical Error Diagnosis | This paper presents the IJCNLP 2017 shared task for Chinese grammatical error diagnosis (CGED) which seeks to identify grammatical error types and their range of occurrence within sentences written by learners of Chinese as foreign language. We describe the task definition, data preparation, performance metrics, and evaluation results. Of the 13 teams registered for this shared task, 5 teams developed the system and submitted a total of 13 runs. We expected this evaluation campaign could lead to the development of more advanced NLP techniques for educational applications, especially for Chinese error detection. All data sets with gold standards and scoring scripts are made publicly available to researchers. | null | https://www.aclweb.org/anthology/I17-4001/ | https://www.aclweb.org/anthology/I17-4001 | [
"Grammatical Error Correction"
] | [] | [] |
dIWpeAPbn0 | https://paperswithcode.com/paper/a-tutorial-on-distributed-non-bayesian | A Tutorial on Distributed (Non-Bayesian) Learning: Problem, Algorithms and Results | We overview some results on distributed learning with focus on a family of
recently proposed algorithms known as non-Bayesian social learning. We consider
different approaches to the distributed learning problem and its algorithmic
solutions for the case of finitely many hypotheses. The original centralized
problem is discussed at first, and then followed by a generalization to the
distributed setting. The results on convergence and convergence rate are
presented for both asymptotic and finite time regimes. Various extensions are
discussed such as those dealing with directed time-varying networks, Nesterov's
acceleration technique and a continuum sets of hypothesis. | 1609.07537 | http://arxiv.org/abs/1609.07537v1 | http://arxiv.org/pdf/1609.07537v1.pdf | [] | [] | [] |
wUfW2on9Eg | https://paperswithcode.com/paper/translucent-players-explaining-cooperative | Translucent Players: Explaining Cooperative Behavior in Social Dilemmas | In the last few decades, numerous experiments have shown that humans do not
always behave so as to maximize their material payoff. Cooperative behavior
when non-cooperation is a dominant strategy (with respect to the material
payoffs) is particularly puzzling. Here we propose a novel approach to explain
cooperation, assuming what Halpern and Pass call translucent players.
Typically, players are assumed to be opaque, in the sense that a deviation by
one player in a normal-form game does not affect the strategies used by other
players. But a player may believe that if he switches from one strategy to
another, the fact that he chooses to switch may be visible to the other
players. For example, if he chooses to defect in Prisoner's Dilemma, the other
player may sense his guilt. We show that by assuming translucent players, we
can recover many of the regularities observed in human behavior in well-studied
games such as Prisoner's Dilemma, Traveler's Dilemma, Bertrand Competition, and
the Public Goods game. | 1606.07533 | http://arxiv.org/abs/1606.07533v1 | http://arxiv.org/pdf/1606.07533v1.pdf | [] | [] | [] |
2_JX0VmsuU | https://paperswithcode.com/paper/polish-multimodal-corpus-a-collection-of | Polish Multimodal Corpus --- a collection of referential gestures | In face to face interaction, people refer to objects and events not only by means of speech but also by means of gesture. The present paper describes building a corpus of referential gestures. The aim is to investigate gestural reference by incorporating insights from semantic ontologies and by employing a more holistic view on referential gestures. The paper's focus is on presenting the data collection procedure and discussing the corpus' design; additionally the first insights from constructing the annotation scheme are described. | null | https://www.aclweb.org/anthology/L12-1431/ | http://www.lrec-conf.org/proceedings/lrec2012/pdf/737_Paper.pdf | [] | [] | [] |
kuVXXeETak | https://paperswithcode.com/paper/investigations-of-the-influences-of-a-cnns | Investigations of the Influences of a CNN's Receptive Field on Segmentation of Subnuclei of Bilateral Amygdalae | Segmentation of objects with various sizes is relatively less explored in medical imaging, and has been very challenging in computer vision tasks in general. We hypothesize that the receptive field of a deep model corresponds closely to the size of object to be segmented, which could critically influence the segmentation accuracy of objects with varied sizes. In this study, we employed "AmygNet", a dual-branch fully convolutional neural network (FCNN) with two different sizes of receptive fields, to investigate the effects of receptive field on segmenting four major subnuclei of bilateral amygdalae. The experiment was conducted on 14 subjects, which are all 3-dimensional MRI human brain images. Since the scale of different subnuclear groups are different, by investigating the accuracy of each subnuclear group while using receptive fields of various sizes, we may find which kind of receptive field is suitable for object of which scale respectively. In the given condition, AmygNet with multiple receptive fields presents great potential in segmenting objects of different sizes. | 1911.02761 | https://arxiv.org/abs/1911.02761v1 | https://arxiv.org/pdf/1911.02761v1.pdf | [] | [] | [] |
k7iyHCiCJO | https://paperswithcode.com/paper/automatic-speech-recognition-and-topic | Automatic Speech Recognition and Topic Identification for Almost-Zero-Resource Languages | Automatic speech recognition (ASR) systems often need to be developed for
extremely low-resource languages to serve end-uses such as audio content
categorization and search. While universal phone recognition is natural to
consider when no transcribed speech is available to train an ASR system in a
language, adapting universal phone models using very small amounts (minutes
rather than hours) of transcribed speech also needs to be studied, particularly
with state-of-the-art DNN-based acoustic models. The DARPA LORELEI program
provides a framework for such very-low-resource ASR studies, and provides an
extrinsic metric for evaluating ASR performance in a humanitarian assistance,
disaster relief setting. This paper presents our Kaldi-based systems for the
program, which employ a universal phone modeling approach to ASR, and describes
recipes for very rapid adaptation of this universal ASR system. The results we
obtain significantly outperform results obtained by many competing approaches
on the NIST LoReHLT 2017 Evaluation datasets. | 1802.08731 | http://arxiv.org/abs/1802.08731v2 | http://arxiv.org/pdf/1802.08731v2.pdf | [
"Speech Recognition"
] | [] | [] |
_vF0HW2wdn | https://paperswithcode.com/paper/inferential-text-generation-with-multiple | Inferential Text Generation with Multiple Knowledge Sources and Meta-Learning | We study the problem of generating inferential texts of events for a variety of commonsense like \textit{if-else} relations. Existing approaches typically use limited evidence from training examples and learn for each relation individually. In this work, we use multiple knowledge sources as fuels for the model. Existing commonsense knowledge bases like ConceptNet are dominated by taxonomic knowledge (e.g., \textit{isA} and \textit{relatedTo} relations), having a limited number of inferential knowledge. We use not only structured commonsense knowledge bases, but also natural language snippets from search-engine results. These sources are incorporated into a generative base model via key-value memory network. In addition, we introduce a meta-learning based multi-task learning algorithm. For each targeted commonsense relation, we regard the learning of examples from other relations as the meta-training process, and the evaluation on examples from the targeted relation as the meta-test process. We conduct experiments on Event2Mind and ATOMIC datasets. Results show that both the integration of multiple knowledge sources and the use of the meta-learning algorithm improve the performance. | 2004.03070 | https://arxiv.org/abs/2004.03070v2 | https://arxiv.org/pdf/2004.03070v2.pdf | [
"Meta-Learning",
"Multi-Task Learning",
"Text Generation"
] | [] | [] |
DM7Ii-wTzn | https://paperswithcode.com/paper/revisiting-the-importance-of-individual-units | Revisiting the Importance of Individual Units in CNNs via Ablation | We revisit the importance of the individual units in Convolutional Neural
Networks (CNNs) for visual recognition. By conducting unit ablation experiments
on CNNs trained on large scale image datasets, we demonstrate that, though
ablating any individual unit does not hurt overall classification accuracy, it
does lead to significant damage on the accuracy of specific classes. This
result shows that an individual unit is specialized to encode information
relevant to a subset of classes. We compute the correlation between the
accuracy drop under unit ablation and various attributes of an individual unit
such as class selectivity and weight L1 norm. We confirm that unit attributes
such as class selectivity are a poor predictor for impact on overall accuracy
as found previously in recent work \cite{morcos2018importance}. However, our
results show that class selectivity along with other attributes are good
predictors of the importance of one unit to individual classes. We evaluate the
impact of random rotation, batch normalization, and dropout to the importance
of units to specific classes. Our results show that units with high selectivity
play an important role in network classification power at the individual class
level. Understanding and interpreting the behavior of these units is necessary
and meaningful. | 1806.02891 | http://arxiv.org/abs/1806.02891v1 | http://arxiv.org/pdf/1806.02891v1.pdf | [] | [
"Dropout"
] | [] |
MNNTGqbHbY | https://paperswithcode.com/paper/backward-forward-algorithm-an-improvement | Backward-Forward Algorithm: An Improvement towards Extreme Learning Machine | The extreme learning machine needs a large number of hidden nodes to generalize a single hidden layer neural network for a given training data-set. The need for more number of hidden nodes suggests that the neural-network is memorizing rather than generalizing the model. Hence, a supervised learning method is described here that uses Moore-Penrose approximation to determine both input-weight and output-weight in two epochs, namely, backward-pass and forward-pass. The proposed technique has an advantage over the back-propagation method in terms of iterations required and is superior to the extreme learning machine in terms of the number of hidden units necessary for generalization. | 1907.10282 | https://arxiv.org/abs/1907.10282v4 | https://arxiv.org/pdf/1907.10282v4.pdf | [] | [] | [] |
qXAIAygCR6 | https://paperswithcode.com/paper/fast-flexible-function-dispatch-in-julia | Fast Flexible Function Dispatch in Julia | Technical computing is a challenging application area for programming
languages to address. This is evinced by the unusually large number of
specialized languages in the area (e.g. MATLAB, R), and the complexity of
common software stacks, often involving multiple languages and custom code
generators. We believe this is ultimately due to key characteristics of the
domain: highly complex operators, a need for extensive code specialization for
performance, and a desire for permissive high-level programming styles allowing
productive experimentation. The Julia language attempts to provide a more
effective structure for this kind of programming by allowing programmers to
express complex polymorphic behaviors using dynamic multiple dispatch over
parametric types. The forms of extension and reuse permitted by this paradigm
have proven valuable for technical computing. We report on how this approach
has allowed domain experts to express useful abstractions while simultaneously
providing a natural path to better performance for high-level technical code. | 1808.03370 | http://arxiv.org/abs/1808.03370v1 | http://arxiv.org/pdf/1808.03370v1.pdf | [] | [] | [] |
Ix3zoSopIP | https://paperswithcode.com/paper/agile-earth-observation-satellite-scheduling | Agile Earth observation satellite scheduling over 20 years: formulations, methods and future directions | Agile satellites with advanced attitude maneuvering capability are the new generation of Earth observation satellites (EOSs). The continuous improvement in satellite technology and decrease in launch cost have boosted the development of agile EOSs (AEOSs). To efficiently employ the increasing orbiting AEOSs, the AEOS scheduling problem (AEOSSP) aiming to maximize the entire observation profit while satisfying all complex operational constraints, has received much attention over the past 20 years. The objectives of this paper are thus to summarize current research on AEOSSP, identify main accomplishments and highlight potential future research directions. To this end, general definitions of AEOSSP with operational constraints are described initially, followed by its three typical variations including different definitions of observation profit, multi-objective function and autonomous model. A detailed literature review from 1997 up to 2019 is then presented in line with four different solution methods, i.e., exact method, heuristic, metaheuristic and machine learning. Finally, we discuss a number of topics worth pursuing in the future. | 2003.06169 | https://arxiv.org/abs/2003.06169v1 | https://arxiv.org/pdf/2003.06169v1.pdf | [] | [
"LINE"
] | [] |
bdLCDv0AZO | https://paperswithcode.com/paper/convolutional-neural-networks-based-automated | Convolutional Neural Networks based automated segmentation and labelling of the lumbar spine X-ray | The aim of this study is to investigate the segmentation accuracies of different segmentation networks trained on 730 manually annotated lateral lumbar spine X-rays. Instance segmentation networks were compared to semantic segmentation networks. The study cohort comprised diseased spines and postoperative images with metallic implants. The average mean accuracy and mean intersection over union (IoU) was up to 3 percent better for the best performing instance segmentation model, the average pixel accuracy and weighted IoU were slightly better for the best performing semantic segmentation model. Moreover, the inferences of the instance segmentation models are easier to implement for further processing pipelines in clinical decision support. | 2004.03364 | https://arxiv.org/abs/2004.03364v1 | https://arxiv.org/pdf/2004.03364v1.pdf | [
"Instance Segmentation",
"Semantic Segmentation"
] | [] | [] |
QaCo9OLesH | https://paperswithcode.com/paper/analysis-of-the-gradient-descent-algorithm | Analysis of the Gradient Descent Algorithm for a Deep Neural Network Model with Skip-connections | The behavior of the gradient descent (GD) algorithm is analyzed for a deep
neural network model with skip-connections. It is proved that in the
over-parametrized regime, for a suitable initialization, with high probability
GD can find a global minimum exponentially fast. Generalization error estimates
along the GD path are also established. As a consequence, it is shown that when
the target function is in the reproducing kernel Hilbert space (RKHS) with a
kernel defined by the initialization, there exist generalizable early-stopping
solutions along the GD path. In addition, it is also shown that the GD path is
uniformly close to the functions given by the related random feature model.
Consequently, in this "implicit regularization" setting, the deep neural
network model deteriorates to a random feature model. Our results hold for
neural networks of any width larger than the input dimension. | 1904.05263 | http://arxiv.org/abs/1904.05263v3 | http://arxiv.org/pdf/1904.05263v3.pdf | [] | [] | [] |
MlHjhfQY_O | https://paperswithcode.com/paper/lsmi-sinkhorn-semi-supervised-squared-loss | LSMI-Sinkhorn: Semi-supervised Squared-Loss Mutual Information Estimation with Optimal Transport | Estimating mutual information is an important machine learning and statistics problem. To estimate the mutual information from data, a common practice is preparing a set of paired samples. However, in some cases, it is difficult to obtain a large number of data pairs. To address this problem, we propose squared-loss mutual information (SMI) estimation using a small number of paired samples and the available unpaired ones. We first represent SMI through the density ratio function, where the expectation is approximated by the samples from marginals and its assignment parameters. The objective is formulated using the optimal transport problem and quadratic programming. Then, we introduce the least-square mutual information-Sinkhorn algorithm (LSMI-Sinkhorn) for efficient optimization. Through experiments, we first demonstrate that the proposed method can estimate the SMI without a large number of paired samples. We also evaluate and show the effectiveness of the proposed LSMI-Sinkhorn on various types of machine learning problems such as image matching and photo album summarization. | 1909.02373 | https://arxiv.org/abs/1909.02373v2 | https://arxiv.org/pdf/1909.02373v2.pdf | [
"Mutual Information Estimation"
] | [] | [] |
YST6gmphYu | https://paperswithcode.com/paper/evaluating-indirect-strategies-for-chinese | Evaluating Indirect Strategies for Chinese-Spanish Statistical Machine Translation | Although, Chinese and Spanish are two of the most spoken languages in the
world, not much research has been done in machine translation for this language
pair. This paper focuses on investigating the state-of-the-art of
Chinese-to-Spanish statistical machine translation (SMT), which nowadays is one
of the most popular approaches to machine translation. For this purpose, we
report details of the available parallel corpus which are Basic Traveller
Expressions Corpus (BTEC), Holy Bible and United Nations (UN). Additionally, we
conduct experimental work with the largest of these three corpora to explore
alternative SMT strategies by means of using a pivot language. Three
alternatives are considered for pivoting: cascading, pseudo-corpus and
triangulation. As pivot language, we use either English, Arabic or French.
Results show that, for a phrase-based SMT system, English is the best pivot
language between Chinese and Spanish. We propose a system output combination
using the pivot strategies which is capable of outperforming the direct
translation strategy. The main objective of this work is motivating and
involving the research community to work in this important pair of languages
given their demographic impact. | 1402.0563 | http://arxiv.org/abs/1402.0563v1 | http://arxiv.org/pdf/1402.0563v1.pdf | [
"Machine Translation"
] | [] | [] |
Wg5Cb5dicj | https://paperswithcode.com/paper/improving-retrieval-modeling-using-cross | Improving Retrieval Modeling Using Cross Convolution Networks And Multi Frequency Word Embedding | To build a satisfying chatbot that has the ability of managing a
goal-oriented multi-turn dialogue, accurate modeling of human conversation is
crucial. In this paper we concentrate on the task of response selection for
multi-turn human-computer conversation with a given context. Previous
approaches show weakness in capturing information of rare keywords that appear
in either or both context and correct response, and struggle with long input
sequences. We propose Cross Convolution Network (CCN) and Multi Frequency word
embedding to address both problems. We train several models using the Ubuntu
Dialogue dataset which is the largest freely available multi-turn based
dialogue corpus. We further build an ensemble model by averaging predictions of
multiple models. We achieve a new state-of-the-art on this dataset with
considerable improvements compared to previous best results. | 1802.05373 | http://arxiv.org/abs/1802.05373v2 | http://arxiv.org/pdf/1802.05373v2.pdf | [
"Chatbot"
] | [
"Convolution"
] | [] |
WF6Ixfs_ee | https://paperswithcode.com/paper/learning-graph-while-training-an-evolving | Learning Graph While Training: An Evolving Graph Convolutional Neural Network | Convolution Neural Networks on Graphs are important generalization and
extension of classical CNNs. While previous works generally assumed that the
graph structures of samples are regular with unified dimensions, in many
applications, they are highly diverse or even not well defined. Under some
circumstances, e.g. chemical molecular data, clustering or coarsening for
simplifying the graphs is hard to be justified chemically. In this paper, we
propose a more general and flexible graph convolution network (EGCN) fed by
batch of arbitrarily shaped data together with their evolving graph Laplacians
trained in supervised fashion. Extensive experiments have been conducted to
demonstrate the superior performance in terms of both the acceleration of
parameter fitting and the significantly improved prediction accuracy on
multiple graph-structured datasets. | 1708.04675 | http://arxiv.org/abs/1708.04675v1 | http://arxiv.org/pdf/1708.04675v1.pdf | [] | [
"Convolution"
] | [] |
Ct6YqeXi9C | https://paperswithcode.com/paper/voar-a-visual-and-integrated-ontology | VOAR: A Visual and Integrated Ontology Alignment Environment | Ontology alignment is a key process for enabling interoperability between ontology-based systems in the Linked Open Data age. From two input ontologies, this process generates an alignment (set of correspondences) between them. In this paper we present VOAR, a new web-based environment for ontology alignment visualization and manipulation. Within this graphical environment, users can manually create/edit correspondences and apply a set of operations on alignments (filtering, merge, difference, etc.). VOAR allows invoking external ontology matching systems that implement a specific alignment interface, so that the generated alignments can be manipulated within the environment. Evaluating multiple alignments together against a reference one can also be carried out, using classical evaluation metrics (precision, recall and f-measure). The status of each correspondence with respect to its presence or absence in reference alignment is visually represented. Overall, the main new aspect of VOAR is the visualization and manipulation of alignments at schema level, in an integrated, visual and web-based environment. | null | https://www.aclweb.org/anthology/L14-1658/ | http://www.lrec-conf.org/proceedings/lrec2014/pdf/851_Paper.pdf | [] | [] | [] |
2pElrHzmtR | https://paperswithcode.com/paper/exploiting-capacity-of-sewer-system-using | Exploiting Capacity of Sewer System Using Unsupervised Learning Algorithms Combined with Dimensionality Reduction | Exploiting capacity of sewer system using decentralized control is a cost
effective mean of minimizing the overflow. Given the size of the real sewer
system, exploiting all the installed control structures in the sewer pipes can
be challenging. This paper presents a divide and conquer solution to implement
decentralized control measures based on unsupervised learning algorithms. A
sewer system is first divided into a number of subcatchments. A series of
natural and built factors that have the impact on sewer system performance is
then collected. Clustering algorithms are then applied to grouping
subcatchments with similar hydraulic hydrologic characteristics. Following
which, principal component analysis is performed to interpret the main features
of sub-catchment groups and identify priority control locations. Overflows
under different control scenarios are compared based on the hydraulic model.
Simulation results indicate that priority control applied to the most suitable
cluster could bring the most profitable result. | 1811.03883 | http://arxiv.org/abs/1811.03883v1 | http://arxiv.org/pdf/1811.03883v1.pdf | [
"Dimensionality Reduction"
] | [] | [] |
wEvY9xOVez | https://paperswithcode.com/paper/enhancing-triplegan-for-semi-supervised | Enhancing TripleGAN for Semi-Supervised Conditional Instance Synthesis and Classification | Learning class-conditional data distributions is crucial for Generative Adversarial Networks (GAN) in semi-supervised learning. To improve both instance synthesis and classification in this setting, we propose an enhanced TripleGAN (EnhancedTGAN) model in this work. We follow the adversarial training scheme of the original TripleGAN, but completely re-design the training targets of the generator and classifier. Specifically, we adopt feature-semantics matching to enhance the generator in learning class-conditional distributions from both the aspects of statistics in the latent space and semantics consistency with respect to the generator and classifier. Since a limited amount of labeled data is not sufficient to determine satisfactory decision boundaries, we include two classifiers, and incorporate collaborative learning into our model to provide better guidance for generator training. The synthesized high-fidelity data can in turn be used for improving classifier training. In the experiments, the superior performance of our approach on multiple benchmark datasets demonstrates the effectiveness of the mutual reinforcement between the generator and classifiers in facilitating semi-supervised instance synthesis and classification.
| null | http://openaccess.thecvf.com/content_CVPR_2019/html/Wu_Enhancing_TripleGAN_for_Semi-Supervised_Conditional_Instance_Synthesis_and_Classification_CVPR_2019_paper.html | http://openaccess.thecvf.com/content_CVPR_2019/papers/Wu_Enhancing_TripleGAN_for_Semi-Supervised_Conditional_Instance_Synthesis_and_Classification_CVPR_2019_paper.pdf | [] | [] | [] |
m37C30v-zj | https://paperswithcode.com/paper/adaptive-iterative-hessian-sketch-via-a | Adaptive Iterative Hessian Sketch via A-Optimal Subsampling | Iterative Hessian sketch (IHS) is an effective sketching method for modeling large-scale data. It was originally proposed by Pilanci and Wainwright (2016; JMLR) based on randomized sketching matrices. However, it is computationally intensive due to the iterative sketch process. In this paper, we analyze the IHS algorithm under the unconstrained least squares problem setting, then propose a deterministic approach for improving IHS via A-optimal subsampling. Our contributions are three-fold: (1) a good initial estimator based on the A-optimal design is suggested; (2) a novel ridged preconditioner is developed for repeated sketching; and (3) an exact line search method is proposed for determining the optimal step length adaptively. Extensive experimental results demonstrate that our proposed A-optimal IHS algorithm outperforms the existing accelerated IHS methods. | 1902.07627 | https://arxiv.org/abs/1902.07627v2 | https://arxiv.org/pdf/1902.07627v2.pdf | [] | [
"LINE"
] | [] |
x52si-3xm1 | https://paperswithcode.com/paper/an-anderson-chebyshev-mixing-method-for | A Fast Anderson-Chebyshev Acceleration for Nonlinear Optimization | Anderson acceleration (or Anderson mixing) is an efficient acceleration method for fixed point iterations $x_{t+1}=G(x_t)$, e.g., gradient descent can be viewed as iteratively applying the operation $G(x) \triangleq x-\alpha\nabla f(x)$. It is known that Anderson acceleration is quite efficient in practice and can be viewed as an extension of Krylov subspace methods for nonlinear problems. In this paper, we show that Anderson acceleration with Chebyshev polynomial can achieve the optimal convergence rate $O(\sqrt{\kappa}\ln\frac{1}{\epsilon})$, which improves the previous result $O(\kappa\ln\frac{1}{\epsilon})$ provided by (Toth and Kelley, 2015) for quadratic functions. Moreover, we provide a convergence analysis for minimizing general nonlinear problems. Besides, if the hyperparameters (e.g., the Lipschitz smooth parameter $L$) are not available, we propose a guessing algorithm for guessing them dynamically and also prove a similar convergence rate. Finally, the experimental results demonstrate that the proposed Anderson-Chebyshev acceleration method converges significantly faster than other algorithms, e.g., vanilla gradient descent (GD), Nesterov's Accelerated GD. Also, these algorithms combined with the proposed guessing algorithm (guessing the hyperparameters dynamically) achieve much better performance. | 1809.02341 | https://arxiv.org/abs/1809.02341v4 | https://arxiv.org/pdf/1809.02341v4.pdf | [] | [] | [] |
V9oY_RMEF3 | https://paperswithcode.com/paper/minimizing-finite-sums-with-the-stochastic | Minimizing Finite Sums with the Stochastic Average Gradient | We propose the stochastic average gradient (SAG) method for optimizing the
sum of a finite number of smooth convex functions. Like stochastic gradient
(SG) methods, the SAG method's iteration cost is independent of the number of
terms in the sum. However, by incorporating a memory of previous gradient
values the SAG method achieves a faster convergence rate than black-box SG
methods. The convergence rate is improved from O(1/k^{1/2}) to O(1/k) in
general, and when the sum is strongly-convex the convergence rate is improved
from the sub-linear O(1/k) to a linear convergence rate of the form O(p^k) for
p \textless{} 1. Further, in many cases the convergence rate of the new method
is also faster than black-box deterministic gradient methods, in terms of the
number of gradient evaluations. Numerical experiments indicate that the new
algorithm often dramatically outperforms existing SG and deterministic gradient
methods, and that the performance may be further improved through the use of
non-uniform sampling strategies. | 1309.2388 | http://arxiv.org/abs/1309.2388v2 | http://arxiv.org/pdf/1309.2388v2.pdf | [] | [] | [] |
trahVogMVF | https://paperswithcode.com/paper/the-en-ru-two-way-integrated-machine | The En-Ru Two-way Integrated Machine Translation System Based on Transformer | Machine translation is one of the most popular areas in natural language processing. WMT is a conference to assess the level of machine translation capabilities of organizations around the world, which is the evaluation activity we participated in. In this review we participated in a two-way translation track from Russian to English and English to Russian. We used official training data, 38 million parallel corpora, and 10 million monolingual corpora. The overall framework we use is the Transformer neural machine translation model, supplemented by data filtering, post-processing, reordering and other related processing methods. The BLEU value of our final translation result from Russian to English is 38.7, ranking 5th, while from English to Russian is 27.8, ranking 10th. | null | https://www.aclweb.org/anthology/W19-5349/ | https://www.aclweb.org/anthology/W19-5349 | [
"Machine Translation"
] | [
"Residual Connection",
"BPE",
"Dense Connections",
"Label Smoothing",
"ReLU",
"Adam",
"Softmax",
"Dropout",
"Multi-Head Attention",
"Layer Normalization",
"Scaled Dot-Product Attention",
"Transformer"
] | [] |
LTqOJcEMd4 | https://paperswithcode.com/paper/proceedings-of-the-thirteenth-conference-on | Proceedings of the Thirteenth Conference on Uncertainty in Artificial Intelligence (1997) | This is the Proceedings of the Thirteenth Conference on Uncertainty in
Artificial Intelligence, which was held in Providence, RI, August 1-3, 1997 | 1304.3846 | http://arxiv.org/abs/1304.3846v1 | http://arxiv.org/pdf/1304.3846v1.pdf | [] | [] | [] |
AqgmFAEfg2 | https://paperswithcode.com/paper/efficient-hinging-hyperplanes-neural-network | Efficient hinging hyperplanes neural network and its application in nonlinear system identification | In this paper, the efficient hinging hyperplanes (EHH) neural network is proposed based on the model of hinging hyperplanes (HH). The EHH neural network is a distributed representation, the training of which involves solving several convex optimization problems and is fast. It is proved that for every EHH neural network, there is an equivalent adaptive hinging hyperplanes (AHH) tree, which was also proposed based on the model of HH and find good applications in system identification. The construction of the EHH neural network includes 2 stages. First the initial structure of the EHH neural network is randomly determined and the Lasso regression is used to choose the appropriate network. To alleviate the impact of randomness, secondly, the stacking strategy is employed to formulate a more general network structure. Different from other neural networks, the EHH neural network has interpretability ability, which can be easily obtained through its ANOVA decomposition (or interaction matrix). The interpretability can then be used as a suggestion for input variable selection. The EHH neural network is applied in nonlinear system identification, the simulation results show that the regression vector selected is reasonable and the identification speed is fast, while at the same time, the simulation accuracy is satisfactory. | 1905.06518 | https://arxiv.org/abs/1905.06518v2 | https://arxiv.org/pdf/1905.06518v2.pdf | [] | [] | [] |
ncs2c4rknQ | https://paperswithcode.com/paper/automated-alertness-and-emotion-detection-for | Automated Alertness and Emotion Detection for Empathic Feedback During E-Learning | In the context of education technology, empathic interaction with the user
and feedback by the learning system using multiple inputs such as video, voice
and text inputs is an important area of research. In this paper, a
nonintrusive, standalone model for intelligent assessment of alertness and
emotional state as well as generation of appropriate feedback has been
proposed. Using the non-intrusive visual cues, the system classifies emotion
and alertness state of the user, and provides appropriate feedback according to
the detected cognitive state using facial expressions, ocular parameters,
postures, and gestures. Assessment of alertness level using ocular parameters
such as PERCLOS and saccadic parameters, emotional state from facial expression
analysis, and detection of both relevant cognitive and emotional states from
upper body gestures and postures has been proposed. Integration of such a
system in e-learning environment is expected to enhance students performance
through interaction, feedback, and positive mood induction. | 1604.00312 | http://arxiv.org/abs/1604.00312v1 | http://arxiv.org/pdf/1604.00312v1.pdf | [] | [] | [] |
iPntTsnQYR | https://paperswithcode.com/paper/generalized-sampling-with-functional | High-resolution signal recovery via generalized sampling and functional principal component analysis | In this paper, we introduce a computational framework for recovering a high-resolution approximation of an unknown function from its low-resolution indirect measurements as well as high-resolution training observations by merging the frameworks of generalized sampling and functional principal component analysis. In particular, we increase the signal resolution via a data driven approach, which models the function of interest as a realization of a random field and leverages a training set of observations generated via the same underlying random process. We study the performance of the resulting estimation procedure and show that high-resolution recovery is indeed possible provided appropriate low-rank and angle conditions hold and provided the training set is sufficiently large relative to the desired resolution. Moreover, we show that the size of the training set can be reduced by leveraging sparse representations of the functional principal components. Furthermore, the effectiveness of the proposed reconstruction procedure is illustrated by various numerical examples. | 2002.08724 | https://arxiv.org/abs/2002.08724v2 | https://arxiv.org/pdf/2002.08724v2.pdf | [] | [] | [] |
eO0b0e_fVz | https://paperswithcode.com/paper/end-to-end-cascaded-u-nets-with-a | End-to-End Cascaded U-Nets with a Localization Network for Kidney Tumor Segmentation | Kidney tumor segmentation emerges as a new frontier of computer vision in medical imaging. This is partly due to its challenging manual annotation and great medical impact. Within the scope of the Kidney Tumor Segmentation Challenge 2019, that is aiming at combined kidney and tumor segmentation, this work proposes a novel combination of 3D U-Nets---collectively denoted TuNet---utilizing the resulting kidney masks for the consecutive tumor segmentation. The proposed method achieves a S{\o}rensen-Dice coefficient score of 0.902 for the kidney, and 0.408 for the tumor segmentation, computed from a five-fold cross-validation on the 210 patients available in the data. | 1910.07521 | https://arxiv.org/abs/1910.07521v1 | https://arxiv.org/pdf/1910.07521v1.pdf | [
"Tumor Segmentation"
] | [] | [] |
BARYVYqXea | https://paperswithcode.com/paper/working-memory-facilitates-reward-modulated | Working memory facilitates reward-modulated Hebbian learning in recurrent neural networks | Reservoir computing is a powerful tool to explain how the brain learns temporal sequences, such as movements, but existing learning schemes are either biologically implausible or too inefficient to explain animal performance. We show that a network can learn complicated sequences with a reward-modulated Hebbian learning rule if the network of reservoir neurons is combined with a second network that serves as a dynamic working memory and provides a spatio-temporal backbone signal to the reservoir. In combination with the working memory, reward-modulated Hebbian learning of the readout neurons performs as well as FORCE learning, but with the advantage of a biologically plausible interpretation of both the learning rule and the learning paradigm. | 1910.10559 | https://arxiv.org/abs/1910.10559v1 | https://arxiv.org/pdf/1910.10559v1.pdf | [] | [] | [] |
o4n8b7mkVj | https://paperswithcode.com/paper/derivative-free-methods-for-policy | Derivative-Free Methods for Policy Optimization: Guarantees for Linear Quadratic Systems | We study derivative-free methods for policy optimization over the class of linear policies. We focus on characterizing the convergence rate of these methods when applied to linear-quadratic systems, and study various settings of driving noise and reward feedback. We show that these methods provably converge to within any pre-specified tolerance of the optimal policy with a number of zero-order evaluations that is an explicit polynomial of the error tolerance, dimension, and curvature properties of the problem. Our analysis reveals some interesting differences between the settings of additive driving noise and random initialization, as well as the settings of one-point and two-point reward feedback. Our theory is corroborated by extensive simulations of derivative-free methods on these systems. Along the way, we derive convergence rates for stochastic zero-order optimization algorithms when applied to a certain class of non-convex problems. | 1812.08305 | https://arxiv.org/abs/1812.08305v3 | https://arxiv.org/pdf/1812.08305v3.pdf | [] | [] | [] |
BUTQGbTAtX | https://paperswithcode.com/paper/unsupervised-feature-learning-for-low-level | Unsupervised Feature Learning for low-level Local Image Descriptors | Unsupervised feature learning has shown impressive results for a wide range
of input modalities, in particular for object classification tasks in computer
vision. Using a large amount of unlabeled data, unsupervised feature learning
methods are utilized to construct high-level representations that are
discriminative enough for subsequently trained supervised classification
algorithms. However, it has never been \emph{quantitatively} investigated yet
how well unsupervised learning methods can find \emph{low-level
representations} for image patches without any additional supervision. In this
paper we examine the performance of pure unsupervised methods on a low-level
correspondence task, a problem that is central to many Computer Vision
applications. We find that a special type of Restricted Boltzmann Machines
(RBMs) performs comparably to hand-crafted descriptors. Additionally, a simple
binarization scheme produces compact representations that perform better than
several state-of-the-art descriptors. | 1301.2840 | http://arxiv.org/abs/1301.2840v4 | http://arxiv.org/pdf/1301.2840v4.pdf | [
"Binarization",
"Object Classification"
] | [] | [] |
J6W5Urddlg | https://paperswithcode.com/paper/generalization-studies-of-neural-network | Generalization Studies of Neural Network Models for Cardiac Disease Detection Using Limited Channel ECG | Acceleration of machine learning research in healthcare is challenged by lack
of large annotated and balanced datasets. Furthermore, dealing with measurement
inaccuracies and exploiting unsupervised data are considered to be central to
improving existing solutions. In particular, a primary objective in predictive
modeling is to generalize well to both unseen variations within the observed
classes, and unseen classes. In this work, we consider such a challenging
problem in machine learning driven diagnosis: detecting a gamut of
cardiovascular conditions (e.g. infarction, dysrhythmia etc.) from limited
channel ECG measurements. Though deep neural networks have achieved
unprecedented success in predictive modeling, they rely solely on
discriminative models that can generalize poorly to unseen classes. We argue
that unsupervised learning can be utilized to construct effective latent spaces
that facilitate better generalization. This work extensively compares the
generalization of our proposed approach against a state-of-the-art deep
learning solution. Our results show significant improvements in F1-scores. | 1901.03295 | http://arxiv.org/abs/1901.03295v1 | http://arxiv.org/pdf/1901.03295v1.pdf | [] | [] | [] |
LxNcAnvl3d | https://paperswithcode.com/paper/object-scene-convolutional-neural-networks | Object-Scene Convolutional Neural Networks for Event Recognition in Images | Event recognition from still images is of great importance for image
understanding. However, compared with event recognition in videos, there are
much fewer research works on event recognition in images. This paper addresses
the issue of event recognition from images and proposes an effective method
with deep neural networks. Specifically, we design a new architecture, called
Object-Scene Convolutional Neural Network (OS-CNN). This architecture is
decomposed into object net and scene net, which extract useful information for
event understanding from the perspective of objects and scene context,
respectively. Meanwhile, we investigate different network architectures for
OS-CNN design, and adapt the deep (AlexNet) and very-deep (GoogLeNet) networks
to the task of event recognition. Furthermore, we find that the deep and
very-deep networks are complementary to each other. Finally, based on the
proposed OS-CNN and comparative study of different network architectures, we
come up with a solution of five-stream CNN for the track of cultural event
recognition at the ChaLearn Looking at People (LAP) challenge 2015. Our method
obtains the performance of 85.5% and ranks the $1^{st}$ place in this
challenge. | 1505.00296 | http://arxiv.org/abs/1505.00296v1 | http://arxiv.org/pdf/1505.00296v1.pdf | [] | [] | [] |
JfZTd1QpcX | https://paperswithcode.com/paper/riemannian-kernel-based-nystrom-method-for | Riemannian kernel based Nyström method for approximate infinite-dimensional covariance descriptors with application to image set classification | In the domain of pattern recognition, using the CovDs (Covariance Descriptors) to represent data and taking the metrics of the resulting Riemannian manifold into account have been widely adopted for the task of image set classification. Recently, it has been proven that infinite-dimensional CovDs are more discriminative than their low-dimensional counterparts. However, the form of infinite-dimensional CovDs is implicit and the computational load is high. We propose a novel framework for representing image sets by approximating infinite-dimensional CovDs in the paradigm of the Nystr\"om method based on a Riemannian kernel. We start by modeling the images via CovDs, which lie on the Riemannian manifold spanned by SPD (Symmetric Positive Definite) matrices. We then extend the Nystr\"om method to the SPD manifold and obtain the approximations of CovDs in RKHS (Reproducing Kernel Hilbert Space). Finally, we approximate infinite-dimensional CovDs via these approximations. Empirically, we apply our framework to the task of image set classification. The experimental results obtained on three benchmark datasets show that our proposed approximate infinite-dimensional CovDs outperform the original CovDs. | 1806.06177 | https://arxiv.org/abs/1806.06177v2 | https://arxiv.org/pdf/1806.06177v2.pdf | [] | [] | [] |
xghNSncVue | https://paperswithcode.com/paper/blind-modulation-classification-based-on-mlp | Blind Modulation Classification based on MLP and PNN | In this work, a pattern recognition system is investigated for blind
automatic classification of digitally modulated communication signals. The
proposed technique is able to discriminate the type of modulation scheme which
is eventually used for demodulation followed by information extraction. The
proposed system is composed of two subsystems namely feature extraction
sub-system (FESS) and classifier sub-system (CSS). The FESS consists of
continuous wavelet transform (CWT) for feature generation and principal
component analysis (PCA) for selection of the feature subset which is rich in
discriminatory information. The CSS uses the selected features to accurately
classify the modulation class of the received signal. The proposed technique
uses probabilistic neural network (PNN) and multilayer perceptron forward
neural network (MLPFN) for comparative study of their recognition ability. PNN
have been found to perform better in terms of classification accuracy as well
as testing and training time than MLPFN. The proposed approach is robust to
presence of phase offset and additive Gaussian noise. | 1605.09441 | http://arxiv.org/abs/1605.09441v1 | http://arxiv.org/pdf/1605.09441v1.pdf | [] | [] | [] |
0yNbQIivDK | https://paperswithcode.com/paper/an-ensemble-deep-learning-based-approach-for | An Ensemble Deep Learning Based Approach for Red Lesion Detection in Fundus Images | Diabetic retinopathy is one of the leading causes of preventable blindness in
the world. Its earliest sign are red lesions, a general term that groups both
microaneurysms and hemorrhages. In daily clinical practice, these lesions are
manually detected by physicians using fundus photographs. However, this task is
tedious and time consuming, and requires an intensive effort due to the small
size of the lesions and their lack of contrast. Computer-assisted diagnosis of
DR based on red lesion detection is being actively explored due to its
improvement effects both in clinicians consistency and accuracy. Several
methods for detecting red lesions have been proposed in the literature, most of
them based on characterizing lesion candidates using hand crafted features, and
classifying them into true or false positive detections. Deep learning based
approaches, by contrast, are scarce in this domain due to the high expense of
annotating the lesions manually. In this paper we propose a novel method for
red lesion detection based on combining both deep learned and domain knowledge.
Features learned by a CNN are augmented by incorporating hand crafted features.
Such ensemble vector of descriptors is used afterwards to identify true lesion
candidates using a Random Forest classifier. We empirically observed that
combining both sources of information significantly improve results with
respect to using each approach separately. Furthermore, our method reported the
highest performance on a per-lesion basis on DIARETDB1 and e-ophtha, and for
screening and need for referral on MESSIDOR compared to a second human expert.
Results highlight the fact that integrating manually engineered approaches with
deep learned features is relevant to improve results when the networks are
trained from lesion-level annotated data. An open source implementation of our
system is publicly available online. | 1706.03008 | http://arxiv.org/abs/1706.03008v2 | http://arxiv.org/pdf/1706.03008v2.pdf | [] | [] | [] |
N3GplP7mRi | https://paperswithcode.com/paper/a-gpu-based-wfst-decoder-with-exact-lattice | A GPU-based WFST Decoder with Exact Lattice Generation | We describe initial work on an extension of the Kaldi toolkit that supports
weighted finite-state transducer (WFST) decoding on Graphics Processing Units
(GPUs). We implement token recombination as an atomic GPU operation in order to
fully parallelize the Viterbi beam search, and propose a dynamic load balancing
strategy for more efficient token passing scheduling among GPU threads. We also
redesign the exact lattice generation and lattice pruning algorithms for better
utilization of the GPUs. Experiments on the Switchboard corpus show that the
proposed method achieves identical 1-best results and lattice quality in
recognition and confidence measure tasks, while running 3 to 15 times faster
than the single process Kaldi decoder. The above results are reported on
different GPU architectures. Additionally we obtain a 46-fold speedup with
sequence parallelism and multi-process service (MPS) in GPU. | 1804.03243 | http://arxiv.org/abs/1804.03243v3 | http://arxiv.org/pdf/1804.03243v3.pdf | [] | [] | [] |
oDyttlr884 | https://paperswithcode.com/paper/deep-elastic-networks-with-model-selection | Deep Elastic Networks with Model Selection for Multi-Task Learning | In this work, we consider the problem of instance-wise dynamic network model selection for multi-task learning. To this end, we propose an efficient approach to exploit a compact but accurate model in a backbone architecture for each instance of all tasks. The proposed method consists of an estimator and a selector. The estimator is based on a backbone architecture and structured hierarchically. It can produce multiple different network models of different configurations in a hierarchical structure. The selector chooses a model dynamically from a pool of candidate models given an input instance. The selector is a relatively small-size network consisting of a few layers, which estimates a probability distribution over the candidate models when an input instance of a task is given. Both estimator and selector are jointly trained in a unified learning framework in conjunction with a sampling-based learning strategy, without additional computation steps. We demonstrate the proposed approach for several image classification tasks compared to existing approaches performing model selection or learning multiple tasks. Experimental results show that our approach gives not only outstanding performance compared to other competitors but also the versatility to perform instance-wise model selection for multiple tasks. | 1909.04860 | https://arxiv.org/abs/1909.04860v1 | https://arxiv.org/pdf/1909.04860v1.pdf | [
"Image Classification",
"Model Selection",
"Multi-Task Learning"
] | [] | [] |
Kwju7tzYbK | https://paperswithcode.com/paper/bio-lstm-a-biomechanically-inspired-recurrent | Bio-LSTM: A Biomechanically Inspired Recurrent Neural Network for 3D Pedestrian Pose and Gait Prediction | In applications such as autonomous driving, it is important to understand, infer, and anticipate the intention and future behavior of pedestrians. This ability allows vehicles to avoid collisions and improve ride safety and quality. This paper proposes a biomechanically inspired recurrent neural network (Bio-LSTM) that can predict the location and 3D articulated body pose of pedestrians in a global coordinate frame, given 3D poses and locations estimated in prior frames with inaccuracy. The proposed network is able to predict poses and global locations for multiple pedestrians simultaneously, for pedestrians up to 45 meters from the cameras (urban intersection scale). The outputs of the proposed network are full-body 3D meshes represented in Skinned Multi-Person Linear (SMPL) model parameters. The proposed approach relies on a novel objective function that incorporates the periodicity of human walking (gait), the mirror symmetry of the human body, and the change of ground reaction forces in a human gait cycle. This paper presents prediction results on the PedX dataset, a large-scale, in-the-wild data set collected at real urban intersections with heavy pedestrian traffic. Results show that the proposed network can successfully learn the characteristics of pedestrian gait and produce accurate and consistent 3D pose predictions. | 1809.03705 | https://arxiv.org/abs/1809.03705v3 | https://arxiv.org/pdf/1809.03705v3.pdf | [
"Autonomous Driving"
] | [] | [] |
hlpiN-lQwD | https://paperswithcode.com/paper/mad-max-affine-spline-insights-into-deep | Mad Max: Affine Spline Insights into Deep Learning | We build a rigorous bridge between deep networks (DNs) and approximation
theory via spline functions and operators. Our key result is that a large class
of DNs can be written as a composition of max-affine spline operators (MASOs),
which provide a powerful portal through which to view and analyze their inner
workings. For instance, conditioned on the input signal, the output of a MASO
DN can be written as a simple affine transformation of the input. This implies
that a DN constructs a set of signal-dependent, class-specific templates
against which the signal is compared via a simple inner product; we explore the
links to the classical theory of optimal classification via matched filters and
the effects of data memorization. Going further, we propose a simple penalty
term that can be added to the cost function of any DN learning algorithm to
force the templates to be orthogonal with each other; this leads to
significantly improved classification performance and reduced overfitting with
no change to the DN architecture. The spline partition of the input signal
space that is implicitly induced by a MASO directly links DNs to the theory of
vector quantization (VQ) and $K$-means clustering, which opens up new geometric
avenue to study how DNs organize signals in a hierarchical fashion. To validate
the utility of the VQ interpretation, we develop and validate a new distance
metric for signals and images that quantifies the difference between their VQ
encodings. (This paper is a significantly expanded version of A Spline Theory
of Deep Learning from ICML 2018.) | 1805.06576 | http://arxiv.org/abs/1805.06576v5 | http://arxiv.org/pdf/1805.06576v5.pdf | [
"Quantization"
] | [] | [] |
viiHPRTjAd | https://paperswithcode.com/paper/a-fast-and-lightweight-system-for | A Fast and Lightweight System for Multilingual Dependency Parsing | We present a multilingual dependency parser with a bidirectional-LSTM (BiLSTM) feature extractor and a multi-layer perceptron (MLP) classifier. We trained our transition-based projective parser in UD version 2.0 datasets without any additional data. The parser is fast, lightweight and effective on big treebanks. In the CoNLL 2017 Shared Task: Multilingual Parsing from Raw Text to Universal Dependencies, the official results show that the macro-averaged LAS F1 score of our system Mengest is 61.33{\%}. | null | https://www.aclweb.org/anthology/K17-3025/ | https://www.aclweb.org/anthology/K17-3025 | [
"Dependency Parsing"
] | [] | [] |
GiElLvoTxg | https://paperswithcode.com/paper/learning-with-random-learning-rates | Learning with Random Learning Rates | Hyperparameter tuning is a bothersome step in the training of deep learning
models. One of the most sensitive hyperparameters is the learning rate of the
gradient descent. We present the 'All Learning Rates At Once' (Alrao)
optimization method for neural networks: each unit or feature in the network
gets its own learning rate sampled from a random distribution spanning several
orders of magnitude. This comes at practically no computational cost. Perhaps
surprisingly, stochastic gradient descent (SGD) with Alrao performs close to
SGD with an optimally tuned learning rate, for various architectures and
problems. Alrao could save time when testing deep learning models: a range of
models could be quickly assessed with Alrao, and the most promising models
could then be trained more extensively. This text comes with a PyTorch
implementation of the method, which can be plugged on an existing PyTorch
model: https://github.com/leonardblier/alrao . | 1810.01322 | http://arxiv.org/abs/1810.01322v3 | http://arxiv.org/pdf/1810.01322v3.pdf | [] | [] | [] |
PpXwdpl82r | https://paperswithcode.com/paper/identification-of-internal-faults-in-indirect | Identification of Internal Faults in Indirect Symmetrical Phase Shift Transformers Using Ensemble Learning | This paper proposes methods to identify 40 different types of internal faults
in an Indirect Symmetrical Phase Shift Transformer (ISPST). The ISPST was
modeled using Power System Computer Aided Design (PSCAD)/ Electromagnetic
Transients including DC (EMTDC). The internal faults were simulated by varying
the transformer tapping, backward and forward phase shifts, loading, and
percentage of winding faulted. Data for 960 cases of each type of fault was
recorded. A series of features were extracted for a, b, and c phases from time,
frequency, time-frequency, and information theory domains. The importance of
the extracted features was evaluated through univariate tests which helped to
reduce the number of features. The selected features were then used for
training five state-of-the-art machine learning classifiers. Extremely Random
Trees and Random Forest, the ensemble-based learners, achieved the accuracy of
98.76% and 97.54% respectively outperforming Multilayer Perceptron (96.13%),
Logistic Regression (93.54%), and Support Vector Machines (92.60%) | 1811.04537 | http://arxiv.org/abs/1811.04537v1 | http://arxiv.org/pdf/1811.04537v1.pdf | [] | [
"Residual Connection",
"BPE",
"Dense Connections",
"Label Smoothing",
"ReLU",
"Adam",
"Softmax",
"Dropout",
"Multi-Head Attention",
"Layer Normalization",
"Scaled Dot-Product Attention",
"Transformer"
] | [] |
tvlO3vucRf | https://paperswithcode.com/paper/constraint-selection-in-metric-learning | Constraint Selection in Metric Learning | A number of machine learning algorithms are using a metric, or a distance, in
order to compare individuals. The Euclidean distance is usually employed, but
it may be more efficient to learn a parametric distance such as Mahalanobis
metric. Learning such a metric is a hot topic since more than ten years now,
and a number of methods have been proposed to efficiently learn it. However,
the nature of the problem makes it quite difficult for large scale data, as
well as data for which classes overlap. This paper presents a simple way of
improving accuracy and scalability of any iterative metric learning algorithm,
where constraints are obtained prior to the algorithm. The proposed approach
relies on a loss-dependent weighted selection of constraints that are used for
learning the metric. Using the corresponding dedicated loss function, the
method clearly allows to obtain better results than state-of-the-art methods,
both in terms of accuracy and time complexity. Some experimental results on
real world, and potentially large, datasets are demonstrating the effectiveness
of our proposition. | 1612.04853 | http://arxiv.org/abs/1612.04853v1 | http://arxiv.org/pdf/1612.04853v1.pdf | [
"Metric Learning"
] | [] | [] |
YBeNSdmMXG | https://paperswithcode.com/paper/chinese-text-in-the-wild | Chinese Text in the Wild | [python3.6] 运用tf实现自然场景文字检测,keras/pytorch实现ctpn+crnn+ctc实现不定长场景文字OCR识别 | 1803.00085 | http://arxiv.org/abs/1803.00085v1 | http://arxiv.org/pdf/1803.00085v1.pdf | [
"Optical Character Recognition"
] | [] | [] |
Me_3GYP3CA | https://paperswithcode.com/paper/sparse-lifting-of-dense-vectors-unifying-word | Sparse Lifting of Dense Vectors: Unifying Word and Sentence Representations | As the first step in automated natural language processing, representing words and sentences is of central importance and has attracted significant research attention. Different approaches, from the early one-hot and bag-of-words representation to more recent distributional dense and sparse representations, were proposed. Despite the successful results that have been achieved, such vectors tend to consist of uninterpretable components and face nontrivial challenge in both memory and computational requirement in practical applications. In this paper, we designed a novel representation model that projects dense word vectors into a higher dimensional space and favors a highly sparse and binary representation of word vectors with potentially interpretable components, while trying to maintain pairwise inner products between original vectors as much as possible. Computationally, our model is relaxed as a symmetric non-negative matrix factorization problem which admits a fast yet effective solution. In a series of empirical evaluations, the proposed model exhibited consistent improvement and high potential in practical applications. | 1911.01625 | https://arxiv.org/abs/1911.01625v1 | https://arxiv.org/pdf/1911.01625v1.pdf | [] | [] | [] |
pAIhLzv0Gr | https://paperswithcode.com/paper/ontology-based-design-of-experiments-on-big | Ontology-based Design of Experiments on Big Data Solutions | Big data solutions are designed to cope with data of huge Volume and wide
Variety, that need to be ingested at high Velocity and have potential Veracity
issues, challenging characteristics that are usually referred to as the "4Vs of
Big Data". In order to evaluate possibly complex big data solutions, stress
tests require to assess a large number of combinations of sub-components
jointly with the possible big data variations. A formalization of the Design of
Experiments (DoE) on big data solutions is aimed at ensuring the
reproducibility of the experiments, facilitating their partitioning in
sub-experiments and guaranteeing the consistency of their outcomes in a global
assessment. In this paper, an ontology-based approach is proposed to support
the evaluation of a big data system in two ways. Firstly, the approach
formalizes a decomposition and recombination of the big data solution, allowing
for the aggregation of component evaluation results at inter-component level.
Secondly, existing work on DoE is translated into an ontology for supporting
the selection of experiments. The proposed ontology-based approach offers the
possibility to combine knowledge from the evaluation domain and the application
domain. It exploits domain and inter-domain specific restrictions on the factor
combinations in order to reduce the number of experiments. Contrary to existing
approaches, the proposed use of ontologies is not limited to the assertional
description and exploitation of past experiments but offers richer
terminological descriptions for the development of a DoE from scratch. As an
application example, a maritime big data solution to the problem of detecting
and predicting vessel suspicious behaviour through mobility analysis is
selected. The article is concluded with a sketch of future works. | 1904.08626 | http://arxiv.org/abs/1904.08626v1 | http://arxiv.org/pdf/1904.08626v1.pdf | [] | [] | [] |
eiNdqbOobk | https://paperswithcode.com/paper/understanding-art-through-multi-modal | Understanding Art through Multi-Modal Retrieval in Paintings | In computer vision, visual arts are often studied from a purely aesthetics
perspective, mostly by analysing the visual appearance of an artistic
reproduction to infer its style, its author, or its representative features. In
this work, however, we explore art from both a visual and a language
perspective. Our aim is to bridge the gap between the visual appearance of an
artwork and its underlying meaning, by jointly analysing its aesthetics and its
semantics. We introduce the use of multi-modal techniques in the field of
automatic art analysis by 1) collecting a multi-modal dataset with fine-art
paintings and comments, and 2) exploring robust visual and textual
representations in artistic images. | 1904.10615 | http://arxiv.org/abs/1904.10615v1 | http://arxiv.org/pdf/1904.10615v1.pdf | [
"Art Analysis"
] | [] | [] |
7soC9FDr67 | https://paperswithcode.com/paper/effectiveness-of-hierarchical-softmax-in | Effectiveness of Hierarchical Softmax in Large Scale Classification Tasks | Typically, Softmax is used in the final layer of a neural network to get a
probability distribution for output classes. But the main problem with Softmax
is that it is computationally expensive for large scale data sets with large
number of possible outputs. To approximate class probability efficiently on
such large scale data sets we can use Hierarchical Softmax. LSHTC datasets were
used to study the performance of the Hierarchical Softmax. LSHTC datasets have
large number of categories. In this paper we evaluate and report the
performance of normal Softmax Vs Hierarchical Softmax on LSHTC datasets. This
evaluation used macro f1 score as a performance measure. The observation was
that the performance of Hierarchical Softmax degrades as the number of classes
increase. | 1812.05737 | http://arxiv.org/abs/1812.05737v1 | http://arxiv.org/pdf/1812.05737v1.pdf | [] | [
"Hierarchical Softmax",
"Softmax"
] | [] |
Xd26jeNTH9 | https://paperswithcode.com/paper/an-empirical-evaluation-of-true-online-td | An Empirical Evaluation of True Online TD(λ) | The true online TD({\lambda}) algorithm has recently been proposed (van
Seijen and Sutton, 2014) as a universal replacement for the popular
TD({\lambda}) algorithm, in temporal-difference learning and reinforcement
learning. True online TD({\lambda}) has better theoretical properties than
conventional TD({\lambda}), and the expectation is that it also results in
faster learning. In this paper, we put this hypothesis to the test.
Specifically, we compare the performance of true online TD({\lambda}) with that
of TD({\lambda}) on challenging examples, random Markov reward processes, and a
real-world myoelectric prosthetic arm. We use linear function approximation
with tabular, binary, and non-binary features. We assess the algorithms along
three dimensions: computational cost, learning speed, and ease of use. Our
results confirm the strength of true online TD({\lambda}): 1) for sparse
feature vectors, the computational overhead with respect to TD({\lambda}) is
minimal; for non-sparse features the computation time is at most twice that of
TD({\lambda}), 2) across all domains/representations the learning speed of true
online TD({\lambda}) is often better, but never worse than that of
TD({\lambda}), and 3) true online TD({\lambda}) is easier to use, because it
does not require choosing between trace types, and it is generally more stable
with respect to the step-size. Overall, our results suggest that true online
TD({\lambda}) should be the first choice when looking for an efficient,
general-purpose TD method. | 1507.00353 | http://arxiv.org/abs/1507.00353v1 | http://arxiv.org/pdf/1507.00353v1.pdf | [] | [] | [] |
dA4E4VKW11 | https://paperswithcode.com/paper/dynamics-of-pedestrian-crossing-decisions | Dynamics of Pedestrian Crossing Decisions Based on Vehicle Trajectories in Large-Scale Simulated and Real-World Data | Humans, as both pedestrians and drivers, generally skillfully navigate
traffic intersections. Despite the uncertainty, danger, and the non-verbal
nature of communication commonly found in these interactions, there are
surprisingly few collisions considering the total number of interactions. As
the role of automation technology in vehicles grows, it becomes increasingly
critical to understand the relationship between pedestrian and driver behavior:
how pedestrians perceive the actions of a vehicle/driver and how pedestrians
make crossing decisions. The relationship between time-to-arrival (TTA) and
pedestrian gap acceptance (i.e., whether a pedestrian chooses to cross under a
given window of time to cross) has been extensively investigated. However, the
dynamic nature of vehicle trajectories in the context of non-verbal
communication has not been systematically explored. Our work provides evidence
that trajectory dynamics, such as changes in TTA, can be powerful signals in
the non-verbal communication between drivers and pedestrians. Moreover, we
investigate these effects in both simulated and real-world datasets, both
larger than have previously been considered in literature to the best of our
knowledge. | 1904.04202 | http://arxiv.org/abs/1904.04202v1 | http://arxiv.org/pdf/1904.04202v1.pdf | [] | [] | [] |
H8lwhDHLJU | https://paperswithcode.com/paper/fusion-of-range-and-thermal-images-for-person | Fusion of Range and Thermal Images for Person Detection | Detecting people in images is a challenging problem. Differences in pose,
clothing and lighting, along with other factors, cause a lot of variation in
their appearance. To overcome these issues, we propose a system based on fused
range and thermal infrared images. These measurements show considerably less
variation and provide more meaningful information. We provide a brief
introduction to the sensor technology used and propose a calibration method.
Several data fusion algorithms are compared and their performance is assessed
on a simulated data set. The results of initial experiments on real data are
analyzed and the measurement errors and the challenges they present are
discussed. The resulting fused data are used to efficiently detect people in a
fixed camera set-up. The system is extended to include person tracking. | 1612.02183 | http://arxiv.org/abs/1612.02183v1 | http://arxiv.org/pdf/1612.02183v1.pdf | [
"Human Detection"
] | [] | [] |
W-VdC-eASI | https://paperswithcode.com/paper/online-risk-bounded-motion-planning-for | Online Risk-Bounded Motion Planning for Autonomous Vehicles in Dynamic Environments | A crucial challenge to efficient and robust motion planning for autonomous
vehicles is understanding the intentions of the surrounding agents. Ignoring
the intentions of the other agents in dynamic environments can lead to risky or
over-conservative plans. In this work, we model the motion planning problem as
a partially observable Markov decision process (POMDP) and propose an online
system that combines an intent recognition algorithm and a POMDP solver to
generate risk-bounded plans for the ego vehicle navigating with a number of
dynamic agent vehicles. The intent recognition algorithm predicts the
probabilistic hybrid motion states of each agent vehicle over a finite horizon
using Bayesian filtering and a library of pre-learned maneuver motion models.
We update the POMDP model with the intent recognition results in real time and
solve it using a heuristic search algorithm which produces policies with
upper-bound guarantees on the probability of near colliding with other dynamic
agents. We demonstrate that our system is able to generate better motion plans
in terms of efficiency and safety in a number of challenging environments
including unprotected intersection left turns and lane changes as compared to
the baseline methods. | 1904.02341 | http://arxiv.org/abs/1904.02341v1 | http://arxiv.org/pdf/1904.02341v1.pdf | [
"Autonomous Vehicles",
"Motion Planning"
] | [] | [] |
xvUD8AlbCd | https://paperswithcode.com/paper/on-the-complexity-of-finding-second-best | On the Complexity of Finding Second-Best Abductive Explanations | While looking for abductive explanations of a given set of manifestations, an
ordering between possible solutions is often assumed. The complexity of
finding/verifying optimal solutions is already known. In this paper we consider
the computational complexity of finding second-best solutions. We consider
different orderings, and consider also different possible definitions of what a
second-best solution is. | 1204.5859 | http://arxiv.org/abs/1204.5859v3 | http://arxiv.org/pdf/1204.5859v3.pdf | [] | [] | [] |
ju-kv6jnJ1 | https://paperswithcode.com/paper/comparative-analysis-of-verbal-alignment-in | Comparative analysis of verbal alignment in human-human and human-agent interactions | Engagement is an important feature in human-human and human-agent interaction. In this paper, we investigate lexical alignment as a cue of engagement, relying on two different corpora : CID and SEMAINE. Our final goal is to build a virtual conversational character that could use alignment strategies to maintain user{'}s engagement. To do so, we investigate two alignment processes : shared vocabulary and other-repetitions. A quantitative and qualitative approach is proposed to characterize these aspects in human-human (CID) and human-operator (SEMAINE) interactions. Our results show that these processes are observable in both corpora, indicating a stable pattern that can be further modelled in conversational agents. | null | https://www.aclweb.org/anthology/L14-1289/ | http://www.lrec-conf.org/proceedings/lrec2014/pdf/327_Paper.pdf | [] | [] | [] |
P-bwRAcPms | https://paperswithcode.com/paper/development-of-a-real-time-colorectal-tumor | Development of a Real-time Colorectal Tumor Classification System for Narrow-band Imaging zoom-videoendoscopy | Colorectal endoscopy is important for the early detection and treatment of
colorectal cancer and is used worldwide. A computer-aided diagnosis (CAD)
system that provides an objective measure to endoscopists during colorectal
endoscopic examinations would be of great value. In this study, we describe a
newly developed CAD system that provides real-time objective measures. Our
system captures the video stream from an endoscopic system and transfers it to
a desktop computer. The captured video stream is then classified by a
pretrained classifier and the results are displayed on a monitor. The
experimental results show that our developed system works efficiently in actual
endoscopic examinations and is medically significant. | 1612.05000 | http://arxiv.org/abs/1612.05000v2 | http://arxiv.org/pdf/1612.05000v2.pdf | [] | [] | [] |
SFF_lndkcI | https://paperswithcode.com/paper/optimality-and-sub-optimality-of-pca-for | Optimality and Sub-optimality of PCA for Spiked Random Matrices and Synchronization | A central problem of random matrix theory is to understand the eigenvalues of
spiked random matrix models, in which a prominent eigenvector is planted into a
random matrix. These distributions form natural statistical models for
principal component analysis (PCA) problems throughout the sciences. Baik, Ben
Arous and P\'ech\'e showed that the spiked Wishart ensemble exhibits a sharp
phase transition asymptotically: when the signal strength is above a critical
threshold, it is possible to detect the presence of a spike based on the top
eigenvalue, and below the threshold the top eigenvalue provides no information.
Such results form the basis of our understanding of when PCA can detect a
low-rank signal in the presence of noise.
However, not all the information about the spike is necessarily contained in
the spectrum. We study the fundamental limitations of statistical methods,
including non-spectral ones. Our results include:
I) For the Gaussian Wigner ensemble, we show that PCA achieves the optimal
detection threshold for a variety of benign priors for the spike. We extend
previous work on the spherically symmetric and i.i.d. Rademacher priors through
an elementary, unified analysis.
II) For any non-Gaussian Wigner ensemble, we show that PCA is always
suboptimal for detection. However, a variant of PCA achieves the optimal
threshold (for benign priors) by pre-transforming the matrix entries according
to a carefully designed function. This approach has been stated before, and we
give a rigorous and general analysis.
III) For both the Gaussian Wishart ensemble and various synchronization
problems over groups, we show that inefficient procedures can work below the
threshold where PCA succeeds, whereas no known efficient algorithm achieves
this. This conjectural gap between what is statistically possible and what can
be done efficiently remains open. | 1609.05573 | http://arxiv.org/abs/1609.05573v2 | http://arxiv.org/pdf/1609.05573v2.pdf | [] | [
"PCA"
] | [] |
z7a_3Z2iWG | https://paperswithcode.com/paper/extreme-value-statistics-for-censored-data | Extreme value statistics for censored data with heavy tails under competing risks | This paper addresses the problem of estimating, in the presence of random
censoring as well as competing risks, the extreme value index of the
(sub)-distribution function associated to one particular cause, in the
heavy-tail case. Asymptotic normality of the proposed estimator (which has the
form of an Aalen-Johansen integral, and is the first estimator proposed in this
context) is established. A small simulation study exhibits its performances for
finite samples. Estimation of extreme quantiles of the cumulative incidence
function is also addressed. | 1701.05458 | http://arxiv.org/abs/1701.05458v1 | http://arxiv.org/pdf/1701.05458v1.pdf | [] | [] | [] |
znuzyqRkOK | https://paperswithcode.com/paper/hybrid-data-clustering-approach-using-k-means | Hybrid data clustering approach using K-Means and Flower Pollination Algorithm | Data clustering is a technique for clustering set of objects into known
number of groups. Several approaches are widely applied to data clustering so
that objects within the clusters are similar and objects in different clusters
are far away from each other. K-Means, is one of the familiar center based
clustering algorithms since implementation is very easy and fast convergence.
However, K-Means algorithm suffers from initialization, hence trapped in local
optima. Flower Pollination Algorithm (FPA) is the global optimization
technique, which avoids trapping in local optimum solution. In this paper, a
novel hybrid data clustering approach using Flower Pollination Algorithm and
K-Means (FPAKM) is proposed. The proposed algorithm results are compared with
K-Means and FPA on eight datasets. From the experimental results, FPAKM is
better than FPA and K-Means. | 1505.03236 | http://arxiv.org/abs/1505.03236v1 | http://arxiv.org/pdf/1505.03236v1.pdf | [] | [] | [] |
rfDaFrHpRg | https://paperswithcode.com/paper/low-shot-learning-with-untrained-neural | Low Shot Learning with Untrained Neural Networks for Imaging Inverse Problems | Employing deep neural networks as natural image priors to solve inverse problems either requires large amounts of data to sufficiently train expressive generative models or can succeed with no data via untrained neural networks. However, very few works have considered how to interpolate between these no- to high-data regimes. In particular, how can one use the availability of a small amount of data (even $5-25$ examples) to one's advantage in solving these inverse problems and can a system's performance increase as the amount of data increases as well? In this work, we consider solving linear inverse problems when given a small number of examples of images that are drawn from the same distribution as the image of interest. Comparing to untrained neural networks that use no data, we show how one can pre-train a neural network with a few given examples to improve reconstruction results in compressed sensing and semantic image recovery problems such as colorization. Our approach leads to improved reconstruction as the amount of available data increases and is on par with fully trained generative models, while requiring less than $1 \%$ of the data needed to train a generative model. | 1910.10797 | https://arxiv.org/abs/1910.10797v1 | https://arxiv.org/pdf/1910.10797v1.pdf | [
"Colorization"
] | [] | [] |
hA2brluDrV | https://paperswithcode.com/paper/pure-strategy-or-mixed-strategy | Pure Strategy or Mixed Strategy? | Mixed strategy EAs aim to integrate several mutation operators into a single
algorithm. However few theoretical analysis has been made to answer the
question whether and when the performance of mixed strategy EAs is better than
that of pure strategy EAs. In theory, the performance of EAs can be measured by
asymptotic convergence rate and asymptotic hitting time. In this paper, it is
proven that given a mixed strategy (1+1) EAs consisting of several mutation
operators, its performance (asymptotic convergence rate and asymptotic hitting
time)is not worse than that of the worst pure strategy (1+1) EA using one
mutation operator; if these mutation operators are mutually complementary, then
it is possible to design a mixed strategy (1+1) EA whose performance is better
than that of any pure strategy (1+1) EA using one mutation operator. | 1112.1517 | http://arxiv.org/abs/1112.1517v4 | http://arxiv.org/pdf/1112.1517v4.pdf | [] | [] | [] |
Q4jcUdyHoQ | https://paperswithcode.com/paper/evolutionary-multitasking-for-multiobjective | Evolutionary Multitasking for Multiobjective Continuous Optimization: Benchmark Problems, Performance Metrics and Baseline Results | In this report, we suggest nine test problems for multi-task multi-objective
optimization (MTMOO), each of which consists of two multiobjective optimization
tasks that need to be solved simultaneously. The relationship between tasks
varies between different test problems, which would be helpful to have a
comprehensive evaluation of the MO-MFO algorithms. It is expected that the
proposed test problems will germinate progress the field of the MTMOO research. | 1706.02766 | http://arxiv.org/abs/1706.02766v1 | http://arxiv.org/pdf/1706.02766v1.pdf | [
"Multiobjective Optimization"
] | [] | [] |
0MPWQR2kb8 | https://paperswithcode.com/paper/improved-coresets-for-kernel-density | Improved Coresets for Kernel Density Estimates | We study the construction of coresets for kernel density estimates. That is
we show how to approximate the kernel density estimate described by a large
point set with another kernel density estimate with a much smaller point set.
For characteristic kernels (including Gaussian and Laplace kernels), our
approximation preserves the $L_\infty$ error between kernel density estimates
within error $\epsilon$, with coreset size $2/\epsilon^2$, but no other aspects
of the data, including the dimension, the diameter of the point set, or the
bandwidth of the kernel common to other approximations. When the dimension is
unrestricted, we show this bound is tight for these kernels as well as a much
broader set.
This work provides a careful analysis of the iterative Frank-Wolfe algorithm
adapted to this context, an algorithm called \emph{kernel herding}. This
analysis unites a broad line of work that spans statistics, machine learning,
and geometry.
When the dimension $d$ is constant, we demonstrate much tighter bounds on the
size of the coreset specifically for Gaussian kernels, showing that it is
bounded by the size of the coreset for axis-aligned rectangles. Currently the
best known constructive bound is $O(\frac{1}{\epsilon} \log^d
\frac{1}{\epsilon})$, and non-constructively, this can be improved by
$\sqrt{\log \frac{1}{\epsilon}}$. This improves the best constant dimension
bounds polynomially for $d \geq 3$. | 1710.04325 | http://arxiv.org/abs/1710.04325v1 | http://arxiv.org/pdf/1710.04325v1.pdf | [] | [
"LINE"
] | [] |
6XCFeR8eaS | https://paperswithcode.com/paper/training-dnn-iot-applications-for-deployment | Training DNN IoT Applications for Deployment On Analog NVM Crossbars | A trend towards energy-efficiency, security and privacy has led to a recent focus on deploying DNNs on microcontrollers. However, limits on compute and memory resources restrict the size and the complexity of the ML models deployable in these systems. Computation-In-Memory architectures based on resistive nonvolatile memory (NVM) technologies hold great promise of satisfying the compute and memory demands of high-performance and low-power, inherent in modern DNNs. Nevertheless, these technologies are still immature and suffer from both the intrinsic analog-domain noise problems and the inability of representing negative weights in the NVM structures, incurring in larger crossbar sizes with concomitant impact on ADCs and DACs. In this paper, we provide a training framework for addressing these challenges and quantitatively evaluate the circuit-level efficiency gains thus accrued. We make two contributions: Firstly, we propose a training algorithm that eliminates the need for tuning individual layers of a DNN ensuring uniformity across layer weights and activations. This ensures analog-blocks that can be reused and peripheral hardware substantially reduced. Secondly, using NAS methods, we propose the use of unipolar-weighted (either all-positive or all-negative weights) matrices/sub-matrices. Weight unipolarity obviates the need for doubling crossbar area leading to simplified analog periphery. We validate our methodology with CIFAR10 and HAR applications by mapping to crossbars using 4-bit and 2-bit devices. We achieve up to 92:91% accuracy (95% floating-point) using 2-bit only-positive weights for HAR. A combination of the proposed techniques leads to 80% area improvement and up to 45% energy reduction. | 1910.13850 | https://arxiv.org/abs/1910.13850v3 | https://arxiv.org/pdf/1910.13850v3.pdf | [
"Quantization"
] | [] | [] |
znnqG7RavX | https://paperswithcode.com/paper/adversarial-ladder-networks | Adversarial Ladder Networks | The use of unsupervised data in addition to supervised data in training
discriminative neural networks has improved the performance of this clas-
sification scheme. However, the best results were achieved with a training
process that is divided in two parts: first an unsupervised pre-training step
is done for initializing the weights of the network and after these weights are
refined with the use of supervised data. On the other hand adversarial noise
has improved the results of clas- sical supervised learning. Recently, a new
neural network topology called Ladder Network, where the key idea is based in
some properties of hierar- chichal latent variable models, has been proposed as
a technique to train a neural network using supervised and unsupervised data at
the same time with what is called semi-supervised learning. This technique has
reached state of the art classification. In this work we add adversarial noise
to the ladder network and get state of the art classification, with several
important conclusions on how adversarial noise can help in addition with new
possible lines of investi- gation. We also propose an alternative to add
adversarial noise to unsu- pervised data. | 1611.02320 | http://arxiv.org/abs/1611.02320v3 | http://arxiv.org/pdf/1611.02320v3.pdf | [
"Latent Variable Models",
"Unsupervised Pre-training"
] | [] | [] |
1sFwUEglDK | https://paperswithcode.com/paper/analysis-of-a-design-pattern-for-teaching | Analysis of a Design Pattern for Teaching with Features and Labels | We study the task of teaching a machine to classify objects using features
and labels. We introduce the Error-Driven-Featuring design pattern for teaching
using features and labels in which a teacher prefers to introduce features only
if they are needed. We analyze the potential risks and benefits of this
teaching pattern through the use of teaching protocols, illustrative examples,
and by providing bounds on the effort required for an optimal machine teacher
using a linear learning algorithm, the most commonly used type of learners in
interactive machine learning systems. Our analysis provides a deeper
understanding of potential trade-offs of using different learning algorithms
and between the effort required for featuring (creating new features) and
labeling (providing labels for objects). | 1611.05950 | http://arxiv.org/abs/1611.05950v1 | http://arxiv.org/pdf/1611.05950v1.pdf | [] | [] | [] |
JgnYLNW15c | https://paperswithcode.com/paper/pznet-efficient-3d-convnet-inference-on | PZnet: Efficient 3D ConvNet Inference on Manycore CPUs | Convolutional nets have been shown to achieve state-of-the-art accuracy in
many biomedical image analysis tasks. Many tasks within biomedical analysis
domain involve analyzing volumetric (3D) data acquired by CT, MRI and
Microscopy acquisition methods. To deploy convolutional nets in practical
working systems, it is important to solve the efficient inference problem.
Namely, one should be able to apply an already-trained convolutional network to
many large images using limited computational resources. In this paper we
present PZnet, a CPU-only engine that can be used to perform inference for a
variety of 3D convolutional net architectures. PZNet outperforms MKL-based CPU
implementations of PyTorch and Tensorflow by more than 3.5x for the popular
U-net architecture. Moreover, for 3D convolutions with low featuremap numbers,
cloud CPU inference with PZnet outperfroms cloud GPU inference in terms of cost
efficiency. | 1903.07525 | http://arxiv.org/abs/1903.07525v1 | http://arxiv.org/pdf/1903.07525v1.pdf | [] | [] | [] |
PQI8sTzcu3 | https://paperswithcode.com/paper/an-experiment-on-measurement-of-pavement | An Experiment on Measurement of Pavement Roughness via Android-Based Smartphones | The study focuses on the experiment of using three different smartphones to collect acceleration data from vibration for the road roughness detection. The Android operating system is used in the application. The study takes place on asphaltic pavement of the expressway system of Thailand, with 9 km distance. The run vehicle has an inertial profiler with laser line sensors to record road roughness according to the IRI. The RMS and Machine Learning (ML) methods are used in this study. There is different ability of each smartphone to detect the vibration, thus different detected values are obtained. The results are compared to the IRI observation. ML method gives better result compared to RMS. This study finds little relationship between IRI and acceleration data from vibration collected from smartphones. | 1907.13120 | https://arxiv.org/abs/1907.13120v1 | https://arxiv.org/pdf/1907.13120v1.pdf | [] | [
"LINE"
] | [] |
5QXXBmRhIR | https://paperswithcode.com/paper/comment-on-adv-bnn-improved-adversarial | Comment on "Adv-BNN: Improved Adversarial Defense through Robust Bayesian Neural Network" | A recent paper by Liu et al. combines the topics of adversarial training and Bayesian Neural Networks (BNN) and suggests that adversarially trained BNNs are more robust against adversarial attacks than their non-Bayesian counterparts. Here, I analyze the proposed defense and suggest that one needs to adjust the adversarial attack to incorporate the stochastic nature of a Bayesian network to perform an accurate evaluation of its robustness. Using this new type of attack I show that there appears to be no strong evidence for higher robustness of the adversarially trained BNNs. | 1907.00895 | https://arxiv.org/abs/1907.00895v1 | https://arxiv.org/pdf/1907.00895v1.pdf | [
"Adversarial Attack",
"Adversarial Defense"
] | [] | [] |
J4IFXqfIQk | https://paperswithcode.com/paper/incremental-adaptation-of-nmt-for | Incremental Adaptation of NMT for Professional Post-editors: A User Study | A common use of machine translation in the industry is providing initial translation hypotheses, which are later supervised and post-edited by a human expert. During this revision process, new bilingual data are continuously generated. Machine translation systems can benefit from these new data, incrementally updating the underlying models under an online learning paradigm. We conducted a user study on this scenario, for a neural machine translation system. The experimentation was carried out by professional translators, with a vast experience in machine translation post-editing. The results showed a reduction in the required amount of human effort needed when post-editing the outputs of the system, improvements in the translation quality and a positive perception of the adaptive system by the users. | 1906.08996 | https://arxiv.org/abs/1906.08996v1 | https://arxiv.org/pdf/1906.08996v1.pdf | [
"Machine Translation"
] | [] | [] |
_YT2dg5UN5 | https://paperswithcode.com/paper/learning-to-define-terms-in-the-software | Learning to Define Terms in the Software Domain | One way to test a person{'}s knowledge of a domain is to ask them to define domain-specific terms. Here, we investigate the task of automatically generating definitions of technical terms by reading text from the technical domain. Specifically, we learn definitions of software entities from a large corpus built from the user forum Stack Overflow. To model definitions, we train a language model and incorporate additional domain-specific information like word co-occurrence, and ontological category information. Our approach improves previous baselines by 2 BLEU points for the definition generation task. Our experiments also show the additional challenges associated with the task and the short-comings of language-model based architectures for definition generation. | null | https://www.aclweb.org/anthology/W18-6122/ | https://www.aclweb.org/anthology/W18-6122 | [
"Knowledge Base Population",
"Language Modelling",
"Relationship Extraction (Distant Supervised)"
] | [] | [] |
BKjOYFehNj | https://paperswithcode.com/paper/inclusion-within-continuous-belief-functions | Inclusion within Continuous Belief Functions | Defining and modeling the relation of inclusion between continuous belief
function may be considered as an important operation in order to study their
behaviors. Within this paper we will propose and present two forms of
inclusion: The strict and the partial one. In order to develop this relation,
we will study the case of consonant belief function. To do so, we will simulate
normal distributions allowing us to model and analyze these relations. Based on
that, we will determine the parameters influencing and characterizing the two
forms of inclusion. | 1501.06705 | http://arxiv.org/abs/1501.06705v1 | http://arxiv.org/pdf/1501.06705v1.pdf | [] | [] | [] |
5cUXnUOPiX | https://paperswithcode.com/paper/distant-supervision-for-relation-extraction-1 | Distant Supervision for Relation Extraction beyond the Sentence Boundary | The growing demand for structured knowledge has led to great interest in
relation extraction, especially in cases with limited supervision. However,
existing distance supervision approaches only extract relations expressed in
single sentences. In general, cross-sentence relation extraction is
under-explored, even in the supervised-learning setting. In this paper, we
propose the first approach for applying distant supervision to cross- sentence
relation extraction. At the core of our approach is a graph representation that
can incorporate both standard dependencies and discourse relations, thus
providing a unifying way to model relations within and across sentences. We
extract features from multiple paths in this graph, increasing accuracy and
robustness when confronted with linguistic variation and analysis error.
Experiments on an important extraction task for precision medicine show that
our approach can learn an accurate cross-sentence extractor, using only a small
existing knowledge base and unlabeled text from biomedical research articles.
Compared to the existing distant supervision paradigm, our approach extracted
twice as many relations at similar precision, thus demonstrating the prevalence
of cross-sentence relations and the promise of our approach. | 1609.04873 | http://arxiv.org/abs/1609.04873v3 | http://arxiv.org/pdf/1609.04873v3.pdf | [
"Relation Extraction"
] | [] | [] |
adyTEEl6F1 | https://paperswithcode.com/paper/the-application-of-differential-privacy-for | The Application of Differential Privacy for Rank Aggregation: Privacy and Accuracy | The potential risk of privacy leakage prevents users from sharing their
honest opinions on social platforms. This paper addresses the problem of
privacy preservation if the query returns the histogram of rankings. The
framework of differential privacy is applied to rank aggregation. The error
probability of the aggregated ranking is analyzed as a result of noise added in
order to achieve differential privacy. Upper bounds on the error rates for any
positional ranking rule are derived under the assumption that profiles are
uniformly distributed. Simulation results are provided to validate the
probabilistic analysis. | 1409.6831 | http://arxiv.org/abs/1409.6831v1 | http://arxiv.org/pdf/1409.6831v1.pdf | [] | [] | [] |
egqfrk8XGu | https://paperswithcode.com/paper/multi-scale-online-learning-and-its | Multi-scale Online Learning and its Applications to Online Auctions | We consider revenue maximization in online auction/pricing problems. A seller
sells an identical item in each period to a new buyer, or a new set of buyers.
For the online posted pricing problem, we show regret bounds that scale with
the best fixed price, rather than the range of the values. We also show regret
bounds that are almost scale free, and match the offline sample complexity,
when comparing to a benchmark that requires a lower bound on the market share.
These results are obtained by generalizing the classical learning from experts
and multi-armed bandit problems to their multi-scale versions. In this version,
the reward of each action is in a different range, and the regret w.r.t. a
given action scales with its own range, rather than the maximum range. | 1705.09700 | http://arxiv.org/abs/1705.09700v2 | http://arxiv.org/pdf/1705.09700v2.pdf | [] | [] | [] |
iX6ULpnhpe | https://paperswithcode.com/paper/unstructured-multi-view-depth-estimation | Unstructured Multi-View Depth Estimation Using Mask-Based Multiplane Representation | This paper presents a novel method, MaskMVS, to solve depth estimation for
unstructured multi-view image-pose pairs. In the plane-sweep procedure, the
depth planes are sampled by histogram matching that ensures covering the depth
range of interest. Unlike other plane-sweep methods, we do not rely on a cost
metric to explicitly build the cost volume, but instead infer a multiplane mask
representation which regularizes the learning. Compared to many previous
approaches, we show that our method is lightweight and generalizes well without
requiring excessive training. We outperform the current state-of-the-art and
show results on the sun3d, scenes11, MVS, and RGBD test data sets. | 1902.02166 | http://arxiv.org/abs/1902.02166v2 | http://arxiv.org/pdf/1902.02166v2.pdf | [
"Depth Estimation"
] | [] | [] |
4CAr41vHwq | https://paperswithcode.com/paper/recovering-block-structured-activations-using | Recovering Block-structured Activations Using Compressive Measurements | We consider the problems of detection and localization of a contiguous block
of weak activation in a large matrix, from a small number of noisy, possibly
adaptive, compressive (linear) measurements. This is closely related to the
problem of compressed sensing, where the task is to estimate a sparse vector
using a small number of linear measurements. Contrary to results in compressed
sensing, where it has been shown that neither adaptivity nor contiguous
structure help much, we show that for reliable localization the magnitude of
the weakest signals is strongly influenced by both structure and the ability to
choose measurements adaptively while for detection neither adaptivity nor
structure reduce the requirement on the magnitude of the signal. We
characterize the precise tradeoffs between the various problem parameters, the
signal strength and the number of measurements required to reliably detect and
localize the block of activation. The sufficient conditions are complemented
with information theoretic lower bounds. | 1209.3431 | http://arxiv.org/abs/1209.3431v2 | http://arxiv.org/pdf/1209.3431v2.pdf | [] | [] | [] |
lUwP6dHKTO | https://paperswithcode.com/paper/max-cost-discrete-function-evaluation-problem | Max-Cost Discrete Function Evaluation Problem under a Budget | We propose novel methods for max-cost Discrete Function Evaluation Problem
(DFEP) under budget constraints. We are motivated by applications such as
clinical diagnosis where a patient is subjected to a sequence of (possibly
expensive) tests before a decision is made. Our goal is to develop strategies
for minimizing max-costs. The problem is known to be NP hard and greedy methods
based on specialized impurity functions have been proposed. We develop a broad
class of \emph{admissible} impurity functions that admit monomials, classes of
polynomials, and hinge-loss functions that allow for flexible impurity design
with provably optimal approximation bounds. This flexibility is important for
datasets when max-cost can be overly sensitive to "outliers." Outliers bias
max-cost to a few examples that require a large number of tests for
classification. We design admissible functions that allow for accuracy-cost
trade-off and result in $O(\log n)$ guarantees of the optimal cost among trees
with corresponding classification accuracy levels. | 1501.02702 | http://arxiv.org/abs/1501.02702v1 | http://arxiv.org/pdf/1501.02702v1.pdf | [] | [] | [] |
3KavIQaVUp | https://paperswithcode.com/paper/integrating-additional-knowledge-into | Integrating Additional Knowledge Into Estimation of Graphical Models | In applications of graphical models, we typically have more information than
just the samples themselves. A prime example is the estimation of brain
connectivity networks based on fMRI data, where in addition to the samples
themselves, the spatial positions of the measurements are readily available.
With particular regard for this application, we are thus interested in ways to
incorporate additional knowledge most effectively into graph estimation. Our
approach to this is to make neighborhood selection receptive to additional
knowledge by strengthening the role of the tuning parameters. We demonstrate
that this concept (i) can improve reproducibility, (ii) is computationally
convenient and efficient, and (iii) carries a lucid Bayesian interpretation. We
specifically show that the approach provides effective estimations of brain
connectivity graphs from fMRI data. However, providing a general scheme for the
inclusion of additional knowledge, our concept is expected to have applications
in a wide range of domains. | 1704.02739 | http://arxiv.org/abs/1704.02739v2 | http://arxiv.org/pdf/1704.02739v2.pdf | [] | [] | [] |
e4S56JA0LE | https://paperswithcode.com/paper/from-averaging-to-acceleration-there-is-only | From Averaging to Acceleration, There is Only a Step-size | We show that accelerated gradient descent, averaged gradient descent and the
heavy-ball method for non-strongly-convex problems may be reformulated as
constant parameter second-order difference equation algorithms, where stability
of the system is equivalent to convergence at rate O(1/n 2), where n is the
number of iterations. We provide a detailed analysis of the eigenvalues of the
corresponding linear dynamical system , showing various oscillatory and
non-oscillatory behaviors, together with a sharp stability result with explicit
constants. We also consider the situation where noisy gradients are available,
where we extend our general convergence result, which suggests an alternative
algorithm (i.e., with different step sizes) that exhibits the good aspects of
both averaging and acceleration. | 1504.01577 | http://arxiv.org/abs/1504.01577v1 | http://arxiv.org/pdf/1504.01577v1.pdf | [] | [] | [] |
rCD_J2HRGk | https://paperswithcode.com/paper/separation-of-target-anatomical-structure-and | Separation of target anatomical structure and occlusions in chest radiographs | Chest radiographs are commonly performed low-cost exams for screening and diagnosis. However, radiographs are 2D representations of 3D structures causing considerable clutter impeding visual inspection and automated image analysis. Here, we propose a Fully Convolutional Network to suppress, for a specific task, undesired visual structure from radiographs while retaining the relevant image information such as lung-parenchyma. The proposed algorithm creates reconstructed radiographs and ground-truth data from high resolution CT-scans. Results show that removing visual variation that is irrelevant for a classification task improves the performance of a classifier when only limited training data are available. This is particularly relevant because a low number of ground-truth cases is common in medical imaging. | 2002.00751 | https://arxiv.org/abs/2002.00751v1 | https://arxiv.org/pdf/2002.00751v1.pdf | [] | [] | [] |
mFxYXDjUkj | https://paperswithcode.com/paper/adapting-neural-single-document-summarization | Adapting Neural Single-Document Summarization Model for Abstractive Multi-Document Summarization: A Pilot Study | Till now, neural abstractive summarization methods have achieved great success for single document summarization (SDS). However, due to the lack of large scale multi-document summaries, such methods can be hardly applied to multi-document summarization (MDS). In this paper, we investigate neural abstractive methods for MDS by adapting a state-of-the-art neural abstractive summarization model for SDS. We propose an approach to extend the neural abstractive model trained on large scale SDS data to the MDS task. Our approach only makes use of a small number of multi-document summaries for fine tuning. Experimental results on two benchmark DUC datasets demonstrate that our approach can outperform a variety of baseline neural models. | null | https://www.aclweb.org/anthology/W18-6545/ | https://www.aclweb.org/anthology/W18-6545 | [
"Abstractive Text Summarization",
"Document Summarization",
"Machine Translation",
"Multi-Document Summarization",
"Text Generation"
] | [] | [] |
YS0jELMsTG | https://paperswithcode.com/paper/task-specific-visual-saliency-prediction-with | Task Specific Visual Saliency Prediction with Memory Augmented Conditional Generative Adversarial Networks | Visual saliency patterns are the result of a variety of factors aside from
the image being parsed, however existing approaches have ignored these. To
address this limitation, we propose a novel saliency estimation model which
leverages the semantic modelling power of conditional generative adversarial
networks together with memory architectures which capture the subject's
behavioural patterns and task dependent factors. We make contributions aiming
to bridge the gap between bottom-up feature learning capabilities in modern
deep learning architectures and traditional top-down hand-crafted features
based methods for task specific saliency modelling. The conditional nature of
the proposed framework enables us to learn contextual semantics and
relationships among different tasks together, instead of learning them
separately for each task. Our studies not only shed light on a novel
application area for generative adversarial networks, but also emphasise the
importance of task specific saliency modelling and demonstrate the plausibility
of fully capturing this context via an augmented memory architecture. | 1803.03354 | http://arxiv.org/abs/1803.03354v1 | http://arxiv.org/pdf/1803.03354v1.pdf | [
"Saliency Prediction"
] | [] | [] |
AUF2sRl87h | https://paperswithcode.com/paper/residual-attention-net-for-superior-cross | Residual Attention Net for Superior Cross-Domain Time Sequence Modeling | We present a novel architecture, residual attention net (RAN), which merges a sequence architecture, universal transformer, and a computer vision architecture, residual net, with a high-way architecture for cross-domain sequence modeling. The architecture aims at addressing the long dependency issue often faced by recurrent-neural-net-based structures. This paper serves as a proof-of-concept for a new architecture, with RAN aiming at providing the model a higher level understanding of sequence patterns. To our best knowledge, we are the first to propose such an architecture. Out of the standard 85 UCR data sets, we have achieved 35 state-of-the-art results with 10 results matching current state-of-the-art results without further model fine-tuning. The results indicate that such architecture is promising in complex, long-sequence modeling and may have vast, cross-domain applications. | 2001.04077 | https://arxiv.org/abs/2001.04077v1 | https://arxiv.org/pdf/2001.04077v1.pdf | [] | [] | [] |
DGXw31FaHQ | https://paperswithcode.com/paper/ambiguity-and-incomplete-information-in | Ambiguity and Incomplete Information in Categorical Models of Language | We investigate notions of ambiguity and partial information in categorical
distributional models of natural language. Probabilistic ambiguity has
previously been studied using Selinger's CPM construction. This construction
works well for models built upon vector spaces, as has been shown in quantum
computational applications. Unfortunately, it doesn't seem to provide a
satisfactory method for introducing mixing in other compact closed categories
such as the category of sets and binary relations. We therefore lack a uniform
strategy for extending a category to model imprecise linguistic information.
In this work we adopt a different approach. We analyze different forms of
ambiguous and incomplete information, both with and without quantitative
probabilistic data. Each scheme then corresponds to a suitable enrichment of
the category in which we model language. We view different monads as
encapsulating the informational behaviour of interest, by analogy with their
use in modelling side effects in computation. Previous results of Jacobs then
allow us to systematically construct suitable bases for enrichment.
We show that we can freely enrich arbitrary dagger compact closed categories
in order to capture all the phenomena of interest, whilst retaining the
important dagger compact closed structure. This allows us to construct a model
with real convex combination of binary relations that makes non-trivial use of
the scalars. Finally we relate our various different enrichments, showing that
finite subconvex algebra enrichment covers all the effects under consideration. | 1701.00660 | http://arxiv.org/abs/1701.00660v1 | http://arxiv.org/pdf/1701.00660v1.pdf | [] | [] | [] |
A60ClWnE3U | https://paperswithcode.com/paper/deep-reinforcement-learning-an-overview-1 | Deep Reinforcement Learning: An Overview | In recent years, a specific machine learning method called deep learning has
gained huge attraction, as it has obtained astonishing results in broad
applications such as pattern recognition, speech recognition, computer vision,
and natural language processing. Recent research has also been shown that deep
learning techniques can be combined with reinforcement learning methods to
learn useful representations for the problems with high dimensional raw data
input. This chapter reviews the recent advances in deep reinforcement learning
with a focus on the most used deep architectures such as autoencoders,
convolutional neural networks and recurrent neural networks which have
successfully been come together with the reinforcement learning framework. | 1806.08894 | http://arxiv.org/abs/1806.08894v1 | http://arxiv.org/pdf/1806.08894v1.pdf | [
"Speech Recognition"
] | [] | [] |
rpKWQbQXBO | https://paperswithcode.com/paper/unsupervised-pidgin-text-generation-by | Unsupervised Pidgin Text Generation By Pivoting English Data and Self-Training | West African Pidgin English is a language that is significantly spoken in West Africa, consisting of at least 75 million speakers. Nevertheless, proper machine translation systems and relevant NLP datasets for pidgin English are virtually absent. In this work, we develop techniques targeted at bridging the gap between Pidgin English and English in the context of natural language generation. %As a proof of concept, we explore the proposed techniques in the area of data-to-text generation. By building upon the previously released monolingual Pidgin English text and parallel English data-to-text corpus, we hope to build a system that can automatically generate Pidgin English descriptions from structured data. We first train a data-to-English text generation system, before employing techniques in unsupervised neural machine translation and self-training to establish the Pidgin-to-English cross-lingual alignment. The human evaluation performed on the generated Pidgin texts shows that, though still far from being practically usable, the pivoting + self-training technique improves both Pidgin text fluency and relevance. | 2003.08272 | https://arxiv.org/abs/2003.08272v1 | https://arxiv.org/pdf/2003.08272v1.pdf | [
"Data-to-Text Generation",
"Machine Translation",
"Text Generation"
] | [] | [] |
-F1igX1lw2 | https://paperswithcode.com/paper/learning-to-hash-tag-videos-with-tag2vec | Learning to Hash-tag Videos with Tag2Vec | User-given tags or labels are valuable resources for semantic understanding
of visual media such as images and videos. Recently, a new type of labeling
mechanism known as hash-tags have become increasingly popular on social media
sites. In this paper, we study the problem of generating relevant and useful
hash-tags for short video clips. Traditional data-driven approaches for tag
enrichment and recommendation use direct visual similarity for label transfer
and propagation. We attempt to learn a direct low-cost mapping from video to
hash-tags using a two step training process. We first employ a natural language
processing (NLP) technique, skip-gram models with neural network training to
learn a low-dimensional vector representation of hash-tags (Tag2Vec) using a
corpus of 10 million hash-tags. We then train an embedding function to map
video features to the low-dimensional Tag2vec space. We learn this embedding
for 29 categories of short video clips with hash-tags. A query video without
any tag-information can then be directly mapped to the vector space of tags
using the learned embedding and relevant tags can be found by performing a
simple nearest-neighbor retrieval in the Tag2Vec space. We validate the
relevance of the tags suggested by our system qualitatively and quantitatively
with a user study. | 1612.04061 | http://arxiv.org/abs/1612.04061v1 | http://arxiv.org/pdf/1612.04061v1.pdf | [] | [] | [] |
BkdoFv3ko8 | https://paperswithcode.com/paper/blenderproc | BlenderProc | BlenderProc is a modular procedural pipeline, which helps in generating real looking images for the training of convolutional neural networks. These can be used in a variety of use cases including segmentation, depth, normal and pose estimation and many others. A key feature of our extension of blender is the simple to use modular pipeline, which was designed to be easily extendable. By offering standard modules, which cover a variety of scenarios, we provide a starting point on which new modules can be created. | 1911.01911 | https://arxiv.org/abs/1911.01911v1 | https://arxiv.org/pdf/1911.01911v1.pdf | [
"3D Object Recognition",
"Depth Image Estimation",
"Instance Segmentation",
"Pose Estimation",
"Semantic Segmentation",
"Surface Normals Estimation"
] | [] | [] |
EDZqYJmTfU | https://paperswithcode.com/paper/pruned-non-local-means | Pruned non-local means | In Non-Local Means (NLM), each pixel is denoised by performing a weighted
averaging of its neighboring pixels, where the weights are computed using image
patches. We demonstrate that the denoising performance of NLM can be improved
by pruning the neighboring pixels, namely, by rejecting neighboring pixels
whose weights are below a certain threshold $\lambda$. While pruning can
potentially reduce pixel averaging in uniform-intensity regions, we demonstrate
that there is generally an overall improvement in the denoising performance. In
particular, the improvement comes from pixels situated close to edges and
corners. The success of the proposed method strongly depends on the choice of
the global threshold $\lambda$, which in turn depends on the noise level and
the image characteristics. We show how Stein's unbiased estimator of the
mean-squared error can be used to optimally tune $\lambda$, at a marginal
computational overhead. We present some representative denoising results to
demonstrate the superior performance of the proposed method over NLM and its
variants. | 1701.08280 | http://arxiv.org/abs/1701.08280v2 | http://arxiv.org/pdf/1701.08280v2.pdf | [
"Denoising"
] | [] | [] |
MTPHm5xPEP | https://paperswithcode.com/paper/value-propagation-networks | Value Propagation Networks | We present Value Propagation (VProp), a set of parameter-efficient
differentiable planning modules built on Value Iteration which can successfully
be trained using reinforcement learning to solve unseen tasks, has the
capability to generalize to larger map sizes, and can learn to navigate in
dynamic environments. We show that the modules enable learning to plan when the
environment also includes stochastic elements, providing a cost-efficient
learning system to build low-level size-invariant planners for a variety of
interactive navigation problems. We evaluate on static and dynamic
configurations of MazeBase grid-worlds, with randomly generated environments of
several different sizes, and on a StarCraft navigation scenario, with more
complex dynamics, and pixels as input. | 1805.11199 | http://arxiv.org/abs/1805.11199v2 | http://arxiv.org/pdf/1805.11199v2.pdf | [
"Starcraft"
] | [] | [] |