text
stringlengths 29
3.31k
| label
sequencelengths 1
11
|
---|---|
The Laplacian representation recently gains increasing attention for
reinforcement learning as it provides succinct and informative representation
for states, by taking the eigenvectors of the Laplacian matrix of the
state-transition graph as state embeddings. Such representation captures the
geometry of the underlying state space and is beneficial to RL tasks such as
option discovery and reward shaping. To approximate the Laplacian
representation in large (or even continuous) state spaces, recent works propose
to minimize a spectral graph drawing objective, which however has infinitely
many global minimizers other than the eigenvectors. As a result, their learned
Laplacian representation may differ from the ground truth. To solve this
problem, we reformulate the graph drawing objective into a generalized form and
derive a new learning objective, which is proved to have eigenvectors as its
unique global minimizer. It enables learning high-quality Laplacian
representations that faithfully approximate the ground truth. We validate this
via comprehensive experiments on a set of gridworld and continuous control
environments. Moreover, we show that our learned Laplacian representations lead
to more exploratory options and better reward shaping. | [
"cs.LG",
"cs.AI",
"stat.ML"
] |
Due to accessible big data collections from consumers, products, and stores,
advanced sales forecasting capabilities have drawn great attention from many
companies especially in the retail business because of its importance in
decision making. Improvement of the forecasting accuracy, even by a small
percentage, may have a substantial impact on companies' production and
financial planning, marketing strategies, inventory controls, supply chain
management, and eventually stock prices. Specifically, our research goal is to
forecast the sales of each product in each store in the near future. Motivated
by tensor factorization methodologies for personalized context-aware
recommender systems, we propose a novel approach called the Advanced Temporal
Latent-factor Approach to Sales forecasting (ATLAS), which achieves accurate
and individualized prediction for sales by building a single
tensor-factorization model across multiple stores and products. Our
contribution is a combination of: tensor framework (to leverage information
across stores and products), a new regularization function (to incorporate
demand dynamics), and extrapolation of tensor into future time periods using
state-of-the-art statistical (seasonal auto-regressive integrated
moving-average models) and machine-learning (recurrent neural networks) models.
The advantages of ATLAS are demonstrated on eight product category datasets
collected by the Information Resource, Inc., where a total of 165 million
weekly sales transactions from more than 1,500 grocery stores over 15,560
products are analyzed. | [
"cs.LG",
"cs.IR",
"stat.ML"
] |
Image segmentation, one of the most critical vision tasks, has been studied
for many years. Most of the early algorithms are unsupervised methods, which
use hand-crafted features to divide the image into many regions. Recently,
owing to the great success of deep learning technology, CNNs based methods show
superior performance in image segmentation. However, these methods rely on a
large number of human annotations, which are expensive to collect. In this
paper, we propose a deep unsupervised method for image segmentation, which
contains the following two stages. First, a Superpixelwise Autoencoder
(SuperAE) is designed to learn the deep embedding and reconstruct a smoothed
image, then the smoothed image is passed to generate superpixels. Second, we
present a novel clustering algorithm called Deep Superpixel Cut (DSC), which
measures the deep similarity between superpixels and formulates image
segmentation as a soft partitioning problem. Via backpropagation, DSC
adaptively partitions the superpixels into perceptual regions. Experimental
results on the BSDS500 dataset demonstrate the effectiveness of the proposed
method. | [
"cs.CV"
] |
This paper considers joint learning of multiple sparse Granger graphical
models to discover underlying common and differential Granger causality (GC)
structures across multiple time series. This can be applied to drawing
group-level brain connectivity inferences from a homogeneous group of subjects
or discovering network differences among groups of signals collected under
heterogeneous conditions. By recognizing that the GC of a single multivariate
time series can be characterized by common zeros of vector autoregressive (VAR)
lag coefficients, a group sparse prior is included in joint regularized
least-squares estimations of multiple VAR models. Group-norm regularizations
based on group- and fused-lasso penalties encourage a decomposition of multiple
networks into a common GC structure, with other remaining parts defined in
individual-specific networks. Prior information about sparseness and sparsity
patterns of desired GC networks are incorporated as relative weights, while a
non-convex group norm in the penalty is proposed to enhance the accuracy of
network estimation in low-sample settings. Extensive numerical results on
simulations illustrated our method's improvements over existing sparse
estimation approaches on GC network sparsity recovery. Our methods were also
applied to available resting-state fMRI time series from the ADHD-200 data sets
to learn the differences of causality mechanisms, called effective brain
connectivity, between adolescents with ADHD and typically developing children.
Our analysis revealed that parts of the causality differences between the two
groups often resided in the orbitofrontal region and areas associated with the
limbic system, which agreed with clinical findings and data-driven results in
previous studies. | [
"cs.LG",
"eess.SP",
"q-bio.NC"
] |
In order to continuously represent molecules, we propose a generative model
in the form of a VAE which is operating on the 2D-graph structure of molecules.
A side predictor is employed to prune the latent space and help the decoder in
generating meaningful adjacency tensor of molecules. Other than the potential
applicability in drug design and property prediction, we show the superior
performance of this technique in comparison to other similar methods based on
the SMILES representation of the molecules with RNN based encoder and decoder. | [
"cs.LG"
] |
By taking into account the properties and limitations of the human visual
system, images can be more efficiently compressed, colors more accurately
reproduced, prints better rendered. To show all these advantages in this paper
new adapted color charts have been created based on technical and visual image
category analysis. A number of tests have been carried out using extreme images
with their key information strictly in dark and light areas. It was shown that
the image categorization using the adapted color charts improves the analysis
of relevant image information with regard to both the image gradation and the
detail reproduction. The images with key information in hi-key areas were also
test printed using the adapted color charts. | [
"cs.CV"
] |
We study an exploration method for model-free RL that generalizes the
counter-based exploration bonus methods and takes into account long term
exploratory value of actions rather than a single step look-ahead. We propose a
model-free RL method that modifies Delayed Q-learning and utilizes the
long-term exploration bonus with provable efficiency. We show that our proposed
method finds a near-optimal policy in polynomial time (PAC-MDP), and also
provide experimental evidence that our proposed algorithm is an efficient
exploration method. | [
"cs.LG",
"stat.ML"
] |
We present a supervised hyperspectral image segmentation algorithm based on a
convex formulation of a marginal maximum a posteriori segmentation with hidden
fields and structure tensor regularization: Segmentation via the Constraint
Split Augmented Lagrangian Shrinkage by Structure Tensor Regularization
(SegSALSA-STR). This formulation avoids the generally discrete nature of
segmentation problems and the inherent NP-hardness of the integer optimization
associated.
We extend the Segmentation via the Constraint Split Augmented Lagrangian
Shrinkage (SegSALSA) algorithm by generalizing the vectorial total variation
prior using a structure tensor prior constructed from a patch-based Jacobian.
The resulting algorithm is convex, time-efficient and highly parallelizable.
This shows the potential of combining hidden fields with convex optimization
through the inclusion of different regularizers. The SegSALSA-STR algorithm is
validated in the segmentation of real hyperspectral images. | [
"cs.CV",
"68"
] |
Structured matrices, such as those derived from Kronecker products (KP), are
effective at compressing neural networks, but can lead to unacceptable accuracy
loss when applied to large models. In this paper, we propose the notion of
doping -- addition of an extremely sparse matrix to a structured matrix. Doping
facilitates additional degrees of freedom for a small number of parameters,
allowing them to independently diverge from the fixed structure. To train LSTMs
with doped structured matrices, we introduce the additional parameter matrix
while slowly annealing its sparsity level. However, we find that performance
degrades as we slowly sparsify the doping matrix, due to co-matrix adaptation
(CMA) between the structured and the sparse matrices. We address this over
dependence on the sparse matrix using a co-matrix dropout regularization (CMR)
scheme. We provide empirical evidence to show that doping, CMA and CMR are
concepts generally applicable to multiple structured matrices (Kronecker
Product, LMF, Hybrid Matrix Decomposition). Additionally, results with doped
kronecker product matrices demonstrate state-of-the-art accuracy at large
compression factors (10 - 25x) across 4 natural language processing
applications with minor loss in accuracy. Doped KP compression technique
outperforms previous state-of-the art compression results by achieving 1.3 -
2.4x higher compression factor at a similar accuracy, while also beating strong
alternatives like pruning and low-rank methods by a large margin (8% or more).
Additionally, we show that doped KP can be deployed on commodity hardware using
the current software stack and achieve 2.5 - 5.5x inference run-time speed-up
over baseline. | [
"cs.LG"
] |
This paper proposes an image dehazing model built with a convolutional neural
network (CNN), called All-in-One Dehazing Network (AOD-Net). It is designed
based on a re-formulated atmospheric scattering model. Instead of estimating
the transmission matrix and the atmospheric light separately as most previous
models did, AOD-Net directly generates the clean image through a light-weight
CNN. Such a novel end-to-end design makes it easy to embed AOD-Net into other
deep models, e.g., Faster R-CNN, for improving high-level task performance on
hazy images. Experimental results on both synthesized and natural hazy image
datasets demonstrate our superior performance than the state-of-the-art in
terms of PSNR, SSIM and the subjective visual quality. Furthermore, when
concatenating AOD-Net with Faster R-CNN and training the joint pipeline from
end to end, we witness a large improvement of the object detection performance
on hazy images. | [
"cs.CV",
"cs.AI"
] |
In the legal domain it is important to differentiate between words in
general, and afterwards to link the occurrences of the same entities. The topic
to solve these challenges is called Named-Entity Linking (NEL). Current
supervised neural networks designed for NEL use publicly available datasets for
training and testing. However, this paper focuses especially on the aspect of
applying transfer learning approach using networks trained for NEL to legal
documents. Experiments show consistent improvement in the legal datasets that
were created from the European Union law in the scope of this research. Using
transfer learning approach, we reached F1-score of 98.90\% and 98.01\% on the
legal small and large test dataset. | [
"cs.LG",
"cs.AI",
"cs.CL",
"stat.ML"
] |
Closed Form is a propagation based matting algorithm, functioning well on
images with good propagation . The deficiency of the Closed Form method is that
for complex areas with poor image propagation , such as hole areas or areas of
long and narrow structures. The right results are usually hard to get. On these
areas, if certain flags are provided, it can improve the effects of matting. In
this paper, we design a matting algorithm by local sampling and the KNN
classifier propagation based matting algorithm. First of all, build the
corresponding features space according to the different components of image
colors to reduce the influence of overlapping between the foreground and
background, and to improve the classification accuracy of KNN classifier.
Second, adaptively use local sampling or using local KNN classifier for
processing based on the pros and cons of the sample performance of unknown
image areas. Finally, based on different treatment methods for the unknown
areas, we will use different weight for augmenting constraints to make the
treatment more effective. In this paper, by combining qualitative observation
and quantitative analysis, we will make evaluation of the experimental results
through online standard set of evaluation tests. It shows that on images with
good propagation , this method is as effective as the Closed Form method, while
on images in complex regions, it can perform even better than Closed Form. | [
"cs.CV"
] |
Many real-world systems problems require reasoning about the long term
consequences of actions taken to configure and manage the system. These
problems with delayed and often sequentially aggregated reward, are often
inherently reinforcement learning problems and present the opportunity to
leverage the recent substantial advances in deep reinforcement learning.
However, in some cases, it is not clear why deep reinforcement learning is a
good fit for the problem. Sometimes, it does not perform better than the
state-of-the-art solutions. And in other cases, random search or greedy
algorithms could outperform deep reinforcement learning. In this paper, we
review, discuss, and evaluate the recent trends of using deep reinforcement
learning in system optimization. We propose a set of essential metrics to guide
future works in evaluating the efficacy of using deep reinforcement learning in
system optimization. Our evaluation includes challenges, the types of problems,
their formulation in the deep reinforcement learning setting, embedding, the
model used, efficiency, and robustness. We conclude with a discussion on open
challenges and potential directions for pushing further the integration of
reinforcement learning in system optimization. | [
"cs.LG",
"cs.AI",
"cs.SY",
"eess.SY"
] |
Circuit design is complicated and requires extensive domain-specific
expertise. One major obstacle stuck on the way to hardware agile development is
the considerably time-consuming process of accurate circuit quality evaluation.
To significantly expedite the circuit evaluation during the translation from
behavioral languages to circuit designs, we formulate it as a
Program-to-Circuit problem, aiming to exploit the representation power of graph
neural networks (GNNs) by representing C/C++ programs as graphs. The goal of
this work is four-fold. First, we build a standard benchmark containing 40k
C/C++ programs, each of which is translated to a circuit design with actual
hardware quality metrics, aiming to facilitate the development of effective
GNNs targeting this high-demand circuit design area. Second, 14
state-of-the-art GNN models are analyzed on the Program-to-Circuit problem. We
identify key design challenges of this problem, which should be carefully
handled but not yet solved by existing GNNs. The goal is to provide
domain-specific knowledge for designing GNNs with suitable inductive biases.
Third, we discuss three sets of real-world benchmarks for GNN generalization
evaluation, and analyze the performance gap between standard programs and the
real-case ones. The goal is to enable transfer learning from limited training
data to real-world large-scale circuit design problems. Fourth, the
Program-to-Circuit problem is a representative within the Program-to-X
framework, a set of program-based analysis problems with various downstream
tasks. The in-depth understanding of strength and weaknesses in applying GNNs
on Program-to-Circuit could largely benefit the entire family of Program-to-X.
Pioneering in this direction, we expect more GNN endeavors to revolutionize
this high-demand Program-to-Circuit problem and to enrich the expressiveness of
GNNs on programs. | [
"cs.LG"
] |
We introduce a novel task, Video Question Generation (Video QG). A Video QG
model automatically generates questions given a video clip and its
corresponding dialogues. Video QG requires a range of skills -- sentence
comprehension, temporal relation, the interplay between vision and language,
and the ability to ask meaningful questions. To address this, we propose a
novel semantic rich cross-modal self-attention (SRCMSA) network to aggregate
the multi-modal and diverse features. To be more precise, we enhance the video
frames semantic by integrating the object-level information, and we jointly
consider the cross-modal attention for the video question generation task.
Excitingly, our proposed model remarkably improves the baseline from 7.58 to
14.48 in the BLEU-4 score on the TVQA dataset. Most of all, we arguably pave a
novel path toward understanding the challenging video input and we provide
detailed analysis in terms of diversity, which ushers the avenues for future
investigations. | [
"cs.CV",
"cs.CL"
] |
The tracking-by-detection framework requires a set of positive and negative
training samples to learn robust tracking models for precise localization of
target objects. However, existing tracking models mostly treat different
samples independently while ignores the relationship information among them. In
this paper, we propose a novel structure-aware deep neural network to overcome
such limitations. In particular, we construct a graph to represent the pairwise
relationships among training samples, and additionally take the natural
language as the supervised information to learn both feature representations
and classifiers robustly. To refine the states of the target and re-track the
target when it is back to view from heavy occlusion and out of view, we
elaborately design a novel subnetwork to learn the target-driven visual
attentions from the guidance of both visual and natural language cues.
Extensive experiments on five tracking benchmark datasets validated the
effectiveness of our proposed method. | [
"cs.CV"
] |
In recent years, semi-supervised learning (SSL) has shown tremendous success
in leveraging unlabeled data to improve the performance of deep learning
models, which significantly reduces the demand for large amounts of labeled
data. Many SSL techniques have been proposed and have shown promising
performance on famous datasets such as ImageNet and CIFAR-10. However, some
exiting techniques (especially data augmentation based) are not suitable for
industrial applications empirically. Therefore, this work proposes the
pseudo-representation labeling, a simple and flexible framework that utilizes
pseudo-labeling techniques to iteratively label a small amount of unlabeled
data and use them as training data. In addition, our framework is integrated
with self-supervised representation learning such that the classifier gains
benefits from representation learning of both labeled and unlabeled data. This
framework can be implemented without being limited at the specific model
structure, but a general technique to improve the existing model. Compared with
the existing approaches, the pseudo-representation labeling is more intuitive
and can effectively solve practical problems in the real world. Empirically, it
outperforms the current state-of-the-art semi-supervised learning methods in
industrial types of classification problems such as the WM-811K wafer map and
the MIT-BIH Arrhythmia dataset. | [
"cs.LG",
"cs.CV",
"stat.ML"
] |
For humans, visual understanding is inherently generative: given a 3D shape,
we can postulate how it would look in the world; given a 2D image, we can infer
the 3D structure that likely gave rise to it. We can thus translate between the
2D visual and 3D structural modalities of a given object. In the context of
computer vision, this corresponds to a learnable module that serves two
purposes: (i) generate a realistic rendering of a 3D object (shape-to-image
translation) and (ii) infer a realistic 3D shape from an image (image-to-shape
translation). In this paper, we learn such a module while being conscious of
the difficulties in obtaining large paired 2D-3D datasets. By leveraging
generative domain translation methods, we are able to define a learning
algorithm that requires only weak supervision, with unpaired data. The
resulting model is not only able to perform 3D shape, pose, and texture
inference from 2D images, but can also generate novel textured 3D shapes and
renders, similar to a graphics pipeline. More specifically, our method (i)
infers an explicit 3D mesh representation, (ii) utilizes example shapes to
regularize inference, (iii) requires only an image mask (no keypoints or camera
extrinsics), and (iv) has generative capabilities. While prior work explores
subsets of these properties, their combination is novel. We demonstrate the
utility of our learned representation, as well as its performance on image
generation and unpaired 3D shape inference tasks. | [
"cs.CV",
"cs.LG",
"I.2.10; I.2.6"
] |
On an artist's profile page, music streaming services frequently recommend a
ranked list of "similar artists" that fans also liked. However, implementing
such a feature is challenging for new artists, for which usage data on the
service (e.g. streams or likes) is not yet available. In this paper, we model
this cold start similar artists ranking problem as a link prediction task in a
directed and attributed graph, connecting artists to their top-k most similar
neighbors and incorporating side musical information. Then, we leverage a graph
autoencoder architecture to learn node embedding representations from this
graph, and to automatically rank the top-k most similar neighbors of new
artists using a gravity-inspired mechanism. We empirically show the flexibility
and the effectiveness of our framework, by addressing a real-world cold start
similar artists ranking problem on a global music streaming service. Along with
this paper, we also publicly release our source code as well as the industrial
graph data from our experiments. | [
"cs.LG",
"cs.IR",
"cs.SI"
] |
With advances in Generative Adversarial Networks (GANs) leading to
dramatically-improved synthetic images and video, there is an increased need
for algorithms which extend traditional forensics to this new category of
imagery. While GANs have been shown to be helpful in a number of computer
vision applications, there are other problematic uses such as `deep fakes'
which necessitate such forensics. Source camera attribution algorithms using
various cues have addressed this need for imagery captured by a camera, but
there are fewer options for synthetic imagery. We address the problem of
attributing a synthetic image to a specific generator in a white box setting,
by inverting the process of generation. This enables us to simultaneously
determine whether the generator produced the image and recover an input which
produces a close match to the synthetic image. | [
"cs.CV",
"cs.LG",
"eess.IV"
] |
Robust Markov Decision Processes (RMDPs) intend to ensure robustness with
respect to changing or adversarial system behavior. In this framework,
transitions are modeled as arbitrary elements of a known and properly
structured uncertainty set and a robust optimal policy can be derived under the
worst-case scenario. In this study, we address the issue of learning in RMDPs
using a Bayesian approach. We introduce the Uncertainty Robust Bellman Equation
(URBE) which encourages safe exploration for adapting the uncertainty set to
new observations while preserving robustness. We propose a URBE-based
algorithm, DQN-URBE, that scales this method to higher dimensional domains. Our
experiments show that the derived URBE-based strategy leads to a better
trade-off between less conservative solutions and robustness in the presence of
model misspecification. In addition, we show that the DQN-URBE algorithm can
adapt significantly faster to changing dynamics online compared to existing
robust techniques with fixed uncertainty sets. | [
"cs.LG",
"cs.AI",
"stat.ML"
] |
Lately, post-training quantization methods have gained considerable
attention, as they are simple to use, and require only a small unlabeled
calibration set. This small dataset cannot be used to fine-tune the model
without significant over-fitting. Instead, these methods only use the
calibration set to set the activations' dynamic ranges. However, such methods
always resulted in significant accuracy degradation, when used below 8-bits
(except on small datasets). Here we aim to break the 8-bit barrier. To this
end, we minimize the quantization errors of each layer separately by optimizing
its parameters over the calibration set. We empirically demonstrate that this
approach is: (1) much less susceptible to over-fitting than the standard
fine-tuning approaches, and can be used even on a very small calibration set;
and (2) more powerful than previous methods, which only set the activations'
dynamic ranges. Furthermore, we demonstrate how to optimally allocate the
bit-widths for each layer, while constraining accuracy degradation or model
compression by proposing a novel integer programming formulation. Finally, we
suggest model global statistics tuning, to correct biases introduced during
quantization. Together, these methods yield state-of-the-art results for both
vision and text models. For instance, on ResNet50, we obtain less than 1\%
accuracy degradation --- with 4-bit weights and activations in all layers, but
the smallest two. We open-sourced our code. | [
"cs.LG",
"stat.ML"
] |
The detection of vehicles in aerial images is widely applied in many
applications. Comparing with object detection in the ground view images,
vehicle detection in aerial images remains a challenging problem because of
small vehicle size, monotone appearance and the complex background. In this
paper, we propose a novel double focal loss convolutional neural network
framework (DFL-CNN). In the proposed framework, the skip connection is used in
the CNN structure to enhance the feature learning. Also, the focal loss
function is used to substitute for conventional cross entropy loss function in
both of the region proposed network and the final classifier. We further
introduce the first large-scale vehicle detection dataset ITCVD with ground
truth annotations for all the vehicles in the scene. We demonstrate the
performance of our model on the existing benchmark DLR 3K dataset as well as
the ITCVD dataset. The experimental results show that our DFL-CNN outperforms
the baselines on vehicle detection. | [
"cs.CV"
] |
Capsule Networks, as alternatives to Convolutional Neural Networks, have been
proposed to recognize objects from images. The current literature demonstrates
many advantages of CapsNets over CNNs. However, how to create explanations for
individual classifications of CapsNets has not been well explored. The widely
used saliency methods are mainly proposed for explaining CNN-based
classifications; they create saliency map explanations by combining activation
values and the corresponding gradients, e.g., Grad-CAM. These saliency methods
require a specific architecture of the underlying classifiers and cannot be
trivially applied to CapsNets due to the iterative routing mechanism therein.
To overcome the lack of interpretability, we can either propose new post-hoc
interpretation methods for CapsNets or modifying the model to have build-in
explanations. In this work, we explore the latter. Specifically, we propose
interpretable Graph Capsule Networks (GraCapsNets), where we replace the
routing part with a multi-head attention-based Graph Pooling approach. In the
proposed model, individual classification explanations can be created
effectively and efficiently. Our model also demonstrates some unexpected
benefits, even though it replaces the fundamental part of CapsNets. Our
GraCapsNets achieve better classification performance with fewer parameters and
better adversarial robustness, when compared to CapsNets. Besides, GraCapsNets
also keep other advantages of CapsNets, namely, disentangled representations
and affine transformation robustness. | [
"cs.CV"
] |
We present a supervised hyperspectral image segmentation algorithm based on a
convex formulation of a marginal maximum a posteriori segmentation with hidden
fields and structure tensor regularization: Segmentation via the Constraint
Split Augmented Lagrangian Shrinkage by Structure Tensor Regularization
(SegSALSA-STR). This formulation avoids the generally discrete nature of
segmentation problems and the inherent NP-hardness of the integer optimization
associated.
We extend the Segmentation via the Constraint Split Augmented Lagrangian
Shrinkage (SegSALSA) algorithm by generalizing the vectorial total variation
prior using a structure tensor prior constructed from a patch-based Jacobian.
The resulting algorithm is convex, time-efficient and highly parallelizable.
This shows the potential of combining hidden fields with convex optimization
through the inclusion of different regularizers. The SegSALSA-STR algorithm is
validated in the segmentation of real hyperspectral images. | [
"cs.CV",
"68"
] |
We study unsupervised video representation learning that seeks to learn both
motion and appearance features from unlabeled video only, which can be reused
for downstream tasks such as action recognition. This task, however, is
extremely challenging due to 1) the highly complex spatial-temporal information
in videos; and 2) the lack of labeled data for training. Unlike the
representation learning for static images, it is difficult to construct a
suitable self-supervised task to well model both motion and appearance
features. More recently, several attempts have been made to learn video
representation through video playback speed prediction. However, it is
non-trivial to obtain precise speed labels for the videos. More critically, the
learnt models may tend to focus on motion pattern and thus may not learn
appearance features well. In this paper, we observe that the relative playback
speed is more consistent with motion pattern, and thus provide more effective
and stable supervision for representation learning. Therefore, we propose a new
way to perceive the playback speed and exploit the relative speed between two
video clips as labels. In this way, we are able to well perceive speed and
learn better motion features. Moreover, to ensure the learning of appearance
features, we further propose an appearance-focused task, where we enforce the
model to perceive the appearance difference between two video clips. We show
that optimizing the two tasks jointly consistently improves the performance on
two downstream tasks, namely action recognition and video retrieval.
Remarkably, for action recognition on UCF101 dataset, we achieve 93.7% accuracy
without the use of labeled data for pre-training, which outperforms the
ImageNet supervised pre-trained model. Code and pre-trained models can be found
at https://github.com/PeihaoChen/RSPNet. | [
"cs.CV"
] |
The task of unsupervised motion retargeting in videos has seen substantial
advancements through the use of deep neural networks. While early works
concentrated on specific object priors such as a human face or body, recent
work considered the unsupervised case. When the source and target videos,
however, are of different shapes, current methods fail. To alleviate this
problem, we introduce JOKR - a JOint Keypoint Representation that captures the
motion common to both the source and target videos, without requiring any
object prior or data collection. By employing a domain confusion term, we
enforce the unsupervised keypoint representations of both videos to be
indistinguishable. This encourages disentanglement between the parts of the
motion that are common to the two domains, and their distinctive appearance and
motion, enabling the generation of videos that capture the motion of the one
while depicting the style of the other. To enable cases where the objects are
of different proportions or orientations, we apply a learned affine
transformation between the JOKRs. This augments the representation to be affine
invariant, and in practice broadens the variety of possible retargeting pairs.
This geometry-driven representation enables further intuitive control, such as
temporal coherence and manual editing. Through comprehensive experimentation,
we demonstrate the applicability of our method to different challenging
cross-domain video pairs. We evaluate our method both qualitatively and
quantitatively, and demonstrate that our method handles various cross-domain
scenarios, such as different animals, different flowers, and humans. We also
demonstrate superior temporal coherency and visual quality compared to
state-of-the-art alternatives, through statistical metrics and a user study.
Source code and videos can be found at https://rmokady.github.io/JOKR/ . | [
"cs.CV"
] |
The signature transform is a 'universal nonlinearity' on the space of
continuous vector-valued paths, and has received attention for use in machine
learning on time series. However, real-world temporal data is typically
observed at discrete points in time, and must first be transformed into a
continuous path before signature techniques can be applied. We make this step
explicit by characterising it as an imputation problem, and empirically assess
the impact of various imputation strategies when applying signature-based
neural nets to irregular time series data. For one of these strategies,
Gaussian process (GP) adapters, we propose an extension~(GP-PoM) that makes
uncertainty information directly available to the subsequent classifier while
at the same time preventing costly Monte-Carlo (MC) sampling. In our
experiments, we find that the choice of imputation drastically affects shallow
signature models, whereas deeper architectures are more robust. Next, we
observe that uncertainty-aware predictions (based on GP-PoM or indicator
imputations) are beneficial for predictive performance, even compared to the
uncertainty-aware training of conventional GP adapters. In conclusion, we have
demonstrated that the path construction is indeed crucial for signature models
and that our proposed strategy leads to competitive performance in general,
while improving robustness of signature models in particular. | [
"cs.LG",
"stat.ML"
] |
In neuroscience, the functional magnetic resonance imaging (fMRI) is a vital
tool to non-invasively access brain activity. Using fMRI, the functional
connectivity (FC) between brain regions can be inferred, which has contributed
to a number of findings of the fundamental properties of the brain. As an
important clinical application of FC, clustering of subjects based on FC
recently draws much attention, which can potentially reveal important
heterogeneity in subjects such as subtypes of psychiatric disorders. In
particular, a multiple-view clustering method is a powerful analytical tool,
which identifies clustering patterns of subjects depending on their FC in
specific brain areas. However, when one applies an existing multiple-view
clustering method to fMRI data, there is a need to simplify the data structure,
independently dealing with elements in a FC matrix, i.e., vectorizing a
correlation matrix. Such a simplification may distort the clustering results.
To overcome this problem, we propose a novel multiple-view clustering method
based on Wishart mixture models, which preserves the correlation matrix
structure without vectorization. The uniqueness of this method is that the
multiple-view clustering of subjects is based on particular networks of nodes
(or regions of interest, ROIs), optimized in a data-driven manner. Hence, it
can identify multiple underlying pairs of associations between a subject
cluster solution and a ROI sub-network. The key assumption of the method is
independence among sub-networks, which is effectively addressed by whitening
correlation matrices. We applied the proposed method to synthetic and fMRI
data, demonstrating the usefulness and power of the proposed method. | [
"stat.ML",
"cs.LG"
] |
Deep learning-based medical image segmentation technology aims at automatic
recognizing and annotating objects on the medical image. Non-local attention
and feature learning by multi-scale methods are widely used to model network,
which drives progress in medical image segmentation. However, those attention
mechanism methods have weakly non-local receptive fields' strengthened
connection for small objects in medical images. Then, the features of important
small objects in abstract or coarse feature maps may be deserted, which leads
to unsatisfactory performance. Moreover, the existing multi-scale methods only
simply focus on different sizes of view, whose sparse multi-scale features
collected are not abundant enough for small objects segmentation. In this work,
a multi-dimensional attention segmentation model with cascade multi-scale
convolution is proposed to predict accurate segmentation for small objects in
medical images. As the weight function, multi-dimensional attention modules
provide coefficient modification for significant/informative small objects
features. Furthermore, The cascade multi-scale convolution modules in each
skip-connection path are exploited to capture multi-scale features in different
semantic depth. The proposed method is evaluated on three datasets: KiTS19,
Pancreas CT of Decathlon-10, and MICCAI 2018 LiTS Challenge, demonstrating
better segmentation performances than the state-of-the-art baselines. | [
"cs.CV"
] |
We study the problem of learning Granger causality between event types from
asynchronous, interdependent, multi-type event sequences. Existing work suffers
from either limited model flexibility or poor model explainability and thus
fails to uncover Granger causality across a wide variety of event sequences
with diverse event interdependency. To address these weaknesses, we propose
CAUSE (Causality from AttribUtions on Sequence of Events), a novel framework
for the studied task. The key idea of CAUSE is to first implicitly capture the
underlying event interdependency by fitting a neural point process, and then
extract from the process a Granger causality statistic using an axiomatic
attribution method. Across multiple datasets riddled with diverse event
interdependency, we demonstrate that CAUSE achieves superior performance on
correctly inferring the inter-type Granger causality over a range of
state-of-the-art methods. | [
"cs.LG",
"stat.ML"
] |
The instance segmentation can be considered an extension of the object
detection problem where bounding boxes are replaced by object contours.
Strictly speaking the problem requires to identify each pixel instance and
class independently of the artifice used for this mean. The advantage of
instance segmentation over the usual object detection lies in the precise
delineation of objects improving object localization. Additionally, object
contours allow the evaluation of partial occlusion with basic image processing
algorithms. This work approaches the instance segmentation problem as an
annotation problem and presents a novel technique to encode and decode ground
truth annotations. We propose a mathematical representation of instances that
any deep semantic segmentation model can learn and generalize. Each individual
instance is represented by a center of mass and a field of vectors pointing to
it. This encoding technique has been denominated Distance to Center of Mass
Encoding (DCME). | [
"cs.CV"
] |
Pretraining general-purpose visual features has become a crucial part of
tackling many computer vision tasks. While one can learn such features on the
extensively-annotated ImageNet dataset, recent approaches have looked at ways
to allow for noisy, fewer, or even no annotations to perform such pretraining.
Starting from the observation that captioned images are easily crawlable, we
argue that this overlooked source of information can be exploited to supervise
the training of visual representations. To do so, motivated by the recent
progresses in language models, we introduce {\em image-conditioned masked
language modeling} (ICMLM) -- a proxy task to learn visual representations over
image-caption pairs. ICMLM consists in predicting masked words in captions by
relying on visual cues. To tackle this task, we propose hybrid models, with
dedicated visual and textual encoders, and we show that the visual
representations learned as a by-product of solving this task transfer well to a
variety of target tasks. Our experiments confirm that image captions can be
leveraged to inject global and localized semantic information into visual
representations. Project website: https://europe.naverlabs.com/icmlm. | [
"cs.CV"
] |
Wind power prediction is of vital importance in wind power utilization. There
have been a lot of researches based on the time series of the wind power or
speed, but In fact, these time series cannot express the temporal and spatial
changes of wind, which fundamentally hinders the advance of wind power
prediction. In this paper, a new kind of feature that can describe the process
of temporal and spatial variation is proposed, namely, Spatio-Temporal
Features. We first map the data collected at each moment from the wind turbine
to the plane to form the state map, namely, the scene, according to the
relative positions. The scene time series over a period of time is a
multi-channel image, i.e. the Spatio-Temporal Features. Based on the
Spatio-Temporal Features, the deep convolutional network is applied to predict
the wind power, achieving a far better accuracy than the existing methods.
Compared with the starge-of-the-art method, the mean-square error (MSE) in our
method is reduced by 49.83%, and the average time cost for training models can
be shortened by a factor of more than 150. | [
"cs.LG",
"stat.ML"
] |
Convolutional Neural Networks (CNNs) have been used successfully across a
broad range of areas including data mining, object detection, and in business.
The dominance of CNNs follows a breakthrough by Alex Krizhevsky which showed
improvements by dramatically reducing the error rate obtained in a general
image classification task from 26.2% to 15.4%. In road safety, CNNs have been
applied widely to the detection of traffic signs, obstacle detection, and lane
departure checking. In addition, CNNs have been used in data mining systems
that monitor driving patterns and recommend rest breaks when appropriate. This
paper presents a driver drowsiness detection system and shows that there are
potential social challenges regarding the application of these techniques, by
highlighting problems in detecting dark-skinned driver's faces. This is a
particularly important challenge in African contexts, where there are more
dark-skinned drivers. Unfortunately, publicly available datasets are often
captured in different cultural contexts, and therefore do not cover all
ethnicities, which can lead to false detections or racially biased models. This
work evaluates the performance obtained when training convolutional neural
network models on commonly used driver drowsiness detection datasets and
testing on datasets specifically chosen for broader representation. Results
show that models trained using publicly available datasets suffer extensively
from over-fitting, and can exhibit racial bias, as shown by testing on a more
representative dataset. We propose a novel visualisation technique that can
assist in identifying groups of people where there might be the potential of
discrimination, using Principal Component Analysis (PCA) to produce a grid of
faces sorted by similarity, and combining these with a model accuracy overlay. | [
"cs.CV"
] |
Standard plane recognition plays an important role in prenatal ultrasound
(US) screening. Automatically recognizing the standard plane along with the
corresponding anatomical structures in US image can not only facilitate US
image interpretation but also improve diagnostic efficiency. In this study, we
build a novel multi-label learning (MLL) scheme to identify multiple standard
planes and corresponding anatomical structures of fetus simultaneously. Our
contribution is three-fold. First, we represent the class correlation by word
embeddings to capture the fine-grained semantic and latent statistical
concurrency. Second, we equip the MLL with a graph convolutional network to
explore the inner and outer relationship among categories. Third, we propose a
novel cluster relabel-based contrastive learning algorithm to encourage the
divergence among ambiguous classes. Extensive validation was performed on our
large in-house dataset. Our approach reports the highest accuracy as 90.25% for
standard planes labeling, 85.59% for planes and structures labeling and mAP as
94.63%. The proposed MLL scheme provides a novel perspective for standard plane
recognition and can be easily extended to other medical image classification
tasks. | [
"cs.CV"
] |
In this work we propose a method based on geometric deep learning to predict
the complete surface of the liver, given a partial point cloud of the organ
obtained during the surgical laparoscopic procedure. We introduce a new data
augmentation technique that randomly perturbs shapes in their frequency domain
to compensate the limited size of our dataset. The core of our method is a
variational autoencoder (VAE) that is trained to learn a latent space for
complete shapes of the liver. At inference time, the generative part of the
model is embedded in an optimisation procedure where the latent representation
is iteratively updated to generate a model that matches the intraoperative
partial point cloud. The effect of this optimisation is a progressive non-rigid
deformation of the initially generated shape. Our method is qualitatively
evaluated on real data and quantitatively evaluated on synthetic data. We
compared with a state-of-the-art rigid registration algorithm, that our method
outperformed in visible areas. | [
"cs.CV",
"cs.LG"
] |
Empowerment is an information-theoretic method that can be used to
intrinsically motivate learning agents. It attempts to maximize an agent's
control over the environment by encouraging visiting states with a large number
of reachable next states. Empowered learning has been shown to lead to complex
behaviors, without requiring an explicit reward signal. In this paper, we
investigate the use of empowerment in the presence of an extrinsic reward
signal. We hypothesize that empowerment can guide reinforcement learning (RL)
agents to find good early behavioral solutions by encouraging highly empowered
states. We propose a unified Bellman optimality principle for empowered reward
maximization. Our empowered reward maximization approach generalizes both
Bellman's optimality principle as well as recent information-theoretical
extensions to it. We prove uniqueness of the empowered values and show
convergence to the optimal solution. We then apply this idea to develop
off-policy actor-critic RL algorithms which we validate in high-dimensional
continuous robotics domains (MuJoCo). Our methods demonstrate improved initial
and competitive final performance compared to model-free state-of-the-art
techniques. | [
"cs.LG",
"cs.AI",
"stat.ML"
] |
Real-world visual recognition problems often exhibit long-tailed
distributions, where the amount of data for learning in different categories
shows significant imbalance. Standard classification models learned on such
data distribution often make biased predictions towards the head classes while
generalizing poorly to the tail classes. In this paper, we present two
effective modifications of CNNs to improve network learning from long-tailed
distribution. First, we present a Class Activation Map Calibration (CAMC)
module to improve the learning and prediction of network classifiers, by
enforcing network prediction based on important image regions. The proposed
CAMC module highlights the correlated image regions across data and reinforces
the representations in these areas to obtain a better global representation for
classification. Furthermore, we investigate the use of normalized classifiers
for representation learning in long-tailed problems. Our empirical study
demonstrates that by simply scaling the outputs of the classifier with an
appropriate scalar, we can effectively improve the classification accuracy on
tail classes without losing the accuracy of head classes. We conduct extensive
experiments to validate the effectiveness of our design and we set new
state-of-the-art performance on five benchmarks, including ImageNet-LT,
Places-LT, iNaturalist 2018, CIFAR10-LT, and CIFAR100-LT. | [
"cs.CV"
] |
Lighting estimation from face images is an important task and has
applications in many areas such as image editing, intrinsic image
decomposition, and image forgery detection. We propose to train a deep
Convolutional Neural Network (CNN) to regress lighting parameters from a single
face image. Lacking massive ground truth lighting labels for face images in the
wild, we use an existing method to estimate lighting parameters, which are
treated as ground truth with unknown noises. To alleviate the effect of such
noises, we utilize the idea of Generative Adversarial Networks (GAN) and
propose a Label Denoising Adversarial Network (LDAN) to make use of synthetic
data with accurate ground truth to help train a deep CNN for lighting
regression on real face images. Experiments show that our network outperforms
existing methods in producing consistent lighting parameters of different faces
under similar lighting conditions. Moreover, our method is 100,000 times faster
in execution time than prior optimization-based lighting estimation approaches. | [
"cs.CV"
] |
Gradient Boosting Machines (GBM) are among the go-to algorithms on tabular
data, which produce state of the art results in many prediction tasks. Despite
its popularity, the GBM framework suffers from a fundamental flaw in its base
learners. Specifically, most implementations utilize decision trees that are
typically biased towards categorical variables with large cardinalities. The
effect of this bias was extensively studied over the years, mostly in terms of
predictive performance. In this work, we extend the scope and study the effect
of biased base learners on GBM feature importance (FI) measures. We show that
although these implementation demonstrate highly competitive predictive
performance, they still, surprisingly, suffer from bias in FI. By utilizing
cross-validated (CV) unbiased base learners, we fix this flaw at a relatively
low computational cost. We demonstrate the suggested framework in a variety of
synthetic and real-world setups, showing a significant improvement in all GBM
FI measures while maintaining relatively the same level of prediction accuracy. | [
"cs.LG",
"stat.ML"
] |
The heavy traffic and related issues have always been concerns for modern
cities. With the help of deep learning and reinforcement learning, people have
proposed various policies to solve these traffic-related problems, such as
smart traffic signal control systems and taxi dispatching systems. People
usually validate these policies in a city simulator, since directly applying
them in the real city introduces real cost. However, these policies validated
in the city simulator may fail in the real city if the simulator is
significantly different from the real world. To tackle this problem, we need to
build a real-like traffic simulation system. Therefore, in this paper, we
propose to learn the human routing model, which is one of the most essential
part in the traffic simulator. This problem has two major challenges. First,
human routing decisions are determined by multiple factors, besides the common
time and distance factor. Second, current historical routes data usually covers
just a small portion of vehicles, due to privacy and device availability
issues. To address these problems, we propose a theory-guided residual network
model, where the theoretical part can emphasize the general principles for
human routing decisions (e.g., fastest route), and the residual part can
capture drivable condition preferences (e.g., local road or highway). Since the
theoretical part is composed of traditional shortest path algorithms that do
not need data to train, our residual network can learn human routing models
from limited data. We have conducted extensive experiments on multiple
real-world datasets to show the superior performance of our model, especially
with small data. Besides, we have also illustrated why our model is better at
recovering real routes through case studies. | [
"cs.LG",
"cs.AI",
"cs.IR"
] |
Using current reinforcement learning methods, it has recently become possible
to learn to play unknown 3D games from raw pixels. In this work, we study the
challenges that arise in such complex environments, and summarize current
methods to approach these. We choose a task within the Doom game, that has not
been approached yet. The goal for the agent is to fight enemies in a 3D world
consisting of five rooms. We train the DQN and LSTM-A3C algorithms on this
task. Results show that both algorithms learn sensible policies, but fail to
achieve high scores given the amount of training. We provide insights into the
learned behavior, which can serve as a valuable starting point for further
research in the Doom domain. | [
"cs.LG",
"cs.AI"
] |
We propose a new equilibrium enforcing method paired with a loss derived from
the Wasserstein distance for training auto-encoder based Generative Adversarial
Networks. This method balances the generator and discriminator during training.
Additionally, it provides a new approximate convergence measure, fast and
stable training and high visual quality. We also derive a way of controlling
the trade-off between image diversity and visual quality. We focus on the image
generation task, setting a new milestone in visual quality, even at higher
resolutions. This is achieved while using a relatively simple model
architecture and a standard training procedure. | [
"cs.LG",
"stat.ML"
] |
We propose directed time series regression, a new approach to estimating
parameters of time-series models for use in certainty equivalent model
predictive control. The approach combines merits of least squares regression
and empirical optimization. Through a computational study involving a
stochastic version of a well known inverted pendulum balancing problem, we
demonstrate that directed time series regression can generate significant
improvements in controller performance over either of the aforementioned
alternatives. | [
"cs.LG",
"cs.SY",
"stat.ML"
] |
The majority of descriptor-based methods for geometric processing of
non-rigid shape rely on hand-crafted descriptors. Recently, learning-based
techniques have been shown effective, achieving state-of-the-art results in a
variety of tasks. Yet, even though these methods can in principle work directly
on raw data, most methods still rely on hand-crafted descriptors at the input
layer. In this work, we wish to challenge this practice and use a neural
network to learn descriptors directly from the raw mesh. To this end, we
introduce two modules into our neural architecture. The first is a local
reference frame (LRF) used to explicitly make the features invariant to rigid
transformations. The second is continuous convolution kernels that provide
robustness to sampling. We show the efficacy of our proposed network in
learning on raw meshes using two cornerstone tasks: shape matching, and human
body parts segmentation. Our results show superior results over baseline
methods that use hand-crafted descriptors. | [
"cs.CV",
"cs.CG"
] |
Recent work has proven the effectiveness of transformers in many computer
vision tasks. However, the performance of transformers in gaze estimation is
still unexplored. In this paper, we employ transformers and assess their
effectiveness for gaze estimation. We consider two forms of vision transformer
which are pure transformers and hybrid transformers. We first follow the
popular ViT and employ a pure transformer to estimate gaze from images. On the
other hand, we preserve the convolutional layers and integrate CNNs as well as
transformers. The transformer serves as a component to complement CNNs. We
compare the performance of the two transformers in gaze estimation. The Hybrid
transformer significantly outperforms the pure transformer in all evaluation
datasets with less parameters. We further conduct experiments to assess the
effectiveness of the hybrid transformer and explore the advantage of
self-attention mechanism. Experiments show the hybrid transformer can achieve
state-of-the-art performance in all benchmarks with pre-training.To facilitate
further research, we release codes and models in
https://github.com/yihuacheng/GazeTR. | [
"cs.CV"
] |
This paper represents an text extraction method from Google maps, GIS
maps/images. Due to an unsupervised approach there is no requirement of any
prior knowledge or training set about the textual and non-textual parts. Fuzzy
CMeans clustering technique is used for image segmentation and Prewitt method
is used to detect the edges. Connected component analysis and gridding
technique enhance the correctness of the results. The proposed method reaches
98.5% accuracy level on the basis of experimental data sets. | [
"cs.CV",
"cs.AI"
] |
Identifying mobility behaviors in rich trajectory data is of great economic
and social interest to various applications including urban planning, marketing
and intelligence. Existing work on trajectory clustering often relies on
similarity measurements that utilize raw spatial and/or temporal information of
trajectories. These measures are incapable of identifying similar moving
behaviors that exhibit varying spatio-temporal scales of movement. In addition,
the expense of labeling massive trajectory data is a barrier to supervised
learning models. To address these challenges, we propose an unsupervised neural
approach for mobility behavior clustering, called the Deep Embedded TrajEctory
ClusTering network (DETECT). DETECT operates in three parts: first it
transforms the trajectories by summarizing their critical parts and augmenting
them with context derived from their geographical locality (e.g., using POIs
from gazetteers). In the second part, it learns a powerful representation of
trajectories in the latent space of behaviors, thus enabling a clustering
function (such as $k$-means) to be applied. Finally, a clustering oriented loss
is directly built on the embedded features to jointly perform feature
refinement and cluster assignment, thus improving separability between mobility
behaviors. Exhaustive quantitative and qualitative experiments on two
real-world datasets demonstrate the effectiveness of our approach for mobility
behavior analyses. | [
"cs.LG",
"stat.ML"
] |
With the maturity of visual detection techniques, we are more ambitious in
describing visual content with open-vocabulary, fine-grained and free-form
language, i.e., the task of image captioning. In particular, we are interested
in generating longer, richer and more fine-grained sentences and paragraphs as
image descriptions. Image captioning can be translated to the task of
sequential language prediction given visual content, where the output sequence
forms natural language description with plausible grammar. However, existing
image captioning methods focus only on language policy while not visual policy,
and thus fail to capture visual context that are crucial for compositional
reasoning such as object relationships (e.g., "man riding horse") and visual
comparisons (e.g., "small(er) cat"). This issue is especially severe when
generating longer sequences such as a paragraph. To fill the gap, we propose a
Context-Aware Visual Policy network (CAVP) for fine-grained image-to-language
generation: image sentence captioning and image paragraph captioning. During
captioning, CAVP explicitly considers the previous visual attentions as
context, and decides whether the context is used for the current word/sentence
generation given the current visual attention. Compared against traditional
visual attention mechanism that only fixes a single visual region at each step,
CAVP can attend to complex visual compositions over time. The whole image
captioning model -- CAVP and its subsequent language policy network -- can be
efficiently optimized end-to-end by using an actor-critic policy gradient
method. We have demonstrated the effectiveness of CAVP by state-of-the-art
performances on MS-COCO and Stanford captioning datasets, using various metrics
and sensible visualizations of qualitative visual context. | [
"cs.CV"
] |
Gastric cancer is the second leading cause of cancer-related deaths
worldwide, and the major hurdle in biomedical image analysis is the
determination of the cancer extent. This assignment has high clinical relevance
and would generally require vast microscopic assessment by pathologists. Recent
advances in deep learning have produced inspiring results on biomedical image
segmentation, while its outcome is reliant on comprehensive annotation. This
requires plenty of labor costs, for the ground truth must be annotated
meticulously by pathologists. In this paper, a reiterative learning framework
was presented to train our network on partial annotated biomedical images, and
superior performance was achieved without any pre-trained or further manual
annotation. We eliminate the boundary error of patch-based model through our
overlapped region forecast algorithm. Through these advisable methods, a mean
intersection over union coefficient (IOU) of 0.883 and mean accuracy of 91.09%
on the partial labeled dataset was achieved, which made us win the 2017 China
Big Data & Artificial Intelligence Innovation and Entrepreneurship
Competitions. | [
"cs.CV"
] |
Recent advances in representation learning on graphs, mainly leveraging graph
convolutional networks, have brought a substantial improvement on many
graph-based benchmark tasks. While novel approaches to learning node embeddings
are highly suitable for node classification and link prediction, their
application to graph classification (predicting a single label for the entire
graph) remains mostly rudimentary, typically using a single global pooling step
to aggregate node features or a hand-designed, fixed heuristic for hierarchical
coarsening of the graph structure. An important step towards ameliorating this
is differentiable graph coarsening---the ability to reduce the size of the
graph in an adaptive, data-dependent manner within a graph neural network
pipeline, analogous to image downsampling within CNNs. However, the previous
prominent approach to pooling has quadratic memory requirements during training
and is therefore not scalable to large graphs. Here we combine several recent
advances in graph neural network design to demonstrate that competitive
hierarchical graph classification results are possible without sacrificing
sparsity. Our results are verified on several established graph classification
benchmarks, and highlight an important direction for future research in
graph-based neural networks. | [
"stat.ML",
"cs.AI",
"cs.LG",
"cs.SI"
] |
In this paper, we propose a very deep fully convolutional encoding-decoding
framework for image restoration such as denoising and super-resolution. The
network is composed of multiple layers of convolution and de-convolution
operators, learning end-to-end mappings from corrupted images to the original
ones. The convolutional layers act as the feature extractor, which capture the
abstraction of image contents while eliminating noises/corruptions.
De-convolutional layers are then used to recover the image details. We propose
to symmetrically link convolutional and de-convolutional layers with skip-layer
connections, with which the training converges much faster and attains a
higher-quality local optimum. First, The skip connections allow the signal to
be back-propagated to bottom layers directly, and thus tackles the problem of
gradient vanishing, making training deep networks easier and achieving
restoration performance gains consequently. Second, these skip connections pass
image details from convolutional layers to de-convolutional layers, which is
beneficial in recovering the original image. Significantly, with the large
capacity, we can handle different levels of noises using a single model.
Experimental results show that our network achieves better performance than all
previously reported state-of-the-art methods. | [
"cs.CV"
] |
Data mining methods have been widely applied in financial markets, with the
purpose of providing suitable tools for prices forecasting and automatic
trading. Particularly, learning methods aim to identify patterns in time series
and, based on such patterns, to recommend buy/sell operations. The objective of
this work is to evaluate the performance of Random Forests, a supervised
learning method based on ensembles of decision trees, for decision support in
stock markets. Preliminary results indicate good rates of successful operations
and good rates of return per operation, providing a strong motivation for
further research in this topic. | [
"stat.ML",
"cs.LG",
"stat.AP"
] |
This paper proposes a novel joint learning and densely-cooperative fusion
(JL-DCF) architecture for RGB-D salient object detection. Existing models
usually treat RGB and depth as independent information and design separate
networks for feature extraction from each. Such schemes can easily be
constrained by a limited amount of training data or over-reliance on an
elaborately-designed training process. In contrast, our JL-DCF learns from both
RGB and depth inputs through a Siamese network. To this end, we propose two
effective components: joint learning (JL), and densely-cooperative fusion
(DCF). The JL module provides robust saliency feature learning, while the
latter is introduced for complementary feature discovery. Comprehensive
experiments on four popular metrics show that the designed framework yields a
robust RGB-D saliency detector with good generalization. As a result, JL-DCF
significantly advances the top-1 D3Net model by an average of ~1.9% (S-measure)
across six challenging datasets, showing that the proposed framework offers a
potential solution for real-world applications and could provide more insight
into the cross-modality complementarity task. The code will be available at
https://github.com/kerenfu/JLDCF/. | [
"cs.CV"
] |
Traditional convolution-based generative adversarial networks synthesize
images based on hierarchical local operations, where long-range dependency
relation is implicitly modeled with a Markov chain. It is still not sufficient
for categories with complicated structures. In this paper, we characterize
long-range dependence with attentive normalization (AN), which is an extension
to traditional instance normalization. Specifically, the input feature map is
softly divided into several regions based on its internal semantic similarity,
which are respectively normalized. It enhances consistency between distant
regions with semantic correspondence. Compared with self-attention GAN, our
attentive normalization does not need to measure the correlation of all
locations, and thus can be directly applied to large-size feature maps without
much computational burden. Extensive experiments on class-conditional image
generation and semantic inpainting verify the efficacy of our proposed module. | [
"cs.CV"
] |
Continual learning models allow to learn and adapt to new changes and tasks
over time. However, in continual and sequential learning scenarios in which the
models are trained using different data with various distributions, neural
networks tend to forget the previously learned knowledge. This phenomenon is
often referred to as catastrophic forgetting. The catastrophic forgetting is an
inevitable problem in continual learning models for dynamic environments. To
address this issue, we propose a method, called Continual Bayesian Learning
Networks (CBLN), which enables the networks to allocate additional resources to
adapt to new tasks without forgetting the previously learned tasks. Using a
Bayesian Neural Network, CBLN maintains a mixture of Gaussian posterior
distributions that are associated with different tasks. The proposed method
tries to optimise the number of resources that are needed to learn each task
and avoids an exponential increase in the number of resources that are involved
in learning multiple tasks. The proposed method does not need to access the
past training data and can choose suitable weights to classify the data points
during the test time automatically based on an uncertainty criterion. We have
evaluated our method on the MNIST and UCR time-series datasets. The evaluation
results show that our method can address the catastrophic forgetting problem at
a promising rate compared to the state-of-the-art models. | [
"cs.LG",
"cs.NE",
"stat.ML"
] |
We propose a novel lightweight generative adversarial network for efficient
image manipulation using natural language descriptions. To achieve this, a new
word-level discriminator is proposed, which provides the generator with
fine-grained training feedback at word-level, to facilitate training a
lightweight generator that has a small number of parameters, but can still
correctly focus on specific visual attributes of an image, and then edit them
without affecting other contents that are not described in the text.
Furthermore, thanks to the explicit training signal related to each word, the
discriminator can also be simplified to have a lightweight structure. Compared
with the state of the art, our method has a much smaller number of parameters,
but still achieves a competitive manipulation performance. Extensive
experimental results demonstrate that our method can better disentangle
different visual attributes, then correctly map them to corresponding semantic
words, and thus achieve a more accurate image modification using natural
language descriptions. | [
"cs.CV",
"cs.CL",
"cs.LG"
] |
We propose a method for the approximation of high- or even
infinite-dimensional feature vectors, which play an important role in
supervised learning. The goal is to reduce the size of the training data,
resulting in lower storage consumption and computational complexity.
Furthermore, the method can be regarded as a regularization technique, which
improves the generalizability of learned target functions. We demonstrate
significant improvements in comparison to the computation of data-driven
predictions involving the full training data set. The method is applied to
classification and regression problems from different application areas such as
image recognition, system identification, and oceanographic time series
analysis. | [
"stat.ML",
"cs.LG"
] |
The Social Internet of Things (SIoT), integration of the Internet of Things
and Social Networks paradigms, has been introduced to build a network of smart
nodes that are capable of establishing social links. In order to deal with
misbehaving service provider nodes, service requestor nodes must evaluate their
trustworthiness levels. In this paper, we propose a novel trust management
mechanism in the SIoT to predict the most reliable service providers for each
service requestor, which leads to reduce the risk of being exposed to malicious
nodes. We model the SIoT with a flexible bipartite graph (containing two sets
of nodes: service providers and service requestors), then build a social
network among the service requestor nodes, using the Hellinger distance.
Afterward, we develop a social trust model using nodes' centrality and
similarity measures to extract trust behaviors among the social network nodes.
Finally, a matrix factorization technique is designed to extract latent
features of SIoT nodes, find trustworthy nodes, and mitigate the data sparsity
and cold start problems. We analyze the effect of parameters in the proposed
trust prediction mechanism on prediction accuracy. The results indicate that
feedbacks from the neighboring nodes of a specific service requestor with high
Hellinger similarity in our mechanism outperforms the best existing methods. We
also show that utilizing the social trust model, which only considers a
similarity measure, significantly improves the accuracy of the prediction
mechanism. Furthermore, we evaluate the effectiveness of the proposed trust
management system through a real-world SIoT use case. Our results demonstrate
that the proposed mechanism is resilient to different types of network attacks,
and it can accurately find the most proper and trustworthy service provider. | [
"cs.LG",
"cs.CR",
"cs.SI",
"stat.ML"
] |
Biomedical research papers use significantly different language and jargon
when compared to typical English text, which reduces the utility of pre-trained
NLP models in this domain. Meanwhile Medline, a database of biomedical
abstracts, introduces nearly a million new documents per-year. Applications
that could benefit from understanding this wealth of publicly available
information, such as scientific writing assistants, chat-bots, or descriptive
hypothesis generation systems, require new domain-centered approaches. A
conditional language model, one that learns the probability of words given some
a priori criteria, is a fundamental building block in many such applications.
We propose a transformer-based conditional language model with a shallow
encoder "condition" stack, and a deep "language model" stack of multi-headed
attention blocks. The condition stack encodes metadata used to alter the output
probability distribution of the language model stack. We sample this
distribution in order to generate biomedical abstracts given only a proposed
title, an intended publication year, and a set of keywords. Using typical
natural language generation metrics, we demonstrate that this proposed approach
is more capable of producing non-trivial relevant entities within the abstract
body than the 1.5B parameter GPT-2 language model. | [
"cs.LG",
"stat.ML"
] |
We present a scalable end-to-end classifier that uses streaming physiological
and medication data to accurately predict the onset of sepsis, a
life-threatening complication from infections that has high mortality and
morbidity. Our proposed framework models the multivariate trajectories of
continuous-valued physiological time series using multitask Gaussian processes,
seamlessly accounting for the high uncertainty, frequent missingness, and
irregular sampling rates typically associated with real clinical data. The
Gaussian process is directly connected to a black-box classifier that predicts
whether a patient will become septic, chosen in our case to be a recurrent
neural network to account for the extreme variability in the length of patient
encounters. We show how to scale the computations associated with the Gaussian
process in a manner so that the entire system can be discriminatively trained
end-to-end using backpropagation. In a large cohort of heterogeneous inpatient
encounters at our university health system we find that it outperforms several
baselines at predicting sepsis, and yields 19.4% and 55.5% improved areas under
the Receiver Operating Characteristic and Precision Recall curves as compared
to the NEWS score currently used by our hospital. | [
"stat.ML",
"stat.AP",
"stat.ME"
] |
Neural controlled differential equations (CDEs) are the continuous-time
analogue of recurrent neural networks, as Neural ODEs are to residual networks,
and offer a memory-efficient continuous-time way to model functions of
potentially irregular time series. Existing methods for computing the forward
pass of a Neural CDE involve embedding the incoming time series into path
space, often via interpolation, and using evaluations of this path to drive the
hidden state. Here, we use rough path theory to extend this formulation.
Instead of directly embedding into path space, we instead represent the input
signal over small time intervals through its \textit{log-signature}, which are
statistics describing how the signal drives a CDE. This is the approach for
solving \textit{rough differential equations} (RDEs), and correspondingly we
describe our main contribution as the introduction of Neural RDEs. This
extension has a purpose: by generalising the Neural CDE approach to a broader
class of driving signals, we demonstrate particular advantages for tackling
long time series. In this regime, we demonstrate efficacy on problems of length
up to 17k observations and observe significant training speed-ups, improvements
in model performance, and reduced memory requirements compared to existing
approaches. | [
"cs.LG",
"cs.AI",
"math.DS",
"stat.ML"
] |
Generative adversarial network (GAN) is among the most popular deep learning
models for learning complex data distributions. However, training a GAN is
known to be a challenging task. This is often attributed to the lack of
correlation between the training progress and the trajectory of the generator
and discriminator losses and the need for the GAN's subjective evaluation. A
recently proposed measure inspired by game theory - the duality gap, aims to
bridge this gap. However, as we demonstrate, the duality gap's capability
remains constrained due to limitations posed by its estimation process. This
paper presents a theoretical understanding of this limitation and proposes a
more dependable estimation process for the duality gap. At the crux of our
approach is the idea that local perturbations can help agents in a zero-sum
game escape non-Nash saddle points efficiently. Through exhaustive
experimentation across GAN models and datasets, we establish the efficacy of
our approach in capturing the GAN training progress with minimal increase to
the computational complexity. Further, we show that our estimate, with its
ability to identify model convergence/divergence, is a potential performance
measure that can be used to tune the hyperparameters of a GAN. | [
"cs.LG",
"cs.AI"
] |
Approximating distributions over complicated manifolds, such as natural
images, are conceptually attractive. The deep latent variable model, trained
using variational autoencoders and generative adversarial networks, is now a
key technique for representation learning. However, it is difficult to unify
these two models for exact latent-variable inference and parallelize both
reconstruction and sampling, partly due to the regularization under the latent
variables, to match a simple explicit prior distribution. These approaches are
prone to be oversimplified, and can only characterize a few modes of the true
distribution. Based on the recently proposed Wasserstein autoencoder (WAE) with
a new regularization as an optimal transport. The paper proposes a stacked
Wasserstein autoencoder (SWAE) to learn a deep latent variable model. SWAE is a
hierarchical model, which relaxes the optimal transport constraints at two
stages. At the first stage, the SWAE flexibly learns a representation
distribution, i.e., the encoded prior; and at the second stage, the encoded
representation distribution is approximated with a latent variable model under
the regularization encouraging the latent distribution to match the explicit
prior. This model allows us to generate natural textual outputs as well as
perform manipulations in the latent space to induce changes in the output
space. Both quantitative and qualitative results demonstrate the superior
performance of SWAE compared with the state-of-the-art approaches in terms of
faithful reconstruction and generation quality. | [
"cs.CV",
"cs.LG"
] |
Extracting detailed 3D information of objects from video data is an important
goal for holistic scene understanding. While recent methods have shown
impressive results when reconstructing meshes of objects from a single image,
results often remain ambiguous as part of the object is unobserved. Moreover,
existing image-based datasets for mesh reconstruction don't permit to study
models which integrate temporal information. To alleviate both concerns we
present SAIL-VOS 3D: a synthetic video dataset with frame-by-frame mesh
annotations which extends SAIL-VOS. We also develop first baselines for
reconstruction of 3D meshes from video data via temporal models. We demonstrate
efficacy of the proposed baseline on SAIL-VOS 3D and Pix3D, showing that
temporal information improves reconstruction quality. Resources and additional
information are available at http://sailvos.web.illinois.edu. | [
"cs.CV",
"cs.GR",
"cs.LG"
] |
Machine learning plays an increasing role in intelligent tutoring systems as
both the amount of data available and specialization among students grow.
Nowadays, these systems are frequently deployed on mobile applications. Users
on such mobile education platforms are dynamic, frequently being added,
accessing the application with varying levels of focus, and changing while
using the service. The education material itself, on the other hand, is often
static and is an exhaustible resource whose use in tasks such as problem
recommendation must be optimized. The ability to update user models with
respect to educational material in real-time is thus essential; however,
existing approaches require time-consuming re-training of user features
whenever new data is added. In this paper, we introduce a neural pedagogical
agent for real-time user modeling in the task of predicting user response
correctness, a central task for mobile education applications. Our model,
inspired by work in natural language processing on sequence modeling and
machine translation, updates user features in real-time via bidirectional
recurrent neural networks with an attention mechanism over embedded
question-response pairs. We experiment on the mobile education application
SantaTOEIC, which has 559k users, 66M response data points as well as a set of
10k study problems each expert-annotated with topic tags and gathered since
2016. Our model outperforms existing approaches over several metrics in
predicting user response correctness, notably out-performing other methods on
new users without large question-response histories. Additionally, our
attention mechanism and annotated tag set allow us to create an interpretable
education platform, with a smart review system that addresses the
aforementioned issue of varied user attention and problem exhaustion. | [
"cs.LG",
"cs.CL",
"stat.ML"
] |
Graph Neural Networks have recently become a prevailing paradigm for various
high-impact graph learning tasks. Existing efforts can be mainly categorized as
spectral-based and spatial-based methods. The major challenge for the former is
to find an appropriate graph filter to distill discriminative information from
input signals for learning. Recently, attempts such as Graph Convolutional
Network (GCN) leverage Chebyshev polynomial truncation to seek an approximation
of graph filters and bridge these two families of methods. It has been shown in
recent studies that GCN and its variants are essentially employing fixed
low-pass filters to perform information denoising. Thus their learning
capability is rather limited and may over-smooth node representations at deeper
layers. To tackle these problems, we develop a novel graph neural network
framework AdaGNN with a well-designed adaptive frequency response filter. At
its core, AdaGNN leverages a simple but elegant trainable filter that spans
across multiple layers to capture the varying importance of different frequency
components for node representation learning. The inherent differences among
different feature channels are also well captured by the filter. As such, it
empowers AdaGNN with stronger expressiveness and naturally alleviates the
over-smoothing problem. We empirically validate the effectiveness of the
proposed framework on various benchmark datasets. Theoretical analysis is also
provided to show the superiority of the proposed AdaGNN. The implementation of
AdaGNN is available at \url{https://github.com/yushundong/AdaGNN}. | [
"cs.LG"
] |
The impact of non verbal behaviour in a hiring decision remains an open
question. Investigating this question is important, as it could provide a
better understanding on how to train candidates for job interviews and make
recruiters be aware of influential non verbal behaviour. This research has
recently been accelerated due to the development of tools for the automatic
analysis of social signals, and the emergence of machine learning methods.
However, these studies are still mainly based on hand engineered features,
which imposes a limit to the discovery of influential social signals. On the
other side, deep learning methods are a promising tool to discover complex
patterns without the necessity of feature engineering. In this paper, we focus
on studying influential non verbal social signals in asynchronous job video
interviews that are discovered by deep learning methods. We use a previously
published deep learning system that aims at inferring the hirability of a
candidate with regard to a sequence of interview questions. One particularity
of this system is the use of attention mechanisms, which aim at identifying the
relevant parts of an answer. Thus, information at a fine-grained temporal level
could be extracted using global (at the interview level) annotations on
hirability. While most of the deep learning systems use attention mechanisms to
offer a quick visualization of slices when a rise of attention occurs, we
perform an in-depth analysis to understand what happens during these moments.
First, we propose a methodology to automatically extract slices where there is
a rise of attention (attention slices). Second, we study the content of
attention slices by comparing them with randomly sampled slices. Finally, we
show that they bear significantly more information for hirability than randomly
sampled slices. | [
"cs.CV",
"cs.HC"
] |
Neural networks are powerful models that have a remarkable ability to extract
patterns that are too complex to be noticed by humans or other machine learning
models. Neural networks are the first class of models that can train end-to-end
systems with large learning capacities. However, we still have the difficult
challenge of designing the neural network, which requires human experience and
a long process of trial and error. As a solution, we can use a neural
architecture search to find the best network architecture for the task at hand.
Existing NAS algorithms generally evaluate the fitness of a new architecture by
fully training from scratch, resulting in the prohibitive computational cost,
even if operated on high-performance computers. In this paper, an end-to-end
offline performance predictor is proposed to accelerate the evaluation of
sampled architectures.
Index Terms- Learning Curve Prediction, Neural Architecture Search,
Reinforcement Learning. | [
"cs.LG"
] |
Deep Reinforcement Learning (DRL) has achieved impressive success in many
applications. A key component of many DRL models is a neural network
representing a Q function, to estimate the expected cumulative reward following
a state-action pair. The Q function neural network contains a lot of implicit
knowledge about the RL problems, but often remains unexamined and
uninterpreted. To our knowledge, this work develops the first mimic learning
framework for Q functions in DRL. We introduce Linear Model U-trees (LMUTs) to
approximate neural network predictions. An LMUT is learned using a novel
on-line algorithm that is well-suited for an active play setting, where the
mimic learner observes an ongoing interaction between the neural net and the
environment. Empirical evaluation shows that an LMUT mimics a Q function
substantially better than five baseline methods. The transparent tree structure
of an LMUT facilitates understanding the network's learned knowledge by
analyzing feature influence, extracting rules, and highlighting the
super-pixels in image inputs. | [
"cs.LG",
"stat.ML"
] |
Colorization is the method of converting an image in grayscale to a fully
color image. There are multiple methods to do the same. Old school methods used
machine learning algorithms and optimization techniques to suggest possible
colors to use. With advances in the field of deep learning, colorization
results have improved consistently with improvements in deep learning
architectures. The latest development in the field of deep learning is the
emergence of generative adversarial networks (GANs) which is used to generate
information and not just predict or classify. As part of this report, 2
architectures of recent papers are reproduced along with a novel architecture
being suggested for general colorization. Following this, we propose the use of
colorization by generating makeup suggestions automatically on a face. To do
this, a dataset consisting of 1000 images has been created. When an image of a
person without makeup is sent to the model, the model first converts the image
to grayscale and then passes it through the suggested GAN model. The output is
a generated makeup suggestion. To develop this model, we need to tweak the
general colorization model to deal only with faces of people. | [
"cs.CV"
] |
Graph neural networks have become an important tool for modeling structured
data. In many real-world systems, intricate hidden information may exist, e.g.,
heterogeneity in nodes/edges, static node/edge attributes, and spatiotemporal
node/edge features. However, most existing methods only take part of the
information into consideration. In this paper, we present the Co-evolved Meta
Graph Neural Network (CoMGNN), which applies meta graph attention to
heterogeneous graphs with co-evolution of node and edge states. We further
propose a spatiotemporal adaption of CoMGNN (ST-CoMGNN) for modeling
spatiotemporal patterns on nodes and edges. We conduct experiments on two
large-scale real-world datasets. Experimental results show that our models
significantly outperform the state-of-the-art methods, demonstrating the
effectiveness of encoding diverse information from different aspects. | [
"cs.LG",
"cs.SI",
"stat.ML"
] |
In this paper, we propose a novel method for projecting data from multiple
modalities to a new subspace optimized for one-class classification. The
proposed method iteratively transforms the data from the original feature space
of each modality to a new common feature space along with finding a joint
compact description of data coming from all the modalities. For data in each
modality, we define a separate transformation to map the data from the
corresponding feature space to the new optimized subspace by exploiting the
available information from the class of interest only. We also propose
different regularization strategies for the proposed method and provide both
linear and non-linear formulations. The proposed Multimodal Subspace Support
Vector Data Description outperforms all the competing methods using data from a
single modality or fusing data from all modalities in four out of five
datasets. | [
"cs.LG",
"stat.ML"
] |
Building neural networks to query a knowledge base (a table) with natural
language is an emerging research topic in deep learning. An executor for table
querying typically requires multiple steps of execution because queries may
have complicated structures. In previous studies, researchers have developed
either fully distributed executors or symbolic executors for table querying. A
distributed executor can be trained in an end-to-end fashion, but is weak in
terms of execution efficiency and explicit interpretability. A symbolic
executor is efficient in execution, but is very difficult to train especially
at initial stages. In this paper, we propose to couple distributed and symbolic
execution for natural language queries, where the symbolic executor is
pretrained with the distributed executor's intermediate execution results in a
step-by-step fashion. Experiments show that our approach significantly
outperforms both distributed and symbolic executors, exhibiting high accuracy,
high learning efficiency, high execution efficiency, and high interpretability. | [
"cs.LG",
"cs.AI",
"cs.CL",
"cs.NE",
"cs.SE"
] |
Co-saliency detection aims to detect common salient objects from a group of
relevant images. Some attempts have been made with the Fully Convolutional
Network (FCN) framework and achieve satisfactory detection results. However,
due to stacking convolution layers and pooling operation, the boundary details
tend to be lost. In addition, existing models often utilize the extracted
features without discrimination, leading to redundancy in representation since
actually not all features are helpful to the final prediction and some even
bring distraction. In this paper, we propose a co-attention module embedded FCN
framework, called as Co-Attention FCN (CA-FCN). Specifically, the co-attention
module is plugged into the high-level convolution layers of FCN, which can
assign larger attention weights on the common salient objects and smaller ones
on the background and uncommon distractors to boost final detection
performance. Extensive experiments on three popular co-saliency benchmark
datasets demonstrate the superiority of the proposed CA-FCN, which outperforms
state-of-the-arts in most cases. Besides, the effectiveness of our new
co-attention module is also validated with ablation studies. | [
"cs.CV"
] |
Machine learning techniques for road networks hold the potential to
facilitate many important transportation applications. Graph Convolutional
Networks (GCNs) are neural networks that are capable of leveraging the
structure of a road network by utilizing information of, e.g., adjacent road
segments. While state-of-the-art GCNs target node classification tasks in
social, citation, and biological networks, machine learning tasks in road
networks differ substantially from such tasks. In road networks, prediction
tasks concern edges representing road segments, and many tasks involve
regression. In addition, road networks differ substantially from the networks
assumed in the GCN literature in terms of the attribute information available
and the network characteristics. Many implicit assumptions of GCNs do therefore
not apply. We introduce the notion of Relational Fusion Network (RFN), a novel
type of GCN designed specifically for machine learning on road networks. In
particular, we propose methods that outperform state-of-the-art GCNs on both a
road segment regression task and a road segment classification task by 32-40%
and 21-24%, respectively. In addition, we provide experimental evidence of the
short-comings of state-of-the-art GCNs in the context of road networks: unlike
our method, they cannot effectively leverage the road network structure for
road segment classification and fail to outperform a regular multi-layer
perceptron. | [
"cs.LG",
"cs.DB",
"stat.ML"
] |
Low attendance levels in medical appointments have been associated with poor
health outcomes and efficiency problems for service providers. To address this
problem, healthcare managers could aim at improving attendance levels or
minimizing the operational impact of no-shows by adapting resource allocation
policies. However, given the uncertainty of patient behaviour, generating
relevant information regarding no-show probabilities could support the
decision-making process for both approaches. In this context many researchers
have used multiple regression models to identify patient and appointment
characteristics than can be used as good predictors for no-show probabilities.
This work develops a Decision Support System (DSS) to support the
implementation of strategies to encourage attendance, for a preventive care
program targeted at underserved communities in Bogot\'a, Colombia. Our
contribution to literature is threefold. Firstly, we assess the effectiveness
of different machine learning approaches to improve the accuracy of regression
models. In particular, Random Forest and Neural Networks are used to model the
problem accounting for non-linearity and variable interactions. Secondly, we
propose a novel use of Layer-wise Relevance Propagation in order to improve the
explainability of neural network predictions and obtain insights from the
modelling step. Thirdly, we identify variables explaining no-show probabilities
in a developing context and study its policy implications and potential for
improving healthcare access. In addition to quantifying relationships reported
in previous studies, we find that income and neighbourhood crime statistics
affect no-show probabilities. Our results will support patient prioritization
in a pilot behavioural intervention and will inform appointment planning
decisions. | [
"cs.LG",
"stat.AP"
] |
In this paper a hierarchical model for pixel clustering and image
segmentation is developed. In the model an image is hierarchically structured.
The original image is treated as a set of nested images, which are capable to
reversibly merge with each other. An object is defined as a structural element
of an image, so that, an image is regarded as a maximal object. The simulating
of none-hierarchical optimal pixel clustering by hierarchical clustering is
studied. To generate a hierarchy of optimized piecewise constant image
approximations, estimated by the standard deviation of approximation from the
image, the conversion of any hierarchy of approximations into the hierarchy
described in relation to the number of intensity levels by convex sequence of
total squared errors is proposed. | [
"cs.CV"
] |
Machine learning pipelines often rely on optimization procedures to make
discrete decisions (e.g., sorting, picking closest neighbors, or shortest
paths). Although these discrete decisions are easily computed, they break the
back-propagation of computational graphs. In order to expand the scope of
learning problems that can be solved in an end-to-end fashion, we propose a
systematic method to transform optimizers into operations that are
differentiable and never locally constant. Our approach relies on
stochastically perturbed optimizers, and can be used readily together with
existing solvers. Their derivatives can be evaluated efficiently, and
smoothness tuned via the chosen noise amplitude. We also show how this
framework can be connected to a family of losses developed in structured
prediction, and give theoretical guarantees for their use in learning tasks. We
demonstrate experimentally the performance of our approach on various tasks. | [
"cs.LG",
"math.OC",
"stat.ML"
] |
Reinforcement learning was carried out in a simulated environment to learn
continuous velocity control over multiple motor axes. This was then applied to
a real-world optical tweezers experiment with the objective of moving a
laser-trapped microsphere to a target location whilst avoiding collisions with
other free-moving microspheres. The concept of training a neural network in a
virtual environment has significant potential in the application of machine
learning for experimental optimization and control, as the neural network can
discover optimal methods for problem solving without the risk of damage to
equipment, and at a speed not limited by movement in the physical environment.
As the neural network treats both virtual and physical environments
equivalently, we show that the network can also be applied to an augmented
environment, where a virtual environment is combined with the physical
environment. This technique may have the potential to unlock capabilities
associated with mixed and augmented reality, such as enforcing safety limits
for machine motion or as a method of inputting observations from additional
sensors. | [
"cs.LG",
"cs.AI",
"cs.RO",
"physics.optics"
] |
Recent research advances in Computer Vision and Natural Language Processing
have introduced novel tasks that are paving the way for solving AI-complete
problems. One of those tasks is called Visual Question Answering (VQA). A VQA
system must take an image and a free-form, open-ended natural language question
about the image, and produce a natural language answer as the output. Such a
task has drawn great attention from the scientific community, which generated a
plethora of approaches that aim to improve the VQA predictive accuracy. Most of
them comprise three major components: (i) independent representation learning
of images and questions; (ii) feature fusion so the model can use information
from both sources to answer visual questions; and (iii) the generation of the
correct answer in natural language. With so many approaches being recently
introduced, it became unclear the real contribution of each component for the
ultimate performance of the model. The main goal of this paper is to provide a
comprehensive analysis regarding the impact of each component in VQA models.
Our extensive set of experiments cover both visual and textual elements, as
well as the combination of these representations in form of fusion and
attention mechanisms. Our major contribution is to identify core components for
training VQA models so as to maximize their predictive performance. | [
"cs.CV",
"cs.CL",
"cs.LG"
] |
Deep neural networks have recently thrived on single image depth estimation.
That being said, current developments on this topic highlight an apparent
compromise between accuracy and network size. This work proposes an accurate
and lightweight framework for monocular depth estimation based on a
self-attention mechanism stemming from salient point detection. Specifically,
we utilize a sparse set of keypoints to train a FuSaNet model that consists of
two major components: Fusion-Net and Saliency-Net. In addition, we introduce a
normalized Hessian loss term invariant to scaling and shear along the depth
direction, which is shown to substantially improve the accuracy. The proposed
method achieves state-of-the-art results on NYU-Depth-v2 and KITTI while using
3.1-38.4 times smaller model in terms of the number of parameters than baseline
approaches. Experiments on the SUN-RGBD further demonstrate the
generalizability of the proposed method. | [
"cs.CV"
] |
We consider interpolation learning in high-dimensional linear regression with
Gaussian data, and prove a generic uniform convergence guarantee on the
generalization error of interpolators in an arbitrary hypothesis class in terms
of the class's Gaussian width. Applying the generic bound to Euclidean norm
balls recovers the consistency result of Bartlett et al. (2020) for
minimum-norm interpolators, and confirms a prediction of Zhou et al. (2020) for
near-minimal-norm interpolators in the special case of Gaussian data. We
demonstrate the generality of the bound by applying it to the simplex,
obtaining a novel consistency result for minimum l1-norm interpolators (basis
pursuit). Our results show how norm-based generalization bounds can explain and
be used to analyze benign overfitting, at least in some settings. | [
"stat.ML",
"cs.LG",
"math.ST",
"stat.TH"
] |
We consider matrix completion for recommender systems from the point of view
of link prediction on graphs. Interaction data such as movie ratings can be
represented by a bipartite user-item graph with labeled edges denoting observed
ratings. Building on recent progress in deep learning on graph-structured data,
we propose a graph auto-encoder framework based on differentiable message
passing on the bipartite interaction graph. Our model shows competitive
performance on standard collaborative filtering benchmarks. In settings where
complimentary feature information or structured data such as a social network
is available, our framework outperforms recent state-of-the-art methods. | [
"stat.ML",
"cs.DB",
"cs.IR",
"cs.LG"
] |
In this work, we first describe a framework for the application of
Reinforcement Learning (RL) control to a radar system that operates in a
congested spectral setting. We then compare the utility of several RL
algorithms through a discussion of experiments performed on Commercial
off-the-shelf (COTS) hardware. Each RL technique is evaluated in terms of
convergence, radar detection performance achieved in a congested spectral
environment, and the ability to share 100MHz spectrum with an uncooperative
communications system. We examine policy iteration, which solves an environment
posed as a Markov Decision Process (MDP) by directly solving for a stochastic
mapping between environmental states and radar waveforms, as well as Deep RL
techniques, which utilize a form of Q-Learning to approximate a parameterized
function that is used by the radar to select optimal actions. We show that RL
techniques are beneficial over a Sense-and-Avoid (SAA) scheme and discuss the
conditions under which each approach is most effective. | [
"cs.LG",
"eess.SP",
"stat.ML"
] |
Rapid pace of generative models has brought about new threats to visual
forensics such as malicious personation and digital copyright infringement,
which promotes works on fake image attribution. Existing works on fake image
attribution mainly rely on a direct classification framework. Without
additional supervision, the extracted features could include many
content-relevant components and generalize poorly. Meanwhile, how to obtain an
interpretable GAN fingerprint to explain the decision remains an open question.
Adopting a multi-task framework, we propose a GAN Fingerprint Disentangling
Network (GFD-Net) to simultaneously disentangle the fingerprint from
GAN-generated images and produce a content-irrelevant representation for fake
image attribution. A series of constraints are provided to guarantee the
stability and discriminability of the fingerprint, which in turn helps
content-irrelevant feature extraction. Further, we perform comprehensive
analysis on GAN fingerprint, providing some clues about the properties of GAN
fingerprint and which factors dominate the fingerprint in GAN architecture.
Experiments show that our GFD-Net achieves superior fake image attribution
performance in both closed-world and open-world testing. We also apply our
method in binary fake image detection and exhibit a significant generalization
ability on unseen generators. | [
"cs.CV"
] |
Appearance-based gaze estimation has achieved significant improvement by
using deep learning. However, many deep learning-based methods suffer from the
vulnerability property, i.e., perturbing the raw image using noise confuses the
gaze estimation models. Although the perturbed image visually looks similar to
the original image, the gaze estimation models output the wrong gaze direction.
In this paper, we investigate the vulnerability of appearance-based gaze
estimation. To our knowledge, this is the first time that the vulnerability of
gaze estimation to be found. We systematically characterized the vulnerability
property from multiple aspects, the pixel-based adversarial attack, the
patch-based adversarial attack and the defense strategy. Our experimental
results demonstrate that the CA-Net shows superior performance against attack
among the four popular appearance-based gaze estimation networks, Full-Face,
Gaze-Net, CA-Net and RT-GENE. This study draws the attention of researchers in
the appearance-based gaze estimation community to defense from adversarial
attacks. | [
"cs.CV"
] |
In unsupervised data generation tasks, besides the generation of a sample
based on previous observations, one would often like to give hints to the model
in order to bias the generation towards desirable metrics. We propose a method
that combines Generative Adversarial Networks (GANs) and reinforcement learning
(RL) in order to accomplish exactly that. While RL biases the data generation
process towards arbitrary metrics, the GAN component of the reward function
ensures that the model still remembers information learned from data. We build
upon previous results that incorporated GANs and RL in order to generate
sequence data and test this model in several settings for the generation of
molecules encoded as text sequences (SMILES) and in the context of music
generation, showing for each case that we can effectively bias the generation
process towards desired metrics. | [
"stat.ML",
"cs.LG"
] |
MEx: Multi-modal Exercises Dataset is a multi-sensor, multi-modal dataset,
implemented to benchmark Human Activity Recognition(HAR) and Multi-modal Fusion
algorithms. Collection of this dataset was inspired by the need for recognising
and evaluating quality of exercise performance to support patients with
Musculoskeletal Disorders(MSD). We select 7 exercises regularly recommended for
MSD patients by physiotherapists and collected data with four sensors a
pressure mat, a depth camera and two accelerometers. The dataset contains three
data modalities; numerical time-series data, video data and pressure sensor
data posing interesting research challenges when reasoning for HAR and Exercise
Quality Assessment. This paper presents our evaluation of the dataset on number
of standard classification algorithms for the HAR task by comparing different
feature representation algorithms for each sensor. These results set a
reference performance for each individual sensor that expose their strengths
and weaknesses for the future tasks. In addition we visualise pressure mat data
to explore the potential of the sensor to capture exercise performance quality.
With the recent advancement in multi-modal fusion, we also believe MEx is a
suitable dataset to benchmark not only HAR algorithms, but also fusion
algorithms of heterogeneous data types in multiple application domains. | [
"cs.CV",
"cs.AI",
"cs.LG",
"eess.SP",
"stat.ML"
] |
In this paper we revisit the method of off-policy corrections for
reinforcement learning (COP-TD) pioneered by Hallak et al. (2017). Under this
method, online updates to the value function are reweighted to avoid divergence
issues typical of off-policy learning. While Hallak et al.'s solution is
appealing, it cannot easily be transferred to nonlinear function approximation.
First, it requires a projection step onto the probability simplex; second, even
though the operator describing the expected behavior of the off-policy learning
algorithm is convergent, it is not known to be a contraction mapping, and
hence, may be more unstable in practice. We address these two issues by
introducing a discount factor into COP-TD. We analyze the behavior of
discounted COP-TD and find it better behaved from a theoretical perspective. We
also propose an alternative soft normalization penalty that can be minimized
online and obviates the need for an explicit projection step. We complement our
analysis with an empirical evaluation of the two techniques in an off-policy
setting on the game Pong from the Atari domain where we find discounted COP-TD
to be better behaved in practice than the soft normalization penalty. Finally,
we perform a more extensive evaluation of discounted COP-TD in 5 games of the
Atari domain, where we find performance gains for our approach. | [
"cs.LG",
"stat.ML"
] |
In this paper we propose a novel easily reproducible technique to attack the
best public Face ID system ArcFace in different shooting conditions. To create
an attack, we print the rectangular paper sticker on a common color printer and
put it on the hat. The adversarial sticker is prepared with a novel algorithm
for off-plane transformations of the image which imitates sticker location on
the hat. Such an approach confuses the state-of-the-art public Face ID model
LResNet100E-IR, ArcFace@ms1m-refine-v2 and is transferable to other Face ID
models. | [
"cs.CV"
] |
Many species have evolved advanced non-visual perception while artificial
systems fall behind. Radar and ultrasound complement camera-based vision but
they are often too costly and complex to set up for very limited information
gain. In nature, sound is used effectively by bats, dolphins, whales, and
humans for navigation and communication. However, it is unclear how to best
harness sound for machine perception. Inspired by bats' echolocation mechanism,
we design a low-cost BatVision system that is capable of seeing the 3D spatial
layout of space ahead by just listening with two ears. Our system emits short
chirps from a speaker and records returning echoes through microphones in an
artificial human pinnae pair. During training, we additionally use a stereo
camera to capture color images for calculating scene depths. We train a model
to predict depth maps and even grayscale images from the sound alone. During
testing, our trained BatVision provides surprisingly good predictions of 2D
visual scenes from two 1D audio signals. Such a sound to vision system would
benefit robot navigation and machine vision, especially in low-light or
no-light conditions. Our code and data are publicly available. | [
"cs.CV",
"cs.RO",
"cs.SD",
"eess.AS"
] |
We present a novel approach to the detection and characterization of edges,
ridges, and blobs in two-dimensional images which exploits the symmetry
properties of directionally sensitive analyzing functions in multiscale systems
that are constructed in the framework of alpha-molecules. The proposed feature
detectors are inspired by the notion of phase congruency, stable in the
presence of noise, and by definition invariant to changes in contrast. We also
show how the behavior of coefficients corresponding to differently scaled and
oriented analyzing functions can be used to obtain a comprehensive
characterization of the geometry of features in terms of local tangent
directions, widths, and heights. The accuracy and robustness of the proposed
measures are validated and compared to various state-of-the-art algorithms in
extensive numerical experiments in which we consider sets of clean and
distorted synthetic images that are associated with reliable ground truths. To
further demonstrate the applicability, we show how the proposed ridge measure
can be used to detect and characterize blood vessels in digital retinal images
and how the proposed blob measure can be applied to automatically count the
number of cell colonies in a Petri dish. | [
"cs.CV"
] |
Graph representation learning embeds nodes in large graphs as low-dimensional
vectors and is of great benefit to many downstream applications. Most embedding
frameworks, however, are inherently transductive and unable to generalize to
unseen nodes or learn representations across different graphs. Although
inductive approaches can generalize to unseen nodes, they neglect different
contexts of nodes and cannot learn node embeddings dually. In this paper, we
present a context-aware unsupervised dual encoding framework, \textbf{CADE}, to
generate representations of nodes by combining real-time neighborhoods with
neighbor-attentioned representation, and preserving extra memory of known
nodes. We exhibit that our approach is effective by comparing to
state-of-the-art methods. | [
"cs.LG",
"stat.ML"
] |
Tensor factorization based models have shown great power in knowledge graph
completion (KGC). However, their performance usually suffers from the
overfitting problem seriously. This motivates various regularizers -- such as
the squared Frobenius norm and tensor nuclear norm regularizers -- while the
limited applicability significantly limits their practical usage. To address
this challenge, we propose a novel regularizer -- namely, DUality-induced
RegulArizer (DURA) -- which is not only effective in improving the performance
of existing models but widely applicable to various methods. The major novelty
of DURA is based on the observation that, for an existing tensor factorization
based KGC model (primal), there is often another distance based KGC model
(dual) closely associated with it. Experiments show that DURA yields consistent
and significant improvements on benchmarks. | [
"cs.LG"
] |
Graph Neural Networks (GNNs) have become a promising approach to machine
learning with graphs. Since existing GNN models are based on flat
message-passing mechanisms, two limitations need to be tackled. One is costly
in encoding global information on the graph topology. The other is failing to
model meso- and macro-level semantics hidden in the graph, such as the
knowledge of institutes and research areas in an academic collaboration
network. To deal with these two issues, we propose a novel Hierarchical
Message-Passing Graph Neural Networks framework. The main idea is to generate a
hierarchical structure that re-organises all nodes in a graph into multi-level
clusters, along with intra- and inter-level edge connections. The derived
hierarchy not only creates shortcuts connecting far-away nodes so that global
information can be efficiently accessed via message passing but also
incorporates meso- and macro-level semantics into the learning of node
embedding. We present the first model to implement this hierarchical
message-passing mechanism, termed Hierarchical Community-aware Graph Neural
Network (HC-GNN), based on hierarchical communities detected from the graph.
Experiments conducted on eight datasets under transductive, inductive, and
few-shot settings exhibit that HC-GNN can outperform state-of-the-art GNN
models in network analysis tasks, including node classification, link
prediction, and community detection. | [
"cs.LG",
"stat.ML"
] |
Generative Adversarial Network(GAN) provides a good generative framework to
produce realistic samples, but suffers from two recognized issues as mode
collapse and unstable training. In this work, we propose to employ explicit
manifold learning as prior to alleviate mode collapse and stabilize training of
GAN. Since the basic assumption of conventional manifold learning fails in case
of sparse and uneven data distribution, we introduce a new target, Minimum
Manifold Coding (MMC), for manifold learning to encourage simple and unfolded
manifold. In essence, MMC is the general case of the shortest Hamiltonian Path
problem and pursues manifold with minimum Riemann volume. Using the
standardized code from MMC as prior, GAN is guaranteed to recover a simple and
unfolded manifold covering all the training data. Our experiments on both the
toy data and real datasets show the effectiveness of MMCGAN in alleviating mode
collapse, stabilizing training, and improving the quality of generated samples. | [
"cs.LG",
"stat.ML"
] |
Simultaneously modeling source code and natural language has many exciting
applications in automated software development and understanding. Pursuant to
achieving such technology, we introduce PyMT5, the Python method text-to-text
transfer transformer, which is trained to translate between all pairs of Python
method feature combinations: a single model that can both predict whole methods
from natural language documentation strings (docstrings) and summarize code
into docstrings of any common style. We present an analysis and modeling effort
of a large-scale parallel corpus of 26 million Python methods and 7.7 million
method-docstring pairs, demonstrating that for docstring and method generation,
PyMT5 outperforms similarly-sized auto-regressive language models (GPT2) which
were English pre-trained or randomly initialized. On the CodeSearchNet test
set, our best model predicts 92.1% syntactically correct method bodies,
achieved a BLEU score of 8.59 for method generation and 16.3 for docstring
generation (summarization), and achieved a ROUGE-L F-score of 24.8 for method
generation and 36.7 for docstring generation. | [
"cs.LG",
"cs.SE"
] |
This paper proposes to make a first step towards compatible and hence
reusable network components. Rather than training networks for different tasks
independently, we adapt the training process to produce network components that
are compatible across tasks. In particular, we split a network into two
components, a features extractor and a target task head, and propose various
approaches to accomplish compatibility between them. We systematically analyse
these approaches on the task of image classification on standard datasets. We
demonstrate that we can produce components which are directly compatible
without any fine-tuning or compromising accuracy on the original tasks.
Afterwards, we demonstrate the use of compatible components on three
applications: Unsupervised domain adaptation, transferring classifiers across
feature extractors with different architectures, and increasing the
computational efficiency of transfer learning. | [
"cs.LG",
"cs.CV",
"stat.ML"
] |