text
stringlengths 29
3.31k
| label
sequencelengths 1
11
|
---|---|
Adversarial training (AT) is one of the most effective ways for improving the
robustness of deep convolution neural networks (CNNs). Just like common network
training, the effectiveness of AT relies on the design of basic network
components. In this paper, we conduct an in-depth study on the role of the
basic ReLU activation component in AT for robust CNNs. We find that the
spatially-shared and input-independent properties of ReLU activation make CNNs
less robust to white-box adversarial attacks with either standard or
adversarial training. To address this problem, we extend ReLU to a novel Sparta
activation function (Spatially attentive and Adversarially Robust Activation),
which enables CNNs to achieve both higher robustness, i.e., lower error rate on
adversarial examples, and higher accuracy, i.e., lower error rate on clean
examples, than the existing state-of-the-art (SOTA) activation functions. We
further study the relationship between Sparta and the SOTA activation
functions, providing more insights about the advantages of our method. With
comprehensive experiments, we also find that the proposed method exhibits
superior cross-CNN and cross-dataset transferability. For the former, the
adversarially trained Sparta function for one CNN (e.g., ResNet-18) can be
fixed and directly used to train another adversarially robust CNN (e.g.,
ResNet-34). For the latter, the Sparta function trained on one dataset (e.g.,
CIFAR-10) can be employed to train adversarially robust CNNs on another dataset
(e.g., SVHN). In both cases, Sparta leads to CNNs with higher robustness than
the vanilla ReLU, verifying the flexibility and versatility of the proposed
method. | [
"cs.LG",
"cs.CV"
] |
Soccer is a sparse rewarding game: any smart or careless action in critical
situations can change the result of the match. Therefore players, coaches, and
scouts are all curious about the best action to be performed in critical
situations, such as the times with a high probability of losing ball possession
or scoring a goal. This work proposes a new state representation for the soccer
game and a batch reinforcement learning to train a smart policy network. This
network gets the contextual information of the situation and proposes the
optimal action to maximize the expected goal for the team. We performed
extensive numerical experiments on the soccer logs made by InStat for 104
European soccer matches. The results show that in all 104 games, the optimized
policy obtains higher rewards than its counterpart in the behavior policy.
Besides, our framework learns policies that are close to the expected behavior
in the real world. For instance, in the optimized policy, we observe that some
actions such as foul, or ball out can be sometimes more rewarding than a shot
in specific situations. | [
"cs.LG"
] |
The availability of large-scale facial databases, together with the
remarkable progresses of deep learning technologies, in particular Generative
Adversarial Networks (GANs), have led to the generation of extremely realistic
fake facial content, raising obvious concerns about the potential for misuse.
Such concerns have fostered the research on manipulation detection methods
that, contrary to humans, have already achieved astonishing results in various
scenarios. In this study, we focus on the synthesis of entire facial images,
which is a specific type of facial manipulation. The main contributions of this
study are four-fold: i) a novel strategy to remove GAN "fingerprints" from
synthetic fake images based on autoencoders is described, in order to spoof
facial manipulation detection systems while keeping the visual quality of the
resulting images; ii) an in-depth analysis of the recent literature in facial
manipulation detection; iii) a complete experimental assessment of this type of
facial manipulation, considering the state-of-the-art fake detection systems
(based on holistic deep networks, steganalysis, and local artifacts), remarking
how challenging is this task in unconstrained scenarios; and finally iv) we
announce a novel public database, named iFakeFaceDB, yielding from the
application of our proposed GAN-fingerprint Removal approach (GANprintR) to
already very realistic synthetic fake images.
The results obtained in our empirical evaluation show that additional efforts
are required to develop robust facial manipulation detection systems against
unseen conditions and spoof techniques, such as the one proposed in this study. | [
"cs.CV"
] |
New generation geostationary satellites make solar reflectance observations
available at a continental scale with unprecedented spatiotemporal resolution
and spectral range. Generating quality land monitoring products requires
correction of the effects of atmospheric scattering and absorption, which vary
in time and space according to geometry and atmospheric composition. Many
atmospheric radiative transfer models, including that of Multi-Angle
Implementation of Atmospheric Correction (MAIAC), are too computationally
complex to be run in real time, and rely on precomputed look-up tables.
Additionally, uncertainty in measurements and models for remote sensing
receives insufficient attention, in part due to the difficulty of obtaining
sufficient ground measurements. In this paper, we present an adaptation of
Bayesian Deep Learning (BDL) to emulation of the MAIAC atmospheric correction
algorithm. Emulation approaches learn a statistical model as an efficient
approximation of a physical model, while machine learning methods have
demonstrated performance in extracting spatial features and learning complex,
nonlinear mappings. We demonstrate stable surface reflectance retrieval by
emulation (R2 between MAIAC and emulator SR are 0.63, 0.75, 0.86, 0.84, 0.95,
and 0.91 for Blue, Green, Red, NIR, SWIR1, and SWIR2 bands, respectively),
accurate cloud detection (86\%), and well-calibrated, geolocated uncertainty
estimates. Our results support BDL-based emulation as an accurate and efficient
(up to 6x speedup) method for approximation atmospheric correction, where
built-in uncertainty estimates stand to open new opportunities for model
assessment and support informed use of SR-derived quantities in multiple
domains. | [
"cs.LG",
"eess.IV",
"stat.ML"
] |
There hardly exists any large-scale datasets with dense optical flow of
non-rigid motion from real-world imagery as of today. The reason lies mainly in
the required setup to derive ground truth optical flows: a series of images
with known camera poses along its trajectory, and an accurate 3D model from a
textured scene. Human annotation is not only too tedious for large databases,
it can simply hardly contribute to accurate optical flow. To circumvent the
need for manual annotation, we propose a framework to automatically generate
optical flow from real-world videos. The method extracts and matches objects
from video frames to compute initial constraints, and applies a deformation
over the objects of interest to obtain dense optical flow fields. We propose
several ways to augment the optical flow variations. Extensive experimental
results show that training on our automatically generated optical flow
outperforms methods that are trained on rigid synthetic data using FlowNet-S,
LiteFlowNet, PWC-Net, and RAFT. Datasets and implementation of our optical flow
generation framework are released at https://github.com/lhoangan/arap_flow | [
"cs.CV"
] |
Short-form video social media shifts away from the traditional media paradigm
by telling the audience a dynamic story to attract their attention. In
particular, different combinations of everyday objects can be employed to
represent a unique scene that is both interesting and understandable. Offered
by the same company, TikTok and Douyin are popular examples of such new media
that has become popular in recent years, while being tailored for different
markets (e.g. the United States and China). The hypothesis that they express
cultural differences together with media fashion and social idiosyncrasy is the
primary target of our research. To that end, we first employ the Faster
Regional Convolutional Neural Network (Faster R-CNN) pre-trained with the
Microsoft Common Objects in COntext (MS-COCO) dataset to perform object
detection. Based on a suite of objects detected from videos, we perform
statistical analysis including label statistics, label similarity, and
label-person distribution. We further use the Two-Stream Inflated 3D ConvNet
(I3D) pre-trained with the Kinetics dataset to categorize and analyze human
actions. By comparing the distributional results of TikTok and Douyin, we
uncover a wealth of similarity and contrast between the two closely related
video social media platforms along the content dimensions of object quantity,
object categories, and human action categories. | [
"cs.CV",
"cs.MM",
"cs.SI"
] |
Robotic surgery has become a powerful tool for performing minimally invasive
procedures, providing advantages in dexterity, precision, and 3D vision, over
traditional surgery. One popular robotic system is the da Vinci surgical
platform, which allows preoperative information to be incorporated into live
procedures using Augmented Reality (AR). Scene depth estimation is a
prerequisite for AR, as accurate registration requires 3D correspondences
between preoperative and intraoperative organ models. In the past decade, there
has been much progress on depth estimation for surgical scenes, such as using
monocular or binocular laparoscopes [1,2]. More recently, advances in deep
learning have enabled depth estimation via Convolutional Neural Networks (CNNs)
[3], but training requires a large image dataset with ground truth depths.
Inspired by [4], we propose a deep learning framework for surgical scene depth
estimation using self-supervision for scalable data acquisition. Our framework
consists of an autoencoder for depth prediction, and a differentiable spatial
transformer for training the autoencoder on stereo image pairs without ground
truth depths. Validation was conducted on stereo videos collected in robotic
partial nephrectomy. | [
"cs.CV",
"cs.RO"
] |
Deep metric learning aims to learn an embedding function, modeled as deep
neural network. This embedding function usually puts semantically similar
images close while dissimilar images far from each other in the learned
embedding space. Recently, ensemble has been applied to deep metric learning to
yield state-of-the-art results. As one important aspect of ensemble, the
learners should be diverse in their feature embeddings. To this end, we propose
an attention-based ensemble, which uses multiple attention masks, so that each
learner can attend to different parts of the object. We also propose a
divergence loss, which encourages diversity among the learners. The proposed
method is applied to the standard benchmarks of deep metric learning and
experimental results show that it outperforms the state-of-the-art methods by a
significant margin on image retrieval tasks. | [
"cs.CV"
] |
Gradual argumentation frameworks represent arguments and their relationships
in a weighted graph. Their graphical structure and intuitive semantics makes
them a potentially interesting tool for interpretable machine learning. It has
been noted recently that their mechanics are closely related to neural
networks, which allows learning their weights from data by standard deep
learning frameworks. As a first proof of concept, we propose a genetic
algorithm to simultaneously learn the structure of argumentative classification
models. To obtain a well interpretable model, the fitness function balances
sparseness and accuracy of the classifier. We discuss our algorithm and present
first experimental results on standard benchmarks from the UCI machine learning
repository. Our prototype learns argumentative classification models that are
comparable to decision trees in terms of learning performance and
interpretability. | [
"cs.LG",
"cs.NE"
] |
A non-parametric low-resolution face recognition model for
resource-constrained environments with limited networking and computing is
proposed in this work. Such environments often demand a small model capable of
being effectively trained on a small number of labeled data samples, with low
training complexity, and low-resolution input images. To address these
challenges, we adopt an emerging explainable machine learning methodology
called successive subspace learning (SSL).SSL offers an explainable
non-parametric model that flexibly trades the model size for verification
performance. Its training complexity is significantly lower since its model is
trained in a one-pass feedforward manner without backpropagation. Furthermore,
active learning can be conveniently incorporated to reduce the labeling cost.
The effectiveness of the proposed model is demonstrated by experiments on the
LFW and the CMU Multi-PIE datasets. | [
"cs.CV"
] |
Social media produces large amounts of contents every day. To help users
quickly capture what they need, keyphrase prediction is receiving a growing
attention. Nevertheless, most prior efforts focus on text modeling, largely
ignoring the rich features embedded in the matching images. In this work, we
explore the joint effects of texts and images in predicting the keyphrases for
a multimedia post. To better align social media style texts and images, we
propose: (1) a novel Multi-Modality Multi-Head Attention (M3H-Att) to capture
the intricate cross-media interactions; (2) image wordings, in forms of optical
characters and image attributes, to bridge the two modalities. Moreover, we
design a unified framework to leverage the outputs of keyphrase classification
and generation and couple their advantages. Extensive experiments on a
large-scale dataset newly collected from Twitter show that our model
significantly outperforms the previous state of the art based on traditional
attention networks. Further analyses show that our multi-head attention is able
to attend information from various aspects and boost classification or
generation in diverse scenarios. | [
"cs.CV",
"cs.CL"
] |
This paper presents a novel Subject-dependent Deep Aging Path (SDAP), which
inherits the merits of both Generative Probabilistic Modeling and Inverse
Reinforcement Learning to model the facial structures and the longitudinal face
aging process of a given subject. The proposed SDAP is optimized using
tractable log-likelihood objective functions with Convolutional Neural Networks
(CNNs) based deep feature extraction. Instead of applying a fixed aging
development path for all input faces and subjects, SDAP is able to provide the
most appropriate aging development path for individual subject that optimizes
the reward aging formulation. Unlike previous methods that can take only one
image as the input, SDAP further allows multiple images as inputs, i.e. all
information of a subject at either the same or different ages, to produce the
optimal aging path for the given subject. Finally, SDAP allows efficiently
synthesizing in-the-wild aging faces. The proposed model is experimented in
both tasks of face aging synthesis and cross-age face verification. The
experimental results consistently show SDAP achieves the state-of-the-art
performance on numerous face aging databases, i.e. FG-NET, MORPH, AginG Faces
in the Wild (AGFW), and Cross-Age Celebrity Dataset (CACD). Furthermore, we
also evaluate the performance of SDAP on large-scale Megaface challenge to
demonstrate the advantages of the proposed solution. | [
"cs.CV"
] |
In computer vision, image segmentation is always selected as a major research
topic by researchers. Due to its vital rule in image processing, there always
arises the need of a better image segmentation method. Clustering is an
unsupervised study with its application in almost every field of science and
engineering. Many researchers used clustering in image segmentation process.
But still there requires improvement of such approaches. In this paper, a novel
approach for clustering based image segmentation is proposed. Here, we give
importance on color space and choose lab for this task. The famous hard
clustering algorithm K-means is used, but as its performance is dependent on
choosing a proper distance measure, so, we go for cosine distance measure. Then
the segmented image is filtered with sobel filter. The filtered image is
analyzed with marker watershed algorithm to have the final segmented result of
our original image. The MSE and PSNR values are evaluated to observe the
performance. | [
"cs.CV"
] |
Slow feature analysis (SFA) is a method for extracting slowly varying driving
forces from quickly varying nonstationary time series. We show here that it is
possible for SFA to detect a component which is even slower than the driving
force itself (e.g. the envelope of a modulated sine wave). It is shown that it
depends on circumstances like the embedding dimension, the time series
predictability, or the base frequency, whether the driving force itself or a
slower subcomponent is detected. We observe a phase transition from one regime
to the other and it is the purpose of this work to quantify the influence of
various parameters on this phase transition. We conclude that what is percieved
as slow by SFA varies and that a more or less fast switching from one regime to
the other occurs, perhaps showing some similarity to human perception. | [
"stat.ML"
] |
User identity linkage is a task of recognizing the identities of the same
user across different social networks (SN). Previous works tackle this problem
via estimating the pairwise similarity between identities from different SN,
predicting the label of identity pairs or selecting the most relevant identity
pair based on the similarity scores. However, most of these methods ignore the
results of previously matched identities, which could contribute to the linkage
in following matching steps. To address this problem, we convert user identity
linkage into a sequence decision problem and propose a reinforcement learning
model to optimize the linkage strategy from the global perspective. Our method
makes full use of both the social network structure and the history matched
identities, and explores the long-term influence of current matching on
subsequent decisions. We conduct experiments on different types of datasets,
the results show that our method achieves better performance than other
state-of-the-art methods. | [
"cs.LG",
"cs.CY",
"cs.SI",
"stat.ML"
] |
In this paper, we present a regression-based pose recognition method using
cascade Transformers. One way to categorize the existing approaches in this
domain is to separate them into 1). heatmap-based and 2). regression-based. In
general, heatmap-based methods achieve higher accuracy but are subject to
various heuristic designs (not end-to-end mostly), whereas regression-based
approaches attain relatively lower accuracy but they have less intermediate
non-differentiable steps. Here we utilize the encoder-decoder structure in
Transformers to perform regression-based person and keypoint detection that is
general-purpose and requires less heuristic design compared with the existing
approaches. We demonstrate the keypoint hypothesis (query) refinement process
across different self-attention layers to reveal the recursive self-attention
mechanism in Transformers. In the experiments, we report competitive results
for pose recognition when compared with the competing regression-based methods. | [
"cs.CV"
] |
This work is the first to employ and adapt the image-to-image translation
concept based on conditional generative adversarial networks (cGAN) towards
learning a forward and an inverse solution operator of partial differential
equations (PDEs). Even though the proposed framework could be applied as a
surrogate model for the solution of any PDEs, here we focus on steady-state
solutions of coupled hydro-mechanical processes in heterogeneous porous media.
Strongly heterogeneous material properties, which translate to the
heterogeneity of coefficients of the PDEs and discontinuous features in the
solutions, require specialized techniques for the forward and inverse solution
of these problems. Additionally, parametrization of the spatially heterogeneous
coefficients is excessively difficult by using standard reduced order modeling
techniques. In this work, we overcome these challenges by employing the
image-to-image translation concept to learn the forward and inverse solution
operators and utilize a U-Net generator and a patch-based discriminator. Our
results show that the proposed data-driven reduced order model has competitive
predictive performance capabilities in accuracy and computational efficiency as
well as training time requirements compared to state-of-the-art data-driven
methods for both forward and inverse problems. | [
"cs.LG",
"cs.NA",
"math.NA"
] |
Tracking a crowd in 3D using multiple RGB cameras is a challenging task. Most
previous multi-camera tracking algorithms are designed for offline setting and
have high computational complexity. Robust real-time multi-camera 3D tracking
is still an unsolved problem. In this work, we propose a novel end-to-end
tracking pipeline, Deep Multi-Camera Tracking (DMCT), which achieves reliable
real-time multi-camera people tracking. Our DMCT consists of 1) a fast and
novel perspective-aware Deep GroudPoint Network, 2) a fusion procedure for
ground-plane occupancy heatmap estimation, 3) a novel Deep Glimpse Network for
person detection and 4) a fast and accurate online tracker. Our design fully
unleashes the power of deep neural network to estimate the "ground point" of
each person in each color image, which can be optimized to run efficiently and
robustly. Our fusion procedure, glimpse network and tracker merge the results
from different views, find people candidates using multiple video frames and
then track people on the fused heatmap. Our system achieves the
state-of-the-art tracking results while maintaining real-time performance.
Apart from evaluation on the challenging WILDTRACK dataset, we also collect two
more tracking datasets with high-quality labels from two different environments
and camera settings. Our experimental results confirm that our proposed
real-time pipeline gives superior results to previous approaches. | [
"cs.CV"
] |
Change detection of heterogeneous remote sensing images is an important and
challenging topic in remote sensing for emergency situation resulting from
nature disaster. Due to the different imaging mechanisms of heterogeneous
sensors, it is difficult to directly compare the images. To address this
challenge, we explore an unsupervised change detection method based on adaptive
local structure consistency (ALSC) between heterogeneous images in this letter,
which constructs an adaptive graph representing the local structure for each
patch in one image domain and then projects this graph to the other image
domain to measure the change level. This local structure consistency exploits
the fact that the heterogeneous images share the same structure information for
the same ground object, which is imaging modality-invariant. To avoid the
leakage of heterogeneous data, the pixelwise change image is calculated in the
same image domain by graph projection. Experiment results demonstrate the
effectiveness of the proposed ALSC based change detection method by comparing
with some state-of-the-art methods. | [
"cs.CV"
] |
Handling missing data is one of the most fundamental problems in machine
learning. Among many approaches, the simplest and most intuitive way is zero
imputation, which treats the value of a missing entry simply as zero. However,
many studies have experimentally confirmed that zero imputation results in
suboptimal performances in training neural networks. Yet, none of the existing
work has explained what brings such performance degradations. In this paper, we
introduce the variable sparsity problem (VSP), which describes a phenomenon
where the output of a predictive model largely varies with respect to the rate
of missingness in the given input, and show that it adversarially affects the
model performance. We first theoretically analyze this phenomenon and propose a
simple yet effective technique to handle missingness, which we refer to as
Sparsity Normalization (SN), that directly targets and resolves the VSP. We
further experimentally validate SN on diverse benchmark datasets, to show that
debiasing the effect of input-level sparsity improves the performance and
stabilizes the training of neural networks. | [
"cs.LG",
"stat.ML"
] |
Sensor-based human activity recognition has become a critical component of
many emerging applications ranging from behavioral medicine to gaming. However,
an unprecedented increase in the diversity of sensor devices in the
Internet-of-Things era has limited the adoption of activity recognition models
for use across different domains. We propose ActiLabel a combinatorial
framework that learns structural similarities among the events in an arbitrary
domain and those of a different domain. The structural similarities are
captured through a graph model, referred to as the it dependency graph, which
abstracts details of activity patterns in low-level signal and feature space.
The activity labels are then autonomously learned by finding an optimal tiered
mapping between the dependency graphs. Extensive experiments based on three
public datasets demonstrate the superiority of ActiLabel over state-of-the-art
transfer learning and deep learning methods. | [
"cs.LG",
"stat.ML"
] |
A novel multi-scale operator for unorganized 3D point clouds is introduced.
The Difference of Normals (DoN) provides a computationally efficient,
multi-scale approach to processing large unorganized 3D point clouds. The
application of DoN in the multi-scale filtering of two different real-world
outdoor urban LIDAR scene datasets is quantitatively and qualitatively
demonstrated. In both datasets the DoN operator is shown to segment large 3D
point clouds into scale-salient clusters, such as cars, people, and lamp posts
towards applications in semi-automatic annotation, and as a pre-processing step
in automatic object recognition. The application of the operator to
segmentation is evaluated on a large public dataset of outdoor LIDAR scenes
with ground truth annotations. | [
"cs.CV"
] |
We introduce Activity Graph Transformer, an end-to-end learnable model for
temporal action localization, that receives a video as input and directly
predicts a set of action instances that appear in the video. Detecting and
localizing action instances in untrimmed videos requires reasoning over
multiple action instances in a video. The dominant paradigms in the literature
process videos temporally to either propose action regions or directly produce
frame-level detections. However, sequential processing of videos is problematic
when the action instances have non-sequential dependencies and/or non-linear
temporal ordering, such as overlapping action instances or re-occurrence of
action instances over the course of the video. In this work, we capture this
non-linear temporal structure by reasoning over the videos as non-sequential
entities in the form of graphs. We evaluate our model on challenging datasets:
THUMOS14, Charades, and EPIC-Kitchens-100. Our results show that our proposed
model outperforms the state-of-the-art by a considerable margin. | [
"cs.CV",
"cs.AI"
] |
This paper proposes RIU-Net (for Range-Image U-Net), the adaptation of a
popular semantic segmentation network for the semantic segmentation of a 3D
LiDAR point cloud. The point cloud is turned into a 2D range-image by
exploiting the topology of the sensor. This image is then used as input to a
U-net. This architecture has already proved its efficiency for the task of
semantic segmentation of medical images. We demonstrate how it can also be used
for the accurate semantic segmentation of a 3D LiDAR point cloud and how it
represents a valid bridge between image processing and 3D point cloud
processing. Our model is trained on range-images built from KITTI 3D object
detection dataset. Experiments show that RIU-Net, despite being very simple,
offers results that are comparable to the state-of-the-art of range-image based
methods. Finally, we demonstrate that this architecture is able to operate at
90fps on a single GPU, which enables deployment for real-time segmentation. | [
"cs.CV"
] |
Neural rendering techniques combining machine learning with geometric
reasoning have arisen as one of the most promising approaches for synthesizing
novel views of a scene from a sparse set of images. Among these, stands out the
Neural radiance fields (NeRF), which trains a deep network to map 5D input
coordinates (representing spatial location and viewing direction) into a volume
density and view-dependent emitted radiance. However, despite achieving an
unprecedented level of photorealism on the generated images, NeRF is only
applicable to static scenes, where the same spatial location can be queried
from different images. In this paper we introduce D-NeRF, a method that extends
neural radiance fields to a dynamic domain, allowing to reconstruct and render
novel images of objects under rigid and non-rigid motions from a \emph{single}
camera moving around the scene. For this purpose we consider time as an
additional input to the system, and split the learning process in two main
stages: one that encodes the scene into a canonical space and another that maps
this canonical representation into the deformed scene at a particular time.
Both mappings are simultaneously learned using fully-connected networks. Once
the networks are trained, D-NeRF can render novel images, controlling both the
camera view and the time variable, and thus, the object movement. We
demonstrate the effectiveness of our approach on scenes with objects under
rigid, articulated and non-rigid motions. Code, model weights and the dynamic
scenes dataset will be released. | [
"cs.CV"
] |
The ability to compare two degenerate probability distributions (i.e. two
probability distributions supported on two distinct low-dimensional manifolds
living in a much higher-dimensional space) is a crucial problem arising in the
estimation of generative models for high-dimensional observations such as those
arising in computer vision or natural language. It is known that optimal
transport metrics can represent a cure for this problem, since they were
specifically designed as an alternative to information divergences to handle
such problematic scenarios. Unfortunately, training generative machines using
OT raises formidable computational and statistical challenges, because of (i)
the computational burden of evaluating OT losses, (ii) the instability and lack
of smoothness of these losses, (iii) the difficulty to estimate robustly these
losses and their gradients in high dimension. This paper presents the first
tractable computational method to train large scale generative models using an
optimal transport loss, and tackles these three issues by relying on two key
ideas: (a) entropic smoothing, which turns the original OT loss into one that
can be computed using Sinkhorn fixed point iterations; (b) algorithmic
(automatic) differentiation of these iterations. These two approximations
result in a robust and differentiable approximation of the OT loss with
streamlined GPU execution. Entropic smoothing generates a family of losses
interpolating between Wasserstein (OT) and Maximum Mean Discrepancy (MMD), thus
allowing to find a sweet spot leveraging the geometry of OT and the favorable
high-dimensional sample complexity of MMD which comes with unbiased gradient
estimates. The resulting computational architecture complements nicely standard
deep network generative models by a stack of extra layers implementing the loss
function. | [
"stat.ML"
] |
Database activity monitoring (DAM) systems are commonly used by organizations
to protect the organizational data, knowledge and intellectual properties. In
order to protect organizations database DAM systems have two main roles,
monitoring (documenting activity) and alerting to anomalous activity. Due to
high-velocity streams and operating costs, such systems are restricted to
examining only a sample of the activity. Current solutions use policies,
manually crafted by experts, to decide which transactions to monitor and log.
This limits the diversity of the data collected. Bandit algorithms, which use
reward functions as the basis for optimization while adding diversity to the
recommended set, have gained increased attention in recommendation systems for
improving diversity.
In this work, we redefine the data sampling problem as a special case of the
multi-armed bandit (MAB) problem and present a novel algorithm, which combines
expert knowledge with random exploration. We analyze the effect of diversity on
coverage and downstream event detection tasks using a simulated dataset. In
doing so, we find that adding diversity to the sampling using the bandit-based
approach works well for this task and maximizing population coverage without
decreasing the quality in terms of issuing alerts about events. | [
"cs.LG",
"cs.AI",
"stat.ML"
] |
Spatial trajectories are ubiquitous and complex signals. Their analysis is
crucial in many research fields, from urban planning to neuroscience. Several
approaches have been proposed to cluster trajectories. They rely on
hand-crafted features, which struggle to capture the spatio-temporal complexity
of the signal, or on Artificial Neural Networks (ANNs) which can be more
efficient but less interpretable. In this paper we present a novel ANN
architecture designed to capture the spatio-temporal patterns characteristic of
a set of trajectories, while taking into account the demographics of the
navigators. Hence, our model extracts markers linked to both behaviour and
demographics. We propose a composite signal analyser (CompSNN) combining three
simple ANN modules. Each of these modules uses different signal representations
of the trajectory while remaining interpretable. Our CompSNN performs
significantly better than its modules taken in isolation and allows to
visualise which parts of the signal were most useful to discriminate the
trajectories. | [
"cs.LG",
"stat.ML"
] |
In recent years, there has been a rapidly expanding focus on explaining the
predictions made by black-box AI systems that handle image and tabular data.
However, considerably less attention has been paid to explaining the
predictions of opaque AI systems handling time series data. In this paper, we
advance a novel model-agnostic, case-based technique -- Native Guide -- that
generates counterfactual explanations for time series classifiers. Given a
query time series, $T_{q}$, for which a black-box classification system
predicts class, $c$, a counterfactual time series explanation shows how $T_{q}$
could change, such that the system predicts an alternative class, $c'$. The
proposed instance-based technique adapts existing counterfactual instances in
the case-base by highlighting and modifying discriminative areas of the time
series that underlie the classification. Quantitative and qualitative results
from two comparative experiments indicate that Native Guide generates
plausible, proximal, sparse and diverse explanations that are better than those
produced by key benchmark counterfactual methods. | [
"cs.LG",
"stat.ML"
] |
We propose a method for meta-learning reinforcement learning algorithms by
searching over the space of computational graphs which compute the loss
function for a value-based model-free RL agent to optimize. The learned
algorithms are domain-agnostic and can generalize to new environments not seen
during training. Our method can both learn from scratch and bootstrap off known
existing algorithms, like DQN, enabling interpretable modifications which
improve performance. Learning from scratch on simple classical control and
gridworld tasks, our method rediscovers the temporal-difference (TD) algorithm.
Bootstrapped from DQN, we highlight two learned algorithms which obtain good
generalization performance over other classical control tasks, gridworld type
tasks, and Atari games. The analysis of the learned algorithm behavior shows
resemblance to recently proposed RL algorithms that address overestimation in
value-based methods. | [
"cs.LG",
"cs.AI",
"cs.NE"
] |
Proximal policy optimization (PPO) is one of the most popular deep
reinforcement learning (RL) methods, achieving state-of-the-art performance
across a wide range of challenging tasks. However, as a model-free RL method,
the success of PPO relies heavily on the effectiveness of its exploratory
policy search. In this paper, we give an in-depth analysis on the exploration
behavior of PPO, and show that PPO is prone to suffer from the risk of lack of
exploration especially under the case of bad initialization, which may lead to
the failure of training or being trapped in bad local optima. To address these
issues, we proposed a novel policy optimization method, named Trust
Region-Guided PPO (TRGPPO), which adaptively adjusts the clipping range within
the trust region. We formally show that this method not only improves the
exploration ability within the trust region but enjoys a better performance
bound compared to the original PPO as well. Extensive experiments verify the
advantage of the proposed method. | [
"cs.LG",
"cs.AI",
"stat.ML"
] |
Salient object detection is a fundamental topic in computer vision. Previous
methods based on RGB-D often suffer from the incompatibility of multi-modal
feature fusion and the insufficiency of multi-scale feature aggregation. To
tackle these two dilemmas, we propose a novel multi-modal and multi-scale
refined network (M2RNet). Three essential components are presented in this
network. The nested dual attention module (NDAM) explicitly exploits the
combined features of RGB and depth flows. The adjacent interactive aggregation
module (AIAM) gradually integrates the neighbor features of high, middle and
low levels. The joint hybrid optimization loss (JHOL) makes the predictions
have a prominent outline. Extensive experiments demonstrate that our method
outperforms other state-of-the-art approaches. | [
"cs.CV"
] |
Although unsupervised person re-identification (RE-ID) has drawn increasing
research attentions due to its potential to address the scalability problem of
supervised RE-ID models, it is very challenging to learn discriminative
information in the absence of pairwise labels across disjoint camera views. To
overcome this problem, we propose a deep model for the soft multilabel learning
for unsupervised RE-ID. The idea is to learn a soft multilabel (real-valued
label likelihood vector) for each unlabeled person by comparing (and
representing) the unlabeled person with a set of known reference persons from
an auxiliary domain. We propose the soft multilabel-guided hard negative mining
to learn a discriminative embedding for the unlabeled target domain by
exploring the similarity consistency of the visual features and the soft
multilabels of unlabeled target pairs. Since most target pairs are cross-view
pairs, we develop the cross-view consistent soft multilabel learning to achieve
the learning goal that the soft multilabels are consistently good across
different camera views. To enable effecient soft multilabel learning, we
introduce the reference agent learning to represent each reference person by a
reference agent in a joint embedding. We evaluate our unified deep model on
Market-1501 and DukeMTMC-reID. Our model outperforms the state-of-the-art
unsupervised RE-ID methods by clear margins. Code is available at
https://github.com/KovenYu/MAR. | [
"cs.CV"
] |
We present recurrent geometry-aware neural networks that integrate visual
information across multiple views of a scene into 3D latent feature tensors,
while maintaining an one-to-one mapping between 3D physical locations in the
world scene and latent feature locations. Object detection, object
segmentation, and 3D reconstruction is then carried out directly using the
constructed 3D feature memory, as opposed to any of the input 2D images. The
proposed models are equipped with differentiable egomotion-aware feature
warping and (learned) depth-aware unprojection operations to achieve
geometrically consistent mapping between the features in the input frame and
the constructed latent model of the scene. We empirically show the proposed
model generalizes much better than geometryunaware LSTM/GRU networks,
especially under the presence of multiple objects and cross-object occlusions.
Combined with active view selection policies, our model learns to select
informative viewpoints to integrate information from by "undoing" cross-object
occlusions, seamlessly combining geometry with learning from experience. | [
"cs.CV"
] |
The $K$-means algorithm is extended to allow for partitioning of skewed
groups. Our algorithm is called TiK-Means and contributes a $K$-means type
algorithm that assigns observations to groups while estimating their
skewness-transformation parameters. The resulting groups and transformation
reveal general-structured clusters that can be explained by inverting the
estimated transformation. Further, a modification of the jump statistic chooses
the number of groups. Our algorithm is evaluated on simulated and real-life
datasets and then applied to a long-standing astronomical dispute regarding the
distinct kinds of gamma ray bursts. | [
"stat.ML",
"astro-ph.HE",
"cs.CV",
"cs.LG",
"stat.AP",
"stat.ME"
] |
In this paper, we propose a set of features called temporal accumulative
features (TAF) for representing and recognizing isolated sign language
gestures. By incorporating sign language specific constructs to better
represent the unique linguistic characteristic of sign language videos, we have
devised an efficient and fast SLR method for recognizing isolated sign language
gestures. The proposed method is an HSV based accumulative video representation
where keyframes based on the linguistic movement-hold model are represented by
different colors. We also incorporate hand shape information and using a small
scale convolutional neural network, demonstrate that sequential modeling of
accumulative features for linguistic subunits improves upon baseline
classification results. | [
"cs.CV"
] |
Although the adoption rate of deep neural networks (DNNs) has tremendously
increased in recent years, a solution for their vulnerability against
adversarial examples has not yet been found. As a result, substantial research
efforts are dedicated to fix this weakness, with many studies typically using a
subset of source images to generate adversarial examples, treating every image
in this subset as equal. We demonstrate that, in fact, not every source image
is equally suited for this kind of assessment. To do so, we devise a
large-scale model-to-model transferability scenario for which we meticulously
analyze the properties of adversarial examples, generated from every suitable
source image in ImageNet by making use of two of the most frequently deployed
attacks. In this transferability scenario, which involves seven distinct DNN
models, including the recently proposed vision transformers, we reveal that it
is possible to have a difference of up to $12.5\%$ in model-to-model
transferability success, $1.01$ in average $L_2$ perturbation, and $0.03$
($8/225$) in average $L_{\infty}$ perturbation when $1,000$ source images are
sampled randomly among all suitable candidates. We then take one of the first
steps in evaluating the robustness of images used to create adversarial
examples, proposing a number of simple but effective methods to identify
unsuitable source images, thus making it possible to mitigate extreme cases in
experimentation and support high-quality benchmarking. | [
"cs.CV",
"cs.CR",
"cs.LG"
] |
Tracking by detection, the dominant approach for online multi-object
tracking, alternates between localization and re-identification steps. As a
result, it strongly depends on the quality of instantaneous observations, often
failing when objects are not fully visible. In contrast, tracking in humans is
underlined by the notion of object permanence: once an object is recognized, we
are aware of its physical existence and can approximately localize it even
under full occlusions. In this work, we introduce an end-to-end trainable
approach for joint object detection and tracking that is capable of such
reasoning. We build on top of the recent CenterTrack architecture, which takes
pairs of frames as input, and extend it to videos of arbitrary length. To this
end, we augment the model with a spatio-temporal, recurrent memory module,
allowing it to reason about object locations and identities in the current
frame using all the previous history. It is, however, not obvious how to train
such an approach. We study this question on a new, large-scale, synthetic
dataset for multi-object tracking, which provides ground truth annotations for
invisible objects, and propose several approaches for supervising tracking
behind occlusions. Our model, trained jointly on synthetic and real data,
outperforms the state of the art on KITTI, and MOT17 datasets thanks to its
robustness to occlusions. | [
"cs.CV"
] |
Deep generative architectures provide a way to model not only images but also
complex, 3-dimensional objects, such as point clouds. In this work, we present
a novel method to obtain meaningful representations of 3D shapes that can be
used for challenging tasks including 3D points generation, reconstruction,
compression, and clustering. Contrary to existing methods for 3D point cloud
generation that train separate decoupled models for representation learning and
generation, our approach is the first end-to-end solution that allows to
simultaneously learn a latent space of representation and generate 3D shape out
of it. Moreover, our model is capable of learning meaningful compact binary
descriptors with adversarial training conducted on a latent space. To achieve
this goal, we extend a deep Adversarial Autoencoder model (AAE) to accept 3D
input and create 3D output. Thanks to our end-to-end training regime, the
resulting method called 3D Adversarial Autoencoder (3dAAE) obtains either
binary or continuous latent space that covers a much wider portion of training
data distribution. Finally, our quantitative evaluation shows that 3dAAE
provides state-of-the-art results for 3D points clustering and 3D object
retrieval. | [
"cs.LG",
"cs.CV",
"stat.ML"
] |
Time series forecasting is an extensively studied subject in statistics,
economics, and computer science. Exploration of the correlation and causation
among the variables in a multivariate time series shows promise in enhancing
the performance of a time series model. When using deep neural networks as
forecasting models, we hypothesize that exploiting the pairwise information
among multiple (multivariate) time series also improves their forecast. If an
explicit graph structure is known, graph neural networks (GNNs) have been
demonstrated as powerful tools to exploit the structure. In this work, we
propose learning the structure simultaneously with the GNN if the graph is
unknown. We cast the problem as learning a probabilistic graph model through
optimizing the mean performance over the graph distribution. The distribution
is parameterized by a neural network so that discrete graphs can be sampled
differentiably through reparameterization. Empirical evaluations show that our
method is simpler, more efficient, and better performing than a recently
proposed bilevel learning approach for graph structure learning, as well as a
broad array of forecasting models, either deep or non-deep learning based, and
graph or non-graph based. | [
"cs.LG",
"stat.ML"
] |
We worked with Nestle SHIELD (Skin Health, Innovation, Education, and
Longevity Development, NSH) to develop a deep learning model that is able to
assess acne severity from selfie images as accurate as dermatologists. The
model was deployed as a mobile application, providing patients an easy way to
assess and track the progress of their acne treatment. NSH acquired 4,700
selfie images for this study and recruited 11 internal dermatologists to label
them in five categories: 1-Clear, 2- Almost Clear, 3-Mild, 4-Moderate,
5-Severe. Using OpenCV to detect facial landmarks we cut specific skin patches
from the selfie images in order to minimize irrelevant background. We then
applied a transfer learning approach by extracting features from the patches
using a ResNet 152 pre-trained model, followed by a fully connected layer
trained to approximate the desired severity rating. To address the problem of
spatial sensitivity of CNN models, we introduce a new image rolling data
augmentation approach, effectively causing acne lesions appeared in more
locations in the training images. Our results demonstrate that this approach
improved the generalization of the CNN model, outperforming more than half of
the panel of human dermatologists on test images. To our knowledge, this is the
first deep learning-based solution for acne assessment using selfie images. | [
"cs.CV"
] |
Automotive Cyber-Physical Systems (ACPS) have attracted a significant amount
of interest in the past few decades, while one of the most critical operations
in these systems is the perception of the environment. Deep learning and,
especially, the use of Deep Neural Networks (DNNs) provides impressive results
in analyzing and understanding complex and dynamic scenes from visual data. The
prediction horizons for those perception systems are very short and inference
must often be performed in real time, stressing the need of transforming the
original large pre-trained networks into new smaller models, by utilizing Model
Compression and Acceleration (MCA) techniques. Our goal in this work is to
investigate best practices for appropriately applying novel weight sharing
techniques, optimizing the available variables and the training procedures
towards the significant acceleration of widely adopted DNNs. Extensive
evaluation studies carried out using various state-of-the-art DNN models in
object detection and tracking experiments, provide details about the type of
errors that manifest after the application of weight sharing techniques,
resulting in significant acceleration gains with negligible accuracy losses. | [
"cs.CV",
"cs.LG"
] |
We develop a novel method, called PoWER-BERT, for improving the inference
time of the popular BERT model, while maintaining the accuracy. It works by: a)
exploiting redundancy pertaining to word-vectors (intermediate encoder outputs)
and eliminating the redundant vectors. b) determining which word-vectors to
eliminate by developing a strategy for measuring their significance, based on
the self-attention mechanism. c) learning how many word-vectors to eliminate by
augmenting the BERT model and the loss function. Experiments on the standard
GLUE benchmark shows that PoWER-BERT achieves up to 4.5x reduction in inference
time over BERT with <1% loss in accuracy. We show that PoWER-BERT offers
significantly better trade-off between accuracy and inference time compared to
prior methods. We demonstrate that our method attains up to 6.8x reduction in
inference time with <1% loss in accuracy when applied over ALBERT, a highly
compressed version of BERT. The code for PoWER-BERT is publicly available at
https://github.com/IBM/PoWER-BERT. | [
"cs.LG",
"cs.CL",
"stat.ML"
] |
Although precision and recall are standard performance measures for anomaly
detection, their statistical properties in sequential detection settings are
poorly understood. In this work, we formalize a notion of precision and recall
with temporal tolerance for point-based anomaly detection in sequential data.
These measures are based on time-tolerant confusion matrices that may be used
to compute time-tolerant variants of many other standard measures. However,
care has to be taken to preserve interpretability. We perform a statistical
simulation study to demonstrate that precision and recall may overestimate the
performance of a detector, when computed with temporal tolerance. To alleviate
this problem, we show how to obtain null distributions for the two measures to
assess the statistical significance of reported results. | [
"cs.LG",
"stat.ML"
] |
This paper designs a technique route to generate high-quality panoramic image
with depth information, which involves two critical research hotspots: fusion
of LiDAR and image data and image stitching. For the fusion of 3D points and
image data, since a sparse depth map can be firstly generated by projecting
LiDAR point onto the RGB image plane based on our reliable calibrated and
synchronized sensors, we adopt a parameter self-adaptive framework to produce
2D dense depth map. For image stitching, optimal seamline for the overlapping
area is searched using a graph-cuts-based method to alleviate the geometric
influence and image blending based on the pyramid multi-band is utilized to
eliminate the photometric effects near the stitching line. Since each pixel is
associated with a depth value, we design this depth value as a radius in the
spherical projection which can further project the panoramic image to the world
coordinate and consequently produces a high-quality measurable panoramic image.
The purposed method is tested on the data from our data collection platform and
presents a satisfactory application prospects. | [
"cs.CV"
] |
A widely established set of unsupervised node embedding methods can be
interpreted as consisting of two distinctive steps: i) the definition of a
similarity matrix based on the graph of interest followed by ii) an explicit or
implicit factorization of such matrix. Inspired by this viewpoint, we propose
improvements in both steps of the framework. On the one hand, we propose to
encode node similarities based on the free energy distance, which interpolates
between the shortest path and the commute time distances, thus, providing an
additional degree of flexibility. On the other hand, we propose a matrix
factorization method based on a loss function that generalizes that of the
skip-gram model with negative sampling to arbitrary similarity matrices.
Compared with factorizations based on the widely used $\ell_2$ loss, the
proposed method can better preserve node pairs associated with higher
similarity scores. Moreover, it can be easily implemented using advanced
automatic differentiation toolkits and computed efficiently by leveraging GPU
resources. Node clustering, node classification, and link prediction
experiments on real-world datasets demonstrate the effectiveness of
incorporating free-energy-based similarities as well as the proposed matrix
factorization compared with state-of-the-art alternatives. | [
"cs.LG",
"cs.SI"
] |
Object detection in streaming images is a major step in different
detection-based applications, such as object tracking, action recognition,
robot navigation, and visual surveillance applications. In mostcases, image
quality is noisy and biased, and as a result, the data distributions are
disturbed and imbalanced. Most object detection approaches, such as the faster
region-based convolutional neural network (Faster RCNN), Single Shot Multibox
Detector with 300x300 inputs (SSD300), and You Only Look Once version 2
(YOLOv2), rely on simple sampling without considering distortions and noise
under real-world changing environments, despite poor object labeling. In this
paper, we propose an Incremental active semi-supervised learning (IASSL)
technology for unseen object detection. It combines batch-based active learning
(AL) and bin-based semi-supervised learning (SSL) to leverage the strong points
of AL's exploration and SSL's exploitation capabilities. A collaborative
sampling method is also adopted to measure the uncertainty and diversity of AL
and the confidence in SSL. Batch-based AL allows us to select more informative,
confident, and representative samples with low cost. Bin-based SSL divides
streaming image samples into several bins, and each bin repeatedly transfers
the discriminative knowledge of convolutional neural network (CNN) deep
learning to the next bin until the performance criterion is reached. IASSL can
overcome noisy and biased labels in unknown, cluttered data distributions. We
obtain superior performance, compared to state-of-the-art technologies such as
Faster RCNN, SSD300, and YOLOv2. | [
"cs.CV",
"cs.LG",
"stat.ML"
] |
Generating images from natural language is one of the primary applications of
recent conditional generative models. Besides testing our ability to model
conditional, highly dimensional distributions, text to image synthesis has many
exciting and practical applications such as photo editing or computer-aided
content creation. Recent progress has been made using Generative Adversarial
Networks (GANs). This material starts with a gentle introduction to these
topics and discusses the existent state of the art models. Moreover, I propose
Wasserstein GAN-CLS, a new model for conditional image generation based on the
Wasserstein distance which offers guarantees of stability. Then, I show how the
novel loss function of Wasserstein GAN-CLS can be used in a Conditional
Progressive Growing GAN. In combination with the proposed loss, the model
boosts by 7.07% the best Inception Score (on the Caltech birds dataset) of the
models which use only the sentence-level visual semantics. The only model which
performs better than the Conditional Wasserstein Progressive Growing GAN is the
recently proposed AttnGAN which uses word-level visual semantics as well. | [
"cs.CV",
"cs.CL"
] |
We propose Unicoder-VL, a universal encoder that aims to learn joint
representations of vision and language in a pre-training manner. Borrow ideas
from cross-lingual pre-trained models, such as XLM and Unicoder, both visual
and linguistic contents are fed into a multi-layer Transformer for the
cross-modal pre-training, where three pre-trained tasks are employed, including
Masked Language Modeling (MLM), Masked Object Classification (MOC) and
Visual-linguistic Matching (VLM). The first two tasks learn context-aware
representations for input tokens based on linguistic and visual contents
jointly. The last task tries to predict whether an image and a text describe
each other. After pretraining on large-scale image-caption pairs, we transfer
Unicoder-VL to caption-based image-text retrieval and visual commonsense
reasoning, with just one additional output layer. We achieve state-of-the-art
or comparable results on both two tasks and show the powerful ability of the
cross-modal pre-training. | [
"cs.CV"
] |
This paper builds on the connection between graph neural networks and
traditional dynamical systems. We propose continuous graph neural networks
(CGNN), which generalise existing graph neural networks with discrete dynamics
in that they can be viewed as a specific discretisation scheme. The key idea is
how to characterise the continuous dynamics of node representations, i.e. the
derivatives of node representations, w.r.t. time. Inspired by existing
diffusion-based methods on graphs (e.g. PageRank and epidemic models on social
networks), we define the derivatives as a combination of the current node
representations, the representations of neighbors, and the initial values of
the nodes. We propose and analyse two possible dynamics on graphs---including
each dimension of node representations (a.k.a. the feature channel) change
independently or interact with each other---both with theoretical
justification. The proposed continuous graph neural networks are robust to
over-smoothing and hence allow us to build deeper networks, which in turn are
able to capture the long-range dependencies between nodes. Experimental results
on the task of node classification demonstrate the effectiveness of our
proposed approach over competitive baselines. | [
"cs.LG",
"stat.ML"
] |
In the real world, agents often have to operate in situations with incomplete
information, limited sensing capabilities, and inherently stochastic
environments, making individual observations incomplete and unreliable.
Moreover, in many situations it is preferable to delay a decision rather than
run the risk of making a bad decision. In such situations it is necessary to
aggregate information before taking an action; however, most state of the art
reinforcement learning (RL) algorithms are biased towards taking actions
\textit{at every time step}, even if the agent is not particularly confident in
its chosen action. This lack of caution can lead the agent to make critical
mistakes, regardless of prior experience and acclimation to the environment.
Motivated by theories of dynamic resolution of uncertainty during decision
making in biological brains, we propose a simple accumulator module which
accumulates evidence in favor of each possible decision, encodes uncertainty as
a dynamic competition between actions, and acts on the environment only when it
is sufficiently confident in the chosen action. The agent makes no decision by
default, and the burden of proof to make a decision falls on the policy to
accrue evidence strongly in favor of a single decision. Our results show that
this accumulator module achieves near-optimal performance on a simple guessing
game, far outperforming deep recurrent networks using traditional, forced
action selection policies. | [
"cs.LG",
"cs.AI",
"stat.ML"
] |
Intelligent behaviour in the real-world requires the ability to acquire new
knowledge from an ongoing sequence of experiences while preserving and reusing
past knowledge. We propose a novel algorithm for unsupervised representation
learning from piece-wise stationary visual data: Variational Autoencoder with
Shared Embeddings (VASE). Based on the Minimum Description Length principle,
VASE automatically detects shifts in the data distribution and allocates spare
representational capacity to new knowledge, while simultaneously protecting
previously learnt representations from catastrophic forgetting. Our approach
encourages the learnt representations to be disentangled, which imparts a
number of desirable properties: VASE can deal sensibly with ambiguous inputs,
it can enhance its own representations through imagination-based exploration,
and most importantly, it exhibits semantically meaningful sharing of latents
between different datasets. Compared to baselines with entangled
representations, our approach is able to reason beyond surface-level statistics
and perform semantically meaningful cross-domain inference. | [
"cs.LG",
"stat.ML"
] |
Taking an image and question as the input of our method, it can output the
text-based answer of the query question about the given image, so called Visual
Question Answering (VQA). There are two main modules in our algorithm. Given a
natural language question about an image, the first module takes the question
as input and then outputs the basic questions of the main given question. The
second module takes the main question, image and these basic questions as input
and then outputs the text-based answer of the main question. We formulate the
basic questions generation problem as a LASSO optimization problem, and also
propose a criterion about how to exploit these basic questions to help answer
main question. Our method is evaluated on the challenging VQA dataset and
yields state-of-the-art accuracy, 60.34% in open-ended task. | [
"cs.CV",
"cs.CL"
] |
The ability of neural networks to continuously learn and adapt to new tasks
while retaining prior knowledge is crucial for many applications. However,
current neural networks tend to forget previously learned tasks when trained on
new ones, i.e., they suffer from Catastrophic Forgetting (CF). The objective of
Continual Learning (CL) is to alleviate this problem, which is particularly
relevant for medical applications, where it may not be feasible to store and
access previously used sensitive patient data. In this work, we propose a
Continual Learning approach for brain segmentation, where a single network is
consecutively trained on samples from different domains. We build upon an
importance driven approach and adapt it for medical image segmentation.
Particularly, we introduce learning rate regularization to prevent the loss of
the network's knowledge. Our results demonstrate that directly restricting the
adaptation of important network parameters clearly reduces Catastrophic
Forgetting for segmentation across domains. | [
"cs.CV"
] |
The adoption of deep learning in healthcare is hindered by their "black box"
nature. In this paper, we explore the RETAIN architecture for the task of
glusose forecasting for diabetic people. By using a two-level attention
mechanism, the recurrent-neural-network-based RETAIN model is interpretable. We
evaluate the RETAIN model on the type-2 IDIAB and the type-1 OhioT1DM datasets
by comparing its statistical and clinical performances against two deep models
and three models based on decision trees. We show that the RETAIN model offers
a very good compromise between accuracy and interpretability, being almost as
accurate as the LSTM and FCN models while remaining interpretable. We show the
usefulness of its interpretable nature by analyzing the contribution of each
variable to the final prediction. It revealed that signal values older than one
hour are not used by the RETAIN model for the 30-minutes ahead of time
prediction of glucose. Also, we show how the RETAIN model changes its behavior
upon the arrival of an event such as carbohydrate intakes or insulin infusions.
In particular, it showed that the patient's state before the event is
particularily important for the prediction. Overall the RETAIN model, thanks to
its interpretability, seems to be a very promissing model for regression or
classification tasks in healthcare. | [
"cs.LG",
"cs.AI",
"stat.ML"
] |
Recent years have witnessed the rapid development of human activity
recognition (HAR) based on wearable sensor data. One can find many practical
applications in this area, especially in the field of health care. Many machine
learning algorithms such as Decision Trees, Support Vector Machine, Naive
Bayes, K-Nearest Neighbor, and Multilayer Perceptron are successfully used in
HAR. Although these methods are fast and easy for implementation, they still
have some limitations due to poor performance in a number of situations. In
this paper, we propose a novel method based on the ensemble learning to boost
the performance of these machine learning methods for HAR. | [
"cs.LG",
"cs.HC",
"eess.SP"
] |
Using heterogeneous depth cameras and 3D scanners in 3D face verification
causes variations in the resolution of the 3D point clouds. To solve this
issue, previous studies use 3D registration techniques. Out of these proposed
techniques, detecting points of correspondence is proven to be an efficient
method given that the data belongs to the same individual. However, if the data
belongs to different persons, the registration algorithms can convert the 3D
point cloud of one person to another, destroying the distinguishing features
between the two point clouds. Another issue regarding the storage size of the
point clouds. That is, if the captured depth image contains around 50 thousand
points in the cloud for a single pose for one individual, then the storage size
of the entire dataset will be in order of giga if not tera bytes. With these
motivations, this work introduces a new technique for 3D point clouds
generation using a neural modeling system to handle the differences caused by
heterogeneous depth cameras, and to generate a new face canonical compact
representation. The proposed system reduces the stored 3D dataset size, and if
required, provides an accurate dataset regeneration. Furthermore, the system
generates neural models for all gallery point clouds and stores these models to
represent the faces in the recognition or verification processes. For the probe
cloud to be verified, a new model is generated specifically for that particular
cloud and is matched against pre-stored gallery model presentations to identify
the query cloud. This work also introduces the utilization of Siamese deep
neural network in 3D face verification using generated model representations as
raw data for the deep network, and shows that the accuracy of the trained
network is comparable all published results on Bosphorus dataset. | [
"cs.CV"
] |
We present the first method to handle curvature regularity in region-based
image segmentation and inpainting that is independent of initialization.
To this end we start from a new formulation of length-based optimization
schemes, based on surface continuation constraints, and discuss the connections
to existing schemes. The formulation is based on a \emph{cell complex} and
considers basic regions and boundary elements. The corresponding optimization
problem is cast as an integer linear program.
We then show how the method can be extended to include curvature regularity,
again cast as an integer linear program. Here, we are considering pairs of
boundary elements to reflect curvature. Moreover, a constraint set is derived
to ensure that the boundary variables indeed reflect the boundary of the
regions described by the region variables.
We show that by solving the linear programming relaxation one gets quite
close to the global optimum, and that curvature regularity is indeed much
better suited in the presence of long and thin objects compared to standard
length regularity. | [
"cs.CV",
"cs.AI",
"math.OC"
] |
Knowledge graph (KG) embeddings learn low-dimensional representations of
entities and relations to predict missing facts. KGs often exhibit hierarchical
and logical patterns which must be preserved in the embedding space. For
hierarchical data, hyperbolic embedding methods have shown promise for
high-fidelity and parsimonious representations. However, existing hyperbolic
embedding methods do not account for the rich logical patterns in KGs. In this
work, we introduce a class of hyperbolic KG embedding models that
simultaneously capture hierarchical and logical patterns. Our approach combines
hyperbolic reflections and rotations with attention to model complex relational
patterns. Experimental results on standard KG benchmarks show that our method
improves over previous Euclidean- and hyperbolic-based efforts by up to 6.1% in
mean reciprocal rank (MRR) in low dimensions. Furthermore, we observe that
different geometric transformations capture different types of relations while
attention-based transformations generalize to multiple relations. In high
dimensions, our approach yields new state-of-the-art MRRs of 49.6% on WN18RR
and 57.7% on YAGO3-10. | [
"cs.LG",
"cs.AI",
"cs.CL",
"stat.ML"
] |
Self-attention has been successfully applied to video representation learning
due to the effectiveness of modeling long range dependencies. Existing
approaches build the dependencies merely by computing the pairwise correlations
along spatial and temporal dimensions simultaneously. However, spatial
correlations and temporal correlations represent different contextual
information of scenes and temporal reasoning. Intuitively, learning spatial
contextual information first will benefit temporal modeling. In this paper, we
propose a separable self-attention (SSA) module, which models spatial and
temporal correlations sequentially, so that spatial contexts can be efficiently
used in temporal modeling. By adding SSA module into 2D CNN, we build a SSA
network (SSAN) for video representation learning. On the task of video action
recognition, our approach outperforms state-of-the-art methods on
Something-Something and Kinetics-400 datasets. Our models often outperform
counterparts with shallower network and fewer modalities. We further verify the
semantic learning ability of our method in visual-language task of video
retrieval, which showcases the homogeneity of video representations and text
embeddings. On MSR-VTT and Youcook2 datasets, video representations learnt by
SSA significantly improve the state-of-the-art performance. | [
"cs.CV"
] |
In high energy physics (HEP), jets are collections of correlated particles
produced ubiquitously in particle collisions such as those at the CERN Large
Hadron Collider (LHC). Machine-learning-based generative models, such as
generative adversarial networks (GANs), have the potential to significantly
accelerate LHC jet simulations. However, despite jets having a natural
representation as a set of particles in momentum-space, a.k.a. a particle
cloud, to our knowledge there exist no generative models applied to such a
dataset. We introduce a new particle cloud dataset (JetNet), and, due to
similarities between particle and point clouds, apply to it existing point
cloud GANs. Results are evaluated using (1) the 1-Wasserstein distance between
high- and low-level feature distributions, (2) a newly developed Fr\'{e}chet
ParticleNet Distance, and (3) the coverage and (4) minimum matching distance
metrics. Existing GANs are found to be inadequate for physics applications,
hence we develop a new message passing GAN (MPGAN), which outperforms existing
point cloud GANs on virtually every metric and shows promise for use in HEP. We
propose JetNet as a novel point-cloud-style dataset for the machine learning
community to experiment with, and set MPGAN as a benchmark to improve upon for
future generative models. | [
"cs.LG",
"hep-ex"
] |
Iris recognition is considered as one of the best biometric methods used for
human identification and verification, this is because of its unique features
that differ from one person to another, and its importance in the security
field. This paper proposes an algorithm for iris recognition and classification
using a system based on Local Binary Pattern and histogram properties as a
statistical approaches for feature extraction, and Combined Learning Vector
Quantization Classifier as Neural Network approach for classification, in order
to build a hybrid model depends on both features. The localization and
segmentation techniques are presented using both Canny edge detection and Hough
Circular Transform in order to isolate an iris from the whole eye image and for
noise detection .Feature vectors results from LBP is applied to a Combined LVQ
classifier with different classes to determine the minimum acceptable
performance, and the result is based on majority voting among several LVQ
classifier. Different iris datasets CASIA, MMU1, MMU2, and LEI with different
extensions and size are presented. Since LBP is working on a grayscale level so
colored iris images should be transformed into a grayscale level. The proposed
system gives a high recognition rate 99.87 % on different iris datasets
compared with other methods. | [
"cs.CV"
] |
Extracting the interaction rules of biological agents from moving sequences
pose challenges in various domains. Granger causality is a practical framework
for analyzing the interactions from observed time-series data; however, this
framework ignores the structures of the generative process in animal behaviors,
which may lead to interpretational problems and sometimes erroneous assessments
of causality. In this paper, we propose a new framework for learning Granger
causality from multi-animal trajectories via augmented theory-based behavioral
models with interpretable data-driven models. We adopt an approach for
augmenting incomplete multi-agent behavioral models described by time-varying
dynamical systems with neural networks. For efficient and interpretable
learning, our model leverages theory-based architectures separating navigation
and motion processes, and the theory-guided regularization for reliable
behavioral modeling. This can provide interpretable signs of Granger-causal
effects over time, i.e., when specific others cause the approach or separation.
In experiments using synthetic datasets, our method achieved better performance
than various baselines. We then analyzed multi-animal datasets of mice, flies,
birds, and bats, which verified our method and obtained novel biological
insights. | [
"cs.LG",
"stat.ML"
] |
Novelty detection is the process of determining whether a query example
differs from the learned training distribution. Previous methods attempt to
learn the representation of the normal samples via generative adversarial
networks (GANs). However, they will suffer from instability training, mode
dropping, and low discriminative ability. Recently, various pretext tasks (e.g.
rotation prediction and clustering) have been proposed for self-supervised
learning in novelty detection. However, the learned latent features are still
low discriminative. We overcome such problems by introducing a novel
decoder-encoder framework. Firstly, a generative network (a.k.a. decoder)
learns the representation by mapping the initialized latent vector to an image.
In particular, this vector is initialized by considering the entire
distribution of training data to avoid the problem of mode-dropping. Secondly,
a contrastive network (a.k.a. encoder) aims to ``learn to compare'' through
mutual information estimation, which directly helps the generative network to
obtain a more discriminative representation by using a negative data
augmentation strategy. Extensive experiments show that our model has
significant superiority over cutting-edge novelty detectors and achieves new
state-of-the-art results on some novelty detection benchmarks, e.g. CIFAR10 and
DCASE. Moreover, our model is more stable for training in a non-adversarial
manner, compared to other adversarial based novelty detection methods. | [
"cs.CV",
"cs.AI"
] |
The efficient treatment of long-range interactions for point clouds is a
challenging problem in many scientific machine learning applications. To
extract global information, one usually needs a large window size, a large
number of layers, and/or a large number of channels. This can often
significantly increase the computational cost. In this work, we present a novel
neural network layer that directly incorporates long-range information for a
point cloud. This layer, dubbed the long-range convolutional (LRC)-layer,
leverages the convolutional theorem coupled with the non-uniform Fourier
transform. In a nutshell, the LRC-layer mollifies the point cloud to an
adequately sized regular grid, computes its Fourier transform, multiplies the
result by a set of trainable Fourier multipliers, computes the inverse Fourier
transform, and finally interpolates the result back to the point cloud. The
resulting global all-to-all convolution operation can be performed in
nearly-linear time asymptotically with respect to the number of input points.
The LRC-layer is a particularly powerful tool when combined with local
convolution as together they offer efficient and seamless treatment of both
short and long range interactions. We showcase this framework by introducing a
neural network architecture that combines LRC-layers with short-range
convolutional layers to accurately learn the energy and force associated with a
$N$-body potential. We also exploit the induced two-level decomposition and
propose an efficient strategy to train the combined architecture with a reduced
number of samples. | [
"stat.ML",
"cs.LG",
"cs.NA",
"math.NA"
] |
Convex Shapes (CS) are common priors for optic disc and cup segmentation in
eye fundus images. It is important to design proper techniques to represent
convex shapes. So far, it is still a problem to guarantee that the output
objects from a Deep Neural Convolution Networks (DCNN) are convex shapes. In
this work, we propose a technique which can be easily integrated into the
commonly used DCNNs for image segmentation and guarantee that outputs are
convex shapes. This method is flexible and it can handle multiple objects and
allow some of the objects to be convex. Our method is based on the dual
representation of the sigmoid activation function in DCNNs. In the dual space,
the convex shape prior can be guaranteed by a simple quadratic constraint on a
binary representation of the shapes. Moreover, our method can also integrate
spatial regularization and some other shape prior using a soft thresholding
dynamics (STD) method. The regularization can make the boundary curves of the
segmentation objects to be simultaneously smooth and convex. We design a very
stable active set projection algorithm to numerically solve our model. This
algorithm can form a new plug-and-play DCNN layer called CS-STD whose outputs
must be a nearly binary segmentation of convex objects. In the CS-STD block,
the convexity information can be propagated to guide the DCNN in both forward
and backward propagation during training and prediction process. As an
application example, we apply the convexity prior layer to the retinal fundus
images segmentation by taking the popular DeepLabV3+ as a backbone network.
Experimental results on several public datasets show that our method is
efficient and outperforms the classical DCNN segmentation methods. | [
"cs.CV"
] |
Lifting is an efficient technique to scale up graphical models generalized to
relational domains by exploiting the underlying symmetries. Concurrently,
neural models are continuously expanding from grid-like tensor data into
structured representations, such as various attributed graphs and relational
databases. To address the irregular structure of the data, the models typically
extrapolate on the idea of convolution, effectively introducing parameter
sharing in their, dynamically unfolded, computation graphs. The computation
graphs themselves then reflect the symmetries of the underlying data, similarly
to the lifted graphical models. Inspired by lifting, we introduce a simple and
efficient technique to detect the symmetries and compress the neural models
without loss of any information. We demonstrate through experiments that such
compression can lead to significant speedups of structured convolutional
models, such as various Graph Neural Networks, across various tasks, such as
molecule classification and knowledge-base completion. | [
"cs.LG",
"cs.AI"
] |
Autonomous robotic systems and self driving cars rely on accurate perception
of their surroundings as the safety of the passengers and pedestrians is the
top priority. Semantic segmentation is one the essential components of
environmental perception that provides semantic information of the scene.
Recently, several methods have been introduced for 3D LiDAR semantic
segmentation. While, they can lead to improved performance, they are either
afflicted by high computational complexity, therefore are inefficient, or lack
fine details of smaller instances. To alleviate this problem, we propose
AF2-S3Net, an end-to-end encoder-decoder CNN network for 3D LiDAR semantic
segmentation. We present a novel multi-branch attentive feature fusion module
in the encoder and a unique adaptive feature selection module with feature map
re-weighting in the decoder. Our AF2-S3Net fuses the voxel based learning and
point-based learning into a single framework to effectively process the large
3D scene. Our experimental results show that the proposed method outperforms
the state-of-the-art approaches on the large-scale SemanticKITTI benchmark,
ranking 1st on the competitive public leaderboard competition upon publication. | [
"cs.CV",
"cs.AI",
"cs.RO"
] |
Exploiting more information from ground truth (GT) images now is a new
research direction for further improving CNN's performance in CT image
segmentation. Previous methods focus on devising the loss function for
fulfilling such a purpose. However, it is rather difficult to devise a general
and optimization-friendly loss function. We here present a novel and practical
method that exploits GT images beyond the loss function. Our insight is that
feature maps of two CNNs trained respectively on GT and CT images should be
similar on some metric space, because they both are used to describe the same
objects for the same purpose. We hence exploit GT images by enforcing such two
CNNs' feature maps to be consistent. We assess the proposed method on two data
sets, and compare its performance to several competitive methods. Extensive
experimental results show that the proposed method is effective, outperforming
all the compared methods. | [
"cs.CV"
] |
We tackle the problem of producing compact models, maximizing their accuracy
for a given model size. A standard solution is to train networks with
Quantization Aware Training, where the weights are quantized during training
and the gradients approximated with the Straight-Through Estimator. In this
paper, we extend this approach to work beyond int8 fixed-point quantization
with extreme compression methods where the approximations introduced by STE are
severe, such as Product Quantization. Our proposal is to only quantize a
different random subset of weights during each forward, allowing for unbiased
gradients to flow through the other weights. Controlling the amount of noise
and its form allows for extreme compression rates while maintaining the
performance of the original model. As a result we establish new
state-of-the-art compromises between accuracy and model size both in natural
language processing and image classification. For example, applying our method
to state-of-the-art Transformer and ConvNet architectures, we can achieve 82.5%
accuracy on MNLI by compressing RoBERTa to 14MB and 80.0 top-1 accuracy on
ImageNet by compressing an EfficientNet-B3 to 3.3MB. | [
"cs.LG",
"stat.ML"
] |
Hyperspectral image (HSI) clustering, which aims at dividing hyperspectral
pixels into clusters, has drawn significant attention in practical
applications. Recently, many graph-based clustering methods, which construct an
adjacent graph to model the data relationship, have shown dominant performance.
However, the high dimensionality of HSI data makes it hard to construct the
pairwise adjacent graph. Besides, abundant spatial structures are often
overlooked during the clustering procedure. In order to better handle the high
dimensionality problem and preserve the spatial structures, this paper proposes
a novel unsupervised approach called spatial-spectral clustering with anchor
graph (SSCAG) for HSI data clustering. The SSCAG has the following
contributions: 1) the anchor graph-based strategy is used to construct a
tractable large graph for HSI data, which effectively exploits all data points
and reduces the computational complexity; 2) a new similarity metric is
presented to embed the spatial-spectral information into the combined adjacent
graph, which can mine the intrinsic property structure of HSI data; 3) an
effective neighbors assignment strategy is adopted in the optimization, which
performs the singular value decomposition (SVD) on the adjacent graph to get
solutions efficiently. Extensive experiments on three public HSI datasets show
that the proposed SSCAG is competitive against the state-of-the-art approaches. | [
"cs.CV"
] |
Hashing has been recognized as an efficient representation learning method to
effectively handle big data due to its low computational complexity and memory
cost. Most of the existing hashing methods focus on learning the
low-dimensional vectorized binary features based on the high-dimensional raw
vectorized features. However, studies on how to obtain preferable binary codes
from the original 2D image features for retrieval is very limited. This paper
proposes a bilinear supervised discrete hashing (BSDH) method based on 2D image
features which utilizes bilinear projections to binarize the image matrix
features such that the intrinsic characteristics in the 2D image space are
preserved in the learned binary codes. Meanwhile, the bilinear projection
approximation and vectorization binary codes regression are seamlessly
integrated together to formulate the final robust learning framework.
Furthermore, a discrete optimization strategy is developed to alternatively
update each variable for obtaining the high-quality binary codes. In addition,
two 2D image features, traditional SURF-based FVLAD feature and CNN-based
AlexConv5 feature are designed for further improving the performance of the
proposed BSDH method. Results of extensive experiments conducted on four
benchmark datasets show that the proposed BSDH method almost outperforms all
competing hashing methods with different input features by different evaluation
protocols. | [
"cs.CV",
"cs.MM"
] |
Long-range and short-range temporal modeling are two complementary and
crucial aspects of video recognition. Most of the state-of-the-arts focus on
short-range spatio-temporal modeling and then average multiple snippet-level
predictions to yield the final video-level prediction. Thus, their video-level
prediction does not consider spatio-temporal features of how video evolves
along the temporal dimension. In this paper, we introduce a novel Dynamic
Segment Aggregation (DSA) module to capture relationship among snippets. To be
more specific, we attempt to generate a dynamic kernel for a convolutional
operation to aggregate long-range temporal information among adjacent snippets
adaptively. The DSA module is an efficient plug-and-play module and can be
combined with the off-the-shelf clip-based models (i.e., TSM, I3D) to perform
powerful long-range modeling with minimal overhead. The final video
architecture, coined as DSANet. We conduct extensive experiments on several
video recognition benchmarks (i.e., Mini-Kinetics-200, Kinetics-400,
Something-Something V1 and ActivityNet) to show its superiority. Our proposed
DSA module is shown to benefit various video recognition models significantly.
For example, equipped with DSA modules, the top-1 accuracy of I3D ResNet-50 is
improved from 74.9% to 78.2% on Kinetics-400. Codes are available at
https://github.com/whwu95/DSANet. | [
"cs.CV",
"cs.MM"
] |
One desired capability for machines is the ability to transfer their
knowledge of one domain to another where data is (usually) scarce. Despite
ample adaptation of transfer learning in various deep learning applications, we
yet do not understand what enables a successful transfer and which part of the
network is responsible for that. In this paper, we provide new tools and
analyses to address these fundamental questions. Through a series of analyses
on transferring to block-shuffled images, we separate the effect of feature
reuse from learning low-level statistics of data and show that some benefit of
transfer learning comes from the latter. We present that when training from
pre-trained weights, the model stays in the same basin in the loss landscape
and different instances of such model are similar in feature space and close in
parameter space. | [
"cs.LG",
"stat.ML"
] |
Computational saliency models for still images have gained significant
popularity in recent years. Saliency prediction from videos, on the other hand,
has received relatively little interest from the community. Motivated by this,
in this work, we study the use of deep learning for dynamic saliency prediction
and propose the so-called spatio-temporal saliency networks. The key to our
models is the architecture of two-stream networks where we investigate
different fusion mechanisms to integrate spatial and temporal information. We
evaluate our models on the DIEM and UCF-Sports datasets and present highly
competitive results against the existing state-of-the-art models. We also carry
out some experiments on a number of still images from the MIT300 dataset by
exploiting the optical flow maps predicted from these images. Our results show
that considering inherent motion information in this way can be helpful for
static saliency estimation. | [
"cs.CV"
] |
Highway traffic modeling and forecasting approaches are critical for
intelligent transportation systems. Recently, deep-learning-based traffic
forecasting methods have emerged as state of the art for a wide range of
traffic forecasting tasks. However, these methods require a large amount of
training data, which needs to be collected over a significant period of time.
This can present a number of challenges for the development and deployment of
data-driven learning methods for highway networks that suffer from lack of
historical data. A promising approach to address this issue is transfer
learning, where a model trained on one part of the highway network can be
adapted for a different part of the highway network. We focus on diffusion
convolutional recurrent neural network (DCRNN), a state-of-the-art graph neural
network for highway network forecasting. It models the complex spatial and
temporal dynamics of the highway network using a graph-based diffusion
convolution operation within a recurrent neural network. DCRNN cannot perform
transfer learning, however, because it learns location-specific traffic
patterns, which cannot be used for unseen regions of the network. To that end,
we develop a new transfer learning approach for DCRNN, where a single model
trained on data-rich regions of the highway network can be used to forecast
traffic on unseen regions of the highway network. We evaluate the ability of
our approach to forecast the traffic on the entire California highway network
with one year of time series data. We show that TL-DCRNN can learn from several
regions of the California highway network and forecast the traffic on the
unseen regions of the network with high accuracy. Moreover, we demonstrate that
TL-DCRNN can learn from San Francisco region traffic data and can forecast
traffic on the Los Angeles region and vice versa. | [
"cs.LG",
"stat.ML"
] |
It is a common paradigm in object detection frameworks to treat all samples
equally and target at maximizing the performance on average. In this work, we
revisit this paradigm through a careful study on how different samples
contribute to the overall performance measured in terms of mAP. Our study
suggests that the samples in each mini-batch are neither independent nor
equally important, and therefore a better classifier on average does not
necessarily mean higher mAP. Motivated by this study, we propose the notion of
Prime Samples, those that play a key role in driving the detection performance.
We further develop a simple yet effective sampling and learning strategy called
PrIme Sample Attention (PISA) that directs the focus of the training process
towards such samples. Our experiments demonstrate that it is often more
effective to focus on prime samples than hard samples when training a detector.
Particularly, On the MSCOCO dataset, PISA outperforms the random sampling
baseline and hard mining schemes, e.g., OHEM and Focal Loss, consistently by
around 2% on both single-stage and two-stage detectors, even with a strong
backbone ResNeXt-101. | [
"cs.CV"
] |
The anomaly detection of time series is a hotspot of time series data mining.
The own characteristics of different anomaly detectors determine the abnormal
data that they are good at. There is no detector can be optimizing in all types
of anomalies. Moreover, it still has difficulties in industrial production due
to problems such as a single detector can't be optimized at different time
windows of the same time series. This paper proposes an adaptive model based on
time series characteristics and selecting appropriate detector and run-time
parameters for anomaly detection, which is called ATSDLN(Adaptive Time Series
Detector Learning Network). We take the time series as the input of the model,
and learn the time series representation through FCN. In order to realize the
adaptive selection of detectors and run-time parameters according to the input
time series, the outputs of FCN are the inputs of two sub-networks: the
detector selection network and the run-time parameters selection network. In
addition, the way that the variable layer width design of the parameter
selection sub-network and the introduction of transfer learning make the model
be with more expandability. Through experiments, it is found that ATSDLN can
select appropriate anomaly detector and run-time parameters, and have strong
expandability, which can quickly transfer. We investigate the performance of
ATSDLN in public data sets, our methods outperform other methods in most cases
with higher effect and better adaptation. We also show experimental results on
public data sets to demonstrate how model structure and transfer learning
affect the effectiveness. | [
"stat.ML",
"cs.LG"
] |
There has recently been significant interest in training reinforcement
learning (RL) agents in vision-based environments. This poses many challenges,
such as high dimensionality and potential for observational overfitting through
spurious correlations. A promising approach to solve both of these problems is
a self-attention bottleneck, which provides a simple and effective framework
for learning high performing policies, even in the presence of distractions.
However, due to poor scalability of attention architectures, these methods do
not scale beyond low resolution visual inputs, using large patches (thus small
attention matrices). In this paper we make use of new efficient attention
algorithms, recently shown to be highly effective for Transformers, and
demonstrate that these new techniques can be applied in the RL setting. This
allows our attention-based controllers to scale to larger visual inputs, and
facilitate the use of smaller patches, even individual pixels, improving
generalization. In addition, we propose a new efficient algorithm approximating
softmax attention with what we call hybrid random features, leveraging the
theory of angular kernels. We show theoretically and empirically that hybrid
random features is a promising approach when using attention for vision-based
RL. | [
"cs.LG",
"cs.AI",
"cs.CV",
"cs.RO"
] |
This research identifies a gap in weakly-labelled multivariate time-series
classification (TSC), where state-of-the-art TSC models do not per-form well.
Weakly labelled time-series are time-series containing noise and significant
redundancies. In response to this gap, this paper proposes an approach of
exploiting context relevance of subsequences from previous subsequences to
improve classification accuracy. To achieve this, state-of-the-art Attention
algorithms are experimented in combination with the top CNN models for TSC (FCN
and ResNet), in an CNN-LSTM architecture. Attention is a popular strategy for
context extraction with exceptional performance in modern sequence-to-sequence
tasks. This paper shows how attention algorithms can be used for improved
weakly labelledTSC by evaluating models on a multivariate EEG time-series
dataset obtained using a commercial Emotiv headsets from participants
performing various activities while driving. These time-series are segmented
into sub-sequences and labelled to allow supervised TSC. | [
"cs.LG"
] |
Many vision-language tasks can be reduced to the problem of sequence
prediction for natural language output. In particular, recent advances in image
captioning use deep reinforcement learning (RL) to alleviate the "exposure
bias" during training: ground-truth subsequence is exposed in every step
prediction, which introduces bias in test when only predicted subsequence is
seen. However, existing RL-based image captioning methods only focus on the
language policy while not the visual policy (e.g., visual attention), and thus
fail to capture the visual context that are crucial for compositional reasoning
such as visual relationships (e.g., "man riding horse") and comparisons (e.g.,
"smaller cat"). To fill the gap, we propose a Context-Aware Visual Policy
network (CAVP) for sequence-level image captioning. At every time step, CAVP
explicitly accounts for the previous visual attentions as the context, and then
decides whether the context is helpful for the current word generation given
the current visual attention. Compared against traditional visual attention
that only fixes a single image region at every step, CAVP can attend to complex
visual compositions over time. The whole image captioning model --- CAVP and
its subsequent language policy network --- can be efficiently optimized
end-to-end by using an actor-critic policy gradient method with respect to any
caption evaluation metric. We demonstrate the effectiveness of CAVP by
state-of-the-art performances on MS-COCO offline split and online server, using
various metrics and sensible visualizations of qualitative visual context. The
code is available at https://github.com/daqingliu/CAVP | [
"cs.CV"
] |
Many real-world applications are associated with structured data, where not
only input but also output has interplay. However, typical classification and
regression models often lack the ability of simultaneously exploring high-order
interaction within input and that within output. In this paper, we present a
deep learning model aiming to generate a powerful nonlinear functional mapping
from structured input to structured output. More specifically, we propose to
integrate high-order hidden units, guided discriminative pretraining, and
high-order auto-encoders for this purpose. We evaluate the model with three
datasets, and obtain state-of-the-art performances among competitive methods.
Our current work focuses on structured output regression, which is a less
explored area, although the model can be extended to handle structured label
classification. | [
"cs.LG",
"cs.NE"
] |
We propose a Reinforcement Learning based approach to approximately solve the
Tree Decomposition (TD) problem. TD is a combinatorial problem, which is
central to the analysis of graph minor structure and computational complexity,
as well as in the algorithms of probabilistic inference, register allocation,
and other practical tasks. Recently, it has been shown that combinatorial
problems can be successively solved by learned heuristics. However, the
majority of existing works do not address the question of the generalization of
learning-based solutions. Our model is based on the graph convolution neural
network (GCN) for learning graph representations. We show that the agent
builton GCN and trained on a single graph using an Actor-Critic method can
efficiently generalize to real-world TD problem instances. We establish that
our method successfully generalizes from small graphs, where TD can be found by
exact algorithms, to large instances of practical interest, while still having
very low time-to-solution. On the other hand, the agent-based approach
surpasses all greedy heuristics by the quality of the solution. | [
"cs.LG",
"stat.ML"
] |
This paper is concerned with a state-space approach to deep Gaussian process
(DGP) regression. We construct the DGP by hierarchically putting transformed
Gaussian process (GP) priors on the length scales and magnitudes of the next
level of Gaussian processes in the hierarchy. The idea of the state-space
approach is to represent the DGP as a non-linear hierarchical system of linear
stochastic differential equations (SDEs), where each SDE corresponds to a
conditional GP. The DGP regression problem then becomes a state estimation
problem, and we can estimate the state efficiently with sequential methods by
using the Markov property of the state-space DGP. The computational complexity
scales linearly with respect to the number of measurements. Based on this, we
formulate state-space MAP as well as Bayesian filtering and smoothing solutions
to the DGP regression problem. We demonstrate the performance of the proposed
models and methods on synthetic non-stationary signals and apply the
state-space DGP to detection of the gravitational waves from LIGO measurements. | [
"stat.ML",
"cs.LG",
"stat.ME"
] |
Sales forecast is an essential task in E-commerce and has a crucial impact on
making informed business decisions. It can help us to manage the workforce,
cash flow and resources such as optimizing the supply chain of manufacturers
etc. Sales forecast is a challenging problem in that sales is affected by many
factors including promotion activities, price changes, and user preferences
etc. Traditional sales forecast techniques mainly rely on historical sales data
to predict future sales and their accuracies are limited. Some more recent
learning-based methods capture more information in the model to improve the
forecast accuracy. However, these methods require case-by-case manual feature
engineering for specific commercial scenarios, which is usually a difficult,
time-consuming task and requires expert knowledge. To overcome the limitations
of existing methods, we propose a novel approach in this paper to learn
effective features automatically from the structured data using the
Convolutional Neural Network (CNN). When fed with raw log data, our approach
can automatically extract effective features from that and then forecast sales
using those extracted features. We test our method on a large real-world
dataset from CaiNiao.com and the experimental results validate the
effectiveness of our method. | [
"cs.LG",
"stat.ML"
] |
We aim to study the modeling limitations of the commonly employed boosted
decision trees classifier. Inspired by the success of large, data-hungry visual
recognition models (e.g. deep convolutional neural networks), this paper
focuses on the relationship between modeling capacity of the weak learners,
dataset size, and dataset properties. A set of novel experiments on the Caltech
Pedestrian Detection benchmark results in the best known performance among
non-CNN techniques while operating at fast run-time speed. Furthermore, the
performance is on par with deep architectures (9.71% log-average miss rate),
while using only HOG+LUV channels as features. The conclusions from this study
are shown to generalize over different object detection domains as demonstrated
on the FDDB face detection benchmark (93.37% accuracy). Despite the impressive
performance, this study reveals the limited modeling capacity of the common
boosted trees model, motivating a need for architectural changes in order to
compete with multi-level and very deep architectures. | [
"cs.CV"
] |
We propose Neural Actor (NA), a new method for high-quality synthesis of
humans from arbitrary viewpoints and under arbitrary controllable poses. Our
method is built upon recent neural scene representation and rendering works
which learn representations of geometry and appearance from only 2D images.
While existing works demonstrated compelling rendering of static scenes and
playback of dynamic scenes, photo-realistic reconstruction and rendering of
humans with neural implicit methods, in particular under user-controlled novel
poses, is still difficult. To address this problem, we utilize a coarse body
model as the proxy to unwarp the surrounding 3D space into a canonical pose. A
neural radiance field learns pose-dependent geometric deformations and pose-
and view-dependent appearance effects in the canonical space from multi-view
video input. To synthesize novel views of high fidelity dynamic geometry and
appearance, we leverage 2D texture maps defined on the body model as latent
variables for predicting residual deformations and the dynamic appearance.
Experiments demonstrate that our method achieves better quality than the
state-of-the-arts on playback as well as novel pose synthesis, and can even
generalize well to new poses that starkly differ from the training poses.
Furthermore, our method also supports body shape control of the synthesized
results. | [
"cs.CV",
"cs.GR",
"cs.LG"
] |
We present INSPIRE, a top-performing general-purpose method for deformable
image registration. INSPIRE extends our existing symmetric registration
framework based on distances combining intensity and spatial information to an
elastic B-splines based transformation model. We also present several
theoretical and algorithmic improvements which provide high computational
efficiency and thereby applicability of the framework in a wide range of real
scenarios. We show that the proposed method delivers both highly accurate as
well as stable and robust registration results. We evaluate the method on a
synthetic dataset created from retinal images, consisting of thin networks of
vessels, where INSPIRE exhibits excellent performance, substantially
outperforming the reference methods. We also evaluate the method on four
benchmark datasets of 3D images of brains, for a total of 2088 pairwise
registrations; a comparison with 15 other state-of-the-art methods reveals that
INSPIRE provides the best overall performance. Code is available at
github.com/MIDA-group/inspire. | [
"cs.CV"
] |
Mobile and ubiquitous sensing of urban air quality has received increased
attention as an economically and operationally viable means to survey
atmospheric environment with high spatial-temporal resolution. This paper
proposes a machine learning based mobile air pollution sensing framework,
called Deep-MAPS, and demonstrates its scientific and financial values in the
following aspects. (1) Based on a network of fixed and mobile air quality
sensors, we perform spatial inference of PM2.5 concentrations in Beijing (3,025
km2, 19 Jun-16 Jul 2018) for a spatial-temporal resolution of 1km-by-1km and 1
hour, with over 85% accuracy. (2) We leverage urban big data to generate
insights regarding the potential cause of pollution, which facilitates
evidence-based sustainable urban management. (3) To achieve such
spatial-temporal coverage and accuracy, Deep-MAPS can save up to 90% hardware
investment, compared with ubiquitous sensing that relies primarily on fixed
sensors. | [
"cs.LG",
"stat.ML"
] |
Human action recognition as an important application of computer vision has
been studied for decades. Among various approaches, skeleton-based methods
recently attract increasing attention due to their robust and superior
performance. However, existing skeleton-based methods ignore the potential
action relationships between different persons, while the action of a person is
highly likely to be impacted by another person especially in complex events. In
this paper, we propose a novel group-skeleton-based human action recognition
method in complex events. This method first utilizes multi-scale
spatial-temporal graph convolutional networks (MS-G3Ds) to extract skeleton
features from multiple persons. In addition to the traditional key point
coordinates, we also input the key point speed values to the networks for
better performance. Then we use multilayer perceptrons (MLPs) to embed the
distance values between the reference person and other persons into the
extracted features. Lastly, all the features are fed into another MS-G3D for
feature fusion and classification. For avoiding class imbalance problems, the
networks are trained with a focal loss. The proposed algorithm is also our
solution for the Large-scale Human-centric Video Analysis in Complex Events
Challenge. Results on the HiEve dataset show that our method can give superior
performance compared to other state-of-the-art methods. | [
"cs.CV"
] |
The spatial convolution layer which is widely used in the Graph Neural
Networks (GNNs) aggregates the feature vector of each node with the feature
vectors of its neighboring nodes. The GNN is not aware of the locations of the
nodes in the global structure of the graph and when the local structures
corresponding to different nodes are similar to each other, the convolution
layer maps all those nodes to similar or same feature vectors in the continuous
feature space. Therefore, the GNN cannot distinguish two graphs if their
difference is not in their local structures. In addition, when the nodes are
not labeled/attributed the convolution layers can fail to distinguish even
different local structures. In this paper, we propose an effective solution to
address this problem of the GNNs. The proposed approach leverages a spatial
representation of the graph which makes the neural network aware of the
differences between the nodes and also their locations in the graph. The
spatial representation which is equivalent to a point-cloud representation of
the graph is obtained by a graph embedding method. Using the proposed approach,
the local feature extractor of the GNN distinguishes similar local structures
in different locations of the graph and the GNN infers the topological
structure of the graph from the spatial distribution of the locally extracted
feature vectors. Moreover, the spatial representation is utilized to simplify
the graph down-sampling problem. A new graph pooling method is proposed and it
is shown that the proposed pooling method achieves competitive or better
results in comparison with the state-of-the-art methods. | [
"cs.LG",
"stat.ML"
] |
The spatio-temporal graph learning is becoming an increasingly important
object of graph study. Many application domains involve highly dynamic graphs
where temporal information is crucial, e.g. traffic networks and financial
transaction graphs. Despite the constant progress made on learning structured
data, there is still a lack of effective means to extract dynamic complex
features from spatio-temporal structures. Particularly, conventional models
such as convolutional networks or recurrent neural networks are incapable of
revealing the temporal patterns in short or long terms and exploring the
spatial properties in local or global scope from spatio-temporal graphs
simultaneously. To tackle this problem, we design a novel multi-scale
architecture, Spatio-Temporal U-Net (ST-UNet), for graph-structured time series
modeling. In this U-shaped network, a paired sampling operation is proposed in
spacetime domain accordingly: the pooling (ST-Pool) coarsens the input graph in
spatial from its deterministic partition while abstracts multi-resolution
temporal dependencies through dilated recurrent skip connections; based on
previous settings in the downsampling, the unpooling (ST-Unpool) restores the
original structure of spatio-temporal graphs and resumes regular intervals
within graph sequences. Experiments on spatio-temporal prediction tasks
demonstrate that our model effectively captures comprehensive features in
multiple scales and achieves substantial improvements over mainstream methods
on several real-world datasets. | [
"cs.LG",
"stat.ML"
] |
Reinforcement learning algorithms need to deal with the exponential growth of
states and actions when exploring optimal control in high-dimensional spaces.
This is known as the curse of dimensionality. By projecting the agent's state
onto a low-dimensional manifold, we can represent the state space in a smaller
and more efficient representation. By using this representation during
learning, the agent can converge to a good policy much faster. We test this
approach in the Mario Benchmarking Domain. When using dimensionality reduction
in Mario, learning converges much faster to a good policy. But, there is a
critical convergence-performance trade-off. By projecting onto a
low-dimensional manifold, we are ignoring important data. In this paper, we
explore this trade-off of convergence and performance. We find that learning in
as few as 4 dimensions (instead of 9), we can improve performance past learning
in the full dimensional space at a faster convergence rate. | [
"cs.LG",
"cs.RO"
] |
Cars can nowadays record several thousands of signals through the CAN bus
technology and potentially provide real-time information on the car, the driver
and the surrounding environment. This paper proposes a new method for the
analysis and classification of driver behavior using a selected subset of CAN
bus signals, specifically gas pedal position, brake pedal pressure, steering
wheel angle, steering wheel momentum, velocity, RPM, frontal and lateral
acceleration. Data has been collected in a completely uncontrolled experiment,
where 64 people drove 10 cars for or a total of over 2000 driving trips without
any type of pre-determined driving instruction on a wide variety of road
scenarios. We propose an unsupervised learning technique that clusters drivers
in different groups, and offers a validation method to test the robustness of
clustering in a wide range of experimental settings. The minimal amount of data
needed to preserve robust driver clustering is also computed. The presented
study provides a new methodology for near-real-time classification of driver
behavior in uncontrolled environments. | [
"cs.LG",
"cs.CY",
"physics.data-an"
] |
Standard methods in deep learning for natural language processing fail to
capture the compositional structure of human language that allows for
systematic generalization outside of the training distribution. However, human
learners readily generalize in this way, e.g. by applying known grammatical
rules to novel words. Inspired by work in neuroscience suggesting separate
brain systems for syntactic and semantic processing, we implement a
modification to standard approaches in neural machine translation, imposing an
analogous separation. The novel model, which we call Syntactic Attention,
substantially outperforms standard methods in deep learning on the SCAN
dataset, a compositional generalization task, without any hand-engineered
features or additional supervision. Our work suggests that separating syntactic
from semantic learning may be a useful heuristic for capturing compositional
structure. | [
"cs.LG",
"cs.CL",
"stat.ML"
] |
Recent years have witnessed the popularity of Graph Neural Networks (GNN) in
various scenarios. To obtain optimal data-specific GNN architectures,
researchers turn to neural architecture search (NAS) methods, which have made
impressive progress in discovering effective architectures in convolutional
neural networks. Two preliminary works, GraphNAS and Auto-GNN, have made first
attempt to apply NAS methods to GNN. Despite the promising results, there are
several drawbacks in expressive capability and search efficiency of GraphNAS
and Auto-GNN due to the designed search space. To overcome these drawbacks, we
propose the SNAG framework (Simplified Neural Architecture search for Graph
neural networks), consisting of a novel search space and a reinforcement
learning based search algorithm. Extensive experiments on real-world datasets
demonstrate the effectiveness of the SNAG framework compared to human-designed
GNNs and NAS methods, including GraphNAS and Auto-GNN. | [
"cs.LG",
"stat.ML"
] |
Cloud segmentation plays a crucial role in image analysis for climate
modeling. Manually labeling the training data for cloud segmentation is
time-consuming and error-prone. We explore to train segmentation networks with
synthetic data due to the natural acquisition of pixel-level labels.
Nevertheless, the domain gap between synthetic and real images significantly
degrades the performance of the trained model. We propose a color space
adaptation method to bridge the gap, by training a color-sensitive generator
and discriminator to adapt synthetic data to real images in color space.
Instead of transforming images by general convolutional kernels, we adopt a set
of closed-form operations to make color-space adjustments while preserving the
labels. We also construct a synthetic-to-real cirrus cloud dataset SynCloud and
demonstrate the adaptation efficacy on the semantic segmentation task of cirrus
clouds. With our adapted synthetic data for training the semantic segmentation,
we achieve an improvement of 6:59% when applied to real images, superior to
alternative methods. | [
"cs.CV"
] |
Time-lapse videos usually contain visually appealing content but are often
difficult and costly to create. In this paper, we present an end-to-end
solution to synthesize a time-lapse video from a single outdoor image using
deep neural networks. Our key idea is to train a conditional generative
adversarial network based on existing datasets of time-lapse videos and image
sequences. We propose a multi-frame joint conditional generation framework to
effectively learn the correlation between the illumination change of an outdoor
scene and the time of the day. We further present a multi-domain training
scheme for robust training of our generative models from two datasets with
different distributions and missing timestamp labels. Compared to alternative
time-lapse video synthesis algorithms, our method uses the timestamp as the
control variable and does not require a reference video to guide the synthesis
of the final output. We conduct ablation studies to validate our algorithm and
compare with state-of-the-art techniques both qualitatively and quantitatively. | [
"cs.CV"
] |
While Gaussian processes (GPs) are the method of choice for regression tasks,
they also come with practical difficulties, as inference cost scales cubic in
time and quadratic in memory. In this paper, we introduce a natural and
expressive way to tackle these problems, by incorporating GPs in sum-product
networks (SPNs), a recently proposed tractable probabilistic model allowing
exact and efficient inference. In particular, by using GPs as leaves of an SPN
we obtain a novel flexible prior over functions, which implicitly represents an
exponentially large mixture of local GPs. Exact and efficient posterior
inference in this model can be done in a natural interplay of the inference
mechanisms in GPs and SPNs. Thereby, each GP is -- similarly as in a mixture of
experts approach -- responsible only for a subset of data points, which
effectively reduces inference cost in a divide and conquer fashion. We show
that integrating GPs into the SPN framework leads to a promising probabilistic
regression model which is: (1) computational and memory efficient, (2) allows
efficient and exact posterior inference, (3) is flexible enough to mix
different kernel functions, and (4) naturally accounts for non-stationarities
in time series. In a variate of experiments, we show that the SPN-GP model can
learn input dependent parameters and hyper-parameters and is on par with or
outperforms the traditional GPs as well as state of the art approximations on
real-world data. | [
"cs.LG",
"stat.ML"
] |
Generative models of natural images have progressed towards high fidelity
samples by the strong leveraging of scale. We attempt to carry this success to
the field of video modeling by showing that large Generative Adversarial
Networks trained on the complex Kinetics-600 dataset are able to produce video
samples of substantially higher complexity and fidelity than previous work. Our
proposed model, Dual Video Discriminator GAN (DVD-GAN), scales to longer and
higher resolution videos by leveraging a computationally efficient
decomposition of its discriminator. We evaluate on the related tasks of video
synthesis and video prediction, and achieve new state-of-the-art Fr\'echet
Inception Distance for prediction for Kinetics-600, as well as state-of-the-art
Inception Score for synthesis on the UCF-101 dataset, alongside establishing a
strong baseline for synthesis on Kinetics-600. | [
"cs.CV",
"cs.LG",
"stat.ML"
] |