Unnamed: 0
int64
0
1.83k
Clean_Title
stringlengths
8
153
Clean_Text
stringlengths
330
2.26k
Clean_Summary
stringlengths
53
295
1,200
Ada-Boundary: Accelerating the DNN Training via Adaptive Boundary Batch Selection
Neural networks can converge faster with help from a smarter batch selection strategy.In this regard, we propose Ada-Boundary, a novel adaptive-batch selection algorithm that constructs an effective mini-batch according to the learning progress of the model.Our key idea is to present confusing samples what the true label is.Thus, the samples near the current decision boundary are considered as the most effective to expedite convergence.Taking advantage of our design, Ada-Boundary maintains its dominance in various degrees of training difficulty.We demonstrate the advantage of Ada-Boundary by extensive experiments using two convolutional neural networks for three benchmark data sets.The experiment results show that Ada-Boundary improves the training time by up to 31.7% compared with the state-of-the-art strategy and by up to 33.5% compared with the baseline strategy.
We suggest a smart batch selection technique called Ada-Boundary.
1,201
Sound event classification using ontology-based neural networks
State of the art sound event classification relies in neural networks to learn the associations between class labels and audio recordings within a dataset.These datasets typically define an ontology to create a structure that relates these sound classes with more abstract super classes.Hence, the ontology serves as a source of domain knowledge representation of sounds.However, the ontology information is rarely considered, and specially under explored to model neural network architectures.We propose two ontology-based neural network architectures for sound event classification.We defined a framework to design simple network architectures that preserve an ontological structure.The networks are trained and evaluated using two of the most common sound event classification datasets.Results show an improvement in classification performance demonstrating the benefits of including the ontological information.
We present ontology-based neural network architectures for sound event classification.
1,202
How much Position Information Do Convolutional Neural Networks Encode?
In contrast to fully connected networks, Convolutional Neural Networks achieve efficiency by learning weights associated with local filters with a finite spatial extent.An implication of this is that a filter may know what it is looking at, but not where it is positioned in the image.Information concerning absolute position is inherently useful, and it is reasonable to assume that deep CNNs may implicitly learn to encode this information if there is a means to do so.In this paper, we test this hypothesis revealing the surprising degree of absolute position information that is encoded in commonly used neural networks.A comprehensive set of experiments show the validity of this hypothesis and shed light on how and where this information is represented while offering clues to where positional information is derived from in deep CNNs.
Our work shows positional information has been implicitly encoded in a network. This information is important for detecting position-dependent features, e.g. semantic and saliency.
1,203
Multi-Task Learning for Semantic Parsing with Cross-Domain Sketch
Semantic parsing which maps a natural language sentence into a formal machine-readable representation of its meaning, is highly constrained by the limited annotated training data.Inspired by the idea of coarse-to-fine, we propose a general-to-detailed neural network by incorporating cross-domain sketch among utterances and their logic forms.For utterances in different domains, the General Network will extract CDS using an encoder-decoder model in a multi-task learning setup.Then for some utterances in a specific domain, the Detailed Network will generate the detailed target parts using sequence-to-sequence architecture with advanced attention to both utterance and generated CDS."Our experiments show that compared to direct multi-task learning, CDS has improved the performance in semantic parsing task which converts users' requests into meaning representation language.", 'We also use experiments to illustrate that CDS works by adding some constraints to the target decoding process, which further proves the effectiveness and rationality of CDS.
General-to-detailed neural network(GDNN) with Multi-Task Learning by incorporating cross-domain sketch(CDS) for semantic parsing
1,204
On Characterizing the Capacity of Neural Networks Using Algebraic Topology
The learnability of different neural architectures can be characterized directly by computable measures of data complexity.In this paper, we reframe the problem of architecture selection as understanding how data determines the most expressive and generalizable architectures suited to that data, beyond inductive bias.After suggesting algebraic topology as a measure for data complexity, we show that the power of a network to express the topological complexity of a dataset in its decision boundary is a strictly limiting factor in its ability to generalize.We then provide the first empirical characterization of the topological capacity of neural networks.Our empirical analysis shows that at every level of dataset complexity, neural networks exhibit topological phase transitions and stratification.This observation allowed us to connect existing theory to empirically driven conjectures on the choice of architectures for a single hidden layer neural networks.
We show that the learnability of different neural architectures can be characterized directly by computable measures of data complexity.
1,205
Augmenting Genetic Algorithms with Deep Neural Networks for Exploring the Chemical Space
Challenges in natural sciences can often be phrased as optimization problems.Machine learning techniques have recently been applied to solve such problems.One example in chemistry is the design of tailor-made organic materials and molecules, which requires efficient methods to explore the chemical space.We present a genetic algorithm that is enhanced with a neural network based discriminator model to improve the diversity of generated molecules and at the same time steer the GA.We show that our algorithm outperforms other generative models in optimization tasks.We furthermore present a way to increase interpretability of genetic algorithms, which helped us to derive design principles
Tackling inverse design via genetic algorithms augmented with deep neural networks.
1,206
How Does BERT Answer Questions? A Layer-Wise Analysis of Transformer Representations
Bidirectional Encoder Representations from Transformers reach state-of-the-art results in a variety of Natural Language Processing tasks.However, understanding of their internal functioning is still insufficient and unsatisfactory."In order to better understand BERT and other Transformer-based models, we present a layer-wise analysis of BERT's hidden states.", 'Unlike previous research, which mainly focuses on explaining Transformer models by their \\hbox weights, we argue that hidden states contain equally valuable information.Specifically, our analysis focuses on models fine-tuned on the task of Question Answering as an example of a complex downstream task.We inspect how QA models transform token vectors in order to find the correct answer.To this end, we apply a set of general and QA-specific probing tasks that reveal the information stored in each representation layer."Our qualitative analysis of hidden state visualizations provides additional insights into BERT's reasoning process.", 'Our results show that the transformations within BERT go through phases that are related to traditional pipeline tasks.The system can therefore implicitly incorporate task-specific information into its token representations."Furthermore, our analysis reveals that fine-tuning has little impact on the models' semantic abilities and that prediction errors can be recognized in the vector representations of even early layers.
We investigate hidden state activations of Transformer Models in Question Answering Tasks.
1,207
Reinforcement and Imitation Learning for Diverse Visuomotor Skills
We propose a general deep reinforcement learning method and apply it to robot manipulation tasks.Our approach leverages demonstration data to assist a reinforcement learning agent in learning to solve a wide range of tasks, mainly previously unsolved.We train visuomotor policies end-to-end to learn a direct mapping from RGB camera inputs to joint velocities.Our experiments indicate that our reinforcement and imitation approach can solve contact-rich robot manipulation tasks that neither the state-of-the-art reinforcement nor imitation learning method can solve alone.We also illustrate that these policies achieved zero-shot sim2real transfer by training with large visual and dynamics variations.
combine reinforcement learning and imitation learning to solve complex robot manipulation tasks from pixels
1,208
BatchEnsemble: an Alternative Approach to Efficient Ensemble and Lifelong Learning
Ensembles, where multiple neural networks are trained individually and their predictions are averaged, have been shown to be widely successful for improving both the accuracy and predictive uncertainty of single neural networks."However, an ensemble's cost for both training and testing increases linearly with the number of networks.", 'In this paper, we propose BatchEnsemble, an ensemble method whose computational and memory costs are significantly lower than typical ensembles.BatchEnsemble achieves this by defining each weight matrix to be the Hadamard product of a shared weight among all ensemble members and a rank-one matrix per member.Unlike ensembles, BatchEnsemble is not only parallelizable across devices, where one device trains one member, but also parallelizable within a device, where multiple ensemble members are updated simultaneously for a given mini-batch.Across CIFAR-10, CIFAR-100, WMT14 EN-DE/EN-FR translation, and contextual bandits tasks, BatchEnsemble yields competitive accuracy and uncertainties as typical ensembles; the speedup at test time is 3X and memory reduction is 3X at an ensemble of size 4.We also apply BatchEnsemble to lifelong learning, where on Split-CIFAR-100, BatchEnsemble yields comparable performance to progressive neural networks while having a much lower computational and memory costs.We further show that BatchEnsemble can easily scale up to lifelong learning on Split-ImageNet which involves 100 sequential learning tasks.
We introduced BatchEnsemble, an efficient method for ensembling and lifelong learning which can be used to improve the accuracy and uncertainty of any neural network like typical ensemble methods.
1,209
Towards Hierarchical Importance Attribution: Explaining Compositional Semantics for Neural Sequence Models
Deep neural networks have achieved impressive performance in handling complicated semantics in natural language, while mostly treated as black boxes.To explain how the model handles compositional semantics of words and phrases, we study the hierarchical explanation problem.We highlight the key challenge is to compute non-additive and context-independent importance for individual words and phrases.We show some prior efforts on hierarchical explanations, e.g. contextual decomposition, do not satisfy the desired properties mathematically, leading to inconsistent explanation quality in different models.In this paper, we propose a formal way to quantify the importance of each word or phrase to generate hierarchical explanations.We modify contextual decomposition algorithms according to our formulation, and propose a model-agnostic explanation algorithm with competitive performance.Human evaluation and automatic metrics evaluation on both LSTM models and fine-tuned BERT Transformer models on multiple datasets show that our algorithms robustly outperform prior works on hierarchical explanations.We show our algorithms help explain compositionality of semantics, extract classification rules, and improve human trust of models.
We propose measurement of phrase importance and algorithms for hierarchical explanation of neural sequence model predictions
1,210
Improving image generative models with human interactions
GANs provide a framework for training generative models which mimic a data distribution.However, in many cases we wish to train a generative model to optimize some auxiliary objective function within the data it generates, such as making more aesthetically pleasing images.In some cases, these objective functions are difficult to evaluate, e.g. they may require human interaction.Here, we develop a system for efficiently training a GAN to increase a generic rate of positive user interactions, for example aesthetic ratings.To do this, we build a model of human behavior in the targeted domain from a relatively small set of interactions, and then use this behavioral model as an auxiliary loss function to improve the generative model.As a proof of concept, we demonstrate that this system is successful at improving positive interaction rates simulated from a variety of objectives, and characterize s
We describe how to improve an image generative model according to a slow- or difficult-to-evaluate objective, such as human feedback, which could have many applications, like making more aesthetic images.
1,211
Selective Self-Training for semi-supervised Learning
Semi-supervised learning is a study that efficiently exploits a large amount of unlabeled data to improve performance in conditions of limited labeled data.Most of the conventional SSL methods assume that the classes of unlabeled data are included in the set of classes of labeled data.In addition, these methods do not sort out useless unlabeled samples and use all the unlabeled data for learning, which is not suitable for realistic situations.In this paper, we propose an SSL method called selective self-training, which selectively decides whether to include each unlabeled sample in the training process.It is also designed to be applied to a more real situation where classes of unlabeled data are different from the ones of the labeled data.For the conventional SSL problems which deal with data where both the labeled and unlabeled samples share the same class categories, the proposed method not only performs comparable to other conventional SSL algorithms but also can be combined with other SSL algorithms.While the conventional methods cannot be applied to the new SSL problems where the separated data do not share the classes, our method does not show any performance degradation even if the classes of unlabeled data are different from those of the labeled data.
Our proposed algorithm does not use all of the unlabeled data for the training, and it rather uses them selectively.
1,212
Latent Constraints: Learning to Generate Conditionally from Unconditional Generative Models
Deep generative neural networks have proven effective at both conditional and unconditional modeling of complex data distributions.Conditional generation enables interactive control, but creating new controls often requires expensive retraining.In this paper, we develop a method to condition generation without retraining the model.By post-hoc learning latent constraints, value functions identify regions in latent space that generate outputs with desired attributes, we can conditionally sample from these regions with gradient-based optimization or amortized actor functions.Combining attribute constraints with a universal “realism” constraint, which enforces similarity to the data distribution, we generate realistic conditional images from an unconditional variational autoencoder.Further, using gradient-based optimization, we demonstrate identity-preserving transformations that make the minimal adjustment in latent space to modify the attributes of an image.Finally, with discrete sequences of musical notes, we demonstrate zero-shot conditional generation, learning latent constraints in the absence of labeled data or a differentiable reward function.
A new approach to conditional generation by constraining the latent space of an unconditional generative model.
1,213
Energy and Policy Considerations for Deep Learning in NLP
Recent progress in hardware and methodology for training neural networks has ushered in a new generation of large networks trained on abundant data.These models have obtained notable gains in accuracy across many NLP tasks.However, these accuracy improvements depend on the availability of exceptionally large computational resources that necessitate similarly substantial energy consumption.As a result these models are costly to train and develop, both financially, due to the cost of hardware and electricity or cloud compute time, and environmentally, due to the carbon footprint required to fuel modern tensor processing hardware.In this paper we bring this issue to the attention of NLP researchers by quantifying the approximate financial and environmental costs of training a variety of recently successful neural network models for NLP.Based on these findings, we propose actionable recommendations to reduce costs and improve equity in NLP research and practice.
We quantify the energy cost in terms of money (cloud credits) and carbon footprint of training recently successful neural network models for NLP. Costs are high.
1,214
PVAE: Learning Disentangled Representations with Intrinsic Dimension via Approximated L0 Regularization
Many models based on the Variational Autoencoder are proposed to achieve disentangled latent variables in inference.However, most current work is focusing on designing powerful disentangling regularizers, while the given number of dimensions for the latent representation at initialization could severely influence the disentanglement.Thus, a pruning mechanism is introduced, aiming at automatically seeking for the intrinsic dimension of the data while promoting disentangled representations.The proposed method is validated on MPI3D and MNIST to be advancing state-of-the-art methods in disentanglement, reconstruction, and robustness.The code is provided on the https://github.com/WeyShi/FYP-of-Disentanglement.
The Pruning VAE is proposed to search for disentangled variables with intrinsic dimension.
1,215
Learning To Generate Reviews and Discovering Sentiment
We explore the properties of byte-level recurrent language models.When given sufficient amounts of capacity, training data, and compute time, the representations learned by these models include disentangled features corresponding to high-level concepts.Specifically, we find a single unit which performs sentiment analysis.These representations, learned in an unsupervised manner, achieve state of the art on the binary subset of the Stanford Sentiment Treebank.They are also very data efficient.When using only a handful of labeled examples, our approach matches the performance of strong baselines trained on full datasets.We also demonstrate the sentiment unit has a direct influence on the generative process of the model.Simply fixing its value to be positive or negative generates samples with the corresponding positive or negative sentiment.
Byte-level recurrent language models learn high-quality domain specific representations of text.
1,216
Revisiting Reweighted Wake-Sleep
Discrete latent-variable models, while applicable in a variety of settings, can often be difficult to learn.Sampling discrete latent variables can result in high-variance gradient estimators for two primary reasons:1) branching on the samples within the model, and2) the lack of a pathwise derivative for the samples.While current state-of-the-art methods employ control-variate schemes for the former and continuous-relaxation methods for the latter, their utility is limited by the complexities of implementing and training effective control-variate schemes and the necessity of evaluating many branch paths in the model.Here, we revisit the Reweighted Wake Sleep algorithm, and through extensive evaluations, show that it circumvents both these issues, outperforming current state-of-the-art methods in learning discrete latent-variable models.Moreover, we observe that, unlike the Importance-weighted Autoencoder, RWS learns better models and inference networks with increasing numbers of particles, and that its benefits extend to continuous latent-variable models as well.Our results suggest that RWS is a competitive, often preferable, alternative for learning deep generative models.
Empirical analysis and explanation of particle-based gradient estimators for approximate inference with deep generative models.
1,217
DeepXML: Scalable & Accurate Deep Extreme Classification for Matching User Queries to Advertiser Bid Phrases
The objective in deep extreme multi-label learning is to jointly learn feature representations and classifiers to automatically tag data points with the most relevant subset of labels from an extremely large label set.Unfortunately, state-of-the-art deep extreme classifiers are either not scalable or inaccurate for short text documents. This paper develops the DeepXML algorithm which addresses both limitations by introducing a novel architecture that splits training of head and tail labels. DeepXML increases accuracyby learning word embeddings on head labels and transferring them through a novel residual connection to data impoverished tail labels; increasing the amount of negative training data available by extending state-of-the-art negative sub-sampling techniques; and re-ranking the set of predicted labels to eliminate the hardest negatives for the original classifier.All of these contributions are implemented efficiently by extending the highly scalable Slice algorithm for pretrained embeddings to learn the proposed DeepXML architecture.As a result, DeepXML could efficiently scale to problems involving millions of labels that were beyond the pale of state-of-the-art deep extreme classifiers as it could be more than 10x faster at training than XML-CNN and AttentionXML.At the same time, DeepXML was also empirically determined to be up to 19% more accurate than leading techniques for matching search engine queries to advertiser bid phrases.
Scalable and accurate deep multi label learning with millions of labels.
1,218
ROBUST ESTIMATION VIA GENERATIVE ADVERSARIAL NETWORKS
Robust estimation under Huber's-contamination model has become an important topic in statistics and theoretical computer science.", "Rate-optimal procedures such as Tukey's median and other estimators based on statistical depth functions are impractical because of their computational intractability.", 'In this paper, we establish an intriguing connection between f-GANs and various depth functions through the lens of f-Learning.Similar to the derivation of f-GAN, we show that these depth functions that lead to rate-optimal robust estimators can all be viewed as variational lower bounds of the total variation distance in the framework of f-Learning.This connection opens the door of computing robust estimators using tools developed for training GANs."In particular, we show that a JS-GAN that uses a neural network discriminator with at least one hidden layer is able to achieve the minimax rate of robust mean estimation under Huber's-contamination model.", 'Interestingly, the hidden layers of the neural net structure in the discriminator class are shown to be necessary for robust estimation.
GANs are shown to provide us a new effective robust mean estimate against agnostic contaminations with both statistical optimality and practical tractability.
1,219
State Space LSTM Models with Particle MCMC Inference
Long Short-Term Memory is one of the most powerful sequence models.Despite the strong performance, however, it lacks the nice interpretability as in state space models.In this paper, we present a way to combine the best of both worlds by introducing State Space LSTM, which generalizes the earlier work of combining topic models with LSTM.However, unlike , we do not make any factorization assumptions in our inference algorithm.We present an efficient sampler based on sequential Monte Carlo method that draws from the joint posterior directly.Experimental results confirms the superiority and stability of this SMC inference algorithm on a variety of domains.
We present State Space LSTM models, a combination of state space models and LSTMs, and propose an inference algorithm based on sequential Monte Carlo.
1,220
From Amortised to Memoised Inference: Combining Wake-Sleep and Variational-Bayes for Unsupervised Few-Shot Program Learning
Given a large database of concepts but only one or a few examples of each, can we learn models for each concept that are not only generalisable, but interpretable?In this work, we aim to tackle this problem through hierarchical Bayesian program induction.We present a novel learning algorithm which can infer concepts as short, generative, stochastic programs, while learning a global prior over programs to improve generalisation and a recognition network for efficient inference.Our algorithm, Wake-Sleep-Remember, combines gradient learning for continuous parameters with neurally-guided search over programs.We show that WSR learns compelling latent programs in two tough symbolic domains: cellular automata and Gaussian process kernels.We also collect and evaluate on a new dataset, Text-Concepts, for discovering structured patterns in natural text data.
We extend the wake-sleep algorithm and use it to learn to learn structured models from few examples,
1,221
DeepProteomics: Protein family classification using Shallow and Deep Networks
The knowledge regarding the function of proteins is necessary as it gives a clear picture of biological processes.Nevertheless, there are many protein sequences found and added to the databases but lacks functional annotation.The laboratory experiments take a considerable amount of time for annotation of the sequences.This arises the need to use computational techniques to classify proteins based on their functions.In our work, we have collected the data from Swiss-Prot containing 40433 proteins which is grouped into 30 families.We pass it to recurrent neural network, long short term memory and gated recurrent unit model and compare it by applying trigram with deep neural network and shallow neural network on the same dataset.Through this approach, we could achieve maximum of around 78% accuracy for the classification of protein families.
Proteins, amino-acid sequences, machine learning, deep learning, recurrent neural network(RNN), long short term memory(LSTM), gated recurrent unit(GRU), deep neural networks
1,222
Robust Spoken Term Detection Automatically Adjusted for a Given Threshold
Spoken term detection is the task of determining whether and where a given word or phrase appears in a given segment of speech.Algorithms for STD are often aimed at maximizing the gap between the scores of positive and negative examples.As such they are focused on ensuring that utterances where the term appears are ranked higher than utterances where the term does not appear.However, they do not determine a detection threshold between the two.In this paper, we propose a new approach for setting an absolute detection threshold for all terms by introducing a new calibrated loss function.The advantage of minimizing this loss function during training is that it aims at maximizing not only the relative ranking scores, but also adjusts the system to use a fixed threshold and thus enhances system robustness and maximizes the detection accuracy rates.We use the new loss function in the structured prediction setting and extend the discriminative keyword spotting algorithm for learning the spoken term detector with a single threshold for all terms.We further demonstrate the effectiveness of the new loss function by applying it on a deep neural Siamese network in a weakly supervised setting for template-based spoken term detection, again with a single fixed threshold.Experiments with the TIMIT, WSJ and Switchboard corpora showed that our approach not only improved the accuracy rates when a fixed threshold was used but also obtained higher Area Under Curve.
Spoken Term Detection, using structured prediction and deep networks, implementing a new loss function that both maximizes AUC and ranks according to a predefined threshold.
1,223
Fair Resource Allocation in Federated Learning
Federated learning involves jointly learning over massively distributed partitions of data generated on remote devices.Naively minimizing an aggregate loss function in such a network may disproportionately advantage or disadvantage some of the devices.In this work, we propose q-Fair Federated Learning, a novel optimization objective inspired by resource allocation strategies in wireless networks that encourages a more fair accuracy distribution across devices in federated networks.To solve q-FFL, we devise a scalable method, q-FedAvg, that can run in federated networks.We validate both the improved fairness and flexibility of q-FFL and the efficiency of q-FedAvg through simulations on federated datasets.
We propose a novel optimization objective that encourages fairness in heterogeneous federated networks, and develop a scalable method to solve it.
1,224
PAIRWISE AUGMENTED GANS WITH ADVERSARIAL RECONSTRUCTION LOSS
We propose a novel autoencoding model called Pairwise Augmented GANs.We train a generator and an encoder jointly and in an adversarial manner.The generator network learns to sample realistic objects.In turn, the encoder network at the same time is trained to map the true data distribution to the prior in latent space.To ensure good reconstructions, we introduce an augmented adversarial reconstruction loss.Here we train a discriminator to distinguish two types of pairs: an object with its augmentation and the one with its reconstruction.We show that such adversarial loss compares objects based on the content rather than on the exact match.We experimentally demonstrate that our model generates samples and reconstructions of quality competitive with state-of-the-art on datasets MNIST, CIFAR10, CelebA and achieves good quantitative results on CIFAR10.
We propose a novel autoencoding model with augmented adversarial reconstruction loss. We intoduce new metric for content-based assessment of reconstructions.
1,225
Bayesian Deep Learning via Stochastic Gradient MCMC with a Stochastic Approximation Adaptation
We propose a robust Bayesian deep learning algorithm to infer complex posteriors with latent variables.Inspired by dropout, a popular tool for regularization and model ensemble, we assign sparse priors to the weights in deep neural networks in order to achieve automatic “dropout” and avoid over-fitting.By alternatively sampling from posterior distribution through stochastic gradient Markov Chain Monte Carlo and optimizing latent variables via stochastic approximation, the trajectory of the target weights is proved to converge to the true posterior distribution conditioned on optimal latent variables.This ensures a stronger regularization on the over-fitted parameter space and more accurate uncertainty quantification on the decisive variables.Simulations from large-p-small-n regressions showcase the robustness of this method when applied to models with latent variables.Additionally, its application on the convolutional neural networks leads to state-of-the-art performance on MNIST and Fashion MNIST datasets and improved resistance to adversarial attacks.
a robust Bayesian deep learning algorithm to infer complex posteriors with latent variables
1,226
Baseline-corrected space-by-time non-negative matrix factorization for decoding single trial population spike trains
Activity of populations of sensory neurons carries stimulus information in both the temporal and the spatial dimensions.This poses the question of how to compactly represent all the information that the population codes carry across all these dimensions."Here, we developed an analytical method to factorize a large number of retinal ganglion cells' spike trains into a robust low-dimensional representation that captures efficiently both their spatial and temporal information.", 'In particular, we extended previously used single-trial space-by-time tensor decomposition based on non-negative matrix factorization to efficiently discount pre-stimulus baseline activity.On data recorded from retinal ganglion cells with strong pre-stimulus baseline, we showed that in situations were the stimulus elicits a strong change in firing rate, our extensions yield a boost in stimulus decoding performance.Our results thus suggest that taking into account the baseline can be important for finding a compact information-rich representation of neural activity.
We extended single-trial space-by-time tensor decomposition based on non-negative matrix factorization to efficiently discount pre-stimulus baseline activity that improves decoding performance on data with non-negligible baselines.
1,227
Generative Adversarial Networks for Extreme Learned Image Compression
We propose a framework for extreme learned image compression based on Generative Adversarial Networks, obtaining visually pleasing images at significantly lower bitrates than previous methods.This is made possible through our GAN formulation of learned compression combined with a generator/decoder which operates on the full-resolution image and is trained in combination with a multi-scale discriminator.Additionally, if a semantic label map of the original image is available, our method can fully synthesize unimportant regions in the decoded image such as streets and trees from the label map, therefore only requiring the storage of the preserved region and the semantic label map.A user study confirms that for low bitrates, our approach is preferred to state-of-the-art methods, even when they use more than double the bits.
GAN-based extreme image compression method using less than half the bits of the SOTA engineered codec while preserving visual quality
1,228
Learning a set of interrelated tasks by using a succession of motor policies for a socially guided intrinsically motivated learner
We propose an active learning algorithmic architecture, capable of organizing its learning process in order to achieve a field of complex tasks by learning sequences of primitive motor policies : Socially Guided Intrinsic Motivation with Procedure Babbling.The learner can generalize over its experience to continuously learn new outcomes, by choosing actively what and how to learn guided by empirical measures of its own progress.In this paper, we are considering the learning of a set of interrelated complex outcomes hierarchically organized.We introduce a new framework called "procedures", which enables the autonomous discovery of how to combine previously learned skills in order to learn increasingly more complex motor policies.Our architecture can actively decide which outcome to focus on and which exploration strategy to apply."Those strategies could be autonomous exploration, or active social guidance, where it relies on the expertise of a human teacher providing demonstrations at the learner's request.", 'We show on a simulated environment that our new architecture is capable of tackling the learning of complex motor policies, to adapt the complexity of its policies to the task at hand.We also show that our "procedures" increases the agent\'s capability to learn complex tasks.
The paper describes a strategic intrinsically motivated learning algorithm which tackles the learning of complex motor policies.
1,229
TPO: TREE SEARCH POLICY OPTIMIZATION FOR CONTINUOUS ACTION SPACES
Monte Carlo Tree Search has achieved impressive results on a range of discrete environments, such as Go, Mario and Arcade games, but it has not yet fulfilled its true potential in continuous domains.In this work, we introduceTPO, a tree search based policy optimization method for continuous environments.TPO takes a hybrid approach to policy optimization. Building the MCTS tree in a continuous action space and updating the policy gradient using off-policy MCTS trajectories are non-trivial.To overcome these challenges, we propose limiting tree search branching factor by drawing only few action samples from the policy distribution and define a new loss function based on the trajectories’ mean and standard deviations. Our approach led to some non-intuitive findings. MCTS training generally requires a large number of samples and simulations.However, we observed that bootstrappingtree search with a pre-trained policy allows us to achieve high quality results with a low MCTS branching factor and few number of simulations.Without the proposed policy bootstrapping, continuous MCTS would require a much larger branching factor and simulation count, rendering it computationally and prohibitively expensive.In our experiments, we use PPO as our baseline policy optimization algorithm.TPO significantly improves the policy on nearly all of our benchmarks. For example, in complex environments such as Humanoid, we achieve a 2.5×improvement over the baseline algorithm.
We use MCTS to further optimize a bootstrapped policy for continuous action spaces under a policy iteration setting.
1,230
On Variational Learning of Controllable Representations for Text without Supervision
The variational autoencoder has found success in modelling the manifold of natural images on certain datasets, allowing meaningful images to be generated while interpolating or extrapolating in the latent code space, but it is unclear whether similar capabilities are feasible for text considering its discrete nature.In this work, we investigate the reason why unsupervised learning of controllable representations fails for text.We find that traditional sequence VAEs can learn disentangled representations through their latent codes to some extent, but they often fail to properly decode when the latent factor is being manipulated, because the manipulated codes often land in holes or vacant regions in the aggregated posterior latent space, which the decoding network is not trained to process.Both as a validation of the explanation and as a fix to the problem, we propose to constrain the posterior mean to a learned probability simplex, and performs manipulation within this simplex.Our proposed method mitigates the latent vacancy problem and achieves the first success in unsupervised learning of controllable representations for text.Empirically, our method significantly outperforms unsupervised baselines and is competitive with strong supervised approaches on text style transfer.Furthermore, when switching the latent factor during a long sentence generation, our proposed framework can often complete the sentence in a seemingly natural way -- a capability that has never been attempted by previous methods.
why previous VAEs on text cannot learn controllable latent representation as on images, as well as a fix to enable the first success towards controlled text generation without supervision
1,231
A Model Cortical Network for Spatiotemporal Sequence Learning and Prediction
In this paper we developed a hierarchical network model, called Hierarchical Prediction Network to understand how spatiotemporal memories might be learned and encoded in a representational hierarchy for predicting future video frames.The model is inspired by the feedforward, feedback and lateral recurrent circuits in the mammalian hierarchical visual system.It assumes that spatiotemporal memories are encoded in the recurrent connections within each level and between different levels of the hierarchy.The model contains a feed-forward path that computes and encodes spatiotemporal features of successive complexity and a feedback path that projects interpretation from a higher level to the level below."Within each level, the feed-forward path and the feedback path intersect in a recurrent gated circuit that integrates their signals as well as the circuit's internal memory states to generate a prediction of the incoming signals.", 'The network learns by comparing the incoming signals with its prediction, updating its internal model of the world by minimizing the prediction errors at each level of the hierarchy in the style of. The network processes data in blocks of video frames rather than a frame-to-frame basis. This allows it to learn relationships among movement patterns, yielding state-of-the-art performance in long range video sequence predictions in benchmark datasets.We observed that hierarchical interaction in the network introduces sensitivity to memories of global movement patterns even in the population representation of the units in the earliest level.Finally, we provided neurophysiological evidence, showing that neurons in the early visual cortex of awake monkeys exhibit very similar sensitivity and behaviors.These findings suggest that predictive self-supervised learning might be an important principle for representational learning in the visual cortex.
A new hierarchical cortical model for encoding spatiotemporal memory and video prediction
1,232
Exploratory Not Explanatory: Counterfactual Analysis of Saliency Maps for Deep RL
Saliency maps are often used to suggest explanations of the behavior of deep rein- forcement learning agents.However, the explanations derived from saliency maps are often unfalsifiable and can be highly subjective.We introduce an empirical approach grounded in counterfactual reasoning to test the hypotheses generated from saliency maps and show that explanations suggested by saliency maps are often not supported by experiments.Our experiments suggest that saliency maps are best viewed as an exploratory tool rather than an explanatory tool.
Proposing a new counterfactual-based methodology to evaluate the hypotheses generated from saliency maps about deep RL agent behavior.
1,233
All Neural Networks are Created Equal
One of the unresolved questions in deep learning is the nature of the solutions that are being discovered.We investigate the collection of solutions reached by the same network architecture, with different random initialization of weights and random mini-batches.These solutions are shown to be rather similar - more often than not, each train and test example is either classified correctly by all the networks, or by none at all. Surprisingly, all the network instances seem to share the same learning dynamics, whereby initially the same train and test examples are correctly recognized by the learned model, followed by other examples which are learned in roughly the same order.When extending the investigation to heterogeneous collections of neural network architectures, once again examples are seen to be learned in the same order irrespective of architecture, although the more powerful architecture may continue to learn and thus achieve higher accuracy.This pattern of results remains true even when the composition of classes in the test set is unrelated to the train set, for example, when using out of sample natural images or even artificial images.To show the robustness of these phenomena we provide an extensive summary of our empirical study, which includes hundreds of graphs describing tens of thousands of networks with varying NN architectures, hyper-parameters and domains.We also discuss cases where this pattern of similarity breaks down, which show that the reported similarity is not an artifact of optimization by gradient descent.Rather, the observed pattern of similarity is characteristic of learning complex problems with big networks.Finally, we show that this pattern of similarity seems to be strongly correlated with effective generalization.
Most neural networks approximate the same classification function, even across architectures, through all stages of learning.
1,234
Wasserstein is all you need
We propose a unified framework for building unsupervised representations of individual objects or entities, by associating with each object both a distributional as well as a point estimate.This is made possible by the use of optimal transport, which allows us to build these associated estimates while harnessing the underlying geometry of the ground space.Our method gives a novel perspective for building rich and powerful feature representations that simultaneously capture uncertainty and interpretability.As a guiding example, we formulate unsupervised representations for text, in particular for sentence representation and entailment detection.Empirical results show strong advantages gained through the proposed framework.This approach can be used for any unsupervised or supervised problem with a co-occurrence structure, such as any sequence data.The key tools underlying the framework are Wasserstein distances and Wasserstein barycenters.
Represent each entity based on its histogram of contexts and then Wasserstein is all you need!
1,235
Interpretable Robust Recommender Systems with Side Information
In this paper, we propose two methods, namely Trace-norm regression and Stable Trace-norm Analysis, to improve performances of recommender systems with side information.Our trace-norm regression approach extracts low-rank latent factors underlying the side information that drives user preference under different context.Furthermore, our novel recommender framework StaTNA not only captures latent low-rank common drivers for user preferences, but also considers idiosyncratic taste for individual users.We compare performances of TNR and StaTNA on the MovieLens datasets against state-of-the-art models, and demonstrate that StaTNA and TNR in general outperforms these methods.
Methodologies for recommender systems with side information based on trace-norm regularization
1,236
Exploration Based Language Learning for Text-Based Games
This work presents an exploration and imitation-learning-based agent capable of state-of-the-art performance in playing text-based computer games.Text-based computer games describe their world to the player through natural language and expect the player to interact with the game using text.These games are of interest as they can be seen as a testbed for language understanding, problem-solving, and language generation by artificial agents.Moreover, they provide a learning environment in which these skills can be acquired through interactions with an environment rather than using fixed corpora.One aspect that makes these games particularly challenging for learning agents is the combinatorially large action space.Existing methods for solving text-based games are limited to games that are either very simple or have an action space restricted to a predetermined set of admissible actions.In this work, we propose to use the exploration approach of Go-Explore for solving text-based games.More specifically, in an initial exploration phase, we first extract trajectories with high rewards, after which we train a policy to solve the game by imitating these trajectories.Our experiments show that this approach outperforms existing solutions in solving text-based games, and it is more sample efficient in terms of the number of interactions with the environment.Moreover, we show that the learned policy can generalize better than existing solutions to unseen games without using any restriction on the action space.
This work presents an exploration and imitation-learning-based agent capable of state-of-the-art performance in playing text-based computer games.
1,237
Deconstructing Lottery Tickets: Zeros, Signs, and the Supermask
The recent “Lottery Ticket Hypothesis” paper by Frankle & Carbin showed that a simple approach to creating sparse networks results in models that are trainable from scratch, but only when starting from the same initial weights.The performance of these networks often exceeds the performance of the non-sparse base model, but for reasons that were not well understood.In this paper we study the three critical components of the Lottery Ticket algorithm, showing that each may be varied significantly without impacting the overall results.Ablating these factors leads to new insights for why LT networks perform as well as they do.We show why setting weights to zero is important, how signs are all you need to make the re-initialized network train, and why masking behaves like training.Finally, we discover the existence of Supermasks, or masks that can be applied to an untrained, randomly initialized network to produce a model with performance far better than chance.
In neural network pruning, zeroing pruned weights is important, sign of initialization is key, and masking can be thought of as training.
1,238
SesameBERT: Attention for Anywhere
Fine-tuning with pre-trained models has achieved exceptional results for many language tasks.In this study, we focused on one such self-attention network model, namely BERT, which has performed well in terms of stacking layers across diverse language-understanding benchmarks.However, in many downstream tasks, information between layers is ignored by BERT for fine-tuning.In addition, although self-attention networks are well-known for their ability to capture global dependencies, room for improvement remains in terms of emphasizing the importance of local contexts.In light of these advantages and disadvantages, this paper proposes SesameBERT, a generalized fine-tuning method that enables the extraction of global information among all layers through Squeeze and Excitation and enriches local information by capturing neighboring contexts via Gaussian blurring.Furthermore, we demonstrated the effectiveness of our approach in the HANS dataset, which is used to determine whether models have adopted shallow heuristics instead of learning underlying generalizations.The experiments revealed that SesameBERT outperformed BERT with respect to GLUE benchmark and the HANS evaluation set.
We proposed SesameBERT, a generalized fine-tuning method that enables the extraction of global information among all layers through Squeeze and Excitation and enriches local information by capturing neighboring contexts via Gaussian blurring.
1,239
Stable Opponent Shaping in Differentiable Games
A growing number of learning methods are actually differentiable games whose players optimise multiple, interdependent objectives in parallel – from GANs and intrinsic curiosity to multi-agent RL.Opponent shaping is a powerful approach to improve learning dynamics in these games, accounting for player influence on others’ updates.Learning with Opponent-Learning Awareness is a recent algorithm that exploits this response and leads to cooperation in settings like the Iterated Prisoner’s Dilemma.Although experimentally successful, we show that LOLA agents can exhibit ‘arrogant’ behaviour directly at odds with convergence.In fact, remarkably few algorithms have theoretical guarantees applying across all games.In this paper we present Stable Opponent Shaping, a new method that interpolates between LOLA and a stable variant named LookAhead.We prove that LookAhead converges locally to equilibria and avoids strict saddles in all differentiable games.SOS inherits these essential guarantees, while also shaping the learning of opponents and consistently either matching or outperforming LOLA experimentally.
Opponent shaping is a powerful approach to multi-agent learning but can prevent convergence; our SOS algorithm fixes this with strong guarantees in all differentiable games.
1,240
Domain-Relevant Embeddings for Question Similarity
The rate at which medical questions are asked online significantly exceeds the capacity of qualified people to answer them, leaving many questions unanswered or inadequately answered.Many of these questions are not unique, and reliable identification of similar questions would enable more efficient and effective question answering schema.While many research efforts have focused on the problem of general question similarity, these approaches do not generalize well to the medical domain, where medical expertise is often required to determine semantic similarity.In this paper, we show how a semi-supervised approach of pre-training a neural network on medical question-answer pairs is a particularly useful intermediate task for the ultimate goal of determining medical question similarity.While other pre-training tasks yield an accuracy below 78.7% on this task, our model achieves an accuracy of 82.6% with the same number of training examples, an accuracy of 80.0% with a much smaller training set, and an accuracy of 84.5% when the full corpus of medical question-answer data is used.
We show that question-answer matching is a particularly good pre-training task for question-similarity and release a dataset for medical question similarity
1,241
Sharing Knowledge in Multi-Task Deep Reinforcement Learning
We study the benefit of sharing representations among tasks to enable the effective use of deep neural networks in Multi-Task Reinforcement Learning.We leverage the assumption that learning from different tasks, sharing common properties, is helpful to generalize the knowledge of them resulting in a more effective feature extraction compared to learning a single task.Intuitively, the resulting set of features offers performance benefits when used by Reinforcement Learning algorithms.We prove this by providing theoretical guarantees that highlight the conditions for which is convenient to share representations among tasks, extending the well-known finite-time bounds of Approximate Value-Iteration to the multi-task setting.In addition, we complement our analysis by proposing multi-task extensions of three Reinforcement Learning algorithms that we empirically evaluate on widely used Reinforcement Learning benchmarks showing significant improvements over the single-task counterparts in terms of sample efficiency and performance.
A study on the benefit of sharing representation in Multi-Task Reinforcement Learning.
1,242
Quaternion Equivariant Capsule Networks for 3D Point Clouds
We present a 3D capsule architecture for processing of point clouds that is equivariant with respect to the SO rotation group, translation and permutation of the unordered input sets.The network operates on a sparse set of local reference frames, computed from an input point cloud and establishes end-to-end equivariance through a novel 3D quaternion group capsule layer, including an equivariant dynamic routing procedure.The capsule layer enables us to disentangle geometry from pose, paving the way for more informative descriptions and a structured latent space.In the process, we theoretically connect the process of dynamic routing between capsules to the well-known Weiszfeld algorithm, a scheme for solving iterative re-weighted least squares problems with provable convergence properties, enabling robust pose estimation between capsule layers.Due to the sparse equivariant quaternion capsules, our architecture allows joint object classification and orientation estimation, which we validate empirically on common benchmark datasets.
Deep architectures for 3D point clouds that are equivariant to SO(3) rotations, as well as translations and permutations.
1,243
Exploring Sentence Vectors Through Automatic Summarization
Vector semantics, especially sentence vectors, have recently been used successfully in many areas of natural language processing.However, relatively little work has explored the internal structure and properties of spaces of sentence vectors.In this paper, we will explore the properties of sentence vectors by studying a particular real-world application: Automatic Summarization.In particular, we show that cosine similarity between sentence vectors and document vectors is strongly correlated with sentence importance and that vector semantics can identify and correct gaps between the sentences chosen so far and the document.In addition, we identify specific dimensions which are linked to effective summaries.To our knowledge, this is the first time specific dimensions of sentence embeddings have been connected to sentence properties.We also compare the features of different methods of sentence embeddings.Many of these insights have applications in uses of sentence embeddings far beyond summarization.
A comparison and detailed analysis of various sentence embedding models through the real-world task of automatic summarization.
1,244
Value Propagation Networks
We present Value Propagation, a parameter-efficient differentiable planning module built on Value Iteration which can successfully be trained in a reinforcement learning fashion to solve unseen tasks, has the capability to generalize to larger map sizes, and can learn to navigate in dynamic environments.We evaluate on configurations of MazeBase grid-worlds, with randomly generated environments of several different sizes.Furthermore, we show that the module enables to learn to plan when the environment also includes stochastic elements, providing a cost-efficient learning system to build low-level size-invariant planners for a variety of interactive navigation problems.
We propose Value Propagation, a novel end-to-end planner which can learn to solve 2D navigation tasks via Reinforcement Learning, and that generalizes to larger and dynamic environments.
1,245
Extracting and Leveraging Feature Interaction Interpretations
Recommendation is a prevalent application of machine learning that affects many users; therefore, it is crucial for recommender models to be accurate and interpretable.In this work, we propose a method to both interpret and augment the predictions of black-box recommender systems.In particular, we propose to extract feature interaction interpretations from a source recommender model and explicitly encode these interactions in a target recommender model, where both source and target models are black-boxes.By not assuming the structure of the recommender system, our approach can be used in general settings. In our experiments, we focus on a prominent use of machine learning recommendation: ad-click prediction.We found that our interaction interpretations are both informative and predictive, i.e., significantly outperforming existing recommender models."What's more, the same approach to interpreting interactions can provide new insights into domains even beyond recommendation.
Proposed a method to extract and leverage interpretations of feature interactions
1,246
Learning Network Parameters in the ReLU Model
Rectified linear units, or ReLUs, have become a preferred activation function for artificial neural networks.In this paper we consider the problem of learning a generative model in the presence of nonlinearity.Given a set of signal vectors, we aim to learn the network parameters, i.e., the matrix, under the model, where is a random bias vector, and^i \\in \\mathbb^kAO\\mathbf$.
We show that it is possible to recover the parameters of a 1-layer ReLU generative model from looking at samples generated by it
1,247
Feat2Vec: Dense Vector Representation for Data with Arbitrary Features
Methods that calculate dense vector representations for features in unstructured data—such as words in a document—have proven to be very successful for knowledge representation.We study how to estimate dense representations when multiple feature types exist within a dataset for supervised learning where explicit labels are available, as well as for unsupervised learning where there are no labels.Feat2Vec calculates embeddings for data with multiple feature types enforcing that all different feature types exist in a common space.In the supervised case, we show that our method has advantages over recently proposed methods; such as enabling higher prediction accuracy, and providing a way to avoid the cold-startproblem.In the unsupervised case, our experiments suggest that Feat2Vec significantly outperforms existing algorithms that do not leverage the structure of the data.We believe that we are the first to propose a method for learning unsuper vised embeddings that leverage the structure of multiple feature types.
Learn dense vector representations of arbitrary types of features in labeled and unlabeled datasets
1,248
Representing Formal Languages: A Comparison Between Finite Automata and Recurrent Neural Networks
We investigate the internal representations that a recurrent neural network uses while learning to recognize a regular formal language.Specifically, we train a RNN on positive and negative examples from a regular language, and ask if there is a simple decoding function that maps states of this RNN to states of the minimal deterministic finite automaton for the language."Our experiments show that such a decoding function indeed exists, and that it maps states of the RNN not to MDFA states, but to states of an obtained by clustering small sets of MDFA states into superstates.", 'A qualitative analysis reveals that the abstraction often has a simple interpretation.Overall, the results suggest a strong structural relationship between internal representations used by RNNs and finite automata, and explain the well-known ability of RNNs to recognize formal grammatical structure.
Finite Automata Can be Linearly decoded from Language-Recognizing RNNs using low coarseness abstraction functions and high accuracy decoders.
1,249
Hardware-aware One-Shot Neural Architecture Search in Coordinate Ascent Framework
Designing accurate and efficient convolutional neural architectures for vast amount of hardware is challenging because hardware designs are complex and diverse.This paper addresses the hardware diversity challenge in Neural Architecture Search.Unlike previous approaches that apply search algorithms on a small, human-designed search space without considering hardware diversity, we propose HURRICANE that explores the automatic hardware-aware search over a much larger search space and a multistep search scheme in coordinate ascent framework, to generate tailored models for different types of hardware.Extensive experiments on ImageNet show that our algorithm consistently achieves a much lower inference latency with a similar or better accuracy than state-of-the-art NAS methods on three types of hardware.Remarkably, HURRICANE achieves a 76.63% top-1 accuracy on ImageNet with a inference latency of only 16.5 ms for DSP, which is a 3.4% higher accuracy and a 6.35x inference speedup than FBNet-iPhoneX.For VPU, HURRICANE achieves a 0.53% higher top-1 accuracy than Proxyless-mobile with a 1.49x speedup.Even for well-studied mobile CPU, HURRICANE achieves a 1.63% higher top-1 accuracy than FBNet-iPhoneX with a comparable inference latency.HURRICANE also reduces the training time by 54.7% on average compared to SinglePath-Oneshot.
We propose HURRICANE to address the challenge of hardware diversity in one-shot neural architecture search
1,250
LDMGAN: Reducing Mode Collapse in GANs with Latent Distribution Matching
Generative Adversarial Networks have shown impressive results in modeling distributions over complicated manifolds such as those of natural images.However, GANs often suffer from mode collapse, which means they are prone to characterize only a single or a few modes of the data distribution.In order to address this problem, we propose a novel framework called LDMGAN.We first introduce Latent Distribution Matching constraint which regularizes the generator by aligning distribution of generated samples with that of real samples in latent space.To make use of such latent space, we propose a regularized AutoEncoder that maps the data distribution to prior distribution in encoded space.Extensive experiments on synthetic data and real world datasets show that our proposed framework significantly improves GAN’s stability and diversity.
We propose an AE-based GAN that alleviates mode collapse in GANs.
1,251
Chameleon: Adaptive Code Optimization for Expedited Deep Neural Network Compilation
Achieving faster execution with shorter compilation time can foster further diversity and innovation in neural networks.However, the current paradigm of executing neural networks either relies on hand-optimized libraries, traditional compilation heuristics, or very recently genetic algorithms and other stochastic methods.These methods suffer from frequent costly hardware measurements rendering them not only too time consuming but also suboptimal.As such, we devise a solution that can learn to quickly adapt to a previously unseen design space for code optimization, both accelerating the search and improving the output performance.This solution dubbed CHAMELEON leverages reinforcement learning whose solution takes fewer steps to converge, and develops an adaptive sampling algorithm that not only focuses on the costly samples on representative points but also uses a domain knowledge inspired logic to improve the samples itself.Experimentation with real hardware shows that CHAMELEON provides 4.45×speed up in optimization time over AutoTVM, while also improving inference time of the modern deep networks by 5.6%.
Reinforcement learning and Adaptive Sampling for Optimized Compilation of Deep Neural Networks.
1,252
Learning Adversarial Grammars for Future Prediction
In this paper, we propose a differentiable adversarial grammar model for future prediction.The objective is to model a formal grammar in terms of differentiable functions and latent representations, so that their learning is possible through standard backpropagation.Learning a formal grammar represented with latent terminals, non-terminals, and productions rules allows capturing sequential structures with multiple possibilities from data.The adversarial grammar is designed so that it can learn stochastic production rules from the data distribution.Being able to select multiple production rules leads to different predicted outcomes, thus efficiently modeling many plausible futures. We confirm the benefit of the adversarial grammar on two diverse tasks: future 3D human pose prediction and future activity prediction.For all settings, the proposed adversarial grammar outperforms the state-of-the-art approaches, being able to predict much more accurately and further in the future, than prior work.
We design a grammar that is learned in an adversarial setting and apply it to future prediction in video.
1,253
Janossy Pooling: Learning Deep Permutation-Invariant Functions for Variable-Size Inputs
We consider a simple and overarching representation for permutation-invariant functions of sequences.Our approach, which we call Janossy pooling, expresses a permutation-invariant function as the average of a permutation-sensitive function applied to all reorderings of the input sequence.This allows us to leverage the rich and mature literature on permutation-sensitive functions to construct novel and flexible permutation-invariant functions.If carried out naively, Janossy pooling can be computationally prohibitive.To allow computational tractability, we consider three kinds of approximations: canonical orderings of sequences, functions with k-order interactions, and stochastic optimization algorithms with random permutations.Our framework unifies a variety of existing work in the literature, and suggests possible modeling and algorithmic extensions.We explore a few in our experiments, which demonstrate improved performance over current state-of-the-art methods.
We propose Janossy pooling, a method for learning deep permutation invariant functions designed to exploit relationships within the input sequence and tractable inference strategies such as a stochastic optimization procedure we call piSGD
1,254
Learning to Balance: Bayesian Meta-Learning for Imbalanced and Out-of-distribution Tasks
While tasks could come with varying the number of instances and classes in realistic settings, the existing meta-learning approaches for few-shot classification assume that number of instances per task and class is fixed.Due to such restriction, they learn to equally utilize the meta-knowledge across all the tasks, even when the number of instances per task and class largely varies.Moreover, they do not consider distributional difference in unseen tasks, on which the meta-knowledge may have less usefulness depending on the task relatedness.To overcome these limitations, we propose a novel meta-learning model that adaptively balances the effect of the meta-learning and task-specific learning within each task.Through the learning of the balancing variables, we can decide whether to obtain a solution by relying on the meta-knowledge or task-specific learning.We formulate this objective into a Bayesian inference framework and tackle it using variational inference.We validate our Bayesian Task-Adaptive Meta-Learning on two realistic task- and class-imbalanced datasets, on which it significantly outperforms existing meta-learning approaches.Further ablation study confirms the effectiveness of each balancing component and the Bayesian learning framework.
A novel meta-learning model that adaptively balances the effect of the meta-learning and task-specific learning, and also class-specific learning within each task.
1,255
Revisiting The Master-Slave Architecture In Multi-Agent Deep Reinforcement Learning
Many tasks in artificial intelligence require the collaboration of multiple agents.We exam deep reinforcement learning for multi-agent domains.Recent research efforts often take the form of two seemingly conflicting perspectives, the decentralized perspective, where each agent is supposed to have its own controller; and the centralized perspective, where one assumes there is a larger model controlling all agents.In this regard, we revisit the idea of the master-slave architecture by incorporating both perspectives within one framework.Such a hierarchical structure naturally leverages advantages from one another.The idea of combining both perspective is intuitive and can be well motivated from many real world systems, however, out of a variety of possible realizations, we highlights three key ingredients, i.e. composed action representation, learnable communication and independent reasoning.With network designs to facilitate these explicitly, our proposal consistently outperforms latest competing methods both in synthetics experiments and when applied to challenging StarCraft micromanagement tasks.
We revisit the idea of the master-slave architecture in multi-agent deep reinforcement learning and outperforms state-of-the-arts.
1,256
When Will Gradient Methods Converge to Max-margin Classifier under ReLU Models?
We study the implicit bias of gradient descent methods in solving a binary classification problem over a linearly separable dataset.The classifier is described by a nonlinear ReLU model and the objective function adopts the exponential loss function.We first characterize the landscape of the loss function and show that there can exist spurious asymptotic local minima besides asymptotic global minima.We then show that gradient descent can converge to either a global or a local max-margin direction, or may diverge from the desired max-margin direction in a general context.For stochastic gradient descent, we show that it converges in expectation to either the global or the local max-margin direction if SGD converges.We further explore the implicit bias of these algorithms in learning a multi-neuron network under certain stationary conditions, and show that the learned classifier maximizes the margins of each sample pattern partition under the ReLU activation.
We study the implicit bias of gradient methods in solving a binary classification problem with nonlinear ReLU models.
1,257
Rethinking the Smaller-Norm-Less-Informative Assumption in Channel Pruning of Convolution Layers
Model pruning has become a useful technique that improves the computational efficiency of deep learning, making it possible to deploy solutions in resource-limited scenarios.A widely-used practice in relevant work assumes that a smaller-norm parameter or feature plays a less informative role at the inference time.In this paper, we propose a channel pruning technique for accelerating the computations of deep convolutional neural networks that does not critically rely on this assumption.Instead, it focuses on direct simplification of the channel-to-channel computation graph of a CNN without the need of performing a computationally difficult and not-always-useful task of making high-dimensional tensors of CNN structured sparse.Our approach takes two stages: first to adopt an end-to-end stochastic training method that eventually forces the outputs of some channels to be constant, and then to prune those constant channels from the original neural network by adjusting the biases of their impacting layers such that the resulting compact model can be quickly fine-tuned.Our approach is mathematically appealing from an optimization perspective and easy to reproduce.We experimented our approach through several image learning benchmarks and demonstrate its interest- ing aspects and competitive performance.
A CNN model pruning method using ISTA and rescaling trick to enforce sparsity of scaling parameters in batch normalization.
1,258
Deep Gradient Boosting -- Layer-wise Input Normalization of Neural Networks
Stochastic gradient descent has been the dominant optimization method for training deep neural networks due to its many desirable properties.One of the more remarkable and least understood quality of SGD is that it generalizes relatively wellon unseen data even when the neural network has millions of parameters.We hypothesize that in certain cases it is desirable to relax its intrinsic generalization properties and introduce an extension of SGD called deep gradient boosting.The key idea of DGB is that back-propagated gradients inferred using the chain rule can be viewed as pseudo-residual targets of a gradient boosting problem.Thus at each layer of a neural network the weight update is calculated by solving the corresponding boosting problem using a linear base learner.The resulting weight update formula can also be viewed as a normalization procedure of the data that arrives at each layer during the forward pass.When implemented as a separate input normalization layer the new architecture shows improved performance on image recognition tasks when compared to the same architecture without normalization layers.As opposed to batch normalization, INN has no learnable parameters however it matches its performance on CIFAR10 and ImageNet classification tasks.
What can we learn about training neural networks if we treat each layer as a gradient boosting problem?
1,259
Measuring and regularizing networks in function space
To optimize a neural network one often thinks of optimizing its parameters, but it is ultimately a matter of optimizing the function that maps inputs to outputs.Since a change in the parameters might serve as a poor proxy for the change in the function, it is of some concern that primacy is given to parameters but that the correspondence has not been tested.Here, we show that it is simple and computationally feasible to calculate distances between functions in a Hilbert space.We examine how typical networks behave in this space, and compare how parameter distances compare to function distances between various points of an optimization trajectory.We find that the two distances are nontrivially related.In particular, the ratio decreases throughout optimization, reaching a steady value around when test error plateaus.We then investigate how the distance could be applied directly to optimization.We first propose that in multitask learning, one can avoid catastrophic forgetting by directly limiting how much the input/output function changes between tasks.Secondly, we propose a new learning rule that constrains the distance a network can travel through-space in any one update.This allows new examples to be learned in a way that minimally interferes with what has previously been learned.These applications demonstrate how one can measure and regularize function distances directly, without relying on parameters or local approximations like loss curvature.
We find movement in function space is not proportional to movement in parameter space during optimization. We propose a new natural-gradient style optimizer to address this.
1,260
The Convex Information Bottleneck Lagrangian
The information bottleneck problem tackles the issue of obtaining relevant compressed representations T of some random variable X for the task of predicting Y. It is defined as a constrained optimization problem which maximizes the information the representation has about the task, I, while ensuring that a minimum level of compression r is achieved <= r).For practical reasons the problem is usually solved by maximizing the IB Lagrangian for many values of the Lagrange multiplier, therefore drawing the IB curve for a given I) and selecting the representation of desired predictability and compression.It is known when Y is a deterministic function of X, the IB curve cannot be explored and other Lagrangians have been proposed to tackle this problem.In this paper we present a general family of Lagrangians which allow for the exploration of the IB curve in all scenarios; prove that if these Lagrangians are used, there is a one-to-one mapping between the Lagrange multiplier and the desired compression rate r for known IB curve shapes, hence, freeing from the burden of solving the optimization problem for many values of the Lagrange multiplier.
We introduce a general family of Lagrangians that allow exploring the IB curve in all scenarios. When these are used, and the IB curve is known, one can optimize directly for a performance/compression level directly.
1,261
Neural Logic Machines
We propose the Neural Logic Machine, a neural-symbolic architecture for both inductive learning and logic reasoning.NLMs exploit the power of both neural networks---as function approximators, and logic programming---as a symbolic processor for objects with properties, relations, logic connectives, and quantifiers. After being trained on small-scale tasks, NLMs can recover lifted rules, and generalize to large-scale tasks.In our experiments, NLMs achieve perfect generalization in a number of tasks, from relational reasoning tasks on the family tree and general graphs, to decision making tasks including sorting arrays, finding shortest paths, and playing the blocks world.Most of these tasks are hard to accomplish for neural networks or inductive logic programming alone.
We propose the Neural Logic Machine (NLM), a neural-symbolic architecture for both inductive learning and logic reasoning.
1,262
Robust Neural Abstractive Summarization Systems and Evaluation against Adversarial Information
Sequence-to-sequence neural models have been actively investigated for abstractive summarization.Nevertheless, existing neural abstractive systems frequently generate factually incorrect summaries and are vulnerable to adversarial information, suggesting a crucial lack of semantic understanding.In this paper, we propose a novel semantic-aware neural abstractive summarization model that learns to generate high quality summaries through semantic interpretation over salient content.A novel evaluation scheme with adversarial samples is introduced to measure how well a model identifies off-topic information, where our model yields significantly better performance than the popular pointer-generator summarizer.Human evaluation also confirms that our system summaries are uniformly more informative and faithful as well as less redundant than the seq2seq model.
We propose a semantic-aware neural abstractive summarization model and a novel automatic summarization evaluation scheme that measures how well a model identifies off-topic information from adversarial samples.
1,263
SuperChat: Dialogue Generation by Transfer Learning from Vision to Language using Two-dimensional Word Embedding and Pretrained ImageNet CNN Models
The recent work of Super Characters method using two-dimensional word embedding achieved state-of-the-art results in text classification tasks, showcasing the promise of this new approach.This paper borrows the idea of Super Characters method and two-dimensional embedding, and proposes a method of generating conversational response for open domain dialogues.The experimental results on a public dataset shows that the proposed SuperChat method generates high quality responses.An interactive demo is ready to show at the workshop.And code will be available at github soon.
Print the input sentence and current response sentence onto an image and use fine-tuned ImageNet CNN model to predict the next response word.
1,264
Unleashing the Potential of CNNs for Interpretable Few-Shot Learning
Convolutional neural networks have been generally acknowledged as one of the driving forces for the advancement of computer vision.Despite their promising performances on many tasks, CNNs still face major obstacles on the road to achieving ideal machine intelligence.One is that CNNs are complex and hard to interpret.Another is that standard CNNs require large amounts of annotated data, which is sometimes very hard to obtain, and it is desirable to be able to learn them from few examples.In this work, we address these limitations of CNNs by developing novel, simple, and interpretable models for few-shot learn- ing. Our models are based on the idea of encoding objects in terms of visual concepts, which are interpretable visual cues represented by the feature vectors within CNNs.We first adapt the learning of visual concepts to the few-shot setting, and then uncover two key properties of feature encoding using visual concepts, which we call category sensitivity and spatial pattern.Motivated by these properties, we present two intuitive models for the problem of few-shot learning.Experiments show that our models achieve competitive performances, while being much more flexible and interpretable than alternative state-of-the-art few-shot learning methods.We conclude that using visual concepts helps expose the natural capability of CNNs for few-shot learning.
We enable ordinary CNNs for few-shot learning by exploiting visual concepts which are interpretable visual cues learnt within CNNs.
1,265
Unsupervised Video-to-Video Translation
Unsupervised image-to-image translation is a recently proposed task of translating an image to a different style or domain given only unpaired image examples at training time.In this paper, we formulate a new task of unsupervised video-to-video translation, which poses its own unique challenges.Translating video implies learning not only the appearance of objects and scenes but also realistic motion and transitions between consecutive frames.We investigate the performance of per-frame video-to-video translation using existing image-to-image translation networks, and propose a spatio-temporal 3D translator as an alternative solution to this problem.We evaluate our 3D method on multiple synthetic datasets, such as moving colorized digits, as well as the realistic segmentation-to-video GTA dataset and a new CT-to-MRI volumetric images translation dataset.Our results show that frame-wise translation produces realistic results on a single frame level but underperforms significantly on the scale of the whole video compared to our three-dimensional translation approach, which is better able to learn the complex structure of video and motion and continuity of object appearance.
Proposed new task, datasets and baselines; 3D Conv CycleGAN preserves object properties across frames; batch structure in frame-level methods matters.
1,266
Building Deep Equivariant Capsule Networks
Capsule networks are constrained by the parameter-expensive nature of their layers, and the general lack of provable equivariance guarantees.We present a variation of capsule networks that aims to remedy this.We identify that learning all pair-wise part-whole relationships between capsules of successive layers is inefficient.Further, we also realise that the choice of prediction networks and the routing mechanism are both key to equivariance.Based on these, we propose an alternative framework for capsule networks that learns to projectively encode the manifold of pose-variations, termed the space-of-variation, for every capsule-type of each layer.This is done using a trainable, equivariant function defined over a grid of group-transformations.Thus, the prediction-phase of routing involves projection into the SOV of a deeper capsule using the corresponding function.As a specific instantiation of this idea, and also in order to reap the benefits of increased parameter-sharing, we use type-homogeneous group-equivariant convolutions of shallower capsules in this phase.We also introduce an equivariant routing mechanism based on degree-centrality.We show that this particular instance of our general model is equivariant, and hence preserves the compositional representation of an input under transformations.We conduct several experiments on standard object-classification datasets that showcase the increased transformation-robustness, as well as general performance, of our model to several capsule baselines.
A new scalable, group-equivariant model for capsule networks that preserves compositionality under transformations, and is empirically more transformation-robust to older capsule network models.
1,267
A Simple yet Effective Baseline for Robust Deep Learning with Noisy Labels
Recently deep neural networks have shown their capacity to memorize training data, even with noisy labels, which hurts generalization performance.To mitigate this issue, we propose a simple but effective method that is robust to noisy labels, even with severe noise. Our objective involves a variance regularization term that implicitly penalizes the Jacobian norm of the neural network on the whole training set, which encourages generalization and prevents overfitting to the corrupted labels.Experiments on noisy benchmarks demonstrate that our approach achieves state-of-the-art performance with a high tolerance to severe noise.
The paper proposed a simple yet effective baseline for learning with noisy labels.
1,268
Has Machine Translation Achieved Human Parity? A Case for Document-level Evaluation
Recent research suggests that neural machine translation achieves parity with professional human translation on the WMT Chinese--English news translation task.We empirically test this claim with alternative evaluation protocols, contrasting the evaluation of single sentences and entire documents.In a pairwise ranking experiment, human raters assessing adequacy and fluency show a stronger preference for human over machine translation when evaluating documents as compared to isolated sentences.Our findings emphasise the need to shift towards document-level evaluation as machine translation improves to the degree that errors which are hard or impossible to spot at the sentence-level become decisive in discriminating quality of different translation outputs.
Raters prefer adequacy in human over machine translation when evaluating entire documents, but not when evaluating single sentences.
1,269
ASYNCHRONOUS MULTI-AGENT GENERATIVE ADVERSARIAL IMITATION LEARNING
Imitation learning aims to inversely learn a policy from expert demonstrations, which has been extensively studied in the literature for both single-agent setting with Markov decision process model, and multi-agent setting with Markov game model.However, existing approaches for general multi-agent Markov games are not applicable to multi-agent extensive Markov games, where agents make asynchronous decisions following a certain order, rather than simultaneous decisions.We propose a novel framework for asynchronous multi-agent generative adversarial imitation learning under general extensive Markov game settings, and the learned expert policies are proven to guarantee subgame perfect equilibrium, a more general and stronger equilibrium than Nash equilibrium.The experiment results demonstrate that compared to state-of-the-art baselines, our AMAGAIL model can better infer the policy of each expert agent using their demonstration data collected from asynchronous decision-making scenarios.
This paper extends the multi-agent generative adversarial imitation learning to extensive-form Markov games.
1,270
Revisiting Self-Training for Neural Sequence Generation
Self-training is one of the earliest and simplest semi-supervised methods.The key idea is to augment the original labeled dataset with unlabeled data paired with the model’s prediction.Self-training has mostly been well-studied to classification problems.However, in complex sequence generation tasks such as machine translation, it is still not clear how self-training woks due to the compositionality of the target space.In this work, we first show that it is not only possible but recommended to apply self-training in sequence generation.Through careful examination of the performance gains, we find that the noise added on the hidden states is critical to the success of self-training, as this acts like a regularizer which forces the model to yield similar predictions for similar inputs from unlabeled data.To further encourage this mechanism, we propose to inject noise to the input space, resulting in a “noisy” version of self-training.Empirical study on standard benchmarks across machine translation and text summarization tasks under different resource settings shows that noisy self-training is able to effectively utilize unlabeled data and improve the baseline performance by large margin.
We revisit self-training as a semi-supervised learning method for neural sequence generation problem, and show that self-training can be quite successful with injected noise.
1,271
The Kanerva Machine: A Generative Distributed Memory
We present an end-to-end trained memory system that quickly adapts to new data and generates samples like them."Inspired by Kanerva's sparse distributed memory, it has a robust distributed reading and writing mechanism.", 'The memory is analytically tractable, which enables optimal on-line compression via a Bayesian update-rule.We formulate it as a hierarchical conditional generative model, where memory provides a rich data-dependent prior distribution.Consequently, the top-down memory and bottom-up perception are combined to produce the code representing an observation.Empirically, we demonstrate that the adaptive memory significantly improves generative models trained on both the Omniglot and CIFAR datasets.Compared with the Differentiable Neural Computer and its variants, our memory model has greater capacity and is significantly easier to train.
A generative memory model that combines slow-learning neural networks and a fast-adapting linear Gaussian model as memory.
1,272
SNIP: SINGLE-SHOT NETWORK PRUNING BASED ON CONNECTION SENSITIVITY
Pruning large neural networks while maintaining their performance is often desirable due to the reduced space and time complexity.In existing methods, pruning is done within an iterative optimization procedure with either heuristically designed pruning schedules or additional hyperparameters, undermining their utility.In this work, we present a new approach that prunes a given network once at initialization prior to training.To achieve this, we introduce a saliency criterion based on connection sensitivity that identifies structurally important connections in the network for the given task.This eliminates the need for both pretraining and the complex pruning schedule while making it robust to architecture variations.After pruning, the sparse network is trained in the standard way.Our method obtains extremely sparse networks with virtually the same accuracy as the reference network on the MNIST, CIFAR-10, and Tiny-ImageNet classification tasks and is broadly applicable to various architectures including convolutional, residual and recurrent networks.Unlike existing methods, our approach enables us to demonstrate that the retained connections are indeed relevant to the given task.
We present a new approach, SNIP, that is simple, versatile and interpretable; it prunes irrelevant connections for a given task at single-shot prior to training and is applicable to a variety of neural network models without modifications.
1,273
Automatic Labeling of Data for Transfer Learning
Transfer learning uses trained weights from a source model as the initial weightsfor the training of a target dataset. A well chosen source with a large numberof labeled data leads to significant improvement in accuracy. We demonstrate atechnique that automatically labels large unlabeled datasets so that they can trainsource models for transfer learning.We experimentally evaluate this method, usinga baseline dataset of human-annotated ImageNet1K labels, against five variationsof this technique. We show that the performance of these automatically trainedmodels come within 17% of baseline on average.
A technique for automatically labeling large unlabeled datasets so that they can train source models for transfer learning and its experimental evaluation.
1,274
Tree-structured Attention Module for Image Classification
Recent studies in attention modules have enabled higher performance in computer vision tasks by capturing global contexts and accordingly attending important features.In this paper, we propose a simple and highly parametrically efficient module named Tree-structured Attention Module which recursively encourages neighboring channels to collaborate in order to produce a spatial attention map as an output.Unlike other attention modules which try to capture long-range dependencies at each channel, our module focuses on imposing non-linearities be- tween channels by utilizing point-wise group convolution.This module not only strengthens representational power of a model but also acts as a gate which controls signal flow.Our module allows a model to achieve higher performance in a highly parameter-efficient manner.We empirically validate the effectiveness of our module with extensive experiments on CIFAR-10/100 and SVHN datasets.With our proposed attention module employed, ResNet50 and ResNet101 models gain 2.3% and 1.2% accuracy improvement with less than 1.5% parameter over- head.Our PyTorch implementation code is publicly available.
Our paper proposes an attention module which captures inter-channel relationships and offers large performance gains.
1,275
Few-Shot Intent Inference via Meta-Inverse Reinforcement Learning
A significant challenge for the practical application of reinforcement learning toreal world problems is the need to specify an oracle reward function that correctly defines a task.Inverse reinforcement learning seeks to avoid this challenge by instead inferring a reward function from expert behavior. While appealing, it can be impractically expensive to collect datasets of demonstrations that cover the variation common in the real world.Thus in practice, IRL must commonly be performed with only a limited set of demonstrations where it can be exceedingly difficult to unambiguously recover a reward function.In this work, we exploit the insight that demonstrations from other tasks can be used to constrain the set of possible reward functions by learning a "prior" that is specifically optimized for the ability to infer expressive reward functions from limited numbers of demonstrations. We demonstrate that our method can efficiently recover rewards from images for novel tasks and provide intuition as to how our approach is analogous to learning a prior.
The applicability of inverse reinforcement learning is often hampered by the expense of collecting expert demonstrations; this paper seeks to broaden its applicability by incorporating prior task information through meta-learning.
1,276
Deepström Networks
Recent work has focused on combining kernel methods and deep learning.With this in mind, we introduce Deepström networks -- a new architecture of neural networks which we use to replace top dense layers of standard convolutional architectures with an approximation of a kernel function by relying on the Nyström approximation.Our approach is easy highly flexible.It is compatible with any kernel function and it allows exploiting multiple kernels.We show that Deepström networks reach state-of-the-art performance on standard datasets like SVHN and CIFAR100.One benefit of the method lies in its limited number of learnable parameters which make it particularly suited for small training set sizes, e.g. from 5 to 20 samples per class.Finally we illustrate two ways of using multiple kernels, including a multiple Deepström setting, that exploits a kernel on each feature map output by the convolutional part of the model.
A new neural architecture where top dense layers of standard convolutional architectures are replaced with an approximation of a kernel function by relying on the Nyström approximation.
1,277
The blood diamond effect in neural art: On ethically troublesome images of the imagenet dataset
The main goal of this short paper is to inform the neural art community at large on the ethical ramifications of using models trained on the imagenet dataset, or using seed images from classes 445 -n02892767- [’bikini, two-piece’] and 459- n02837789- [’brassiere, bra, bandeau’] of the same.We discovered that many of the images belong to these classes were verifiably pornographic, shot in a non-consensual setting, voyeuristic and also entailed underage nudity.Akin to the and nexuses, we posit there is a similar moral conundrum at play here and would like to instigate a conversation amongst the neural artists in the community.
There's non-consensual and pornographic images in the ImageNet dataset
1,278
Inductive and Unsupervised Representation Learning on Graph Structured Objects
Inductive and unsupervised graph learning is a critical technique for predictive or information retrieval tasks where label information is difficult to obtain.It is also challenging to make graph learning inductive and unsupervised at the same time, as learning processes guided by reconstruction error based loss functions inevitably demand graph similarity evaluation that is usually computationally intractable.In this paper, we propose a general framework SEED for inductive and unsupervised representation learning on graph structured objects.Instead of directly dealing with the computational challenges raised by graph similarity evaluation, given an input graph, the SEED framework samples a number of subgraphs whose reconstruction errors could be efficiently evaluated, encodes the subgraph samples into a collection of subgraph vectors, and employs the embedding of the subgraph vector distribution as the output vector representation for the input graph.By theoretical analysis, we demonstrate the close connection between SEED and graph isomorphism.Using public benchmark datasets, our empirical study suggests the proposed SEED framework is able to achieve up to 10% improvement, compared with competitive baseline methods.
This paper proposed a novel framework for graph similarity learning in inductive and unsupervised scenario.
1,279
Adversarial Training of Neural Encoding Models on Population Spike Trains
Neural population responses to sensory stimuli can exhibit both nonlinear stimulus- dependence and richly structured shared variability.Here, we show how adversarial training can be used to optimize neural encoding models to capture both the deterministic and stochastic components of neural population data.To account for the discrete nature of neural spike trains, we use the REBAR method to estimate unbiased gradients for adversarial optimization of neural encoding models.We illustrate our approach on population recordings from primary visual cortex.We show that adding latent noise-sources to a convolutional neural network yields a model which captures both the stimulus-dependence and noise correlations of the population activity.
We show how neural encoding models can be trained to capture both the signal and spiking variability of neural population data using GANs.
1,280
Weakly Supervised Clustering by Exploiting Unique Class Count
A weakly supervised learning based clustering framework is proposed in this paper.As the core of this framework, we introduce a novel multiple instance learning task based on a bag level label called unique class count, which is the number of unique classes among all instances inside the bag.In this task, no annotations on individual instances inside the bag are needed during training of the models.We mathematically prove that with a perfect ucc classifier, perfect clustering of individual instances inside the bags is possible even when no annotations on individual instances are given during training.We have constructed a neural network based ucc classifier and experimentally shown that the clustering performance of our framework with our weakly supervised ucc classifier is comparable to that of fully supervised learning models where labels for all instances are known.Furthermore, we have tested the applicability of our framework to a real world task of semantic segmentation of breast cancer metastases in histological lymph node sections and shown that the performance of our weakly supervised framework is comparable to the performance of a fully supervised Unet model.
A weakly supervised learning based clustering framework performs comparable to that of fully supervised learning models by exploiting unique class count.
1,281
Per-Tensor Fixed-Point Quantization of the Back-Propagation Algorithm
The high computational and parameter complexity of neural networks makes their training very slow and difficult to deploy on energy and storage-constrained comput- ing systems.Many network complexity reduction techniques have been proposed including fixed-point implementation.However, a systematic approach for design- ing full fixed-point training and inference of deep neural networks remains elusive.We describe a precision assignment methodology for neural network training in which all network parameters, i.e., activations and weights in the feedforward path, gradients and weight accumulators in the feedback path, are assigned close to minimal precision.The precision assignment is derived analytically and enables tracking the convergence behavior of the full precision training, known to converge a priori.Thus, our work leads to a systematic methodology of determining suit- able precision for fixed-point training.The near optimality of the resulting precision assignment is validated empirically for four networks on the CIFAR-10, CIFAR-100, and SVHN datasets.The complexity reduction arising from our approach is compared with other fixed-point neural network designs.
We analyze and determine the precision requirements for training neural networks when all tensors, including back-propagated signals and weight accumulators, are quantized to fixed-point format.
1,282
Prototypical Examples in Deep Learning: Metrics, Characteristics, and Utility
Machine learning research has investigated prototypes: examples that are representative of the behavior to be learned.We systematically evaluate five methods for identifying prototypes, both ones previously introduced as well as new ones we propose, finding all of them to provide meaningful but different interpretations.Through a human study, we confirm that all five metrics are well matched to human intuition.Examining cases where the metrics disagree offers an informative perspective on the properties of data and algorithms used in learning, with implications for data-corpus construction, efficiency, adversarial robustness, interpretability, and other ML aspects.In particular, we confirm that the "train on hard" curriculum approach can improve accuracy on many datasets and tasks, but that it is strictly worse when there are many mislabeled or ambiguous examples.
We can identify prototypical and outlier examples in machine learning that are quantifiably very different, and make use of them to improve many aspects of neural networks.
1,283
Sparse Deep Scattering Croisé Network
In this work, we propose the Sparse Deep Scattering Croisé Network a novel architecture based on the Deep Scattering Network.The DSN is achieved by cascading wavelet transform convolutions with a complex modulus and a time-invariant operator.We extend this work by first,crossing multiple wavelet family transforms to increase the feature diversity while avoiding any learning.Thus providing a more informative latent representation and benefit from the development of highly specialized wavelet filters over the last decades.Beside, by combining all the different wavelet representations, we reduce the amount of prior information needed regarding the signals at hand.Secondly, we develop an optimal thresholding strategy for over-complete filter banks that regularizes the network and controls instabilities such as inherent non-stationary noise in the signal.Our systematic and principled solution sparsifies the latent representation of the network by acting as a local mask distinguishing between activity and noise.Thus, we propose to enhance the DSN by increasing the variance of the scattering coefficients representation as well as improve its robustness with respect to non-stationary noise.We show that our new approach is more robust and outperforms the DSN on a bird detection task.
We propose to enhance the Deep Scattering Network in order to improve control and stability of any given machine learning pipeline by proposing a continuous wavelet thresholding scheme
1,284
(How) Can AI Bots Lie?
Recent work on explanation generation for decision-making problems has viewed the explanation process as one of model reconciliation where an AI agent brings the human mental model to the same page with regards to a task at hand.This formulation succinctly captures many possible types of explanations, as well as explicitly addresses the various properties -- e.g. the social aspects, contrastiveness, and selectiveness -- of explanations studied in social sciences among human-human interactions.However, it turns out that the same process can be hijacked into producing "alternative explanations" -- i.e. explanations that are not true but still satisfy all the properties of a proper explanation.In previous work, we have looked at how such explanations may be perceived by the human in the loop and alluded to one possible way of generating them.In this paper, we go into more details of this curious feature of the model reconciliation process and discuss similar implications to the overall notion of explainable decision-making.
Model Reconciliation is an established framework for plan explanations, but can be easily hijacked to produce lies.
1,285
Optimistic Acceleration for Optimization
We consider new variants of optimization algorithms.Our algorithms are based on the observation that mini-batch of stochastic gradients in consecutive iterations do not change drastically and consequently may be predictable.Inspired by the similar setting in online learning literature called Optimistic Online learning, we propose two new optimistic algorithms for AMSGrad and Adam, respectively, by exploiting the predictability of gradients. The new algorithms combine the idea of momentum method, adaptive gradient method, and algorithms in Optimistic Online learning, which leads to speed up in training deep neural nets in practice.
We consider new variants of optimization algorithms for training deep nets.
1,286
Dual-module Inference for Efficient Recurrent Neural Networks
Using Recurrent Neural Networks in sequence modeling tasks is promising in delivering high-quality results but challenging to meet stringent latency requirements because of the memory-bound execution pattern of RNNs.We propose a big-little dual-module inference to dynamically skip unnecessary memory access and computation to speedup RNN inference.Leveraging the error-resilient feature of nonlinear activation functions used in RNNs, we propose to use a lightweight little module that approximates the original RNN layer, which is referred to as the big module, to compute activations of the insensitive region that are more error-resilient.The expensive memory access and computation of the big module can be reduced as the results are only used in the sensitive region.Our method can reduce the overall memory access by 40% on average and achieve 1.54x to 1.75x speedup on CPU-based server platform with negligible impact on model quality.
We accelerate RNN inference by dynamically reducing redundant memory access using a mixture of accurate and approximate modules.
1,287
Fine-grained Entity Recognition with Reduced False Negatives and Large Type Coverage
Fine-grained Entity Recognition is the task of detecting and classifying entity mentions to a large set of types spanning diverse domains such as biomedical, finance and sports. We observe that when the type set spans several domains, detection of entity mention becomes a limitation for supervised learning models. The primary reason being lack of dataset where entity boundaries are properly annotated while covering a large spectrum of entity types. Our work directly addresses this issue. We propose Heuristics Allied with Distant Supervision framework to automatically construct a quality dataset suitable for the FgER task. HAnDS framework exploits the high interlink among Wikipedia and Freebase in a pipelined manner, reducing annotation errors introduced by naively using distant supervision approach. Using HAnDS framework, we create two datasets, one suitable for building FgER systems recognizing up to 118 entity types based on the FIGER type hierarchy and another for up to 1115 entity types based on the TypeNet hierarchy. Our extensive empirical experimentation warrants the quality of the generated datasets. Along with this, we also provide a manually annotated dataset for benchmarking FgER systems.
We initiate a push towards building ER systems to recognize thousands of types by providing a method to automatically construct suitable datasets based on the type hierarchy.
1,288
INVOCMAP: MAPPING METHOD NAMES TO METHOD INVOCATIONS VIA MACHINE LEARNING
Implementing correct method invocation is an important task for software developers.However, this is challenging work, since the structure of method invocation can be complicated.In this paper, we propose InvocMap, a code completion tool allows developers to obtain an implementation of multiple method invocations from a list of method names inside code context.InvocMap is able to predict the nested method invocations which their names didn’t appear in the list of input method names given by developers.To achieve this, we analyze the Method Invocations by four levels of abstraction.We build a Machine Translation engine to learn the mapping from the first level to the third level of abstraction of multiple method invocations, which only requires developers to manually add local variables from generated expression to get the final code.We evaluate our proposed approach on six popular libraries: JDK, Android, GWT, Joda-Time, Hibernate, and Xstream.With the training corpus of 2.86 million method invocations extracted from 1000 Java Github projects and the testing corpus extracted from 120 online forums code snippets, InvocMap achieves the accuracy rate up to 84 in F1- score depending on how much information of context provided along with method names, that shows its potential for auto code completion.
This paper proposes a theory of classifying Method Invocations by different abstraction levels and conducting a statistical approach for code completion from method name to method invocation.
1,289
Hiding Objects from Detectors: Exploring Transferrable Adversarial Patterns
Adversaries in neural networks have drawn much attention since their first debut.While most existing methods aim at deceiving image classification models into misclassification or crafting attacks for specific object instances in the object setection tasks, we focus on creating universal adversaries to fool object detectors and hide objects from the detectors.The adversaries we examine are universal in three ways: They are not specific for specific object instances; They are image-independent; They can further transfer to different unknown models.To achieve this, we propose two novel techniques to improve the transferability of the adversaries: and .Both techniques prove to simplify the patterns of generated adversaries, and ultimately result in higher transferability.
We focus on creating universal adversaries to fool object detectors and hide objects from the detectors.
1,290
DNN Feature Map Compression using Learned Representation over GF(2)
In this paper, we introduce a method to compress intermediate feature maps of deep neural networks to decrease memory storage and bandwidth requirements during inference.Unlike previous works, the proposed method is based on converting fixed-point activations into vectors over the smallest GF finite field followed by nonlinear dimensionality reduction layers embedded into a DNN.Such an end-to-end learned representation finds more compact feature maps by exploiting quantization redundancies within the fixed-point activations along the channel or spatial dimensions.We apply the proposed network architecture to the tasks of ImageNet classification and PASCAL VOC object detection.Compared to prior approaches, the conducted experiments show a factor of 2 decrease in memory requirements with minor degradation in accuracy while adding only bitwise computations.
Feature map compression method that converts quantized activations into binary vectors followed by nonlinear dimensionality reduction layers embedded into a DNN
1,291
Label Smoothing and Logit Squeezing: A Replacement for Adversarial Training?
Adversarial training is one of the strongest defenses against adversarial attacks, but it requires adversarial examples to be generated for every mini-batch during optimization. The expense of producing these examples during training often precludes adversarial training from use on complex image datasets.In this study, we explore the mechanisms by which adversarial training improves classifier robustness, and show that these mechanisms can be effectively mimicked using simple regularization methods, including label smoothing and logit squeezing. Remarkably, using these simple regularization methods in combination with Gaussian noise injection, we are able to achieve strong adversarial robustness -- often exceeding that of adversarial training -- using no adversarial examples.
Achieving strong adversarial robustness comparable to adversarial training without training on adversarial examples
1,292
Meta-Learning Update Rules for Unsupervised Representation Learning
A major goal of unsupervised learning is to discover data representations that are useful for subsequent tasks, without access to supervised labels during training.Typically, this involves minimizing a surrogate objective, such as the negative log likelihood of a generative model, with the hope that representations useful for subsequent tasks will arise as a side effect.In this work, we propose instead to directly target later desired tasks by meta-learning an unsupervised learning rule which leads to representations useful for those tasks. Specifically, we target semi-supervised classification performance, and we meta-learn an algorithm -- an unsupervised weight update rule -- that produces representations useful for this task.Additionally, we constrain our unsupervised update rule to a be a biologically-motivated, neuron-local function, which enables it to generalize to different neural network architectures, datasets, and data modalities.We show that the meta-learned update rule produces useful features and sometimes outperforms existing unsupervised learning techniques.We further show that the meta-learned unsupervised update rule generalizes to train networks with different widths, depths, and nonlinearities.It also generalizes to train on data with randomly permuted input dimensions and even generalizes from image datasets to a text task.
We learn an unsupervised learning algorithm that produces useful representations from a set of supervised tasks. At test-time, we apply this algorithm to new tasks without any supervision and show performance comparable to a VAE.
1,293
In Support of Over-Parametrization in Deep Reinforcement Learning: an Empirical Study
There is significant recent evidence in supervised learning that, in the over-parametrized setting, wider networks achieve better test error.In other words, the bias-variance tradeoff is not directly observable when increasing network width arbitrarily.We investigate whether a corresponding phenomenon is present in reinforcement learning.We experiment on four OpenAI Gym environments, increasing the width of the value and policy networks beyond their prescribed values.Our empirical results lend support to this hypothesis.However, tuning the hyperparameters of each network width separately remains as important future work in environments/algorithms where the optimal hyperparameters vary noticably across widths, confounding the results when the same hyperparameters are used for all widths.
Over-parametrization in width seems to help in deep reinforcement learning, just as it does in supervised learning.
1,294
CZ-GEM: A FRAMEWORK FOR DISENTANGLED REPRESENTATION LEARNING
Learning disentangled representations of data is one of the central themes in unsupervised learning in general and generative modelling in particular. In this work, we tackle a slightly more intricate scenario where the observations are generated from a conditional distribution of some known control variate and some latent noise variate. To this end, we present a hierarchical model and a training method that leverages some of the recent developments in likelihood-based and likelihood-free generative models. We show that by formulation, CZ-GEM introduces the right inductive biases that ensure the disentanglement of the control from the noise variables, while also keeping the components of the control variate disentangled.This is achieved without compromising on the quality of the generated samples.Our approach is simple, general, and can be applied both in supervised and unsupervised settings.
Hierarchical generative model (hybrid of VAE and GAN) that learns a disentangled representation of data without compromising the generative quality.
1,295
Deep Active Learning for Named Entity Recognition
Deep learning has yielded state-of-the-art performance on many natural language processing tasks including named entity recognition.However, this typically requires large amounts of labeled data.In this work, we demonstrate that the amount of labeled training data can be drastically reduced when deep learning is combined with active learning.While active learning is sample-efficient, it can be computationally expensive since it requires iterative retraining.To speed this up, we introduce a lightweight architecture for NER, viz., the CNN-CNN-LSTM model consisting of convolutional character and word encoders and a long short term memory tag decoder.The model achieves nearly state-of-the-art performance on standard datasets for the task while being computationally much more efficient than best performing models.We carry out incremental active learning, during the training process, and are able to nearly match state-of-the-art performance with just 25% of the original training data.
We introduce a lightweight architecture for named entity recognition and carry out incremental active learning, which is able to match state-of-the-art performance with just 25% of the original training data.
1,296
GQ-Net: Training Quantization-Friendly Deep Networks
Network quantization is a model compression and acceleration technique that has become essential to neural network deployment.Most quantization methods per- form fine-tuning on a pretrained network, but this sometimes results in a large loss in accuracy compared to the original network.We introduce a new technique to train quantization-friendly networks, which can be directly converted to an accurate quantized network without the need for additional fine-tuning.Our technique allows quantizing the weights and activations of all network layers down to 4 bits, achieving high efficiency and facilitating deployment in practical settings.Com- pared to other fully quantized networks operating at 4 bits, we show substantial improvements in accuracy, for example 66.68% top-1 accuracy on ImageNet using ResNet-18, compared to the previous state-of-the-art accuracy of 61.52% Louizos et al. and a full precision reference accuracy of 69.76%.We performed a thorough set of experiments to test the efficacy of our method and also conducted ablation studies on different aspects of the method and techniques to improve training stability and accuracy.Our codebase and trained models are available on GitHub.
We train accurate fully quantized networks using a loss function maximizing full precision model accuracy and minimizing the difference between the full precision and quantized networks.
1,297
Connectivity Learning in Multi-Branch Networks
While much of the work in the design of convolutional networks over the last five years has revolved around the empirical investigation of the importance of depth, filter sizes, and number of feature channels, recent studies have shown that branching, i.e., splitting the computation along parallel but distinct threads and then aggregating their outputs, represents a new promising dimension for significant improvements in performance.To combat the complexity of design choices in multi-branch architectures, prior work has adopted simple strategies, such as a fixed branching factor, the same input being fed to all parallel branches, and an additive combination of the outputs produced by all branches at aggregation points.In this work we remove these predefined choices and propose an algorithm to learn the connections between branches in the network.Instead of being chosen a priori by the human designer, the multi-branch connectivity is learned simultaneously with the weights of the network by optimizing a single loss function defined with respect to the end task."We demonstrate our approach on the problem of multi-class image classification using four different datasets where it yields consistently higher accuracy compared to the state-of-the-art ResNeXt multi-branch network given the same learning capacity.
In this paper we introduced an algorithm to learn the connectivity of deep multi-branch networks. The approach is evaluated on image categorization where it consistently yields accuracy gains over state-of-the-art models that use fixed connectivity.
1,298
Discovery of Natural Language Concepts in Individual Units of CNNs
Although deep convolutional networks have achieved improved performance in many natural language tasks, they have been treated as black boxes because they are difficult to interpret.Especially, little is known about how they represent language in their intermediate layers.In an attempt to understand the representations of deep convolutional networks trained on language tasks, we show that individual units are selectively responsive to specific morphemes, words, and phrases, rather than responding to arbitrary and uninterpretable patterns.In order to quantitatively analyze such intriguing phenomenon, we propose a concept alignment method based on how units respond to replicated text.We conduct analyses with different architectures on multiple datasets for classification and translation tasks and provide new insights into how deep models understand natural language.
We show that individual units in CNN representations learned in NLP tasks are selectively responsive to natural language concepts.
1,299
Learning Temporal Abstraction with Information-theoretic Constraints for Hierarchical Reinforcement Learning
Applying reinforcement learning to real-world problems will require reasoning about action-reward correlation over long time horizons.Hierarchical reinforcement learning methods handle this by dividing the task into hierarchies, often with hand-tuned network structure or pre-defined subgoals.We propose a novel HRL framework TAIC, which learns the temporal abstraction from past experience or expert demonstrations without task-specific knowledge.We formulate the temporal abstraction problem as learning latent representations of action sequences and present a novel approach of regularizing the latent space by adding information-theoretic constraints.Specifically, we maximize the mutual information between the latent variables and the state changes.A visualization of the latent space demonstrates that our algorithm learns an effective abstraction of the long action sequences.The learned abstraction allows us to learn new tasks on higher level more efficiently.We convey a significant speedup in convergence over benchmark learning problems.These results demonstrate that learning temporal abstractions is an effective technique in increasing the convergence rate and sample efficiency of RL algorithms.
We propose a novel HRL framework, in which we formulate the temporal abstraction problem as learning a latent representation of action sequence.