input
stringlengths 331
3.18k
| output
sequence |
---|---|
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['In this paper, we introduce a novel method to interpret recurrent neural networks (RNNs), particularly long short-term memory networks (LSTMs) at the cellular level.', 'We propose a systematic pipeline for interpreting individual hidden state dynamics within the network using response characterization methods.', "The ranked contribution of individual cells to the network's output is computed by analyzing a set of interpretable metrics of their decoupled step and sinusoidal responses.", "As a result, our method is able to uniquely identify neurons with insightful dynamics, quantify relationships between dynamical properties and test accuracy through ablation analysis, and interpret the impact of network capacity on a network's dynamical distribution.", 'Finally, we demonstrate generalizability and scalability of our method by evaluating a series of different benchmark sequential datasets.']
### SUMMARY:
| [
"Introducing the response charactrization method for interpreting cell dynamics in learned long short-term memory (LSTM) networks. "
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Decades of research on the neural code underlying spatial navigation have revealed a diverse set of neural response properties.', 'The Entorhinal Cortex (EC) of the mammalian brain contains a rich set of spatial correlates, including grid cells which encode space using tessellating patterns.', 'However, the mechanisms and functional significance of these spatial representations remain largely mysterious.', 'As a new way to understand these neural representations, we trained recurrent neural networks (RNNs) to perform navigation tasks in 2D arenas based on velocity inputs.', 'Surprisingly, we find that grid-like spatial response patterns emerge in trained networks, along with units that exhibit other spatial correlates, including border cells and band-like cells.', 'All these different functional types of neurons have been observed experimentally.', 'The order of the emergence of grid-like and border cells is also consistent with observations from developmental studies.', 'Together, our results suggest that grid cells, border cells and others as observed in EC may be a natural solution for representing space efficiently given the predominant recurrent connections in the neural circuits.\n']
### SUMMARY:
| [
"To our knowledge, this is the first study to show how neural representations of space, including grid-like cells and border cells as observed in the brain, could emerge from training a recurrent neural network to perform navigation tasks."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Source separation for music is the task of isolating contributions, or stems, from different instruments recorded individually and arranged together to form a song.', 'Such components include voice, bass, drums and any other accompaniments.', 'While end-to-end models that directly generate the waveform are state-of-the-art in many audio synthesis problems, the best multi-instrument source separation models generate masks on the magnitude spectrum and achieve performances far above current end-to-end, waveform-to-waveform models.', 'We present an in-depth analysis of a new architecture, which we will refer to as Demucs, based on a (transposed) convolutional autoencoder, with a bidirectional LSTM at the bottleneck layer and skip-connections as in U-Networks (Ronneberger et al., 2015).', 'Compared to the state-of-the-art waveform-to-waveform model, Wave-U-Net (Stoller et al., 2018), the main features of our approach in addition of the bi-LSTM are the use of trans-posed convolution layers instead of upsampling-convolution blocks, the use of gated linear units, exponentially growing the number of channels with depth and a new careful initialization of the weights. ', 'Results on the MusDB dataset show that our architecture achieves a signal-to-distortion ratio (SDR) nearly 2.2 points higher than the best waveform-to-waveform competitor (from 3.2 to 5.4 SDR).', 'This makes our model match the state-of-the-art performances on this dataset, bridging the performance gap between models that operate on the spectrogram and end-to-end approaches.']
### SUMMARY:
| [
"We match the performance of spectrogram based model with a model trained end-to-end in the waveform domain"
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Although challenging, strategy profile evaluation in large connected learner networks is crucial for enabling the next wave of machine learning applications.', 'Recently, $\\alpha$-Rank, an evolutionary algorithm, has been proposed as a solution for ranking joint policy profiles in multi-agent systems.', '$\\alpha$-Rank claimed scalability through a polynomial time implementation with respect to the total number of pure strategy profiles.', 'In this paper, we formally prove that such a claim is not grounded.', 'In fact, we show that $\\alpha$-Rank exhibits an exponential complexity in number of agents, hindering its application beyond a small finite number of joint profiles.', 'Realizing such a limitation, we contribute by proposing a scalable evaluation protocol that we title $\\alpha^{\\alpha}$-Rank.', 'Our method combines evolutionary dynamics with stochastic optimization and double oracles for \\emph{truly} scalable ranking with linear (in number of agents) time and memory complexities.', 'Our contributions allow us, for the first time, to conduct large-scale evaluation experiments of multi-agent systems, where we show successful results on large joint strategy profiles with sizes in the order of $\\mathcal{O}(2^{25})$ (i.e., $\\approx \\text{$33$ million strategies}$) -- a setting not evaluable using current techniques.']
### SUMMARY:
| [
"We provide a scalable solution to multi-agent evaluation with linear rate complexity in both time and memory in terms of number of agents"
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Deep neural networks are known to be annotation-hungry.', 'Numerous efforts have been devoted to reducing the annotation cost when learning with deep networks.', 'Two prominent directions include learning with noisy labels and semi-supervised learning by exploiting unlabeled data.', 'In this work, we propose DivideMix, a novel framework for learning with noisy labels by leveraging semi-supervised learning techniques.', 'In particular, DivideMix models the per-sample loss distribution with a mixture model to dynamically divide the training data into a labeled set with clean samples and an unlabeled set with noisy samples, and trains the model on both the labeled and unlabeled data in a semi-supervised manner.', 'To avoid confirmation bias, we simultaneously train two diverged networks where each network uses the dataset division from the other network.', 'During the semi-supervised training phase, we improve the MixMatch strategy by performing label co-refinement and label co-guessing on labeled and unlabeled samples, respectively.', 'Experiments on multiple benchmark datasets demonstrate substantial improvements over state-of-the-art methods.', 'Code is available at https://github.com/LiJunnan1992/DivideMix .']
### SUMMARY:
| [
"We propose a novel semi-supervised learning approach with SOTA performance on combating learning with noisy labels."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['We present a new algorithm to train a robust neural network against adversarial attacks. \n', 'Our algorithm is motivated by the following two ideas.', 'First, although recent work has demonstrated that fusing randomness can improve the robustness of neural networks (Liu 2017), we noticed that adding noise blindly to all the layers is not the optimal way to incorporate randomness. \n', 'Instead, we model randomness under the framework of Bayesian Neural Network (BNN) to formally learn the posterior distribution of models in a scalable way.', 'Second, we formulate the mini-max problem in BNN to learn the best model distribution under adversarial attacks, leading to an adversarial-trained Bayesian neural net.', 'Experiment results demonstrate that the proposed algorithm achieves state-of-the-art performance under strong attacks.', 'On CIFAR-10 with VGG network, our model leads to 14% accuracy improvement compared with adversarial training (Madry 2017) and random self-ensemble (Liu, 2017) under PGD attack with 0.035 distortion, and the gap becomes even larger on a subset of ImageNet.']
### SUMMARY:
| [
"We design an adversarial training method to Bayesian neural networks, showing a much stronger defense to white-box adversarial attacks"
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Federated learning distributes model training among a multitude of agents, who, guided by privacy concerns, perform training using their local data but share only model parameter updates, for iterative aggregation at the server.', 'In this work, we explore the threat of model poisoning attacks on federated learning initiated by a single, non-colluding malicious agent where the adversarial objective is to cause the model to misclassify a set of chosen inputs with high confidence.', "We explore a number of strategies to carry out this attack, starting with simple boosting of the malicious agent's update to overcome the effects of other agents' updates.", 'To increase attack stealth, we propose an alternating minimization strategy, which alternately optimizes for the training loss and the adversarial objective.', "We follow up by using parameter estimation for the benign agents' updates to improve on attack success.", 'Finally, we use a suite of interpretability techniques to generate visual explanations of model decisions for both benign and malicious models and show that the explanations are nearly visually indistinguishable.', 'Our results indicate that even a highly constrained adversary can carry out model poisoning attacks while simultaneously maintaining stealth, thus highlighting the vulnerability of the federated learning setting and the need to develop effective defense strategies.']
### SUMMARY:
| [
"Effective model poisoning attacks on federated learning able to cause high-confidence targeted misclassification of desired inputs"
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Despite rapid advances in speech recognition, current models remain brittle to superficial perturbations to their inputs.', 'Small amounts of noise can destroy the performance of an otherwise state-of-the-art model.', 'To harden models against background noise, practitioners often perform data augmentation, adding artificially-noised examples to the training set, carrying over the original label.', "In this paper, we hypothesize that a clean example and its superficially perturbed counterparts shouldn't merely map to the same class--- they should map to the same representation.", 'We propose invariant-representation-learning (IRL): At each training iteration, for each training example, we sample a noisy counterpart.', 'We then apply a penalty term to coerce matched representations at each layer (above some chosen layer).', 'Our key results, demonstrated on the LibriSpeech dataset are the following:', "(i) IRL significantly reduces character error rates (CER)on both `clean' (3.3% vs 6.5%) and `other' (11.0% vs 18.1%) test sets;", "(ii) on several out-of-domain noise settings (different from those seen during training), IRL's benefits are even more pronounced.", 'Careful ablations confirm that our results are not simply due to shrinking activations at the chosen layers.']
### SUMMARY:
| [
" In this paper, we hypothesize that superficially perturbed data points shouldn’t merely map to the same class---they should map to the same representation."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['In this paper we design a harmonic acoustic model for pitch detection.', 'This model arranges conventional convolution and sparse convolution in a way such that the global harmonic patterns captured by sparse convolution are composed of the enough number of local patterns captured by layers of conventional convolution.', 'When trained on the MAPS dataset, the harmonic model outperforms all existing pitch detection systems trained on the same dataset.', 'Most impressively, when trained on MAPS with simple data augmentation, the harmonic model with an LSTM layer on top surpasses an up-to-date, more complex pitch detection system trained on the MAESTRO dataset to which complicated data augmentation is applied and whose training split is an order-of-magnitude larger than the training split of MAPS.', 'The harmonic model has demonstrated potential to be used for advanced automatic music transcription (AMT) systems.']
### SUMMARY:
| [
"harmonic acoustic model"
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Learning domain-invariant representation is a dominant approach for domain generalization.', 'However, previous methods based on domain invariance overlooked the underlying dependency of classes on domains, which is responsible for the trade-off between classification accuracy and the invariance.', 'This study proposes a novel method {\\em adversarial feature learning under accuracy constraint (AFLAC)}, which maximizes domain invariance within a range that does not interfere with accuracy.', 'Empirical validations show that the performance of AFLAC is superior to that of baseline methods, supporting the importance of considering the dependency and the efficacy of the proposed method to overcome the problem.']
### SUMMARY:
| [
"Address the trade-off caused by the dependency of classes on domains by improving domain adversarial nets"
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Recent advances in deep generative models have lead to remarkable progress in synthesizing high quality images.', 'Following their successful application in image processing and representation learning, an important next step is to consider videos.', 'Learning generative models of video is a much harder task, requiring a model to capture the temporal dynamics of a scene, in addition to the visual presentation of objects.', 'While recent generative models of video have had some success, current progress is hampered by the lack of qualitative metrics that consider visual quality, temporal coherence, and diversity of samples.', 'To this extent we propose Fréchet Video Distance (FVD), a new metric for generative models of video based on FID.', 'We contribute a large-scale human study, which confirms that FVD correlates well with qualitative human judgment of generated videos.']
### SUMMARY:
| [
"We propose FVD: a new metric for generative models of video based on FID. A large-scale human study confirms that FVD correlates well with qualitative human judgment of generated videos."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Despite advances in deep learning, artificial neural networks do not learn the same way as humans do.', 'Today, neural networks can learn multiple tasks when trained on them jointly, but cannot maintain performance on learnt tasks when tasks are presented one at a time -- this phenomenon called catastrophic forgetting is a fundamental challenge to overcome before neural networks can learn continually from incoming data.', 'In this work, we derive inspiration from human memory to develop an architecture capable of learning continuously from sequentially incoming tasks, while averting catastrophic forgetting.', 'Specifically, our model consists of a dual memory architecture to emulate the complementary learning systems (hippocampus and the neocortex) in the human brain and maintains a consolidated long-term memory via generative replay of past experiences.', 'We', '(i) substantiate our claim that replay should be generative,', '(ii) show the benefits of generative replay and dual memory via experiments, and', '(iii) demonstrate improved performance retention even for small models with low capacity.', 'Our architecture displays many important characteristics of the human memory and provides insights on the connection between sleep and learning in humans.']
### SUMMARY:
| [
"A dual memory architecture inspired from human brain to learn sequentially incoming tasks, while averting catastrophic forgetting."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Most research on lifelong learning applies to images or games, but not language.\n', 'We present LAMOL, a simple yet effective method for lifelong language learning (LLL) based on language modeling.\n', 'LAMOL replays pseudo-samples of previous tasks while requiring no extra memory or model capacity.\n', 'Specifically, LAMOL is a language model that simultaneously learns to solve the tasks and generate training samples.\n', 'When the model is trained for a new task, it generates pseudo-samples of previous tasks for training alongside data for the new task.\n', 'The results show that LAMOL prevents catastrophic forgetting without any sign of intransigence and can perform five very different language tasks sequentially with only one model. \n', 'Overall, LAMOL outperforms previous methods by a considerable margin and is only 2-3% worse than multitasking, which is usually considered the LLL upper bound.\n', 'The source code is available at https://github.com/jojotenya/LAMOL.']
### SUMMARY:
| [
"Language modeling for lifelong language learning."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Deep learning natural language processing models often use vector word embeddings, such as word2vec or GloVe, to represent words.', 'A discrete sequence of words can be much more easily integrated with downstream neural layers if it is represented as a sequence of continuous vectors.', 'Also, semantic relationships between words, learned from a text corpus, can be encoded in the relative configurations of the embedding vectors.', 'However, storing and accessing embedding vectors for all words in a dictionary requires large amount of space, and may stain systems with limited GPU memory.', 'Here, we used approaches inspired by quantum computing to propose two related methods, word2ket and word2ketXS, for storing word embedding matrix during training and inference in a highly efficient way.', 'Our approach achieves a hundred-fold or more reduction in the space required to store the embeddings with almost no relative drop in accuracy in practical natural language processing tasks.']
### SUMMARY:
| [
"We use ideas from quantum computing to proposed word embeddings that utilize much fewer trainable parameters."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['One of the distinguishing aspects of human language is its compositionality, which allows us to describe complex environments with limited vocabulary.', 'Previously, it has been shown that neural network agents can learn to communicate in a highly structured, possibly compositional language based on disentangled input (e.g. hand- engineered features).', 'Humans, however, do not learn to communicate based on well-summarized features.', 'In this work, we train neural agents to simultaneously develop visual perception from raw image pixels, and learn to communicate with a sequence of discrete symbols.', 'The agents play an image description game where the image contains factors such as colors and shapes.', 'We train the agents using the obverter technique where an agent introspects to generate messages that maximize its own understanding.', 'Through qualitative analysis, visualization and a zero-shot test, we show that the agents can develop, out of raw image pixels, a language with compositional properties, given a proper pressure from the environment.']
### SUMMARY:
| [
"We train neural network agents to develop a language with compositional properties from raw pixel input."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['The ability to forecast a set of likely yet diverse possible future behaviors of an agent (e.g., future trajectories of a pedestrian) is essential for safety-critical perception systems (e.g., autonomous vehicles).', 'In particular, a set of possible future behaviors generated by the system must be diverse to account for all possible outcomes in order to take necessary safety precautions.', 'It is not sufficient to maintain a set of the most likely future outcomes because the set may only contain perturbations of a dominating single outcome (major mode).', 'While generative models such as variational autoencoders (VAEs) have been shown to be a powerful tool for learning a distribution over future trajectories, randomly drawn samples from the learned implicit likelihood model may not be diverse -- the likelihood model is derived from the training data distribution and the samples will concentrate around the major mode of the data.', 'In this work, we propose to learn a diversity sampling function (DSF) that generates a diverse yet likely set of future trajectories.', 'The DSF maps forecasting context features to a set of latent codes which can be decoded by a generative model (e.g., VAE) into a set of diverse trajectory samples.', 'Concretely, the process of identifying the diverse set of samples is posed as DSF parameter estimation.', 'To learn the parameters of the DSF, the diversity of the trajectory samples is evaluated by a diversity loss based on a determinantal point process (DPP).', 'Gradient descent is performed over the DSF parameters, which in turn moves the latent codes of the sample set to find an optimal set of diverse yet likely trajectories.', 'Our method is a novel application of DPPs to optimize a set of items (forecasted trajectories) in continuous space.', 'We demonstrate the diversity of the trajectories produced by our approach on both low-dimensional 2D trajectory data and high-dimensional human motion data.']
### SUMMARY:
| [
"We learn a diversity sampling function with DPPs to obtain a diverse set of samples from a generative model."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['There is mounting evidence that pretraining can be valuable for neural network language understanding models, but we do not yet have a clear understanding of how the choice of pretraining objective affects the type of linguistic information that models learn.', 'With this in mind, we compare four objectives---language modeling, translation, skip-thought, and autoencoding---on their ability to induce syntactic and part-of-speech information, holding constant the genre and quantity of training data.', 'We find that representations from language models consistently perform best on our syntactic auxiliary prediction tasks, even when trained on relatively small amounts of data, which suggests that language modeling may be the best data-rich pretraining task for transfer learning applications requiring syntactic information.', 'We also find that a randomly-initialized, frozen model can perform strikingly well on our auxiliary tasks, but that this effect disappears when the amount of training data for the auxiliary tasks is reduced.']
### SUMMARY:
| [
"Representations from language models consistently perform better than translation encoders on syntactic auxiliary prediction tasks."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['We consider the problem of generating configurations that satisfy physical constraints for optimal material nano-pattern design, where multiple (and often conflicting) properties need to be simultaneously satisfied. ', 'Consider, for example, the trade-off between thermal resistance, electrical conductivity, and mechanical stability needed to design a nano-porous template with optimal thermoelectric efficiency. ', 'To that end, we leverage the posterior regularization framework andshow that this constraint satisfaction problem can be formulated as sampling froma Gibbs distribution. ', 'The main challenges come from the black-box nature ofthose physical constraints, since they are obtained via solving highly non-linearPDEs.', 'To overcome those difficulties, we introduce Surrogate-based Constrained Langevin dynamics for black-box sampling.', 'We explore two surrogate approaches.', 'The first approach exploits zero-order approximation of gradients in the Langevin Sampling and we refer to it as Zero-Order Langevin.', 'In practice, this approach can be prohibitive since we still need to often query the expensive PDE solvers.', 'The second approach approximates the gradients in the Langevin dynamics with deep neural networks, allowing us an efficient sampling strategy using the surrogate model.', 'We prove the convergence of those two approaches when the target distribution is log-concave and smooth.', 'We show the effectiveness of both approaches in designing optimal nano-porous material configurations, where the goal is to produce nano-pattern templates with low thermal conductivity and reasonable mechanical stability.']
### SUMMARY:
| [
"We propose surrogate based Constrained Langevin sampling with application in nano-porous material configuration design."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['There is growing interest in geometrically-inspired embeddings for learning hierarchies, partial orders, and lattice structures, with natural applications to transitive relational data such as entailment graphs.', 'Recent work has extended these ideas beyond deterministic hierarchies to probabilistically calibrated models, which enable learning from uncertain supervision and inferring soft-inclusions among concepts, while maintaining the geometric inductive bias of hierarchical embedding models.', 'We build on the Box Lattice model of Vilnis et al. (2018), which showed promising results in modeling soft-inclusions through an overlapping hierarchy of sets, parameterized as high-dimensional hyperrectangles (boxes).', 'However, the hard edges of the boxes present difficulties for standard gradient based optimization; that work employed a special surrogate function for the disjoint case, but we find this method to be fragile. ', 'In this work, we present a novel hierarchical embedding model, inspired by a relaxation of box embeddings into parameterized density functions using Gaussian convolutions over the boxes.', 'Our approach provides an alternative surrogate to the original lattice measure that improves the robustness of optimization in the disjoint case, while also preserving the desirable properties with respect to the original lattice.', 'We demonstrate increased or matching performance on WordNet hypernymy prediction, Flickr caption entailment, and a MovieLens-based market basket dataset.', 'We show especially marked improvements in the case of sparse data, where many conditional probabilities should be low, and thus boxes should be nearly disjoint.']
### SUMMARY:
| [
"Improve hierarchical embedding models using kernel smoothing"
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['We present a weakly-supervised data augmentation approach to improve Named Entity Recognition (NER) in a challenging domain: extracting biomedical entities (e.g., proteins) from the scientific literature.', 'First, we train a neural NER (NNER) model over a small seed of fully-labeled examples.', 'Second, we use a reference set of entity names (e.g., proteins in UniProt) to identify entity mentions with high precision, but low recall, on an unlabeled corpus.', 'Third, we use the NNER model to assign weak labels to the corpus.', 'Finally, we retrain our NNER model iteratively over the augmented training set, including the seed, the reference-set examples, and the weakly-labeled examples, which results in refined labels.', 'We show empirically that this augmented bootstrapping process significantly improves NER performance, and discuss the factors impacting the efficacy of the approach.']
### SUMMARY:
| [
"Augmented bootstrapping approach combining information from a reference set with iterative refinements of soft labels to improve Name Entity Recognition from biomedical literature."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Quantum machine learning methods have the potential to facilitate learning using extremely large datasets.', 'While the availability of data for training machine learning models is steadily increasing, oftentimes it is much easier to collect feature vectors that to obtain the corresponding labels.', 'One of the approaches for addressing this issue is to use semi-supervised learning, which leverages not only the labeled samples, but also unlabeled feature vectors.', 'Here, we present a quantum machine learning algorithm for training Semi-Supervised Kernel Support Vector Machines.', 'The algorithm uses recent advances in quantum sample-based Hamiltonian simulation to extend the existing Quantum LS-SVM algorithm to handle the semi-supervised term in the loss, while maintaining the same quantum speedup as the Quantum LS-SVM.']
### SUMMARY:
| [
"We extend quantum SVMs to semi-supervised setting, to deal with the likely problem of many missing class labels in huge datasets."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Deep neural networks have become the state-of-the-art models in numerous machine learning tasks.', 'However, general guidance to network architecture design is still missing.', 'In our work, we bridge deep neural network design with numerical differential equations.', 'We show that many effective networks, such as ResNet, PolyNet, FractalNet and RevNet, can be interpreted as different numerical discretizations of differential equations.', 'This finding brings us a brand new perspective on the design of effective deep architectures.', 'We can take advantage of the rich knowledge in numerical analysis to guide us in designing new and potentially more effective deep networks.', 'As an example, we propose a linear multi-step architecture (LM-architecture) which is inspired by the linear multi-step method solving ordinary differential equations.', 'The LM-architecture is an effective structure that can be used on any ResNet-like networks.', 'In particular, we demonstrate that LM-ResNet and LM-ResNeXt (i.e. the networks obtained by applying the LM-architecture on ResNet and ResNeXt respectively) can achieve noticeably higher accuracy than ResNet and ResNeXt on both CIFAR and ImageNet with comparable numbers of trainable parameters.', 'In particular, on both CIFAR and ImageNet, LM-ResNet/LM-ResNeXt can significantly compress (>50%) the original networks while maintaining a similar performance.', 'This can be explained mathematically using the concept of modified equation from numerical analysis.', 'Last but not least, we also establish a connection between stochastic control and noise injection in the training process which helps to improve generalization of the networks.', 'Furthermore, by relating stochastic training strategy with stochastic dynamic system, we can easily apply stochastic training to the networks with the LM-architecture.', 'As an example, we introduced stochastic depth to LM-ResNet and achieve significant improvement over the original LM-ResNet on CIFAR10.']
### SUMMARY:
| [
"This paper bridges deep network architectures with numerical (stochastic) differential equations. This new perspective enables new designs of more effective deep neural networks."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Transforming a graphical user interface screenshot created by a designer into computer code is a typical task conducted by a developer in order to build customized software, websites, and mobile applications.', 'In this paper, we show that deep learning methods can be leveraged to train a model end-to-end to automatically generate code from a single input image with over 77% of accuracy for three different platforms (i.e. iOS, Android and web-based technologies).']
### SUMMARY:
| [
"CNN and LSTM to generate markup-like code describing graphical user interface images."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Computer vision tasks such as image classification, image retrieval and few-shot learning are currently dominated by Euclidean and spherical embeddings, so that the final decisions about class belongings or the degree of similarity are made using linear hyperplanes, Euclidean distances, or spherical geodesic distances (cosine similarity).', 'In this work, we demonstrate that in many practical scenarios hyperbolic embeddings provide a better alternative.']
### SUMMARY:
| [
"We show that hyperbolic embeddings are useful for high-level computer vision tasks, especially for few-shot classification."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['High-dimensional time series are common in many domains.', 'Since human cognition is not optimized to work well in high-dimensional spaces, these areas could benefit from interpretable low-dimensional representations.', 'However, most representation learning algorithms for time series data are difficult to interpret.', 'This is due to non-intuitive mappings from data features to salient properties of the representation and non-smoothness over time.\n', 'To address this problem, we propose a new representation learning framework building on ideas from interpretable discrete dimensionality reduction and deep generative modeling.', 'This framework allows us to learn discrete representations of time series, which give rise to smooth and interpretable embeddings with superior clustering performance.', 'We introduce a new way to overcome the non-differentiability in discrete representation learning and present a gradient-based version of the traditional self-organizing map algorithm that is more performant than the original.', 'Furthermore, to allow for a probabilistic interpretation of our method, we integrate a Markov model in the representation space.\n', 'This model uncovers the temporal transition structure, improves clustering performance even further and provides additional explanatory insights as well as a natural representation of uncertainty.\n', 'We evaluate our model in terms of clustering performance and interpretability on static (Fashion-)MNIST data, a time series of linearly interpolated (Fashion-)MNIST images, a chaotic Lorenz attractor system with two macro states, as well as on a challenging real world medical time series application on the eICU data set.', 'Our learned representations compare favorably with competitor methods and facilitate downstream tasks on the real world data.']
### SUMMARY:
| [
"We present a method to learn interpretable representations on time series using ideas from variational autoencoders, self-organizing maps and probabilistic models."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['We propose Significance-Offset Convolutional Neural Network, a deep convolutional network architecture for regression of multivariate asynchronous time series. ', 'The model is inspired by standard autoregressive (AR) models and gating mechanisms used in recurrent neural networks. ', 'It involves an AR-like weighting system, where the final predictor is obtained as a weighted sum of adjusted regressors, while the weights are data-dependent functions learnt through a convolutional network.', 'The architecture was designed for applications on asynchronous time series and is evaluated on such datasets: a hedge fund proprietary dataset of over 2 million quotes for a credit derivative index, an artificially generated noisy autoregressive series and household electricity consumption dataset. ', 'The pro-posed architecture achieves promising results as compared to convolutional and recurrent neural networks.', 'The code for the numerical experiments and the architecture implementation will be shared online to make the research reproducible.']
### SUMMARY:
| [
"Convolutional architecture for learning data-dependent weights for autoregressive forecasting of time series."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['MixUp is a data augmentation scheme in which pairs of training samples and their corresponding labels are mixed using linear coefficients.', 'Without label mixing, MixUp becomes a more conventional scheme: input samples are moved but their original labels are retained.', 'Because samples are preferentially moved in the direction of other classes \\iffalse -- which are typically clustered in input space -- \\fi we refer to this method as directional adversarial training, or DAT.', 'We show that under two mild conditions, MixUp asymptotically convergences to a subset of DAT.', 'We define untied MixUp (UMixUp), a superset of MixUp wherein training labels are mixed with different linear coefficients to those of their corresponding samples.', 'We show that under the same mild conditions, untied MixUp converges to the entire class of DAT schemes.', 'Motivated by the understanding that UMixUp is both a generalization of MixUp and a form of adversarial training, we experiment with different datasets and loss functions to show that UMixUp provides improved performance over MixUp.', 'In short, we present a novel interpretation of MixUp as belonging to a class highly analogous to adversarial training, and on this basis we introduce a simple generalization which outperforms MixUp.']
### SUMMARY:
| [
"We present a novel interpretation of MixUp as belonging to a class highly analogous to adversarial training, and on this basis we introduce a simple generalization which outperforms MixUp"
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Plan recognition aims to look for target plans to best explain the observed actions based on plan libraries and/or domain models.', 'Despite the success of previous approaches on plan recognition, they mostly rely on correct action observations. \n', 'Recent advances in visual activity recognition have the potential of enabling applications such as automated video surveillance.', 'Effective approaches for such problems would require the ability to recognize the plans of agents from video information.', 'Traditional plan recognition algorithms rely on access to detailed planning domain models.', 'One recent promising direction involves learning approximate (or shallow) domain models directly from the observed activity sequences.', 'Such plan recognition approaches expect observed action sequences as inputs.', 'However, visual inference results are often noisy and uncertain, typically represented as a distribution over possible actions.', 'In this work, we develop a visual plan recognition framework that recognizes plans with an approximate domain model learned from uncertain visual data.']
### SUMMARY:
| [
"Handling Uncertainty in Visual Perception for Plan Recognition"
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['We consider the task of answering complex multi-hop questions using a corpus as a virtual knowledge base (KB).', 'In particular, we describe a neural module, DrKIT, that traverses textual data like a virtual KB, softly following paths of relations between mentions of entities in the corpus.', 'At each step the operation uses a combination of sparse-matrix TFIDF indices and maximum inner product search (MIPS) on a special index of contextual representations.', 'This module is differentiable, so the full system can be trained completely end-to-end using gradient based methods, starting from natural language inputs.', 'We also describe a pretraining scheme for the index mention encoder by generating hard negative examples using existing knowledge bases.', 'We show that DrKIT improves accuracy by 9 points on 3-hop questions in the MetaQA dataset, cutting the gap between text-based and KB-based state-of-the-art by 70%.', 'DrKIT is also very efficient, processing upto 10x more queries per second than existing state-of-the-art QA systems.']
### SUMMARY:
| [
"Differentiable multi-hop access to a textual knowledge base of indexed contextual representations"
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['In spite of their great success, traditional factorization algorithms typically do not support features (e.g., Matrix Factorization), or their complexity scales quadratically with the number of features (e.g, Factorization Machine).', 'On the other hand, neural methods allow large feature sets, but are often designed for a specific application.', 'We propose novel deep factorization methods that allow efficient and flexible feature representation.', 'For example, we enable describing items with natural language with complexity linear to the vocabulary size—this enables prediction for unseen items and avoids the cold start problem.', 'We show that our architecture can generalize some previously published single-purpose neural architectures.', 'Our experiments suggest improved training times and accuracy compared to shallow methods.']
### SUMMARY:
| [
"Scalable general-purpose factorization algorithm-- also helps to circumvent cold start problem."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Augmented Reality (AR) can assist with physical tasks such as object assembly through the use of “situated instructions”.', 'These instructions can be in the form of videos, pictures, text or guiding animations, where the most helpful media among these is highly dependent on both the user and the nature of the task.', 'Our work supports the authoring of AR tutorials for assembly tasks with little overhead beyond simply performing the task itself.', 'The presented system, AuthAR reduces the time and effort required to build interactive AR tutorials by automatically generating key components of the AR tutorial while the author is assembling the physical pieces.', 'Further, the system guides authors through the process of adding videos, pictures, text and animations to the tutorial.', 'This concurrent assembly and tutorial generation approach allows for authoring of portable tutorials that fit the preferences of different end users.']
### SUMMARY:
| [
"We present a mixed media assembly tutorial authoring system that streamlines creation of videos, images, text and dynamic instructions in situ."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Monitoring patients in ICU is a challenging and high-cost task.', "Hence, predicting the condition of patients during their ICU stay can help provide better acute care and plan the hospital's resources.", 'There has been continuous progress in machine learning research for ICU management, and most of this work has focused on using time series signals recorded by ICU instruments.', 'In our work, we show that adding clinical notes as another modality improves the performance of the model for three benchmark tasks: in-hospital mortality prediction, modeling decompensation, and length of stay forecasting that play an important role in ICU management.', 'While the time-series data is measured at regular intervals, doctor notes are charted at irregular times, making it challenging to model them together.', 'We propose a method to model them jointly, achieving considerable improvement across benchmark tasks over baseline time-series model.']
### SUMMARY:
| [
"We demostarte that using clinical notes in conjuntion with ICU instruments data improves the perfomance on ICU management benchmark tasks"
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Existing works in deep Multi-Agent Reinforcement Learning (MARL) mainly focus on coordinating cooperative agents to complete certain tasks jointly.', 'However, in many cases of the real world, agents are self-interested such as employees in a company and clubs in a league.', 'Therefore, the leader, i.e., the manager of the company or the league, needs to provide bonuses to followers for efficient coordination, which we call expensive coordination.', 'The main difficulties of expensive coordination are that', "i) the leader has to consider the long-term effect and predict the followers' behaviors when assigning bonuses and", "ii) the complex interactions between followers make the training process hard to converge, especially when the leader's policy changes with time.", 'In this work, we address this problem through an event-based deep RL approach.', 'Our main contributions are threefold.', "(1) We model the leader's decision-making process as a semi-Markov Decision Process and propose a novel multi-agent event-based policy gradient to learn the leader's long-term policy.", "(2) We exploit the leader-follower consistency scheme to design a follower-aware module and a follower-specific attention module to predict the followers' behaviors and make accurate response to their behaviors.", "(3) We propose an action abstraction-based policy gradient algorithm to reduce the followers' decision space and thus accelerate the training process of followers.", 'Experiments in resource collections, navigation, and the predator-prey game reveal that our approach outperforms the state-of-the-art methods dramatically.']
### SUMMARY:
| [
"We propose an event-based policy gradient to train the leader and an action abstraction policy gradient to train the followers in leader-follower Markov game."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Recent work has studied the emergence of language among deep reinforcement learning agents that must collaborate to solve a task.', 'Of particular interest are the factors that cause language to be compositional---i.e., express meaning by combining words which themselves have meaning.', 'Evolutionary linguists have found that in addition to structural priors like those already studied in deep learning, the dynamics of transmitting language from generation to generation contribute significantly to the emergence of compositionality.', 'In this paper, we introduce these cultural evolutionary dynamics into language emergence by periodically replacing agents in a population to create a knowledge gap, implicitly inducing cultural transmission of language.', 'We show that this implicit cultural transmission encourages the resulting languages to exhibit better compositional generalization.']
### SUMMARY:
| [
"We use cultural transmission to encourage compositionality in languages that emerge from interactions between neural agents."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Based on our observation that there exists a dramatic drop for the singular values of the fully connected layers or a single feature map of the convolutional layer, and that the dimension of the concatenated feature vector almost equals the summation of the dimension on each feature map, we propose a singular value decomposition (SVD) based approach to estimate the dimension of the deep manifolds for a typical convolutional neural network VGG19.', 'We choose three categories from the ImageNet, namely Persian Cat, Container Ship and Volcano, and determine the local dimension of the deep manifolds of the deep layers through the tangent space of a target image.', 'Through several augmentation methods, we found that the Gaussian noise method is closer to the intrinsic dimension, as by adding random noise to an image we are moving in an arbitrary dimension, and when the rank of the feature matrix of the augmented images does not increase we are very close\n', 'to the local dimension of the manifold.', 'We also estimate the dimension of the deep manifold based on the tangent space for each of the maxpooling layers.', 'Our results show that the dimensions of different categories are close to each other and decline quickly along the convolutional layers and fully connected layers.', 'Furthermore, we show that the dimensions decline quickly inside the Conv5 layer.', 'Our work provides new insights for the intrinsic structure of deep neural networks and helps unveiling the inner organization of the black box of deep neural networks.']
### SUMMARY:
| [
"We propose a SVD based method to explore the local dimension of activation manifold in deep neural networks."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Large pre-trained Transformers such as BERT have been tremendously effective for many NLP tasks.', ' However, inference in these large-capacity models is prohibitively slow and expensive', '. Transformers are essentially a stack of self-attention layers which encode each input position using the entire input sequence as its context', '. However, we find that it may not be necessary to apply this expensive sequence-wide self-attention over at all layers', '. Based on this observation, we propose a decomposition to a pre-trained Transformer that allows the lower layers to process segments of the input independently enabling parallelism and caching', '. We show that the information loss due to this decomposition can be recovered in the upper layers with auxiliary supervision during fine-tuning', '. We evaluate de-composition with pre-trained BERT models on five different paired-input tasks in question answering, sentence similarity, and natural language inference', '. Results show that decomposition enables faster inference (up to 4x), significant memory reduction (up to 70%) while retaining most (up to 99%) of the original performance', '. We will release the code at<anonymized url>.']
### SUMMARY:
| [
"Inference in large Transformers is expensive due to the self-attention in multiple layers. We show a simple decomposition technique can yield a faster, low memory-footprint model that is just as accurate of the original models."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Exploration while learning representations is one of the main challenges Deep\n', 'Reinforcement Learning (DRL) faces today.', 'As the learned representation is dependant in the observed data, the exploration strategy has a crucial role.', 'The popular DQN algorithm has improved significantly the capabilities of Reinforcement\n', 'Learning (RL) algorithms to learn state representations from raw data, yet, it uses\n', 'a naive exploration strategy which is statistically inefficient.', 'The Randomized\n', 'Least Squares Value Iteration (RLSVI) algorithm (Osband et al., 2016), on the\n', 'other hand, explores and generalizes efficiently via linearly parameterized value\n', 'functions.', 'However, it is based on hand-designed state representation that requires\n', 'prior engineering work for every environment.', 'In this paper, we propose a Deep\n', 'Learning adaptation for RLSVI.', 'Rather than using hand-design state representation, we use a state representation that is being learned directly from the data by a\n', 'DQN agent.', 'As the representation is being optimized during the learning process,\n', 'a key component for the suggested method is a likelihood matching mechanism,\n', 'which adapts to the changing representations.', 'We demonstrate the importance of\n', 'the various properties of our algorithm on a toy problem and show that our method\n', 'outperforms DQN in five Atari benchmarks, reaching competitive results with the\n', 'Rainbow algorithm.']
### SUMMARY:
| [
"A Deep Learning adaptation of Randomized Least Squares Value Iteration"
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['The complexity of large-scale neural networks can lead to poor understanding of their internal details.', 'We show that this opaqueness provides an opportunity for adversaries to embed unintended functionalities into the network in the form of Trojan horse attacks.', 'Our novel framework hides the existence of a malicious network within a benign transport network.', 'Our attack is flexible, easy to execute, and difficult to detect.', "We prove theoretically that the malicious network's detection is computationally infeasible and demonstrate empirically that the transport network does not compromise its disguise.", 'Our attack exposes an important, previously unknown loophole that unveils a new direction in machine learning security.']
### SUMMARY:
| [
"Parameters of a trained neural network can be permuted to produce a completely separate model for a different task, enabling the embedding of Trojan horse networks inside another network."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['In this paper, we introduce Random Path Generative Adversarial Network (RPGAN) --- an alternative scheme of GANs that can serve as a tool for generative model analysis.', 'While the latent space of a typical GAN consists of input vectors, randomly sampled from the standard Gaussian distribution, the latent space of RPGAN consists of random paths in a generator network.', 'As we show, this design allows to associate different layers of the generator with different regions of the latent space, providing their natural interpretability.', 'With experiments on standard benchmarks, we demonstrate that RPGAN reveals several interesting insights about roles that different layers play in the image generation process.', 'Aside from interpretability, the RPGAN model also provides competitive generation quality and allows efficient incremental learning on new data.']
### SUMMARY:
| [
"We introduce an alternative GAN design based on random routes in generator, which can serve as a tool for generative models interpretability."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Deep artificial neural networks can achieve an extremely small difference between training and test accuracies on identically distributed training and test sets, which is a standard measure of generalization.', 'However, the training and test sets may not be sufficiently representative of the empirical sample set, which consists of real-world input samples.', 'When samples are drawn from an underrepresented or unrepresented subset during inference, the gap between the training and inference accuracies can be significant.', 'To address this problem, we first reformulate a classification algorithm as a procedure for searching for a source code that maps input features to classes.', 'We then derive a necessary and sufficient condition for generalization using a universal cognitive similarity metric, namely information distance, based on Kolmogorov complexity.', 'Using this condition, we formulate an optimization problem to learn a more general classification function.', 'To achieve this end, we extend the input features by concatenating encodings of them, and then train the classifier on the extended features.', 'As an illustration of this idea, we focus on image classification, where we use channel codes on the input features as a systematic way to improve the degree to which the training and test sets are representative of the empirical sample set.', 'To showcase our theoretical findings, considering that corrupted or perturbed input features belong to the empirical sample set, but typically not to the training and test sets, we demonstrate through extensive systematic experiments that, as a result of learning a more general classification function, a model trained on encoded input features is significantly more robust to common corruptions, e.g., Gaussian and shot noise, as well as adversarial perturbations, e.g., those found via projected gradient descent, than the model trained on uncoded input features.']
### SUMMARY:
| [
"We present a theoretical and experimental framework for defining, understanding, and achieving generalization, and as a result robustness, in deep learning by drawing on algorithmic information theory and coding theory."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Many approaches to causal discovery are limited by their inability to discriminate between Markov equivalent graphs given only observational data.', 'We formulate causal discovery as a marginal likelihood based Bayesian model selection problem.', 'We adopt a parameterization based on the notion of the independence of causal mechanisms which renders Markov equivalent graphs distinguishable.', 'We complement this with an empirical Bayesian approach to setting priors so that the actual underlying causal graph is assigned a higher marginal likelihood than its alternatives.', 'Adopting a Bayesian approach also allows for straightforward modeling of unobserved confounding variables, for which we provide a variational algorithm to approximate the marginal likelihood, since this desirable feat renders the computation of the marginal likelihood intractable.', 'We believe that the Bayesian approach to causal discovery both allows the rich methodology of Bayesian inference to be used in various difficult aspects of this problem and provides a unifying framework to causal discovery research.', 'We demonstrate promising results in experiments conducted on real data, supporting our modeling approach and our inference methodology.']
### SUMMARY:
| [
"We cast causal structure discovery as a Bayesian model selection in a way that allows us to discriminate between Markov equivalent graphs to identify the unique causal graph."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
[' The goal of compressed sensing is to learn a structured signal $x$\n from a limited number of noisy linear measurements $y', "\\approx Ax$. In\n traditional compressed sensing, ``structure'' is represented by\n sparsity in some known basis. ", 'Inspired by the success of deep\n learning in modeling images, recent work starting with~\\cite{BDJP17}\n has instead considered structure to come from a generative model\n $G: \\R^k \\to \\R^n$. We present two results establishing the\n difficulty of this latter task, showing that existing bounds are\n tight.\n\n ', 'First, we provide a lower bound matching the~\\cite{BDJP17} upper\n bound for compressed sensing from $L$-Lipschitz generative models\n $G$. In particular, there exists such a function that requires\n roughly $\\Omega(k \\log L)$ linear measurements for sparse recovery\n to be possible. ', 'This holds even for the more relaxed goal of\n \\emph{nonuniform} recovery.\n\n ', 'Second, we show that generative models generalize sparsity as a\n representation of structure. ', 'In particular, we construct a\n ReLU-based neural network $G: \\R^{2k} \\to \\R^n$ with $O(1)$ layers\n and $O(kn)$ activations per layer, such that the range of $G$\n contains all $k$-sparse vectors.\n']
### SUMMARY:
| [
"Lower bound for compressed sensing w/ generative models that matches known upper bounds"
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['We hypothesize that end-to-end neural image captioning systems work seemingly well because they exploit and learn ‘distributional similarity’ in a multimodal feature space, by mapping a test image to similar training images in this space and generating a caption from the same space.', 'To validate our hypothesis, we focus on the ‘image’ side of image captioning, and vary the input image representation but keep the RNN text generation model of a CNN-RNN constant.', 'We propose a sparse bag-of-objects vector as an interpretable representation to investigate our distributional similarity hypothesis.', 'We found that image captioning models', '(i) are capable of separating structure from noisy input representations;', '(ii) experience virtually no significant performance loss when a high dimensional representation is compressed to a lower dimensional space;', '(iii) cluster images with similar visual and linguistic information together;', '(iv) are heavily reliant on test sets with a similar distribution as the training set;', '(v) repeatedly generate the same captions by matching images and ‘retrieving’ a caption in the joint visual-textual space.', 'Our experiments all point to one fact: that our distributional similarity hypothesis holds.', 'We conclude that, regardless of the image representation, image captioning systems seem to match images and generate captions in a learned joint image-text semantic subspace.\n']
### SUMMARY:
| [
"This paper presents an empirical analysis on the role of different types of image representations and probes the properties of these representations for the task of image captioning."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['We present a sequence-to-action parsing approach for the natural language to SQL task that incrementally fills the slots of a SQL query with feasible actions from a pre-defined inventory.', 'To account for the fact that typically there are multiple correct SQL queries with the same or very similar semantics, we draw inspiration from syntactic parsing techniques and propose to train our sequence-to-action models with non-deterministic oracles.', 'We evaluate our models on the WikiSQL dataset and achieve an execution accuracy of 83.7% on the test set, a 2.1% absolute improvement over the models trained with traditional static oracles assuming a single correct target SQL query.', 'When further combined with the execution-guided decoding strategy, our model sets a new state-of-the-art performance at an execution accuracy of 87.1%.']
### SUMMARY:
| [
"We design incremental sequence-to-action parsers for text-to-SQL task and achieve SOTA results. We further improve by using non-deterministic oracles to allow multiple correct action sequences. "
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['We propose Efficient Neural Architecture Search (ENAS), a faster and less expensive approach to automated model design than previous methods.', 'In ENAS, a controller learns to discover neural network architectures by searching for an optimal path within a larger model.', 'The controller is trained with policy gradient to select a path that maximizes the expected reward on the validation set.', 'Meanwhile the model corresponding to the selected path is trained to minimize the cross entropy loss.', 'On the Penn Treebank dataset, ENAS can discover a novel architecture thats achieves a test perplexity of 57.8, which is state-of-the-art among automatic model design methods on Penn Treebank.', 'On the CIFAR-10 dataset, ENAS can design novel architectures that achieve a test error of 2.89%, close to the 2.65% achieved by standard NAS (Zoph et al., 2017).', 'Most importantly, our experiments show that ENAS is more than 10x faster and 100x less resource-demanding than NAS.']
### SUMMARY:
| [
"An approach that speeds up neural architecture search by 10x, whilst using 100x less computing resource."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Nowadays, deep neural networks (DNNs) have become the main instrument for machine learning tasks within a wide range of domains, including vision, NLP, and speech.', 'Meanwhile, in an important case of heterogenous tabular data, the advantage of DNNs over shallow counterparts remains questionable.', 'In particular, there is no sufficient evidence that deep learning machinery allows constructing methods that outperform gradient boosting decision trees (GBDT), which are often the top choice for tabular problems.', 'In this paper, we introduce Neural Oblivious Decision Ensembles (NODE), a new deep learning architecture, designed to work with any tabular data.', 'In a nutshell, the proposed NODE architecture generalizes ensembles of oblivious decision trees, but benefits from both end-to-end gradient-based optimization and the power of multi-layer hierarchical representation learning.', 'With an extensive experimental comparison to the leading GBDT packages on a large number of tabular datasets, we demonstrate the advantage of the proposed NODE architecture, which outperforms the competitors on most of the tasks.', 'We open-source the PyTorch implementation of NODE and believe that it will become a universal framework for machine learning on tabular data.']
### SUMMARY:
| [
"We propose a new DNN architecture for deep learning on tabular data"
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
["Person re-identification (re-ID) aims at identifying the same persons' images across different cameras.", 'However, domain diversities between different datasets pose an evident challenge for adapting the re-ID model trained on one dataset to another one.', 'State-of-the-art unsupervised domain adaptation methods for person re-ID transferred the learned knowledge from the source domain by optimizing with pseudo labels created by clustering algorithms on the target domain.', 'Although they achieved state-of-the-art performances, the inevitable label noise caused by the clustering procedure was ignored.', "Such noisy pseudo labels substantially hinders the model's capability on further improving feature representations on the target domain.", 'In order to mitigate the effects of noisy pseudo labels, we propose to softly refine the pseudo labels in the target domain by proposing an unsupervised framework, Mutual Mean-Teaching (MMT), to learn better features from the target domain via off-line refined hard pseudo labels and on-line refined soft pseudo labels in an alternative training manner. ', 'In addition, the common practice is to adopt both the classification loss and the triplet loss jointly for achieving optimal performances in person re-ID models.', 'However, conventional triplet loss cannot work with softly refined labels.', 'To solve this problem, a novel soft softmax-triplet loss is proposed to support learning with soft pseudo triplet labels for achieving the optimal domain adaptation performance.', 'The proposed MMT framework achieves considerable improvements of 14.4%, 18.2%, 13.1% and 16.4% mAP on Market-to-Duke, Duke-to-Market, Market-to-MSMT and Duke-to-MSMT unsupervised domain adaptation tasks.']
### SUMMARY:
| [
"A framework that conducts online refinement of pseudo labels with a novel soft softmax-triplet loss for unsupervised domain adaptation on person re-identification."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['We present the first end-to-end verifier of audio classifiers.', 'Compared to existing methods, our approach enables analysis of both, the entire audio processing stage as well as recurrent neural network architectures (e.g., LSTM).', 'The audio processing is verified using novel convex relaxations tailored to feature extraction operations used in audio (e.g., Fast Fourier Transform) while recurrent architectures are certified via a novel binary relaxation for the recurrent unit update.', 'We show the verifier scales to large networks while computing significantly tighter bounds than existing methods for common audio classification benchmarks: on the challenging Google Speech Commands dataset we certify 95% more inputs than the interval approximation (only prior scalable method), for a perturbation of -90dB.']
### SUMMARY:
| [
"We present the first approach to certify robustness of neural networks against noise-based perturbations in the audio domain."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Since deep neural networks are over-parameterized, they can memorize noisy examples.', 'We address such memorizing issue in the presence of annotation noise.', 'From the fact that deep neural networks cannot generalize neighborhoods of the features acquired via memorization, we hypothesize that noisy examples do not consistently incur small losses on the network under a certain perturbation.', 'Based on this, we propose a novel training method called Learning with Ensemble Consensus (LEC) that prevents overfitting noisy examples by eliminating them using the consensus of an ensemble of perturbed networks.', 'One of the proposed LECs, LTEC outperforms the current state-of-the-art methods on noisy MNIST, CIFAR-10, and CIFAR-100 in an efficient manner.']
### SUMMARY:
| [
"This work presents a method of generating and using ensembles effectively to identify noisy examples in the presence of annotation noise. "
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Recovering sparse conditional independence graphs from data is a fundamental problem in machine learning with wide applications.', 'A popular formulation of the problem is an $\\ell_1$ regularized maximum likelihood estimation.', 'Many convex optimization algorithms have been designed to solve this formulation to recover the graph structure.', 'Recently, there is a surge of interest to learn algorithms directly based on data, and in this case, learn to map empirical covariance to the sparse precision matrix.', 'However, it is a challenging task in this case, since the symmetric positive definiteness (SPD) and sparsity of the matrix are not easy to enforce in learned algorithms, and a direct mapping from data to precision matrix may contain many parameters.', 'We propose a deep learning architecture, GLAD, which uses an Alternating Minimization (AM) algorithm as our model inductive bias, and learns the model parameters via supervised learning.', 'We show that GLAD learns a very compact and effective model for recovering sparse graphs from data.']
### SUMMARY:
| [
"A data-driven learning algorithm based on unrolling the Alternating Minimization optimization for sparse graph recovery."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Counterfactual regret minimization (CFR) is a fundamental and effective technique for solving Imperfect Information Games (IIG).', 'However, the original CFR algorithm only works for discrete states and action spaces, and the resulting strategy is maintained as a tabular representation.', 'Such tabular representation limits the method from being directly applied to large games.', 'In this paper, we propose a double neural representation for the IIGs, where one neural network represents the cumulative regret, and the other represents the average strategy. ', 'Such neural representations allow us to avoid manual game abstraction and carry out end-to-end optimization.', 'To make the learning efficient, we also developed several novel techniques including a robust sampling method and a mini-batch Monte Carlo Counterfactual Regret Minimization (MCCFR) method, which may be of independent interests. ', 'Empirically, on games tractable to tabular approaches, neural strategies trained with our algorithm converge comparably to their tabular counterparts, and significantly outperform those based on deep reinforcement learning. ', 'On extremely large games with billions of decision nodes, our approach achieved strong performance while using hundreds of times less memory than the tabular CFR.', "On head-to-head matches of hands-up no-limit texas hold'em, our neural agent beat the strong agent ABS-CFR by $9.8\\pm4.1$ chips per game.", "It's a successful application of neural CFR in large games.\n"]
### SUMMARY:
| [
"We proposed a double neural framework to solve large-scale imperfect information game. "
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['We present the first verification that a neural network for perception tasks produces\n', 'a correct output within a specified tolerance for every input of interest.', 'We define\n', 'correctness relative to a specification which identifies 1) a state space consisting of\n', 'all relevant states of the world and 2) an observation process that produces neural\n', 'network inputs from the states of the world.', 'Tiling the state and input spaces with\n', 'a finite number of tiles, obtaining ground truth bounds from the state tiles and\n', 'network output bounds from the input tiles, then comparing the ground truth and\n', 'network output bounds delivers an upper bound on the network output error for\n', 'any input of interest.', 'Results from two case studies highlight the ability of our\n', 'technique to deliver tight error bounds for all inputs of interest and show how the\n', 'error bounds vary over the state and input spaces.']
### SUMMARY:
| [
"We present the first verification that a neural network for perception tasks produces a correct output within a specified tolerance for every input of interest. "
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Deep generative models have achieved remarkable progress in recent years.', 'Despite this progress, quantitative evaluation and comparison of generative models remains as one of the important challenges.', 'One of the most popular metrics for evaluating generative models is the log-likelihood.', 'While the direct computation of log-likelihood can be intractable, it has been recently shown that the log-likelihood of some of the most interesting generative models such as variational autoencoders (VAE) or generative adversarial networks (GAN) can be efficiently estimated using annealed importance sampling (AIS).', 'In this work, we argue that the log-likelihood metric by itself cannot represent all the different performance characteristics of generative models, and propose to use rate distortion curves to evaluate and compare deep generative models.', 'We show that we can approximate the entire rate distortion curve using one single run of AIS for roughly the same computational cost as a single log-likelihood estimate.', 'We evaluate lossy compression rates of different deep generative models such as VAEs, GANs (and its variants) and adversarial autoencoders (AAE) on MNIST and CIFAR10, and arrive at a number of insights not obtainable from log-likelihoods alone.']
### SUMMARY:
| [
"We study rate distortion approximations for evaluating deep generative models, and show that rate distortion curves provide more insights about the model than the log-likelihood alone while requiring roughly the same computational cost."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Although reinforcement learning methods can achieve impressive results in simulation, the real world presents two major challenges: generating samples is exceedingly expensive, and unexpected perturbations or unseen situations cause proficient but specialized policies to fail at test time.', 'Given that it is impractical to train separate policies to accommodate all situations the agent may see in the real world, this work proposes to learn how to quickly and effectively adapt online to new tasks.', 'To enable sample-efficient learning, we consider learning online adaptation in the context of model-based reinforcement learning.', 'Our approach uses meta-learning to train a dynamics model prior such that, when combined with recent data, this prior can be rapidly adapted to the local context.', 'Our experiments demonstrate online adaptation for continuous control tasks on both simulated and real-world agents.', 'We first show simulated agents adapting their behavior online to novel terrains, crippled body parts, and highly-dynamic environments.', "We also illustrate the importance of incorporating online adaptation into autonomous agents that operate in the real world by applying our method to a real dynamic legged millirobot: We demonstrate the agent's learned ability to quickly adapt online to a missing leg, adjust to novel terrains and slopes, account for miscalibration or errors in pose estimation, and compensate for pulling payloads."]
### SUMMARY:
| [
"A model-based meta-RL algorithm that enables a real robot to adapt online in dynamic environments"
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Model-free deep reinforcement learning approaches have shown superhuman performance in simulated environments (e.g., Atari games, Go, etc).', 'During training, these approaches often implicitly construct a latent space that contains key information for decision making.', 'In this paper, we learn a forward model on this latent space and apply it to model-based planning in miniature Real-time Strategy game with incomplete information (MiniRTS).', 'We first show that the latent space constructed from existing actor-critic models contains relevant information of the game, and design training procedure to learn forward models.', 'We also show that our learned forward model can predict meaningful future state and is usable for latent space Monte-Carlo Tree Search (MCTS), in terms of win rates against rule-based agents.']
### SUMMARY:
| [
"The paper analyzes the latent space learned by model-free approaches in a miniature incomplete information game, trains a forward model in the latent space and apply it to Monte-Carlo Tree Search, yielding positive performance."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Several state of the art convolutional networks rely on inter-connecting different layers to ease the flow of information and gradient between their input and output layers.', 'These techniques have enabled practitioners to successfully train deep convolutional networks with hundreds of layers.', 'Particularly, a novel way of interconnecting layers was introduced as the Dense Convolutional Network (DenseNet) and has achieved state of the art performance on relevant image recognition tasks.', 'Despite their notable empirical success, their theoretical understanding is still limited.', 'In this work, we address this problem by analyzing the effect of layer interconnection on the overall expressive power of a convolutional network.', 'In particular, the connections used in DenseNet are compared with other types of inter-layer connectivity.', 'We carry out a tensor analysis on the expressive power inter-connections on convolutional arithmetic circuits (ConvACs) and relate our results to standard convolutional networks.', 'The analysis leads to performance bounds and practical guidelines for design of ConvACs.', 'The generalization of these results are discussed for other kinds of convolutional networks via generalized tensor decompositions.']
### SUMMARY:
| [
"We analyze the expressive power of the connections used in DenseNets via tensor decompositions."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['We consider the following central question in the field of Deep Reinforcement Learning (DRL):', 'How can we use implicit human feedback to accelerate and optimize the training of a DRL algorithm?', 'State-of-the-art methods rely on any human feedback to be provided explicitly, requiring the active participation of humans (e.g., expert labeling, demonstrations, etc.).', 'In this work, we investigate an alternative paradigm, where non-expert humans are silently observing (and assessing) the agent interacting with the environment.', "The human's intrinsic reactions to the agent's behavior is sensed as implicit feedback by placing electrodes on the human scalp and monitoring what are known as event-related electric potentials.", "The implicit feedback is then used to augment the agent's learning in the RL tasks.", 'We develop a system to obtain and accurately decode the implicit human feedback (specifically error-related event potentials) for state-action pairs in an Atari-type environment.', 'As a baseline contribution, we demonstrate the feasibility of capturing error-potentials of a human observer watching an agent learning to play several different Atari-games using an electroencephalogram (EEG) cap, and then decoding the signals appropriately and using them as an auxiliary reward function to a DRL algorithm with the intent of accelerating its learning of the game.', 'Building atop the baseline, we then make the following novel contributions in our work:\n(i) We argue that the definition of error-potentials is generalizable across different environments; specifically we show that error-potentials of an observer can be learned for a specific game, and the definition used as-is for another game without requiring re-learning of the error-potentials. \n', '(ii) We propose two different frameworks to combine recent advances in DRL into the error-potential based feedback system in a sample-efficient manner, allowing humans to provide implicit feedback while training in the loop, or prior to the training of the RL agent.\n', '(iii) Finally, we scale the implicit human feedback (via ErrP) based RL to reasonably complex environments (games) and demonstrate the significance of our approach through synthetic and real user experiments.\n']
### SUMMARY:
| [
"We use implicit human feedback (via error-potentials, EEG) to accelerate and optimize the training of a DRL algorithm, in a practical manner."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Deep learning has demonstrated abilities to learn complex structures, but they can be restricted by available data.', 'Recently, Consensus Networks (CNs) were proposed to alleviate data sparsity by utilizing features from multiple modalities, but they too have been limited by the size of labeled data.', 'In this paper, we extend CN to Transductive Consensus Networks (TCNs), suitable for semi-supervised learning.', 'In TCNs, different modalities of input are compressed into latent representations, which we encourage to become indistinguishable during iterative adversarial training.', 'To understand TCNs two mechanisms, consensus and classification, we put forward its three variants in ablation studies on these mechanisms.', 'To further investigate TCN models, we treat the latent representations as probability distributions and measure their similarities as the negative relative Jensen-Shannon divergences.', 'We show that a consensus state beneficial for classification desires a stable but imperfect similarity between the representations.', 'Overall, TCNs outperform or align with the best benchmark algorithms given 20 to 200 labeled samples on the Bank Marketing and the DementiaBank datasets.']
### SUMMARY:
| [
"TCN for multimodal semi-supervised learning + ablation study of its mechanisms + interpretations of latent representations"
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Several first order stochastic optimization methods commonly used in the Euclidean domain such as stochastic gradient descent (SGD), accelerated gradient descent or variance reduced methods have already been adapted to certain Riemannian settings.', 'However, some of the most popular of these optimization tools - namely Adam, Adagrad and the more recent Amsgrad - remain to be generalized to Riemannian manifolds.', 'We discuss the difficulty of generalizing such adaptive schemes to the most agnostic Riemannian setting, and then provide algorithms and convergence proofs for geodesically convex objectives in the particular case of a product of Riemannian manifolds, in which adaptivity is implemented across manifolds in the cartesian product.', 'Our generalization is tight in the sense that choosing the Euclidean space as Riemannian manifold yields the same algorithms and regret bounds as those that were already known for the standard algorithms.', 'Experimentally, we show faster convergence and to a lower train loss value for Riemannian adaptive methods over their corresponding baselines on the realistic task of embedding the WordNet taxonomy in the Poincare ball.']
### SUMMARY:
| [
"Adapting Adam, Amsgrad, Adagrad to Riemannian manifolds. "
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['We study the problem of defending deep neural network approaches for image classification from physically realizable attacks.', 'First, we demonstrate that the two most scalable and effective methods for learning robust models, adversarial training with PGD attacks and randomized smoothing, exhibit very limited effectiveness against three of the highest profile physical attacks.', 'Next, we propose a new abstract adversarial model, rectangular occlusion attacks, in which an adversary places a small adversarially crafted rectangle in an image, and develop two approaches for efficiently computing the resulting adversarial examples.', 'Finally, we demonstrate that adversarial training using our new attack yields image classification models that exhibit high robustness against the physically realizable attacks we study, offering the first effective generic defense against such attacks.']
### SUMMARY:
| [
"Defending Against Physically Realizable Attacks on Image Classification"
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Continual learning is the problem of sequentially learning new tasks or knowledge while protecting previously acquired knowledge.', 'However, catastrophic forgetting poses a grand challenge for neural networks performing such learning process.', 'Thus, neural networks that are deployed in the real world often struggle in scenarios where the data distribution is non-stationary (concept drift), imbalanced, or not always fully available, i.e., rare edge cases.', 'We propose a Differentiable Hebbian Consolidation model which is composed of a Differentiable Hebbian Plasticity (DHP) Softmax layer that adds a rapid learning plastic component (compressed episodic memory) to the fixed (slow changing) parameters of the softmax output layer; enabling learned representations to be retained for a longer timescale.', 'We demonstrate the flexibility of our method by integrating well-known task-specific synaptic consolidation methods to penalize changes in the slow weights that are important for each target task.', 'We evaluate our approach on the Permuted MNIST, Split MNIST and Vision Datasets Mixture benchmarks, and introduce an imbalanced variant of Permuted MNIST --- a dataset that combines the challenges of class imbalance and concept drift.', 'Our proposed model requires no additional hyperparameters and outperforms comparable baselines by reducing forgetting.']
### SUMMARY:
| [
"Hebbian plastic weights can behave as a compressed episodic memory storage in neural networks and with the combination of task-specific synaptic consolidation can improve the ability to alleviate catastrophic forgetting in continual learning."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['In order to choose a neural network architecture that will be effective for a particular modeling problem, one must understand the limitations imposed by each of the potential options.', 'These limitations are typically described in terms of information theoretic bounds, or by comparing the relative complexity needed to approximate example functions between different architectures.', 'In this paper, we examine the topological constraints that the architecture of a neural network imposes on the level sets of all the functions that it is able to approximate.', 'This approach is novel for both the nature of the limitations and the fact that they are independent of network depth for a broad family of activation functions.']
### SUMMARY:
| [
"This paper proves that skinny neural networks cannot approximate certain functions, no matter how deep they are."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Program verification offers a framework for ensuring program correctness and therefore systematically eliminating different classes of bugs.', 'Inferring loop invariants is one of the main challenges behind automated verification of real-world programs which often contain many loops.', 'In this paper, we present Continuous Logic Network (CLN), a novel neural architecture for automatically learning loop invariants directly from program execution traces.', 'Unlike existing neural networks, CLNs can learn precise and explicit representations of formulas in Satisfiability Modulo Theories (SMT) for loop invariants from program execution traces.', 'We develop a new sound and complete semantic mapping for assigning SMT formulas to continuous truth values that allows CLNs to be trained efficiently.', 'We use CLNs to implement a new inference system for loop invariants, CLN2INV, that significantly outperforms existing approaches on the popular Code2Inv dataset.', 'CLN2INV is the first tool to solve all 124 theoretically solvable problems in the Code2Inv dataset.', 'Moreover, CLN2INV takes only 1.1 second on average for each problem, which is 40 times faster than existing approaches.', 'We further demonstrate that CLN2INV can even learn 12 significantly more complex loop invariants than the ones required for the Code2Inv dataset.']
### SUMMARY:
| [
"We introduce the Continuous Logic Network (CLN), a novel neural architecture for automatically learning loop invariants and general SMT formulas."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Single cell RNA sequencing (scRNAseq) technology enables quantifying gene expression profiles by individual cells within cancer.', 'Dimension reduction methods have been commonly used for cell clustering analysis and visualization of the data.', 'Current dimension reduction methods tend overly eliminate the expression variations correspond to less dominating characteristics, such we fail to find the homogenious properties of cancer development.', 'In this paper, we proposed a new and clustering analysis method for scRNAseq data, namely BBSC, via implementing a binarization of the gene expression profile into on/off frequency changes with a Boolean matrix factorization.', 'The low rank representation of expression matrix recovered by BBSC increase the resolution in identifying distinct cell types or functions.', 'Application of BBSC on two cancer scRNAseq data successfully discovered both homogeneous and heterogeneous cancer cell clusters.', 'Further finding showed potential in preventing cancer progression.']
### SUMMARY:
| [
"Our finding shed lights in preventing cancer progression"
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['While normalizing flows have led to significant advances in modeling high-dimensional continuous distributions, their applicability to discrete distributions remains unknown.', 'In this paper, we show that flows can in fact be extended to discrete events---and under a simple change-of-variables formula not requiring log-determinant-Jacobian computations.', 'Discrete flows have numerous applications.', 'We display proofs of concept under 2 flow architectures: discrete autoregressive flows enable bidirectionality, allowing for example tokens in text to depend on both left-to-right and right-to-left contexts in an exact language model; and discrete bipartite flows (i.e., with layer structure from RealNVP) enable parallel generation such as exact nonautoregressive text modeling.']
### SUMMARY:
| [
"We extend autoregressive flows and RealNVP to discrete data."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['We present a Deep Neural Network with Spike Assisted Feature Extraction (SAFE-DNN) to improve robustness of classification under stochastic perturbation of inputs.', 'The proposed network augments a DNN with unsupervised learning of low-level features using spiking neuron network (SNN) with Spike-Time-Dependent-Plasticity (STDP).', 'The complete network learns to ignore local perturbation while performing global feature detection and classification.', 'The experimental results on CIFAR-10 and ImageNet subset demonstrate improved noise robustness for multiple DNN architectures without sacrificing accuracy on clean images.']
### SUMMARY:
| [
"A noise robust deep learning architecture."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Neural embeddings have been used with great success in Natural Language Processing (NLP) where they provide compact representations that encapsulate word similarity and attain state-of-the-art performance in a range of linguistic tasks.', 'The success of neural embeddings has prompted significant amounts of research into applications in domains other than language.', 'One such domain is graph-structured data, where embeddings of vertices can be learned that encapsulate vertex similarity and improve performance on tasks including edge prediction and vertex labelling.', 'For both NLP and graph-based tasks, embeddings in high-dimensional Euclidean spaces have been learned.\n', 'However, recent work has shown that the appropriate isometric space for embedding complex networks is not the flat Euclidean space, but a negatively curved hyperbolic space.', 'We present a new concept that exploits these recent insights and propose learning neural embeddings of graphs in hyperbolic space.', 'We provide experimental evidence that hyperbolic embeddings significantly outperform Euclidean embeddings on vertex classification tasks for several real-world public datasets.']
### SUMMARY:
| [
"We learn neural embeddings of graphs in hyperbolic instead of Euclidean space"
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['The International Competition on Knowledge Engineering for Planning and Scheduling (ICKEPS) plays a pivotal role in fostering the development of new Knowledge Engineering (KE) tools, and in emphasising the importance of principled approaches for all the different KE aspects that are needed for the successful long-term use of planning in real-world applications. \n', 'In this paper, as an exercise in synthesis and for the sake of stimulating thoughts and discussion, we review the format of previous ICKEPS, to suggest alternative formats for future competitions, ideally to motivate someone to step up and organise the next ones.']
### SUMMARY:
| [
"Ideas for future ICKEPS"
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['We show that generating English Wikipedia articles can be approached as a multi-\n', 'document summarization of source documents.', 'We use extractive summarization\n', 'to coarsely identify salient information and a neural abstractive model to generate\n', 'the article.', 'For the abstractive model, we introduce a decoder-only architecture\n', 'that can scalably attend to very long sequences, much longer than typical encoder-\n', 'decoder architectures used in sequence transduction.', 'We show that this model can\n', 'generate fluent, coherent multi-sentence paragraphs and even whole Wikipedia\n', 'articles.', 'When given reference documents, we show it can extract relevant factual\n', 'information as reflected in perplexity, ROUGE scores and human evaluations.']
### SUMMARY:
| [
"We generate Wikipedia articles abstractively conditioned on source document text."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Abstract Stochastic gradient descent (SGD) and Adam are commonly used to optimize deep neural networks, but choosing one usually means making tradeoffs between speed, accuracy and stability.', 'Here we present an intuition for why the tradeoffs exist as well as a method for unifying the two in a continuous way.', 'This makes it possible to control the way models are trained in much greater detail.', 'We show that for default parameters, the new algorithm equals or outperforms SGD and Adam across a range of models for image classification tasks and outperforms SGD for language modeling tasks.']
### SUMMARY:
| [
"An algorithm for unifying SGD and Adam and empirical study of its performance"
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['The use of imitation learning to learn a single policy for a complex task that has multiple modes or hierarchical structure can be challenging.', 'In fact, previous work has shown that when the modes are known, learning separate policies for each mode or sub-task can greatly improve the performance of imitation learning.', 'In this work, we discover the interaction between sub-tasks from their resulting state-action trajectory sequences using a directed graphical model.', 'We propose a new algorithm based on the generative adversarial imitation learning framework which automatically learns sub-task policies from unsegmented demonstrations.', 'Our approach maximizes the directed information flow in the graphical model between sub-task latent variables and their generated trajectories.', 'We also show how our approach connects with the existing Options framework, which is commonly used to learn hierarchical policies.']
### SUMMARY:
| [
"Learning Hierarchical Policies from Unsegmented Demonstrations using Directed Information"
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['The Convolutional Neural Network (CNN) has been successfully applied in many fields during recent decades; however it lacks the ability to utilize prior domain knowledge when dealing with many realistic problems.', 'We present a framework called Geometric Operator Convolutional Neural Network (GO-CNN) that uses domain knowledge, wherein the kernel of the first convolutional layer is replaced with a kernel generated by a geometric operator function.', 'This framework integrates many conventional geometric operators, which allows it to adapt to a diverse range of problems.', 'Under certain conditions, we theoretically analyze the convergence and the bound of the generalization errors between GO-CNNs and common CNNs.', 'Although the geometric operator convolution kernels have fewer trainable parameters than common convolution kernels, the experimental results indicate that GO-CNN performs more accurately than common CNN on CIFAR-10/100.', 'Furthermore, GO-CNN reduces dependence on the amount of training examples and enhances adversarial stability.']
### SUMMARY:
| [
"Traditional image processing algorithms are combined with Convolutional Neural Networks,a new neural network."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Determinantal point processes (DPPs) is an effective tool to deliver diversity on multiple machine learning and computer vision tasks.', 'Under deep learning framework, DPP is typically optimized via approximation, which is not straightforward and has some conflict with diversity requirement.', 'We note, however, there has been no deep learning paradigms to optimize DPP directly since it involves matrix inversion which may result in highly computational instability.', 'This fact greatly hinders the wide use of DPP on some specific objectives where DPP serves as a term to measure the feature diversity.', 'In this paper, we devise a simple but effective algorithm to address this issue to optimize DPP term directly expressed with L-ensemble in spectral domain over gram matrix, which is more flexible than learning on parametric kernels.', 'By further taking into account some geometric constraints, our algorithm seeks to generate valid sub-gradients of DPP term in case when the DPP gram matrix is not invertible (no gradients exist in this case).', 'In this sense, our algorithm can be easily incorporated with multiple deep learning tasks.', 'Experiments show the effectiveness of our algorithm, indicating promising performance for practical learning problems.']
### SUMMARY:
| [
"We proposed a specific back-propagation method via proper spectral sub-gradient to integrate determinantal point process to deep learning framework."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['The quality of a machine translation system depends largely on the availability of sizable parallel corpora.', 'For the recently popular Neural Machine Translation (NMT) framework, data sparsity problem can become even more severe.', 'With large amount of tunable parameters, the NMT model may overfit to the existing language pairs while failing to understand the general diversity in language.', 'In this paper, we advocate to broadcast every sentence pair as two groups of similar sentences to incorporate more diversity in language expressions, which we name as parallel cluster.', 'Then we define a more general cluster-to-cluster correspondence score and train our model to maximize this score.', 'Since direct maximization is difficult, we derive its lower-bound as our surrogate objective, which is found to generalize point-point Maximum Likelihood Estimation (MLE) and point-to-cluster Reward Augmented Maximum Likelihood (RAML) algorithms as special cases.', 'Based on this novel objective function, we delineate four potential systems to realize our cluster-to-cluster framework and test their performances in three recognized translation tasks, each task with forward and reverse translation directions.', 'In each of the six experiments, our proposed four parallel systems have consistently proved to outperform the MLE baseline, RL (Reinforcement Learning) and RAML systems significantly.', 'Finally, we have performed case study to empirically analyze the strength of the cluster-to-cluster NMT framework.']
### SUMMARY:
| [
"We invent a novel cluster-to-cluster framework for NMT training, which can better understand the both source and target language diversity."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['The capability of making interpretable and self-explanatory decisions is essential for developing responsible machine learning systems.', 'In this work, we study the learning to explain the problem in the scope of inductive logic programming (ILP).', 'We propose Neural Logic Inductive Learning (NLIL), an efficient differentiable ILP framework that learns first-order logic rules that can explain the patterns in the data.', 'In experiments, compared with the state-of-the-art models, we find NLIL is able to search for rules that are x10 times longer while remaining x3 times faster.', 'We also show that NLIL can scale to large image datasets, i.e. Visual Genome, with 1M entities.']
### SUMMARY:
| [
"An efficient differentiable ILP model that learns first-order logic rules that can explain the data."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Neural network-based classifiers parallel or exceed human-level accuracy on many common tasks and are used in practical systems.', 'Yet, neural networks are susceptible to adversarial examples, carefully perturbed inputs that cause networks to misbehave in arbitrarily chosen ways.', 'When generated with standard methods, these examples do not consistently fool a classifier in the physical world due to a combination of viewpoint shifts, camera noise, and other natural transformations.', 'Adversarial examples generated using standard techniques require complete control over direct input to the classifier, which is impossible in many real-world systems.\n\n', 'We introduce the first method for constructing real-world 3D objects that consistently fool a neural network across a wide distribution of angles and viewpoints.', 'We present a general-purpose algorithm for generating adversarial examples that are robust across any chosen distribution of transformations.', 'We demonstrate its application in two dimensions, producing adversarial images that are robust to noise, distortion, and affine transformation.', 'Finally, we apply the algorithm to produce arbitrary physical 3D-printed adversarial objects, demonstrating that our approach works end-to-end in the real world.', 'Our results show that adversarial examples are a practical concern for real-world systems.\n']
### SUMMARY:
| [
"We introduce a new method for synthesizing adversarial examples robust in the physical world and use it to fabricate the first 3D adversarial objects."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Representations of sets are challenging to learn because operations on sets should be permutation-invariant.', 'To this end, we propose a Permutation-Optimisation module that learns how to permute a set end-to-end.', 'The permuted set can be further processed to learn a permutation-invariant representation of that set, avoiding a bottleneck in traditional set models.', "We demonstrate our model's ability to learn permutations and set representations with either explicit or implicit supervision on four datasets, on which we achieve state-of-the-art results: number sorting, image mosaics, classification from image mosaics, and visual question answering.\n"]
### SUMMARY:
| [
"Learn how to permute a set, then encode permuted set with RNN to obtain a set representation."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['The physical design of a robot and the policy that controls its motion are inherently coupled.', 'However, existing approaches largely ignore this coupling, instead choosing to alternate between separate design and control phases, which requires expert intuition throughout and risks convergence to suboptimal designs.', 'In this work, we propose a method that jointly optimizes over the physical design of a robot and the corresponding control policy in a model-free fashion, without any need for expert supervision.', 'Given an arbitrary robot morphology, our method maintains a distribution over the design parameters and uses reinforcement learning to train a neural network controller.', 'Throughout training, we refine the robot distribution to maximize the expected reward.', 'This results in an assignment to the robot parameters and neural network policy that are jointly optimal.', 'We evaluate our approach in the context of legged locomotion, and demonstrate that it discovers novel robot designs and walking gaits for several different morphologies, achieving performance comparable to or better than that of hand-crafted designs.']
### SUMMARY:
| [
"Use deep reinforcement learning to design the physical attributes of a robot jointly with a control policy."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Deep learning enables training of large and flexible function approximators from scratch at the cost of large amounts of data.', 'Applications of neural networks often consider learning in the context of a single task.', 'However, in many scenarios what we hope to learn is not just a single task, but a model that can be used to solve multiple different tasks.', 'Such multi-task learning settings have the potential to improve data efficiency and generalization by sharing data and representations across tasks.', 'However, in some challenging multi-task learning settings, particularly in reinforcement learning, it is very difficult to learn a single model that can solve all the tasks while realizing data efficiency and performance benefits.', 'Learning each of the tasks independently from scratch can actually perform better in such settings, but it does not benefit from the representation sharing that multi-task learning can potentially provide.', 'In this work, we develop an approach that endows a single model with the ability to represent both extremes: joint training and independent training.', 'To this end, we introduce matrix-interleaving (Mint), a modification to standard neural network models that projects the activations for each task into a different learned subspace, represented by a per-task and per-layer matrix.', 'By learning these matrices jointly with the other model parameters, the optimizer itself can decide how much to share representations between tasks.', 'On three challenging multi-task supervised learning and reinforcement learning problems with varying degrees of shared task structure, we find that this model consistently matches or outperforms joint training and independent training, combining the best elements of both.']
### SUMMARY:
| [
"We propose an approach that endows a single model with the ability to represent both extremes: joint training and independent training, which leads to effective multi-task learning."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Training agents to operate in one environment often yields overfitted models that are unable to generalize to the changes in that environment.', 'However, due to the numerous variations that can occur in the real-world, the agent is often required to be robust in order to be useful.', 'This has not been the case for agents trained with reinforcement learning (RL) algorithms.', 'In this paper, we investigate the overfitting of RL agents to the training environments in visual navigation tasks.', 'Our experiments show that deep RL agents can overfit even when trained on multiple environments simultaneously. \n', 'We propose a regularization method which combines RL with supervised learning methods by adding a term to the RL objective that would encourage the invariance of a policy to variations in the observations that ought not to affect the action taken.', 'The results of this method, called invariance regularization, show an improvement in the generalization of policies to environments not seen during training.\n']
### SUMMARY:
| [
"We propose a regularization term that, when added to the reinforcement learning objective, allows the policy to maximize the reward and simultaneously learn to be invariant to the irrelevant changes within the input.."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Though visual information has been introduced for enhancing neural machine translation (NMT), its effectiveness strongly relies on the availability of large amounts of bilingual parallel sentence pairs with manual image annotations.', 'In this paper, we present a universal visual representation learned over the monolingual corpora with image annotations, which overcomes the lack of large-scale bilingual sentence-image pairs, thereby extending image applicability in NMT.', 'In detail, a group of images with similar topics to the source sentence will be retrieved from a light topic-image lookup table learned over the existing sentence-image pairs, and then is encoded as image representations by a pre-trained ResNet.', 'An attention layer with a gated weighting is to fuse the visual information and text information as input to the decoder for predicting target translations.', 'In particular, the proposed method enables the visual information to be integrated into large-scale text-only NMT in addition to the multimodel NMT.', "Experiments on four widely used translation datasets, including the WMT'16 English-to-Romanian, WMT'14 English-to-German, WMT'14 English-to-French, and Multi30K, show that the proposed approach achieves significant improvements over strong baselines."]
### SUMMARY:
| [
"This work proposed a universal visual representation for neural machine translation (NMT) using retrieved images with similar topics to source sentence, extending image applicability in NMT."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['This paper introduces a novel framework for learning algorithms to solve online combinatorial optimization problems.', 'Towards this goal, we introduce a number of key ideas from traditional algorithms and complexity theory.', 'First, we draw a new connection between primal-dual methods and reinforcement learning.', 'Next, we introduce the concept of adversarial distributions (universal and high-entropy training sets), which are distributions that encourage the learner to find algorithms that work well in the worst case.', 'We test our new ideas on a number of optimization problem such as the AdWords problem, the online knapsack problem, and the secretary problem.', 'Our results indicate that the models have learned behaviours that are consistent with the traditional optimal algorithms for these problems.']
### SUMMARY:
| [
"By combining ideas from traditional algorithms design and reinforcement learning, we introduce a novel framework for learning algorithms that solve online combinatorial optimization problems."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
["Despite their popularity and successes, deep neural networks are poorly understood theoretically and treated as 'black box' systems.", 'Using a functional view of these networks gives us a useful new lens with which to understand them.', 'This allows us us to theoretically or experimentally probe properties of these networks, including the effect of standard initializations, the value of depth, the underlying loss surface, and the origins of generalization.', 'One key result is that generalization results from smoothness of the functional approximation, combined with a flat initial approximation.', 'This smoothness increases with number of units, explaining why massively overparamaterized networks continue to generalize well.']
### SUMMARY:
| [
"A functional approach reveals that flat initialization, preserved by gradient descent, leads to generalization ability."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['It is well-known that deeper neural networks are harder to train than shallower ones.', 'In this short paper, we use the (full) eigenvalue spectrum of the Hessian to explore how the loss landscape changes as the network gets deeper, and as residual connections are added to the architecture.', 'Computing a series of quantitative measures on the Hessian spectrum, we show that the Hessian eigenvalue distribution in deeper networks has substantially heavier tails (equivalently, more outlier eigenvalues), which makes the network harder to optimize with first-order methods.', 'We show that adding residual connections mitigates this effect substantially, suggesting a mechanism by which residual connections improve training.']
### SUMMARY:
| [
"Network depth increases outlier eigenvalues in the Hessian. Residual connections mitigate this."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['In the context of optimization, a gradient of a neural network indicates the amount a specific weight should change with respect to the loss.', 'Therefore, small gradients indicate a good value of the weight that requires no change and can be kept frozen during training.', 'This paper provides an experimental study on the importance of a neural network weights, and to which extent do they need to be updated.', 'We wish to show that starting from the third epoch, freezing weights which have no informative gradient and are less likely to be changed during training, results in a very slight drop in the overall accuracy (and in sometimes better).', 'We experiment on the MNIST, CIFAR10 and Flickr8k datasets using several architectures (VGG19,\n', 'ResNet-110 and DenseNet-121).', 'On CIFAR10, we show that freezing 80% of the VGG19 network parameters from the third epoch onwards results in 0.24% drop in accuracy, while freezing 50% of Resnet-110 parameters results in 0.9% drop in accuracy and finally freezing 70% of Densnet-121 parameters results in 0.57% drop in accuracy.', 'Furthermore, to experiemnt with real-life applications, we train an image captioning model with attention mechanism on the Flickr8k dataset using LSTM networks, freezing 60% of the parameters from the third epoch onwards, resulting in a better BLEU-4 score than the fully trained model.', 'Our source code can be found in the appendix.']
### SUMMARY:
| [
"An experimental paper that proves the amount of redundant weights that can be freezed from the third epoch only, with only a very slight drop in accuracy."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Like humans, deep networks learn better when samples are organized and introduced in a meaningful order or curriculum.', 'While conventional approaches to curriculum learning emphasize the difficulty of samples as the core incremental strategy, it forces networks to learn from small subsets of data while introducing pre-computation overheads.', 'In this work, we propose Learning with Incremental Labels and Adaptive Compensation (LILAC), which introduces a novel approach to curriculum learning.', 'LILAC emphasizes incrementally learning labels instead of incrementally learning difficult samples.', 'It works in two distinct phases: first, in the incremental label introduction phase, we unmask ground-truth labels in fixed increments during training, to improve the starting point from which networks learn.', 'In the adaptive compensation phase, we compensate for failed predictions by adaptively altering the target vector to a smoother distribution.', 'We evaluate LILAC against the closest comparable methods in batch and curriculum learning and label smoothing, across three standard image benchmarks, CIFAR-10, CIFAR-100, and STL-10.', 'We show that our method outperforms batch learning with higher mean recognition accuracy as well as lower standard deviation in performance consistently across all benchmarks.', 'We further extend LILAC to state-of-the-art performance across CIFAR-10 using simple data augmentation while exhibiting label order invariance among other important properties.']
### SUMMARY:
| [
"A novel approach to curriculum learning by incrementally learning labels and adaptively smoothing labels for mis-classified samples which boost average performance and decreases standard deviation."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Words in natural language follow a Zipfian distribution whereby some words are frequent but most are rare.', "Learning representations for words in the ``long tail'' of this distribution requires enormous amounts of data. \n", 'Representations of rare words trained directly on end tasks are usually poor, requiring us to pre-train embeddings on external data, or treat all rare words as out-of-vocabulary words with a unique representation.', 'We provide a method for predicting embeddings of rare words on the fly from small amounts of auxiliary data with a network trained end-to-end for the downstream task.', 'We show that this improves results against baselines where embeddings are trained on the end task for reading comprehension, recognizing textual entailment and language modeling.\n']
### SUMMARY:
| [
"We propose a method to deal with rare words by computing their embedding from definitions."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['The capability of reliably detecting out-of-distribution samples is one of the key factors in deploying a good classifier, as the test distribution always does not match with the training distribution in most real-world applications.', 'In this work, we propose a deep generative classifier which is effective to detect out-of-distribution samples as well as classify in-distribution samples, by integrating the concept of Gaussian discriminant analysis into deep neural networks.', 'Unlike the discriminative (or softmax) classifier that only focuses on the decision boundary partitioning its latent space into multiple regions, our generative classifier aims to explicitly model class-conditional distributions as separable Gaussian distributions.', 'Thereby, we can define the confidence score by the distance between a test sample and the center of each distribution.', 'Our empirical evaluation on multi-class images and tabular data demonstrate that the generative classifier achieves the best performances in distinguishing out-of-distribution samples, and also it can be generalized well for various types of deep neural networks.']
### SUMMARY:
| [
"This paper proposes a deep generative classifier which is effective to detect out-of-distribution samples as well as classify in-distribution samples, by integrating the concept of Gaussian discriminant analysis into deep neural networks."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['One of the most prevalent symptoms among the elderly population, dementia, can be detected by classifiers trained on linguistic features extracted from narrative transcripts.', 'However, these linguistic features are impacted in a similar but different fashion by the normal aging process.', 'Aging is therefore a confounding factor, whose effects have been hard for machine learning classifiers to isolate. \n\n', 'In this paper, we show that deep neural network (DNN) classifiers can infer ages from linguistic features, which is an entanglement that could lead to unfairness across age groups.', 'We show this problem is caused by undesired activations of v-structures in causality diagrams, and it could be addressed with fair representation learning.', 'We build neural network classifiers that learn low-dimensional representations reflecting the impacts of dementia yet discarding the effects of age.', 'To evaluate these classifiers, we specify a model-agnostic score $\\Delta_{eo}^{(N)}$ measuring how classifier results are disentangled from age.', 'Our best models outperform baseline neural network classifiers in disentanglement, while compromising accuracy by as little as 2.56\\% and 2.25\\% on DementiaBank and the Famous People dataset respectively.']
### SUMMARY:
| [
"Show that age confounds cognitive impairment detection + solve with fair representation learning + propose metrics and models."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Much of the focus in the design of deep neural networks had been on improving accuracy, leading to more powerful yet highly complex network architectures that are difficult to deploy in practical scenarios. ', 'As a result, there has been a recent interest in the design of quantitative metrics for evaluating deep neural networks that accounts for more than just model accuracy as the sole indicator of network performance. ', 'In this study, we continue the conversation towards universal metrics for evaluating the performance of deep neural networks for practical on-device edge usage by introducing NetScore, a new metric designed specifically to provide a quantitative assessment of the balance between accuracy, computational complexity, and network architecture complexity of a deep neural network. ', 'In what is one of the largest comparative analysis between deep neural networks in literature, the NetScore metric, the top-1 accuracy metric, and the popular information density metric were compared across a diverse set of 60 different deep convolutional neural networks for image classification on the ImageNet Large Scale Visual Recognition Challenge (ILSVRC 2012) dataset. ', 'The evaluation results across these three metrics for this diverse set of networks are presented in this study to act as a reference guide for practitioners in the field. ']
### SUMMARY:
| [
"We introduce NetScore, new metric designed to provide a quantitative assessment of the balance between accuracy, computational complexity, and network architecture complexity of a deep neural network."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Deep Neural Networks (DNNs) are vulnerable to adversarial attacks, especially white-box targeted attacks.', 'This paper studies the problem of how aggressive white-box targeted attacks can be to go beyond widely used Top-1 attacks.', 'We propose to learn ordered Top-k attacks (k>=1), which enforce the Top-k predicted labels of an adversarial example to be the k (randomly) selected and ordered labels (the ground-truth label is exclusive).', 'Two methods are presented.', 'First, we extend the vanilla Carlini-Wagner (C&W) method and use it as a strong baseline.', 'Second, we present an adversarial distillation framework consisting of two components:', '(i) Computing an adversarial probability distribution for any given ordered Top-$k$ targeted labels.', '(ii) Learning adversarial examples by minimizing the Kullback-Leibler (KL) divergence between the adversarial distribution and the predicted distribution, together with the perturbation energy penalty.', 'In computing adversarial distributions, we explore how to leverage label semantic similarities, leading to knowledge-oriented attacks.', 'In experiments, we test Top-k (k=1,2,5,10) attacks in the ImageNet-1000 val dataset using two popular DNNs trained with the clean ImageNet-1000 train dataset, ResNet-50 and DenseNet-121.', 'Overall, the adversarial distillation approach obtains the best results, especially by large margin when computation budget is limited..', "It reduces the perturbation energy consistently with the same attack success rate on all the four k's, and improve the attack success rate by large margin against the modified C&W method for k=10. "]
### SUMMARY:
| [
"ordered Top-k adversarial attacks"
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Neural message passing algorithms for semi-supervised classification on graphs have recently achieved great success.', 'However, for classifying a node these methods only consider nodes that are a few propagation steps away and the size of this utilized neighborhood is hard to extend.', 'In this paper, we use the relationship between graph convolutional networks (GCN) and PageRank to derive an improved propagation scheme based on personalized PageRank.', 'We utilize this propagation procedure to construct a simple model, personalized propagation of neural predictions (PPNP), and its fast approximation, APPNP.', "Our model's training time is on par or faster and its number of parameters on par or lower than previous models.", 'It leverages a large, adjustable neighborhood for classification and can be easily combined with any neural network.', 'We show that this model outperforms several recently proposed methods for semi-supervised classification in the most thorough study done so far for GCN-like models.', 'Our implementation is available online.']
### SUMMARY:
| [
"Personalized propagation of neural predictions (PPNP) improves graph neural networks by separating them into prediction and propagation via personalized PageRank."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Most of recent work in cross-lingual word embeddings is severely Anglocentric.', 'The vast majority of lexicon induction evaluation dictionaries are between English and another language, and the English embedding space is selected by default as the hub when learning in a multilingual setting.', 'With this work, however, we challenge these practices.', 'First, we show that the choice of hub language can significantly impact downstream lexicon induction performance.', 'Second, we both expand the current evaluation dictionary collection to include all language pairs using triangulation, and also create new dictionaries for under-represented languages.', 'Evaluating established methods over all these language pairs sheds light into their suitability and presents new challenges for the field.', 'Finally, in our analysis we identify general guidelines for strong cross-lingual embeddings baselines, based on more than just Anglocentric experiments.']
### SUMMARY:
| [
"The choice of the hub (target) language affects the quality of cross-lingual embeddings, which shouldn't be evaluated only on English-centric dictionaries."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Interpreting generative adversarial network (GAN) training as approximate divergence minimization has been theoretically insightful, has spurred discussion, and has lead to theoretically and practically interesting extensions such as f-GANs and Wasserstein GANs.', 'For both classic GANs and f-GANs, there is an original variant of training and a "non-saturating" variant which uses an alternative form of generator update.', 'The original variant is theoretically easier to study, but the alternative variant frequently performs better and is recommended for use in practice.', 'The alternative generator update is often regarded as a simple modification to deal with optimization issues, and it appears to be a common misconception that the two variants minimize the same divergence.', 'In this short note we derive the divergences approximately minimized by the original and alternative variants of GAN and f-GAN training.', 'This highlights important differences between the two variants.', 'For example, we show that the alternative variant of KL-GAN training actually minimizes the reverse KL divergence, and that the alternative variant of conventional GAN training minimizes a "softened" version of the reverse KL.', 'We hope these results may help to clarify some of the theoretical discussion surrounding the divergence minimization view of GAN training.']
### SUMMARY:
| [
"Typical GAN training doesn't optimize Jensen-Shannon, but something like a reverse KL divergence."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['REINFORCE can be used to train models in structured prediction settings to directly optimize the test-time objective.', 'However, the common case of sampling one prediction per datapoint (input) is data-inefficient.', 'We show that by drawing multiple samples (predictions) per datapoint, we can learn with significantly less data, as we freely obtain a REINFORCE baseline to reduce variance.', 'Additionally we derive a REINFORCE estimator with baseline, based on sampling without replacement.', 'Combined with a recent technique to sample sequences without replacement using Stochastic Beam Search, this improves the training procedure for a sequence model that predicts the solution to the Travelling Salesman Problem.']
### SUMMARY:
| [
"We show that by drawing multiple samples (predictions) per input (datapoint), we can learn with less data as we freely obtain a REINFORCE baseline."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['Reinforcement learning (RL) is a powerful technique to train an agent to perform a task. ', 'However, an agent that is trained using RL is only capable of achieving the single task that is specified via its reward function. ', 'Such an approach does not scale well to settings in which an agent needs to perform a diverse set of tasks, such as navigating to varying positions in a room or moving objects to varying locations. ', 'Instead, we propose a method that allows an agent to automatically discover the range of tasks that it is capable of performing in its environment. ', 'We use a generator network to propose tasks for the agent to try to achieve, each task being specified as reaching a certain parametrized subset of the state-space. ', 'The generator network is optimized using adversarial training to produce tasks that are always at the appropriate level of difficulty for the agent. ', 'Our method thus automatically produces a curriculum of tasks for the agent to learn. ', 'We show that, by using this framework, an agent can efficiently and automatically learn to perform a wide set of tasks without requiring any prior knowledge of its environment (Videos and code available at: https://sites.google.com/view/goalgeneration4rl).', 'Our method can also learn to achieve tasks with sparse rewards, which pose significant challenges for traditional RL methods.']
### SUMMARY:
| [
"We efficiently solve multi-task problems with an automatic curriculum generation algorithm based on a generative model that tracks the learning agent's performance."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['A wide range of defenses have been proposed to harden neural networks against adversarial attacks.', 'However, a pattern has emerged in which the majority of adversarial defenses are quickly broken by new attacks. ', 'Given the lack of success at generating robust defenses, we are led to ask a fundamental question: Are adversarial attacks inevitable?\n', 'This paper analyzes adversarial examples from a theoretical perspective, and identifies fundamental bounds on the susceptibility of a classifier to adversarial attacks. ', 'We show that, for certain classes of problems, adversarial examples are inescapable. ', "Using experiments, we explore the implications of theoretical guarantees for real-world problems and discuss how factors such as dimensionality and image complexity limit a classifier's robustness against adversarial examples.\n\n"]
### SUMMARY:
| [
"This paper identifies classes of problems for which adversarial examples are inescapable, and derives fundamental bounds on the susceptibility of any classifier to adversarial examples. "
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['For computer vision applications, prior works have shown the efficacy of reducing numeric precision of model parameters (network weights) in deep neural networks.', 'Activation maps, however, occupy a large memory footprint during both the training and inference step when using mini-batches of inputs.', 'One way to reduce this large memory footprint is to reduce the precision of activations.', 'However, past works have shown that reducing the precision of activations hurts model accuracy.', 'We study schemes to train networks from scratch using reduced-precision activations without hurting accuracy.', 'We reduce the precision of activation maps (along with model parameters) and increase the number of filter maps in a layer, and find that this scheme matches or surpasses the accuracy of the baseline full-precision network.', 'As a result, one can significantly improve the execution efficiency (e.g. reduce dynamic memory footprint, memory band- width and computational energy) and speed up the training and inference process with appropriate hardware support.', 'We call our scheme WRPN -- wide reduced-precision networks.', 'We report results and show that WRPN scheme is better than previously reported accuracies on ILSVRC-12 dataset while being computationally less expensive compared to previously reported reduced-precision networks.']
### SUMMARY:
| [
"Lowering precision (to 4-bits, 2-bits and even binary) and widening the filter banks gives as accurate network as those obtained with FP32 weights and activations."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['We investigate methods for semi-supervised learning (SSL) of a neural linear-chain conditional random field (CRF) for Named Entity Recognition (NER) by treating the tagger as the amortized variational posterior in a generative model of text given tags.', 'We first illustrate how to incorporate a CRF in a VAE, enabling end-to-end training on semi-supervised data.', 'We then investigate a series of increasingly complex deep generative models of tokens given tags enabled by end-to-end optimization, comparing the proposed models against supervised and strong CRF SSL baselines on the Ontonotes5 NER dataset.', 'We find that our best proposed model consistently improves performance by $\\approx 1\\%$ F1 in low- and moderate-resource regimes and easily addresses degenerate model behavior in a more difficult, partially supervised setting.']
### SUMMARY:
| [
"We embed a CRF in a VAE of tokens and NER tags for semi-supervised learning and show improvements in low-resource settings."
] |
### Instruction->PROVIDE ME WITH SUMMARY FOR THE GIVEN INPUT WHILE KEEPING THE MOST IMPORTANT DETAILS INTACT:
['To make deep neural networks feasible in resource-constrained environments (such as mobile devices), it is beneficial to quantize models by using low-precision weights.', 'One common technique for quantizing neural networks is the straight-through gradient method, which enables back-propagation through the quantization mapping.', 'Despite its empirical success, little is understood about why the straight-through gradient method works.\n', 'Building upon a novel observation that the straight-through gradient method is in fact identical to the well-known Nesterov’s dual-averaging algorithm on a quantization constrained optimization problem, we propose a more principled alternative approach, called ProxQuant , that formulates quantized network training as a regularized learning problem instead and optimizes it via the prox-gradient method.', 'ProxQuant does back-propagation on the underlying full-precision vector and applies an efficient prox-operator in between stochastic gradient steps to encourage quantizedness.', 'For quantizing ResNets and LSTMs, ProxQuant outperforms state-of-the-art results on binary quantization and is on par with state-of-the-art on multi-bit quantization.', 'For binary quantization, our analysis shows both theoretically and experimentally that ProxQuant is more stable than the straight-through gradient method (i.e. BinaryConnect), challenging the indispensability of the straight-through gradient method and providing a powerful alternative.']
### SUMMARY:
| [
"A principled framework for model quantization using the proximal gradient method."
] |