ICLR2019 Poster 文章汇总, 共478 papers

【ICLR2019】Poster 论文汇总-编程知识网


Convolutional Neural Networks on Non-uniform Geometrical Signals Using Euclidean Spectral Transformation

Keywords:Non-uniform Fourier Transform, 3D Learning, CNN, surface reconstruction
TL;DR:We use non-Euclidean Fourier Transformation of shapes defined by a simplicial complex for deep learning, achieving significantly better results than point-based sampling techiques used in current 3D learning literature.

Augmented Cyclic Adversarial Learning for Low Resource Domain Adaptation

Keywords:Domain adaptation, generative adversarial network, cyclic adversarial learning, speech
TL;DR:A new cyclic adversarial learning augmented with auxiliary task model which improves domain adaptation performance in low resource supervised and unsupervised situations

Variance Networks: When Expectation Does Not Meet Your Expectations

Keywords:deep learning, variational inference, variational dropout
TL;DR:It is possible to learn a zero-centered Gaussian distribution over the weights of a neural network by learning only variances, and it works surprisingly well.

Initialized Equilibrium Propagation for Backprop-Free Training

Keywords:credit assignment, energy-based models, biologically plausible learning
TL;DR:We train a feedforward network without backprop by using an energy-based model to provide local targets

Explaining Image Classifiers by Counterfactual Generation

Keywords:Explainability, Interpretability, Generative Models, Saliency Map, Machine Learning, Deep Learning
TL;DR:We compute saliency by using a strong generative model to efficiently marginalize over plausible alternative inputs, revealing concentrated pixel areas that preserve label information.

SNIP: SINGLE-SHOT NETWORK PRUNING BASED ON CONNECTION SENSITIVITY

Keywords:neural network pruning, connection sensitivity
TL;DR:We present a new approach, SNIP, that is simple, versatile and interpretable; it prunes irrelevant connections for a given task at single-shot prior to training and is applicable to a variety of neural network models without modifications.

Diagnosing and Enhancing VAE Models

Keywords:variational autoencoder, generative models
TL;DR:We closely analyze the VAE objective function and draw novel conclusions that lead to simple enhancements.

Disjoint Mapping Network for Cross-modal Matching of Voices and Faces

Keywords:None
TL;DR:None

Automatically Composing Representation Transformations as a Means for Generalization

Keywords:compositionality, deep learning, metareasoning
TL;DR:We explore the problem of compositional generalization and propose a means for endowing neural network architectures with the ability to compose themselves to solve these problems.

Visual Reasoning by Progressive Module Networks

Keywords:None
TL;DR:None

Bayesian Deep Convolutional Networks with Many Channels are Gaussian Processes

Keywords:Deep Convolutional Neural Networks, Gaussian Processes, Bayesian
TL;DR:Finite-width SGD trained CNNs vs. infinitely wide fully Bayesian CNNs. Who wins?

Learning to Learn without Forgetting by Maximizing Transfer and Minimizing Interference

Keywords:None
TL;DR:None

Sparse Dictionary Learning by Dynamical Neural Networks

Keywords:None
TL;DR:None

Eidetic 3D LSTM: A Model for Video Prediction and Beyond

Keywords:None
TL;DR:None

ALISTA: Analytic Weights Are As Good As Learned Weights in LISTA

Keywords:None
TL;DR:None

Three Mechanisms of Weight Decay Regularization

Keywords:Generalization, Regularization, Optimization
TL;DR:We investigate weight decay regularization for different optimizers and identify three distinct mechanisms by which weight decay improves generalization.

Learning Multimodal Graph-to-Graph Translation for Molecule Optimization

Keywords:graph-to-graph translation, graph generation, molecular optimization
TL;DR:We introduce a graph-to-graph encoder-decoder framework for learning diverse graph translations.

A Data-Driven and Distributed Approach to Sparse Signal Representation and Recovery

Keywords:Sparsity, Compressive Sensing, Convolutional Network
TL;DR:We use deep learning techniques to solve the sparse signal representation and recovery problem.

On the Minimal Supervision for Training Any Binary Classifier from Only Unlabeled Data

Keywords:learning from only unlabeled data, empirical risk minimization, unbiased risk estimator
TL;DR:Three class priors are all you need to train deep models from only U data, while any two should not be enough.

Neural Logic Machines

Keywords:Neural-Symbolic Computation, Rule Induction, First-Order Logic
TL;DR:We propose the Neural Logic Machine (NLM), a neural-symbolic architecture for both inductive learning and logic reasoning.

Neural Speed Reading with Structural-Jump-LSTM

Keywords:natural language processing, speed reading, recurrent neural network, classification
TL;DR:We propose a new model for neural speed reading that utilizes the inherent punctuation structure of a text to define effective jumping and skipping behavior.

Rigorous Agent Evaluation: An Adversarial Approach to Uncover Catastrophic Failures

Keywords:agent evaluation, adversarial examples, robustness, safety, reinforcement learning
TL;DR:We show that rare but catastrophic failures may be missed entirely by random testing, which poses issues for safe deployment. Our proposed approach for adversarial testing fixes this.

Woulda, Coulda, Shoulda: Counterfactually-Guided Policy Search

Keywords:None
TL;DR:None

signSGD via Zeroth-Order Oracle

Keywords:nonconvex optimization, zeroth-order algorithm, black-box adversarial attack
TL;DR:We design and analyze a new zeroth-order stochastic optimization algorithm, ZO-signSGD, and demonstrate its connection and application to black-box adversarial attacks in robust deep learning

Preventing Posterior Collapse with delta-VAEs

Keywords:Posterior Collapse, VAE, Autoregressive Models
TL;DR: Avoid posterior collapse by lower bounding the rate.

Algorithmic Framework for Model-based Deep Reinforcement Learning with Theoretical Guarantees

Keywords:model-based reinforcement learning, sample efficiency, deep reinforcement learning
TL;DR:We design model-based reinforcement learning algorithms with theoretical guarantees and achieve state-of-the-art results on Mujuco benchmark tasks when one million or fewer samples are permitted.

Knowledge Flow: Improve Upon Your Teachers

Keywords:Transfer Learning, Reinforcement Learning
TL;DR:‘Knowledge Flow’ trains a deep net (student) by injecting information from multiple nets (teachers). The student is independent upon training and performs very well on learned tasks irrespective of the setting (reinforcement or supervised learning).

Directed-Info GAIL: Learning Hierarchical Policies from Unsegmented Demonstrations using Directed Information

Keywords:Imitation Learning, Reinforcement Learning, Deep Learning
TL;DR:Learning Hierarchical Policies from Unsegmented Demonstrations using Directed Information

A Max-Affine Spline Perspective of Recurrent Neural Networks

Keywords:RNN, max-affine spline operators
TL;DR:We provide new insights and interpretations of RNNs from a max-affine spline operators perspective.

Learning to Navigate the Web

Keywords:navigating web pages, reinforcement learning, q learning, curriculum learning, meta training
TL;DR:We train reinforcement learning policies using reward augmentation, curriculum learning, and meta-learning to successfully navigate web pages.

Training for Faster Adversarial Robustness Verification via Inducing ReLU Stability

Keywords:verification, adversarial robustness, adversarial examples, stability, deep learning, regularization
TL;DR:We develop methods to train deep neural models that are both robust to adversarial perturbations and whose robustness is significantly easier to verify.

Learning to Learn with Conditional Class Dependencies

Keywords:meta-learning, learning to learn, few-shot learning
TL;DR:CAML is an instance of MAML with conditional class dependencies.

Hierarchical Visuomotor Control of Humanoids

Keywords:hierarchical reinforcement learning, motor control, motion capture
TL;DR:Solve tasks involving vision-guided humanoid locomotion, reusing locomotion behavior from motion capture data.

Unsupervised Adversarial Image Reconstruction

Keywords:None
TL;DR:None

Max-MIG: an Information Theoretic Approach for Joint Learning from Crowds

Keywords:None
TL;DR:None

AutoLoss: Learning Discrete Schedule for Alternate Optimization

Keywords:Meta Learning, AutoML, Optimization Schedule
TL;DR:We propose a unified formulation for iterative alternate optimization and develop AutoLoss, a framework to automatically learn and generate optimization schedules.

Learning what and where to attend

Keywords:Attention models, human feature importance, object recognition, cognitive science
TL;DR:A large-scale dataset for training attention models for object recognition leads to more accurate, interpretable, and human-like object recognition.

ROBUST ESTIMATION VIA GENERATIVE ADVERSARIAL NETWORKS

Keywords:robust statistics, neural networks, minimax rate, data depth, contamination model, Tukey median, GAN
TL;DR:GANs are shown to provide us a new effective robust mean estimate against agnostic contaminations with both statistical optimality and practical tractability.

INVASE: Instance-wise Variable Selection using Neural Networks

Keywords:None
TL;DR:None

Meta-Learning with Latent Embedding Optimization

Keywords:meta-learning, few-shot, miniImageNet, tieredImageNet, hypernetworks, generative, latent embedding, optimization
TL;DR:Latent Embedding Optimization (LEO) is a novel gradient-based meta-learner with state-of-the-art performance on the challenging 5-way 1-shot and 5-shot miniImageNet and tieredImageNet classification tasks.

Non-vacuous Generalization Bounds at the ImageNet Scale: a PAC-Bayesian Compression Approach

Keywords:generalization, deep-learning, pac-bayes
TL;DR:We obtain non-vacuous generalization bounds on ImageNet-scale deep neural networks by combining an original PAC-Bayes bound and an off-the-shelf neural network compression method.

Learning to Represent Edits

Keywords:None
TL;DR:None

Neural Probabilistic Motor Primitives for Humanoid Control

Keywords:Motor Primitives, Distillation, Reinforcement Learning, Continuous Control, Humanoid Control, Motion Capture, One-Shot Imitation
TL;DR:Neural Probabilistic Motor Primitives compress motion capture tracking policies into one flexible model capable of one-shot imitation and reuse as a low-level controller.

Differentiable Perturb-and-Parse: Semi-Supervised Parsing with a Structured Variational Autoencoder

Keywords:differentiable dynamic programming, variational auto-encoder, dependency parsing, semi-supervised learning
TL;DR:Differentiable dynamic programming over perturbed input weights with application to semi-supervised VAE

Janossy Pooling: Learning Deep Permutation-Invariant Functions for Variable-Size Inputs

Keywords:representation learning, permutation invariance, set functions, feature pooling
TL;DR:We propose Janossy pooling, a method for learning deep permutation invariant functions designed to exploit relationships within the input sequence and tractable inference strategies such as a stochastic optimization procedure we call piSGD

An Empirical Study of Example Forgetting during Deep Neural Network Learning

Keywords:catastrophic forgetting, sample weighting, deep generalization
TL;DR:We show that catastrophic forgetting occurs within what is considered to be a single task and find that examples that are not prone to forgetting can be removed from the training set without loss of generalization.

RNNs implicitly implement tensor-product representations

Keywords:tensor-product representations, compositionality, neural network interpretability, recurrent neural networks
TL;DR:RNNs implicitly implement tensor-product representations, a principled and interpretable method for representing symbolic structures in continuous space.

Learning To Solve Circuit-SAT: An Unsupervised Differentiable Approach

Keywords:Neuro-Symbolic Methods, Circuit Satisfiability, Neural SAT Solver, Graph Neural Networks
TL;DR:We propose a neural framework that can learn to solve the Circuit Satisfiability problem from (unlabeled) circuit instances.

Dynamic Channel Pruning: Feature Boosting and Suppression

Keywords:dynamic network, faster CNNs, channel pruning
TL;DR:We make convolutional layers run faster by dynamically boosting and suppressing channels in feature computation.

signSGD with Majority Vote is Communication Efficient and Fault Tolerant

Keywords:large-scale learning, distributed systems, communication efficiency, convergence rate analysis, robust optimisation
TL;DR:Workers send gradient signs to the server, and the update is decided by majority vote. We show that this algorithm is convergent, communication efficient and fault tolerant, both in theory and in practice.

Bounce and Learn: Modeling Scene Dynamics with Real-World Bounces

Keywords:None
TL;DR:None

K for the Price of 1: Parameter-efficient Multi-task and Transfer Learning

Keywords:deep learning, mobile, transfer learning, multi-task learning, computer vision, small models, imagenet, inception, batch normalization
TL;DR:A novel and practically effective method to adapt pretrained neural networks to new tasks by retraining a minimal (e.g., less than 2%) number of parameters

Towards Metamerism via Foveated Style Transfer

Keywords:Metamerism, foveation, perception, style transfer, psychophysics
TL;DR:We introduce a novel feed-forward framework to generate visual metamers

Post Selection Inference with Incomplete Maximum Mean Discrepancy Estimator

Keywords:None
TL;DR:None

Emergent Coordination Through Competition

Keywords:Multi-agent learning, Reinforcement Learning
TL;DR:We introduce a new MuJoCo soccer environment for continuous multi-agent reinforcement learning research, and show that population-based training of independent reinforcement learners can learn cooperative behaviors

Prior Convictions: Black-box Adversarial Attacks with Bandits and Priors

Keywords:adversarial examples, gradient estimation, black-box attacks, model-based optimization, bandit optimization
TL;DR:We present a unifying view on black-box adversarial attacks as a gradient estimation problem, and then present a framework (based on bandits optimization) to integrate priors into gradient estimation, leading to significantly increased performance.

Sample Efficient Imitation Learning for Continuous Control

Keywords:Imitation Learning, Continuous Control, Reinforcement Learning, Inverse Reinforcement Learning, Conditional Generative Adversarial Network
TL;DR:In this paper, we proposed a model-free, off-policy IL algorithm for continuous control. Experimental results showed that our algorithm achieves competitive results with GAIL while significantly reducing the environment interactions.

Generative Code Modeling with Graphs

Keywords:Generative Model, Source Code, Graph Learning
TL;DR:Representing programs as graphs including semantics helps when generating programs

Critical Learning Periods in Deep Networks

Keywords:Critical Period, Deep Learning, Information Theory, Artificial Neuroscience, Information Plasticity
TL;DR:Sensory deficits in early training phases can lead to irreversible performance loss in both artificial and neuronal networks, suggesting information phenomena as the common cause, and point to the importance of the initial transient and forgetting.

CEM-RL: Combining evolutionary and gradient-based methods for policy search

Keywords:evolution strategy, deep reinforcement learning
TL;DR:We propose a new combination of evolution strategy and deep reinforcement learning which takes the best of both worlds

LanczosNet: Multi-Scale Deep Graph Convolutional Networks

Keywords:None
TL;DR:None

Excessive Invariance Causes Adversarial Vulnerability

Keywords:Generalization, Adversarial Examples, Invariance, Information Theory, Invertible Networks
TL;DR:We show deep networks are not only too sensitive to task-irrelevant changes of their input, but also too invariant to a wide range of task-relevant changes, thus making vast regions in input space vulnerable to adversarial attacks.

Hindsight policy gradients

Keywords:reinforcement learning, policy gradients, multi-goal reinforcement learning
TL;DR:We introduce the capacity to exploit information about the degree to which an arbitrary goal has been achieved while another goal was intended to policy gradient methods.

Adaptive Gradient Methods with Dynamic Bound of Learning Rate

Keywords:Optimization, SGD, Adam, Generalization
TL;DR:Novel variants of optimization methods that combine the benefits of both adaptive and non-adaptive methods.

Decoupled Weight Decay Regularization

Keywords:None
TL;DR:None

Optimistic mirror descent in saddle-point problems: Going the extra (gradient) mile

Keywords:Mirror descent, extra-gradient, generative adversarial networks, saddle-point problems
TL;DR:We show how the inclusion of an extra-gradient step in first-order GAN training methods can improve stability and lead to improved convergence results.

DialogWAE: Multimodal Response Generation with Conditional Wasserstein Auto-Encoder

Keywords:None
TL;DR:None

No Training Required: Exploring Random Encoders for Sentence Classification

Keywords:None
TL;DR:None

Neural Graph Evolution: Automatic Robot Design

Keywords:Reinforcement learning, graph neural networks, robotics, deep learning, transfer learning
TL;DR:Automatic robotic design search with graph neural networks

Function Space Particle Optimization for Bayesian Neural Networks

Keywords:None
TL;DR:None

Structured Adversarial Attack: Towards General Implementation and Better Interpretability

Keywords:None
TL;DR:None

Spherical CNNs on Unstructured Grids

Keywords:Spherical CNN, unstructured grid, panoramic, semantic segmentation, parameter efficiency
TL;DR:We present a new CNN kernel for unstructured grids for spherical signals, and show significant accuracy and parameter efficiency gain on tasks such as 3D classfication and omnidirectional image segmentation.

Optimal Transport Maps For Distribution Preserving Operations on Latent Spaces of Generative Models

Keywords:generative models, optimal transport, distribution preserving operations
TL;DR:We propose a framework for modifying the latent space operations such that the distribution mismatch between the resulting outputs and the prior distribution the generative model was trained on is fully eliminated.

Deep Lagrangian Networks: Using Physics as Model Prior for Deep Learning

Keywords:Deep Model Learning, Robot Control
TL;DR:This paper introduces a physics prior for Deep Learning and applies the resulting network topology for model-based control.

Accumulation Bit-Width Scaling For Ultra-Low Precision Training Of Deep Networks

Keywords:reduced precision floating-point, partial sum accumulation bit-width, deep learning, training
TL;DR:We present an analytical framework to determine accumulation bit-width requirements in all three deep learning training GEMMs and verify the validity and tightness of our method via benchmarking experiments.

Deep Convolutional Networks as shallow Gaussian Processes

Keywords:Gaussian process, CNN, ResNet, Bayesian
TL;DR:We show that CNNs and ResNets with appropriate priors on the parameters are Gaussian processes in the limit of infinitely many convolutional filters.

Unsupervised Domain Adaptation for Distance Metric Learning

Keywords:domain adaptation, distance metric learning, face recognition
TL;DR:A new theory of unsupervised domain adaptation for distance metric learning and its application to face recognition across diverse ethnicity variations.

A comprehensive, application-oriented study of catastrophic forgetting in DNNs

Keywords:incremental learning, deep neural networks, catatrophic forgetting, sequential learning
TL;DR:We check DNN models for catastrophic forgetting using a new evaluation scheme that reflects typical application conditions, with surprising results.

Posterior Attention Models for Sequence to Sequence Learning

Keywords:posterior inference, attention, seq2seq learning, translation
TL;DR:Computing attention based on posterior distribution leads to more meaningful attention and better performance

Generative Question Answering: Learning to Answer the Whole Question

Keywords:Question answering, question generation, reasoning, squad, clevr
TL;DR:Question answering models that model the joint distribution of questions and answers can learn more than discriminative models

Diversity and Depth in Per-Example Routing Models

Keywords:conditional computation, routing models, depth
TL;DR:Per-example routing models benefit from architectural diversity, but still struggle to scale to a large number of routing decisions.

Selfless Sequential Learning

Keywords:Lifelong learning, Continual Learning, Sequential learning, Regularization
TL;DR:A regularization strategy for improving the performance of sequential learning

M^3RL: Mind-aware Multi-agent Management Reinforcement Learning

Keywords:Multi-agent Reinforcement Learning, Deep Reinforcement Learning
TL;DR:We propose Mind-aware Multi-agent Management Reinforcement Learning (M^3RL) for training a manager to motivate self-interested workers to achieve optimal collaboration by assigning suitable contracts to them.

The Deep Weight Prior

Keywords:deep learning, variational inference, prior distributions
TL;DR:The deep weight prior learns a generative model for kernels of convolutional neural networks, that acts as a prior distribution while training on new datasets.

Efficient Multi-Objective Neural Architecture Search via Lamarckian Evolution

Keywords:Neural Architecture Search, AutoML, AutoDL, Deep Learning, Evolutionary Algorithms, Multi-Objective Optimization
TL;DR:We propose a method for efficient Multi-Objective Neural Architecture Search based on Lamarckian inheritance and evolutionary algorithms.

Quaternion Recurrent Neural Networks

Keywords:None
TL;DR:None

Adversarial Audio Synthesis

Keywords:audio, waveform, spectrogram, GAN, adversarial, WaveGAN, SpecGAN
TL;DR:Learning to synthesize raw waveform audio with GANs

Preconditioner on Matrix Lie Group for SGD

Keywords:preconditioner, stochastic gradient descent, Newton method, Fisher information, natural gradient, Lie group
TL;DR:We propose a new framework for preconditioner learning, derive new forms of preconditioners and learning methods, and reveal the relationship to methods like RMSProp, Adam, Adagrad, ESGD, KFAC, batch normalization, etc.

Learning to Screen for Fast Softmax Inference on Large Vocabulary Neural Networks

Keywords:None
TL;DR:None

Adaptive Posterior Learning: few-shot learning with a surprise-based memory module

Keywords:metalearning, memory, few-shot, relational, self-attention, classification, sequential, reasoning, working memory, episodic memory
TL;DR:We introduce a model which generalizes quickly from few observations by storing surprising information and attending over the most relevant data at each time point.

Probabilistic Planning with Sequential Monte Carlo methods

Keywords:control as inference, probabilistic planning, sequential monte carlo, model based reinforcement learning
TL;DR:Leveraging control as inference and Sequential Monte Carlo methods, we proposed a probabilistic planning algorithm.

Plan Online, Learn Offline: Efficient Learning and Exploration via Model-Based Control

Keywords:deep reinforcement learning, exploration, model-based
TL;DR:We propose a framework that incorporates planning for efficient exploration and learning in complex environments.

DHER: Hindsight Experience Replay for Dynamic Goals

Keywords:None
TL;DR:None

FlowQA: Grasping Flow in History for Conversational Machine Comprehension

Keywords:Machine Comprehension, Conversational Agent, Natural Language Processing, Deep Learning
TL;DR:We propose the Flow mechanism and an end-to-end architecture, FlowQA, that achieves SotA on two conversational QA datasets and a sequential instruction understanding task.

Learning to Design RNA

Keywords:matter engineering, bioinformatics, rna design, reinforcement learning, meta learning, neural architecture search, hyperparameter optimization
TL;DR:We learn to solve the RNA Design problem with reinforcement learning using meta learning and autoML approaches.

Robust Conditional Generative Adversarial Networks

Keywords:conditional GAN, unsupervised pathway, autoencoder, robustness
TL;DR:We introduce a new type of conditional GAN, which aims to leverage structure in the target space of the generator. We augment the generator with a new, unsupervised pathway to learn the target structure.

Top-Down Neural Model For Formulae

Keywords:logic, formula, recursive neural networks, recurrent neural networks
TL;DR:A top-down approach how to recursively represent propositional formulae by neural networks is presented.

Cost-Sensitive Robustness against Adversarial Examples

Keywords:Certified robustness, Adversarial examples, Cost-sensitive learning
TL;DR:A general method for training certified cost-sensitive robust classifier against adversarial perturbations

The role of over-parametrization in generalization of neural networks

Keywords:Generalization, Over-Parametrization, Neural Networks, Deep Learning
TL;DR:We suggest a generalization bound that could partly explain the improvement in generalization with over-parametrization.

Diffusion Scattering Transforms on Graphs

Keywords:graph neural networks, deep learning, stability, scattering transforms, convolutional neural networks
TL;DR:Stability of scattering transform representations of graph data to deformations of the underlying graph support.

Capsule Graph Neural Network

Keywords:CapsNet, Graph embedding, GNN
TL;DR:Inspired by CapsNet, we propose a novel architecture for graph embeddings on the basis of node features extracted from GNN.

Energy-Constrained Compression for Deep Neural Networks via Weighted Sparse Projection and Layer Input Masking

Keywords:None
TL;DR:None

Emerging Disentanglement in Auto-Encoder Based Unsupervised Image Content Transfer

Keywords:Image-to-image Translation, Disentanglement, Autoencoders, Faces
TL;DR:An image to image translation method which adds to one image the content of another thereby creating a new image.

SGD Converges to Global Minimum in Deep Learning via Star-convex Path

Keywords:None
TL;DR:None

Toward Understanding the Impact of Staleness in Distributed Machine Learning

Keywords:None
TL;DR:None

Transfer Learning for Sequences via Learning to Collocate

Keywords:transfer learning, recurrent neural network, attention, natural language processing
TL;DR:Transfer learning for sequence via learning to align cell-level information across domains.

Learning Procedural Abstractions and Evaluating Discrete Latent Temporal Structure

Keywords:None
TL;DR:None

Unsupervised Speech Recognition via Segmental Empirical Output Distribution Matching

Keywords:None
TL;DR:None

Adversarial Attacks on Graph Neural Networks via Meta Learning

Keywords:graph mining, adversarial attacks, meta learning, graph neural networks, node classification
TL;DR:We use meta-gradients to attack the training procedure of deep neural networks for graphs.

Maximal Divergence Sequential Autoencoder for Binary Software Vulnerability Detection

Keywords:Vulnerabilities Detection, Sequential Auto-Encoder, Separable Representation
TL;DR:We propose a novel method named Maximal Divergence Sequential Auto-Encoder that leverages Variational AutoEncoder representation for binary code vulnerability detection.

Neural Program Repair by Jointly Learning to Localize and Repair

Keywords:neural program repair, neural program embeddings, pointer networks
TL;DR:Multi-headed Pointer Networks for jointly learning to localize and repair Variable Misuse bugs

Information-Directed Exploration for Deep Reinforcement Learning

Keywords:reinforcement learning, exploration, information directed sampling
TL;DR:We develop a practical extension of Information-Directed Sampling for Reinforcement Learning, which accounts for parametric uncertainty and heteroscedasticity in the return distribution for exploration.

Attention, Learn to Solve Routing Problems!

Keywords:learning, routing problems, heuristics, attention, reinforce, travelling salesman problem, vehicle routing problem, orienteering problem, prize collecting travelling salesman problem
TL;DR:Attention based model trained with REINFORCE with greedy rollout baseline to learn heuristics with competitive results on TSP and other routing problems

L2-Nonexpansive Neural Networks

Keywords:None
TL;DR:None

Improving Generalization and Stability of Generative Adversarial Networks

Keywords:GAN, generalization, gradient penalty, zero centered, convergence
TL;DR:We propose a zero-centered gradient penalty for improving generalization and stability of GANs

Adaptive Input Representations for Neural Language Modeling

Keywords:Neural language modeling
TL;DR:Variable capacity input word embeddings and SOTA on WikiText-103, Billion Word benchmarks.

Neural Persistence: A Complexity Measure for Deep Neural Networks Using Algebraic Topology

Keywords:Algebraic topology, persistent homology, network complexity, neural network
TL;DR:We develop a new topological complexity measure for deep neural networks and demonstrate that it captures their salient properties.

Efficient Augmentation via Data Subsampling

Keywords:data augmentation, invariance, subsampling, influence
TL;DR:Selectively augmenting difficult to classify points results in efficient training.

Neural TTS Stylization with Adversarial and Collaborative Games

Keywords:Text-To-Speech synthesis, GANs
TL;DR:a generative adversarial network for style modeling in a text-to-speech system

Optimal Control Via Neural Networks: A Convex Approach

Keywords:None
TL;DR:None

CBOW Is Not All You Need: Combining CBOW with the Compositional Matrix Space Model

Keywords:Text representation learning, Sentence embedding, Efficient training scheme, word2vec
TL;DR:We present a novel training scheme for efficiently obtaining order-aware sentence representations.

Stochastic Optimization of Sorting Networks via Continuous Relaxations

Keywords:continuous relaxations, sorting, permutation, stochastic computation graphs, Plackett-Luce
TL;DR:We provide a continuous relaxation to the sorting operator, enabling end-to-end, gradient-based stochastic optimization.

Adaptivity of deep ReLU network for learning in Besov and mixed smooth Besov spaces: optimal rate and curse of dimensionality

Keywords:None
TL;DR:None

Generating Multiple Objects at Spatially Distinct Locations

Keywords:controllable image generation, text-to-image synthesis, generative model, generative adversarial network, gan
TL;DR:Extend GAN architecture to obtain control over locations and identities of multiple objects within generated images.

Near-Optimal Representation Learning for Hierarchical Reinforcement Learning

Keywords:representation hierarchy reinforcement learning
TL;DR:We translate a bound on sub-optimality of representations to a practical training objective in the context of hierarchical reinforcement learning.

Understanding Composition of Word Embeddings via Tensor Decomposition

Keywords:word embeddings, semantic composition, tensor decomposition
TL;DR:We present a generative model for compositional word embeddings that captures syntactic relations, and provide empirical verification and evaluation.

Structured Neural Summarization

Keywords:Summarization, Graphs, Source Code
TL;DR:One simple trick to improve sequence models: Compose them with a graph model

Graph Wavelet Neural Network

Keywords:graph convolution, graph wavelet transform, graph Fourier transform, semi-supervised learning
TL;DR:We present graph wavelet neural network (GWNN), a novel graph convolutional neural network (CNN), leveraging graph wavelet transform to address the shortcoming of previous spectral graph CNN methods that depend on graph Fourier transform.

A rotation-equivariant convolutional neural network model of primary visual cortex

Keywords:rotation equivariance, equivariance, primary visual cortex, V1, neuroscience, system identification
TL;DR:A rotation-equivariant CNN model of V1 that outperforms previous models and suggest functional groupings of V1 neurons.

Supervised Community Detection with Line Graph Neural Networks

Keywords:community detection, graph neural networks, belief propagation, energy landscape, non-backtracking matrix
TL;DR:We propose a novel graph neural network architecture based on the non-backtracking matrix defined over the edge adjacencies and demonstrate its effectiveness in community detection tasks on graphs.

Multiple-Attribute Text Rewriting

Keywords:controllable text generation, generative models, conditional generative models, style transfer
TL;DR:A system for rewriting text conditioned on multiple controllable attributes

Wasserstein Barycenter Model Ensembling

Keywords:Wasserstein barycenter model ensembling
TL;DR:we propose to use Wasserstein barycenters for semantic model ensembling

Policy Transfer with Strategy Optimization

Keywords:transfer learning, reinforcement learning, modeling error, strategy optimization
TL;DR:We propose a policy transfer algorithm that can overcome large and challenging discrepancies in the system dynamics such as latency, actuator modeling error, etc.

code2seq: Generating Sequences from Structured Representations of Code

Keywords:source code, programs, code2seq
TL;DR:We leverage the syntactic structure of source code to generate natural language sequences.

Predict then Propagate: Graph Neural Networks meet Personalized PageRank

Keywords:Graph, GCN, GNN, Neural network, Graph neural network, Message passing neural network, Semi-supervised classification, Semi-supervised learning, PageRank, Personalized PageRank
TL;DR:Personalized propagation of neural predictions (PPNP) improves graph neural networks by separating them into prediction and propagation via personalized PageRank.

Slimmable Neural Networks

Keywords:Slimmable neural networks, mobile deep learning, accuracy-efficiency trade-offs
TL;DR:We present a simple and general method to train a single neural network executable at different widths (number of channels in a layer), permitting instant and adaptive accuracy-efficiency trade-offs at runtime.

Analysing Mathematical Reasoning Abilities of Neural Models

Keywords:mathematics, dataset, algebraic, reasoning
TL;DR:A dataset for testing mathematical reasoning (and algebraic generalization), and results on current sequence-to-sequence models.

RotDCF: Decomposition of Convolutional Filters for Rotation-Equivariant Deep Networks

Keywords:None
TL;DR:None

Execution-Guided Neural Program Synthesis

Keywords:None
TL;DR:None

Dynamic Sparse Graph for Efficient Deep Learning

Keywords:Sparsity, compression, training, acceleration
TL;DR:We construct dynamic sparse graph via dimension-reduction search to reduce compute and memory cost in both DNN training and inference.

Fixup Initialization: Residual Learning Without Normalization

Keywords:deep learning, residual networks, initialization, batch normalization, layer normalization
TL;DR:All you need to train deep residual networks is a good initialization; normalization layers are not necessary.

ProbGAN: Towards Probabilistic GAN with Theoretical Guarantees

Keywords:Generative Adversarial Networks, Bayesian Deep Learning, Mode Collapse, Inception Score, Generator, Discriminator, CIFAR-10, STL-10, ImageNet
TL;DR:A novel probabilistic treatment for GAN with theoretical guarantee.

Exploration by random network distillation

Keywords:reinforcement learning, exploration, curiosity
TL;DR:A simple exploration bonus is introduced and achieves state of the art performance in 3 hard exploration Atari games.

Unsupervised Learning of the Set of Local Maxima

Keywords:None
TL;DR:None

On the Convergence of A Class of Adam-Type Algorithms for Non-Convex Optimization

Keywords:nonconvex optimization, Adam, convergence analysis
TL;DR:We analyze convergence of Adam-type algorithms and provide mild sufficient conditions to guarantee their convergence, we also show violating the conditions can makes an algorithm diverge.

Minimum Divergence vs. Maximum Margin: an Empirical Comparison on Seq2Seq Models

Keywords:None
TL;DR:None

GANSynth: Adversarial Neural Audio Synthesis

Keywords:GAN, Audio, WaveNet, NSynth, Music
TL;DR:High-quality audio synthesis with GANs

Sliced Wasserstein Auto-Encoders

Keywords:optimal transport, Wasserstein distances, auto-encoders, unsupervised learning
TL;DR:In this paper we use the sliced-Wasserstein distance to shape the latent distribution of an auto-encoder into any samplable prior distribution.

Learning Two-layer Neural Networks with Symmetric Inputs

Keywords:Neural Network, Optimization, Symmetric Inputs, Moment-of-moments
TL;DR:We give an algorithm for learning a two-layer neural network with symmetric input distribution.

Learning to Understand Goal Specifications by Modelling Reward

Keywords:instruction following, reward modelling, language understanding
TL;DR:We propose AGILE, a framework for training agents to perform instructions from examples of respective goal-states.

Do Deep Generative Models Know What They Don’t Know?

Keywords:None
TL;DR:None

Identifying and Controlling Important Neurons in Neural Machine Translation

Keywords:neural machine translation, individual neurons, unsupervised, analysis, correlation, translation control, distributivity, localization
TL;DR:Unsupervised methods for finding, analyzing, and controlling important neurons in NMT

Representing Formal Languages: A Comparison Between Finite Automata and Recurrent Neural Networks

Keywords:Language recognition, Recurrent Neural Networks, Representation Learning, deterministic finite automaton, automaton
TL;DR:Finite Automata Can be Linearly decoded from Language-Recognizing RNNs using low coarseness abstraction functions and high accuracy decoders.

Visual Explanation by Interpretation: Improving Visual Feedback Capabilities of Deep Neural Networks

Keywords:model explanation, model interpretation, explainable ai, evaluation
TL;DR:Interpretation by Identifying model-learned features that serve as indicators for the task of interest. Explain model decisions by highlighting the response of these features in test data. Evaluate explanations objectively with a controlled dataset.

Don’t let your Discriminator be fooled

Keywords:GAN, generative models, computer vision
TL;DR:A discriminator that is not easily fooled by adversarial example makes GAN training more robust and leads to a smoother objective.

Latent Convolutional Models

Keywords:latent models, convolutional networks, unsupervised learning, deep learning, modeling natural images, image restoration
TL;DR:We present a new deep latent model of natural images that can be trained from unlabeled datasets and can be utilized to solve various image restoration tasks.

A Universal Music Translation Network

Keywords:None
TL;DR:None

How to train your MAML

Keywords:meta-learning, deep-learning, few-shot learning, supervised learning, neural-networks, stochastic optimization
TL;DR:MAML is great, but it has many problems, we solve many of those problems and as a result we learn most hyper parameters end to end, speed-up training and inference and set a new SOTA in few-shot learning

Learning a SAT Solver from Single-Bit Supervision

Keywords:sat, search, graph neural network, theorem proving, proof
TL;DR:We train a graph network to predict boolean satisfiability and show that it learns to search for solutions, and that the solutions it finds can be decoded from its activations.

Learning Representations of Sets through Optimized Permutations

Keywords:sets, representation learning, permutation invariance
TL;DR:Learn how to permute a set, then encode permuted set with RNN to obtain a set representation.

Big-Little Net: An Efficient Multi-Scale Feature Representation for Visual and Speech Recognition

Keywords:None
TL;DR:None

Unsupervised Hyper-alignment for Multilingual Word Embeddings

Keywords:None
TL;DR:None

Visual Semantic Navigation using Scene Priors

Keywords:None
TL;DR:None

NOODL: Provable Online Dictionary Learning and Sparse Coding

Keywords:provable dictionary learning, sparse coding, support recovery, iterative hard thresholding, matrix factorization, neural architectures, noodl
TL;DR:We present a provable algorithm for exactly recovering both factors of the dictionary learning model.

Stochastic Gradient/Mirror Descent: Minimax Optimality and Implicit Regularization

Keywords:None
TL;DR:None

Active Learning with Partial Feedback

Keywords:None
TL;DR:None

Gradient descent aligns the layers of deep linear networks

Keywords:None
TL;DR:None

Data-Dependent Coresets for Compressing Neural Networks with Applications to Generalization Bounds

Keywords:None
TL;DR:None

On the loss landscape of a class of deep neural networks with no bad local valleys

Keywords:None
TL;DR:None

DOM-Q-NET: Grounded RL on Structured Language

Keywords:Reinforcement Learning, Web Navigation, Graph Neural Networks
TL;DR:Graph-based Deep Q Network for Web Navigation

Boosting Robustness Certification of Neural Networks

Keywords:Robustness certification, Adversarial Attacks, Abstract Interpretation, MILP Solvers, Verification of Neural Networks
TL;DR:We refine the over-approximation results from incomplete verifiers using MILP solvers to prove more robustness properties than state-of-the-art.

Learning To Simulate

Keywords:Simulation in machine learning, reinforcement learning, policy gradients, image rendering
TL;DR:We propose an algorithm that automatically adjusts parameters of a simulation engine to generate training data for a neural network such that validation accuracy is maximized.

Towards Understanding Regularization in Batch Normalization

Keywords:None
TL;DR:None

The Laplacian in RL: Learning Representations with Efficient Approximations

Keywords:Laplacian, reinforcement learning, representation
TL;DR:We propose a scalable method to approximate the eigenvectors of the Laplacian in the reinforcement learning context and we show that the learned representations can improve the performance of an RL agent.

Predicting the Generalization Gap in Deep Networks with Margin Distributions

Keywords:Deep learning, large margin, generalization bounds, generalization gap.
TL;DR:We develop a new scheme to predict the generalization gap in deep networks with high accuracy.

Adversarial Imitation via Variational Inverse Reinforcement Learning

Keywords:Our method introduces the empowerment-regularized maximum-entropy inverse reinforcement learning to learn near-optimal rewards and policies from expert demonstrations.
TL;DR:Inverse Reinforcement Learning, Imitation learning, Variational lnference, Learning from demonstrations

Reasoning About Physical Interactions with Object-Oriented Prediction and Planning

Keywords:structured scene representation, predictive models, intuitive physics, self-supervised learning
TL;DR:We present a framework for learning object-centric representations suitable for planning in tasks that require an understanding of physics.

LayoutGAN: Generating Graphic Layouts with Wireframe Discriminators

Keywords:None
TL;DR:None

Learning Mixed-Curvature Representations in Product Spaces

Keywords:embeddings, non-Euclidean geometry, manifolds, geometry of data
TL;DR:Product manifold embedding spaces with heterogenous curvature yield improved representations compared to traditional embedding spaces for a variety of structures.

StrokeNet: A Neural Painting Environment

Keywords:image generation, differentiable model, reinforcement learning, deep learning, model based
TL;DR:StrokeNet is a novel architecture where the agent is trained to draw by strokes on a differentiable simulation of the environment, which could effectively exploit the power of back-propagation.

Harmonizing Maximum Likelihood with GANs for Multimodal Conditional Generation

Keywords:conditional GANs, conditional image generation, multimodal generation, reconstruction loss, maximum likelihood estimation, moment matching
TL;DR:We prove that the mode collapse in conditional GANs is largely attributed to a mismatch between reconstruction loss and GAN loss and introduce a set of novel loss functions as alternatives for reconstruction loss.

Measuring Compositionality in Representation Learning

Keywords:compositionality, representation learning, evaluation
TL;DR:This paper proposes a simple procedure for evaluating compositional structure in learned representations, and uses the procedure to explore the role of compositionality in four learning problems.

Benchmarking Neural Network Robustness to Common Corruptions and Perturbations

Keywords:robustness, benchmark, convnets, perturbations
TL;DR:We propose ImageNet-C to measure classifier corruption robustness and ImageNet-P to measure perturbation robustness

ADef: an Iterative Algorithm to Construct Adversarial Deformations

Keywords:Adversarial examples, deformations, deep neural networks, computer vision
TL;DR:We propose a new, efficient algorithm to construct adversarial examples by means of deformations, rather than additive perturbations.

Discriminator-Actor-Critic: Addressing Sample Inefficiency and Reward Bias in Adversarial Imitation Learning

Keywords:deep learning, reinforcement learning, imitation learning, adversarial learning
TL;DR:We address sample inefficiency and reward bias in adversarial imitation learning algorithms such as GAIL and AIRL.

Doubly Reparameterized Gradient Estimators for Monte Carlo Objectives

Keywords:variational autoencoder, reparameterization trick, IWAE, VAE, RWS, JVI
TL;DR:Doubly reparameterized gradient estimators provide unbiased variance reduction which leads to improved performance.

Learning Recurrent Binary/Ternary Weights

Keywords:Quantized Recurrent Neural Network, Hardware Implementation, Deep Learning
TL;DR:We propose high-performance LSTMs with binary/ternary weights, that can greatly reduce implementation complexity

Learning concise representations for regression by evolving networks of trees

Keywords:regression, stochastic optimization, evolutionary compution, feature engineering
TL;DR:Representing the network architecture as a set of syntax trees and optimizing their structure leads to accurate and concise regression models.

Efficient Training on Very Large Corpora via Gramian Estimation

Keywords:similarity learning, pairwise learning, matrix factorization, Gramian estimation, variance reduction, neural embedding models, recommender systems
TL;DR:We develop efficient methods to train neural embedding models with a dot-product structure, by reformulating the objective function in terms of generalized Gram matrices, and maintaining estimates of those matrices.

MAE: Mutual Posterior-Divergence Regularization for Variational AutoEncoders

Keywords:None
TL;DR:None

Residual Non-local Attention Networks for Image Restoration

Keywords:Non-local network, attention network, image restoration, residual learning
TL;DR:New state-of-the-art framework for image restoration

Meta-Learning For Stochastic Gradient MCMC

Keywords:Meta Learning, MCMC
TL;DR:This paper proposes a method to automate the design of stochastic gradient MCMC proposal using meta learning approach.

Systematic Generalization: What Is Required and Can It Be Learned?

Keywords:systematic generalization, language understanding, visual questions answering, neural module networks
TL;DR:We show that modular structured models are the best in terms of systematic generalization and that their end-to-end versions don’t generalize as well.

Efficient Lifelong Learning with A-GEM

Keywords:Lifelong Learning, Continual Learning, Catastrophic Forgetting, Few-shot Transfer
TL;DR:An efficient lifelong learning algorithm that provides a better trade-off between accuracy and time/ memory complexity compared to other algorithms.

Multi-step Retriever-Reader Interaction for Scalable Open-domain Question Answering

Keywords:Open domain Question Answering, Reinforcement Learning, Query reformulation
TL;DR:Paragraph retriever and machine reader interacts with each other via reinforcement learning to yield large improvements on open domain datasets

Double Viterbi: Weight Encoding for High Compression Ratio and Fast On-Chip Reconstruction for Deep Neural Network

Keywords:quantization, pruning, memory footprint, model compression, sparse matrix
TL;DR:We present a new weight encoding scheme which enables high compression ratio and fast sparse-to-dense matrix conversion.

Overcoming the Disentanglement vs Reconstruction Trade-off via Jacobian Supervision

Keywords:disentangling, autoencoders, jacobian, face manipulation
TL;DR:A method for learning image representations that are good for both disentangling factors of variation and obtaining faithful reconstructions.

RotatE: Knowledge Graph Embedding by Relational Rotation in Complex Space

Keywords:knowledge graph embedding, knowledge graph completion, adversarial sampling
TL;DR:A new state-of-the-art approach for knowledge graph embedding.

Guiding Policies with Language via Meta-Learning

Keywords:meta-learning, language grounding, interactive
TL;DR:We propose a meta-learning method for interactively correcting policies with natural language.

AdaShift: Decorrelation and Convergence of Adaptive Learning Rate Methods

Keywords:optimizer, Adam, convergence, decorrelation
TL;DR:We analysis and solve the non-convergence issue of Adam.

AD-VAT: An Asymmetric Dueling mechanism for learning Visual Active Tracking

Keywords:Active tracking, reinforcement learning, adversarial learning, multi agent
TL;DR:We propose AD-VAT, where the tracker and the target object, viewed as two learnable agents, are opponents and can mutually enhance during training.

Marginal Policy Gradients: A Unified Family of Estimators for Bounded Action Spaces with Applications

Keywords:None
TL;DR:None

On Self Modulation for Generative Adversarial Networks

Keywords:unsupervised learning, generative adversarial networks, deep generative modelling
TL;DR:A simple GAN modification that improves performance across many losses, architectures, regularization schemes, and datasets.

Off-Policy Evaluation and Learning from Logged Bandit Feedback: Error Reduction via Surrogate Policy

Keywords:None
TL;DR:None

Subgradient Descent Learns Orthogonal Dictionaries

Keywords:Dictionary learning, Sparse coding, Non-convex optimization, Theory
TL;DR:Efficient dictionary learning by L1 minimization via a novel analysis of the non-convex non-smooth geometry.

ClariNet: Parallel Wave Generation in End-to-End Text-to-Speech

Keywords:None
TL;DR:None

MARGINALIZED AVERAGE ATTENTIONAL NETWORK FOR WEAKLY-SUPERVISED LEARNING

Keywords:feature aggregation, weakly supervised learning, temporal action localization
TL;DR:A novel marginalized average attentional network for weakly-supervised temporal action localization

Towards GAN Benchmarks Which Require Generalization

Keywords:evaluation, generative adversarial networks, adversarial divergences
TL;DR:We argue that GAN benchmarks must require a large sample from the model to penalize memorization and investigate whether neural network divergences have this property.

A Closer Look at Few-shot Classification

Keywords:few shot classification, meta-learning
TL;DR: A detailed empirical study in few-shot classification that revealing challenges in standard evaluation setting and showing a new direction.

Meta-Learning Probabilistic Inference for Prediction

Keywords:probabilistic models, approximate inference, few-shot learning, meta-learning
TL;DR:Novel framework for meta-learning that unifies and extends a broad class of existing few-shot learning methods. Achieves strong performance on few-shot learning benchmarks without requiring iterative test-time inference.

Deep reinforcement learning with relational inductive biases

Keywords:relational reasoning, reinforcement learning, graph neural networks, starcraft, generalization, inductive bias
TL;DR:Relational inductive biases improve out-of-distribution generalization capacities in model-free reinforcement learning agents

Relaxed Quantization for Discretized Neural Networks

Keywords:Quantization, Compression, Neural Networks, Efficiency
TL;DR:We introduce a technique that allows for gradient based training of quantized neural networks.

Tree-Structured Recurrent Switching Linear Dynamical Systems for Multi-Scale Modeling

Keywords:None
TL;DR:None

STCN: Stochastic Temporal Convolutional Networks

Keywords:latent variables, variational inference, temporal convolutional networks, sequence modeling, auto-regressive modeling
TL;DR:We combine the computational advantages of temporal convolutional architectures with the expressiveness of stochastic latent variables.

Soft Q-Learning with Mutual-Information Regularization

Keywords:None
TL;DR:None

On the Turing Completeness of Modern Neural Network Architectures

Keywords:Transformer, NeuralGPU, Turing completeness
TL;DR:We show that the Transformer architecture and the Neural GPU are Turing complete.

Improving Differentiable Neural Computers Through Memory Masking, De-allocation, and Link Distribution Sharpness Control

Keywords:None
TL;DR:None

Evaluating Robustness of Neural Networks with Mixed Integer Programming

Keywords:verification, adversarial robustness, adversarial examples, deep learning
TL;DR:We efficiently verify the robustness of deep neural models with over 100,000 ReLUs, certifying more samples than the state-of-the-art and finding more adversarial examples than a strong first-order attack.

Random mesh projectors for inverse problems

Keywords:imaging, inverse problems, subspace projections, random Delaunay triangulations, CNN, geophysics, regularization
TL;DR:We solve ill-posed inverse problems with scarce ground truth examples by estimating an ensemble of random projections of the model instead of the model itself.

Multi-Agent Dual Learning

Keywords:None
TL;DR:None

Complement Objective Training

Keywords:optimization, entropy, image recognition, natural language understanding, adversarial attacks, deep learning
TL;DR:We propose Complement Objective Training (COT), a new training paradigm that optimizes both the primary and complement objectives for effectively learning the parameters of neural networks.

Mode Normalization

Keywords:Deep Learning, Expert Models, Normalization, Computer Vision
TL;DR:We present a novel normalization method for deep neural networks that is robust to multi-modalities in intermediate feature distributions.

Detecting Egregious Responses in Neural Sequence-to-sequence Models

Keywords:Deep Learning, Natural Language Processing, Adversarial Attacks, Dialogue Response Generation
TL;DR:This paper aims to provide an empirical answer to the question of whether well-trained dialogue response model can output malicious responses.

Learning Actionable Representations with Goal Conditioned Policies

Keywords:Representation Learning, Reinforcement Learning
TL;DR:Learning state representations which capture factors necessary for control

Verification of Non-Linear Specifications for Neural Networks

Keywords:None
TL;DR:None

Generating Liquid Simulations with Deformation-aware Neural Networks

Keywords:Learning weighting and deformations of space-time data sets for highly efficient approximations of liquid behavior.
TL;DR:deformation learning, spatial transformer networks, fluid simulation

DyRep: Learning Representations over Dynamic Graphs

Keywords:Dynamic Graphs, Representation Learning, Dynamic Processes, Temporal Point Process, Attention, Latent Representation
TL;DR:Models Representation Learning over dynamic graphs as latent hidden process bridging two observed processes of Topological Evolution of and Interactions on dynamic graphs.

Trellis Networks for Sequence Modeling

Keywords:sequence modeling, language modeling, recurrent networks, convolutional networks, trellis networks
TL;DR:Trellis networks are a new sequence modeling architecture that bridges recurrent and convolutional models and sets a new state of the art on word- and character-level language modeling.

Scalable Unbalanced Optimal Transport using Generative Adversarial Networks

Keywords:unbalanced optimal transport, generative adversarial networks, population modeling
TL;DR:We propose new methodology for unbalanced optimal transport using generative adversarial networks.

Solving the Rubik’s Cube with Approximate Policy Iteration

Keywords:reinforcement learning, Rubik’s Cube, approximate policy iteration, deep learning, deep reinforcement learning
TL;DR:We solve the Rubik’s Cube with pure reinforcement learning

Variance Reduction for Reinforcement Learning in Input-Driven Environments

Keywords:reinforcement learning, policy gradient, input-driven environments, variance reduction, baseline
TL;DR:For environments dictated partially by external input processes, we derive an input-dependent baseline that provably reduces the variance for policy gradient methods and improves the policy performance in a wide range of RL tasks.

Model-Predictive Policy Learning with Uncertainty Regularization for Driving in Dense Traffic

Keywords:model-based reinforcement learning, stochastic video prediction, autonomous driving
TL;DR:A model-based RL approach which uses a differentiable uncertainty penalty to learn driving policies from purely observational data.

GAN Dissection: Visualizing and Understanding Generative Adversarial Networks

Keywords:GAN representations are examined in detail, and sets of representation units are found that control the generation of semantic concepts in the output.
TL;DR:GANs, representation, interpretability, causality

Improving MMD-GAN Training with Repulsive Loss Function

Keywords:generative adversarial nets, loss function, maximum mean discrepancy, image generation, unsupervised learning
TL;DR:Rearranging the terms in maximum mean discrepancy yields a much better loss function for the discriminator of generative adversarial nets

Deterministic PAC-Bayesian generalization bounds for deep networks via generalizing noise-resilience

Keywords:generalization, PAC-Bayes, SGD, learning theory, implicit regularization
TL;DR:We provide a PAC-Bayes based generalization guarantee for uncompressed, deterministic deep networks by generalizing noise-resilience of the network on the training data to the test data.

Recall Traces: Backtracking Models for Efficient Reinforcement Learning

Keywords:Model free RL, Variational Inference
TL;DR:A backward model of previous (state, action) given the next state, i.e. P(s_t, a_t | s_{t+1}), can be used to simulate additional trajectories terminating at states of interest! Improves RL learning efficiency.

Stable Recurrent Models

Keywords:stability, gradient descent, non-convex optimization, recurrent neural networks
TL;DR:Stable recurrent models can be approximated by feed-forward networks and empirically perform as well as unstable models on benchmark tasks.

The Limitations of Adversarial Training and the Blind-Spot Attack

Keywords:Adversarial Examples, Adversarial Training, Blind-Spot Attack
TL;DR:We show that even the strongest adversarial training methods cannot defend against adversarial examples crafted on slightly scaled and shifted test images.

Efficiently testing local optimality and escaping saddles for ReLU networks

Keywords:local optimality, second-order stationary point, escaping saddle points, nondifferentiability, ReLU, empirical risk
TL;DR:A theoretical algorithm for testing local optimality and extracting descent directions at nondifferentiable points of empirical risks of one-hidden-layer ReLU networks.

ProxylessNAS: Direct Neural Architecture Search on Target Task and Hardware

Keywords:Neural Architecture Search, Efficient Neural Networks
TL;DR:Proxy-less neural architecture search for directly learning architectures on large-scale target task (ImageNet) while reducing the cost to the same level of normal training.

Hierarchical Reinforcement Learning via Advantage-Weighted Information Maximization

Keywords:Hierarchical reinforcement learning, Representation learning, Continuous control
TL;DR:This paper presents a hierarchical reinforcement learning framework based on deterministic option policies and mutual information maximization.

Generalizable Adversarial Training via Spectral Normalization

Keywords:None
TL;DR:None

Adversarial Domain Adaptation for Stable Brain-Machine Interfaces

Keywords:Brain-Machine Interfaces, Domain Adaptation, Adversarial Networks
TL;DR:We implement an adversarial domain adaptation network to stabilize a fixed Brain-Machine Interface against gradual changes in the recorded neural signals.

Deep Online Learning Via Meta-Learning: Continual Adaptation for Model-Based RL

Keywords:None
TL;DR:None

Deep Anomaly Detection with Outlier Exposure

Keywords:confidence, uncertainty, anomaly, robustness
TL;DR:OE teaches anomaly detectors to learn heuristics for detecting unseen anomalies; experiments are in classification, density estimation, and calibration in NLP and vision settings; we do not tune on test distribution samples, unlike previous work

Contingency-Aware Exploration in Reinforcement Learning

Keywords:Reinforcement Learning, Exploration, Contingency-Awareness
TL;DR:We investigate contingency-awareness and controllable aspects in exploration and achieve state-of-the-art performance on Montezuma’s Revenge without expert demonstrations.

Context-adaptive Entropy Model for End-to-end Optimized Image Compression

Keywords:image compression, deep learning, entropy model
TL;DR:Context-adaptive entropy model for use in end-to-end optimized image compression, which significantly improves compression performance

Variational Discriminator Bottleneck: Improving Imitation Learning, Inverse RL, and GANs by Constraining Information Flow

Keywords:reinforcement learning, generative adversarial networks, imitation learning, inverse reinforcement learning, information bottleneck
TL;DR:Regularizing adversarial learning with an information bottleneck, applied to imitation learning, inverse reinforcement learning, and generative adversarial networks.

Meta-learning with differentiable closed-form solvers

Keywords:few-shot learning, one-shot learning, meta-learning, deep learning, ridge regression, classification
TL;DR:We propose a meta-learning approach for few-shot classification that achieves strong performance at high-speed by back-propagating through the solution of fast solvers, such as ridge regression or logistic regression.

Learning Self-Imitating Diverse Policies

Keywords:Reinforcement-learning, Imitation-learning, Ensemble-training
TL;DR:Policy optimization by using past good rollouts from the agent; learning shaped rewards via divergence minimization; SVPG with JS-kernel for population-based exploration.

ProxQuant: Quantized Neural Networks via Proximal Operators

Keywords:Model quantization, Optimization, Regularization
TL;DR:A principled framework for model quantization using the proximal gradient method, with empirical evaluation and theoretical convergence analyses.

Universal Transformers

Keywords:sequence-to-sequence, rnn, transformer, machine translation, language understanding, learning to execute
TL;DR:We introduce the Universal Transformer, a self-attentive parallel-in-time recurrent sequence model that outperforms Transformers and LSTMs on a wide range of sequence-to-sequence tasks, including machine translation.

Learning to Adapt in Dynamic, Real-World Environments through Meta-Reinforcement Learning

Keywords:meta-learning, reinforcement learning, meta reinforcement learning, online adaptation
TL;DR:A model-based meta-RL algorithm that enables a real robot to adapt online in dynamic environments

L-Shapley and C-Shapley: Efficient Model Interpretation for Structured Data

Keywords:Model Interpretation, Feature Selection
TL;DR:We develop two linear-complexity algorithms for model-agnostic model interpretation based on the Shapley value, in the settings where the contribution of features to the target is well-approximated by a graph-structured factorization.

Discovery of Natural Language Concepts in Individual Units of CNNs

Keywords:interpretability of deep neural networks, natural language representation
TL;DR:We show that individual units in CNN representations learned in NLP tasks are selectively responsive to natural language concepts.

Towards the first adversarially robust neural network model on MNIST

Keywords:None
TL;DR:None

Discriminator Rejection Sampling

Keywords:GANs, rejection sampling
TL;DR:We use a GAN discriminator to perform an approximate rejection sampling scheme on the output of the GAN generator.

Harmonic Unpaired Image-to-image Translation

Keywords:unpaired image-to-image translation, cyclegan, smoothness constraint
TL;DR:Smooth regularization over sample graph for unpaired image-to-image translation results in significantly improved consistency

Universal Successor Features Approximators

Keywords:None
TL;DR:None

Gradient Descent Provably Optimizes Over-parameterized Neural Networks

Keywords:theory, non-convex optimization, overparameterization, gradient descent
TL;DR:We prove gradient descent achieves zero training loss with a linear rate on over-parameterized neural networks.

Opportunistic Learning: Budgeted Cost-Sensitive Learning from Data Streams

Keywords:Cost-Aware Learning, Feature Acquisition, Reinforcement Learning, Stream Learning, Deep Q-Learning
TL;DR:An online algorithm for cost-aware feature acquisition and prediction

DARTS: Differentiable Architecture Search

Keywords:deep learning, autoML, neural architecture search, image classification, language modeling
TL;DR:We propose a differentiable architecture search algorithm for both convolutional and recurrent networks, achieving competitive performance with the state of the art using orders of magnitude less computation resources.

Feature-Wise Bias Amplification

Keywords:None
TL;DR:None

The relativistic discriminator: a key element missing from standard GAN

Keywords:Improving the quality and stability of GANs using a relativistic discriminator; IPM GANs (such as WGAN-GP) are a special case.
TL;DR:AI, deep learning, generative models, GAN

Understanding and Improving Interpolation in Autoencoders via an Adversarial Regularizer

Keywords:We propose a regularizer that improves interpolation and autoencoders and show that it also improves the learned representation for downstream tasks.
TL;DR:autoencoders, interpolation, unsupervised learning, representation learning, adversarial learning

Quasi-hyperbolic momentum and Adam for deep learning

Keywords:sgd, momentum, nesterov, adam, qhm, qhadam, optimization
TL;DR:Mix plain SGD and momentum (or do something similar with Adam) for great profit.

Local SGD Converges Fast and Communicates Little

Keywords:optimization, communication, theory, stochastic gradient descent, SGD, mini-batch, local SGD, parallel restart SGD, distributed training
TL;DR:We prove that parallel local SGD achieves linear speedup with much lesser communication than parallel mini-batch SGD.

Learning Finite State Representations of Recurrent Policy Networks

Keywords:recurrent neural networks, finite state machine, quantization, interpretability, autoencoder, moore machine, reinforcement learning, imitation learning, representation, Atari, Tomita
TL;DR:Extracting a finite state machine from a recurrent neural network via quantization for the purpose of interpretability with experiments on Atari.

Multilingual Neural Machine Translation with Knowledge Distillation

Keywords:NMT, Multilingual NMT, Knowledge Distillation
TL;DR:We proposed a knowledge distillation based method to boost the accuracy of multilingual neural machine translation.

MisGAN: Learning from Incomplete Data with Generative Adversarial Networks

Keywords:generative models, missing data
TL;DR:This paper presents a GAN-based framework for learning the distribution from high-dimensional incomplete data.

A Direct Approach to Robust Deep Learning Using Adversarial Networks

Keywords:deep learning, adversarial learning, generative adversarial networks
TL;DR:Jointly train an adversarial noise generating network with a classification network to provide better robustness to adversarial attacks.

Combinatorial Attacks on Binarized Neural Networks

Keywords:binarized neural networks, combinatorial optimization, integer programming
TL;DR:Gradient-based attacks on binarized neural networks are not effective due to the non-differentiability of such networks; Our IPROP algorithm solves this problem using integer optimization

Exemplar Guided Unsupervised Image-to-Image Translation with Semantic Consistency

Keywords:image-to-image translation, image generation, domain adaptation
TL;DR:We propose the Exemplar Guided & Semantically Consistent Image-to-image Translation (EGSC-IT) network which conditions the translation process on an exemplar image in the target domain.

ARM: Augment-REINFORCE-Merge Gradient for Stochastic Binary Networks

Keywords:Antithetic sampling, variable augmentation, deep discrete latent variable models, variance reduction, variational auto-encoder
TL;DR:An unbiased and low-variance gradient estimator for discrete latent variable models

Building Dynamic Knowledge Graphs from Text using Machine Reading Comprehension

Keywords:None
TL;DR:None

Information asymmetry in KL-regularized RL

Keywords:Deep Reinforcement Learning, Continuous Control, RL as Inference
TL;DR:Limiting state information for the default policy can improvement performance, in a KL-regularized RL framework where both agent and default policy are optimized together

TimbreTron: A WaveNet(CycleGAN(CQT(Audio))) Pipeline for Musical Timbre Transfer

Keywords:Generative models, Timbre Transfer, Wavenet, CycleGAN
TL;DR:We present the TimbreTron, a pipeline for perfoming high-quality timbre transfer on musical waveforms using CQT-domain style transfer.

Whitening and Coloring Batch Transform for GANs

Keywords:None
TL;DR:None

Learnable Embedding Space for Efficient Neural Architecture Compression

Keywords:Network Compression, Neural Architecture Search, Bayesian Optimization, Architecture Embedding
TL;DR:We propose a method to incrementally learn an embedding space over the domain of network architectures, to enable the careful selection of architectures for evaluation during compressed architecture search.

On the Sensitivity of Adversarial Robustness to Input Data Distributions

Keywords:adversarial robustness, adversarial training, PGD training, adversarial perturbation, input data distribution
TL;DR:Robustness performance of PGD trained models are sensitive to semantics-preserving transformation of image datasets, which implies the trickiness of evaluation of robust learning algorithms in practice.

Minimal Images in Deep Neural Networks: Fragile Object Recognition in Natural Images

Keywords:None
TL;DR:None

A Statistical Approach to Assessing Neural Network Robustness

Keywords:neural network verification, multi-level splitting, formal verification
TL;DR:We introduce a statistical approach to assessing neural network robustness that provides an informative notion of how robust a network is, rather than just the conventional binary assertion of whether or not of property is violated.

Improving Sequence-to-Sequence Learning via Optimal Transport

Keywords:None
TL;DR:None

PATE-GAN: Generating Synthetic Data with Differential Privacy Guarantees

Keywords:None
TL;DR:None

Integer Networks for Data Compression with Latent-Variable Models

Keywords:data compression, variational models, network quantization
TL;DR:We train variational models with quantized networks for computational determinism. This enables using them for cross-platform data compression.

Value Propagation Networks

Keywords:Reinforcement Learning, Value Iteration, Navigation, Convolutional Neural Networks, Learning to plan
TL;DR:We present planners based on convnets that are sample-efficient and that generalize to larger instances of navigation and pathfinding problems.

Bayesian Policy Optimization for Model Uncertainty

Keywords:Bayes-Adaptive Markov Decision Process, Model Uncertainty, Bayes Policy Optimization
TL;DR:We formulate model uncertainty in Reinforcement Learning as a continuous Bayes-Adaptive Markov Decision Process and present a method for practical and scalable Bayesian policy optimization.

Variational Bayesian Phylogenetic Inference

Keywords:Bayesian phylogenetic inference, Variational inference, Subsplit Bayesian networks
TL;DR:The first variational Bayes formulation of phylogenetic inference, a challenging inference problem over structures with intertwined discrete and continuous components

LEARNING FACTORIZED REPRESENTATIONS FOR OPEN-SET DOMAIN ADAPTATION

Keywords:None
TL;DR:None

On the Universal Approximability and Complexity Bounds of Quantized ReLU Neural Networks

Keywords:Quantized Neural Networks, Universial Approximability, Complexity Bounds, Optimal Bit-width
TL;DR:This paper proves the universal approximability of quantized ReLU neural networks and puts forward the complexity bound given arbitrary error.

Learning Localized Generative Models for 3D Point Clouds via Graph Convolution

Keywords:A GAN using graph convolution operations with dynamically computed graphs from hidden features
TL;DR:GAN, graph convolution, point clouds

ACCELERATING NONCONVEX LEARNING VIA REPLICA EXCHANGE LANGEVIN DIFFUSION

Keywords:None
TL;DR:None

Dynamically Unfolding Recurrent Restorer: A Moving Endpoint Control Method for Image Restoration

Keywords:image restoration, differential equation
TL;DR:We propose a novel method to handle image degradations of different levels by learning a diffusion terminal time. Our model can generalize to unseen degradation level and different noise statistic.

Bias-Reduced Uncertainty Estimation for Deep Neural Classifiers

Keywords:Uncertainty estimation, Deep learning
TL;DR:We use snapshots from the training process to improve any uncertainty estimation method of a DNN classifier.

CAMOU: Learning Physical Vehicle Camouflages to Adversarially Attack Detectors in the Wild

Keywords:Adversarial Attack, Object Detection, Synthetic Simulation
TL;DR:We propose a method to learn physical vehicle camouflage to adversarially attack object detectors in the wild. We find our camouflage effective and transferable.

Learning Latent Superstructures in Variational Autoencoders for Deep Multidimensional Clustering

Keywords:latent tree model, variational autoencoder, deep learning, latent variable model, bayesian network, structure learning, stepwise em, message passing, graphical model, multidimensional clustering, unsupervised learning
TL;DR:We investigate a variant of variational autoencoders where there is a superstructure of discrete latent variables on top of the latent features.

Learning Programmatically Structured Representations with Perceptor Gradients

Keywords:None
TL;DR:None

Variational Autoencoders with Jointly Optimized Latent Dependency Structure

Keywords:deep generative models, structure learning
TL;DR:We propose a method for learning latent dependency structure in variational autoencoders.

The Unusual Effectiveness of Averaging in GAN Training

Keywords:None
TL;DR:None

Beyond Pixel Norm-Balls: Parametric Adversaries using an Analytically Differentiable Renderer

Keywords:adversarial examples, norm-balls, differentiable renderer
TL;DR:Enabled by a novel differentiable renderer, we propose a new metric that has real-world implications for evaluating adversarial machine learning algorithms, resolving the lack of realism of the existing metric based on pixel norms.

Diversity is All You Need: Learning Skills without a Reward Function

Keywords:reinforcement learning, unsupervised learning, skill discovery
TL;DR:We propose an algorithm for learning useful skills without a reward function, and show how these skills can be used to solve downstream tasks.

Supervised Policy Update for Deep Reinforcement Learning

Keywords:Deep Reinforcement Learning
TL;DR:first posing and solving the sample efficiency optimization problem in the non-parameterized policy space, and then solving a supervised regression problem to find a parameterized policy that is near the optimal non-parameterized policy.

Learning sparse relational transition models

Keywords:Deictic reference, relational model, rule-based transition model
TL;DR:A new approach that learns a representation for describing transition models in complex uncertaindomains using relational rules.

Learning to Schedule Communication in Multi-agent Reinforcement Learning

Keywords:None
TL;DR:None

Hierarchical RL Using an Ensemble of Proprioceptive Periodic Policies

Keywords:None
TL;DR:None

Multi-class classification without multi-class labels

Keywords:None
TL;DR:None

What do you learn from context? Probing for sentence structure in contextualized word representations

Keywords:natural language processing, word embeddings, transfer learning, interpretability
TL;DR:We probe for sentence structure in ELMo and related contextual embedding models. We find existing models efficiently encode syntax and show evidence of long-range dependencies, but only offer small improvements on semantic tasks.

Spectral Inference Networks: Unifying Deep and Spectral Learning

Keywords:spectral learning, unsupervised learning, manifold learning, dimensionality reduction
TL;DR:We show how to learn spectral decompositions of linear operators with deep learning, and use it for unsupervised learning without a generative model.

PeerNets: Exploiting Peer Wisdom Against Adversarial Attacks

Keywords:None
TL;DR:None

Attentive Neural Processes

Keywords:Neural Processes, Conditional Neural Processes, Stochastic Processes, Regression, Attention
TL;DR:A model for regression that learns conditional distributions of a stochastic process, by incorporating attention into Neural Processes.

Representation Degeneration Problem in Training Natural Language Generation Models

Keywords:None
TL;DR:None

Hierarchical interpretations for neural network predictions

Keywords:interpretability, natural language processing, computer vision
TL;DR:We introduce and validate hierarchical local interpretations, the first technique to automatically search for and display important interactions for individual predictions made by LSTMs and CNNs.

Spreading vectors for similarity search

Keywords:dimensionality reduction, similarity search, indexing, differential entropy
TL;DR:We learn a neural network that uniformizes the input distribution, which leads to competitive indexing performance in high-dimensional space

A Convergence Analysis of Gradient Descent for Deep Linear Neural Networks

Keywords:Deep Learning, Learning Theory, Non-Convex Optimization
TL;DR:We analyze gradient descent for deep linear neural networks, providing a guarantee of convergence to global optimum at a linear rate.

Feed-forward Propagation in Probabilistic Neural Networks with Categorical and Max Layers

Keywords:probabilistic neural network, uncertainty, dropout, bayesian, softmax, argmax, logsumexp
TL;DR:Approximating mean and variance of the NN output over noisy input / dropout / uncertain parameters. Analytic approximations for argmax, softmax and max layers.

Measuring and regularizing networks in function space

Keywords:function space, Hilbert space, empirical characterization, multitask learning, catastrophic forgetting, optimization, natural gradient
TL;DR:We find movement in function space is not proportional to movement in parameter space during optimization. We propose a new natural-gradient style optimizer to address this.

Fluctuation-dissipation relations for stochastic gradient descent

Keywords:stochastic gradient descent, adaptive method, loss surface, Hessian
TL;DR:We prove fluctuation-dissipation relations for SGD, which can be used to (i) adaptively set learning rates and (ii) probe loss surfaces.

Poincare Glove: Hyperbolic Word Embeddings

Keywords:word embeddings, hyperbolic spaces, poincare ball, hypernymy, analogy, similarity, gaussian embeddings
TL;DR:We embed words in the hyperbolic space and make the connection with the Gaussian word embeddings.

Episodic Curiosity through Reachability

Keywords:deep learning, reinforcement learning, curiosity, exploration, episodic memory
TL;DR:We propose a novel model of curiosity based on episodic memory and the ideas of reachability which allows us to overcome the known “couch-potato” issues of prior work.

Phase-Aware Speech Enhancement with Deep Complex U-Net

Keywords:speech enhancement, deep learning, complex neural networks, phase estimation
TL;DR:This paper proposes a novel complex masking method for speech enhancement along with a loss function for efficient phase estimation.

Generative predecessor models for sample-efficient imitation learning

Keywords:None
TL;DR:None

Adaptive Estimators Show Information Compression in Deep Neural Networks

Keywords:deep neural networks, mutual information, information bottleneck, noise, L2 regularization
TL;DR:We developed robust mutual information estimates for DNNs and used them to observe compression in networks with non-saturating activation functions

Multilingual Neural Machine Translation With Soft Decoupled Encoding

Keywords:None
TL;DR:None

Approximating CNNs with Bag-of-local-Features models works surprisingly well on ImageNet

Keywords:interpretability, representation learning, bag of features, deep learning, object recognition
TL;DR:Aggregating class evidence from many small image patches suffices to solve ImageNet, yields more interpretable models and can explain aspects of the decision-making of popular DNNs.

Reward Constrained Policy Optimization

Keywords:reinforcement learning, markov decision process, constrained markov decision process, deep learning
TL;DR:For complex constraints in which it is not easy to estimate the gradient, we use the discounted penalty as a guiding signal. We prove that under certain assumptions it converges to a feasible solution.

On the Relation Between the Sharpest Directions of DNN Loss and the SGD Step Length

Keywords:optimization, generalization, theory of deep learning, SGD, hessian
TL;DR:SGD is steered early on in training towards a region in which its step is too large compared to curvature, which impacts the rest of training.

Modeling the Long Term Future in Model-Based Reinforcement Learning

Keywords:model-based reinforcement learning, variation inference
TL;DR:incorporating, in the model, latent variables that encode future content improves the long-term prediction accuracy, which is critical for better planning in model-based RL.

Understanding Straight-Through Estimator in Training Activation Quantized Neural Nets

Keywords:straight-through estimator, quantized activation, binary neuron
TL;DR:We make theoretical justification for the concept of straight-through estimator.

DISTRIBUTIONAL CONCAVITY REGULARIZATION FOR GANS

Keywords:None
TL;DR:None

LeMoNADe: Learned Motif and Neuronal Assembly Detection in calcium imaging videos

Keywords:VAE, unsupervised learning, neuronal assemblies, calcium imaging analysis
TL;DR:We present LeMoNADe, an end-to-end learned motif detection method directly operating on calcium imaging videos.

Competitive experience replay

Keywords:reinforcement learning, sparse reward, goal-based learning
TL;DR:a novel method to learn with sparse reward using adversarial reward re-labeling

Multi-Domain Adversarial Learning

Keywords:multi-domain learning, domain adaptation, adversarial learning, H-divergence, deep representation learning, high-content microscopy
TL;DR:Adversarial Domain adaptation and Multi-domain learning: a new loss to handle multi- and single-domain classes in the semi-supervised setting.

ProMP: Proximal Meta-Policy Search

Keywords:Meta-Reinforcement Learning, Meta-Learning, Reinforcement-Learning
TL;DR:A novel and theoretically grounded meta-reinforcement learning algorithm

Don’t Settle for Average, Go for the Max: Fuzzy Sets and Max-Pooled Word Vectors

Keywords:word vectors, sentence representations, distributed representations, fuzzy sets, bag-of-words, unsupervised learning, word vector compositionality, max-pooling, Jaccard index
TL;DR:Max-pooled word vectors with fuzzy Jaccard set similarity are an extremely competitive baseline for semantic similarity; we propose a simple dynamic variant that performs even better.

Stable Opponent Shaping in Differentiable Games

Keywords:multi-agent learning, multiple interacting losses, opponent shaping, exploitation, convergence
TL;DR:Opponent shaping is a powerful approach to multi-agent learning but can prevent convergence; our SOS algorithm fixes this with strong guarantees in all differentiable games.

A Mean Field Theory of Batch Normalization

Keywords:theory, batch normalization, mean field theory, trainability
TL;DR:Batch normalization causes exploding gradients in vanilla feedforward networks.

Learning Exploration Policies for Navigation

Keywords:None
TL;DR:None

Distribution-Interpolation Trade off in Generative Models

Keywords:generative models, latent distribution, Cauchy distribution, interpolations
TL;DR:We theoretically prove that linear interpolations are unsuitable for analysis of trained implicit generative models.

Learning to Describe Scenes with Programs

Keywords:Structured scene representations, program synthesis
TL;DR:We present scene programs, a structured scene representation that captures both low-level object appearance and high-level regularity in the scene.

Visceral Machines: Risk-Aversion in Reinforcement Learning with Intrinsic Physiological Rewards

Keywords:Reinforcement Learning, Simulation, Affective Computing
TL;DR:We present a novel approach to reinforcement learning that leverages a task-independent intrinsic reward function trained on peripheral pulse measurements that are correlated with human autonomic nervous system responses.

Deep Frank-Wolfe For Neural Network Optimization

Keywords:optimization, conditional gradient, Frank-Wolfe, SVM
TL;DR:We train neural networks by locally linearizing them and using a linear SVM solver (Frank-Wolfe) at each iteration.

LEARNING TO PROPAGATE LABELS: TRANSDUCTIVE PROPAGATION NETWORK FOR FEW-SHOT LEARNING

Keywords:few-shot learning, meta-learning, label propagation, manifold learning
TL;DR:We propose a novel meta-learning framework for transductive inference that classifies the entire test set at once to alleviate the low-data problem.

Improving the Generalization of Adversarial Training with Domain Adaptation

Keywords:adversarial training, domain adaptation, adversarial example, deep learning
TL;DR:We propose a novel adversarial training with domain adaptation method that significantly improves the generalization ability on adversarial examples from different attacks.

Dimensionality Reduction for Representing the Knowledge of Probabilistic Models

Keywords:metric learning, distance learning, dimensionality reduction, bound guarantees
TL;DR:dimensionality reduction for cases where examples can be represented as soft probability distributions

Learning protein sequence embeddings using information from structure

Keywords:sequence embedding, sequence alignment, RNN, LSTM, protein structure, amino acid sequence, contextual embeddings, transmembrane prediction
TL;DR:We present a method for learning protein sequence embedding models using structural information in the form of global structural similarity between proteins and within protein residue-residue contacts.

Variational Smoothing in Recurrent Neural Network Language Models

Keywords:None
TL;DR:None

Biologically-Plausible Learning Algorithms Can Scale to Large Datasets

Keywords:biologically plausible learning algorithm, ImageNet, sign-symmetry, feedback alignment
TL;DR:Biologically plausible learning algorithms, particularly sign-symmetry, work well on ImageNet

Coarse-grain Fine-grain Coattention Network for Multi-evidence Question Answering

Keywords:question answering, reading comprehension, nlp, natural language processing, attention, representation learning
TL;DR:A new state-of-the-art model for multi-evidence question answering using coarse-grain fine-grain hierarchical attention.

Learning a Meta-Solver for Syntax-Guided Program Synthesis

Keywords:Syntax-guided Synthesis, Context Free Grammar, Logical Specification, Representation Learning, Meta Learning, Reinforcement Learning
TL;DR:We propose a meta-learning framework that learns a transferable policy from only weak supervision to solve synthesis tasks with different logical specifications and grammars.

Towards Robust, Locally Linear Deep Networks

Keywords:robust derivatives, transparency, interpretability
TL;DR:A scalable algorithm to establish robust derivatives of deep networks w.r.t. the inputs.

How Important is a Neuron

Keywords:None
TL;DR:None

Learning to Make Analogies by Contrasting Abstract Relational Structure

Keywords:cognitive science, analogy, psychology, cognitive theory, cognition, abstraction, generalization
TL;DR:The most robust capacity for analogical reasoning is induced when networks learn analogies by contrasting abstract relational structures in their input domains.

Learning what you can do before doing anything

Keywords:unsupervised learning, vision, motion, action space, video prediction, variational models
TL;DR:We learn a representation of an agent’s action space from pure visual observations. We use a recurrent latent variable approach with a novel composability loss.

Learning Grid Cells as Vector Representation of Self-Position Coupled with Matrix Representation of Self-Motion

Keywords:None
TL;DR:None

Universal Stagewise Learning for Non-Convex Problems with Convergence on Averaged Solutions

Keywords:None
TL;DR:None

Invariant and Equivariant Graph Networks

Keywords:graph learning, equivariance, deep learning
TL;DR:The paper provides a full characterization of permutation invariant and equivariant linear layers for graph data.

Robustness May Be at Odds with Accuracy

Keywords:adversarial examples, robust machine learning, robust optimization, deep feature representations
TL;DR:We show that adversarial robustness might come at the cost of standard classification performance, but also yields unexpected benefits.

Feature Intertwiner for Object Detection

Keywords:feature learning, computer vision, deep learning
TL;DR:(Camera-ready version) A feature intertwiner module to leverage features from one accurate set to help the learning of another less reliable set.

Adversarial Reprogramming of Neural Networks

Keywords:Adversarial, Neural Networks, Machine Learning Security
TL;DR:We introduce the first instance of adversarial attacks that reprogram the target model to perform a task chosen by the attacker—without the attacker needing to specify or compute the desired output for each test-time input.

G-SGD: Optimizing ReLU Neural Networks in its Positively Scale-Invariant Space

Keywords:None
TL;DR:None

From Hard to Soft: Understanding Deep Network Nonlinearities via Vector Quantization and Statistical Inference

Keywords:Spline, Vector Quantization, Inference, Nonlinearities, Deep Network
TL;DR:Reformulate deep networks nonlinearities from a vector quantization scope and bridge most known nonlinearities together.

Aggregated Momentum: Stability Through Passive Damping

Keywords:momentum, optimization, deep learning, neural networks
TL;DR:We introduce a simple variant of momentum optimization which is able to outperform classical momentum, Nesterov, and Adam on deep learning tasks with minimal hyperparameter tuning.

Variational Autoencoder with Arbitrary Conditioning

Keywords:unsupervised learning, generative models, conditional variational autoencoder, variational autoencoder, missing features multiple imputation, inpainting
TL;DR:We propose an extension of conditional variational autoencoder that allows conditioning on an arbitrary subset of the features and sampling the remaining ones.

Time-Agnostic Prediction: Predicting Predictable Video Frames

Keywords:visual prediction, subgoal generation, bottleneck states, time-agnostic
TL;DR:In visual prediction tasks, letting your predictive model choose which times to predict does two things: (i) improves prediction quality, and (ii) leads to semantically coherent “bottleneck state” predictions, which are useful for planning.

A Closer Look at Deep Learning Heuristics: Learning rate restarts, Warmup and Distillation

Keywords:deep learning heuristics, learning rate restarts, learning rate warmup, knowledge distillation, mode connectivity, SVCCA
TL;DR:We use empirical tools of mode connectivity and SVCCA to investigate neural network training heuristics of learning rate restarts, warmup and knowledge distillation.

Self-Monitoring Navigation Agent via Auxiliary Progress Estimation

Keywords:visual grounding, textual grounding, instruction-following, navigation agent
TL;DR:We propose a self-monitoring agent for the Vision-and-Language Navigation task.

Kernel Change-point Detection with Auxiliary Deep Generative Models

Keywords:deep kernel learning, generative models, kernel two-sample test, time series change-point detection
TL;DR:In this paper, we propose KL-CPD, a novel kernel learning framework for time series CPD that optimizes a lower bound of test power via an auxiliary generative model as a surrogate to the abnormal distribution.

Unsupervised Learning via Meta-Learning

Keywords:unsupervised learning, meta-learning
TL;DR:An unsupervised learning method that uses meta-learning to enable efficient learning of downstream image classification tasks, outperforming state-of-the-art methods.

Auxiliary Variational MCMC

Keywords:None
TL;DR:None

Neural network gradient-based learning of black-box function interfaces

Keywords:neural networks, black box functions, gradient descent
TL;DR:Training DNNs to interface w black box functions wo intermediate labels by using an estimator sub-network that can be replaced with the black box after training

Self-Tuning Networks: Bilevel Optimization of Hyperparameters using Structured Best-Response Functions

Keywords:hyperparameter optimization, game theory, optimization
TL;DR:We use a hypernetwork to predict optimal weights given hyperparameters, and jointly train everything together.

Unsupervised Control Through Non-Parametric Discriminative Rewards

Keywords:deep reinforcement learning, goals, UVFA, mutual information
TL;DR:Unsupervised reinforcement learning method for learning a policy to robustly achieve perceptually specified goals.

Interpolation-Prediction Networks for Irregularly Sampled Time Series

Keywords:irregular sampling, multivariate time series, supervised learning, interpolation, missing data
TL;DR:This paper presents a new deep learning architecture for addressing the problem of supervised learning with sparse and irregularly sampled multivariate time series.

Riemannian Adaptive Optimization Methods

Keywords:Riemannian optimization, adaptive, hyperbolic, curvature, manifold, adam, amsgrad, adagrad, rsgd, convergence
TL;DR:Adapting Adam, Amsgrad, Adagrad to Riemannian manifolds.

Minimal Random Code Learning: Getting Bits Back from Compressed Model Parameters

Keywords:compression, neural networks, bits-back argument, Bayesian, Shannon, information theory
TL;DR:This paper proposes an effective method to compress neural networks based on recent results in information theory.

Characterizing Audio Adversarial Examples Using Temporal Dependency

Keywords:audio adversarial example, mitigation, detection, machine learning
TL;DR:Adversarial audio discrimination using temporal dependency

Equi-normalization of Neural Networks

Keywords:convolutional neural networks, Normalization, Sinkhorn, Regularization
TL;DR:Fast iterative algorithm to balance the energy of a network while staying in the same functional equivalence class

Generalized Tensor Models for Recurrent Neural Networks

Keywords:expressive power, recurrent neural networks, Tensor-Train decomposition
TL;DR:Analysis of expressivity and generality of recurrent neural networks with ReLu nonlinearities using Tensor-Train decomposition.

Wizard of Wikipedia: Knowledge-Powered Conversational Agents

Keywords:dialogue, knowledge, language, conversation
TL;DR:We build knowledgeable conversational agents by conditioning on Wikipedia + a new supervised task.

Are adversarial examples inevitable?

Keywords:adversarial examples, neural networks, security
TL;DR:This paper identifies classes of problems for which adversarial examples are inescapable, and derives fundamental bounds on the susceptibility of any classifier to adversarial examples.

A Variational Inequality Perspective on Generative Adversarial Networks

Keywords:optimization, variational inequality, games, saddle point, extrapolation, averaging, extragradient, generative modeling, generative adversarial network
TL;DR:We cast GANs in the variational inequality framework and import techniques from this literature to optimize GANs better; we give algorithmic extensions and empirically test their performance for training GANs.

Learning-Based Frequency Estimation Algorithms

Keywords:streaming algorithms, heavy-hitters, Count-Min, Count-Sketch
TL;DR:Data stream algorithms can be improved using deep learning, while retaining performance guarantees.

From Language to Goals: Inverse Reinforcement Learning for Vision-Based Instruction Following

Keywords:inverse reinforcement learning, language grounding, instruction following, language-based learning
TL;DR:We ground language commands in a high-dimensional visual environment by learning language-conditioned rewards using inverse reinforcement learning.

Backpropamine: training self-modifying neural networks with differentiable neuromodulated plasticity

Keywords:meta-learning, reinforcement learning, plasticity, neuromodulation, Hebbian learning, recurrent neural networks
TL;DR:Neural networks can be trained to modify their own connectivity, improving their online learning performance on challenging tasks.

Recurrent Experience Replay in Distributed Reinforcement Learning

Keywords:RNN, LSTM, experience replay, distributed training, reinforcement learning
TL;DR:Investigation on combining recurrent neural networks and experience replay leading to state-of-the-art agent on both Atari-57 and DMLab-30 using single set of hyper-parameters.

A Generative Model For Electron Paths

Keywords:Molecules, Reaction Prediction, Graph Neural Networks, Deep Generative Models
TL;DR:A generative model for reaction prediction that learns the mechanistic electron steps of a reaction directly from raw reaction data.

Modeling Uncertainty with Hedged Instance Embeddings

Keywords:uncertainty, instance embedding, metric learning, probabilistic embedding
TL;DR:The paper proposes using probability distributions instead of points for instance embeddings tasks such as recognition and verification.

Beyond Greedy Ranking: Slate Optimization via List-CVAE

Keywords:CVAE, VAE, recommendation system, slate optimization, whole page optimization
TL;DR:We used a CVAE type model structure to learn to directly generate slates/whole pages for recommendation systems.

Stochastic Prediction of Multi-Agent Interactions from Partial Observations

Keywords:Dynamics modeling, partial observations, multi-agent interactions, predictive models
TL;DR:We present a method which learns to integrate temporal information and ambiguous visual information in the context of interacting agents.

GamePad: A Learning Environment for Theorem Proving

Keywords:Theorem proving, ITP, systems, neural embeddings
TL;DR:We introduce a system called GamePad to explore the application of machine learning methods to theorem proving in the Coq proof assistant.

GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding

Keywords:natural language understanding, multi-task learning, evaluation
TL;DR:We present a multi-task benchmark and analysis platform for evaluating generalization in natural language understanding systems.

On Computation and Generalization of Generative Adversarial Networks under Spectrum Control

Keywords:None
TL;DR:None

Large-Scale Study of Curiosity-Driven Learning

Keywords:exploration, curiosity, intrinsic reward, no extrinsic reward, unsupervised, no-reward, skills
TL;DR:An agent trained only with curiosity, and no extrinsic reward, does surprisingly well on 54 popular environments, including the suite of Atari games, Mario etc.

Unsupervised Discovery of Parts, Structure, and Dynamics

Keywords:Self-Supervised Learning, Visual Prediction, Hierarchical Models
TL;DR:Learning object parts, hierarchical structure, and dynamics by watching how they move

Music Transformer: Generating Music with Long-Term Structure

Keywords:music generation
TL;DR:We show the first successful use of Transformer in generating music that exhibits long-term structure.

BabyAI: A Platform to Study the Sample Efficiency of Grounded Language Learning

Keywords:language, learning, efficiency, imitation learning, reinforcement learning
TL;DR:We present the BabyAI platform for studying data efficiency of language learning with a human in the loop

Analyzing Inverse Problems with Invertible Neural Networks

Keywords:Inverse problems, Neural Networks, Uncertainty, Invertible Neural Networks
TL;DR:To analyze inverse problems with Invertible Neural Networks

RelGAN: Relational Generative Adversarial Networks for Text Generation

Keywords:None
TL;DR:None

The Singular Values of Convolutional Layers

Keywords:singular values, operator norm, convolutional layers, regularization
TL;DR:We characterize the singular values of the linear transformation associated with a standard 2D multi-channel convolutional layer, enabling their efficient computation.

An Empirical study of Binary Neural Networks’ Optimisation

Keywords:None
TL;DR:None

Approximability of Discriminators Implies Diversity in GANs

Keywords:Theory, Generative adversarial networks, Mode collapse, Generalization
TL;DR:GANs can in principle learn distributions sample-efficiently, if the discriminator class is compact and has strong distinguishing power against the particular generator class.

Learning Embeddings into Entropic Wasserstein Spaces

Keywords:Embedding, Wasserstein, Sinkhorn, Optimal Transport
TL;DR:We show that Wasserstein spaces are good targets for embedding data with complex semantic structure.

DeepOBS: A Deep Learning Optimizer Benchmark Suite

Keywords:deep learning, optimization
TL;DR:We provide a software package that drastically simplifies, automates, and improves the evaluation of deep learning optimizers.

InfoBot: Transfer and Exploration via the Information Bottleneck

Keywords:Information bottleneck, policy transfer, policy generalization, exploration
TL;DR:Training agents with goal-policy information bottlenecks promotes transfer and yields a powerful exploration bonus

The Comparative Power of ReLU Networks and Polynomial Kernels in the Presence of Sparse Latent Structure

Keywords:theory, representational power, universal approximators, polynomial kernels, latent sparsity, beyond worst case, separation result
TL;DR:Beyond-worst-case analysis of the representational power of ReLU nets & polynomial kernels — in particular in the presence of sparse latent structure.

Learning Implicitly Recurrent CNNs Through Parameter Sharing

Keywords:deep learning, architecture search, computer vision
TL;DR:We propose a method that enables CNN folding to create recurrent connections

Learning Particle Dynamics for Manipulating Rigid Bodies, Deformable Objects, and Fluids

Keywords:Dynamics modeling, Control, Particle-Based Representation
TL;DR:Learning particle dynamics with dynamic interaction graphs for simulating and control rigid bodies, deformable objects, and fluids.

Regularized Learning for Domain Adaptation under Label Shifts

Keywords:Deep Learning, Domain Adaptation, Label Shift, Importance Weights, Generalization
TL;DR:A practical and provably guaranteed approach for training efficiently classifiers in the presence of label shifts between Source and Target data sets

Von Mises-Fisher Loss for Training Sequence to Sequence Models with Continuous Outputs

Keywords:Language Generation, Regression, Word Embeddings, Machine Translation
TL;DR:Language generation using seq2seq models which produce word embeddings instead of a softmax based distribution over the vocabulary at each step enabling much faster training while maintaining generation quality

Relational Forward Models for Multi-Agent Learning

Keywords:multi-agent reinforcement learning, relational reasoning, forward models
TL;DR:Relational Forward Models for multi-agent learning make accurate predictions of agents’ future behavior, they produce intepretable representations and can be used inside agents.

Imposing Category Trees Onto Word-Embeddings Using A Geometric Construction

Keywords:category tree, word-embeddings, geometry
TL;DR:we show a geometric method to perfectly encode categroy tree information into pre-trained word-embeddings.

Two-Timescale Networks for Nonlinear Value Function Approximation

Keywords:Reinforcement learning, policy evaluation, nonlinear function approximation
TL;DR:We propose an architecture for learning value functions which allows the use of any linear policy evaluation algorithm in tandem with nonlinear feature learning.

Diversity-Sensitive Conditional Generative Adversarial Networks

Keywords:Conditional Generative Adversarial Network, mode-collapse, multi-modal generation, image-to-image translation, image in-painting, video prediction
TL;DR:We propose a simple and general approach that avoids a mode collapse problem in various conditional GANs.

Query-Efficient Hard-label Black-box Attack: An Optimization-based Approach

Keywords:None
TL;DR:None

Rethinking the Value of Network Pruning

Keywords:In structured network pruning, fine-tuning a pruned model only gives comparable performance with training it from scratch.
TL;DR:network pruning, network compression, architecture search, train from scratch

Hyperbolic Attention Networks

Keywords:Hyperbolic Geometry, Attention Methods, Reasoning on Graphs, Relation Learning, Scale Free Graphs, Transformers, Power Law
TL;DR:We propose to incorporate inductive biases and operations coming from hyperbolic geometry to improve the attention mechanism of the neural networks.

Learning from Positive and Unlabeled Data with a Selection Bias

Keywords:None
TL;DR:None

Adv-BNN: Improved Adversarial Defense through Robust Bayesian Neural Network

Keywords:None
TL;DR:None

Optimal Completion Distillation for Sequence Learning

Keywords:Sequence Learning, Edit Distance, Speech Recognition, Deep Reinforcement Learning
TL;DR:Optimal Completion Distillation (OCD) is a training procedure for optimizing sequence to sequence models based on edit distance which achieves state-of-the-art on end-to-end Speech Recognition tasks.

Caveats for information bottleneck in deterministic scenarios

Keywords:Information bottleneck behaves in surprising ways whenever the output is a deterministic function of the input.
TL;DR:information bottleneck, supervised learning, deep learning, information theory

Deep Learning 3D Shapes Using Alt-az Anisotropic 2-Sphere Convolution

Keywords:Spherical Convolution, Geometric deep learning, 3D shape analysis
TL;DR:A method for applying deep learning to 3D surfaces using their spherical descriptors and alt-az anisotropic convolution on 2-sphere.

Small nonlinearities in activation functions create bad local minima in neural networks

Keywords:spurious local minima, loss surface, optimization landscape, neural network
TL;DR:We constructively prove that even the slightest nonlinear activation functions introduce spurious local minima, for general datasets and activation functions.

Information Theoretic lower bounds on negative log likelihood

Keywords:latent variable modeling, rate-distortion theory, log likelihood bounds
TL;DR:Use rate-distortion theory to bound how much a latent variable model can be improved

Preferences Implicit in the State of the World

Keywords:Preference learning, Inverse reinforcement learning, Inverse optimal stochastic control, Maximum entropy reinforcement learning, Apprenticeship learning
TL;DR:When a robot is deployed in an environment that humans have been acting in, the state of the environment is already optimized for what humans want, and we can use this to infer human preferences.

A Kernel Random Matrix-Based Approach for Sparse PCA

Keywords:None
TL;DR:None

Bayesian Prediction of Future Street Scenes using Synthetic Likelihoods

Keywords:bayesian inference, segmentation, anticipation, multi-modality
TL;DR:Dropout based Bayesian inference is extended to deal with multi-modality and is evaluated on scene anticipation tasks.

There Are Many Consistent Explanations of Unlabeled Data: Why You Should Average

Keywords:semi-supervised learning, computer vision, classification, consistency regularization, flatness, weight averaging, stochastic weight averaging
TL;DR:Consistency-based models for semi-supervised learning do not converge to a single point but continue to explore a diverse set of plausible solutions on the perimeter of a flat region. Weight averaging helps improve generalization performance.

Large-Scale Answerer in Questioner’s Mind for Visual Dialog Question Generation

Keywords:None
TL;DR:None

Graph HyperNetworks for Neural Architecture Search

Keywords:None
TL;DR:None

DELTA: DEEP LEARNING TRANSFER USING FEATURE MAP WITH ATTENTION FOR CONVOLUTIONAL NETWORKS

Keywords:transfer learning, deep learning, regularization, attention, cnn
TL;DR:improving deep transfer learning with regularization using attention based feature maps

textTOvec: DEEP CONTEXTUALIZED NEURAL AUTOREGRESSIVE TOPIC MODELS OF LANGUAGE WITH DISTRIBUTED COMPOSITIONAL PRIOR

Keywords:neural topic model, natural language processing, text representation, language modeling, information retrieval, deep learning
TL;DR:Unified neural model of topic and language modeling to introduce language structure in topic models for contextualized topic vectors

Amortized Bayesian Meta-Learning

Keywords:variational inference, meta-learning, few-shot learning, uncertainty quantification
TL;DR:We propose a meta-learning method which efficiently amortizes hierarchical variational inference across training episodes.

Probabilistic Recursive Reasoning for Multi-Agent Reinforcement Learning

Keywords:Multi-agent Reinforcement Learning, Recursive Reasoning
TL;DR:We proposed a novel probabilisitic recursive reasoning (PR2) framework for multi-agent deep reinforcement learning tasks.

Learning Neural PDE Solvers with Convergence Guarantees

Keywords:Partial differential equation, deep learning
TL;DR:We learn a fast neural solver for PDEs that has convergence guarantees.

A new dog learns old tricks: RL finds classic optimization algorithms

Keywords:reinforcement learning, algorithms, adwords, knapsack, secretary
TL;DR:By combining ideas from traditional algorithms design and reinforcement learning, we introduce a novel framework for learning algorithms that solve online combinatorial optimization problems.

Deep Graph Infomax

Keywords:Unsupervised Learning, Graph Neural Networks, Graph Convolutions, Mutual Information, Infomax, Deep Learning
TL;DR:A new method for unsupervised representation learning on graphs, relying on maximizing mutual information between local and global representations in a graph. State-of-the-art results, competitive with supervised learning.

Theoretical Analysis of Auto Rate-Tuning by Batch Normalization

Keywords:batch normalization, scale invariance, learning rate, stationary point
TL;DR:We give a theoretical analysis of the ability of batch normalization to automatically tune learning rates, in the context of finding stationary points for a deep learning objective.

Per-Tensor Fixed-Point Quantization of the Back-Propagation Algorithm

Keywords:deep learning, reduced precision, fixed-point, quantization, back-propagation algorithm
TL;DR:We analyze and determine the precision requirements for training neural networks when all tensors, including back-propagated signals and weight accumulators, are quantized to fixed-point format.

FUNCTIONAL VARIATIONAL BAYESIAN NEURAL NETWORKS

Keywords:functional variational inference, Bayesian neural networks, stochastic processes
TL;DR:We perform functional variational inference on the stochastic processes defined by Bayesian neural networks.

NADPEx: An on-policy temporally consistent exploration method for deep reinforcement learning

Keywords:None
TL;DR:None

SPIGAN: Privileged Adversarial Learning from Simulation

Keywords:domain adaptation, GAN, semantic segmentation, simulation, privileged information
TL;DR:An unsupervised sim-to-real domain adaptation method for semantic segmentation using privileged information from a simulator with GAN-based image translation.

Generating Multi-Agent Trajectories using Programmatic Weak Supervision

Keywords:deep learning, generative models, imitation learning, hierarchical methods, data programming, weak supervision, spatiotemporal
TL;DR:We blend deep generative models with programmatic weak supervision to generate coordinated multi-agent trajectories of significantly higher quality than previous baselines.

Label super-resolution networks

Keywords:weakly supervised segmentation, land cover mapping, medical imaging
TL;DR:Super-resolving coarse labels into pixel-level labels, applied to aerial imagery and medical scans.

ANYTIME MINIBATCH: EXPLOITING STRAGGLERS IN ONLINE DISTRIBUTED OPTIMIZATION

Keywords:distributed optimization, gradient descent, minibatch, stragglers
TL;DR:Accelerate distributed optimization by exploiting stragglers.

Sample Efficient Adaptive Text-to-Speech

Keywords:few shot, meta learning, text to speech, wavenet
TL;DR:Sample efficient algorithms to adapt a text-to-speech model to a new voice style with the state-of-the-art performance.

Practical lossless compression with latent variables using bits back coding

Keywords:compression, variational auto-encoders, deep latent gaussian models, lossless compression, latent variables, approximate inference, variational inference
TL;DR:We do lossless compression of large image datasets using a VAE, beat existing compression algorithms.

Kernel RNN Learning (KeRNL)

Keywords:RNNs, Biologically plausible learning rules, Algorithm, Neural Networks, Supervised Learning
TL;DR:A biologically plausible learning rule for training recurrent neural networks

Deep, Skinny Neural Networks are not Universal Approximators

Keywords:This paper proves that skinny neural networks cannot approximate certain functions, no matter how deep they are.
TL;DR:neural network, universality, expressability

Large Scale Graph Learning From Smooth Signals

Keywords:None
TL;DR:None

Overcoming Catastrophic Forgetting for Continual Learning via Model Adaptation

Keywords:None
TL;DR:None

Analysis of Quantized Models

Keywords:weight quantization, gradient quantization, distributed learning
TL;DR:In this paper, we studied efficient training of loss-aware weight-quantized networks with quantized gradient in a distributed environment, both theoretically and empirically.

Deep learning generalizes because the parameter-function map is biased towards simple functions

Keywords:generalization, deep learning theory, PAC-Bayes, Gaussian processes, parameter-function map, simplicity bias
TL;DR:The parameter-function map of deep networks is hugely biased; this can explain why they generalize. We use PAC-Bayes and Gaussian processes to obtain nonvacuous bounds.

Learning when to Communicate at Scale in Multiagent Cooperative and Competitive Tasks

Keywords:multiagent, communication, competitive, cooperative, continuous, emergent, reinforcement learning
TL;DR:We introduce IC3Net, a single network which can be used to train agents in cooperative, competitive and mixed scenarios. We also show that agents can learn when to communicate using our model.

Synthetic Datasets for Neural Program Synthesis

Keywords:None
TL;DR:None

DPSNet: End-to-end Deep Plane Sweep Stereo

Keywords:Deep Learning, Stereo, Depth, Geometry
TL;DR:A convolution neural network for multi-view stereo matching whose design is inspired by best practices of traditional geometry-based approaches

Conditional Network Embeddings

Keywords:Network embedding, graph embedding, learning node representations, link prediction, multi-label classification of nodes
TL;DR:We introduce a network embedding method that accounts for prior information about the network, yielding superior empirical performance.

Defensive Quantization: When Efficiency Meets Robustness

Keywords:defensive quantization, model quantization, adversarial attack, efficiency, robustness
TL;DR:We designed a novel quantization methodology to jointly optimize the efficiency and robustness of deep learning models.

GO Gradient for Expectation-Based Objectives

Keywords:generalized reparameterization gradient, variance reduction, non-reparameterizable, discrete random variable, GO gradient, general and one-sample gradient, expectation-based objective, variable nabla, statistical back-propagation, hierarchical, graphical model
TL;DR:a Rep-like gradient for non-reparameterizable continuous/discrete distributions; further generalized to deep probabilistic models, yielding statistical back-propagation

h-detach: Modifying the LSTM Gradient Towards Better Optimization

Keywords:LSTM, Optimization, Long term dependencies, Back-propagation through time
TL;DR:A simple algorithm to improve optimization and handling of long term dependencies in LSTM

An analytic theory of generalization dynamics and transfer learning in deep linear networks

Keywords:Generalization, Theory, Transfer, Multi-task, Linear
TL;DR:We provide many insights into neural network generalization from the theoretically tractable linear case.

Differentiable Learning-to-Normalize via Switchable Normalization

Keywords:None
TL;DR:None

SOM-VAE: Interpretable Discrete Representation Learning on Time Series

Keywords:deep learning, self-organizing map, variational autoencoder, representation learning, time series, machine learning, interpretability
TL;DR:We present a method to learn interpretable representations on time series using ideas from variational autoencoders, self-organizing maps and probabilistic models.

Hierarchical Generative Modeling for Controllable Speech Synthesis

Keywords:speech synthesis, representation learning, deep generative model, sequence-to-sequence model
TL;DR:Building a TTS model with Gaussian Mixture VAEs enables fine-grained control of speaking style, noise condition, and more.

Learning Factorized Multimodal Representations

Keywords:multimodal learning, representation learning
TL;DR:We propose a model to learn factorized multimodal representations that are discriminative, generative, and interpretable.

Composing Complex Skills by Learning Transition Policies

Keywords:reinforcement learning, hierarchical reinforcement learning, continuous control, modular framework
TL;DR:Transition policies enable agents to compose complex skills by smoothly connecting previously acquired primitive skills.

Human-level Protein Localization with Convolutional Neural Networks

Keywords:None
TL;DR:None

Environment Probing Interaction Policies

Keywords:None
TL;DR:None

Lagging Inference Networks and Posterior Collapse in Variational Autoencoders

Keywords:variational autoencoders, posterior collapse, generative models
TL;DR:To address posterior collapse in VAEs, we propose a novel yet simple training procedure that aggressively optimizes inference network with more updates. This new training procedure mitigates posterior collapse and leads to a better VAE model.

A2BCD: Asynchronous Acceleration with Optimal Complexity

Keywords:asynchronous, optimization, parallel, accelerated, complexity
TL;DR:We prove the first-ever convergence proof of an asynchronous accelerated algorithm that attains a speedup.

Learning to Infer and Execute 3D Shape Programs

Keywords:Program Synthesis, 3D Shape Modeling, Self-supervised Learning
TL;DR:We propose 3D shape programs, a structured, compositional shape representation. Our model learns to infer and execute shape programs to explain 3D shapes.

Deep Decoder: Concise Image Representations from Untrained Non-convolutional Networks

Keywords:natural image model, image prior, under-determined neural networks, untrained network, non-convolutional network, denoising, inverse problem
TL;DR:We introduce an underparameterized, nonconvolutional, and simple deep neural network that can, without training, effectively represent natural images and solve image processing tasks like compression and denoising competitively.

SNAS: stochastic neural architecture search

Keywords:None
TL;DR:None

Revealing interpretable object representations from human behavior

Keywords:category representation, sparse coding, representation learning, interpretable representations
TL;DR:Human behavioral judgments are used to obtain sparse and interpretable representations of objects that generalize to other tasks

AntisymmetricRNN: A Dynamical System View on Recurrent Neural Networks

Keywords:None
TL;DR:None

Global-to-local Memory Pointer Networks for Task-Oriented Dialogue

Keywords:pointer networks, memory networks, task-oriented dialogue systems, natural language processing
TL;DR:GLMP: Global memory encoder (context RNN, global pointer) and local memory decoder (sketch RNN, local pointer) that share external knowledge (MemNN) are proposed to strengthen response generation in task-oriented dialogue.

InstaGAN: Instance-aware Image-to-Image Translation

Keywords:Image-to-Image Translation, Generative Adversarial Networks
TL;DR:We propose a novel method to incorporate the set of instance attributes for image-to-image translation.

Deep Layers as Stochastic Solvers

Keywords:deep networks, optimization
TL;DR:A framework that links deep network layers to stochastic optimization algorithms; can be used to improve model accuracy and inform network design.

Learning Multi-Level Hierarchies with Hindsight

Keywords:Hierarchical Reinforcement Learning, Reinforcement Learning, Deep Reinforcement Learning
TL;DR:We introduce the first Hierarchical RL approach to successfully learn 3-level hierarchies in parallel in tasks with continuous state and action spaces.