title
stringlengths 19
143
| url
stringlengths 41
43
| detail_url
stringlengths 41
43
| authors
stringlengths 9
347
| tags
stringclasses 3
values | abstract
stringlengths 457
2.38k
| pdf
stringlengths 71
71
|
---|---|---|---|---|---|---|
A Critique of Self-Expressive Deep Subspace Clustering | https://openreview.net/forum?id=FOyuZ26emy | https://openreview.net/forum?id=FOyuZ26emy | Benjamin David Haeffele,Chong You,Rene Vidal | ICLR 2021,Poster | Subspace clustering is an unsupervised clustering technique designed to cluster data that is supported on a union of linear subspaces, with each subspace defining a cluster with dimension lower than the ambient space. Many existing formulations for this problem are based on exploiting the self-expressive property of linear subspaces, where any point within a subspace can be represented as linear combination of other points within the subspace. To extend this approach to data supported on a union of non-linear manifolds, numerous studies have proposed learning an embedding of the original data using a neural network which is regularized by a self-expressive loss function on the data in the embedded space to encourage a union of linear subspaces prior on the data in the embedded space. Here we show that there are a number of potential flaws with this approach which have not been adequately addressed in prior work. In particular, we show the model formulation is often ill-posed in that it can lead to a degenerate embedding of the data, which need not correspond to a union of subspaces at all and is poorly suited for clustering. We validate our theoretical results experimentally and also repeat prior experiments reported in the literature, where we conclude that a significant portion of the previously claimed performance benefits can be attributed to an ad-hoc post processing step rather than the deep subspace clustering model. | https://openreview.net/pdf/b695594106c11bb76d5334e21cd6ae28728ec857.pdf |
Explaining by Imitating: Understanding Decisions by Interpretable Policy Learning | https://openreview.net/forum?id=unI5ucw_Jk | https://openreview.net/forum?id=unI5ucw_Jk | Alihan Hüyük,Daniel Jarrett,Cem Tekin,Mihaela van der Schaar | ICLR 2021,Poster | Understanding human behavior from observed data is critical for transparency and accountability in decision-making. Consider real-world settings such as healthcare, in which modeling a decision-maker’s policy is challenging—with no access to underlying states, no knowledge of environment dynamics, and no allowance for live experimentation. We desire learning a data-driven representation of decision- making behavior that (1) inheres transparency by design, (2) accommodates partial observability, and (3) operates completely offline. To satisfy these key criteria, we propose a novel model-based Bayesian method for interpretable policy learning (“Interpole”) that jointly estimates an agent’s (possibly biased) belief-update process together with their (possibly suboptimal) belief-action mapping. Through experiments on both simulated and real-world data for the problem of Alzheimer’s disease diagnosis, we illustrate the potential of our approach as an investigative device for auditing, quantifying, and understanding human decision-making behavior. | https://openreview.net/pdf/87d4ea3176fd5ca710f4f3d549c2f1c25b5898cf.pdf |
For self-supervised learning, Rationality implies generalization, provably | https://openreview.net/forum?id=Srmggo3b3X6 | https://openreview.net/forum?id=Srmggo3b3X6 | Yamini Bansal,Gal Kaplun,Boaz Barak | ICLR 2021,Poster | We prove a new upper bound on the generalization gap of classifiers that are obtained by first using self-supervision to learn a representation $r$ of the training~data, and then fitting a simple (e.g., linear) classifier $g$ to the labels. Specifically, we show that (under the assumptions described below) the generalization gap of such classifiers tends to zero if $\mathsf{C}(g) \ll n$, where $\mathsf{C}(g)$ is an appropriately-defined measure of the simple classifier $g$'s complexity, and $n$ is the number of training samples. We stress that our bound is independent of the complexity of the representation $r$.
We do not make any structural or conditional-independence assumptions on the representation-learning task, which can use the same training dataset that is later used for classification. Rather, we assume that the training procedure satisfies certain natural noise-robustness (adding small amount of label noise causes small degradation in performance) and rationality (getting the wrong label is not better than getting no label at all) conditions that widely hold across many standard architectures.
We also conduct an extensive empirical study of the generalization gap and the quantities used in our assumptions for a variety of self-supervision based algorithms, including SimCLR, AMDIM and BigBiGAN, on the CIFAR-10 and ImageNet datasets. We show that, unlike standard supervised classifiers, these algorithms display small generalization gap, and the bounds we prove on this gap are often non vacuous. | https://openreview.net/pdf/111c52ee7a399d9253f0dc005866fa6a4b705dc7.pdf |
Directed Acyclic Graph Neural Networks | https://openreview.net/forum?id=JbuYF437WB6 | https://openreview.net/forum?id=JbuYF437WB6 | Veronika Thost,Jie Chen | ICLR 2021,Poster | Graph-structured data ubiquitously appears in science and engineering. Graph neural networks (GNNs) are designed to exploit the relational inductive bias exhibited in graphs; they have been shown to outperform other forms of neural networks in scenarios where structure information supplements node features. The most common GNN architecture aggregates information from neighborhoods based on message passing. Its generality has made it broadly applicable. In this paper, we focus on a special, yet widely used, type of graphs---DAGs---and inject a stronger inductive bias---partial ordering---into the neural network design. We propose the directed acyclic graph neural network, DAGNN, an architecture that processes information according to the flow defined by the partial order. DAGNN can be considered a framework that entails earlier works as special cases (e.g., models for trees and models updating node representations recurrently), but we identify several crucial components that prior architectures lack. We perform comprehensive experiments, including ablation studies, on representative DAG datasets (i.e., source code, neural architectures, and probabilistic graphical models) and demonstrate the superiority of DAGNN over simpler DAG architectures as well as general graph architectures. | https://openreview.net/pdf/435a89f1a5c579c1c5fe7ba3ebef81c09224e075.pdf |
My Body is a Cage: the Role of Morphology in Graph-Based Incompatible Control | https://openreview.net/forum?id=N3zUDGN5lO | https://openreview.net/forum?id=N3zUDGN5lO | Vitaly Kurin,Maximilian Igl,Tim Rocktäschel,Wendelin Boehmer,Shimon Whiteson | ICLR 2021,Poster | Multitask Reinforcement Learning is a promising way to obtain models with better performance, generalisation, data efficiency, and robustness. Most existing work is limited to compatible settings, where the state and action space dimensions are the same across tasks. Graph Neural Networks (GNN) are one way to address incompatible environments, because they can process graphs of arbitrary size. They also allow practitioners to inject biases encoded in the structure of the input graph. Existing work in graph-based continuous control uses the physical morphology of the agent to construct the input graph, i.e., encoding limb features as node labels and using edges to connect the nodes if their corresponded limbs are physically connected.
In this work, we present a series of ablations on existing methods that show that morphological information encoded in the graph does not improve their performance. Motivated by the hypothesis that any benefits GNNs extract from the graph structure are outweighed by difficulties they create for message passing, we also propose Amorpheus, a transformer-based approach. Further results show that, while Amorpheus ignores the morphological information that GNNs encode, it nonetheless substantially outperforms GNN-based methods. | https://openreview.net/pdf/130809fe4dd2abe36b6f0c395f8cc2f51174bc4d.pdf |
Incremental few-shot learning via vector quantization in deep embedded space | https://openreview.net/forum?id=3SV-ZePhnZM | https://openreview.net/forum?id=3SV-ZePhnZM | Kuilin Chen,Chi-Guhn Lee | ICLR 2021,Poster | The capability of incrementally learning new tasks without forgetting old ones is a challenging problem due to catastrophic forgetting. This challenge becomes greater when novel tasks contain very few labelled training samples. Currently, most methods are dedicated to class-incremental learning and rely on sufficient training data to learn additional weights for newly added classes. Those methods cannot be easily extended to incremental regression tasks and could suffer from severe overfitting when learning few-shot novel tasks. In this study, we propose a nonparametric method in deep embedded space to tackle incremental few-shot learning problems. The knowledge about the learned tasks are compressed into a small number of quantized reference vectors. The proposed method learns new tasks sequentially by adding more reference vectors to the model using few-shot samples in each novel task. For classification problems, we employ the nearest neighbor scheme to make classification on sparsely available data and incorporate intra-class variation, less forgetting regularization and calibration of reference vectors to mitigate catastrophic forgetting. In addition, the proposed learning vector quantization (LVQ) in deep embedded space can be customized as a kernel smoother to handle incremental few-shot regression tasks. Experimental results demonstrate that the proposed method outperforms other state-of-the-art methods in incremental learning. | https://openreview.net/pdf/e34aea7744f0d7fa06719626e37f52420d1dc247.pdf |
Revisiting Few-sample BERT Fine-tuning | https://openreview.net/forum?id=cO1IH43yUF | https://openreview.net/forum?id=cO1IH43yUF | Tianyi Zhang,Felix Wu,Arzoo Katiyar,Kilian Q Weinberger,Yoav Artzi | ICLR 2021,Poster | This paper is a study of fine-tuning of BERT contextual representations, with focus on commonly observed instabilities in few-sample scenarios. We identify several factors that cause this instability: the common use of a non-standard optimization method with biased gradient estimation; the limited applicability of significant parts of the BERT network for down-stream tasks; and the prevalent practice of using a pre-determined, and small number of training iterations. We empirically test the impact of these factors, and identify alternative practices that resolve the commonly observed instability of the process. In light of these observations, we re-visit recently proposed methods to improve few-sample fine-tuning with BERT and re-evaluate their effectiveness. Generally, we observe the impact of these methods diminishes significantly with our modified process. | https://openreview.net/pdf/b9891ff9bbfe3c50cf752711eb45ea789cd534aa.pdf |
Linear Convergent Decentralized Optimization with Compression | https://openreview.net/forum?id=84gjULz1t5 | https://openreview.net/forum?id=84gjULz1t5 | Xiaorui Liu,Yao Li,Rongrong Wang,Jiliang Tang,Ming Yan | ICLR 2021,Poster | Communication compression has become a key strategy to speed up distributed optimization. However, existing decentralized algorithms with compression mainly focus on compressing DGD-type algorithms. They are unsatisfactory in terms of convergence rate, stability, and the capability to handle heterogeneous data. Motivated by primal-dual algorithms, this paper proposes the first \underline{L}in\underline{EA}r convergent \underline{D}ecentralized algorithm with compression, LEAD. Our theory describes the coupled dynamics of the inexact primal and dual update as well as compression error, and we provide the first consensus error bound in such settings without assuming bounded gradients. Experiments on convex problems validate our theoretical analysis, and empirical study on deep neural nets shows that LEAD is applicable to non-convex problems. | https://openreview.net/pdf/2c18c598153405d899eb172db4db2e112b25c66f.pdf |
Learning from Demonstration with Weakly Supervised Disentanglement | https://openreview.net/forum?id=Ldau9eHU-qO | https://openreview.net/forum?id=Ldau9eHU-qO | Yordan Hristov,Subramanian Ramamoorthy | ICLR 2021,Poster | Robotic manipulation tasks, such as wiping with a soft sponge, require control from multiple rich sensory modalities. Human-robot interaction, aimed at teach- ing robots, is difficult in this setting as there is potential for mismatch between human and machine comprehension of the rich data streams. We treat the task of interpretable learning from demonstration as an optimisation problem over a probabilistic generative model. To account for the high-dimensionality of the data, a high-capacity neural network is chosen to represent the model. The latent variables in this model are explicitly aligned with high-level notions and concepts that are manifested in a set of demonstrations. We show that such alignment is best achieved through the use of labels from the end user, in an appropriately restricted vocabulary, in contrast to the conventional approach of the designer picking a prior over the latent variables. Our approach is evaluated in the context of two table-top robot manipulation tasks performed by a PR2 robot – that of dabbing liquids with a sponge (forcefully pressing a sponge and moving it along a surface) and pouring between different containers. The robot provides visual information, arm joint positions and arm joint efforts. We have made videos of the tasks and data available - see supplementary materials at: https://sites.google.com/view/weak-label-lfd. | https://openreview.net/pdf/49356704899c9011f84ed5a7f4cb44bbe43ae737.pdf |
Incorporating Symmetry into Deep Dynamics Models for Improved Generalization | https://openreview.net/forum?id=wta_8Hx2KD | https://openreview.net/forum?id=wta_8Hx2KD | Rui Wang,Robin Walters,Rose Yu | ICLR 2021,Poster | Recent work has shown deep learning can accelerate the prediction of physical dynamics relative to numerical solvers. However, limited physical accuracy and an inability to generalize under distributional shift limit its applicability to the real world. We propose to improve accuracy and generalization by incorporating symmetries into convolutional neural networks. Specifically, we employ a variety of methods each tailored to enforce a different symmetry. Our models are both theoretically and experimentally robust to distributional shift by symmetry group transformations and enjoy favorable sample complexity. We demonstrate the advantage of our approach on a variety of physical dynamics including Rayleigh–Bénard convection and real-world ocean currents and temperatures. Compare with image or text applications, our work is a significant step towards applying equivariant neural networks to high-dimensional systems with complex dynamics. | https://openreview.net/pdf/0baeb937c065c6c783ae239c7722374340b6497e.pdf |
The Risks of Invariant Risk Minimization | https://openreview.net/forum?id=BbNIbVPJ-42 | https://openreview.net/forum?id=BbNIbVPJ-42 | Elan Rosenfeld,Pradeep Kumar Ravikumar,Andrej Risteski | ICLR 2021,Poster | Invariant Causal Prediction (Peters et al., 2016) is a technique for out-of-distribution generalization which assumes that some aspects of the data distribution vary across the training set but that the underlying causal mechanisms remain constant. Recently, Arjovsky et al. (2019) proposed Invariant Risk Minimization (IRM), an objective based on this idea for learning deep, invariant features of data which are a complex function of latent variables; many alternatives have subsequently been suggested. However, formal guarantees for all of these works are severely lacking. In this paper, we present the first analysis of classification under the IRM objective—as well as these recently proposed alternatives—under a fairly natural and general model. In the linear case, we show simple conditions under which the optimal solution succeeds or, more often, fails to recover the optimal invariant predictor. We furthermore present the very first results in the non-linear regime: we demonstrate that IRM can fail catastrophically unless the test data is sufficiently similar to the training distribution—this is precisely the issue that it was intended to solve. Thus, in this setting we find that IRM and its alternatives fundamentally do not improve over standard Empirical Risk Minimization. | https://openreview.net/pdf/28dd76f69ec8c36b1e5610943745438b4d6ca498.pdf |
Scaling Symbolic Methods using Gradients for Neural Model Explanation | https://openreview.net/forum?id=V5j-jdoDDP | https://openreview.net/forum?id=V5j-jdoDDP | Subham Sekhar Sahoo,Subhashini Venugopalan,Li Li,Rishabh Singh,Patrick Riley | ICLR 2021,Poster | Symbolic techniques based on Satisfiability Modulo Theory (SMT) solvers have been proposed for analyzing and verifying neural network properties, but their usage has been fairly limited owing to their poor scalability with larger networks. In this work, we propose a technique for combining gradient-based methods with symbolic techniques to scale such analyses and demonstrate its application for model explanation. In particular, we apply this technique to identify minimal regions in an input that are most relevant for a neural network's prediction. Our approach uses gradient information (based on Integrated Gradients) to focus on a subset of neurons in the first layer, which allows our technique to scale to large networks. The corresponding SMT constraints encode the minimal input mask discovery problem such that after masking the input, the activations of the selected neurons are still above a threshold. After solving for the minimal masks, our approach scores the mask regions to generate a relative ordering of the features within the mask. This produces a saliency map which explains" where a model is looking" when making a prediction. We evaluate our technique on three datasets-MNIST, ImageNet, and Beer Reviews, and demonstrate both quantitatively and qualitatively that the regions generated by our approach are sparser and achieve higher saliency scores compared to the gradient-based methods alone. Code and examples are at - https://github.com/google-research/google-research/tree/master/smug_saliency | https://openreview.net/pdf/d1b9da67f960e7d1f4843fe8b37b4e9cb6cbbc3b.pdf |
Contextual Dropout: An Efficient Sample-Dependent Dropout Module | https://openreview.net/forum?id=ct8_a9h1M | https://openreview.net/forum?id=ct8_a9h1M | XINJIE FAN,Shujian Zhang,Korawat Tanwisuth,Xiaoning Qian,Mingyuan Zhou | ICLR 2021,Poster | Dropout has been demonstrated as a simple and effective module to not only regularize the training process of deep neural networks, but also provide the uncertainty estimation for prediction. However, the quality of uncertainty estimation is highly dependent on the dropout probabilities. Most current models use the same dropout distributions across all data samples due to its simplicity. Despite the potential gains in the flexibility of modeling uncertainty, sample-dependent dropout, on the other hand, is less explored as it often encounters scalability issues or involves non-trivial model changes. In this paper, we propose contextual dropout with an efficient structural design as a simple and scalable sample-dependent dropout module, which can be applied to a wide range of models at the expense of only slightly increased memory and computational cost. We learn the dropout probabilities with a variational objective, compatible with both Bernoulli dropout and Gaussian dropout. We apply the contextual dropout module to various models with applications to image classification and visual question answering and demonstrate the scalability of the method with large-scale datasets, such as ImageNet and VQA 2.0. Our experimental results show that the proposed method outperforms baseline methods in terms of both accuracy and quality of uncertainty estimation. | https://openreview.net/pdf/e60af0056960e7c5e82ee3d15ad9c6dd3577788d.pdf |
Contrastive Learning with Hard Negative Samples | https://openreview.net/forum?id=CR1XOQ0UTh- | https://openreview.net/forum?id=CR1XOQ0UTh- | Joshua David Robinson,Ching-Yao Chuang,Suvrit Sra,Stefanie Jegelka | ICLR 2021,Poster | We consider the question: how can you sample good negative examples for contrastive learning? We argue that, as with metric learning, learning contrastive representations benefits from hard negative samples (i.e., points that are difficult to distinguish from an anchor point). The key challenge toward using hard negatives is that contrastive methods must remain unsupervised, making it infeasible to adopt existing negative sampling strategies that use label information. In response, we develop a new class of unsupervised methods for selecting hard negative samples where the user can control the amount of hardness. A limiting case of this sampling results in a representation that tightly clusters each class, and pushes different classes as far apart as possible. The proposed method improves downstream performance across multiple modalities, requires only few additional lines of code to implement, and introduces no computational overhead.
| https://openreview.net/pdf/abbdb606c8b0133d5a3f1cb9b239cb75f5530c20.pdf |
Debiasing Concept-based Explanations with Causal Analysis | https://openreview.net/forum?id=6puUoArESGp | https://openreview.net/forum?id=6puUoArESGp | Mohammad Taha Bahadori,David Heckerman | ICLR 2021,Poster | Concept-based explanation approach is a popular model interpertability tool because it expresses the reasons for a model's predictions in terms of concepts that are meaningful for the domain experts. In this work, we study the problem of the concepts being correlated with confounding information in the features. We propose a new causal prior graph for modeling the impacts of unobserved variables and a method to remove the impact of confounding information and noise using a two-stage regression technique borrowed from the instrumental variable literature. We also model the completeness of the concepts set and show that our debiasing method works when the concepts are not complete. Our synthetic and real-world experiments demonstrate the success of our method in removing biases and improving the ranking of the concepts in terms of their contribution to the explanation of the predictions. | https://openreview.net/pdf/6c474aec6a470f6431f665cd26196adb69dee2cd.pdf |
Risk-Averse Offline Reinforcement Learning | https://openreview.net/forum?id=TBIzh9b5eaz | https://openreview.net/forum?id=TBIzh9b5eaz | Núria Armengol Urpí,Sebastian Curi,Andreas Krause | ICLR 2021,Poster | Training Reinforcement Learning (RL) agents in high-stakes applications might be too prohibitive due to the risk associated to exploration. Thus, the agent can only use data previously collected by safe policies. While previous work considers optimizing the average performance using offline data, we focus on optimizing a risk-averse criteria, namely the CVaR. In particular, we present the Offline Risk-Averse Actor-Critic (O-RAAC), a model-free RL algorithm that is able to learn risk-averse policies in a fully offline setting. We show that O-RAAC learns policies with higher CVaR than risk-neutral approaches in different robot control tasks. Furthermore, considering risk-averse criteria guarantees distributional robustness of the average performance with respect to particular distribution shifts. We demonstrate empirically that in the presence of natural distribution-shifts, O-RAAC learns policies with good average performance.
| https://openreview.net/pdf/be66503efdccf10359c7112b8c3732b28564db16.pdf |
Parameter Efficient Multimodal Transformers for Video Representation Learning | https://openreview.net/forum?id=6UdQLhqJyFD | https://openreview.net/forum?id=6UdQLhqJyFD | Sangho Lee,Youngjae Yu,Gunhee Kim,Thomas Breuel,Jan Kautz,Yale Song | ICLR 2021,Poster | The recent success of Transformers in the language domain has motivated adapting it to a multimodal setting, where a new visual model is trained in tandem with an already pretrained language model. However, due to the excessive memory requirements from Transformers, existing work typically fixes the language model and train only the vision module, which limits its ability to learn cross-modal information in an end-to-end manner. In this work, we focus on reducing the parameters of multimodal Transformers in the context of audio-visual video representation learning. We alleviate the high memory requirement by sharing the parameters of Transformers across layers and modalities; we decompose the Transformer into modality-specific and modality-shared parts so that the model learns the dynamics of each modality both individually and together, and propose a novel parameter sharing scheme based on low-rank approximation. We show that our approach reduces parameters of the Transformers up to 97%, allowing us to train our model end-to-end from scratch. We also propose a negative sampling approach based on an instance similarity measured on the CNN embedding space that our model learns together with the Transformers. To demonstrate our approach, we pretrain our model on 30-second clips (480 frames) from Kinetics-700 and transfer it to audio-visual classification tasks. | https://openreview.net/pdf/2cccd8332d73cd6ac7e33027594d30f19af464d5.pdf |
Tradeoffs in Data Augmentation: An Empirical Study | https://openreview.net/forum?id=ZcKPWuhG6wy | https://openreview.net/forum?id=ZcKPWuhG6wy | Raphael Gontijo-Lopes,Sylvia Smullin,Ekin Dogus Cubuk,Ethan Dyer | ICLR 2021,Poster | Though data augmentation has become a standard component of deep neural network training, the underlying mechanism behind the effectiveness of these techniques remains poorly understood. In practice, augmentation policies are often chosen using heuristics of distribution shift or augmentation diversity. Inspired by these, we conduct an empirical study to quantify how data augmentation improves model generalization. We introduce two interpretable and easy-to-compute measures: Affinity and Diversity. We find that augmentation performance is predicted not by either of these alone but by jointly optimizing the two. | https://openreview.net/pdf/3d2fc5aa6c81e18581fdf14b971478263093ec81.pdf |
Concept Learners for Few-Shot Learning | https://openreview.net/forum?id=eJIJF3-LoZO | https://openreview.net/forum?id=eJIJF3-LoZO | Kaidi Cao,Maria Brbic,Jure Leskovec | ICLR 2021,Poster | Developing algorithms that are able to generalize to a novel task given only a few labeled examples represents a fundamental challenge in closing the gap between machine- and human-level performance. The core of human cognition lies in the structured, reusable concepts that help us to rapidly adapt to new tasks and provide reasoning behind our decisions. However, existing meta-learning methods learn complex representations across prior labeled tasks without imposing any structure on the learned representations. Here we propose COMET, a meta-learning method that improves generalization ability by learning to learn along human-interpretable concept dimensions. Instead of learning a joint unstructured metric space, COMET learns mappings of high-level concepts into semi-structured metric spaces, and effectively combines the outputs of independent concept learners. We evaluate our model on few-shot tasks from diverse domains, including fine-grained image classification, document categorization and cell type annotation on a novel dataset from a biological domain developed in our work. COMET significantly outperforms strong meta-learning baselines, achieving 6-15% relative improvement on the most challenging 1-shot learning tasks, while unlike existing methods providing interpretations behind the model's predictions. | https://openreview.net/pdf/4502f92e9d047dea9aa42deceb3ff6b44a74942e.pdf |
Neural Mechanics: Symmetry and Broken Conservation Laws in Deep Learning Dynamics | https://openreview.net/forum?id=q8qLAbQBupm | https://openreview.net/forum?id=q8qLAbQBupm | Daniel Kunin,Javier Sagastuy-Brena,Surya Ganguli,Daniel LK Yamins,Hidenori Tanaka | ICLR 2021,Poster | Understanding the dynamics of neural network parameters during training is one of the key challenges in building a theoretical foundation for deep learning. A central obstacle is that the motion of a network in high-dimensional parameter space undergoes discrete finite steps along complex stochastic gradients derived from real-world datasets. We circumvent this obstacle through a unifying theoretical framework based on intrinsic symmetries embedded in a network's architecture that are present for any dataset. We show that any such symmetry imposes stringent geometric constraints on gradients and Hessians, leading to an associated conservation law in the continuous-time limit of stochastic gradient descent (SGD), akin to Noether's theorem in physics. We further show that finite learning rates used in practice can actually break these symmetry induced conservation laws. We apply tools from finite difference methods to derive modified gradient flow, a differential equation that better approximates the numerical trajectory taken by SGD at finite learning rates. We combine modified gradient flow with our framework of symmetries to derive exact integral expressions for the dynamics of certain parameter combinations. We empirically validate our analytic expressions for learning dynamics on VGG-16 trained on Tiny ImageNet. Overall, by exploiting symmetry, our work demonstrates that we can analytically describe the learning dynamics of various parameter combinations at finite learning rates and batch sizes for state of the art architectures trained on any dataset. | https://openreview.net/pdf/5abc5ac75318bd0ee0b7d7d6a805c51471a672e0.pdf |
Fair Mixup: Fairness via Interpolation | https://openreview.net/forum?id=DNl5s5BXeBn | https://openreview.net/forum?id=DNl5s5BXeBn | Ching-Yao Chuang,Youssef Mroueh | ICLR 2021,Poster | Training classifiers under fairness constraints such as group fairness, regularizes the disparities of predictions between the groups. Nevertheless, even though the constraints are satisfied during training, they might not generalize at evaluation time. To improve the generalizability of fair classifiers, we propose fair mixup, a new data augmentation strategy for imposing the fairness constraint. In particular, we show that fairness can be achieved by regularizing the models on paths of interpolated samples between the groups. We use mixup, a powerful data augmentation strategy to generate these interpolates. We analyze fair mixup and empirically show that it ensures a better generalization for both accuracy and fairness measurement in tabular, vision, and language benchmarks. | https://openreview.net/pdf/26fcabb2173a9b48339615540d74ed1bc1c6281e.pdf |
Heating up decision boundaries: isocapacitory saturation, adversarial scenarios and generalization bounds | https://openreview.net/forum?id=UwGY2qjqoLD | https://openreview.net/forum?id=UwGY2qjqoLD | Bogdan Georgiev,Lukas Franken,Mayukh Mukherjee | ICLR 2021,Poster | In the present work we study classifiers' decision boundaries via Brownian motion processes in ambient data space and associated probabilistic techniques. Intuitively, our ideas correspond to placing a heat source at the decision boundary and observing how effectively the sample points warm up. We are largely motivated by the search for a soft measure that sheds further light on the decision boundary's geometry. En route, we bridge aspects of potential theory and geometric analysis (Maz'ya 2011, Grigor'Yan and Saloff-Coste 2002) with active fields of ML research such as adversarial examples and generalization bounds. First, we focus on the geometric behavior of decision boundaries in the light of adversarial attack/defense mechanisms. Experimentally, we observe a certain capacitory trend over different adversarial defense strategies: decision boundaries locally become flatter as measured by isoperimetric inequalities (Ford et al 2019); however, our more sensitive heat-diffusion metrics extend this analysis and further reveal that some non-trivial geometry invisible to plain distance-based methods is still preserved. Intuitively, we provide evidence that the decision boundaries nevertheless retain many persistent "wiggly and fuzzy" regions on a finer scale.
Second, we show how Brownian hitting probabilities translate to soft generalization bounds which are in turn connected to compression and noise stability (Arora et al 2018), and these bounds are significantly stronger if the decision boundary has controlled geometric features. | https://openreview.net/pdf/e8a429ae487de8dda95c91d6c5401966de30ba95.pdf |
Influence Functions in Deep Learning Are Fragile | https://openreview.net/forum?id=xHKVVHGDOEk | https://openreview.net/forum?id=xHKVVHGDOEk | Samyadeep Basu,Phil Pope,Soheil Feizi | ICLR 2021,Poster | Influence functions approximate the effect of training samples in test-time predictions and have a wide variety of applications in machine learning interpretability and uncertainty estimation. A commonly-used (first-order) influence function can be implemented efficiently as a post-hoc method requiring access only to the gradients and Hessian of the model. For linear models, influence functions are well-defined due to the convexity of the underlying loss function and are generally accurate even across difficult settings where model changes are fairly large such as estimating group influences. Influence functions, however, are not well-understood in the context of deep learning with non-convex loss functions. In this paper, we provide a comprehensive and large-scale empirical study of successes and failures of influence functions in neural network models trained on datasets such as Iris, MNIST, CIFAR-10 and ImageNet. Through our extensive experiments, we show that the network architecture, its depth and width, as well as the extent of model parameterization and regularization techniques have strong effects in the accuracy of influence functions. In particular, we find that (i) influence estimates are fairly accurate for shallow networks, while for deeper networks the estimates are often erroneous; (ii) for certain network architectures and datasets, training with weight-decay regularization is important to get high-quality influence estimates; and (iii) the accuracy of influence estimates can vary significantly depending on the examined test points. These results suggest that in general influence functions in deep learning are fragile and call for developing improved influence estimation methods to mitigate these issues in non-convex setups. | https://openreview.net/pdf/5f4c0d6ade226a87db8c80975d4487e28ec11d9e.pdf |
Few-Shot Learning via Learning the Representation, Provably | https://openreview.net/forum?id=pW2Q2xLwIMD | https://openreview.net/forum?id=pW2Q2xLwIMD | Simon Shaolei Du,Wei Hu,Sham M. Kakade,Jason D. Lee,Qi Lei | ICLR 2021,Poster | This paper studies few-shot learning via representation learning, where one uses $T$ source tasks with $n_1$ data per task to learn a representation in order to reduce the sample complexity of a target task for which there is only $n_2 (\ll n_1)$ data. Specifically, we focus on the setting where there exists a good common representation between source and target, and our goal is to understand how much a sample size reduction is possible. First, we study the setting where this common representation is low-dimensional and provide a risk bound of $\tilde{O}(\frac{dk}{n_1T} + \frac{k}{n_2})$ on the target task for the linear representation class; here $d$ is the ambient input dimension and $k (\ll d)$ is the dimension of the representation. This result bypasses the $\Omega(\frac{1}{T})$ barrier under the i.i.d. task assumption, and can capture the desired property that all $n_1T$ samples from source tasks can be \emph{pooled} together for representation learning. We further extend this result to handle a general representation function class and obtain a similar result. Next, we consider the setting where the common representation may be high-dimensional but is capacity-constrained (say in norm); here, we again demonstrate the advantage of representation learning in both high-dimensional linear regression and neural networks, and show that representation learning can fully utilize all $n_1T$ samples from source tasks. | https://openreview.net/pdf/d552a527585afe0600e439411a5e9152ce459182.pdf |
Self-Supervised Learning of Compressed Video Representations | https://openreview.net/forum?id=jMPcEkJpdD | https://openreview.net/forum?id=jMPcEkJpdD | Youngjae Yu,Sangho Lee,Gunhee Kim,Yale Song | ICLR 2021,Poster | Self-supervised learning of video representations has received great attention. Existing methods typically require frames to be decoded before being processed, which increases compute and storage requirements and ultimately hinders large-scale training. In this work, we propose an efficient self-supervised approach to learn video representations by eliminating the expensive decoding step. We use a three-stream video architecture that encodes I-frames and P-frames of a compressed video. Unlike existing approaches that encode I-frames and P-frames individually, we propose to jointly encode them by establishing bidirectional dynamic connections across streams. To enable self-supervised learning, we propose two pretext tasks that leverage the multimodal nature (RGB, motion vector, residuals) and the internal GOP structure of compressed videos. The first task asks our network to predict zeroth-order motion statistics in a spatio-temporal pyramid; the second task asks correspondence types between I-frames and P-frames after applying temporal transformations. We show that our approach achieves competitive performance on compressed video recognition both in supervised and self-supervised regimes.
| https://openreview.net/pdf/32a283d43ef9e69a3039580759776862664a9146.pdf |
Flowtron: an Autoregressive Flow-based Generative Network for Text-to-Speech Synthesis | https://openreview.net/forum?id=Ig53hpHxS4 | https://openreview.net/forum?id=Ig53hpHxS4 | Rafael Valle,Kevin J. Shih,Ryan Prenger,Bryan Catanzaro | ICLR 2021,Poster | In this paper we propose Flowtron: an autoregressive flow-based generative network for text-to-speech synthesis with style transfer and speech variation. Flowtron borrows insights from Autoregressive Flows and revamps Tacotron 2 in order to provide high-quality and expressive mel-spectrogram synthesis. Flowtron is optimized by maximizing the likelihood of the training data, which makes training simple and stable. Flowtron learns an invertible mapping of data to a latent space that can be used to modulate many aspects of speech synthesis (timbre, expressivity, accent). Our mean opinion scores (MOS) show that Flowtron matches state-of-the-art TTS models in terms of speech quality. We provide results on speech variation, interpolation over time between samples and style transfer between seen and unseen speakers. Code and pre-trained models are publicly available at \href{https://github.com/NVIDIA/flowtron}{https://github.com/NVIDIA/flowtron}. | https://openreview.net/pdf/f022be71a045af955f0b061b9fb50186c102f407.pdf |
R-GAP: Recursive Gradient Attack on Privacy | https://openreview.net/forum?id=RSU17UoKfJF | https://openreview.net/forum?id=RSU17UoKfJF | Junyi Zhu,Matthew B. Blaschko | ICLR 2021,Poster | Federated learning frameworks have been regarded as a promising approach to break the dilemma between demands on privacy and the promise of learning from large collections of distributed data. Many such frameworks only ask collaborators to share their local update of a common model, i.e. gradients with respect to locally stored data, instead of exposing their raw data to other collaborators. However, recent optimization-based gradient attacks show that raw data can often be accurately recovered from gradients. It has been shown that minimizing the Euclidean distance between true gradients and those calculated from estimated data is often effective in fully recovering private data. However, there is a fundamental lack of theoretical understanding of how and when gradients can lead to unique recovery of original data. Our research fills this gap by providing a closed-form recursive procedure to recover data from gradients in deep neural networks. We name it Recursive Gradient Attack on Privacy (R-GAP). Experimental results demonstrate that R-GAP works as well as or even better than optimization-based approaches at a fraction of the computation under certain conditions. Additionally, we propose a Rank Analysis method, which can be used to estimate the risk of gradient attacks inherent in certain network architectures, regardless of whether an optimization-based or closed-form-recursive attack is used. Experimental results demonstrate the utility of the rank analysis towards improving the network's security. Source code is available for download from https://github.com/JunyiZhu-AI/R-GAP. | https://openreview.net/pdf/943e8621d1ff1fb8b9daa8cc68876c327966099c.pdf |
Why resampling outperforms reweighting for correcting sampling bias with stochastic gradients | https://openreview.net/forum?id=iQQK02mxVIT | https://openreview.net/forum?id=iQQK02mxVIT | Jing An,Lexing Ying,Yuhua Zhu | ICLR 2021,Poster | A data set sampled from a certain population is biased if the subgroups of the population are sampled at proportions that are significantly different from their underlying proportions. Training machine learning models on biased data sets requires correction techniques to compensate for the bias. We consider two commonly-used techniques, resampling and reweighting, that rebalance the proportions of the subgroups to maintain the desired objective function. Though statistically equivalent, it has been observed that resampling outperforms reweighting when combined with stochastic gradient algorithms. By analyzing illustrative examples, we explain the reason behind this phenomenon using tools from dynamical stability and stochastic asymptotics. We also present experiments from regression, classification, and off-policy prediction to demonstrate that this is a general phenomenon. We argue that it is imperative to consider the objective function design and the optimization algorithm together while addressing the sampling bias.
| https://openreview.net/pdf/0fa836fe2e9df3b7a0bc598d12aa906302c5bf9b.pdf |
Learning with Instance-Dependent Label Noise: A Sample Sieve Approach | https://openreview.net/forum?id=2VXyy9mIyU3 | https://openreview.net/forum?id=2VXyy9mIyU3 | Hao Cheng,Zhaowei Zhu,Xingyu Li,Yifei Gong,Xing Sun,Yang Liu | ICLR 2021,Poster | Human-annotated labels are often prone to noise, and the presence of such noise will degrade the performance of the resulting deep neural network (DNN) models. Much of the literature (with several recent exceptions) of learning with noisy labels focuses on the case when the label noise is independent of features. Practically, annotations errors tend to be instance-dependent and often depend on the difficulty levels of recognizing a certain task. Applying existing results from instance-independent settings would require a significant amount of estimation of noise rates. Therefore, providing theoretically rigorous solutions for learning with instance-dependent label noise remains a challenge. In this paper, we propose CORES$^{2}$ (COnfidence REgularized Sample Sieve), which progressively sieves out corrupted examples. The implementation of CORES$^{2}$ does not require specifying noise rates and yet we are able to provide theoretical guarantees of CORES$^{2}$ in filtering out the corrupted examples. This high-quality sample sieve allows us to treat clean examples and the corrupted ones separately in training a DNN solution, and such a separation is shown to be advantageous in the instance-dependent noise setting. We demonstrate the performance of CORES$^{2}$ on CIFAR10 and CIFAR100 datasets with synthetic instance-dependent label noise and Clothing1M with real-world human noise. As of independent interests, our sample sieve provides a generic machinery for anatomizing noisy datasets and provides a flexible interface for various robust training techniques to further improve the performance. Code is available at https://github.com/UCSC-REAL/cores. | https://openreview.net/pdf/d1f9063b86dd15432acc0f429ea3d22cbb41ae74.pdf |
Unsupervised Audiovisual Synthesis via Exemplar Autoencoders | https://openreview.net/forum?id=43VKWxg_Sqr | https://openreview.net/forum?id=43VKWxg_Sqr | Kangle Deng,Aayush Bansal,Deva Ramanan | ICLR 2021,Poster | We present an unsupervised approach that converts the input speech of any individual into audiovisual streams of potentially-infinitely many output speakers. Our approach builds on simple autoencoders that project out-of-sample data onto the distribution of the training set. We use exemplar autoencoders to learn the voice, stylistic prosody, and visual appearance of a specific target exemplar speech. In contrast to existing methods, the proposed approach can be easily extended to an arbitrarily large number of speakers and styles using only 3 minutes of target audio-video data, without requiring any training data for the input speaker. To do so, we learn audiovisual bottleneck representations that capture the structured linguistic content of speech. We outperform prior approaches on both audio and video synthesis.
| https://openreview.net/pdf/6310fc99e7f1359a15e30437011f7577dab60ff1.pdf |
Single-Timescale Actor-Critic Provably Finds Globally Optimal Policy | https://openreview.net/forum?id=pqZV_srUVmK | https://openreview.net/forum?id=pqZV_srUVmK | Zuyue Fu,Zhuoran Yang,Zhaoran Wang | ICLR 2021,Poster | We study the global convergence and global optimality of actor-critic, one of the most popular families of reinforcement learning algorithms. While most existing works on actor-critic employ bi-level or two-timescale updates, we focus on the more practical single-timescale setting, where the actor and critic are updated simultaneously. Specifically, in each iteration, the critic update is obtained by applying the Bellman evaluation operator only once while the actor is updated in the policy gradient direction computed using the critic. Moreover, we consider two function approximation settings where both the actor and critic are represented by linear or deep neural networks. For both cases, we prove that the actor sequence converges to a globally optimal policy at a sublinear $O(K^{-1/2})$ rate, where $K$ is the number of iterations. To the best of our knowledge, we establish the rate of convergence and global optimality of single-timescale actor-critic with linear function approximation for the first time. Moreover, under the broader scope of policy optimization with nonlinear function approximation, we prove that actor-critic with deep neural network finds the globally optimal policy at a sublinear rate for the first time. | https://openreview.net/pdf/23d733db99e61b484d06180f5dfbb470789b7ff0.pdf |
Wandering within a world: Online contextualized few-shot learning | https://openreview.net/forum?id=oZIvHV04XgC | https://openreview.net/forum?id=oZIvHV04XgC | Mengye Ren,Michael Louis Iuzzolino,Michael Curtis Mozer,Richard Zemel | ICLR 2021,Poster | We aim to bridge the gap between typical human and machine-learning environments by extending the standard framework of few-shot learning to an online, continual setting. In this setting, episodes do not have separate training and testing phases, and instead models are evaluated online while learning novel classes. As in the real world, where the presence of spatiotemporal context helps us retrieve learned skills in the past, our online few-shot learning setting also features an underlying context that changes throughout time. Object classes are correlated within a context and inferring the correct context can lead to better performance. Building upon this setting, we propose a new few-shot learning dataset based on large scale indoor imagery that mimics the visual
experience of an agent wandering within a world. Furthermore, we convert popular few-shot learning approaches into online versions and we also propose a new model that can make use of spatiotemporal contextual information from the recent past. | https://openreview.net/pdf/c7d39970b7f4089067f9662d8cebe4f73c71d1bb.pdf |
Neural Delay Differential Equations | https://openreview.net/forum?id=Q1jmmQz72M2 | https://openreview.net/forum?id=Q1jmmQz72M2 | Qunxi Zhu,Yao Guo,Wei Lin | ICLR 2021,Poster | Neural Ordinary Differential Equations (NODEs), a framework of continuous-depth neural networks, have been widely applied, showing exceptional efficacy in coping with some representative datasets. Recently, an augmented framework has been successfully developed for conquering some limitations emergent in application of the original framework. Here we propose a new class of continuous-depth neural networks with delay, named as Neural Delay Differential Equations (NDDEs), and, for computing the corresponding gradients, we use the adjoint sensitivity method to obtain the delayed dynamics of the adjoint. Since the differential equations with delays are usually seen as dynamical systems of infinite dimension possessing more fruitful dynamics, the NDDEs, compared to the NODEs, own a stronger capacity of nonlinear representations. Indeed, we analytically validate that the NDDEs are of universal approximators, and further articulate an extension of the NDDEs, where the initial function of the NDDEs is supposed to satisfy ODEs. More importantly, we use several illustrative examples to demonstrate the outstanding capacities of the NDDEs and the NDDEs with ODEs' initial value. More precisely, (1) we successfully model the delayed dynamics where the trajectories in the lower-dimensional phase space could be mutually intersected, while the traditional NODEs without any argumentation are not directly applicable for such modeling, and (2) we achieve lower loss and higher accuracy not only for the data produced synthetically by complex models but also for the real-world image datasets, i.e., CIFAR10, MNIST and SVHN. Our results on the NDDEs reveal that appropriately articulating the elements of dynamical systems into the network design is truly beneficial to promoting the network performance. | https://openreview.net/pdf/39dbf04903dd5453991dacd8dcb2333de44b3a6e.pdf |
Learning Parametrised Graph Shift Operators | https://openreview.net/forum?id=0OlrLvrsHwQ | https://openreview.net/forum?id=0OlrLvrsHwQ | George Dasoulas,Johannes F. Lutzeyer,Michalis Vazirgiannis | ICLR 2021,Poster | In many domains data is currently represented as graphs and therefore, the graph representation of this data becomes increasingly important in machine learning. Network data is, implicitly or explicitly, always represented using a graph shift operator (GSO) with the most common choices being the adjacency, Laplacian matrices and their normalisations. In this paper, a novel parametrised GSO (PGSO) is proposed, where specific parameter values result in the most commonly used GSOs and message-passing operators in graph neural network (GNN) frameworks. The PGSO is suggested as a replacement of the standard GSOs that are used in state-of-the-art GNN architectures and the optimisation of the PGSO parameters is seamlessly included in the model training. It is proved that the PGSO has real eigenvalues and a set of real eigenvectors independent of the parameter values and spectral bounds on the PGSO are derived. PGSO parameters are shown to adapt to the sparsity of the graph structure in a study on stochastic blockmodel networks, where they are found to automatically replicate the GSO regularisation found in the literature. On several real-world datasets the accuracy of state-of-the-art GNN architectures is improved by the inclusion of the PGSO in both node- and graph-classification tasks. | https://openreview.net/pdf/e6905a755d143d191885c3b5fef08eb1af58ebb1.pdf |
DOP: Off-Policy Multi-Agent Decomposed Policy Gradients | https://openreview.net/forum?id=6FqKiVAdI3Y | https://openreview.net/forum?id=6FqKiVAdI3Y | Yihan Wang,Beining Han,Tonghan Wang,Heng Dong,Chongjie Zhang | ICLR 2021,Poster | Multi-agent policy gradient (MAPG) methods recently witness vigorous progress. However, there is a significant performance discrepancy between MAPG methods and state-of-the-art multi-agent value-based approaches. In this paper, we investigate causes that hinder the performance of MAPG algorithms and present a multi-agent decomposed policy gradient method (DOP). This method introduces the idea of value function decomposition into the multi-agent actor-critic framework. Based on this idea, DOP supports efficient off-policy learning and addresses the issue of centralized-decentralized mismatch and credit assignment in both discrete and continuous action spaces. We formally show that DOP critics have sufficient representational capability to guarantee convergence. In addition, empirical evaluations on the StarCraft II micromanagement benchmark and multi-agent particle environments demonstrate that DOP outperforms both state-of-the-art value-based and policy-based multi-agent reinforcement learning algorithms. Demonstrative videos are available at https://sites.google.com/view/dop-mapg/. | https://openreview.net/pdf/3f81a57de106341a8eae39f9a550980090df7241.pdf |
Unsupervised Discovery of 3D Physical Objects from Video | https://openreview.net/forum?id=lf7st0bJIA5 | https://openreview.net/forum?id=lf7st0bJIA5 | Yilun Du,Kevin A. Smith,Tomer Ullman,Joshua B. Tenenbaum,Jiajun Wu | ICLR 2021,Poster | We study the problem of unsupervised physical object discovery. While existing frameworks aim to decompose scenes into 2D segments based off each object's appearance, we explore how physics, especially object interactions, facilitates disentangling of 3D geometry and position of objects from video, in an unsupervised manner. Drawing inspiration from developmental psychology, our Physical Object Discovery Network (POD-Net) uses both multi-scale pixel cues and physical motion cues to accurately segment observable and partially occluded objects of varying sizes, and infer properties of those objects. Our model reliably segments objects on both synthetic and real scenes. The discovered object properties can also be used to reason about physical events. | https://openreview.net/pdf/b246a3d6a8c606639649d7945c479176a1f965c7.pdf |
Exploring Balanced Feature Spaces for Representation Learning | https://openreview.net/forum?id=OqtLIabPTit | https://openreview.net/forum?id=OqtLIabPTit | Bingyi Kang,Yu Li,Sa Xie,Zehuan Yuan,Jiashi Feng | ICLR 2021,Poster | Existing self-supervised learning (SSL) methods are mostly applied for training representation models from artificially balanced datasets (e.g., ImageNet). It is unclear how well they will perform in the practical scenarios where datasets are often imbalanced w.r.t. the classes. Motivated by this question, we conduct a series of studies on the performance of self-supervised contrastive learning and supervised learning methods over multiple datasets where training instance distributions vary from a balanced one to a long-tailed one. Our findings are quite intriguing. Different from supervised methods with large performance drop, the self-supervised contrastive learning methods perform stably well even when the datasets are heavily imbalanced. This motivates us to explore the balanced feature spaces learned by contrastive learning, where the feature representations present similar linear separability w.r.t. all the classes. Our further experiments reveal that a representation model generating a balanced feature space can generalize better than that yielding an imbalanced one across multiple settings. Inspired by these insights, we develop a novel representation learning method, called $k$-positive contrastive learning. It effectively combines strengths of the supervised method and the contrastive learning method to learn representations that are both discriminative and balanced. Extensive experiments demonstrate its superiority on multiple recognition tasks. Remarkably, it achieves new state-of-the-art on challenging long-tailed recognition benchmarks. Code and models will be released. | https://openreview.net/pdf/a6e3728ce8efd94d339778f7a7bd85042c5df475.pdf |
Auxiliary Learning by Implicit Differentiation | https://openreview.net/forum?id=n7wIfYPdVet | https://openreview.net/forum?id=n7wIfYPdVet | Aviv Navon,Idan Achituve,Haggai Maron,Gal Chechik,Ethan Fetaya | ICLR 2021,Poster | Training neural networks with auxiliary tasks is a common practice for improving the performance on a main task of interest.
Two main challenges arise in this multi-task learning setting: (i) designing useful auxiliary tasks; and (ii) combining auxiliary tasks into a single coherent loss. Here, we propose a novel framework, AuxiLearn, that targets both challenges based on implicit differentiation. First, when useful auxiliaries are known, we propose learning a network that combines all losses into a single coherent objective function. This network can learn non-linear interactions between tasks. Second, when no useful auxiliary task is known, we describe how to learn a network that generates a meaningful, novel auxiliary task. We evaluate AuxiLearn in a series of tasks and domains, including image segmentation and learning with attributes in the low data regime, and find that it consistently outperforms competing methods. | https://openreview.net/pdf/2571967094ddf29d39923e16c68d6ddb7f5bb72e.pdf |
On the Bottleneck of Graph Neural Networks and its Practical Implications | https://openreview.net/forum?id=i80OPhOCVH2 | https://openreview.net/forum?id=i80OPhOCVH2 | Uri Alon,Eran Yahav | ICLR 2021,Poster | Since the proposal of the graph neural network (GNN) by Gori et al. (2005) and Scarselli et al. (2008), one of the major problems in training GNNs was their struggle to propagate information between distant nodes in the graph.
We propose a new explanation for this problem: GNNs are susceptible to a bottleneck when aggregating messages across a long path. This bottleneck causes the over-squashing of exponentially growing information into fixed-size vectors.
As a result, GNNs fail to propagate messages originating from distant nodes and perform poorly when the prediction task depends on long-range interaction.
In this paper, we highlight the inherent problem of over-squashing in GNNs:
we demonstrate that the bottleneck hinders popular GNNs from fitting long-range signals in the training data;
we further show that GNNs that absorb incoming edges equally, such as GCN and GIN, are more susceptible to over-squashing than GAT and GGNN;
finally, we show that prior work, which extensively tuned GNN models of long-range problems, suffers from over-squashing, and that breaking the bottleneck improves their state-of-the-art results without any tuning or additional weights.
Our code is available at https://github.com/tech-srl/bottleneck/ . | https://openreview.net/pdf/d8fb680febbb7eb04a6ffa0380ec045684628ebc.pdf |
The Role of Momentum Parameters in the Optimal Convergence of Adaptive Polyak's Heavy-ball Methods | https://openreview.net/forum?id=L7WD8ZdscQ5 | https://openreview.net/forum?id=L7WD8ZdscQ5 | Wei Tao,Sheng Long,Gaowei Wu,Qing Tao | ICLR 2021,Poster | The adaptive stochastic gradient descent (SGD) with momentum has been widely adopted in deep learning as well as convex optimization. In practice, the last iterate is commonly used as the final solution. However, the available regret analysis and the setting of constant momentum parameters only guarantee the optimal convergence of the averaged solution. In this paper, we fill this theory-practice gap by investigating the convergence of the last iterate (referred to as {\it individual convergence}), which is a more difficult task than convergence analysis of the averaged solution. Specifically, in the constrained convex cases, we prove that the adaptive Polyak's Heavy-ball (HB) method, in which the step size is only updated using the exponential moving average strategy, attains an individual convergence rate of $O(\frac{1}{\sqrt{t}})$, as opposed to that of $O(\frac{\log t}{\sqrt {t}})$ of SGD, where $t$ is the number of iterations. Our new analysis not only shows how the HB momentum and its time-varying weight help us to achieve the acceleration in convex optimization but also gives valuable hints how the momentum parameters should be scheduled in deep learning. Empirical results validate the correctness of our convergence analysis in optimizing convex functions and demonstrate the improved performance of the adaptive HB methods in training deep networks. | https://openreview.net/pdf/88887f8c02852ef3a9f82b1ae8a010345d221175.pdf |
Entropic gradient descent algorithms and wide flat minima | https://openreview.net/forum?id=xjXg0bnoDmS | https://openreview.net/forum?id=xjXg0bnoDmS | Fabrizio Pittorino,Carlo Lucibello,Christoph Feinauer,Gabriele Perugini,Carlo Baldassi,Elizaveta Demyanenko,Riccardo Zecchina | ICLR 2021,Poster | The properties of flat minima in the empirical risk landscape of neural networks have been debated for some time. Increasing evidence suggests they possess better generalization capabilities with respect to sharp ones. In this work we first discuss the relationship between alternative measures of flatness: The local entropy, which is useful for analysis and algorithm development, and the local energy, which is easier to compute and was shown empirically in extensive tests on state-of-the-art networks to be the best predictor of generalization capabilities. We show semi-analytically in simple controlled scenarios that these two measures correlate strongly with each other and with generalization. Then, we extend the analysis to the deep learning scenario by extensive numerical validations. We study two algorithms, Entropy-SGD and Replicated-SGD, that explicitly include the local entropy in the optimization objective. We devise a training schedule by which we consistently find flatter minima (using both flatness measures), and improve the generalization error for common architectures (e.g. ResNet, EfficientNet). | https://openreview.net/pdf/0b359e4e5d157768feb17f39a57619c3f110d21f.pdf |
SEDONA: Search for Decoupled Neural Networks toward Greedy Block-wise Learning | https://openreview.net/forum?id=XLfdzwNKzch | https://openreview.net/forum?id=XLfdzwNKzch | Myeongjang Pyeon,Jihwan Moon,Taeyoung Hahn,Gunhee Kim | ICLR 2021,Poster | Backward locking and update locking are well-known sources of inefficiency in backpropagation that prevent from concurrently updating layers. Several works have recently suggested using local error signals to train network blocks asynchronously to overcome these limitations. However, they often require numerous iterations of trial-and-error to find the best configuration for local training, including how to decouple network blocks and which auxiliary networks to use for each block. In this work, we propose a differentiable search algorithm named SEDONA to automate this process. Experimental results show that our algorithm can consistently discover transferable decoupled architectures for VGG and ResNet variants, and significantly outperforms the ones trained with end-to-end backpropagation and other state-of-the-art greedy-leaning methods in CIFAR-10, Tiny-ImageNet and ImageNet. | https://openreview.net/pdf/0b88f9a4f997ec52f81dce9a0410a6b5646e18a8.pdf |
not-MIWAE: Deep Generative Modelling with Missing not at Random Data | https://openreview.net/forum?id=tu29GQT0JFy | https://openreview.net/forum?id=tu29GQT0JFy | Niels Bruun Ipsen,Pierre-Alexandre Mattei,Jes Frellsen | ICLR 2021,Poster | When a missing process depends on the missing values themselves, it needs to be explicitly modelled and taken into account while doing likelihood-based inference. We present an approach for building and fitting deep latent variable models (DLVMs) in cases where the missing process is dependent on the missing data. Specifically, a deep neural network enables us to flexibly model the conditional distribution of the missingness pattern given the data. This allows for incorporating prior information about the type of missingness (e.g.~self-censoring) into the model. Our inference technique, based on importance-weighted variational inference, involves maximising a lower bound of the joint likelihood. Stochastic gradients of the bound are obtained by using the reparameterisation trick both in latent space and data space. We show on various kinds of data sets and missingness patterns that explicitly modelling the missing process can be invaluable. | https://openreview.net/pdf/cd42c1e09d98ef209caa6b63d5a67a5273108128.pdf |
Reweighting Augmented Samples by Minimizing the Maximal Expected Loss | https://openreview.net/forum?id=9G5MIc-goqB | https://openreview.net/forum?id=9G5MIc-goqB | Mingyang Yi,Lu Hou,Lifeng Shang,Xin Jiang,Qun Liu,Zhi-Ming Ma | ICLR 2021,Poster | Data augmentation is an effective technique to improve the generalization of deep neural networks. However, previous data augmentation methods usually treat the augmented samples equally without considering their individual impacts on the model. To address this, for the augmented samples from the same training example, we propose to assign different weights to them. We construct the maximal expected loss which is the supremum over any reweighted loss on augmented samples. Inspired by adversarial training, we minimize this maximal expected loss (MMEL) and obtain a simple and interpretable closed-form solution: more attention should be paid to augmented samples with large loss values (i.e., harder examples). Minimizing this maximal expected loss enables the model to perform well under any reweighting strategy. The proposed method can generally be applied on top of any data augmentation methods. Experiments are conducted on both natural language understanding tasks with token-level data augmentation, and image classification tasks with commonly-used image augmentation techniques like random crop and horizontal flip. Empirical results show that the proposed method improves the generalization performance of the model. | https://openreview.net/pdf/e4c9206df0bf95f0a614a0ac2cc2acd08f5125ce.pdf |
Learning to Represent Action Values as a Hypergraph on the Action Vertices | https://openreview.net/forum?id=Xv_s64FiXTv | https://openreview.net/forum?id=Xv_s64FiXTv | Arash Tavakoli,Mehdi Fatemi,Petar Kormushev | ICLR 2021,Poster | Action-value estimation is a critical component of many reinforcement learning (RL) methods whereby sample complexity relies heavily on how fast a good estimator for action value can be learned. By viewing this problem through the lens of representation learning, good representations of both state and action can facilitate action-value estimation. While advances in deep learning have seamlessly driven progress in learning state representations, given the specificity of the notion of agency to RL, little attention has been paid to learning action representations. We conjecture that leveraging the combinatorial structure of multi-dimensional action spaces is a key ingredient for learning good representations of action. To test this, we set forth the action hypergraph networks framework---a class of functions for learning action representations in multi-dimensional discrete action spaces with a structural inductive bias. Using this framework we realise an agent class based on a combination with deep Q-networks, which we dub hypergraph Q-networks. We show the effectiveness of our approach on a myriad of domains: illustrative prediction problems under minimal confounding effects, Atari 2600 games, and discretised physical control benchmarks. | https://openreview.net/pdf/1ffba349384a4d39e139b3deafa3e80e587e2dc1.pdf |
Shapley Explanation Networks | https://openreview.net/forum?id=vsU0efpivw | https://openreview.net/forum?id=vsU0efpivw | Rui Wang,Xiaoqian Wang,David I. Inouye | ICLR 2021,Poster | Shapley values have become one of the most popular feature attribution explanation methods. However, most prior work has focused on post-hoc Shapley explanations, which can be computationally demanding due to its exponential time complexity and preclude model regularization based on Shapley explanations during training. Thus, we propose to incorporate Shapley values themselves as latent representations in deep models thereby making Shapley explanations first-class citizens in the modeling paradigm. This intrinsic explanation approach enables layer-wise explanations, explanation regularization of the model during training, and fast explanation computation at test time. We define the Shapley transform that transforms the input into a Shapley representation given a specific function. We operationalize the Shapley transform as a neural network module and construct both shallow and deep networks, called ShapNets, by composing Shapley modules. We prove that our Shallow ShapNets compute the exact Shapley values and our Deep ShapNets maintain the missingness and accuracy properties of Shapley values. We demonstrate on synthetic and real-world datasets that our ShapNets enable layer-wise Shapley explanations, novel Shapley regularizations during training, and fast computation while maintaining reasonable performance. Code is available at https://github.com/inouye-lab/ShapleyExplanationNetworks. | https://openreview.net/pdf/12770a255a18bc6c25ce69bdda082fd0d7a8cc87.pdf |
NBDT: Neural-Backed Decision Tree | https://openreview.net/forum?id=mCLVeEpplNE | https://openreview.net/forum?id=mCLVeEpplNE | Alvin Wan,Lisa Dunlap,Daniel Ho,Jihan Yin,Scott Lee,Suzanne Petryk,Sarah Adel Bargal,Joseph E. Gonzalez | ICLR 2021,Poster | Machine learning applications such as finance and medicine demand accurate and justifiable predictions, barring most deep learning methods from use. In response, previous work combines decision trees with deep learning, yielding models that (1) sacrifice interpretability for accuracy or (2) sacrifice accuracy for interpretability. We forgo this dilemma by jointly improving accuracy and interpretability using Neural-Backed Decision Trees (NBDTs). NBDTs replace a neural network's final linear layer with a differentiable sequence of decisions and a surrogate loss. This forces the model to learn high-level concepts and lessens reliance on highly-uncertain decisions, yielding (1) accuracy: NBDTs match or outperform modern neural networks on CIFAR, ImageNet and better generalize to unseen classes by up to 16%. Furthermore, our surrogate loss improves the original model's accuracy by up to 2%. NBDTs also afford (2) interpretability: improving human trustby clearly identifying model mistakes and assisting in dataset debugging. Code and pretrained NBDTs are at https://github.com/alvinwan/neural-backed-decision-trees. | https://openreview.net/pdf/9083652f9da499bb55946d63a399c08093258e1c.pdf |
Denoising Diffusion Implicit Models | https://openreview.net/forum?id=St1giarCHLP | https://openreview.net/forum?id=St1giarCHLP | Jiaming Song,Chenlin Meng,Stefano Ermon | ICLR 2021,Poster | Denoising diffusion probabilistic models (DDPMs) have achieved high quality image generation without adversarial training, yet they require simulating a Markov chain for many steps in order to produce a sample. To accelerate sampling, we present denoising diffusion implicit models (DDIMs), a more efficient class of iterative implicit probabilistic models with the same training procedure as DDPMs. In DDPMs, the generative process is defined as the reverse of a particular Markovian diffusion process. We generalize DDPMs via a class of non-Markovian diffusion processes that lead to the same training objective. These non-Markovian processes can correspond to generative processes that are deterministic, giving rise to implicit models that produce high quality samples much faster. We empirically demonstrate that DDIMs can produce high quality samples $10 \times$ to $50 \times$ faster in terms of wall-clock time compared to DDPMs, allow us to trade off computation for sample quality, perform semantically meaningful image interpolation directly in the latent space, and reconstruct observations with very low error. | https://openreview.net/pdf/d01650b6472e18084dcf272952f04a11ffbd6065.pdf |
GShard: Scaling Giant Models with Conditional Computation and Automatic Sharding | https://openreview.net/forum?id=qrwe7XHTmYb | https://openreview.net/forum?id=qrwe7XHTmYb | Dmitry Lepikhin,HyoukJoong Lee,Yuanzhong Xu,Dehao Chen,Orhan Firat,Yanping Huang,Maxim Krikun,Noam Shazeer,Zhifeng Chen | ICLR 2021,Poster | Neural network scaling has been critical for improving the model quality in many real-world machine learning applications with vast amounts of training data and compute. Although this trend of scaling is affirmed to be a sure-fire approach for better model quality, there are challenges on the path such as the computation cost,ease of programming, and efficient implementation on parallel devices. In this paper we demonstrate conditional computation as a remedy to the above mentioned impediments, and demonstrate its efficacy and utility. We make extensive use of GShard, a module composed of a set of lightweight annotation APIs and an extension to the XLA compiler to enable large scale models with up to trillions of parameters. GShard and conditional computation enable us to scale up multilingual neural machine translation Transformer model with Sparsely-Gated Mixture-of-Experts. We demonstrate that such a giant model with 600 billion parameters can efficiently be trained on 2048 TPU v3 cores in 4 days to achieve far superior quality for translation from 100 languages to English compared to the prior art. | https://openreview.net/pdf/cdb90e31da8446076492c5aef0c8c6ae35dd472a.pdf |
What Should Not Be Contrastive in Contrastive Learning | https://openreview.net/forum?id=CZ8Y3NzuVzO | https://openreview.net/forum?id=CZ8Y3NzuVzO | Tete Xiao,Xiaolong Wang,Alexei A Efros,Trevor Darrell | ICLR 2021,Poster | Recent self-supervised contrastive methods have been able to produce impressive transferable visual representations by learning to be invariant to different data augmentations. However, these methods implicitly assume a particular set of representational invariances (e.g., invariance to color), and can perform poorly when a downstream task violates this assumption (e.g., distinguishing red vs. yellow cars). We introduce a contrastive learning framework which does not require prior knowledge of specific, task-dependent invariances. Our model learns to capture varying and invariant factors for visual representations by constructing separate embedding spaces, each of which is invariant to all but one augmentation. We use a multi-head network with a shared backbone which captures information across each augmentation and alone outperforms all baselines on downstream tasks. We further find that the concatenation of the invariant and varying spaces performs best across all tasks we investigate, including coarse-grained, fine-grained, and few-shot downstream classification tasks, and various data corruptions. | https://openreview.net/pdf/2b5543cc1e932e02f5311e6077b7743dd8c446f5.pdf |
On Data-Augmentation and Consistency-Based Semi-Supervised Learning | https://openreview.net/forum?id=7FNqrcPtieT | https://openreview.net/forum?id=7FNqrcPtieT | Atin Ghosh,Alexandre H. Thiery | ICLR 2021,Poster | Recently proposed consistency-based Semi-Supervised Learning (SSL) methods such as the Pi-model, temporal ensembling, the mean teacher, or the virtual adversarial training, achieve the state of the art results in several SSL tasks. These methods can typically reach performances that are comparable to their fully supervised counterparts while using only a fraction of labelled examples. Despite these methodological advances, the understanding of these methods is still relatively limited. To make progress, we analyse (variations of) the Pi-model in settings where analytically tractable results can be obtained. We establish links with Manifold Tangent Classifiers and demonstrate that the quality of the perturbations is key to obtaining reasonable SSL performances. Furthermore, we propose a simple extension of the Hidden Manifold Model that naturally incorporates data-augmentation schemes and offers a tractable framework for understanding SSL methods. | https://openreview.net/pdf/b336e63b29c715b76018ba42cb29f2592569ab56.pdf |
Diverse Video Generation using a Gaussian Process Trigger | https://openreview.net/forum?id=Qm7R_SdqTpT | https://openreview.net/forum?id=Qm7R_SdqTpT | Gaurav Shrivastava,Abhinav Shrivastava | ICLR 2021,Poster | Generating future frames given a few context (or past) frames is a challenging task. It requires modeling the temporal coherence of videos as well as multi-modality in terms of diversity in the potential future states. Current variational approaches for video generation tend to marginalize over multi-modal future outcomes. Instead, we propose to explicitly model the multi-modality in the future outcomes and leverage it to sample diverse futures. Our approach, Diverse Video Generator, uses a GP to learn priors on future states given the past and maintains a probability distribution over possible futures given a particular sample. We leverage the changes in this distribution over time to control the sampling of diverse future states by estimating the end of on-going sequences. In particular, we use the variance of GP over the output function space to trigger a change in the action sequence. We achieve state-of-the-art results on diverse future frame generation in terms of reconstruction quality and diversity of the generated sequences. | https://openreview.net/pdf/fa17e9ba0b11fb7c0b424feda300e370d82df3eb.pdf |
Monotonic Kronecker-Factored Lattice | https://openreview.net/forum?id=0pxiMpCyBtr | https://openreview.net/forum?id=0pxiMpCyBtr | William Taylor Bakst,Nobuyuki Morioka,Erez Louidor | ICLR 2021,Poster | It is computationally challenging to learn flexible monotonic functions that guarantee model behavior and provide interpretability beyond a few input features, and in a time where minimizing resource use is increasingly important, we must be able to learn such models that are still efficient. In this paper we show how to effectively and efficiently learn such functions using Kronecker-Factored Lattice ($\mathrm{KFL}$), an efficient reparameterization of flexible monotonic lattice regression via Kronecker product. Both computational and storage costs scale linearly in the number of input features, which is a significant improvement over existing methods that grow exponentially. We also show that we can still properly enforce monotonicity and other shape constraints. The $\mathrm{KFL}$ function class consists of products of piecewise-linear functions, and the size of the function class can be further increased through ensembling. We prove that the function class of an ensemble of $M$ base $\mathrm{KFL}$ models strictly increases as $M$ increases up to a certain threshold. Beyond this threshold, every multilinear interpolated lattice function can be expressed. Our experimental results demonstrate that $\mathrm{KFL}$ trains faster with fewer parameters while still achieving accuracy and evaluation speeds comparable to or better than the baseline methods and preserving monotonicity guarantees on the learned model. | https://openreview.net/pdf/86fd2e9cf39fac3f9c7f37e71683f60eb7e3e575.pdf |
Auto Seg-Loss: Searching Metric Surrogates for Semantic Segmentation | https://openreview.net/forum?id=MJAqnaC2vO1 | https://openreview.net/forum?id=MJAqnaC2vO1 | Hao Li,Chenxin Tao,Xizhou Zhu,Xiaogang Wang,Gao Huang,Jifeng Dai | ICLR 2021,Poster | Designing proper loss functions is essential in training deep networks. Especially in the field of semantic segmentation, various evaluation metrics have been proposed for diverse scenarios. Despite the success of the widely adopted cross-entropy loss and its variants, the mis-alignment between the loss functions and evaluation metrics degrades the network performance. Meanwhile, manually designing loss functions for each specific metric requires expertise and significant manpower. In this paper, we propose to automate the design of metric-specific loss functions by searching differentiable surrogate losses for each metric. We substitute the non-differentiable operations in the metrics with parameterized functions, and conduct parameter search to optimize the shape of loss surfaces. Two constraints are introduced to regularize the search space and make the search efficient. Extensive experiments on PASCAL VOC and Cityscapes demonstrate that the searched surrogate losses outperform the manually designed loss functions consistently. The searched losses can generalize well to other datasets and networks. Code shall be released at https://github.com/fundamentalvision/Auto-Seg-Loss. | https://openreview.net/pdf/c95e4fe3eeed14d6b9ea3a3421f111e68f1ba477.pdf |
Combining Physics and Machine Learning for Network Flow Estimation | https://openreview.net/forum?id=l0V53bErniB | https://openreview.net/forum?id=l0V53bErniB | Arlei Lopes da Silva,Furkan Kocayusufoglu,Saber Jafarpour,Francesco Bullo,Ananthram Swami,Ambuj Singh | ICLR 2021,Poster | The flow estimation problem consists of predicting missing edge flows in a network (e.g., traffic, power, and water) based on partial observations. These missing flows depend both on the underlying \textit{physics} (edge features and a flow conservation law) as well as the observed edge flows. This paper introduces an optimization framework for computing missing edge flows and solves the problem using bilevel optimization and deep learning. More specifically, we learn regularizers that depend on edge features (e.g., number of lanes in a road, the resistance of a power line) using neural networks. Empirical results show that our method accurately predicts missing flows, outperforming the best baseline, and is able to capture relevant physical properties in traffic and power networks. | https://openreview.net/pdf/9dc2744a465941220de07cf308acf822ec8aaa64.pdf |
NeMo: Neural Mesh Models of Contrastive Features for Robust 3D Pose Estimation | https://openreview.net/forum?id=pmj131uIL9H | https://openreview.net/forum?id=pmj131uIL9H | Angtian Wang,Adam Kortylewski,Alan Yuille | ICLR 2021,Poster | 3D pose estimation is a challenging but important task in computer vision. In this work, we show that standard deep learning approaches to 3D pose estimation are not robust to partial occlusion. Inspired by the robustness of generative vision models to partial occlusion, we propose to integrate deep neural networks with 3D generative representations of objects into a unified neural architecture that we term NeMo. In particular, NeMo learns a generative model of neural feature activations at each vertex on a dense 3D mesh. Using differentiable rendering we estimate the 3D object pose by minimizing the reconstruction error between NeMo and the feature representation of the target image. To avoid local optima in the reconstruction loss, we train the feature extractor to maximize the distance between the individual feature representations on the mesh using contrastive learning. Our extensive experiments on PASCAL3D+, occluded-PASCAL3D+ and ObjectNet3D show that NeMo is much more robust to partial occlusion compared to standard deep networks, while retaining competitive performance on non-occluded data. Interestingly, our experiments also show that NeMo performs reasonably well even when the mesh representation only crudely approximates the true object geometry with a cuboid, hence revealing that the detailed 3D geometry is not needed for accurate 3D pose estimation. | https://openreview.net/pdf/8b6175a9362d0d2b28d4e33e82e7dbcef30c90a0.pdf |
How to Find Your Friendly Neighborhood: Graph Attention Design with Self-Supervision | https://openreview.net/forum?id=Wi5KUNlqWty | https://openreview.net/forum?id=Wi5KUNlqWty | Dongkwan Kim,Alice Oh | ICLR 2021,Poster | Attention mechanism in graph neural networks is designed to assign larger weights to important neighbor nodes for better representation. However, what graph attention learns is not understood well, particularly when graphs are noisy. In this paper, we propose a self-supervised graph attention network (SuperGAT), an improved graph attention model for noisy graphs. Specifically, we exploit two attention forms compatible with a self-supervised task to predict edges, whose presence and absence contain the inherent information about the importance of the relationships between nodes. By encoding edges, SuperGAT learns more expressive attention in distinguishing mislinked neighbors. We find two graph characteristics influence the effectiveness of attention forms and self-supervision: homophily and average degree. Thus, our recipe provides guidance on which attention design to use when those two graph characteristics are known. Our experiment on 17 real-world datasets demonstrates that our recipe generalizes across 15 datasets of them, and our models designed by recipe show improved performance over baselines. | https://openreview.net/pdf/9a9b994ea7dc4b59415ee1780753f7061f19eea4.pdf |
MiCE: Mixture of Contrastive Experts for Unsupervised Image Clustering | https://openreview.net/forum?id=gV3wdEOGy_V | https://openreview.net/forum?id=gV3wdEOGy_V | Tsung Wei Tsai,Chongxuan Li,Jun Zhu | ICLR 2021,Poster | We present Mixture of Contrastive Experts (MiCE), a unified probabilistic clustering framework that simultaneously exploits the discriminative representations learned by contrastive learning and the semantic structures captured by a latent mixture model. Motivated by the mixture of experts, MiCE employs a gating function to partition an unlabeled dataset into subsets according to the latent semantics and multiple experts to discriminate distinct subsets of instances assigned to them in a contrastive learning manner. To solve the nontrivial inference and learning problems caused by the latent variables, we further develop a scalable variant of the Expectation-Maximization (EM) algorithm for MiCE and provide proof of the convergence. Empirically, we evaluate the clustering performance of MiCE on four widely adopted natural image datasets. MiCE achieves significantly better results than various previous methods and a strong contrastive learning baseline. | https://openreview.net/pdf/44a32c550457ebb4f9a6863753563e1f3b2740db.pdf |
Anytime Sampling for Autoregressive Models via Ordered Autoencoding | https://openreview.net/forum?id=TSRTzJnuEBS | https://openreview.net/forum?id=TSRTzJnuEBS | Yilun Xu,Yang Song,Sahaj Garg,Linyuan Gong,Rui Shu,Aditya Grover,Stefano Ermon | ICLR 2021,Poster | Autoregressive models are widely used for tasks such as image and audio generation. The sampling process of these models, however, does not allow interruptions and cannot adapt to real-time computational resources. This challenge impedes the deployment of powerful autoregressive models, which involve a slow sampling process that is sequential in nature and typically scales linearly with respect to the data dimension. To address this difficulty, we propose a new family of autoregressive models that enables anytime sampling. Inspired by Principal Component Analysis, we learn a structured representation space where dimensions are ordered based on their importance with respect to reconstruction. Using an autoregressive model in this latent space, we trade off sample quality for computational efficiency by truncating the generation process before decoding into the original data space. Experimentally, we demonstrate in several image and audio generation tasks that sample quality degrades gracefully as we reduce the computational budget for sampling. The approach suffers almost no loss in sample quality (measured by FID) using only 60\% to 80\% of all latent dimensions for image data. Code is available at https://github.com/Newbeeer/Anytime-Auto-Regressive-Model. | https://openreview.net/pdf/1367785ee095d8df274476d18c53a21ad0d67173.pdf |
Initialization and Regularization of Factorized Neural Layers | https://openreview.net/forum?id=KTlJT1nof6d | https://openreview.net/forum?id=KTlJT1nof6d | Mikhail Khodak,Neil A. Tenenholtz,Lester Mackey,Nicolo Fusi | ICLR 2021,Poster | Factorized layers—operations parameterized by products of two or more matrices—occur in a variety of deep learning contexts, including compressed model training, certain types of knowledge distillation, and multi-head self-attention architectures. We study how to initialize and regularize deep nets containing such layers, examining two simple, understudied schemes, spectral initialization and Frobenius decay, for improving their performance. The guiding insight is to design optimization routines for these networks that are as close as possible to that of their well-tuned, non-decomposed counterparts; we back this intuition with an analysis of how the initialization and regularization schemes impact training with gradient descent, drawing on modern attempts to understand the interplay of weight-decay and batch-normalization. Empirically, we highlight the benefits of spectral initialization and Frobenius decay across a variety of settings. In model compression, we show that they enable low-rank methods to significantly outperform both unstructured sparsity and tensor methods on the task of training low-memory residual networks; analogs of the schemes also improve the performance of tensor decomposition techniques. For knowledge distillation, Frobenius decay enables a simple, overcomplete baseline that yields a compact model from over-parameterized training without requiring retraining with or pruning a teacher network. Finally, we show how both schemes applied to multi-head attention lead to improved performance on both translation and unsupervised pre-training. | https://openreview.net/pdf/d22bc639b3e05ed1c862db4eaa41726d1f24406d.pdf |
CO2: Consistent Contrast for Unsupervised Visual Representation Learning | https://openreview.net/forum?id=U4XLJhqwNF1 | https://openreview.net/forum?id=U4XLJhqwNF1 | Chen Wei,Huiyu Wang,Wei Shen,Alan Yuille | ICLR 2021,Poster | Contrastive learning has recently been a core for unsupervised visual representation learning. Without human annotation, the common practice is to perform an instance discrimination task: Given a query image crop, label crops from the same image as positives, and crops from other randomly sampled images as negatives. An important limitation of this label assignment is that it can not reflect the heterogeneous similarity of the query crop to crops from other images, but regarding them as equally negative. To address this issue, inspired by consistency regularization in semi-supervised learning, we propose Consistent Contrast (CO2), which introduces a consistency term into unsupervised contrastive learning framework. The consistency term takes the similarity of the query crop to crops from other images as unlabeled, and the corresponding similarity of a positive crop as a pseudo label. It then encourages consistency between these two similarities. Empirically, CO2 improves Momentum Contrast (MoCo) by 2.9% top-1 accuracy on ImageNet linear protocol, 3.8% and 1.1% top-5 accuracy on 1% and 10% labeled semi-supervised settings. It also transfers to image classification, object detection, and semantic segmentation on PASCAL VOC. This shows that CO2 learns better visual representations for downstream tasks. | https://openreview.net/pdf/3bf139a485a1e2e65ce2dc0b66f592976f4cfe61.pdf |
DC3: A learning method for optimization with hard constraints | https://openreview.net/forum?id=V1ZHVxJ6dSS | https://openreview.net/forum?id=V1ZHVxJ6dSS | Priya L. Donti,David Rolnick,J Zico Kolter | ICLR 2021,Poster | Large optimization problems with hard constraints arise in many settings, yet classical solvers are often prohibitively slow, motivating the use of deep networks as cheap "approximate solvers." Unfortunately, naive deep learning approaches typically cannot enforce the hard constraints of such problems, leading to infeasible solutions. In this work, we present Deep Constraint Completion and Correction (DC3), an algorithm to address this challenge. Specifically, this method enforces feasibility via a differentiable procedure, which implicitly completes partial solutions to satisfy equality constraints and unrolls gradient-based corrections to satisfy inequality constraints. We demonstrate the effectiveness of DC3 in both synthetic optimization tasks and the real-world setting of AC optimal power flow, where hard constraints encode the physics of the electrical grid. In both cases, DC3 achieves near-optimal objective values while preserving feasibility. | https://openreview.net/pdf/07736947eb069cd731941458cb558b9bbec98ba1.pdf |
Enforcing robust control guarantees within neural network policies | https://openreview.net/forum?id=5lhWG3Hj2By | https://openreview.net/forum?id=5lhWG3Hj2By | Priya L. Donti,Melrose Roderick,Mahyar Fazlyab,J Zico Kolter | ICLR 2021,Poster | When designing controllers for safety-critical systems, practitioners often face a challenging tradeoff between robustness and performance. While robust control methods provide rigorous guarantees on system stability under certain worst-case disturbances, they often yield simple controllers that perform poorly in the average (non-worst) case. In contrast, nonlinear control methods trained using deep learning have achieved state-of-the-art performance on many control tasks, but often lack robustness guarantees. In this paper, we propose a technique that combines the strengths of these two approaches: constructing a generic nonlinear control policy class, parameterized by neural networks, that nonetheless enforces the same provable robustness criteria as robust control. Specifically, our approach entails integrating custom convex-optimization-based projection layers into a neural network-based policy. We demonstrate the power of this approach on several domains, improving in average-case performance over existing robust control methods and in worst-case stability over (non-robust) deep RL methods. | https://openreview.net/pdf/c31f3f12091e950c7b2a58373e1af5225f1b5377.pdf |
Robust and Generalizable Visual Representation Learning via Random Convolutions | https://openreview.net/forum?id=BVSM0x3EDK6 | https://openreview.net/forum?id=BVSM0x3EDK6 | Zhenlin Xu,Deyi Liu,Junlin Yang,Colin Raffel,Marc Niethammer | ICLR 2021,Poster | While successful for various computer vision tasks, deep neural networks have shown to be vulnerable to texture style shifts and small perturbations to which humans are robust. In this work, we show that the robustness of neural networks can be greatly improved through the use of random convolutions as data augmentation. Random convolutions are approximately shape-preserving and may distort local textures. Intuitively, randomized convolutions create an infinite number of new domains with similar global shapes but random local texture. Therefore, we explore using outputs of multi-scale random convolutions as new images or mixing them with the original images during training. When applying a network trained with our approach to unseen domains, our method consistently improves the performance on domain generalization benchmarks and is scalable to ImageNet. In particular, in the challenging scenario of generalizing to the sketch domain in PACS and to ImageNet-Sketch, our method outperforms state-of-art methods by a large margin. More interestingly, our method can benefit downstream tasks by providing a more robust pretrained visual representation. | https://openreview.net/pdf/10040feb356410f9e8b8ee9678ecb702e0534990.pdf |
WrapNet: Neural Net Inference with Ultra-Low-Precision Arithmetic | https://openreview.net/forum?id=3SqrRe8FWQ- | https://openreview.net/forum?id=3SqrRe8FWQ- | Renkun Ni,Hong-min Chu,Oscar Castaneda,Ping-yeh Chiang,Christoph Studer,Tom Goldstein | ICLR 2021,Poster | Low-precision neural networks represent both weights and activations with few bits, drastically reducing the cost of multiplications. Meanwhile, these products are accumulated using high-precision (typically 32-bit) additions. Additions dominate the arithmetic complexity of inference in quantized (e.g., binary) nets, and high precision is needed to avoid overflow. To further optimize inference, we propose WrapNet, an architecture that adapts neural networks to use low-precision (8-bit) additions while achieving classification accuracy comparable to their 32-bit counterparts. We achieve resilience to low-precision accumulation by inserting a cyclic activation layer that makes results invariant to overflow. We demonstrate the efficacy of our approach using both software and hardware platforms. | https://openreview.net/pdf/f4b38b314f5630e45d41c5ff39eace1a2b8afca6.pdf |
Rank the Episodes: A Simple Approach for Exploration in Procedurally-Generated Environments | https://openreview.net/forum?id=MtEE0CktZht | https://openreview.net/forum?id=MtEE0CktZht | Daochen Zha,Wenye Ma,Lei Yuan,Xia Hu,Ji Liu | ICLR 2021,Poster | Exploration under sparse reward is a long-standing challenge of model-free reinforcement learning. The state-of-the-art methods address this challenge by introducing intrinsic rewards to encourage exploration in novel states or uncertain environment dynamics. Unfortunately, methods based on intrinsic rewards often fall short in procedurally-generated environments, where a different environment is generated in each episode so that the agent is not likely to visit the same state more than once. Motivated by how humans distinguish good exploration behaviors by looking into the entire episode, we introduce RAPID, a simple yet effective episode-level exploration method for procedurally-generated environments. RAPID regards each episode as a whole and gives an episodic exploration score from both per-episode and long-term views. Those highly scored episodes are treated as good exploration behaviors and are stored in a small ranking buffer. The agent then imitates the episodes in the buffer to reproduce the past good exploration behaviors. We demonstrate our method on several procedurally-generated MiniGrid environments, a first-person-view 3D Maze navigation task from MiniWorld, and several sparse MuJoCo tasks. The results show that RAPID significantly outperforms the state-of-the-art intrinsic reward strategies in terms of sample efficiency and final performance. The code is available at https://github.com/daochenzha/rapid | https://openreview.net/pdf/1a3486946250b679d39b93e6b12ff912023a3955.pdf |
Stochastic Security: Adversarial Defense Using Long-Run Dynamics of Energy-Based Models | https://openreview.net/forum?id=gwFTuzxJW0 | https://openreview.net/forum?id=gwFTuzxJW0 | Mitch Hill,Jonathan Craig Mitchell,Song-Chun Zhu | ICLR 2021,Poster | The vulnerability of deep networks to adversarial attacks is a central problem for deep learning from the perspective of both cognition and security. The current most successful defense method is to train a classifier using adversarial images created during learning. Another defense approach involves transformation or purification of the original input to remove adversarial signals before the image is classified. We focus on defending naturally-trained classifiers using Markov Chain Monte Carlo (MCMC) sampling with an Energy-Based Model (EBM) for adversarial purification. In contrast to adversarial training, our approach is intended to secure highly vulnerable pre-existing classifiers. To our knowledge, no prior defensive transformation is capable of securing naturally-trained classifiers, and our method is the first to validate a post-training defense approach that is distinct from current successful defenses which modify classifier training.
The memoryless behavior of long-run MCMC sampling will eventually remove adversarial signals, while metastable behavior preserves consistent appearance of MCMC samples after many steps to allow accurate long-run prediction. Balancing these factors can lead to effective purification and robust classification. We evaluate adversarial defense with an EBM using the strongest known attacks against purification. Our contributions are 1) an improved method for training EBM's with realistic long-run MCMC samples for effective purification, 2) an Expectation-Over-Transformation (EOT) defense that resolves ambiguities for evaluating stochastic defenses and from which the EOT attack naturally follows, and 3) state-of-the-art adversarial defense for naturally-trained classifiers and competitive defense compared to adversarial training on CIFAR-10, SVHN, and CIFAR-100. Our code and pre-trained models are available at https://github.com/point0bar1/ebm-defense. | https://openreview.net/pdf/f72ab5e343005a27c47de932d2ee35163cb3c617.pdf |
Nearest Neighbor Machine Translation | https://openreview.net/forum?id=7wCBOfJ8hJM | https://openreview.net/forum?id=7wCBOfJ8hJM | Urvashi Khandelwal,Angela Fan,Dan Jurafsky,Luke Zettlemoyer,Mike Lewis | ICLR 2021,Poster | We introduce $k$-nearest-neighbor machine translation ($k$NN-MT), which predicts tokens with a nearest-neighbor classifier over a large datastore of cached examples, using representations from a neural translation model for similarity search. This approach requires no additional training and scales to give the decoder direct access to billions of examples at test time, resulting in a highly expressive model that consistently improves performance across many settings. Simply adding nearest-neighbor search improves a state-of-the-art German-English translation model by 1.5 BLEU. $k$NN-MT allows a single model to be adapted to diverse domains by using a domain-specific datastore, improving results by an average of 9.2 BLEU over zero-shot transfer, and achieving new state-of-the-art results---without training on these domains. A massively multilingual model can also be specialized for particular language pairs, with improvements of 3 BLEU for translating from English into German and Chinese. Qualitatively, $k$NN-MT is easily interpretable; it combines source and target context to retrieve highly relevant examples. | https://openreview.net/pdf/8c353a40fa6aa514c510ad3cc89f331af070681b.pdf |
Learning Hyperbolic Representations of Topological Features | https://openreview.net/forum?id=yqPnIRhHtZv | https://openreview.net/forum?id=yqPnIRhHtZv | Panagiotis Kyriakis,Iordanis Fostiropoulos,Paul Bogdan | ICLR 2021,Poster | Learning task-specific representations of persistence diagrams is an important problem in topological data analysis and machine learning. However, current state of the art methods are restricted in terms of their expressivity as they are focused on Euclidean representations. Persistence diagrams often contain features of infinite persistence (i.e., essential features) and Euclidean spaces shrink their importance relative to non-essential features because they cannot assign infinite distance to finite points. To deal with this issue, we propose a method to learn representations of persistence diagrams on hyperbolic spaces, more specifically on the Poincare ball. By representing features of infinite persistence infinitesimally close to the boundary of the ball, their distance to non-essential features approaches infinity, thereby their relative importance is preserved. This is achieved without utilizing extremely high values for the learnable parameters, thus the representation can be fed into downstream optimization methods and trained efficiently in an end-to-end fashion. We present experimental results on graph and image classification tasks and show that the performance of our method is on par with or exceeds the performance of other state of the art methods.
| https://openreview.net/pdf/90d9d8bd76c09f1561cba895e6574b5a137974f8.pdf |
Selectivity considered harmful: evaluating the causal impact of class selectivity in DNNs | https://openreview.net/forum?id=8nl0k08uMi | https://openreview.net/forum?id=8nl0k08uMi | Matthew L Leavitt,Ari S. Morcos | ICLR 2021,Poster | The properties of individual neurons are often analyzed in order to understand the biological and artificial neural networks in which they're embedded. Class selectivity—typically defined as how different a neuron's responses are across different classes of stimuli or data samples—is commonly used for this purpose. However, it remains an open question whether it is necessary and/or sufficient for deep neural networks (DNNs) to learn class selectivity in individual units. We investigated the causal impact of class selectivity on network function by directly regularizing for or against class selectivity. Using this regularizer to reduce class selectivity across units in convolutional neural networks increased test accuracy by over 2% in ResNet18 and 1% in ResNet50 trained on Tiny ImageNet. For ResNet20 trained on CIFAR10 we could reduce class selectivity by a factor of 2.5 with no impact on test accuracy, and reduce it nearly to zero with only a small (~2%) drop in test accuracy. In contrast, regularizing to increase class selectivity significantly decreased test accuracy across all models and datasets. These results indicate that class selectivity in individual units is neither sufficient nor strictly necessary, and can even impair DNN performance. They also encourage caution when focusing on the properties of single units as representative of the mechanisms by which DNNs function. | https://openreview.net/pdf/17afa91e20d55de3a216e07f59e5ab74d27027a6.pdf |
Integrating Categorical Semantics into Unsupervised Domain Translation | https://openreview.net/forum?id=IMPA6MndSXU | https://openreview.net/forum?id=IMPA6MndSXU | Samuel Lavoie-Marchildon,Faruk Ahmed,Aaron Courville | ICLR 2021,Poster | While unsupervised domain translation (UDT) has seen a lot of success recently, we argue that mediating its translation via categorical semantic features could broaden its applicability. In particular, we demonstrate that categorical semantics improves the translation between perceptually different domains sharing multiple object categories. We propose a method to learn, in an unsupervised manner, categorical semantic features (such as object labels) that are invariant of the source and target domains. We show that conditioning the style encoder of unsupervised domain translation methods on the learned categorical semantics leads to a translation preserving the digits on MNIST$\leftrightarrow$SVHN and to a more realistic stylization on Sketches$\to$Reals. | https://openreview.net/pdf/63644971703ef2d045a2404b8db5e211dd90c275.pdf |
Zero-shot Synthesis with Group-Supervised Learning | https://openreview.net/forum?id=8wqCDnBmnrT | https://openreview.net/forum?id=8wqCDnBmnrT | Yunhao Ge,Sami Abu-El-Haija,Gan Xin,Laurent Itti | ICLR 2021,Poster | Visual cognition of primates is superior to that of artificial neural networks in its ability to “envision” a visual object, even a newly-introduced one, in different attributes including pose, position, color, texture, etc. To aid neural networks to envision objects with different attributes, we propose a family of objective functions, expressed on groups of examples, as a novel learning framework that we term Group-Supervised Learning (GSL). GSL allows us to decompose inputs into a disentangled representation with swappable components, that can be recombined to synthesize new samples. For instance, images of red boats & blue cars can be decomposed and recombined to synthesize novel images of red cars. We propose an implementation based on auto-encoder, termed group-supervised zero-shot synthesis network (GZS-Net) trained with our learning framework, that can produce a high-quality red car even if no such example is witnessed during training. We test our model and learning framework on existing benchmarks, in addition to a new dataset that we open-source. We qualitatively and quantitatively demonstrate that GZS-Net trained with GSL outperforms state-of-the-art methods | https://openreview.net/pdf/5cc8c0d37d4aa9db1645e1181b60428d215c25b4.pdf |
VA-RED$^2$: Video Adaptive Redundancy Reduction | https://openreview.net/forum?id=g21u6nlbPzn | https://openreview.net/forum?id=g21u6nlbPzn | Bowen Pan,Rameswar Panda,Camilo Luciano Fosco,Chung-Ching Lin,Alex J Andonian,Yue Meng,Kate Saenko,Aude Oliva,Rogerio Feris | ICLR 2021,Poster | Performing inference on deep learning models for videos remains a challenge due to the large amount of computational resources required to achieve robust recognition. An inherent property of real-world videos is the high correlation of information across frames which can translate into redundancy in either temporal or spatial feature maps of the models, or both. The type of redundant features depends on the dynamics and type of events in the video: static videos have more temporal redundancy while videos focusing on objects tend to have more channel redundancy. Here we present a redundancy reduction framework, termed VA-RED$^2$, which is input-dependent. Specifically, our VA-RED$^2$ framework uses an input-dependent policy to decide how many features need to be computed for temporal and channel dimensions. To keep the capacity of the original model, after fully computing the necessary features, we reconstruct the remaining redundant features from those using cheap linear operations. We learn the adaptive policy jointly with the network weights in a differentiable way with a shared-weight mechanism, making it highly efficient. Extensive experiments on multiple video datasets and different visual tasks show that our framework achieves $20\% - 40\%$ reduction in computation (FLOPs) when compared to state-of-the-art methods without any performance loss. Project page: http://people.csail.mit.edu/bpan/va-red/. | https://openreview.net/pdf/474acfaf4c205394a38367bb1c69399a31b1264e.pdf |
BERTology Meets Biology: Interpreting Attention in Protein Language Models | https://openreview.net/forum?id=YWtLZvLmud7 | https://openreview.net/forum?id=YWtLZvLmud7 | Jesse Vig,Ali Madani,Lav R. Varshney,Caiming Xiong,richard socher,Nazneen Rajani | ICLR 2021,Poster | Transformer architectures have proven to learn useful representations for protein classification and generation tasks. However, these representations present challenges in interpretability. In this work, we demonstrate a set of methods for analyzing protein Transformer models through the lens of attention. We show that attention: (1) captures the folding structure of proteins, connecting amino acids that are far apart in the underlying sequence, but spatially close in the three-dimensional structure, (2) targets binding sites, a key functional component of proteins, and (3) focuses on progressively more complex biophysical properties with increasing layer depth. We find this behavior to be consistent across three Transformer architectures (BERT, ALBERT, XLNet) and two distinct protein datasets. We also present a three-dimensional visualization of the interaction between attention and protein structure. Code for visualization and analysis is available at https://github.com/salesforce/provis. | https://openreview.net/pdf/8fe71cea00c77fd4bce646070f7b96ef9300e91d.pdf |
Trajectory Prediction using Equivariant Continuous Convolution | https://openreview.net/forum?id=J8_GttYLFgr | https://openreview.net/forum?id=J8_GttYLFgr | Robin Walters,Jinxi Li,Rose Yu | ICLR 2021,Poster | Trajectory prediction is a critical part of many AI applications, for example, the safe operation of autonomous vehicles. However, current methods are prone to making inconsistent and physically unrealistic predictions. We leverage insights from fluid dynamics to overcome this limitation by considering internal symmetry in real-world trajectories. We propose a novel model, Equivariant Continous COnvolution (ECCO) for improved trajectory prediction. ECCO uses rotationally-equivariant continuous convolutions to embed the symmetries of the system. On both vehicle and pedestrian trajectory datasets, ECCO attains competitive accuracy with significantly fewer parameters. It is also more sample efficient, generalizing automatically from few data points in any orientation. Lastly, ECCO improves generalization with equivariance, resulting in more physically consistent predictions. Our method provides a fresh perspective towards increasing trust and transparency in deep learning models. Our code and data can be found at https://github.com/Rose-STL-Lab/ECCO. | https://openreview.net/pdf/e66067f5a20544270d6679390629804b037a33d7.pdf |
Unsupervised Meta-Learning through Latent-Space Interpolation in Generative Models | https://openreview.net/forum?id=XOjv2HxIF6i | https://openreview.net/forum?id=XOjv2HxIF6i | Siavash Khodadadeh,Sharare Zehtabian,Saeed Vahidian,Weijia Wang,Bill Lin,Ladislau Boloni | ICLR 2021,Poster | Several recently proposed unsupervised meta-learning approaches rely on synthetic meta-tasks created using techniques such as random selection, clustering and/or augmentation. In this work, we describe a novel approach that generates meta-tasks using generative models. The proposed family of algorithms generate pairs of in-class and out-of-class samples from the latent space in a principled way, allowing us to create synthetic classes forming the training and validation data of a meta-task. We find that the proposed approach, LAtent Space Interpolation Unsupervised Meta-learning (LASIUM), outperforms or is competitive with current unsupervised learning baselines on few-shot classification tasks on the most widely used benchmark datasets. | https://openreview.net/pdf/8242bb1cb5c7dc59b17a736254778eca82c71c06.pdf |
Heteroskedastic and Imbalanced Deep Learning with Adaptive Regularization | https://openreview.net/forum?id=mEdwVCRJuX4 | https://openreview.net/forum?id=mEdwVCRJuX4 | Kaidi Cao,Yining Chen,Junwei Lu,Nikos Arechiga,Adrien Gaidon,Tengyu Ma | ICLR 2021,Poster | Real-world large-scale datasets are heteroskedastic and imbalanced --- labels have varying levels of uncertainty and label distributions are long-tailed. Heteroskedasticity and imbalance challenge deep learning algorithms due to the difficulty of distinguishing among mislabeled, ambiguous, and rare examples. Addressing heteroskedasticity and imbalance simultaneously is under-explored. We propose a data-dependent regularization technique for heteroskedastic datasets that regularizes different regions of the input space differently. Inspired by the theoretical derivation of the optimal regularization strength in a one-dimensional nonparametric classification setting, our approach adaptively regularizes the data points in higher-uncertainty, lower-density regions more heavily. We test our method on several benchmark tasks, including a real-world heteroskedastic and imbalanced dataset, WebVision. Our experiments corroborate our theory and demonstrate a significant improvement over other methods in noise-robust deep learning. | https://openreview.net/pdf/99d4ce0622fd824cd067efb8dda5611146eb4070.pdf |
Implicit Under-Parameterization Inhibits Data-Efficient Deep Reinforcement Learning | https://openreview.net/forum?id=O9bnihsFfXU | https://openreview.net/forum?id=O9bnihsFfXU | Aviral Kumar,Rishabh Agarwal,Dibya Ghosh,Sergey Levine | ICLR 2021,Poster | We identify an implicit under-parameterization phenomenon in value-based deep RL methods that use bootstrapping: when value functions, approximated using deep neural networks, are trained with gradient descent using iterated regression onto target values generated by previous instances of the value network, more gradient updates decrease the expressivity of the current value network. We char- acterize this loss of expressivity via a drop in the rank of the learned value net- work features, and show that this typically corresponds to a performance drop. We demonstrate this phenomenon on Atari and Gym benchmarks, in both offline and online RL settings. We formally analyze this phenomenon and show that it results from a pathological interaction between bootstrapping and gradient-based optimization. We further show that mitigating implicit under-parameterization by controlling rank collapse can improve performance. | https://openreview.net/pdf/766d703650bcb86c35c53681b89ad0a2aba30504.pdf |
Bowtie Networks: Generative Modeling for Joint Few-Shot Recognition and Novel-View Synthesis | https://openreview.net/forum?id=ESG-DMKQKsD | https://openreview.net/forum?id=ESG-DMKQKsD | Zhipeng Bao,Yu-Xiong Wang,Martial Hebert | ICLR 2021,Poster | We propose a novel task of joint few-shot recognition and novel-view synthesis: given only one or few images of a novel object from arbitrary views with only category annotation, we aim to simultaneously learn an object classifier and generate images of that type of object from new viewpoints. While existing work copes with two or more tasks mainly by multi-task learning of shareable feature representations, we take a different perspective. We focus on the interaction and cooperation between a generative model and a discriminative model, in a way that facilitates knowledge to flow across tasks in complementary directions. To this end, we propose bowtie networks that jointly learn 3D geometric and semantic representations with a feedback loop. Experimental evaluation on challenging fine-grained recognition datasets demonstrates that our synthesized images are realistic from multiple viewpoints and significantly improve recognition performance as ways of data augmentation, especially in the low-data regime. | https://openreview.net/pdf/7991edacbb1e82b6f57da8010cb4bbb939dd5461.pdf |
Saliency is a Possible Red Herring When Diagnosing Poor Generalization | https://openreview.net/forum?id=c9-WeM-ceB | https://openreview.net/forum?id=c9-WeM-ceB | Joseph D Viviano,Becks Simpson,Francis Dutil,Yoshua Bengio,Joseph Paul Cohen | ICLR 2021,Poster | Poor generalization is one symptom of models that learn to predict target variables using spuriously-correlated image features present only in the training distribution instead of the true image features that denote a class. It is often thought that this can be diagnosed visually using attribution (aka saliency) maps. We study if this assumption is correct. In some prediction tasks, such as for medical images, one may have some images with masks drawn by a human expert, indicating a region of the image containing relevant information to make the prediction. We study multiple methods that take advantage of such auxiliary labels, by training networks to ignore distracting features which may be found outside of the region of interest. This mask information is only used during training and has an impact on generalization accuracy depending on the severity of the shift between the training and test distributions. Surprisingly, while these methods improve generalization performance in the presence of a covariate shift, there is no strong correspondence between the correction of attribution towards the features a human expert have labelled as important and generalization performance. These results suggest that the root cause of poor generalization may not always be spatially defined, and raise questions about the utility of masks as 'attribution priors' as well as saliency maps for explainable predictions. | https://openreview.net/pdf/11d962b7010ee77620458bfc746f11370a23e2dc.pdf |
What Can You Learn From Your Muscles? Learning Visual Representation from Human Interactions | https://openreview.net/forum?id=Qm8UNVCFdh | https://openreview.net/forum?id=Qm8UNVCFdh | Kiana Ehsani,Daniel Gordon,Thomas Hai Dang Nguyen,Roozbeh Mottaghi,Ali Farhadi | ICLR 2021,Poster | Learning effective representations of visual data that generalize to a variety of downstream tasks has been a long quest for computer vision. Most representation learning approaches rely solely on visual data such as images or videos. In this paper, we explore a novel approach, where we use human interaction and attention cues to investigate whether we can learn better representations compared to visual-only representations. For this study, we collect a dataset of human interactions capturing body part movements and gaze in their daily lives. Our experiments show that our ``"muscly-supervised" representation that encodes interaction and attention cues outperforms a visual-only state-of-the-art method MoCo (He et al.,2020), on a variety of target tasks: scene classification (semantic), action recognition (temporal), depth estimation (geometric), dynamics prediction (physics) and walkable surface estimation (affordance). Our code and dataset are available at: https://github.com/ehsanik/muscleTorch. | https://openreview.net/pdf/68e63865a158882a87332ea2d2b8f196693b3f62.pdf |
Conservative Safety Critics for Exploration | https://openreview.net/forum?id=iaO86DUuKi | https://openreview.net/forum?id=iaO86DUuKi | Homanga Bharadhwaj,Aviral Kumar,Nicholas Rhinehart,Sergey Levine,Florian Shkurti,Animesh Garg | ICLR 2021,Poster | Safe exploration presents a major challenge in reinforcement learning (RL): when active data collection requires deploying partially trained policies, we must ensure that these policies avoid catastrophically unsafe regions, while still enabling trial and error learning. In this paper, we target the problem of safe exploration in RL, by learning a conservative safety estimate of environment states through a critic, and provably upper bound the likelihood of catastrophic failures at every training iteration. We theoretically characterize the tradeoff between safety and policy improvement, show that the safety constraints are satisfied with high probability during training, derive provable convergence guarantees for our approach which is no worse asymptotically then standard RL, and empirically demonstrate the efficacy of the proposed approach on a suite of challenging navigation, manipulation, and locomotion tasks. Our results demonstrate that the proposed approach can achieve competitive task performance, while incurring significantly lower catastrophic failure rates during training as compared to prior methods. Videos are at this URL https://sites.google.com/view/conservative-safety-critics/ | https://openreview.net/pdf/31cfa17ce6b5a4dd1c7e3bf3ce4c025642d3199e.pdf |
Multi-Time Attention Networks for Irregularly Sampled Time Series | https://openreview.net/forum?id=4c0J6lwQ4_ | https://openreview.net/forum?id=4c0J6lwQ4_ | Satya Narayan Shukla,Benjamin Marlin | ICLR 2021,Poster | Irregular sampling occurs in many time series modeling applications where it presents a significant challenge to standard deep learning models. This work is motivated by the analysis of physiological time series data in electronic health records, which are sparse, irregularly sampled, and multivariate. In this paper, we propose a new deep learning framework for this setting that we call Multi-Time Attention Networks. Multi-Time Attention Networks learn an embedding of continuous time values and use an attention mechanism to produce a fixed-length representation of a time series containing a variable number of observations. We investigate the performance of this framework on interpolation and classification tasks using multiple datasets. Our results show that the proposed approach performs as well or better than a range of baseline and recently proposed models while offering significantly faster training times than current state-of-the-art methods. | https://openreview.net/pdf/35e8d4bd3fd59a402e488637f16e0fec9e36706a.pdf |
Graph Traversal with Tensor Functionals: A Meta-Algorithm for Scalable Learning | https://openreview.net/forum?id=6DOZ8XNNfGN | https://openreview.net/forum?id=6DOZ8XNNfGN | Elan Sopher Markowitz,Keshav Balasubramanian,Mehrnoosh Mirtaheri,Sami Abu-El-Haija,Bryan Perozzi,Greg Ver Steeg,Aram Galstyan | ICLR 2021,Poster | Graph Representation Learning (GRL) methods have impacted fields from chemistry to social science. However, their algorithmic implementations are specialized to specific use-cases e.g. "message passing" methods are run differently from "node embedding" ones. Despite their apparent differences, all these methods utilize the graph structure, and therefore, their learning can be approximated with stochastic graph traversals. We propose Graph Traversal via Tensor Functionals (GTTF), a unifying meta-algorithm framework for easing the implementation of diverse graph algorithms and enabling transparent and efficient scaling to large graphs. GTTF is founded upon a data structure (stored as a sparse tensor) and a stochastic graph traversal algorithm (described using tensor operations). The algorithm is a functional that accept two functions, and can be specialized to obtain a variety of GRL models and objectives, simply by changing those two functions. We show for a wide class of methods, our algorithm learns in an unbiased fashion and, in expectation, approximates the learning as if the specialized implementations were run directly.
With these capabilities, we scale otherwise non-scalable methods to set state-of-the-art on large graph datasets while being more efficient than existing GRL libraries -- with only a handful of lines of code for each method specialization. | https://openreview.net/pdf/32f7d4bfb27e1afd27046d2e43aa02a64afd5952.pdf |
What they do when in doubt: a study of inductive biases in seq2seq learners | https://openreview.net/forum?id=YmA86Zo-P_t | https://openreview.net/forum?id=YmA86Zo-P_t | Eugene Kharitonov,Rahma Chaabouni | ICLR 2021,Poster | Sequence-to-sequence (seq2seq) learners are widely used, but we still have only limited knowledge about what inductive biases shape the way they generalize. We address that by investigating how popular seq2seq learners generalize in tasks that have high ambiguity in the training data. We use four new tasks to study learners' preferences for memorization, arithmetic, hierarchical, and compositional reasoning. Further, we connect to Solomonoff's theory of induction and propose to use description length as a principled and sensitive measure of inductive biases. In our experimental study, we find that LSTM-based learners can learn to perform counting, addition, and multiplication by a constant from a single training example. Furthermore, Transformer and LSTM-based learners show a bias toward the hierarchical induction over the linear one, while CNN-based learners prefer the opposite. The latter also show a bias toward a compositional generalization over memorization. Finally, across all our experiments, description length proved to be a sensitive measure of inductive biases. | https://openreview.net/pdf/585c196f903498736cc63f2856f42b9c3fd4139d.pdf |
Tomographic Auto-Encoder: Unsupervised Bayesian Recovery of Corrupted Data | https://openreview.net/forum?id=YtMG5ex0ou | https://openreview.net/forum?id=YtMG5ex0ou | Francesco Tonolini,Pablo Garcia Moreno,Andreas Damianou,Roderick Murray-Smith | ICLR 2021,Poster | We propose a new probabilistic method for unsupervised recovery of corrupted data. Given a large ensemble of degraded samples, our method recovers accurate posteriors of clean values, allowing the exploration of the manifold of possible reconstructed data and hence characterising the underlying uncertainty. In this set-ting, direct application of classical variational methods often gives rise to collapsed densities that do not adequately explore the solution space. Instead, we derive our novel reduced entropy condition approximate inference method that results in rich posteriors. We test our model in a data recovery task under the common setting of missing values and noise, demonstrating superior performance to existing variational methods for imputation and de-noising with different real data sets. We further show higher classification accuracy after imputation, proving the advantage of propagating uncertainty to downstream tasks with our model. | https://openreview.net/pdf/1a034aa157281f3316e3bb5608ae9c2bb7e13e30.pdf |
OPAL: Offline Primitive Discovery for Accelerating Offline Reinforcement Learning | https://openreview.net/forum?id=V69LGwJ0lIN | https://openreview.net/forum?id=V69LGwJ0lIN | Anurag Ajay,Aviral Kumar,Pulkit Agrawal,Sergey Levine,Ofir Nachum | ICLR 2021,Poster | Reinforcement learning (RL) has achieved impressive performance in a variety of online settings in which an agent’s ability to query the environment for transitions and rewards is effectively unlimited. However, in many practical applications, the situation is reversed: an agent may have access to large amounts of undirected offline experience data, while access to the online environment is severely limited. In this work, we focus on this offline setting. Our main insight is that, when presented with offline data composed of a variety of behaviors, an effective way to leverage this data is to extract a continuous space of recurring and temporally extended primitive behaviors before using these primitives for downstream task learning. Primitives extracted in this way serve two purposes: they delineate the behaviors that are supported by the data from those that are not, making them useful for avoiding distributional shift in offline RL; and they provide a degree of temporal abstraction, which reduces the effective horizon yielding better learning in theory, and improved offline RL in practice. In addition to benefiting offline policy optimization, we show that performing offline primitive learning in this way can also be leveraged for improving few-shot imitation learning as well as exploration and transfer in online RL on a variety of benchmark domains. Visualizations and code are available at https://sites.google.com/view/opal-iclr | https://openreview.net/pdf/b57bd58e05e17b49da003dded33a75506b9ddbac.pdf |
Knowledge distillation via softmax regression representation learning | https://openreview.net/forum?id=ZzwDy_wiWv | https://openreview.net/forum?id=ZzwDy_wiWv | Jing Yang,Brais Martinez,Adrian Bulat,Georgios Tzimiropoulos | ICLR 2021,Poster | This paper addresses the problem of model compression via knowledge distillation. We advocate for a method that optimizes the output feature of the penultimate layer of the student network and hence is directly related to representation learning. Previous distillation methods which typically impose direct feature matching between the student and the teacher do not take into account the classification problem at hand. On the contrary, our distillation method decouples representation learning and classification and utilizes the teacher's pre-trained classifier to train the student's penultimate layer feature. In particular, for the same input image, we wish the teacher's and student's feature to produce the same output when passed through the teacher's classifier which is achieved with a simple $L_2$ loss. Our method is extremely simple to implement and straightforward to train and is shown to consistently outperform previous state-of-the-art methods over a large set of experimental settings including different (a) network architectures, (b) teacher-student capacities, (c) datasets, and (d) domains. The code will be available at \url{https://github.com/jingyang2017/KD_SRRL}. | https://openreview.net/pdf/94c700e0e7f29d564e416cc3b56b6b350c29aafb.pdf |
Large Associative Memory Problem in Neurobiology and Machine Learning | https://openreview.net/forum?id=X4y_10OX-hX | https://openreview.net/forum?id=X4y_10OX-hX | Dmitry Krotov,John J. Hopfield | ICLR 2021,Poster | Dense Associative Memories or modern Hopfield networks permit storage and reliable retrieval of an exponentially large (in the dimension of feature space) number of memories. At the same time, their naive implementation is non-biological, since it seemingly requires the existence of many-body synaptic junctions between the neurons. We show that these models are effective descriptions of a more microscopic (written in terms of biological degrees of freedom) theory that has additional (hidden) neurons and only requires two-body interactions between them. For this reason our proposed microscopic theory is a valid model of large associative memory with a degree of biological plausibility. The dynamics of our network and its reduced dimensional equivalent both minimize energy (Lyapunov) functions. When certain dynamical variables (hidden neurons) are integrated out from our microscopic theory, one can recover many of the models that were previously discussed in the literature, e.g. the model presented in "Hopfield Networks is All You Need" paper. We also provide an alternative derivation of the energy function and the update rule proposed in the aforementioned paper and clarify the relationships between various models of this class. | https://openreview.net/pdf/707e3c461fd1c2e94f05cad3278e6aaa3a32f2de.pdf |
Remembering for the Right Reasons: Explanations Reduce Catastrophic Forgetting | https://openreview.net/forum?id=tHgJoMfy6nI | https://openreview.net/forum?id=tHgJoMfy6nI | Sayna Ebrahimi,Suzanne Petryk,Akash Gokul,William Gan,Joseph E. Gonzalez,Marcus Rohrbach,trevor darrell | ICLR 2021,Poster | The goal of continual learning (CL) is to learn a sequence of tasks without suffering from the phenomenon of catastrophic forgetting. Previous work has shown that leveraging memory in the form of a replay buffer can reduce performance degradation on prior tasks. We hypothesize that forgetting can be further reduced when the model is encouraged to remember the \textit{evidence} for previously made decisions. As a first step towards exploring this hypothesis, we propose a simple novel training paradigm, called Remembering for the Right Reasons (RRR), that additionally stores visual model explanations for each example in the buffer and ensures the model has ``the right reasons'' for its predictions by encouraging its explanations to remain consistent with those used to make decisions at training time. Without this constraint, there is a drift in explanations and increase in forgetting as conventional continual learning algorithms learn new tasks. We demonstrate how RRR can be easily added to any memory or regularization-based approach and results in reduced forgetting, and more importantly, improved model explanations. We have evaluated our approach in the standard and few-shot settings and observed a consistent improvement across various CL approaches using different architectures and techniques to generate model explanations and demonstrated our approach showing a promising connection between explainability and continual learning. Our code is available at \url{https://github.com/SaynaEbrahimi/Remembering-for-the-Right-Reasons}. | https://openreview.net/pdf/c9a83cab23e5aa3aa692ee2b8ff385ac2cd2da0b.pdf |
Deep Networks and the Multiple Manifold Problem | https://openreview.net/forum?id=O-6Pm_d_Q- | https://openreview.net/forum?id=O-6Pm_d_Q- | Sam Buchanan,Dar Gilboa,John Wright | ICLR 2021,Poster | We study the multiple manifold problem, a binary classification task modeled on applications in machine vision, in which a deep fully-connected neural network is trained to separate two low-dimensional submanifolds of the unit sphere. We provide an analysis of the one-dimensional case, proving for a simple manifold configuration that when the network depth $L$ is large relative to certain geometric and statistical properties of the data, the network width $n$ grows as a sufficiently large polynomial in $L$, and the number of i.i.d. samples from the manifolds is polynomial in $L$, randomly-initialized gradient descent rapidly learns to classify the two manifolds perfectly with high probability. Our analysis demonstrates concrete benefits of depth and width in the context of a practically-motivated model problem: the depth acts as a fitting resource, with larger depths corresponding to smoother networks that can more readily separate the class manifolds, and the width acts as a statistical resource, enabling concentration of the randomly-initialized network and its gradients. The argument centers around the "neural tangent kernel" of Jacot et al. and its role in the nonasymptotic analysis of training overparameterized neural networks; to this literature, we contribute essentially optimal rates of concentration for the neural tangent kernel of deep fully-connected ReLU networks, requiring width $n \geq L\,\mathrm{poly}(d_0)$ to achieve uniform concentration of the initial kernel over a $d_0$-dimensional submanifold of the unit sphere $\mathbb{S}^{n_0-1}$, and a nonasymptotic framework for establishing generalization of networks trained in the "NTK regime" with structured data. The proof makes heavy use of martingale concentration to optimally treat statistical dependencies across layers of the initial random network. This approach should be of use in establishing similar results for other network architectures. | https://openreview.net/pdf/56f1d814ac75024a949f12d01b0c52e5a0e108f3.pdf |
Interpreting Knowledge Graph Relation Representation from Word Embeddings | https://openreview.net/forum?id=gLWj29369lW | https://openreview.net/forum?id=gLWj29369lW | Carl Allen,Ivana Balazevic,Timothy Hospedales | ICLR 2021,Poster | Many models learn representations of knowledge graph data by exploiting its low-rank latent structure, encoding known relations between entities and enabling unknown facts to be inferred. To predict whether a relation holds between entities, embeddings are typically compared in the latent space following a relation-specific mapping. Whilst their predictive performance has steadily improved, how such models capture the underlying latent structure of semantic information remains unexplained. Building on recent theoretical understanding of word embeddings, we categorise knowledge graph relations into three types and for each derive explicit requirements of their representations. We show that empirical properties of relation representations and the relative performance of leading knowledge graph representation methods are justified by our analysis. | https://openreview.net/pdf/1feb4e4d391a73dd7a1de57536d4621967f6cd43.pdf |
Learning perturbation sets for robust machine learning | https://openreview.net/forum?id=MIDckA56aD | https://openreview.net/forum?id=MIDckA56aD | Eric Wong,J Zico Kolter | ICLR 2021,Poster | Although much progress has been made towards robust deep learning, a significant gap in robustness remains between real-world perturbations and more narrowly defined sets typically studied in adversarial defenses. In this paper, we aim to bridge this gap by learning perturbation sets from data, in order to characterize real-world effects for robust training and evaluation. Specifically, we use a conditional generator that defines the perturbation set over a constrained region of the latent space. We formulate desirable properties that measure the quality of a learned perturbation set, and theoretically prove that a conditional variational autoencoder naturally satisfies these criteria. Using this framework, our approach can generate a variety of perturbations at different complexities and scales, ranging from baseline spatial transformations, through common image corruptions, to lighting variations. We measure the quality of our learned perturbation sets both quantitatively and qualitatively, finding that our models are capable of producing a diverse set of meaningful perturbations beyond the limited data seen during training. Finally, we leverage our learned perturbation sets to train models which are empirically and certifiably robust to adversarial image corruptions and adversarial lighting variations, while improving generalization on non-adversarial data. All code and configuration files for reproducing the experiments as well as pretrained model weights can be found at https://github.com/locuslab/perturbation_learning. | https://openreview.net/pdf/b2d7a9e83f89aa52841bff6e66a954313729ecb3.pdf |
A Trainable Optimal Transport Embedding for Feature Aggregation and its Relationship to Attention | https://openreview.net/forum?id=ZK6vTvb84s | https://openreview.net/forum?id=ZK6vTvb84s | Grégoire Mialon,Dexiong Chen,Alexandre d'Aspremont,Julien Mairal | ICLR 2021,Poster | We address the problem of learning on sets of features, motivated by the need of performing pooling operations in long biological sequences of varying sizes, with long-range dependencies, and possibly few labeled data. To address this challenging task, we introduce a parametrized representation of fixed size, which embeds and then aggregates elements from a given input set according to the optimal transport plan between the set and a trainable reference. Our approach scales to large datasets and allows end-to-end training of the reference, while also providing a simple unsupervised learning mechanism with small computational cost. Our aggregation technique admits two useful interpretations: it may be seen as a mechanism related to attention layers in neural networks, or it may be seen as a scalable surrogate of a classical optimal transport-based kernel. We experimentally demonstrate the effectiveness of our approach on biological sequences, achieving state-of-the-art results for protein fold recognition and detection of chromatin profiles tasks, and, as a proof of concept, we show promising results for processing natural language sequences. We provide an open-source implementation of our embedding that can be used alone or as a module in larger learning models at https://github.com/claying/OTK. | https://openreview.net/pdf/209f1d71b4e8e73a59634e771ff1c3a43d72b849.pdf |
RODE: Learning Roles to Decompose Multi-Agent Tasks | https://openreview.net/forum?id=TTUVg6vkNjK | https://openreview.net/forum?id=TTUVg6vkNjK | Tonghan Wang,Tarun Gupta,Anuj Mahajan,Bei Peng,Shimon Whiteson,Chongjie Zhang | ICLR 2021,Poster | Role-based learning holds the promise of achieving scalable multi-agent learning by decomposing complex tasks using roles. However, it is largely unclear how to efficiently discover such a set of roles. To solve this problem, we propose to first decompose joint action spaces into restricted role action spaces by clustering actions according to their effects on the environment and other agents. Learning a role selector based on action effects makes role discovery much easier because it forms a bi-level learning hierarchy: the role selector searches in a smaller role space and at a lower temporal resolution, while role policies learn in significantly reduced primitive action-observation spaces. We further integrate information about action effects into the role policies to boost learning efficiency and policy generalization. By virtue of these advances, our method (1) outperforms the current state-of-the-art MARL algorithms on 9 of the 14 scenarios that comprise the challenging StarCraft II micromanagement benchmark and (2) achieves rapid transfer to new environments with three times the number of agents. Demonstrative videos can be viewed at https://sites.google.com/view/rode-marl. | https://openreview.net/pdf/b3099a13254cda0cd68fedf29a0227d9684bc973.pdf |
Scalable Bayesian Inverse Reinforcement Learning | https://openreview.net/forum?id=4qR3coiNaIv | https://openreview.net/forum?id=4qR3coiNaIv | Alex James Chan,Mihaela van der Schaar | ICLR 2021,Poster | Bayesian inference over the reward presents an ideal solution to the ill-posed nature of the inverse reinforcement learning problem. Unfortunately current methods generally do not scale well beyond the small tabular setting due to the need for an inner-loop MDP solver, and even non-Bayesian methods that do themselves scale often require extensive interaction with the environment to perform well, being inappropriate for high stakes or costly applications such as healthcare. In this paper we introduce our method, Approximate Variational Reward Imitation Learning (AVRIL), that addresses both of these issues by jointly learning an approximate posterior distribution over the reward that scales to arbitrarily complicated state spaces alongside an appropriate policy in a completely offline manner through a variational approach to said latent reward. Applying our method to real medical data alongside classic control simulations, we demonstrate Bayesian reward inference in environments beyond the scope of current methods, as well as task performance competitive with focused offline imitation learning algorithms. | https://openreview.net/pdf/a94f190029f9ebd5affafc141a770843d501a8a2.pdf |
Learning "What-if" Explanations for Sequential Decision-Making | https://openreview.net/forum?id=h0de3QWtGG | https://openreview.net/forum?id=h0de3QWtGG | Ioana Bica,Daniel Jarrett,Alihan Hüyük,Mihaela van der Schaar | ICLR 2021,Poster | Building interpretable parameterizations of real-world decision-making on the basis of demonstrated behavior--i.e. trajectories of observations and actions made by an expert maximizing some unknown reward function--is essential for introspecting and auditing policies in different institutions. In this paper, we propose learning explanations of expert decisions by modeling their reward function in terms of preferences with respect to ``"what if'' outcomes: Given the current history of observations, what would happen if we took a particular action? To learn these cost-benefit tradeoffs associated with the expert's actions, we integrate counterfactual reasoning into batch inverse reinforcement learning. This offers a principled way of defining reward functions and explaining expert behavior, and also satisfies the constraints of real-world decision-making---where active experimentation is often impossible (e.g. in healthcare). Additionally, by estimating the effects of different actions, counterfactuals readily tackle the off-policy nature of policy evaluation in the batch setting, and can naturally accommodate settings where the expert policies depend on histories of observations rather than just current states. Through illustrative experiments in both real and simulated medical environments, we highlight the effectiveness of our batch, counterfactual inverse reinforcement learning approach in recovering accurate and interpretable descriptions of behavior. | https://openreview.net/pdf/8007445a08db8481ba43eaac1f01ad2c9495810b.pdf |
Learning advanced mathematical computations from examples | https://openreview.net/forum?id=-gfhS00XfKj | https://openreview.net/forum?id=-gfhS00XfKj | Francois Charton,Amaury Hayat,Guillaume Lample | ICLR 2021,Poster | Using transformers over large generated datasets, we train models to learn mathematical properties of differential systems, such as local stability, behavior at infinity and controllability. We achieve near perfect prediction of qualitative characteristics, and good approximations of numerical features of the system. This demonstrates that neural networks can learn to perform complex computations, grounded in advanced theory, from examples, without built-in mathematical knowledge. | https://openreview.net/pdf/fc6de8bba436a4a56d0914d23d125bfb2eb84682.pdf |
Repurposing Pretrained Models for Robust Out-of-domain Few-Shot Learning | https://openreview.net/forum?id=qkLMTphG5-h | https://openreview.net/forum?id=qkLMTphG5-h | Namyeong Kwon,Hwidong Na,Gabriel Huang,Simon Lacoste-Julien | ICLR 2021,Poster | Model-agnostic meta-learning (MAML) is a popular method for few-shot learning but assumes that we have access to the meta-training set. In practice, training on the meta-training set may not always be an option due to data privacy concerns, intellectual property issues, or merely lack of computing resources. In this paper, we consider the novel problem of repurposing pretrained MAML checkpoints to solve new few-shot classification tasks. Because of the potential distribution mismatch, the original MAML steps may no longer be optimal. Therefore we propose an alternative meta-testing procedure and combine MAML gradient steps with adversarial training and uncertainty-based stepsize adaptation. Our method outperforms "vanilla" MAML on same-domain and cross-domains benchmarks using both SGD and Adam optimizers and shows improved robustness to the choice of base stepsize. | https://openreview.net/pdf/d7cdb4aa01e48fb3d0a82cabe99b8fe0b1c57f47.pdf |
Empirical Analysis of Unlabeled Entity Problem in Named Entity Recognition | https://openreview.net/forum?id=5jRVa89sZk | https://openreview.net/forum?id=5jRVa89sZk | Yangming Li,lemao liu,Shuming Shi | ICLR 2021,Poster | In many scenarios, named entity recognition (NER) models severely suffer from unlabeled entity problem, where the entities of a sentence may not be fully annotated. Through empirical studies performed on synthetic datasets, we find two causes of performance degradation. One is the reduction of annotated entities and the other is treating unlabeled entities as negative instances. The first cause has less impact than the second one and can be mitigated by adopting pretraining language models. The second cause seriously misguides a model in training and greatly affects its performances. Based on the above observations, we propose a general approach, which can almost eliminate the misguidance brought by unlabeled entities. The key idea is to use negative sampling that, to a large extent, avoids training NER models with unlabeled entities. Experiments on synthetic datasets and real-world datasets show that our model is robust to unlabeled entity problem and surpasses prior baselines. On well-annotated datasets, our model is competitive with the state-of-the-art method. | https://openreview.net/pdf/3caa712b0d1c2caaf7b3578b53b9f9c46e78db74.pdf |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.