From aed5b081b4f96fb5eec3e837b9e82d500da0c017 Mon Sep 17 00:00:00 2001 From: Sungwoo Kim Date: Sat, 10 Aug 2024 10:58:40 -0400 Subject: [PATCH] add papers --- .../Declarative Recursive Computation on an RDBMS | 1 + ... for Certifiable Defense against Patch Attacks | 1 + ...usible 3D Human Models to Ambiguous Image Data | 1 + ...3D Self-Supervised Methods for Medical Imaging | 1 + .../3D Shape Reconstruction from Vision and Touch | 1 + ...nse Large-Scale Networks: Design and Inference | 1 + ...g Algorithm and Applications to Auction Design | 1 + ... Nonparametrics View into Deep Representations | 1 + ...spective on Training Speed and Model Selection | 1 + ...eralization in Grounded Language Understanding | 1 + ...sible Neural Network for Slow Feature Analysis | 1 + ...oolean Task Algebra for Reinforcement Learning | 1 + .../A Catalyst Framework for Minimax Optimization | 1 + ...A Causal View on Robustness of Neural Networks | 1 + ...ithms for General Instrumental Variable Models | 1 + .../A Closer Look at Accuracy vs. Robustness | 1 + ...the Training Strategy for Modern Meta-Learning | 1 + ...Combinatorial Perspective on Transfer Learning | 1 + ...n between Private Learning and Online Learning | 1 + ...ror Descent Approach to Sparse Phase Retrieval | 1 + ...m for Simulations of Multi-modal Distributions | 1 + ...Assembly and Viral Quasispecies Reconstruction | 1 + ...rithm for Training Generative Adversarial Nets | 1 + ... to Domain-Invariant Learning in Deep Networks | 1 + ...opic Model without the Reparametrization Trick | 1 + ...tral Limit Theorem for Shallow Neural Networks | 1 + ...air Classifier Using Kernel Density Estimation | 1 + ... for Nonconvex Sparse Constrained Optimization | 1 + ...nalysis of Two Time-Scale Actor-Critic Methods | 1 + ...iors with Adaptive Smoothing and Game Encoding | 1 + ...s of Additive Adversarial Attacks and Defenses | 1 + ...aximization Algorithm with Endogenous Sampling | 1 + ... Framework for Solving Integer Linear Programs | 1 + ...eneral Method for Robust Learning from Batches | 1 + ... Kernel Analysis for Two-layer Neural Networks | 1 + ...roup-Theoretic Framework for Data Augmentation | 2 ++ .../A Limitation of the PAC-Bayes Framework | 2 ++ ...Code for Distributional Reinforcement Learning | 1 + ...al Networks Based on Watson's Perceptual Model | 1 + ... and Its Application to Co-occurrence Matrices | 1 + ...o Off-Policy Evaluation in Average-Reward MDPs | 1 + ...Approach to Kernel Conditional Mean Embeddings | 1 + ...nalysis for Stein Variational Gradient Descent | 1 + ...r Constrained Optimization in Graphical Models | 1 + ...ategy to Solve Hard Sokoban Planning Instances | 1 + ...thm to Reduce the Support of Discrete Measures | 1 + ...ification and Localisation in Object Detection | 1 + ...l EM Algorithm for Incomplete Panel Count Data | 1 + ...vacy-Preserving Collaborative Machine Learning | 1 + ...r Learning Optimal Multivariate Decision Trees | 1 + .../neurips/A Self-Tuning Actor-Critic Algorithm | 1 + ...mple Language Model for Task-Oriented Dialogue | 1 + ... for Faster Optimization and Local Exploration | 1 + ...ion with Adversarial or Stochastic Constraints | 1 + ...gorithm for Nonconvex-Concave Min-Max Problems | 1 + ... Energy Distance for Parallel Speech Synthesis | 1 + ... Low-bitwidth Training of Deep Neural Networks | 1 + ...ask-Agnostic Sample Design in Machine Learning | 1 + ...l EstimatoR Expectation Maximization Algorithm | 1 + ...dy on Encodings for Neural Architecture Search | 2 ++ ...A Theoretical Framework for Target Propagation | 1 + ... Bound and Efficient Reduction for Swap Regret | 1 + ...pological Filter for Learning with Label Noise | 1 + ... Convergence Analysis of Q-Learning Algorithms | 1 + .../A Unified View of Label Shift Estimation | 1 + ...of Optimism in Episodic Reinforcement Learning | 1 + ...works for Expressing Probability Distributions | 1 + ... for Learning from Positive and Unlabeled Data | 1 + ...al view of compositional zero-shot recognition | 2 ++ ...zation formulation for multivariate regression | 1 + ...ement using multi-agent reinforcement learning | 1 + .../neurips/A kernel test for quasi-independence | 1 + ... automatic differentiation in machine learning | 1 + ...thematical theory of cooperative communication | 1 + ...an-field analysis of two-player zero-sum games | 1 + ...carve a desired function into a neural network | 1 + ... Q-learning with linear function approximation | 1 + ...zed linear models of noisy interacting neurons | 1 + ...ariational form of the Schatten-$p$ quasi-norm | 1 + ...rithm for learning nonparametric causal graphs | 1 + .../A shooting formulation of deep learning | 1 + ...mates local non-Hebbian learning in the cortex | 1 + ...ymbolic regression exploiting graph modularity | 1 + ... Based Identity Swapping for Forgery Detection | 1 + ...Expanding Receptive Field for Dense Prediction | 1 + ...aset: Anonymized Videos from Diverse Countries | 1 + ...anguage Models with Progressive Layer Dropping | 1 + .../Acceleration with a Ball Optimization Oracle | 1 + ...alized Odds by Resampling Sensitive Attributes | 1 + ...iction: Experiment Selection through Stability | 1 + ...rning of Causal DAGs via Directed Clique Trees | 1 + ... Stepsizes by the Belief in Observed Gradients | 1 + ...o Share For Efficient Deep Multi-Task Learning | 1 + ...tive Tensor Program Compilation Made Efficient | 1 + .../Adam with Bandit Sampling for Deep Learning | 1 + ...Allow Identification of Optimized Neural Codes | 1 + .../Adapting Neural Architectures Between Domains | 1 + ...ting to Misspecification in Contextual Bandits | 1 + ...ization for Model-Based Reinforcement Learning | 2 ++ ...al Interference: A Maximum Likelihood Approach | 2 ++ ...ve Gradient Quantization for Data-Parallel SGD | 1 + ...onal Recurrent Network for Traffic Forecasting | 1 + ...zation and Sampling with Decreasing Step-Sizes | 1 + ...dels for Efficient Pairwise Sequence Alignment | 1 + ...line Estimation of Piecewise Polynomial Trends | 1 + ...ive Probing Policies for Shortest Path Routing | 1 + .../2020/neurips/Adaptive Reduced Rank Regression | 1 + ...e Sampling for Stochastic Risk-Averse Learning | 1 + ...tive Shrinkage Estimation for Streaming Graphs | 1 + ...ox Adversarial Attacks using Normalizing Flows | 1 + ... Flows, Importance Weighting, and Optimization | 1 + .../Adversarial Attacks on Deep Graph Matching | 1 + ...versarial Attacks on Linear Contextual Bandits | 1 + ...ns: Regret Lower Bound and No-regret Algorithm | 0 data/2020/neurips/Adversarial Blocking Bandits | 1 + ...cing Through Robust Rank-One Matrix Completion | 1 + ...stributional Training for Robust Deep Learning | 1 + data/2020/neurips/Adversarial Example Games | 1 + ...dversarial Learning for Robust Deep Clustering | 1 + ...rsarial Robustness of Supervised Sparse Coding | 1 + ...versarial Self-Supervised Contrastive Learning | 1 + ...Imitation Learning without Policy Optimization | 1 + ...Sparse Transformer for Time Series Forecasting | 1 + ...ng for One-Shot Unsupervised Domain Adaptation | 1 + ...of Data-dependent Operator Norm Regularization | 1 + ...eight Perturbation Helps Robust Generalization | 1 + ...robustness via robust low rank representations | 4 ++++ ...st Few-Shot Learning: A Meta-Learning Approach | 1 + ... Streaming Algorithms via Differential Privacy | 1 + ...semble of Discrete Undirected Graphical Models | 1 + ...s on Approximation Error and Sample Complexity | 1 + ...rning of a Single Neuron with Gradient Descent | 1 + .../Agnostic Learning with Multiple Objectives | 1 + ...emble Knowledge Distillation in Gradient Space | 1 + ...ect causal knowledge: a probabilistic approach | 1 + .../All Word Embeddings from One Embedding | 1 + ...transitions in sparse spiked matrix estimation | 1 + ... Learningvia Reference-Advantage Decomposition | 1 + .../neurips/Almost Surely Stable Deep Dynamics | 1 + ...n Analysis of SVD for Deep Rotation Estimation | 1 + ...mental Algorithm for Contextual Linear Bandits | 1 + ...fficient Adversarial Attack for Tree Ensembles | 1 + ... Evolutionary and Gradient-based Policy Search | 1 + ...ent Framework for Clustered Federated Learning | 1 + ...lgorithms for Combinatorial and Linear Bandits | 1 + ... and Non-Uniform Sampling in Experience Replay | 1 + ...ch to Transfer Learning with Dynamics Mismatch | 1 + ...y Gradient and Natural Policy Gradient Methods | 1 + ...s of Stochastic Gradient Descent with Momentum | 1 + ... Elimination Algorithm for Learning a Best Arm | 1 + ... Estimator for Learning with Augmented Classes | 1 + ...nformation-Theoretic Perceptual Quality Metric | 1 + ...etworks dynamics for hinge loss classification | 1 + ...tion of stagewise convex optimization problems | 1 + ...rning approach for parametric modal regression | 1 + .../An operator view of policy gradient methods | 1 + ...son Sampling for Stochastic Partial Monitoring | 1 + ...ian in Shallow ReLU Models: A Tale of Symmetry | 1 + ...tion-Maximization for Deep Generative Networks | 1 + ...cations of Common Entropy for Causal Inference | 1 + ...oximate Cross-Validation for Structured Models | 1 + ...lidation with Low-Rank Data in High Dimensions | 1 + ...ained Learning with Lagrange Multiplier Models | 1 + ...nce Reduction for Reparameterization Gradients | 1 + ... Ability to Solve the Symbol Grounding Problem | 1 + ...g: A Framework for Multi-Organization Learning | 1 + ...l knowledge into model-agnostic explainability | 1 + ...eling Based on the Smooth Wasserstein Distance | 1 + ...rs neural network in the random features model | 1 + ...ly Optimal Exact Minibatch Metropolis-Hastings | 1 + ...es, You Really Can Backdoor Federated Learning | 1 + ...ment Learning Model for Traffic Signal Control | 1 + ...n implement reward-based error backpropagation | 1 + ...ibute Prototype Network for Zero-Shot Learning | 1 + ...ompression for Reliable Network Interpretation | 1 + ...udio Generation for a Silent Performance Video | 1 + ... Machine Learning: How Private is Private SGD? | 1 + data/2020/neurips/Auto Learning Attention | 1 + ... Architecture Search for Panoptic Segmentation | 1 + ...ient Algorithm for Block Stacking Style Search | 1 + ... Selection for Secure Neural Network Inference | 2 ++ ...ze for Data-Parallel Distributed Deep Learning | 1 + ...coders that don't overfit towards the Identity | 1 + .../Autofocused oracles for model-based design | 1 + ...Curriculum Learning through Value Disagreement | 1 + ...s for Scalable Certified Robustness and Beyond | 1 + ...ity-aware Surrogates for Optimization Problems | 1 + data/2020/neurips/Autoregressive Score Matching | 1 + ...ary Task Reweighting for Minimum-data Learning | 1 + data/2020/neurips/AvE: Assistance via Empowerment | 1 + ...iding Side Effects By Considering Future Tasks | 1 + .../Avoiding Side Effects in Complex Environments | 1 + .../Axioms for Learning from Pairwise Comparisons | 1 + ...Learning for Batch Deep Reinforcement Learning | 1 + ...nce: Fast and Robust Inference with Early Exit | 1 + ...BOSS: Bayesian Optimization over String Spaces | 1 + .../BRP-NAS: Prediction-based NAS using GCNs | 1 + ...proves Transferability of Adversarial Examples | 1 + ...Bad Global Minima Exist and SGD Can Reach Them | 1 + ...eta-Softmax for Long-Tailed Visual Recognition | 1 + data/2020/neurips/Bandit Linear Control | 1 + ...it Samplers for Training Graph Neural Networks | 1 + ...e k-Medoids Clustering via Multi-Armed Bandits | 1 + ...pproach to search over molecule synthesis DAGs | 1 + ...ollapse for randomly initialised deep networks | 1 + .../Batched Coarse Ranking in Multi-Armed Bandits | 1 + data/2020/neurips/Baxter Permutation Process | 1 + ...onal Learning for Multi-omics Data Integration | 1 + ... Loss Functions and the Scoring Function Class | 1 + data/2020/neurips/Bayesian Attention Modules | 1 + ...yesian Bits: Unifying Quantization and Pruning | 1 + ...g with Zero-Inflated Poisson Bayesian Networks | 1 + ...n Deep Ensembles via the Neural Tangent Kernel | 1 + ... a Probabilistic Perspective of Generalization | 1 + ...ning for the Few-Shot Setting via Deep Kernels | 1 + ...type Mean Field Multi-agent Imitation Learning | 1 + .../Bayesian Optimization for Iterative Learning | 1 + .../Bayesian Optimization of Risk Measures | 1 + ...c Numerical Integration with Tree-Based Models | 1 + data/2020/neurips/Bayesian Pseudocoresets | 1 + ...ian Robust Optimization for Imitation Learning | 1 + ...n-adaptive neural network optimization methods | 1 + .../neurips/Belief Propagation Neural Networks | 1 + ...overy in POMDPs using the Value of Information | 1 + ...odels over time, and the Neural-Adjoint method | 1 + ...ng Interpretability in Time Series Predictions | 1 + ...ulti-Hop Logical Reasoning in Knowledge Graphs | 1 + ... Pedestrian Detection from Another Perspective | 1 + ...trix Regret via Parameter-Free Online Learning | 1 + ...r Set Representations For Relational Reasoning | 1 + ...rks: Current Limitations and Effective Designs | 1 + ... Interactive Summaries of Actionable Recourses | 1 + ...ng for Over-parameterized Tensor Decomposition | 1 + ...ntees with Arbitrary Adversarial Test Examples | 1 + ...CNNs and humans by measuring error consistency | 1 + ...Processes Improve the Predictive Uncertainties | 1 + ...r Learning Energy-based Latent Variable Models | 1 + ...regret bounds for adversarial bandits and MDPs | 2 ++ ... Convolutional Poisson Gamma Dynamical Systems | 1 + .../Big Bird: Transformers for Longer Sequences | 1 + ...sed Models are Strong Semi-Supervised Learners | 1 + ...ough dynamic inversion of feedforward networks | 1 + ...Inspired Mechanisms for Adversarial Robustness | 1 + ...ing: A Functional Optimization Based Framework | 1 + ... Optimization with Local Generative Surrogates | 1 + ...odels using generative evolutionary algorithms | 1 + ...ideo Temporal Consistency via Deep Video Prior | 1 + ...e Scene Representations from Unlabelled Images | 1 + ...or Efficient Monte-Carlo Bayesian Optimization | 1 + ...dversarial Training with Hypersphere Embedding | 1 + ...tive: New Schemes with Faster Worst-Case Rates | 1 + ...t - A New Approach to Self-Supervised Learning | 1 + data/2020/neurips/Bootstrapping neural processes | 1 + ...ry thickness and robustness in learning models | 1 + ... Embedding Model for Knowledge Base Completion | 1 + ... Langevin Dynamics for Non-Convex Optimization | 1 + ...ng the Communication-Privacy-Accuracy Trilemma | 1 + ...Reinforcement Learning with a Generative Model | 1 + ...ty for Model-Based Deep Reinforcement Learning | 1 + ...One-shot Neural Architecture Search with BONAS | 1 + ...eural networks with structural message-passing | 1 + ...tine Resilient Distributed Multi-Task Learning | 1 + ...arization via Auxiliary Causal Graph Discovery | 1 + ...tworks with Scalable and Consistent Estimation | 1 + ...ural Architecture Search for Image Restoration | 1 + ...t Embeddings from Narrated Instructional Video | 1 + ...sformer for Video-Text Representation Learning | 1 + .../COPT: Coordinated Optimal Transport on Graphs | 1 + ...g Sequential Data via Causal Optimal Transport | 1 + ...: Communication-efficient SGD with Error Reset | 1 + ...Learning on Distributionally Shifted Instances | 1 + ...cal Spatiotemporal Point Cloud Representations | 1 + ...able Regression using Maximum Mean Discrepancy | 1 + .../Calibrating CNNs for Lifelong Learning | 1 + ...ibrating Deep Neural Networks using Focal Loss | 1 + ... General Sum Partially Observable Markov Games | 1 + ...Can Graph Neural Networks Count Substructures? | 1 + ...ess with Unlabeled Data and Bayesian Inference | 1 + ...Stochastic Convex Optimization as a Case Study | 1 + ...alizable Branching Heuristic for a SAT Solver? | 1 + ...ning Learn Representation? A Mean-Field Theory | 3 +++ ... Backpropagation in Predictive Coding Networks | 1 + ...ense weakly-supervised category reconstruction | 1 + ...caded Text Generation with Markov Transformers | 1 + ...Unknown Targets: Characterization and Learning | 1 + ...usal Discovery in Physical Systems from Videos | 1 + .../Causal Estimation with Functional Confounders | 1 + ...on for Weakly-Supervised Semantic Segmentation | 1 + ...plain Individual Predictions of Complex Models | 2 ++ .../Causal analysis of Covid-19 Spread in Germany | 1 + ...y Robust Detection of Out-of-Distribution Data | 1 + ...Image Transformations via Randomized Smoothing | 1 + .../neurips/Certified Monotonic Neural Networks | 1 + ...Graph Classification under Topological Attacks | 1 + ...Certifying Confidence via Randomized Smoothing | 1 + .../Certifying Strategyproof Auction Networks | 1 + ...Optimism: Volume Analysis of Learning in Games | 3 +++ ...licies: Where to Intervene and What to Observe | 1 + ... of candidate learning rules for deep networks | 1 + data/2020/neurips/Choice Bandits | 1 + ... Adversarial Learning across Spherical Circles | 1 + ...s, Generalized Linear Models, and Evolvability | 1 + ...lassification with Valid and Adaptive Coverage | 1 + ...ntization Gap: PixelCNN as a Single-Layer Flow | 1 + data/2020/neurips/Co-Tuning for Transfer Learning | 1 + ...xposure Maximization in Online Social Networks | 1 + ...ution Networks for Co-Salient Object Detection | 1 + ...ltimodal Image Representation for Registration | 1 + .../neurips/CoSE: Compositional Stroke Embeddings | 1 + ...For Function-Level Binary Source Code Matching | 1 + ...Matrix Multiplication For Straggler Mitigation | 1 + .../neurips/CogLTX: Applying BERT to Long Texts | 1 + ...sign for COVID-19 Using Deep Generative Models | 1 + ...erarchical Multi-Label Classification Networks | 1 + ...ICE: Off-Policy Confidence Interval Estimation | 1 + ...actical Private Mean and Covariance Estimation | 1 + ...anguage GANs with Cautious Sampling Strategies | 1 + ...heir Application to Public Health Intervention | 1 + data/2020/neurips/Collegial Ensembles | 1 + ...usions: A Statistics-based Computational Model | 1 + ...ing and Search for Imperfect-Information Games | 1 + ...evolving graphs with a dynamical Bethe-Hessian | 1 + ...dinality semidefinite programming\342\200\251" | 1 + ...ormative model for higher-order brain activity | 1 + .../neurips/Comparator-Adaptive Convex Bandits | 1 + ...Understanding Gradient Flow in Phase Retrieval | 1 + .../neurips/Compositional Explanations of Neurons | 1 + ...eralization by Learning Analytical Expressions | 1 + ...eralization via Neural-Symbolic Stack Machines | 1 + ...nal Visual Generation with Energy Based Models | 1 + ...ing via Fine-Grained Dense Feature Composition | 1 + ...llation for Weakly-Supervised Object Detection | 1 + ...t Representations with Relative Entropy Coding | 1 + ... Selective Inference using Dynamic Programming | 1 + ...ve Information-Theoretic Generalization Bounds | 1 + ...nce sequences for sampling without replacement | 1 + ...ormal Symplectic and Relativistic Optimization | 1 + ...ion in Infinite-Horizon Reinforcement Learning | 1 + ...timization over Positive Semidefinite Matrices | 1 + data/2020/neurips/Consequences of Misaligned AI | 1 + ... Q-Learning for Offline Reinforcement Learning | 1 + ...r Certified Robustness of Smoothed Classifiers | 1 + ...etric Mixture Models from Grouped Observations | 1 + ...sifiers for Complex Objectives and Constraints | 1 + ...l Relation Learning for Zero-Shot Segmentation | 1 + ...re selection for analytic deep neural networks | 2 ++ ... for Compressed Sensing with Generative Priors | 1 + ...arning in concave-convex and knapsack settings | 1 + ...rence with Geometric Jensen-Shannon Divergence | 1 + ... and Coordination in Recommendation Ecosystems | 1 + ...es: Multi-Agent Learning with Side Information | 1 + ...tion in Auctions via Mixed Integer Programming | 1 + ...by Functional Regularisation of Memorable Past | 1 + ...nual Learning in Low-rank Orthogonal Subspaces | 1 + ...l Primitives : Skill Discovery via Reset-Games | 1 + ...Mixed Sequence of Similar and Dissimilar Tasks | 1 + ...nce based Adaptive Group Sparse Regularization | 1 + .../Continuous Meta-Learning without Tasks | 1 + ...View Synthesis without Target View Supervision | 1 + ...Continuous Regularized Wasserstein Barycenters | 1 + ...bmodular Maximization: Beyond DR-Submodularity | 1 + data/2020/neurips/Continuous Surface Embeddings | 1 + ...tive Learning for Conditional Image Generation | 1 + ...Contrastive Learning with Adversarial Examples | 1 + ...al image segmentation with limited annotations | 1 + ...oving BERT with Span-based Dynamic Convolution | 1 + ... Convolutional Networks on Large Random Graphs | 1 + ...sk-Specific Adaptation over Partial Parameters | 1 + ...tion based on global lower second-order models | 1 + ...Convolutional Generation of Textured 3D Meshes | 1 + ...Tensor-Train LSTM for Spatio-Temporal Learning | 1 + ...tive Heterogeneous Deep Reinforcement Learning | 1 + .../Cooperative Multi-player Bandit Optimization | 1 + .../neurips/Coresets for Near-Convex Functions | 2 ++ .../Coresets for Regressions with Panel Data | 1 + ...g of Deep Neural Networks against Noisy Labels | 0 ...imization for Continual Learning and Streaming | 1 + .../Correlation Robust Influence Maximization | 1 + ...ence learning via linearly-invariant embedding | 1 + ...e-Guided Learning of Monotonic Neural Networks | 1 + ...or Weakly-Supervised Vision-Language Grounding | 1 + ...a Augmentation using Locally Factored Dynamics | 1 + ...Counterfactual Prediction for Bundle Treatment | 1 + ...rfactual Predictions under Runtime Confounding | 1 + ...nd-Language Navigation: Unravelling the Unseen | 1 + ...rks Are Universal Diffeomorphism Approximators | 1 + ... Paths For One-Shot Neural Architecture Search | 1 + data/2020/neurips/Critic Regularized Regression | 1 + ...raph Neural Network for Image Super-Resolution | 1 + ...trieval for Iterative Self-Supervised Training | 1 + ...validation Confidence Intervals for Test Error | 1 + ...ransformers: spatially-aware few-shot transfer | 1 + ...tructured Bandits Beyond Asymptotic Optimality | 1 + data/2020/neurips/Curriculum By Smoothing | 1 + ...rriculum Learning by Dynamic Instance Hardness | 1 + ...for multilevel budgeted combinatorial problems | 1 + ...ation to Prevent Distortion in Graph Embedding | 1 + ... Self-Supervised Video Representation Learning | 1 + ...us Optimization for Learning Bayesian Networks | 1 + ...: Learning local features with policy gradient | 1 + ...es for Enhanced Robust Generation of Ensembles | 1 + ... Continual Learning: a Strong, Simple Baseline | 1 + ...Simple Strategy For Neural Machine Translation | 1 + ...ing Text by Fingerprinting Language Generation | 1 + data/2020/neurips/Debiased Contrastive Learning | 1 + ...stic Gradient Descent to handle missing values | 1 + ... Surrogate Sketching and Scaled Regularization | 1 + .../Debugging Tests for Model Explanations | 1 + ...tralized Accelerated Proximal Gradient Descent | 1 + ...alized Langevin Dynamics for Bayesian Learning | 1 + ...ion Approximation and its Finite-Time Analysis | 1 + ...o characterize their generalization properties | 1 + ...on-Making with Auto-Encoding Variational Bayes | 1 + ...terfactual Explanations and Strategic Behavior | 1 + data/2020/neurips/Deep Archimedean Copulas | 1 + data/2020/neurips/Deep Automodulators | 1 + ...iant Wasserstein Distributional Classification | 1 + .../2020/neurips/Deep Direct Likelihood Knockoffs | 1 + ...Energy-based Modeling of Discrete-Time Physics | 1 + data/2020/neurips/Deep Evidential Regression | 1 + ...phical model for improved animal pose tracking | 1 + ...ion Learning for Bimanual Robotic Manipulation | 1 + .../Deep Metric Learning with Spherical Embedding | 1 + .../Deep Multimodal Fusion by Channel Exchanging | 1 + ...d Particle Filters for Time Series Forecasting | 1 + ...ed Hierarchical Attention for Text-based Games | 1 + .../Deep Reinforcement and InfoMax Learning | 1 + ...odeling via Graph Poisson Gamma Belief Network | 1 + ...ed Shape Correspondence with Optimal Transport | 1 + ...ep Smoothing of the Implied Volatility Surface | 1 + data/2020/neurips/Deep Statistical Solvers | 1 + ... Models for Tractable Counterfactual Inference | 1 + ...eep Subspace Clustering with Data Augmentation | 1 + .../Deep Transformation-Invariant Clustering | 1 + .../neurips/Deep Transformers with Latent Depth | 1 + .../Deep Variational Instance Segmentation | 1 + ...iener Meets Deep Learning for Image Deblurring | 1 + ...ive inference agents using Monte-Carlo methods | 1 + ...he time evolution of the Neural Tangent Kernel | 1 + ...ruction of strange attractors from time series | 1 + ...to-Image Translation by Transferring from GANs | 1 + ...nerative Network for Vector Graphics Animation | 1 + ...Learned Spectral Total Variation Decomposition | 1 + ...nd Cooperation in Nonstochastic Linear Bandits | 1 + ...l Networks using Structured Response Jacobians | 1 + ...m in Semi-supervised Video Object Segmentation | 1 + ...ural population data from multiple brain areas | 1 + ...Demystifying Orthogonal Monte Carlo and Beyond | 1 + ... A Provable Defense for Pretrained Classifiers | 1 + .../Denoising Diffusion Probabilistic Models | 1 + ...rning Transformation Synchronization on Graphs | 1 + .../neurips/Depth Uncertainty in Neural Networks | 1 + .../Design Space for Graph Neural Networks | 1 + ...s and Recognizing Physical Contact in the Wild | 1 + ... from Neural Networks via Topological Analysis | 0 ...rtified Object Detection with Median Smoothing | 0 ...imization over a Matroid in Nearly Linear Time | 1 + ...a: Learning Visual Dialog Agents from VQA Data | 1 + ...tial Operators and Algebraic Multigrid Pooling | 1 + ...e Augmentation for Data-Efficient GAN Training | 1 + ...able Causal Discovery from Interventional Data | 1 + ...Parallel Multi-Objective Bayesian Optimization | 1 + ...ifferentiable Meta-Learning of Bandit Policies | 1 + ... Equivalent Space with Exploration Enhancement | 1 + .../Differentiable Top-k with Optimal Transport | 1 + ...Private Clustering: Tight Approximation Ratios | 2 ++ ...ifferentially-Private Federated Linear Bandits | 1 + .../Digraph Inception Convolutional Networks | 1 + ...o Modern Deep Learning Tasks and Architectures | 1 + ...mization of Policies in Discrete Action Spaces | 1 + .../Directional Pruning of Deep Neural Networks | 1 + ...nal convergence and alignment in deep learning | 1 + .../Dirichlet Graph Variational Autoencoder | 1 + ...Gradient Estimator for Binary Latent Variables | 1 + ...forcement Learning via Distribution Correction | 1 + .../Discovering Reinforcement Learning Algorithms | 1 + ...odels from Deep Learning with Inductive Biases | 1 + ...covering conflicting groups in signed networks | 1 + ...ation via Self-supervised Audiovisual Matching | 1 + ...Ground Truth in Segmentation of Medical Images | 1 + .../neurips/Disentangling by Subspace Diffusion | 1 + ... Learning for Accurate Optical Flow Estimation | 1 + data/2020/neurips/Dissecting Neural ODEs | 1 + ...ral Networks for Graph Representation Learning | 1 + ...istributed Distillation for On-Device Learning | 1 + ... Communicate Less and Resist Byzantine Workers | 1 + ...ta: Bridging Median- and Mean-Based Algorithms | 1 + ...-label for Imbalanced Semi-supervised Learning | 1 + .../Distribution Matching for Crowd Counting | 1 + ...ion sets, confidence intervals and calibration | 1 + ...with IPMs and links to Regularization and GANs | 1 + .../Distributionally Robust Federated Averaging | 1 + ...st Local Non-parametric Conditional Estimation | 1 + ...obust Parametric Maximum Likelihood Estimation | 1 + ...ioning with Context-Object Split Latent Spaces | 1 + ...versification for White- and Black-box Attacks | 1 + ...e Bayesian Optimization With Batch Evaluations | 1 + ...rially Robust ImageNet Models Transfer Better? | 1 + ...tion Learning Help Neural Architecture Search? | 1 + ... as a Problem of Inference on Graphical Models | 1 + ...tribution Matching and Generalized Label Shift | 1 + ...fication with Linear-Dependency Regularization | 2 ++ ...main Generalization via Entropy Regularization | 1 + ...Gradient Estimation for Deterministic Policies | 1 + .../neurips/Dual Instrumental Variable Regression | 1 + ...ense against Lp and non-Lp Adversarial Attacks | 1 + ... for Transition Matrix in Label-noise Learning | 1 + ...ntralized Optimization with Variance Reduction | 1 + .../Dual-Resolution Correspondence Networks | 1 + ...Factorization Based Knowledge Graph Completion | 1 + ...RT: Dynamic BERT with Adaptive Width and Depth | 1 + ...nd Verbal Narrations in Knowledge-rich Domains | 1 + .../Dynamic Regret of Convex and Smooth Functions | 1 + ...cy Optimization in Non-Stationary Environments | 1 + data/2020/neurips/Dynamic Submodular Maximization | 1 + ...ted memory resources in reinforcement learning | 1 + ...ent descent in Gaussian mixture classification | 1 + ...rization Prevents Memorization of Noisy Labels | 1 + ...s Under Extreme Budget and Network Constraints | 1 + ... Faster Regularized Least-Squares Optimization | 1 + ...ity in Population Based Reinforcement Learning | 1 + ...ms for Device Placement of DNN Graph Operators | 1 + ...d On A Unified View Of $K$-means And Ratio-cut | 1 + ...r Stretched Mixtures: Landscape and Optimality | 1 + ...ent Contextual Bandits with Continuous Actions | 1 + ...ed High-Dimensional Distributions via Learning | 6 ++++++ ...xact Verification of Binarized Neural Networks | 1 + ...inforcement Learning via Bayesian Optimization | 1 + ... Objects with Constrained Adversarial Networks | 1 + ...fficient Learning of Discrete Graphical Models | 1 + ...ve Models via Finite-Difference Score Matching | 1 + ...sian Variational Inference for Neural Networks | 1 + ...e and Structured Latent Variables via Sparsity | 1 + ... through Optimistic Policy Search and Planning | 1 + ...ian Optimization via One-Shot Multi-Step Trees | 1 + ... Dimensionality Reduction via Gradient Descent | 1 + ...e MDPs with Weak Linear Function Approximation | 1 + ...tion-free Algorithms for Saddle Point Problems | 1 + ...parse Deep Learning with Theoretical Guarantee | 1 + ...sparse halfspaces with arbitrary bounded noise | 1 + ... of neural tuning during naturalistic behavior | 1 + ...ased inference for binary and multi-class MRFs | 1 + ...presentation Learning in Class-Imbalanced Data | 1 + ...: Protecting SignSGD against Byzantine Attacks | 1 + ...t Transfer via Unsupervised Environment Design | 1 + ...n from Randomized Uncertain Social Preferences | 1 + .../Empirical Likelihood for Contextual Bandits | 1 + ... via memory-efficient semidefinite programming | 1 + .../End-to-End Learning and Intervention in Games | 1 + .../Energy-based Out-of-distribution Detection | 1 + ... for Robust Model Fusion in Federated Learning | 2 ++ ...ophysical models with Bayesian Neural Networks | 1 + .../Ensuring Fairness Beyond the Training Data | 1 + ...nce: Identifiability and Finite Sample Results | 1 + ...Unbalanced Gaussian Measures has a Closed Form | 1 + ...ergence of iterative methods for eigenproblems | 1 + ...uivariant Networks for Hierarchical Structures | 0 ...ework for Combinatorial Optimization on Graphs | 1 + ... Bounds of Imitating Policies and Environments | 1 + ...int Faster under Interpolation-like Conditions | 1 + .../Escaping the Gravitational Pull of Softmax | 1 + ...ural Representations of Uncertain Environments | 1 + ...rom Heavy-Tailed Noise via Self-Avoiding Walks | 1 + ...ing Data Influence by Tracing Gradient Descent | 1 + ...ability with polylogarithmic sample complexity | 2 ++ ...ventions using Generative Adversarial Networks | 1 + .../Estimating weighted areas under the ROC curve | 1 + ...mation of Skill Distribution from a Tournament | 1 + ...aluating Attribution for Graph Neural Networks | 1 + ...g Teamwork Using Cooperative Game Abstractions | 1 + ...with Hybrid-Cylindrical-Spherical Voxelization | 1 + ...Spaces in Conditional Variational Autoencoders | 1 + ...y Prediction with Dynamic Relational Reasoning | 1 + ...al Planning for Vision-and-Language Navigation | 1 + .../Evolving Normalization-Activation Layers | 1 + ... of Mangled Clusters with Same-Cluster Queries | 1 + ...cit regularization via surrogate random design | 1 + ... the Local Lipschitz Constant of ReLU Networks | 1 + .../Exchangeable Neural ODE for Set Modeling | 1 + data/2020/neurips/Exemplar Guided Active Learning | 1 + ...rest Neighbor Retrieval, and Data Augmentation | 1 + ...zation to Train Compact Convolutional Networks | 1 + ...imental design for MRI by greedy policy search | 1 + ...ing for Offline Policy Learning and Evaluation | 1 + data/2020/neurips/Explainable Voting | 1 + ...ear Classifiers with Polynomial Time and Delay | 1 + ...it Regularisation in Gaussian Noise Injections | 1 + ...ative-free Optimization and Continuous Bandits | 1 + ... Fair and Transferable Representation Learning | 1 + ...rogate Gap in Online Multiclass Classification | 1 + ...ual patterns to learn from partial annotations | 1 + ...radient Methods with Variable Stepsize Scaling | 1 + ...st Neighbour and Its Improved Convergence Rate | 1 + ...y and Representation Learning of Low Rank MDPs | 1 + data/2020/neurips/Factor Graph Grammars | 1 + data/2020/neurips/Factor Graph Neural Networks | 1 + .../Factorizable Graph Convolutional Networks | 1 + ...ocesses: K-Shot Prediction of Neural Responses | 1 + data/2020/neurips/Fair Hierarchical Clustering | 2 ++ ...ple Decision Making Through Soft Interventions | 1 + .../neurips/Fair Performance Metric Elicitation | 1 + ... and recalibration with statistical guarantees | 1 + .../Fair regression with Wasserstein barycenters | 1 + ... help exact inference in structured prediction | 1 + ...bmodular Maximization: Algorithms and Hardness | 1 + ...verlapping Groups; a Probabilistic Perspective | 0 ...hics through Adversarially Reweighted Learning | 1 + ...Faithful Embeddings for Knowledge Base Queries | 1 + ...con: Fast Spectral Inference on Encrypted Data | 1 + ... Maximization Subject to a Knapsack Constraint | 1 + ...namics on Manifold: Geodesics meet Log-Sobolev | 1 + ...Distributionally Robust Support Vector Machine | 1 + data/2020/neurips/Fast Fourier Convolution | 1 + ...o Gaussian Processes and Bayesian Optimization | 1 + .../Fast Transformers with Clustered Attention | 1 + .../Fast Unbalanced Optimal Transport on a Tree | 1 + ...nd Accurate $k$-means++ via Rejection Sampling | 1 + ... Temporal Point Processes with Triangular Maps | 1 + ...Fast geometric learning with symbolic matrices | 1 + ...ls for Tabular Data via Augmented Distillation | 1 + ...aster DBSCAN via subsampled similarity queries | 1 + ...ergence Analysis of Discretized Langevin MCMC" | 1 + ...or Point Methods for Tall Wide Linear Programs | 1 + ...stance Estimation with the Sinkhorn Divergence | 1 + .../Feature Importance Ranking for Deep Learning | 1 + ...ave Shifted via Conditional Distribution Tests | 1 + ...hmic framework for fast federated optimization | 1 + ...erated Accelerated Stochastic Gradient Descent | 1 + ...ed Bayesian Optimization via Thompson Sampling | 1 + .../Federated Principal Component Analysis | 1 + ...ject Detection with Adversarial-Paced Learning | 1 + ...e Generation with Elastic Weight Consolidation | 1 + ...ning with Meta-Analogical Contrastive Learning | 1 + ...etric Learning Perspective Using Fewer Proxies | 1 + ...mes: Continuous Time Analysis and Applications | 2 ++ ...wise Learning for Multi-field Categorical Data | 1 + ... Behavioral Cloning from Observation Histories | 1 + ...vex Linearly Constrained Optimization Problems | 1 + ...gy of Decision Boundaries with Active Learning | 1 + ...Fine-Grained Dynamic Head for Object Detection | 1 + ...c Reconstruction via Biodiversity Optimization | 1 + data/2020/neurips/Finite Continuum-Armed Bandits | 1 + ...s Infinite Neural Networks: an Empirical Study | 1 + .../Finite-Time Analysis for Double Q-learning | 1 + ...tion with Multiple Plays and Markovian Rewards | 1 + ...a General Approach for Growing Neural Networks | 1 + ...Order Constrained Optimization in Policy Space | 1 + ...for Large-Scale Market Equilibrium Computation | 1 + ...vised Learning with Consistency and Confidence | 1 + ...ers: Computational Hardness and Fast Algorithm | 1 + .../FleXOR: Trainable Fractional Quantization | 1 + ...xtures of non-overlapping exponential families | 1 + ...neous manifold learning and density estimation | 1 + ...proves Information Transfer in Visual Features | 1 + ...t Parallel Algorithms for Smooth Minimax Games | 1 + ...Forethought and Hindsight in Credit Assignment | 1 + ... Depth Estimators with MED Probability Volumes | 1 + ...Frequency Functions in Low Dimensional Domains | 1 + ...everage Scores and Approximate Kernel Learning | 3 +++ ...Discrepancies in Deep Network Generated Images | 1 + ...stability of deep learning models for genomics | 1 + ...rally and Spatially for Efficient DNN Training | 1 + ...ann Machines to Neural Networks and Back Again | 1 + ...s to Decisions: Using Lookahead Regularization | 1 + ...s and Back: Hyperbolic Hierarchical Clustering | 1 + ...ML Prediction APIs more accurately and cheaply | 1 + ...oder using Efficient Spatially Varying Kernels | 1 + ...orithm for Constrained Submodular Optimization | 1 + ...on Learning: A Unified Theoretical Perspective | 1 + ...l Redundancy for Efficient Language Processing | 1 + ... Outlier Detection with Deep Generative Models | 1 + ...ing rule derived from backpropagation of error | 1 + data/2020/neurips/GAN Memory with No Forgetting | 1 + ...NSpace: Discovering Interpretable GAN Controls | 1 + ...ing \"When to Sample\" from \"How to Sample\"" | 1 + ...inatorial Algorithms over Billion-sized Graphs | 1 + ...ph Neural Networks against Adversarial Attacks | 1 + ...orrespondence Volumes into Your Neural Network | 1 + ...PS-Net: Graph-based Photometric Stereo Network | 1 + ... for Extremely Fast Large-Scale Classification | 1 + ...e Radiance Fields for 3D-Aware Image Synthesis | 1 + ...for Learning Differentially Private Generators | 1 + ...rence Learning for Infinite-Horizon Prediction | 1 + data/2020/neurips/Gaussian Gated Linear Networks | 1 + ...ion of the Thermodynamic Variational Objective | 1 + ...unctions for Causal Effect Estimation from IVs | 0 ...ty of Soft Interventions: Completeness Results | 1 + ... Bayesian Filtering via Sequential Monte Carlo | 1 + ...radient Descent for Non-Convex Metric Learning | 1 + ...tion by infinite dimensional Langevin dynamics | 1 + ...proaching Bayes error with convex optimization | 1 + data/2020/neurips/Generalized Boosting | 0 ...uted Bounding Boxes for Dense Object Detection | 1 + ...neralized Hindsight for Reinforcement Learning | 1 + ...n for Estimating Latent Variable Causal Graphs | 1 + ...ed Leverage Score Sampling for Neural Networks | 4 ++++ ...ubgoals in Hierarchical Reinforcement Learning | 1 + ...rs for Progressive Matrices Intelligence Tests | 1 + ...ve 3D Part Assembly via Dynamic Graph Learning | 1 + .../neurips/Generative Neurosymbolic Machines | 1 + ...rom Single-view Semantics to Novel-view Images | 1 + ...e causal explanations of black-box classifiers | 1 + ...Functions for Single-view Human Reconstruction | 1 + ...Geometric All-way Boolean Tensor Decomposition | 1 + ...metric Dataset Distances via Optimal Transport | 1 + .../Geometric Exploration for Online Control | 1 + data/2020/neurips/Gibbs Sampling with People | 1 + ...ing Spatial Redundancy in Image Classification | 1 + ...Class of Nonconvex-Nonconcave Minimax Problems | 1 + ... One Wide Layer Followed by Pyramidal Topology | 1 + ... Text-to-Speech via Monotonic Alignment Search | 1 + ...raining Deep Neural Networks on Encrypted Data | 2 ++ ... Structures with Conditional Generative Models | 1 + ...Regularization Method for Deep Neural Networks | 1 + .../neurips/Gradient Boosted Normalizing Flows | 1 + ...ient Estimation with Stochastic Softmax Tricks | 1 + ...rized V-Learning for Dynamic Treatment Regimes | 1 + .../Gradient Surgery for Multi-Task Learning | 1 + .../neurips/Gradient-EM Bayesian Meta-Learning | 1 + ...o Unsupervised Graph Matching Network Learning | 1 + ...N: Deep 3D Texture Synthesis From 2D Exemplars | 1 + ...aph Cross Networks with Vertex Infomax Pooling | 1 + .../neurips/Graph Geometry Interaction Learning | 1 + data/2020/neurips/Graph Information Bottleneck | 1 + .../Graph Meta Learning via Local Subgraphs | 1 + ...ork for Transferable Active Learning on Graphs | 1 + ...etworks for Semi-Supervised Learning on Graphs | 1 + ...c Neural Networks for Semi-supervised Learning | 1 + ...d the Transferability of Graph Neural Networks | 1 + ...Solution for Visual Learning of Robotic Grasps | 1 + ...ogarithmic Number of Winning Tickets is Enough | 1 + ... inference with structure-exploiting lazy maps | 1 + ...ol: Distortion-Aware Sparse Adversarial Attack | 1 + .../Group Contextual Encoding for 3D Point Clouds | 1 + ...: Federated Learning of Large CNNs at the Edge | 1 + ...roup-Fair Online Allocation in Continuous Time | 1 + ... Evaluating and Enhancing Adversarial Defenses | 1 + ...olecular Optimization with Genetic Exploration | 1 + ...naptic plasticity with Hebbian Memory Networks | 1 + ...trace-Weighted Quantization of Neural Networks | 1 + ...earest Neighbor Search on Heterogeneous Memory | 1 + ...ating and Decomposing Human-Object Interaction | 1 + ...HRN: A Holistic Approach to One Class Learning | 1 + ...: Pruning Adversarially Robust Neural Networks | 1 + ...nference for latent Gaussian models and beyond | 0 ...issing Data with Graph Representation Learning | 1 + ...sis for Cross-domain Shape Similarity Learning | 1 + .../Hard Negative Mixing for Contrastive Learning | 1 + .../Hard Shape-Constrained Kernel Machines | 1 + ... Learning Neural Networks with Natural Weights | 1 + ...y Tails, and Generalization in Neural Networks | 1 + ...xt Polarity Classification & Data Augmentation | 1 + ...aster convergence of external and swap regrets | 4 ++++ data/2020/neurips/Heuristic Domain Adaptation | 1 + ...r Efficient and High Fidelity Speech Synthesis | 1 + ...ent Memory with Optimal Polynomial Projections | 1 + ...ess Priors for Bayesian Neural Network Weights | 1 + .../Hierarchical Granularity Transfer Learning | 1 + ...l Architecture Search for Deep Stereo Matching | 1 + ...Generating Diverse Videos from a Single Sample | 1 + ...g for Compositional Generalization in Language | 1 + .../neurips/Hierarchical Quantized Autoencoders | 1 + ...ierarchical nucleation in deep neural networks | 1 + ...or Exploratory Search in Morphogenetic Systems | 1 + ...n Optimization via Nested Riemannian Manifolds | 1 + ...wn Context Rewards using Bayesian Optimization | 1 + .../High-Dimensional Sparse Linear Bandits | 1 + .../High-Fidelity Generative Image Compression | 1 + .../neurips/High-Throughput Synchronous Deep RL | 1 + ...f deep neural network models of visual cortex" | 1 + ...correlated time series with latent confounders | 1 + ...r-Order Certification For Randomized Smoothing | 1 + ...r-Order Spectral Clustering of Directed Graphs | 1 + ...ction for Maximizing Expected Order Statistics | 2 ++ ...riminative features on deep network boundaries | 1 + ...udy of Deep Neural Network Explanation Methods | 1 + ...pharmaceutical Interventions against COVID-19? | 1 + ...air decisions fare in long-term qualification? | 1 + ...rpretable Attribution for Feature Interactions | 1 + ...eneralisation Ability of Deep Neural Networks? | 0 ...distinguish graphs with graph neural networks? | 1 + ...itial point worth in Low-rank Matrix Recovery? | 1 + ...verparameterized Convolutional Neural Networks | 1 + ... Action-Gradient-Estimator Policy Optimization | 1 + ...e Image to 3D Human via Cross-View Consistency | 1 + ...ith Hybrid Similarity Measure and Triplet Loss | 1 + .../neurips/Hybrid Models for Learning to Branch | 1 + ...inimax Problems with Nonconvex-Linear Function | 1 + ...ersolvers: Toward Fast Continuous-Depth Models | 1 + ...epresentations and Feature Attribution Mapping | 1 + ...nergy-Based Deep Models Based on Nonlinear ICA | 1 + ... Correlation Network for Co-Saliency Detection | 1 + ...alized Accelerated Augmented Lagrangian Method | 1 + ...nt Neural Architecture Search by Sparse Coding | 1 + ...nference Failure with Uncertainty-Aware Models | 1 + ...Learning Rules From Neural Network Observables | 1 + ...d Data using the Area Under the Margin Ranking | 1 + ...n activity with Gaussian process factor models | 1 + ... Algorithm Configuration from an Infinite Pool | 1 + ...ion: Initialization Scale vs Training Accuracy | 1 + ...Implicit Distributional Reinforcement Learning | 1 + data/2020/neurips/Implicit Graph Neural Networks | 1 + ...esentations with Periodic Activation Functions | 1 + .../neurips/Implicit Rank-Minimizing Autoencoder | 1 + ... Deep Learning May Not Be Explainable by Norms | 1 + ... Results for Grammar-Compressed Linear Algebra | 3 +++ ...rithms for Convex-Concave Minimax Optimization | 1 + ...lar Maximization via First-order Regret Bounds | 1 + ...lipping Algorithms for Non-convex Optimization | 1 + ...uarantees for k-means++ and k-means++ Parallel | 1 + ...for Incremental Autonomous Exploration in MDPs | 1 + ...es for Episodic Memory-based Lifelong Learning | 1 + ...ues for Training Score-Based Generative Models | 1 + ... Phylogenetic Inference with Normalizing Flows | 1 + ...Column Subset Selection and the Nystrom method | 1 + ...o-Augment via Augmentation-Wise Weight Sharing | 1 + ...bability Ratio Clipping and Sample Reweighting | 1 + ...forcement Learning with Mixture Regularization | 1 + ...proving Inference for Neural Image Compression | 1 + ...dentifiability in Probabilistic Box Embeddings | 1 + ... Tasks with Human Gaze-Guided Neural Attention | 1 + ...twork Training in Low Dimensional Random Bases | 1 + ... Sequential Decision Making and ML Predictions | 1 + ...-Constrained Kidney Exchange via Pre-Screening | 1 + ...y Bounds for (Natural) Actor-Critic Algorithms | 1 + ...ctor Technique with Renyi Differential Privacy | 1 + ... with accuracy versus uncertainty optimization | 1 + ...mmon corruptions by covariate shift adaptation | 1 + ...In search of robust measures of generalization | 1 + ... into Parallel Sequence Decoding with Adapters | 1 + ...Output Constraints in Bayesian Neural Networks | 1 + ...Reasoning Communication into Emergent Language | 1 + ...Methods for Competitive Reinforcement Learning | 1 + data/2020/neurips/Inductive Quantum Embedding | 1 + ...on for Cross-scenario 3D Human Pose Estimation | 1 + data/2020/neurips/Inference for Batched Bandits | 1 + ...ing learning rules from animal decision-making | 1 + ...ented Online Planning for Complex Environments | 1 + ...Information Maximization for Few-Shot Learning | 0 ...l Learning from Missing-Not-At-Random Feedback | 1 + ...tic Regret Bounds for Online Nonlinear Control | 1 + ...ion theoretic limits of learning a sparse rule | 1 + ...Task Selection for Meta-Reinforcement Learning | 1 + .../neurips/Input-Aware Dynamic Backdoor Attack | 1 + ...d Approximations to Profile Maximum Likelihood | 1 + data/2020/neurips/Instance Selection for GANs | 1 + ...based Generalization in Reinforcement Learning | 1 + ...via approximate inverse sensitivity mechanisms | 1 + data/2020/neurips/Instance-wise Feature Grouping | 1 + ...rning, Automatically Synthesize Fast Gradients | 1 + ...terferometer by a reinforcement learning agent | 1 + ...t Solving for LP-based prediction+optimisation | 1 + ... Speed Up Gradients Propagation in Neural ODEs | 1 + ...ble Sequence Learning for Covid-19 Forecasting | 1 + ...olicies from Heterogeneous User Demonstrations | 1 + ...ng fMRI responses to continuous natural speech | 1 + ...ent Architecture for Knowledge Graph Embedding | 1 + .../2020/neurips/Interventional Few-Shot Learning | 1 + ...for Calibration of Multi-Class Neural Networks | 1 + ...ocessing Methods for Debiasing Neural Networks | 1 + ...ducing Routing Uncertainty in Capsule Networks | 11 +++++++++++ data/2020/neurips/Inverse Learning of Symmetries | 1 + ...ially Observable Continuous Nonlinear Dynamics | 1 + ...arameterization: Revisiting the Gumbel-Softmax | 1 + ... is it to break privacy in federated learning? | 1 + ...anguage Models Using Causal Mediation Analysis | 1 + ...rizon RL More Difficult Than Short Horizon RL? | 1 + ...ient for Feature-based Reinforcement Learning? | 1 + ...ndispensable for training deep neural network? | 1 + ...al Networks: Better and Robust Node Embeddings | 1 + ...JAX MD: A Framework for Differentiable Physics | 1 + ...ntrastive Learning with Infinite Possibilities | 1 + ...agent Collaboration with Imperfect Information | 1 + data/2020/neurips/Joints in Random Forests | 1 + ...ep Multitask Models with Gradient Sign Dropout | 1 + ...ation Algorithm for $k$-center Fair Clustering | 1 + ...n for User Behavior Modeling in CTR Prediction | 1 + ... Estimator: Risk Prediction from Training Data | 1 + ...ressive Distillation for Adder Neural Networks | 1 + ... Roof: Handling Billions of Points Efficiently | 1 + ...ble 3-factor Hebbian learning in deep networks | 1 + ... Facial Expression and Action Unit Recognition | 1 + ...k Bound, Data Efficiency and Imperfect Teacher | 1 + ... Reinforcement Learning for Continuous Control | 1 + ...k for Single Image Super-resolution and Beyond | 1 + ...ard Better Generalization and Local Elasticity | 1 + ...from scratch with multi-modal self-supervision | 1 + ...ble signal propagation in feedforward networks | 1 + .../neurips/Language Models are Few-Shot Learners | 1 + ...proach for Multiscale Language Representations | 1 + ...Entity Relationship Graph for Agent Navigation | 1 + ... Imagine Goals in Curiosity Driven Exploration | 1 + ...mitation Learning for Robot Manipulation Tasks | 1 + ...or Vision-and-Language Representation Learning | 1 + ...thods for Distributionally Robust Optimization | 1 + data/2020/neurips/Latent Bandits Revisited | 1 + ...Analysis of High-Dimensional Neural Recordings | 1 + .../Latent Template Induction with Gumbel-CRFs | 1 + ...Models For Intrinsically Motivated Exploration | 1 + ... Systematically Reason Over Implicit Knowledge | 1 + ...Learnability with Indirect Supervision Signals | 1 + ...bout Objects by Learning to Interact with Them | 1 + ...for Interaction Exploration in 3D Environments | 1 + .../Learning Agent Representations for Ice Hockey | 1 + ...ugmented Energy Minimization via Speed Scaling | 1 + .../Learning Bounds for Risk-sensitive Learning | 1 + ...fects via Weighted Empirical Risk Minimization | 1 + ...ng Certified Individually Fair Representations | 1 + ...able Energy Surrogates for PDE Order Reduction | 1 + ...mpositional Rules via Neural Program Synthesis | 1 + ... from Irregularly-Sampled Partial Observations | 1 + ...ep Attribution Priors Based On Prior Knowledge | 1 + ...mable Tetrahedral Meshes for 3D Reconstruction | 1 + ...ble Programs with Admissible Neural Heuristics | 1 + ... Differential Equations that are Easy to Solve | 1 + ...odels via Auxiliary-variable Local Exploration | 1 + ... and Group Structure of Dynamical Environments | 1 + ...ed Representations of Videos with Missing Data | 1 + ...the Principle of Maximal Coding Rate Reduction | 1 + ...elief Graphs to Generalize on Text-Based Games | 1 + .../Learning Feature Sparse Principal Subspace | 1 + ...consistent with Local Contrastive Explanations | 2 ++ ... Structure With A Finite-State Automaton Layer | 1 + ...idance Rewards with Trajectory-space Smoothing | 1 + ...Cooperative Multi-Agent Reinforcement Learning | 1 + ...Topology-Varying Dense 3D Shape Correspondence | 1 + ...rred Communication for Multi-Agent Cooperation | 1 + ...ariances in Neural Networks from Training Data | 1 + .../Learning Invariants through Soft Unification | 1 + .../Learning Kernel Tests Without Data Splitting | 1 + ...Learning Latent Space Energy-Based Prior Model | 1 + ...earning Linear Programs from Optimal Decisions | 1 + .../Learning Loss for Test-Time Augmentation | 1 + ...d Implicitly via Explicit Heat-Kernel Learning | 1 + ...ication through Structured Attentive Reasoning | 1 + ...Target Coverage in Directional Sensor Networks | 1 + data/2020/neurips/Learning Mutational Semantics | 1 + ...ons of Multi-Object Scenes from Multiple Views | 1 + ...ions with the Decodable Information Bottleneck | 1 + .../Learning Parities with Neural Networks | 1 + ...g Physical Constraints with Neural Projections | 1 + ...sical Graph Representations from Visual Scenes | 1 + ...sentations from Audio-Visual Spatial Alignment | 1 + ...oltzmann Machines with Sparse Latent Variables | 1 + ... Knowledge with Reverse Reinforcement Learning | 1 + data/2020/neurips/Learning Rich Rankings | 1 + ...bust Decision Policies from Observational Data | 1 + ...box Optimization using Monte Carlo Tree Search | 1 + ...malization for Generative Adversarial Networks | 1 + ...aphical Models without Condition Number Bounds | 2 ++ ...Learning Sparse Prototypes for Text Generation | 1 + .../Learning Strategic Network Emergence Games | 1 + .../Learning Strategy-Aware Linear Classifiers | 1 + ...ons From Untrusted Batches: Faster and Simpler | 2 ++ ...lities and Equilibria in Non-Truthful Auctions | 1 + ...r drawing by efficient motor program induction | 1 + ...Learning by Minimizing the Sum of Ranked Range | 1 + ...al functions via multiplicative weight updates | 1 + ...g discrete distributions with infinite support | 1 + ...rete distributions: user vs item-level privacy | 3 +++ ...ndent representations with synaptic plasticity | 1 + .../neurips/Learning from Aggregate Observations | 1 + ...: De-biasing Classifier from Biased Classifier | 0 ... Proportions: A Mutual Contamination Framework | 1 + ...rom Mixtures of Private and Public Populations | 3 +++ ...d Unlabeled Data with Arbitrary Positive Shift | 1 + ... high-dimensional neural activity using pi-VAE | 1 + ...Discrete Graphical Models with Neural Networks | 1 + ...Black-Box: The pursuit of interpretable models | 1 + ...iologically plausible local wiring constraints | 1 + .../Learning the Geometry of Wave-Based Imaging | 1 + ...uadratic Regulator from Nonlinear Observations | 1 + .../neurips/Learning to Adapt to Evolving Domains | 1 + .../Learning to Approximate a Bregman Divergence | 1 + ...r Decoding of Sparse Graph-Based Channel Codes | 1 + ...hop Scheduling via Deep Reinforcement Learning | 1 + ...uction Pointer Attention Graph Neural Networks | 1 + ...sductive Few-shot Out-of-Graph Link Prediction | 1 + .../Learning to Incentivize Other Learning Agents | 1 + .../Learning to Learn Variational Semantic Memory | 1 + ...ng to Learn with Feedback and Local Plasticity | 1 + ...to Mutate with Hypergradient Guided Population | 1 + ...ent Surfaces by Self-supervised Spherical CNNs | 1 + ... Diplomacy with Best Response Policy Iteration | 1 + ...Play Sequential Games versus Unknown Opponents | 1 + ...rove Theorems by Learning to Generate Theorems | 1 + ...Forecast Tasks for Clinical Outcome Prediction | 0 ...ping Rewards: A New Approach of Reward Shaping | 1 + ...ficiently for causally near-optimal treatments | 1 + ... regularised problems with unrolled algorithms | 1 + .../Learning to summarize with human feedback | 0 ...plications to Variational and Ensemble methods | 1 + ...arning with Differentiable Pertubed Optimizers | 0 ...ued Kernels in Reproducing Kernel Krein Spaces | 1 + ...ning without Sparsity and Low-Rank Assumptions | 1 + ...kovian Data: Fundamental Limits and Algorithms | 2 ++ ...of KL Regularization in Reinforcement Learning | 1 + ...vex Optimization via Gradient-based Algorithms | 1 + ...t Need Complex Weight Posterior Approximations | 1 + ...olicies for Faster Training Without Forgetting | 1 + ...al Networks for Text-Guided Image Manipulation | 1 + ...n Detection Score For Variational Auto-encoder | 1 + ... on Testing Structural Changes in Ising Models | 1 + ...Limits to Depth Efficiencies of Self-Attention | 1 + ...esentations and Unsupervised Action Estimation | 1 + ...ical Systems as a Core Computational Primitive | 1 + ...e Sinkhorn Divergences using Positive Features | 1 + ...near-Sample Learning of Low-Rank Distributions | 1 + .../Linearly Converging Error Compensated SGD | 1 + ...rovably Robust Training by Laplacian Smoothing | 1 + ...-Certifiable Training with a Tight Outer Bound | 1 + ... Mean Estimation via Iterative Multi-Filtering | 1 + ...ning to Sounds of Silence for Speech Denoising | 1 + ...oCo: Local Contrastive Representation Learning | 1 + ...entially Private (Contextual) Bandits Learning | 1 + ...butions is faster using interactive mechanisms | 1 + ...Locally-Adaptive Nonparametric Online Learning | 1 + ...and Quantifiable Neural Distribution Alignment | 1 + .../neurips/Logarithmic Pruning is All You Need | 1 + ... Partially Observable Linear Dynamical Systems | 1 + ... with Goal-Conditioned Hierarchical Predictors | 1 + ...od and Removing the Bad Momentum Causal Effect | 1 + ... Pose and Shape for 3D Human Mesh Registration | 1 + ...-Resampling with Spatially Stochastic Networks | 1 + ... Awareness to Task Embedding for Meta Learning | 1 + .../MCUNet: Tiny Deep Learning on IoT Devices | 1 + ...ks: Group Symmetries in Reinforcement Learning | 1 + ...Ensemble Imbalanced Learning with MEta-SAmpler | 1 + ...ural Networks by Maximizing the Minimal Angles | 1 + .../MOPO: Model-based Offline Policy Optimization | 1 + ...eL: Model-Based Offline Reinforcement Learning | 1 + ...rmuted Pre-training for Language Understanding | 1 + .../MRI Banding Removal via Adversarial Training | 1 + ...Shot Video Object Segmentation Efficient Again | 1 + ...chastic Control (Almost) as Easy as Stochastic | 2 ++ ... non-Euclidean latent structure in neural data | 1 + .../Manifold structure in graph embeddings | 1 + ... in Continuous or Large Discrete Action Spaces | 1 + ... Insufficient for Explaining Gradient Boosting | 1 + ...etion with Hierarchical Graph Side Information | 1 + ...Inference and Estimation in Multi-Layer Models | 1 + ...rn Gaussian Processes on Riemannian Manifolds" | 1 + ...ion for Improved Generalization and Robustness | 1 + ...al Distribution Shifts in Image Classification | 1 + ...n in Neural Proof Generation with Transformers | 1 + ...oned Policies for Learning from Sparse Rewards | 1 + ...r Dynamical Systems for Prediction and Control | 1 + ...MeshSDF: Differentiable Iso-Surface Extraction | 4 ++++ .../Meta-Consolidation for Continual Learning | 3 +++ ...t Learning with an Objective Discovered Online | 1 + .../Meta-Learning Requires Meta-Augmentation | 1 + ...Prediction with Convolutional Neural Processes | 1 + ... through Hebbian Plasticity in Random Networks | 1 + .../Meta-Learning with Adaptive Hyperparameters | 1 + data/2020/neurips/Meta-Neighborhoods | 1 + ...from Tasks with Heterogeneous Attribute Spaces | 1 + ...-trained agents implement Bayes-optimal agents | 1 + ...izer for Heterogeneous Tasks and Architectures | 1 + ...cal General-purpose Clean-label Data Poisoning | 1 + ...taSDF: Meta-Learning Signed Distance Functions | 1 + ...ic-Free Individual Fairness in Online Learning | 1 + ...nd: Regularization, Approximation and Numerics | 1 + ...nostic Compression of Pre-Trained Transformers | 1 + ... Stochastic Approximate Proximal Point Methods | 1 + ...cal SGD for Heterogeneous Distributed Learning | 1 + .../Minimax Bounds for Generalized Linear Models | 1 + ...ation with 0-1 Loss and Performance Guarantees | 1 + ... Networks of Excitatory and Inhibitory Neurons | 1 + ...inimax Estimation of Conditional Moment Models | 1 + ...th Linear and One-hidden Layer Neural Networks | 1 + ... Estimation of Heterogeneous Treatment Effects | 1 + ...nline Convex Optimization: No Phase Transition | 1 + ... Off-Policy Evaluation and Policy Optimization | 4 ++++ ...l Learning via Instance-Aware Parameterization | 1 + ...eer Review via Randomized Reviewer Assignments | 1 + ...for Learning Models from Mixture Distributions | 1 + ...lo for Mixed Discrete and Continuous Variables | 1 + .../Model Agnostic Multilevel Explanations | 1 + .../Model Class Reliance for Random Forests | 1 + .../neurips/Model Fusion via Optimal Transport | 1 + ...sting Resolution, Depth and Width for TinyNets | 1 + ...uction System via Automated Online Experiments | 1 + ...ction in Contextual Stochastic Bandit Problems | 1 + ...rkov Games with Near-Optimal Sample Complexity | 1 + ...-based Adversarial Meta-Reinforcement Learning | 1 + ...ptimization with Unsupervised Model Adaptation | 1 + ...emi-Markov Decision Processes with Neural ODEs | 1 + ...astic Processes with Dynamic Normalizing Flows | 1 + .../Modeling Noisy Annotations for Crowd Counting | 1 + ... in Neuroimaging Studies through MultiView ICA | 3 +++ ...tion in the Brain via Zero-Shot MEG Prediction | 1 + ...ng and Optimization Trade-off in Meta-learning | 1 + ...Attention for Immune Repertoire Classification | 1 + .../neurips/Modular Meta-Learning with Shrinkage | 1 + ...rating Momentum into Recurrent Neural Networks | 1 + .../Monotone operator equilibrium networks | 1 + ...ment Pruning: Adaptive Sparsity by Fine-Tuning | 1 + ...Compression of LiDAR using Deep Entropy Models | 1 + ...Bayesian Optimization via Deep Neural Networks | 1 + ...lti-Plane Program Induction with 3D Box Priors | 1 + ...with Probabilistic Safety Barrier Certificates | 1 + data/2020/neurips/Multi-Stage Influence Function | 1 + ...einforcement Learning with Soft Modularization | 1 + ...s for On-Device Contactless Vitals Measurement | 1 + ...ajectory Prediction with Fuzzy Query Attention | 1 + ...gent active perception with prediction rewards | 1 + .../Multi-label Contrastive Predictive Coding | 1 + ...bset accuracy really conflict with each other? | 1 + ...t Estimation and Automatic Structure Discovery | 1 + ...ch Reinforcement Learning with Metric Learning | 1 + ...i-task Causal Learning with Gaussian Processes | 1 + ...antic Map Memory using Multi-Object Navigation | 1 + ...y Estimation for Label-Efficient Deep Learning | 1 + ...hical Partitioning and Data-dependent Grouping | 1 + ...e Learning Utilizing Jensen-Shannon-Divergence | 1 + ...al Generalization in Visual Question Answering | 1 + ...istence Image for Topological Machine Learning | 1 + ... for Parametric Partial Differential Equations | 1 + .../neurips/Multiscale Deep Equilibrium Models | 1 + ...ction by Disentangling Geometry and Appearance | 1 + .../neurips/Munchausen Reinforcement Learning | 1 + ...sivity as a challenge for deep neural networks | 1 + data/2020/neurips/Myersonian Regression | 1 + ...E: A Deep Hierarchical Variational Autoencoder | 1 + ...zing Flows with Sublinear Parameter Complexity | 1 + data/2020/neurips/Natural Graph Networks | 1 + ...thod for Constrained Markov Decision Processes | 1 + .../Near-Optimal Comparison Based Clustering | 1 + ...-Optimal Reinforcement Learning with Self-Play | 1 + ... Halfspaces and ReLUs under Gaussian Marginals | 1 + ...work Diffusions via Neural Mean-Field Dynamics | 1 + ...n memorization with two-layers neural networks | 0 ...on with Conditional Invertible Neural Networks | 1 + ...ng for supervised learning with missing values | 1 + data/2020/neurips/Neural Anisotropy Directions | 1 + .../Neural Architecture Generator Optimization | 1 + ... Evaluating Safety-Critical Autonomous Systems | 1 + data/2020/neurips/Neural Complexity Measures | 1 + ...fferential Equations for Irregular Time Series | 1 + ... Policies for End-to-End Sensorimotor Learning | 1 + ...ution Engines: Learning to Execute Subroutines | 1 + ...ral FFTs for Universal Texture Image Synthesis | 1 + ...eural Manifold Ordinary Differential Equations | 1 + ...nifold Mesh Generation via Diffeomorphic Flows | 1 + ...i-Relational Ordered and Recursive Hypergraphs | 1 + ...l Methods for Point-wise Dependency Estimation | 1 + ... to Learn Periodic Functions and How to Fix It | 1 + ...ization with (almost) no Over-Parameterization | 2 ++ ...al Networks with Recurrent Generative Feedback | 1 + ...th Small Weights and Depth-Separation Barriers | 1 + data/2020/neurips/Neural Non-Rigid Tracking | 1 + ...derstanding the role of gates in deep learning | 1 + data/2020/neurips/Neural Power Units | 1 + ...al Sparse Representation for Image Restoration | 1 + data/2020/neurips/Neural Sparse Voxel Fields | 1 + ...Neural Star Domain as Primitive Representation | 1 + ...ural Topographic Factor Analysis for fMRI Data | 1 + ...Distance Fields for Implicit Function Learning | 1 + .../neurips/Neural encoding with visual attention | 1 + ...euron Merging: Compensating for Pruned Neurons | 1 + ...n Shapley: Discovering the Responsible Neurons | 1 + ...uctured Pruning using Polarization Regularizer | 1 + .../neurips/Neuronal Gaussian Process Regression | 1 + ...nt Learning with Formally Verified Exploration | 1 + ...lic Transformers for Multi-Agent Communication | 1 + ... Self-Selection Bias in Sampling for Sortition | 1 + ...Parallel GPU Task Scheduling for Deep Learning | 1 + ...ness in Coarse-Grained Classification Problems | 1 + ...mics for Extensive-Form Correlated Equilibrium | 1 + ...ing and Mixed Nash Equilibria: They Do Not Mix | 1 + ... Competitions under Consumer Reference Effects | 1 + ... Labels via Meta Transformed Network Embedding | 1 + ...t Low-Rank Representations of Complex Networks | 2 ++ ...ve Estimation for Multivariate Point Processes | 1 + ...ng A Self-Supervised Bound for Image Denoising | 1 + ...Learns Halfspaces with Adversarial Label Noise | 1 + ...sion for Distributional Reinforcement Learning | 0 .../neurips/Non-Euclidean Universal Approximation | 1 + .../Non-Stochastic Control with Bandit Feedback | 1 + ...n-parametric Models for Non-negative Functions | 1 + ...ying latent dynamical structure in neural data | 1 + ... Spiked Matrix Recovery with Generative Priors | 1 + ...ng under Laplacian Constrained Graphical Model | 1 + ... Filters for Multivariate Time Series Analysis | 1 + ...ing to Weight Data in Semi-supervised Learning | 1 + ...ational Space for Sample Efficient Exploration | 1 + ...rtial Differential Equations via Deep Learning | 1 + ...iversal Approximability of Sparse Transformers | 1 + ...t-of-Distribution Detection and Classification | 1 + ... Optimal Transport Approach for Topic Modeling | 1 + ...ation using Goal-Oriented Semantic Exploration | 1 + .../Object-Centric Learning with Slot Attention | 1 + data/2020/neurips/Ode to an ODE | 1 + ... for External Validity under a Covariate Shift | 1 + ...licy Evaluation via the Regularized Lagrangian | 1 + ...ff-Policy Imitation Learning from Observations | 1 + ...rval Estimation with Lipschitz Value Iteration | 1 + ...uential Decisions Under Unobserved Confounding | 1 + ...itation Learning with a Misspecified Simulator | 1 + .../On 1 n neural representation and robustness | 1 + ...aptive Attacks to Adversarial Example Defenses | 1 + data/2020/neurips/On Adaptive Distance Estimation | 2 ++ ...ept-Based Explanations in Deep Neural Networks | 1 + ...ergence and Generalization of Dropout Training | 1 + ...ghbor Classifiers over Feature Transformations | 1 + ...fferentiation for Non-Differentiable Functions | 1 + ...iciency in Hierarchical Reinforcement Learning | 1 + data/2020/neurips/On Infinite-Width Hypernetworks | 1 + ...Ising Models under Huber's Contamination Model | 1 + .../neurips/On Numerosity of Deep Neural Networks | 1 + data/2020/neurips/On Power Laws in Deep Ensembles | 2 ++ .../neurips/On Regret with Multiple Best Arms | 1 + ...nt Learning with Linear Function Approximation | 1 + ...econd Order Behaviour in Augmented Neural ODEs | 1 + data/2020/neurips/On Testing of Samplers | 3 +++ ...onvergence and Low-Norm Interpolation Learning | 1 + .../On Warm-Starting Neural Network Training | 1 + ...king via sorting by estimated expected utility | 1 + ...hastic Gradient Descent in Non-Convex Problems | 1 + ...egularized Approximate Value Iteration Schemes | 1 + ...vate Learnability beyond Binary Classification | 1 + ...rmality of Randomized Midpoint Sampling Method | 1 + ...he Error Resistance of Hinge-Loss Minimization | 1 + ...roximate Inference in Bayesian Neural Networks | 1 + ...dentifying Challenges and How to Overcome Them | 1 + .../neurips/On the Modularity of Hypernetworks | 1 + ...Power of Louvain in the Stochastic Block Model | 1 + ...y and DAG Constraints for Learning Linear DAGs | 1 + ...between the Laplace and Neural Tangent Kernels | 1 + ...ning: A Case Study on Linear Quadratic Systems | 1 + ...fer Learning: The Importance of Task Diversity | 1 + ... Certifying Robustness to Adversarial Examples | 1 + ...ff between Adversarial and Backdoor Robustness | 1 + ...ribution Testing: An Example of Goodhart's Law | 1 + ... neural networks and the stability of learning | 1 + ...nd molecular wave function with poor basis set | 1 + ...s: when and why the tangent kernel is constant | 1 + ...deoff between Robustness and Accuracy for Free | 1 + ...ably Robust Geometric Perception with Outliers | 1 + ...ew-Shot Extrapolation via Structured MaxEnt RL | 1 + .../One-bit Supervision for Image Classification | 1 + ...ple Guided Object Representation Disassembling | 1 + ...line Agnostic Boosting via Regret Minimization | 3 +++ ...quential Selection with Contextual Information | 1 + ...ti-shop Ski Rental with Machine Learned Advice | 1 + ...ference for Boundedly Rational Planning Agents | 1 + data/2020/neurips/Online Bayesian Persuasion | 1 + ... Optimization Over Erdos-Renyi Random Networks | 1 + ...sed Visual Tracking via Reinforcement Learning | 1 + ... (OSAKA): a New Approach to Continual Learning | 1 + ...ence Maximization under Linear Threshold Model | 1 + ...Contextual Bandits using Gated Linear Networks | 1 + ...ine Learning with Primary and Secondary Losses | 1 + .../Online Linear Optimization with Many Hints | 1 + ...MAP Inference of Determinantal Point Processes | 1 + ...Online Matrix Completion with Side Information | 1 + ...c Learning for Off-Policy Actor-Critic Methods | 1 + ...nline Multitask Learning with Long-Term Memory | 1 + ...nnectivity Estimation with Noisy Group Testing | 1 + ...on-Convex Optimization with Imperfect Feedback | 1 + ...timization with Memory and Competitive Control | 1 + .../Online Planning with Lookahead Policies | 1 + ...nline Robust Regression via SGD on the l1 loss | 1 + ...ptimal Transport distances from sample streams | 1 + data/2020/neurips/Online Structured Meta-learning | 1 + ... learning with dynamics: A minimax perspective | 3 +++ ...hmark: Datasets for Machine Learning on Graphs | 1 + ... Maximize Simultaneously Recorded Neuron Yield | 1 + ... Multi-Armed Bandits with Heavy Tailed Rewards | 1 + ... - Smoothness Tradeoffs for Soft-Max Functions | 1 + ...imal Best-arm Identification in Linear Bandits | 1 + ...escent Ascent Methods for Min-Max Optimization | 1 + ...h the Subsampled Randomized Hadamard Transform | 1 + .../Optimal Learning from Verified Training Data | 1 + ...ogarithmic Over-Parameterization is Sufficient | 1 + ...the Number of Unseen Species with Multiplicity | 1 + ...ation under Minimal Distributional Assumptions | 1 + ...exity of Secure Stochastic Convex Optimization | 1 + ...-offs for Learning-Augmented Online Algorithms | 1 + ...dient Estimator for Importance-Weighted Bounds | 1 + ...and Strongly Convex Decentralized Optimization | 1 + ...ceiving a Learning Leader in Stackelberg Games | 1 + ...Coherent Non-monotone Variational Inequalities | 1 + ...plication to Multi-scale Graph Neural Networks | 1 + ...l Networks with Quadratic Activation Functions | 6 ++++++ ...imizing Mode Connectivity via Neuron Alignment | 1 + ...ng Neural Networks via Koopman Operator Theory | 1 + ... offering using an individual treatment effect | 1 + ... task-computation to enable continual learning | 1 + ...stimation with Subgaussian Rates via Stability | 1 + ...nalysis Overcoming the Curse of Dimensionality | 1 + ...rmless for Basis Pursuit, But Only to a Degree | 1 + .../PAC-Bayes Analysis Beyond the Usual Bounds | 2 ++ ...es Learning Bounds for Sample-Dependent Priors | 1 + ...yesian Bound for the Conditional Value at Risk | 1 + ...loration for Provable Policy Gradient Learning | 1 + .../PEP: Parameter Ensembling by Perturbation | 1 + ...l Model Explanations for Graph Neural Networks | 1 + ...NET: Parametric Inference of Point Cloud Edges | 1 + ...S: Neuro-Symbolic Program Learning from Videos | 1 + ...pological Layer based on Persistent Landscapes | 0 ...inuous Space MDPs with Non-Asymptotic Analysis | 1 + .../POMDPs in Continuous Time and Discrete Spaces | 1 + ...ith Multiple Optima for Reinforcement Learning | 1 + .../PRANK: motion Prediction based on RANKing | 1 + .../Parabolic Approximation Line Search for DNNs | 1 + ...rameterized Explainer for Graph Neural Network | 1 + ...ation for Unsupervised Visual Feature learning | 1 + ... Noise: Towards Instance-dependent Label Noise | 1 + .../neurips/Partially View-aligned Clustering | 1 + ...MRI with Self-Supervised Learning\342\200\213" | 1 + ...volution and Pooling for Graph Neural Networks | 1 + ...ient Estimators for Stochastic Binary Networks | 1 + ...ing penalty for smooth and log-concave targets | 1 + ...mechanism for differentially private selection | 1 + ...lized Federated Learning with Moreau Envelopes | 1 + ...ntees: A Model-Agnostic Meta-Learning Approach | 1 + ...ard and Strict Blackbox Attack Transferability | 1 + ...tatistical and computational phase transitions | 1 + ...ing Approximate Nash Equilibria in Large Games | 1 + ...tive for Domain Adaptive Semantic Segmentation | 1 + ...lanning With Sparse Rewards and Multiple Goals | 1 + ...Processes with Gap-Dependent Sample Complexity | 1 + ...bjective Functions: Going Beyond Total Rewards | 1 + ...ection in high-dimensional neural spike trains | 1 + data/2020/neurips/Pointer Graph Networks | 1 + ... Improvement via Imitation of Multiple Oracles | 1 + ...Form Games with Public Chance Moves and Beyond | 1 + ...: An End-to-End Learning and Control Framework | 1 + ...ed Gradient for Model Quantization and Pruning | 1 + ...erarchical Data Augmentation for Deep Networks | 1 + ...ut OOD Samples via Density-Based Pseudo-Counts | 2 ++ ...sterior Re-calibration for Imbalanced Datasets | 1 + ...ion Compression in Decentralized Deep Learning | 1 + ...ctical No-box Adversarial Attacks against DNNs | 1 + ...wton Methods for Training Deep Neural Networks | 1 + data/2020/neurips/Pre-training via Paraphrasing | 1 + ...: Low-rank approximation and randomized Newton | 1 + .../Predicting Training Time Without Training | 1 + .../Prediction with Corrupted Expert Advice | 1 + ...dictive Information Accelerates Learning in RL | 1 + ...d neural networks with noise, chaos and delays | 1 + ...ce is free with the jackknife+-after-bootstrap | 1 + ...ultiple criteria: A game-theoretic perspective | 1 + ...forcement Learning with Finite-Time Guarantees | 1 + ...roximal Stochastic Gradient Langevin Algorithm | 1 + ...Primal-Dual Mesh Convolutional Neural Networks | 1 + ...cipal Neighbourhood Aggregation for Graph Nets | 1 + .../Privacy Amplification via Random Check-Ins | 1 + ...ity Testing for High-Dimensional Distributions | 1 + ...onstruction and Reducing the Sample Complexity | 1 + .../neurips/Probabilistic Active Meta-Learning | 1 + ...ational Inference in Discrete Graphical Models | 1 + data/2020/neurips/Probabilistic Fair Clustering | 1 + ...heoretical Limits and Practical Approximations | 1 + ...babilistic Linear Solvers for Machine Learning | 1 + ...on Estimation with Matrix Fisher Distributions | 1 + ... Forecasting with Shape and Temporal Diversity | 0 ...bly Approximately Correct Constrained Learning | 1 + ...rnability and Compressibility of Distributions | 1 + ...Program Synthesis with Pragmatic Communication | 1 + .../Projected Stein Variational Gradient Descent | 1 + ...ethod and Optimal Nonsmooth Frank-Wolfe Method | 1 + ...sserstein Distance and Riemannian Optimization | 1 + ...ion in Multi-Agent Deep Reinforcement Learning | 1 + ...a a Simple and Efficient Regularization Method | 1 + ...on: Predicting Attention with Future Attention | 1 + ...of a Structured Tensor via Dictionary Learning | 1 + ...lapping Community Detection in Weighted Graphs | 1 + .../Provably Consistent Partial-Label Learning | 1 + ...forcement Learning Using Unsupervised Learning | 1 + ...tural Equation Models: An Adversarial Approach | 1 + ...y Efficient Neural GTD for Off-Policy Learning | 0 ...ter Optimization with Population-Based Bandits | 1 + ...with Kernel and Neural Function Approximations | 0 ...gnostic Navigation with Linear Value Iteration | 1 + ...inforcement Learning Without Great Exploration | 0 data/2020/neurips/Provably Robust Metric Learning | 1 + ...aptive reinforcement learning in metric spaces | 1 + .../Proximal Mapping for Deep Regularization | 1 + ...trix Perspective Function and its Applications | 1 + data/2020/neurips/Pruning Filter in Filter | 1 + ...y data by iteratively conserving synaptic flow | 1 + ...g at Cloud Scale with Microsoft Floating Point | 1 + ...lic Programming for Automated Machine Learning | 1 + ...l Concepts Emerging in Representation Learning | 1 + ... Measures: Beating the Curse of Dimensionality | 1 + ...for Wasserstein-Approximate Gaussian Processes | 1 + ...ation of Chaos for SGD in Wide Neural Networks | 1 + data/2020/neurips/Quantized Variational Inference | 1 + ...evant mechanism for sequential decision-making | 1 + ...on Attention Network for Semantic Segmentation | 1 + ...Transient Tasks for Continual Image Captioning | 1 + ...Scene Synthesis Using Structured Latent Spaces | 1 + ... Benchmarks for Offline Reinforcement Learning | 0 ...n-linear Pooling for RAM Constrained Inference | 1 + ... Sample-based Keypoint Detector and Descriptor | 1 + ... Data Augmentation with a Reduced Search Space | 1 + .../Random Reshuffling is Not Always Better | 1 + ...ffling: Simple Analysis with Vast Improvements | 1 + .../neurips/Random Walk Graph Neural Networks | 1 + ...ession: A more efficient and powerful solution | 1 + ...Projection Alternative to the Softmax Function | 1 + ...rmulation of Wasserstein Discriminant Analysis | 1 + data/2020/neurips/Rational neural networks | 1 + ...ngs for High-Dimensional Bayesian Optimization | 1 + .../Real World Games Look Like Spinning Tops | 1 + ...-Time Dynamical Systems using Polynomial Forms | 1 + ...ersarial Learning via Characteristic Functions | 1 + ...mization Analyses: The Intrinsic Learning Rate | 1 + ...rative Objectives For Counterfactual Reasoning | 1 + ...ages from Brain Activity by Shape-Semantic GAN | 1 + ...e linear classifiers from mixture of responses | 1 + .../neurips/Recurrent Quantum Neural Networks | 1 + ...ls for Multiple Interacting Neural Populations | 1 + ...cursive Inference for Variational Autoencoders | 1 + ...lly Robust Learning to Non-Robust PAC Learning | 1 + ...ability using Self-Supervised Object Proposals | 1 + ...sion with reject option and application to kNN | 1 + ...Online Learning with Relative-Lipschitz Losses | 2 ++ .../Regret in Online Recommendation Systems | 1 + ...s recover the principal components, eventually | 1 + ...Black-box Models for Improved Interpretability | 1 + ...rds Permutation Invariance In Recurrent Models | 2 ++ ...mization with Neighborhood-Controlled Grammars | 1 + ...Learning for Control with Multiple Frequencies | 1 + ...ter Regret Bounds for the Non-Episodic Setting | 1 + .../Reinforcement Learning with Augmented Data | 1 + ...ial Actions: An Application to Vehicle Routing | 1 + .../Reinforcement Learning with Feedback Graphs | 1 + ...fficient Approach via Bounded Eluder Dimension | 1 + ...enchmark for Grounding Spatial Relations in 3D | 1 + ...s for Object Detection via Transformer Decoder | 1 + ...he Jacobian term in unsupervised deep learning | 1 + ...e Graph Neural Networks via Robust Aggregation | 1 + ...ularization by Maximizing Functional Entropies | 1 + ...fication Meets Regression for Object Detection | 1 + ...ameterizing Mirror Descent as Gradient Descent | 1 + ...amics for Bayesian Learning on Large Datasets" | 1 + ... Outcomes to Optimize Individualized Treatment | 1 + ...escuing neural spike train models from bad MLE | 1 + ...ts Recurrent Kernels and Structured Transforms | 1 + ...ortable Deep Neural Networks without Shortcuts | 1 + ...havior Imitation and Extended Motion Synthesis | 1 + ...mplexity Algorithm for Online Restless Bandits | 1 + ...ative Information in Few-Shot Object Detection | 1 + ...ing for Deep Learning under Distribution Shift | 1 + ...able Tree Filter for Generic Feature Transform | 1 + .../Rethinking Pre-training and Self-training | 1 + .../Rethinking pooling in graph neural networks | 1 + ...Labels for Improving Class-Imbalanced Learning | 1 + ...d Generation for Knowledge-Intensive NLP Tasks | 1 + ...mpose Retrosynthesis Prediction Like A Chemist | 1 + ...ions to a hierarchical inference task for mice | 1 + ...Polytopes: Strict Complementarity and Sparsity | 1 + ...ing for Automatic Neural Channel Number Search | 1 + ...e Spectrum Approximation of Gaussian Processes | 1 + ...Propagation Using Graph Convolutional Networks | 1 + ...oice: A unifying formalism for reward learning | 1 + ...RL: Hindsight Inference for Policy Improvement | 1 + ...tions by Following Eigenvectors of the Hessian | 1 + .../Riemannian Continuous Normalizing Flows | 1 + ...g: Near-Optimal Risk-Sample Tradeoff in Regret | 1 + ...g Bias using Cumulative Distribution Functions | 1 + ...dversarial Perturbations on State Observations | 1 + ...bust Density Estimation under Besov IPM Losses | 1 + ...ust Disentanglement of a Few Factors at a Time | 1 + ...arning: The Case of Affine Distribution Shifts | 1 + ...stimation in Nearly-Matrix Multiplication Time | 1 + ...for Mixed Linear Regression with Small Batches | 1 + ... Reinforcement Learning with Model Uncertainty | 1 + ... Reweighting of the Graph Connection Laplacian | 1 + ...s in Generative Modeling and Domain Adaptation | 1 + ...ation for Fairness with Noisy Protected Groups | 1 + ...Persistence Diagrams using Reproducing Kernels | 1 + ...e-Training by Adversarial Contrastive Learning | 1 + ...obust Quantization: One Model to Rule Them All | 1 + ...earning Rates for Double Over-parameterization | 1 + ...atment Effects with Uncertainty Quantification | 1 + ...ia Adversarial training with Langevin Dynamics | 1 + .../Robust Sequence Submodular Maximization | 1 + ...nalysis and Width-Independent Schatten Packing | 1 + ...stimation Made Simple, via Regret Minimization | 3 +++ ...ust compressed sensing using generative models | 1 + ...bust large-margin learning in hyperbolic space | 1 + ...chastic Optimization for Variational Inference | 1 + ...trol of Linear Systems: beyond Quadratic Costs | 1 + ...tic Gradient Descent using Biased Expectations | 1 + ...sian Neural Networks to Gradient-Based Attacks | 1 + ...ty Detection to Random Geometric Perturbations | 1 + data/2020/neurips/Rotated Binary Neural Network | 1 + ...bal Representation Learning for 3D Point Cloud | 1 + ... Self-Attention via Sparse Adaptive Connection | 1 + ...ic Control for Reliable Neural Network Pruning | 1 + ...ce 3D Object Reconstruction from Static Images | 1 + ...oto-Translation Equivariant Attention Networks | 1 + ...pplications in Radar and Satellite Meteorology | 1 + ...mponent convexity and large epoch requirements | 1 + ...ced Network For Spatial Description Resolution | 1 + ...nknown dynamical systems with long-term memory | 1 + ...fficient Attention using Asymmetric Clustering | 1 + ...SOLOv2: Dynamic and Fast Instance Segmentation | 1 + ... Simple Temporal Regularization For Neural ODE | 1 + ...nforced Multivariate Recurrent Neural Networks | 1 + ..., Robust, Fast Distribution Learning Algorithm | 1 + ...in gradient flow of the chi-squared divergence | 1 + ...einforcement Learning via Curriculum Induction | 1 + ...rning: Sharper Analysis and Variance Reduction | 1 + ...ty of Uniform Convergence for Multicalibration | 1 + ...cement Learning via Low-Rank Matrix Estimation | 1 + ...ffective dimension for regression on manifolds | 1 + ...Deep Generative Models via Weighted Retraining | 1 + ...Reinforcement Learning of Undercomplete POMDPs | 1 + ...ling from a k-DPP without looking at all items | 1 + ...ecomposable Generative Adversarial Recommender | 1 + ...ng Methods: Random Tickets can Win the Jackpot | 1 + ...able Belief Propagation via Relaxed Scheduling | 0 ... Neural Networks via Bidirectional Propagation | 1 + ...ning for Networked Systems with Average Reward | 1 + ...r Communication-Efficient Distributed Training | 1 + ...Oversmoothness in Graph Convolutional Networks | 1 + ...r Low-Bit Weights in Quantized Neural Networks | 1 + ...x Optimization via Perturbed Gradient Tracking | 0 ...Bayesian Bounds for the Weighted Majority Vote | 1 + ... Matching Problems with Machine Learned Advice | 0 ...nd Seldonian Reinforcement Learning Algorithms | 1 + ...xplore: Curiosity via Audio-Visual Association | 1 + ...e Training: beyond Empirical Risk Minimization | 1 + ...03\251 from Focused and Defocused Image Pairs" | 1 + ...tion Amplifies Regularization in Hilbert Space | 1 + ...tillation as Instance-Specific Label Smoothing | 1 + ...earning via Generalized Lower Bound Q-learning | 1 + ...ations for Improving Gaze and Head Redirection | 1 + .../Self-Paced Deep Reinforcement Learning | 1 + ...f-Supervised Few-Shot Learning on Point Clouds | 1 + ...-Supervised Generative Adversarial Compression | 1 + ...Learning by Cross-Modal Audio-Video Clustering | 1 + .../Self-Supervised MultiModal Versatile Networks | 1 + ...lational Reasoning for Representation Learning | 1 + .../neurips/Self-Supervised Relationship Probing | 1 + ...esentation Learning from Hierarchical Grouping | 1 + ...Hybrid Memory for Domain Adaptive Object Re-ID | 1 + ...rning with Meta-paths for Heterogeneous Graphs | 1 + ... Co-Training for Video Representation Learning | 1 + ...upervised learning through the eyes of a child | 1 + ...ids Using Spurious Features Under Domain Shift | 1 + ...c Visual Navigation by Watching YouTube Videos | 1 + .../Semi-Supervised Neural Architecture Search | 1 + ...rning via Confidence-Rated Margin Maximization | 1 + ...ation for Lipschitz Constants of ReLU Networks | 1 + ...Analysis of Bias Due to Unobserved Confounding | 1 + ...perimental Design with Variable Cost Structure | 1 + .../neurips/Set2Graph: Learning Graphs From Sets | 1 + ...w: Learnable Deformation Flows Among 3D Shapes | 0 ...-Critic for Multi-Agent Reinforcement Learning | 1 + ...er Learning for analyzing multi-site fMRI data | 1 + ...ReLU Networks with Precise Dependence on Depth | 5 +++++ ...rgence bounds through empirical centralization | 1 + ... an Application to Noisy, Iterative Algorithms | 1 + ...er Generalization Bounds for Pairwise Learning | 1 + .../ShiftAddNet: A Hardware-Inspired Deep Network | 1 + ...r Binary Integer and Online Linear Programming | 0 ...rministic Deep Learning via Distance Awareness | 1 + ... Sparse k-means Clustering via Feature Ranking | 1 + ... Sampling for Implicit Collaborative Filtering | 1 + ...NNs Improves Robustness to Image Perturbations | 1 + ...ce and Metric Learning from Paired Comparisons | 1 + ...dversarial Episodic MDPs with Known Transition | 1 + ...orn Barycenter via Functional Gradient Descent | 5 +++++ ...inkhorn Natural Gradient for Generative Models | 1 + ...ion: From Global Inference to Local Adjustment | 1 + ...ng Window Algorithms for k-Clustering Problems | 1 + ...h Equilibrium Certificates in Very Large Games | 1 + ... And Consistent Probabilistic Regression Trees | 1 + ... of Online and Differentially Private Learning | 1 + .../Smoothed Geometry for Robust Attribution | 1 + ...ing User Contributions in Differential Privacy | 1 + .../SnapBoost: A Heterogeneous Boosting Machine | 1 + ...t Contrastive Learning for Visual Localization | 1 + ...ic Framework for Normalizing Flow on Manifolds | 1 + ...max Deep Double Deterministic Policy Gradients | 1 + ...Physics to Interact with Iterative PDE-Solvers | 1 + ...me Correspondence as a Contrastive Random Walk | 1 + .../Sparse Graphical Memory for Robust Planning | 1 + data/2020/neurips/Sparse Learning with CART | 1 + ...put Measures for Nonstationary Kernel Learning | 1 + ...arse Symplectically Integrated Neural Networks | 1 + .../neurips/Sparse Weight Activation Training | 1 + .../Sparse and Continuous Attention Mechanisms | 1 + ...angent Kernel for linear-width neural networks | 1 + ...twork for Multivariate Time-series Forecasting | 1 + ...Bayes for high dimensional logistic regression | 1 + data/2020/neurips/Spin-Weighted Spherical CNNs | 1 + ...ic Gradient Descent on Nonsmooth Convex Losses | 2 ++ .../Stable and expressive recurrent vision models | 1 + .../Stage-wise Conservative Linear Bandits | 1 + ...ynamic Deterministic Markov Decision Processes | 1 + ...s for Uncertainty Calibration in Deep Learning | 1 + ...ompson Sampling for Combinatorial Semi-Bandits | 1 + ...of Distributed Nearest Neighbor Classification | 1 + ...l Transport posed as Learning Kernel Embedding | 1 + ...l Properties of Sliced Probability Divergences | 1 + ...rce imaging with desparsified mutli-task Lasso | 0 ...al-Query Lower Bounds via Functional Gradients | 1 + ...te Analysis of Episodic Reinforcement Learning | 1 + ...ighbors in Supervised Dimensionality Reduction | 1 + ...Repulsive Dynamics: Benefits From Past Samples | 1 + ...Stochastic Deep Gaussian Processes over Graphs | 1 + ...elated Settings: A Study on Gaussian Processes | 1 + ...orcement Learning with a Latent Variable Model | 1 + data/2020/neurips/Stochastic Normalization | 1 + data/2020/neurips/Stochastic Normalizing Flows | 1 + ...astic Optimization for Performative Prediction | 3 +++ ...Tailed Noise via Accelerated Gradient Clipping | 1 + ...astic Optimization with Laggard Data Pipelines | 1 + ...ic Nonconvex-Strongly-Concave Minimax Problems | 1 + ...ing Spatially Correlated Aleatoric Uncertainty | 1 + data/2020/neurips/Stochastic Stein Discrepancies | 1 + ...earning Rate for Multiscale Objective Function | 1 + ...hannel Pruning via Deep Reinforcement Learning | 1 + ...r Misinformation Prevention in Social Networks | 1 + ...Learning by Energy-based Distribution Matching | 1 + ...onstituency Parsing with Graph Neural Networks | 1 + ...supervised learning and local graph clustering | 1 + ...nvolutions for Efficient Neural Network Design | 1 + ...tured Prediction for Conditional Meta-Learning | 1 + ...Bayesian Optimisation in Unknown Search Spaces | 1 + ...or Efficient Non-Parametric Bandit Exploration | 1 + data/2020/neurips/Subgraph Neural Networks | 1 + ...ubgroup-based Rank-1 Lattice Quasi-Monte Carlo | 1 + ...modular Maximization Through Barrier Functions | 1 + data/2020/neurips/Submodular Meta-Learning | 1 + ...nt Communication With Temporal Message Control | 1 + ...on using principal optimal transport direction | 1 + ... A Generic Loss for Robust Curriculum Learning | 1 + data/2020/neurips/Supermasks in Superposition | 1 + data/2020/neurips/Supervised Contrastive Learning | 1 + ...tions to Bridge the Gap between VAEs and Flows | 1 + ...apping Autoencoder for Deep Image Manipulation | 1 + ...ng Learning Algorithms with Synthetic Datasets | 1 + ...earning to Repair for Neural Program Synthesis | 1 + ...Synthesizing Tasks for Block-based Programming | 1 + ...hetic Data Generators - Sequential and Private | 1 + ...nstraints: A Circuit Model of the Inner Retina | 1 + ...Semantic Pyramid for Sign Language Translation | 1 + ...roblem in Heterogeneous Federated Optimization | 1 + ...ete Integration via the Boon of Dimensionality | 2 ++ ...l Perturbations for Monocular Depth Prediction | 1 + ... Inference of Gaussian Process Hyperparameters | 1 + ...with an Infinite Mixture of Gaussian Processes | 1 + .../neurips/Task-Oriented Feature Distillation | 1 + .../Task-Robust Model-Agnostic Meta-Learning | 1 + ...agnostic Exploration in Reinforcement Learning | 1 + ...s Sample-Efficient Natural Language Generation | 1 + .../2020/neurips/Teaching a GAN What Not to Learn | 1 + .../neurips/Telescoping Density-Ratio Estimation | 1 + ...ical Hypothesis Generation via Risk Estimation | 1 + ...ckpropagation for Deep Spiking Neural Networks | 1 + ...mporal Variability in Implicit Online Learning | 1 + .../2020/neurips/Tensor Completion Made Practical | 1 + .../neurips/Testing Determinantal Point Processes | 1 + ...re Interpolation for Probing Visual Perception | 1 + ...ty of Maximizing a Gross Substitutes Valuation | 1 + ...ning for Biased Regularization and Fine Tuning | 1 + ...All-or-Nothing Phenomenon in Sparse Tensor PCA | 1 + .../The Autoencoding Variational Autoencoder | 1 + .../neurips/The Complete Lasso Tradeoff Diagram | 1 + ...per Learning of Halfspaces with Agnostic Noise | 1 + ... of Silence: Speech Separation by Localization | 1 + ...on Relaxations for Neural Network Verification | 1 + ...on Exponential and Generalized Sylvester Flows | 1 + ... Macroscopic Prediction via Microscopic Models | 1 + ...hted TriHard Loss for Person Re-Identification | 1 + ...The Discrete Gaussian for Differential Privacy | 2 ++ .../The Diversified Ensemble Neural Network | 1 + ...l Privacy: Private Counting with Minimal Space | 1 + ...n-Stability Tradeoff In Neural Network Pruning | 1 + ...h Nonlinear Observations and Generative Priors | 1 + ...nge: Detecting Hate Speech in Multimodal Memes | 1 + ...al Correlation on Learning Some Deep Functions | 1 + ...Model-Based Behavior in Reinforcement Learning | 1 + ...icket Hypothesis for Pre-trained BERT Networks | 1 + .../The MAGICAL Benchmark for Robust Imitation | 1 + .../The Mean-Squared Error of Double Q-Learning | 1 + .../2020/neurips/The NetHack Learning Environment | 1 + ... Texture Bias in Convolutional Neural Networks | 1 + ...Pitfalls of Simplicity Bias in Neural Networks | 1 + ...tts-Ising model for discrete multivariate data | 1 + ...isons for Actively Learning Linear Classifiers | 1 + .../The Power of Predictions in Online Control | 1 + ...-Dual method for Learning Augmented Algorithms | 1 + .../The Smoothed Possibility of Social Choice | 1 + ...cal Complexity of Early-Stopped Mirror Descent | 2 ++ .../neurips/The Strong Screening Rule for SLOPE | 3 +++ ...arly-Time Learning Dynamics of Neural Networks | 1 + ...inciple for Model-Based Reinforcement Learning | 1 + .../The Wasserstein Proximal Gradient Algorithm | 1 + ...ndomness and structure during learning in RNNs | 1 + ...f approximation rates for deep neural networks | 1 + ...ames: When is price of anarchy too optimistic? | 2 ++ ...sification: A High-dimensional Asymptotic View | 1 + ...rized Differential Network Architecture Search | 1 + ...ology Design for Cross-Silo Federated Learning | 1 + ...oordinate Selection Solver for Sparse Learning | 1 + ...r Regret Bounds for Adversarial Linear Bandits | 1 + ...dient Descent under the Noiseless Linear Model | 1 + ...s for no-regret learning in multi-player games | 1 + .../neurips/Time-Reversal Symmetric ODE Network | 1 + ... using Temporal Hierarchical One-Class Network | 1 + ...ot Parameters for Efficient On-Device Learning | 1 + .../Top-KAST: Top-K Always Sparse Training | 1 + ...g GAN Performance by Throwing Away Bad Samples | 1 + ...arning Approach to Sequential Conformer Search | 1 + ...d the Fundamental Limits of Imitation Learning | 1 + ...er Generalization of Adaptive Gradient Methods | 1 + ... Analysis of Random Forests for Classification | 0 ...etworks using Decentralized Mixture-of-Experts | 1 + ...tworks with Differentiable Group Normalization | 1 + ...standing with Explanations as Latent Variables | 1 + .../Towards Learning Convolutions from Scratch | 1 + ...tween In-Domain & Out-of-Distribution Examples | 1 + ...Learning in Factored Markov Decision Processes | 1 + ...l Adversarial Attacks on Graph Neural Networks | 1 + .../neurips/Towards Neural Programming Interfaces | 1 + ...ll MOBA Games with Deep Reinforcement Learning | 1 + ...wards Problem-dependent Optimal Learning Rates | 1 + ...afe Policy Improvement for Non-Stationary MDPs | 1 + ...ards Scalable Bayesian Learning of Causal DAGs | 1 + ... Problem Solving by Iterative Homogeneous GNNs | 0 ... Generalizes Better Than Adam in Deep Learning | 1 + ...l Learning: Benefits of Neural Representations | 1 + ...Towards a Better Global Loss Landscape of GANs | 1 + ...al Characterization of Bounded-Memory Learning | 2 ++ ... differentially private causal graph discovery | 1 + ...sentation Learning for Information Obfuscation | 1 + ...acy: Data Debugging in Collaborative Filtering | 1 + ...oupling Locations of Weights from Their Values | 1 + ...rks by Solving Ordinary Differential Equations | 1 + ...erative Adversarial Networks with Limited Data | 1 + .../neurips/Training Linear Finite-State Machines | 1 + ...neck for Competitive Generative Classification | 1 + ...ng Stronger Baselines for Learning to Optimize | 1 + ...amics Generalization in Reinforcement Learning | 1 + ...fer Learning via \342\204\2231 Regularization" | 0 ...h Lower Bias and Variance in Domain Adaptation | 1 + ...Transferable Graph Optimizers for ML Compilers | 1 + ...e! I am a low dimensional Hyperbolic Embedding | 1 + ...ds of overfitting: where & why do they appear? | 1 + ...Truncated Linear Regression in High Dimensions | 1 + ... Is Confident: Masked Model-based Actor-Critic | 2 ++ .../Truthful Data Acquisition via Peer Prediction | 1 + ...et: Single View Reconstruction in Object Space | 1 + ...iscovering of Constructive Solid Geometry Tree | 1 + ...raphy, Watermarking, and Light Field Messaging | 1 + ...el Capacity Weakly Supervised Object Detection | 1 + ...ecision 4-bit Training of Deep Neural Networks | 1 + .../Ultrahyperbolic Representation Learning | 1 + ... a Modulo Image for High Dynamic Range Imaging | 1 + data/2020/neurips/Unbalanced Sobolev Descent | 1 + ...y Aware Semi-Supervised Learning on Graph Data | 1 + ...y Quantification for Inferring Hawkes Networks | 1 + ...e Learning for Zero-Shot Semantic Segmentation | 1 + ...Self-training for Few-shot Text Classification | 1 + ...me-Varying fMRI Data using Cubical Persistence | 1 + ...ough Hierarchies of Distributions and Features | 1 + ...tural Gradient Descent in Wide Neural Networks | 1 + ...tanding Deep Architecture with Reasoning Layer | 0 ...res A Fine-Grained Bias-Variance Decomposition | 1 + ...ontributions With Additive Importance Measures | 1 + ...ipping in Private SGD: A Geometric Perspective | 1 + ...ring the Network with Stochastic Architectures | 1 + ...anding and Improving Fast Adversarial Training | 1 + ...g spiking networks through convex optimization | 1 + ...Role of Training Regimes in Continual Learning | 1 + ...'s functions for optimized reservoir computing | 1 + ...nating Optimization for Blind Super Resolution | 1 + ...sed Learning Rules for Spiking Neural Networks | 1 + ...sal Domain Adaptation through Self Supervision | 1 + .../Universal Function Approximation on Graphs | 1 + ...duction via a higher-order splitting criterion | 2 ++ .../Universally Quantized Neural Compression | 1 + ...lgorithms in Multi-Armed Bandit with Many Arms | 0 ...sed Data Augmentation for Consistency Training | 1 + ...tations with Compositional Energy-Based Models | 1 + ...vised Learning of Dense Visual Representations | 1 + ...ynamics from Images for Prediction and Control | 1 + ...ect Landmarks via Self-Training Correspondence | 1 + ...al Features by Contrasting Cluster Assignments | 1 + ...resentation Learning by Invariance Propagation | 1 + ...Template Matching for Semi-Supervised Learning | 1 + ...nd Separation Using Mixture Invariant Training | 1 + ...rvised Text Generation by Learning from Search | 1 + ...upervised Translation of Programming Languages | 1 + ...ntric video generation and decomposition in 3D | 1 + ...d self-attention in artificial neural networks | 0 ...cement Learning for CMDP with Adversarial Loss | 1 + ...Sequence Models for Continuous-Time Event Data | 1 + ...nt neural network structure and prune synapses | 1 + ...rative Model for Heterogeneous Mixed Type Data | 1 + ...and Semi-supervised Learning to Tabular Domain | 1 + .../2020/neurips/Value-driven Hindsight Modelling | 1 + ...e Gradient Estimator for Variational Inference | 1 + ...ted Dual Averaging for Finite-Sum Optimization | 1 + ...Random Coordinate Descent-Langevin Monte Carlo | 3 +++ ... Learning: Non-Asymptotic Convergence Analysis | 1 + .../neurips/Variational Amodal Object Completion | 1 + ...al Bayesian Monte Carlo with Noisy Likelihoods | 1 + data/2020/neurips/Variational Bayesian Unlearning | 1 + ...Absence of Graph Data and Adversarial Settings | 1 + ... Reinforcement Learning with General Utilities | 1 + ...eo Frame Interpolation without Temporal Priors | 1 + ...e Feature Bank and Uncertain-Region Refinement | 1 + ...escent Directions for Constrained Minimization | 1 + ...riational Inference for Bayesian Deep Learning | 1 + ...tein Distances for Stereo Disparity Estimation | 1 + ...urring the Vision of Your Deep Neural Networks | 1 + ... Training of High Resolution Normalizing Flows | 1 + .../Weak Form Generalized Hamiltonian Learning | 1 + ...rvised Deep Functional Maps for Shape Matching | 0 ...inforcement Learning for Controllable Behavior | 1 + ...on for Deep Multi-Agent Reinforcement Learning | 1 + ...Towards scalable higher-order graph embeddings | 1 + ...ston-Watkins Hinge Loss and Ordered Partitions | 1 + ...ning Agent Behaviour through Intended Outcomes | 1 + ...etworks Learn When Trained With Random Labels? | 1 + ...Makes for Good Views for Contrastive Learning? | 1 + ...overing the Long Tail via Influence Estimation | 2 ++ .../neurips/What if Neural Networks had SVDs? | 1 + ...hat is being transferred in transfer learning? | 1 + ...xploring datasets, architectures, and training | 1 + ...re importance for time-series black-box models | 1 + .../When Counterpoint Meets Chinese Folk Melodies | 1 + ... Do Neural Networks Outperform Kernel Methods? | 1 + ...essment using Compartmental Gaussian Processes | 1 + ...etworks? - A Neural Tangent Kernel Perspective | 1 + ... Flows Fail to Detect Out-of-Distribution Data | 1 + ...re Adaptive Methods Good for Attention Models? | 1 + ...ing the Lottery with Continuous Sparsification | 1 + ... Improving Consistency of Deep Learning Models | 1 + ...r Approximation for Neural Network Compression | 2 ++ ...bury Transformations for Deep Generative Flows | 1 + ...orst-Case Analysis for Randomly Collected Data | 1 + ...AL: Explicit Calibration for Survival Analysis | 1 + ...ecretly Suffice Multi-Source Domain Adaptation | 1 + ...hould Use Discriminator Driven Latent Sampling | 1 + ...Learning With Nonlinear Function Approximation | 1 + ...esource Knowledge-Grounded Dialogue Generation | 1 + .../neurips/f-Divergence Variational Inference | 1 + ... for Generative Adversarial Imitation Learning | 1 + ...-Supervised Learning of Speech Representations | 1 + ...Exploration from Decentralized Learning Agents | 1 + ...th Correspondence Learning and Mesh Refinement | 1 + ...e Voxel-to-BEV Tracker for Sparse Point Clouds | 1 + ...Scene Perception via Probabilistic Programming | 1 + ...nerative Model for Structure-Based Drug Design | 1 + ...Recommender Systems in a Two-sided Marketplace | 1 + ...with Control in the Presence of Subpopulations | 1 + ...to Reasoning and Learning in Intuitive Physics | 1 + ... to Solve Combinatorial Optimization on Graphs | 1 + ...eural Network Sampler with Near-Optimal Regret | 1 + ...A Causal Lens for Controllable Text Generation | 1 + ...rem for Differentially Private Query Answering | 1 + ...case Behavior of Multi-armed Bandit Algorithms | 1 + ...Circuit Operations for Probabilistic Inference | 1 + ...ely Tight Analysis of Gradient Descent for PCA | 1 + ... for Learning Exponential Family Distributions | 1 + ...g Agent for Model-Based Reinforcement Learning | 1 + ...ndom-Order No-Substitution k-Median Clustering | 1 + .../A Continuous Mapping For Augmentation Design | 1 + ...ch for Training Variational Autoencoder Priors | 1 + ...s of Gradient Descent on Graph Neural Networks | 1 + ...al Estimation with Deep Latent Variable Models | 1 + ...lgorithm with Order-Optimal Regret Performance | 1 + ...lized Algorithm for Nonconvex Minimax Problems | 1 + ...lgorithm with Applications in Machine Learning | 1 + .../A Framework to Learn with Interpretation | 1 + data/2021/neurips/A Gang of Adversarial Bandits | 1 + ... Mixture Model for Multi-Label Active Learning | 1 + ...of Neural Collapse with Unconstrained Features | 1 + ...ural Calibration via Sensitivity Decomposition | 1 + ...on and Its Role in Making Gradients Small Fast | 1 + .../A Gradient Method for Multilevel Optimization | 1 + ...rge-scale Dynamic Pickup and Delivery Problems | 1 + ...n Application to Function-On-Scalar Regression | 1 + ...st of Independence for Cluster-correlated Data | 1 + ...garithm for Multi-Agent Reinforcement Learning | 1 + ... Robust Features for Targeted Transfer Attacks | 1 + ...nsferability in Multi-source Transfer Learning | 1 + ...n Entropy Framework for Reinforcement Learning | 1 + ...ist Approach to Offline Reinforcement Learning | 1 + ...Multi-Implicit Neural Representation for Fonts | 1 + ... for Debiasing Trained Machine Learning Models | 1 + ...astic Bilevel Optimization via Double-Momentum | 1 + ...k for Fast and Accurate Online Decision-Making | 1 + ...or Robust Acceleration in the Hyperbolic Plane | 1 + ...rithm for Positive Semidefinite Factorizations | 1 + ...e Algorithm for Independent Component Analysis | 1 + ... Note on Sparse Generalized Eigenvalue Problem | 1 + ...A PAC-Bayes Analysis of Adversarial Robustness | 1 + ...Inference from Differential Equations and Data | 1 + ...d Framework for Unsupervised Domain Adaptation | 1 + ...ing Method for Episodic Reinforcement Learning | 1 + ...Collection Strategy for Reinforcement Learning | 1 + ...proach to Learning-Augmented Online Algorithms | 1 + ...ata-oblivious and Data-aware Poisoning Attacks | 1 + ...el for Shape-Accurate 3D-Aware Image Synthesis | 1 + ... Algorithm for Distributed Convex Optimization | 1 + ...r Prediction+Programming with Soft Constraints | 0 ...l Analysis of Fine-tuning with Linear Teachers | 1 + ...rtion-Perception Tradeoff in Wasserstein Space | 1 + ...Method for Contrastive Representation Learning | 1 + .../A Topological Perspective on Causal Inference | 1 + ...ding Model for Hyperspectral Image Restoration | 1 + ... Online Learning via Blackwell Approachability | 1 + ...ied View of cGANs with and without Classifiers | 1 + ...A Universal Law of Robustness via Isoperimetry | 1 + ...ion-Based Generative Models and Score Matching | 1 + ...rks Can Improve Out-of-Distribution Robustness | 1 + ...ual method with adaptivity to local smoothness | 1 + ... neural population responses to natural images | 1 + ...nonparametric Bayesian model for whole genomes | 1 + ...rea recurrent network model of decision-making | 1 + ...al change problems with statistical guarantees | 1 + ...ributions based on optimal weak mass transport | 1 + ...ling-based circuit for optimal decision making | 1 + ...ptures feature learning effects in finite CNNs | 1 + ... examples on random two-layers neural networks | 1 + ... unified framework for bandit multiple testing | 1 + ...veals ongoing modulation of neural variability | 1 + ...oximate posterior for the deep Wishart process | 1 + ...for Learning Human Shape, Appearance, and Pose | 1 + ... for Class-imbalanced Semi-supervised Learning | 1 + ... DeCompressed Training of Deep Neural Networks | 1 + ...vation Compression with Guaranteed Convergence | 1 + ...ing of Negative Transfer in Continual Learning | 1 + ... Efficient Point Cloud Representation Learning | 1 + ...essive Transformers for Indoor Scene Synthesis | 1 + ...st for Detecting Heteroscedastic Relationships | 1 + ...ficient Method to Find N: M Transposable Masks | 1 + ...ratic Optimization with Reinforcement Learning | 1 + ...t Learning via Parameterized Action Primitives | 1 + ...ity for Multi-Objective Reinforcement Learning | 1 + ...cumulative Poisoning Attacks on Real-time Data | 1 + ...oud Registration with Robust Optimal Transport | 1 + ...ately Solving Rod Dynamics with Graph Learning | 1 + ...n and Knowledge Transfer in Continual Learning | 1 + ...ance with Bessel-Convolutional Neural Networks | 1 + ...r decoding by probabilistic manifold alignment | 1 + .../Action-guided 3D Human Motion Prediction | 1 + ...sport Problem without Bidirectional Connection | 1 + ... 3D Shape Reconstruction from Vision and Touch | 1 + ...s Accuracy Surface Over Attribute Combinations | 1 + ...Active Learning of Convex Halfspaces on Graphs | 1 + data/2021/neurips/Active Offline Policy Selection | 1 + .../Active clustering for labeling training data | 1 + ...iables Given Only Response Variable Observable | 1 + ...Populations via a Generative Model of Policies | 1 + ... and growth conditions in private optimization | 1 + ...e Conformal Inference Under Distribution Shift | 1 + .../Adaptive Data Augmentation on Temporal Graphs | 0 .../neurips/Adaptive Denoising via GainTuning | 1 + .../Adaptive Diffusion in Graph Neural Networks | 1 + ... Minimizing Estimation Bias via Error Feedback | 1 + ...ex Minimization without Lipschitz Requirements | 1 + data/2021/neurips/Adaptive Machine Unlearning | 1 + ...aptive Online Packing-guided Search for POMDPs | 1 + ...radient Methods for Structured Neural Networks | 1 + ...inimization: Learning to Adapt to Domain Shift | 1 + ...ptive Sampling for Minimax Fair Classification | 1 + ...n from neural networks through interpretations | 1 + .../Adder Attention for Vision Transformer | 1 + ...erformance Inconsistency in Federated Learning | 1 + ...ated Errors in Neural Networks for Time Series | 1 + ...k Generation Empowered by Min-Max Optimization | 1 + ...eraging the Power of Geometric Transformations | 1 + ...on Graph Classifiers via Bayesian Optimisation | 0 .../Adversarial Examples Make Strong Poisons | 1 + ...sifiers Based on Higher-Order Voronoi Diagrams | 1 + ...l Examples in Multi-Layer Random ReLU Networks | 1 + .../neurips/Adversarial Feature Desensitization | 1 + ...entation to Improve Graph Contrastive Learning | 1 + ...ntrinsic Motivation for Reinforcement Learning | 1 + ...Neuron Pruning Purifies Backdoored Deep Models | 1 + ...on with Doubly Non-negative Weighting Matrices | 1 + ...rial Reweighting for Partial Domain Adaptation | 1 + ...rial Robustness with Non-uniform Perturbations | 1 + ...stness with Semi-Infinite Constrained Learning | 1 + ... A Teacher-Guided Curriculum Learning Approach | 0 ...resentation Learning for Domain Generalization | 1 + ...s Transfer Learning via Better Representations | 1 + ...oint Cloud Recognition Using Self-Supervisions | 1 + .../Adversarially Robust Change Point Detection | 1 + ...ng for security-constrained optimal power flow | 1 + ... Observability for Deep Reinforcement Learning | 1 + ...rning with Low-Rank MDPs and Rich Observations | 1 + ... Instabilities of Accelerated Gradient Descent | 1 + ...of an unsupervised feature selection algorithm | 1 + .../Alias-Free Generative Adversarial Networks | 1 + ...esentation Learning with Momentum Distillation | 1 + ... Learning for Efficient Image Super-Resolution | 1 + ...etection via Object-Level Contrastive Learning | 1 + ...ology for Self-Adaptive 3D Human Pose Recovery | 1 + ...ention by Matching Key and Query Distributions | 1 + ...beling for Training Better Vision Transformers | 1 + ...onfigurations Using a Differentiable Surrogate | 1 + ...ional Inference for Simple Hierarchical Models | 1 + ...n-convex Regime: Asymptotic Normality and Bias | 1 + ...Provably-Fair Welfare-Centric Machine Learning | 1 + ...hastic Linear Bandits with General Constraints | 1 + ...ramework for Multiagent Reinforcement Learning | 1 + ... Generalization with Empirical Risk Minimizers | 1 + ... of Adder Neural Networks for Object Detection | 1 + ...rithm: Minibatching and Interpolation Learning | 1 + ...e Generalization Error for the Gibbs Algorithm | 1 + ...morization Capacity of Deep Threshold Networks | 1 + ...Realizable MDP with Constant Suboptimality Gap | 1 + ...and Words: Towards Disentanglement in The Wild | 1 + ...tion under Without-replacement Sampling Orders | 0 ...nt Tracking for Decentralized Machine Learning | 1 + ...ets That Fixes Their Asymptotic Overconfidence | 1 + ...tion-theoretic Approach to Distribution Shifts | 1 + ...Robust Optimization with Non-convex Objectives | 1 + ... for Stochastic Canonical Correlation Analysis | 1 + ...ple is a Price of Privacy-Preserving Microdata | 1 + ...of Ermakov-Zolotukhin quadrature using kernels | 1 + ...rithm for difference-of-squares classification | 1 + ... Algorithm: Designing a Unified Sequence Model | 1 + ...gnal Recovery under a Generalized Linear Model | 1 + ...layer neural networks via the resolvent method | 1 + ...ucture and Rank of Neural Network Hessian Maps | 1 + ...adigmatic High-Dimensional Non-Convex Problems | 1 + ...distillable Teachers in Knowledge Distillation | 1 + ... of SGLD Using Properties of Gaussian Channels | 1 + ...sal Queries With the Maximum Causal Set Effect | 1 + ...arning: Training Clean Models on Poisoned Data | 1 + ... of Label Differential Privacy: PATE and ALIBI | 1 + ... Minimization for Cardinality-Based Components | 1 + ...ization of convex functions with outlier noise | 1 + ...ing the Permanent with Deep Rejection Sampling | 1 + ...rbitrary Conditional Distributions with Energy | 1 + ...air? An Empirical Study of Fixed-Seed Training | 1 + .../Are Transformers more robust than CNNs? | 1 + ... Diffusion: Learning Categorical Distributions | 1 + ...nal-external Learning and Contrastive Learning | 1 + ...ssing Fairness in the Presence of Missing Data | 1 + ...ith Transformers for Video Object Segmentation | 1 + .../Associative Memories via Predictive Coding | 1 + ...Effect Identification with Multi-Armed Bandits | 1 + ...icy Evaluation with Misspecified Linear Models | 1 + ...on learning in finite Bayesian neural networks | 1 + ...Applications to Inference with Model Selection | 1 + .../Asynchronous Decentralized Online Learning | 1 + ...ntralized SGD with Quantized and Local Updates | 1 + ...hastic Optimization Robust to Arbitrary Delays | 1 + ...tention Approximates Sparse Distributed Memory | 1 + .../Attention Bottlenecks for Multimodal Fusion | 1 + ...ct Embeddings Enables Complex Visual Reasoning | 1 + ...iction Models for Data Minimization Compliance | 1 + ...on of Random Augmentations for Robust Training | 1 + .../Augmented Shortcuts for Vision Transformers | 1 + ...aph for Unsupervised Medical Report Generation | 1 + ...: Optimized Loss Functions for Imbalanced Data | 1 + ... Neural Network with Explicit Link Information | 1 + ...Autobahn: Automorphism-based Graph Neural Nets | 1 + ...o-Correlation for Long-Term Series Forecasting | 1 + ...ry of Adaptive Attacks on Adversarial Defenses | 1 + .../neurips/Automated Dynamic Mechanism Design | 1 + ...n for Generalization in Reinforcement Learning | 1 + ...scovery with Lie Algebra Convolutional Network | 1 + ...Automatic Unsupervised Outlier Model Selection | 1 + ...aphic Optimization: A Dynamic Barrier Approach | 0 ...morphic Equivalence-aware Graph Neural Network | 1 + ...s Reinforcement Learning via Subgoal Curricula | 1 + ...rage-Reward Learning and Planning with Options | 1 + ...dimension-free convergence of gradient descent | 1 + ...: Evaluating Generated Text as Text Generation | 1 + ... Spanning Trees for Complex Constrained Domain | 1 + ...ional Approaches for Bayesian Causal Discovery | 1 + ...k for Coupons Allocation in E-commerce Market" | 1 + ... Structures Dynamically for Continual Learning | 1 + ... the goals, preferences, and actions of others | 1 + ...th Imperceptible Input and Latent Modification | 1 + ...e Prediction Updates: A Probabilistic Approach | 1 + ...omprehensive Metric for Point Cloud Completion | 0 ...Hop Reasoning at Scale via Condensed Retrieval | 1 + ...Bandit Learning with Delayed Impact of Actions | 1 + data/2021/neurips/Bandit Phase Retrieval | 1 + .../neurips/Bandit Quickest Changepoint Detection | 1 + .../Bandits with Knapsacks beyond the Worst Case | 1 + data/2021/neurips/Bandits with many optimal arms | 1 + data/2021/neurips/Batch Active Learning at Scale | 1 + ...ptimization with Deep Auto-Regressive Networks | 1 + ...alizes Representations in Deep Random Networks | 1 + ...-all Architecture Search with Robust Quantizer | 1 + data/2021/neurips/Batched Thompson Sampling | 1 + ...ertainty Quantification for Causal Data Fusion | 1 + .../Bayesian Adaptation for Covariate Shift | 0 data/2021/neurips/Bayesian Bellman Operators | 1 + .../Bayesian Optimization of Function Networks | 1 + ...ian Optimization with High-Dimensional Outputs | 1 + ...fied priors with applications to meta-learning | 1 + ...aph Neural Networks via Confidence Calibration | 1 + ...rom the Void: Unsupervised Active Pre-Training | 1 + ...for Offline Multi-Agent Reinforcement Learning | 1 + ...f RL Problems, and Sample-Efficient Algorithms | 1 + ...t Pessimism for Offline Reinforcement Learning | 1 + .../Beltrami Flow and Neural Diffusion on Graphs | 1 + ...lassification: All Roads Lead to Interpolation | 1 + ...h Spectral Filters via Bernstein Approximation | 1 + ...ntification in Contaminated Stochastic Bandits | 0 ...ly Optimal Submodular Maximization in Parallel | 1 + .../Best-case lower bounds in online learning | 1 + ...nts for Neural Network Robustness Verification | 1 + ... Algorithms for Individually Fair k-Clustering | 1 + ...Delusive Adversaries with Adversarial Training | 1 + ...ausal Discovery Benchmarks May Be Easy to Game | 1 + ...t Feedback in Online Multiclass Classification | 1 + ...nderstanding of Normalization in Deep Learning | 1 + ...hods for Calibrated Uncertainty Quantification | 1 + ...Analysis into Nonparametric Density Estimation | 1 + ...oncordant losses, via iterative regularization | 1 + ...ret Bounds for Episodic Reinforcement Learning | 1 + ...onparametric Tensor Completion via Sign Series | 1 + ...l Biases in Popular Generative Language Models | 1 + ...Bias and variance of the Bayesian-mean decoder | 1 + .../neurips/Biological key-value memory networks | 0 .../2021/neurips/Black Box Probabilistic Numerics | 1 + ...lending for Arbitrary Stylized Face Generation | 1 + ...Blending Anti-Aliasing into Vision Transformer | 1 + ...oosting Approach for Continual Learning of VAE | 1 + ...rovably Efficient Bootstrapped Value Iteration | 1 + .../neurips/Boost Neural Networks by Checkpoints | 1 + data/2021/neurips/Boosted CVaR Classification | 1 + data/2021/neurips/Boosting with Multiple Sources | 1 + ...tstrap Your Object Detector via Mixed Training | 1 + .../Bootstrapping the Error of Oja's Algorithm | 1 + ... energy-based models with bidirectional bounds | 1 + ... Dilemma of Medical Image-to-image Translation | 1 + ...l Gradient Methods Using MaxIP Data-structures | 1 + ...hm for Bandits with Super Heavy-Tailed Payoffs | 1 + ...gret-Optimal Model-Free Reinforcement Learning | 3 +++ ...ed barrier for cross-device federated learning | 1 + ... Construction with Deep Reinforcement Learning | 1 + ... Generative Models via Neural Stein Estimators | 1 + ...the-wild Data for Incremental Object Detection | 1 + ...ng and Imitation Learning: A Tale of Pessimism | 1 + ...and PAC-Bayes Theory in Few-Shot Meta-Learning | 1 + ... the Imitation Gap by Adaptive Insubordination | 1 + ... real-time flow prediction on neural manifolds | 1 + ...l Network Training via Boundary Example Mining | 1 + ...ByPE-VAE: Bayesian Pseudocoresets Exemplar VAE | 1 + ...on Modules for Generative Adversarial Networks | 1 + ...ex Optimization with Communication Compression | 1 + ...ith Continuous Augmented Positional Embeddings | 1 + ...etic-REINFORCE Multi-Sample Gradient Estimator | 1 + ...egation Transformers for Visual Correspondence | 1 + ...sion using a pseudo-Lagrange multiplier method | 1 + ...VS: Context-aware Controllable Video Synthesis | 1 + ...ence-based Pruning for Compact Neural Networks | 1 + ...Learning for Semi-Supervised Domain Adaptation | 1 + .../CLIP-It! Language-Guided Video Summarization | 1 + ... reInforcement Learning On sub-Task curriculum | 1 + ... Text Sequences for Language Model Pretraining | 1 + ...bject and Hand Embedding Segmentation In Video | 0 ...vative Offline Model-Based Policy Optimization | 1 + ...s Based on Patient Disease Class, Sex, and Age | 1 + ...odels for Probabilistic Time Series Imputation | 1 + ...s: A Novel Approach to Multi-Class Calibration | 1 + ...nd Consistency of Adversarial Surrogate Losses | 1 + ... Targets for Interventions in Neural Circuits? | 1 + ...ancing Label Noise Rates Considered Beneficial | 1 + ... Easy to Hard Problems with Recurrent Networks | 1 + ...contrastive learning avoid shortcut solutions? | 1 + ...sentation of syntactic structure in the brain? | 1 + ...sification networks know what they don't know? | 1 + ...ation loss? Quasiconvexity in ridge regression | 1 + ... and Adversarial Robustness of Neural Networks | 1 + ...es: Self-Supervised Capsules in Canonical Pose | 1 + ...arned Geometric Embeddings for Directed Graphs | 1 + ...ith self-supervised hyperbolic representations | 1 + ...ned submodular maximization for random streams | 1 + .../Cardinality-Regularized Hawkes-Granger Model | 1 + ...n The Emergence Of Compositional Communication | 1 + ...ic Data Leakage in Vertical Federated Learning | 0 ... to Generate Audio from a Single Short Example | 1 + .../Causal Abstractions of Neural Networks | 1 + .../Causal Bandits with Unknown Graph Structure | 1 + ...sal Effect Inference for Structured Treatments | 1 + .../Causal Identification with Matrix Equations | 1 + ...or Event Pairs in Multivariate Point Processes | 1 + ...Improving Efficiency in Reinforcement Learning | 1 + ... Navigation by Continuous-time Neural Networks | 1 + ...nfer Treatment-Effects from Observational Data | 1 + ...y in Shared Multi-Agent Reinforcement Learning | 1 + ...obustness for Networks with Structured Outputs | 1 + ...stance Representation for Scene Text Detection | 1 + ...ss to Programmable Data Bias in Decision Trees | 1 + ...ties in High Dimensional Variational Inference | 1 + ...on via Multivariate Singular Spectrum Analysis | 1 + .../Channel Permutations for N: M Sparsity | 1 + ...Of-Distribution Shifts in Deep Metric Learning | 1 + ...lure modes in physics-informed neural networks | 1 + .../Characterizing the risk of fairwashing | 1 + ...ace of Solutions for Recurrent Neural Networks | 1 + ...Vision Transformers: An End-to-End Exploration | 1 + ...nett Inequality for the Weighted Majority Vote | 1 + .../Choose a Transformer: Fourier or Galerkin | 1 + ...ca: Stochastic ReLUs for Private Deep Learning | 1 + ...lications in Adversarial Detection and Defense | 1 + ...ass-Incremental Learning via Dual Augmentation | 1 + ... Reconstruction of Dynamic Objects from Videos | 1 + .../neurips/Clockwork Variational Autoencoders | 1 + ...ochastic Gradient Methods for Bilevel Problems | 1 + ...-making: A case study on organ transplantation | 1 + ...Clustering Effect of Adversarial Robust Models | 0 ...in Inference-based Deep Reinforcement Learning | 1 + ...ion Transformer for Protein Contact Prediction | 1 + ...g Convolution and Attention for All Data Sizes | 1 + ...espondences for Robust PointCloud Registration | 1 + ...l Architecture Inspired by Continued Fractions | 0 ...oarse-to-fine Animal Pose and Shape Estimation | 1 + ... Tool for the Training of Deep Neural Networks | 1 + ...ring Text-to-Image Generation via Transformers | 1 + .../Collaborating with Humans without Human Data | 1 + ...ive Causal Discovery with Atomic Interventions | 1 + ...ogeneous, Asynchronous and Nonconvex Learning) | 1 + ...ertainty in Multi-Agent Trajectory Forecasting | 1 + ...ariational Bounds for Bayesian Neural Networks | 1 + ... Learning by Region Uncertainty Quantification | 1 + ... Segmentation: A Fully Differentiable Approach | 1 + ...re Exploration with Bottleneck Reward Function | 1 + ...ntion Transformer with Sparse Computation Cost | 1 + ...ilities via Confusion Matrices and Calibration | 1 + ...ayesian Optimization over Combinatorial Spaces | 1 + ...ous-time Models with Linear State Space Layers | 1 + ...ient SGD: From Local SGD to One-Shot Averaging | 1 + ...Efficient Low-Rank Hypercomplex Adapter Layers | 1 + ...onconvex-Strongly-Concave Min-Max Optimization | 1 + ...namical Systems with ODE-based Random Features | 1 + ...forcement Learning from Logical Specifications | 1 + ...ompositional Transformers for Scene Generation | 1 + ...nowledge Distillation with Causal Intervention | 1 + .../neurips/Compressed Video Contrastive Learning | 1 + ...termining the Optimal Layer-wise Decomposition | 1 + .../neurips/Compressive Visual Representations | 1 + .../neurips/Computer-Aided Design as Language | 1 + ... for Multi-Hop Reasoning over Knowledge Graphs | 1 + ...er sub-Gaussian and sub-exponential conditions | 1 + ...itional Generation Using Polynomial Expansions | 1 + ...ks for Mesh-Based Modeling of Physical Systems | 1 + ... Gaussian Processes for Online Decision-making | 1 + ...ng from Demonstrations with Varying Optimality | 1 + ...or-Induced Multi-Source Free Domain Adaptation | 1 + ...verse Gradient Descent for Multi-task learning | 1 + data/2021/neurips/Conformal Bayesian Computation | 1 + ...formal Prediction using Conditional Histograms | 1 + .../neurips/Conformal Time-series Forecasting | 1 + ...meter-Free Convex-Concave Saddle-Point Solving | 1 + ... for Multi-Task Offline Reinforcement Learning | 1 + ... Offline Distributional Reinforcement Learning | 1 + ...y Regularization for Variational Auto-Encoders | 1 + ... and Sparse Regression with Oblivious Outliers | 1 + ...n-Parametric Methods for Maximizing Robustness | 1 + ...orks on Critical and Under-Represented Classes | 1 + .../Constrained Robust Submodular Partitioning | 1 + .../Container: Context Aggregation Networks | 1 + ...ations and Low-Regret Cutting-Plane Algorithms | 1 + ...tion with Self-attention for Visual Re-ranking | 1 + .../neurips/Continual Auxiliary Task Learning | 1 + ...ontinual Learning via Local Module Composition | 1 + ...Benchmark For Continual Reinforcement Learning | 1 + ...ic Gradient Descents, and of Gossip Algorithms | 0 ...oubly Constrained Batch Reinforcement Learning | 1 + data/2021/neurips/Continuous Latent Process Flows | 1 + .../neurips/Continuous Mean-Covariance Bandits | 1 + ... Discrete Optimization of Deep Neural Networks | 1 + ...modelling using non-parametric point processes | 1 + ...Video Domain Adaptation with Background Mixing | 1 + data/2021/neurips/Contrastive Active Inference | 1 + ...ervised Learning with Extremely Limited Labels | 1 + data/2021/neurips/Contrastive Laplacian Eigenmaps | 1 + .../Contrastive Learning for Neural Topic Model | 1 + ...ning of Global and Local Video Representations | 0 ...rcement Learning of Symbolic Reasoning Domains | 1 + ...isentangled Sequential Variational Autoencoder | 1 + ...ntrol Variates for Slate Off-Policy Evaluation | 1 + ...tinuous Optimization with Multiple Constraints | 1 + ...ling Neural Networks with Rule Representations | 1 + ...Gradient Descent under Infinite Noise Variance | 1 + ...hms for constrained weakly convex optimization | 1 + ...nvex Polytope Trees and its Application to VAE | 0 .../Convex-Concave Min-Max Stackelberg Games | 1 + ... Convolutional Network Robustness and Training | 1 + ...h Asynchronous Agents and Constrained Feedback | 1 + .../Coordinated Proximal Policy Optimization | 1 + ...r Classification - Simplified and Strengthened | 1 + .../Coresets for Clustering with Missing Values | 3 +++ .../Coresets for Decision Trees of Signals | 1 + .../neurips/Coresets for Time Series Clustering | 1 + ...ng with Applications to Recovering Communities | 1 + .../neurips/Corruption Robust Active Learning | 1 + ...er Network for Cortical Surface Reconstruction | 0 ...ellar networks as decoupling neural interfaces | 1 + ...and Strategic Incentives in Allocation Markets | 1 + ... Policy Refinement Using Bayesian Optimization | 1 + ...Counterfactual Explanations Can Be Manipulated | 1 + ...n Sequential Decision Making Under Uncertainty | 1 + ...o Spurious Correlations in Text Classification | 0 ...kelihood Estimation for Training Deep Networks | 1 + ...dient Estimators for Discrete Latent Variables | 1 + ...nd Edge Learning via Dynamic Graph Propagation | 1 + ...timation Without Private Covariance Estimation | 1 + data/2021/neurips/Credal Self-Supervised Learning | 1 + ...ent Through Broadcasting a Global Error Vector | 1 + ... Neural Networks through Deep Feedback Control | 1 + ...r Cost-Efficient Visual Reinforcement Learning | 1 + ...o-localization with Layer-to-Layer Transformer | 1 + ...Multi-Party Computation Meets Machine Learning | 1 + ...ng via Demonstrations: Theory and Applications | 1 + ...ngled Recommendation with Noisy Multi-feedback | 1 + ...um Learning for Vision-and-Language Navigation | 1 + .../neurips/Curriculum Offline Imitating Learning | 1 + .../Cycle Self-Training for Domain Adaptation | 1 + ...ing Models for Few-Shot Conditional Generation | 1 + ... Data Using Causally-Aware Generative Networks | 1 + ...Material with a Hybrid Differentiable Renderer | 1 + ...: Spatial Invariance and Neural Tangent Kernel | 1 + ...e-Training Objective for Programming Languages | 1 + ... Method for Detecting Misclassification Errors | 1 + ...supervised Learning with A Few Labeled Samples | 1 + .../DRIVE: One-bit Distributed Mean Estimation | 1 + ... SLAM for Monocular, Stereo, and RGB-D Cameras | 1 + ...ware Low-rank Compression for Large NLP Models | 1 + ...perts with Applications to Multi-Task Learning | 1 + ...: Acceleration, Convergence, and Stabilization | 1 + ...Bayesian Model Averaging under Covariate Shift | 1 + .../Data Augmentation Can Improve Robustness | 1 + ... Compression for Cooperative Networked Control | 1 + .../neurips/Data driven semi-supervised learning | 1 + ...t) Augmentations: A Lottery Ticket Perspective | 1 + ...stance Generation from Instance Discrimination | 1 + ...on with Infinitely Wide Convolutional Networks | 1 + ...CMC dynamics with the diffusion Stein operator | 1 + ...n Fixed-Confidence Linear Top-m Identification | 1 + ...Answering from Feature and Sample Perspectives | 1 + ...ugmentation for GAN Training with Limited Data | 1 + ...centralized Learning in Online Queuing Systems | 1 + ...entralized Q-learning in Zero-sum Markov Games | 1 + ...: Reinforcement Learning via Sequence Modeling | 1 + ...onditional Downscaling with Gaussian Processes | 1 + .../Deconvolutional Networks on Graph Data | 1 + ...g the Depth and Scope of Graph Neural Networks | 1 + ...y Complex Wordplay Puzzles as a Target for NLP | 1 + ...e and Efficient Exploration with Deep Networks | 1 + ...ssian Mixture Model for Constrained Clustering | 1 + .../neurips/Deep Contextual Video Compression | 1 + ...icit Duration Switching Models for Time Series | 1 + ...p Extended Hazard Models for Survival Analysis | 1 + ...xtrapolation for Attribute-Enhanced Generation | 1 + ...cy Evaluation in Continuous Treatment Settings | 1 + ...earning Through the Lens of Example Difficulty | 1 + ...: Finding Important Examples Early in Training | 1 + .../Deep Learning with Label Differential Privacy | 1 + ...ntation for High-Resolution 3D Shape Synthesis | 1 + ...ent Temporal and Spatial Analysis of fMRI Data | 1 + ...g via Fusing Physical and Chemical Information | 1 + ...Deep Networks Provably Classify Data on Curves | 1 + ...as Point Estimates for Deep Gaussian Processes | 1 + ...ication to Confounded Bandit Policy Evaluation | 1 + ...rning at the Edge of the Statistical Precipice | 1 + ...p Residual Learning in Spiking Neural Networks | 1 + ...issimilarities as Powerful Visual Fingerprints | 1 + ...e-Carlo Planning in Reconnaissance Blind Chess | 1 + ...n using selective backpropagation through time | 1 + ...of model smoothness in anisotropic Besov space | 1 + ...d Expectation-Maximization for Blind Inversion | 1 + ...nication Framework for Federated Deep Learning | 1 + ...omposition of What and When Across Time Scales | 1 + ...ameter-Efficient Convolutional Neural Networks | 1 + ... Highly Structured and Sparse Linear Transform | 1 + ...e Communication Latency for Federated Learning | 0 ...ds Practical Control in Cyber-Physical Systems | 1 + .../Demystifying and Generalizing BinaryConnect | 1 + data/2021/neurips/Denoising Normalizing Flow | 1 + .../Dense Keypoints via Multiview Supervision | 0 ...e Unsupervised Learning for Video Segmentation | 1 + .../neurips/Densely connected normalizing flows | 1 + ... Implicit Regularization and Sample Complexity | 1 + ...ments for Stochastic Contextual Linear Bandits | 1 + ...rfactual Generators using Deep Model Inversion | 1 + ... Event Sequences with Temporal Point Processes | 1 + ...on Unlabeled Data with Self-training Ensembles | 1 + ...tyle: Exploring Behavioral Stylometry in Chess | 1 + ...hlights in Videos via Natural Language Queries | 0 ...istribution Shifts in Bayesian Online Learning | 1 + ...al polynomials for sampling minibatches in SGD | 1 + ...BS: Differentiable Bayesian Structure Learning | 1 + ...ance Sampling and the Perils of Gradient Noise | 1 + ...lberg Models of Combinatorial Congestion Games | 1 + .../neurips/Differentiable Learning Under Triage | 1 + .../Differentiable Multiple Shooting Layers | 1 + ...ondecomposable Functions using Linear Programs | 1 + .../2021/neurips/Differentiable Quality Diversity | 1 + ...entiable Simulation of Soft Multi-body Systems | 1 + ...t-Descent for Training Spiking Neural Networks | 0 .../neurips/Differentiable Spline Approximations | 1 + ...ferentiable Synthesis of Program Architectures | 1 + ...d Feature Selection based on a Gated Laplacian | 1 + ...erentiable rendering with perturbed optimizers | 1 + ... Langevin Diffusion and Noisy Gradient Descent | 1 + ...Differential Privacy Over Riemannian Manifolds | 1 + ...ical Risk Minimization under the Fairness Lens | 1 + ...sian Optimization with Distributed Exploration | 1 + ...tially Private Learning with Adaptive Clipping | 1 + .../Differentially Private Model Personalization | 1 + ...ivate Multi-Armed Bandits in the Shuffle Model | 1 + ...erentially Private Sampling from Distributions | 1 + ... New Results in Convex and Non-Convex Settings | 1 + .../Differentially Private n-gram Extraction | 1 + .../Diffusion Models Beat GANs on Image Synthesis | 1 + data/2021/neurips/Diffusion Normalizing Flow | 1 + ...plications to Score-Based Generative Modeling" | 1 + .../Dimension-free empirical entropy estimation | 1 + ...sionality Reduction for Wasserstein Barycenter | 1 + ...ect Multi-view Multi-person 3D Pose Estimation | 1 + .../neurips/Directed Graph Contrastive Learning | 1 + .../2021/neurips/Directed Probabilistic Watershed | 1 + ...ve Latent Network Models Of Neural Populations | 1 + ... on Molecular Graphs via Synthetic Coordinates | 1 + ...rained Learning for Deep Graph Neural Networks | 1 + ...tworks with Hierarchical Voting Transformation | 1 + ...ions for Spatio-Temporal Graph Neural Networks | 1 + ...scovering and Achieving Goals via World Models | 1 + ...Discovery of Options via Meta-Learned Subgoals | 1 + .../neurips/Discrete-Valued Neural Communication | 1 + .../Disentangled Contrastive Learning on Graphs | 1 + ... from Noisy Data with Structured Nonlinear ICA | 1 + ...ion and the Prior in the Cold Posterior Effect | 1 + ...ncertainty Estimation Without Harming Accuracy | 1 + ...Process in Linear Graph Convolutional Networks | 1 + ...stilling Image Classifiers in Object Detectors | 1 + ...icit Drug Trafficker Detection on Social Media | 1 + ...tilling Object Detectors with Feature Richness | 1 + ...Adversarial Examples by Information Bottleneck | 1 + ...stributed Deep Learning In Open Collaborations | 1 + ...les per User: Sharp Rates and Phase Transition | 1 + ...achine Learning with Sparse Heterogeneous Data | 1 + ... Component Analysis with Limited Communication | 1 + ...ed Saddle-Point Problems Under Data Similarity | 0 ...ero-Order Optimization under Adversarial Noise | 1 + ...gression: discrete, continuous, and in between | 1 + ... for Learning Uncertain Neural Dynamics Models | 1 + ...earning for Multi-Dimensional Reward Functions | 1 + .../Distributionally Robust Imitation Learning | 1 + ..., Quantization Effects, and Frontier Integrals | 1 + ...Message Passing for Attribute with Heterophily | 1 + ...ve Learning with Strictly Proper Scoring Rules | 1 + ...Diversity Matters When Learning From Ensembles | 1 + ...ing Tasks Require Different Appearance Models? | 1 + ...t Gradients Highlight Discriminative Features? | 1 + ...ers Work? A Continuous Wasserstein-2 Benchmark | 1 + ...Really Perform Badly for Graph Representation? | 1 + ...ormers See Like Convolutional Neural Networks? | 1 + ...l Networks Really Help Adversarial Robustness? | 1 + .../Does Knowledge Distillation Really Work? | 1 + ...p Training Over-parameterized Neural Networks? | 1 + ...mitigate biases caused by subpopulation shift? | 1 + ...ation Learning: What Transformations to Learn? | 1 + ...n Learning with Domain Density Transformations | 1 + ...N: M sparse schemes from dense neural networks | 1 + ...ate Generative Models with Sinkhorn Divergence | 1 + ...Machine Learning for Dynamic Treatment Effects | 1 + ...n for Local Treatment Effects with Instruments | 1 + ...y Robust Thompson Sampling with Linear Payoffs | 1 + ... the strange case of off-policy policy updates | 1 + ...Are Found within Randomly Initialized Networks | 1 + ...rvised Approach for Generating Neural Activity | 1 + ...gnal Between Sequences While Dropping Outliers | 1 + ...se the Expressiveness of Graph Neural Networks | 1 + ...mizing the Adaptive Regret of Convex Functions | 1 + ...ation of Sparse Variational Gaussian Processes | 1 + ...ype Network for Generalized Zero-Shot Learning | 1 + .../Dual-stream Network for Visual Recognition | 1 + .../DualNet: Continual Learning, Fast and Slow | 1 + .../Dueling Bandits with Adversarial Sleeping | 1 + .../neurips/Dueling Bandits with Team Comparisons | 1 + ...ce Learning for Reversible Machine Translation | 1 + ...form for Holistic Next-Generation Benchmarking | 1 + ... via De-Sparsified Orthogonal Matching Pursuit | 1 + ...tleneck for Robust Self-Supervised Exploration | 1 + ...ure from a spatial-temporal transmission model | 1 + .../neurips/Dynamic Causal Bayesian Optimization | 1 + ...omain Few-Shot Recognition with Unlabeled Data | 1 + ...ynamic Grained Encoder for Vision Transformers | 1 + .../Dynamic Inference with Neural Interpreters | 1 + ...ders for High-Resolution Semantic Segmentation | 1 + ...ization and Relay for Video Action Recognition | 1 + data/2021/neurips/Dynamic Resolution Network | 1 + ...e Screening for Norm-Regularized Least Squares | 1 + data/2021/neurips/Dynamic Trace Estimation | 1 + ...ntiable Physics Models from Video and Language | 1 + data/2021/neurips/Dynamic influence maximization | 1 + ...ulti-agent communication with natural language | 1 + ...Transformers with Dynamic Token Sparsification | 1 + ...sserstein Barycenters for Time-series Modeling | 1 + ...entum Methods on Large-scale, Quadratic Models | 1 + ...inematic policy for egocentric pose estimation | 1 + .../neurips/E(n) Equivariant Normalizing Flows | 1 + ...xplaining Deep Reinforcement Learning Policies | 1 + ... Better, and Practically Faster Error Feedback | 1 + ...Efficient Infinite-Depth Graph Neural Networks | 1 + ...ploration through Learned Language Abstraction | 1 + ...arly Convolutions Help Transformers See Better | 1 + .../Early-stopped neural networks are consistent | 1 + .../Edge Representation Learning with Hypergraphs | 1 + ...EditGAN: High-Precision Semantic Image Editing | 1 + ...a classifier by rewriting its prediction rules | 1 + ...rization by Kernelized Proximal Regularization | 1 + ...sian Process Classification by Error Reduction | 1 + ... Neural Networks with General ReLU Activations | 1 + ...ture learning via local Markov boundary search | 1 + ...terialization and Offloading for Training DNNs | 1 + data/2021/neurips/Efficient Equivariant Network | 1 + ...ion, Allocation, and Triangular Discrimination | 1 + ...lization with Distributionally Robust Learning | 1 + ...ning of Discrete-Continuous Computation Graphs | 1 + ... Ascent Methods for Nonsmooth Minimax Problems | 1 + ...orward and Backward Propagation Sparsification | 1 + ... of Causal Effects by Deciding What to Observe | 1 + ...ssment of Neural Network Corruption Robustness | 1 + ...ining of Retrieval Models using Negative Cache | 1 + ...ing of Visual Transformers with Small Datasets | 1 + ... Linear Regression with Unknown Noise Variance | 1 + ...ficient and Accurate Gradients for Neural SDEs | 1 + .../Efficient and Local Parallel Random Walks | 1 + ...ned sampling via the mirror-Langevin algorithm | 1 + ...tio-temporal regression models in neuroimaging | 1 + ... random fields under sparse linear constraints | 1 + ...tifying Task Groupings for Multi-Task Learning | 1 + ...ng One Hidden Layer ReLU Networks From Queries | 0 ...iple of Loss Landscape of Deep Neural Networks | 1 + .../Emergent Communication of Generalizations | 1 + ...ication under Varying Sizes and Connectivities | 0 ...gent Discrete Communication in Semantic Spaces | 1 + ...via Just-in-Time Compilation and Vectorization | 1 + ...ge Style via Adversarial Feature Perturbations | 1 + ...volutional Features for Texture Representation | 1 + ...d Retriever for Open-Domain Question Answering | 1 + data/2021/neurips/End-to-End Weak Supervision | 1 + ...nd-to-end Multi-modal Video Temporal Grounding | 1 + ...ata-driven regularization for inverse problems | 1 + .../Ensembling Graph Predictions for AMR Parsing | 1 + ...ntropic Desired Dynamics for Intrinsic Control | 1 + ...Entropy-based adaptive Hamiltonian Monte Carlo | 1 + ...Zero-Shot Compositional Reinforcement Learning | 1 + ...ent Learning with Curiosity-driven Exploration | 1 + ...hines: The One-Sided Quasi-Perfect Equilibrium | 1 + ... the learning of Restricted Boltzmann Machines | 1 + data/2021/neurips/Equivariant Manifold Flows | 1 + ...Compensated Distributed SGD Can Be Accelerated | 1 + ...r compensation for variance reduced algorithms | 1 + ...s by a simple gradient-descent based algorithm | 1 + .../Escaping Saddle Points with Compressed SGD | 1 + ...radients of the Data Distribution by Denoising | 1 + ...reatment Effects via Single-cause Perturbation | 1 + ...ting the Long-Term Effects of Novel Treatments | 1 + ...the Unique Information of Continuous Variables | 1 + ...Performance Estimators of Neural Architectures | 1 + ...ion Attacks and Defenses in Federated Learning | 1 + ...Classification Models Against Bayes Optimality | 1 + ...el performance under worst-case subpopulations | 1 + ...ms for Learned and Rule-Based Agents in Hanabi | 1 + ...en Regularization Imposed by Self-Distillation | 1 + ...imodal Distributions in Deep Generative Models | 1 + ... Meta-Learning and Hyperparameter Optimization | 1 + ...Large-Scale Benchmark for Evolving Soft Robots | 1 + ...he Exponential Mechanism with Artificial Atoms | 1 + ...stributions of finite Bayesian neural networks | 1 + .../Excess Capacity and Backdoor Poisoning | 1 + ...o Vision with Cross-Modal Contrastive Learning | 1 + ...eter Optimization via Partial Dependence Plots | 1 + ...tent Representations with a Corpus of Examples | 1 + ...rhinal cortex with task-driven neural networks | 1 + ...sed Data Augmentation for Image Classification | 1 + ...eward Design for Reinforcement Learning Agents | 1 + ...e gradient descent training of neural networks | 1 + ...' Theorem to Compare Probability Distributions | 1 + ...in Secure Cross-Platform Social Recommendation | 1 + ...ific Features to Enhance Domain Generalization | 1 + ...ethods Globally: Adaptive Sample Size Approach | 1 + ... Under Utility Constraints in Sequential Games | 1 + ...ploiting a Zoo of Checkpoints for Unseen Tasks | 1 + ...od Structure for Source-free Domain Adaptation | 1 + ...petition: Convergence with Bounded Rationality | 1 + ...s of Adversarially Robust Deep Neural Networks | 1 + ...r Weakly-Supervised Audio-Visual Video Parsing | 1 + ...ensic Dental Identification with Deep Learning | 1 + ...riational Autoencoder for Interaction Modeling | 1 + ...ng the Limits of Out-of-Distribution Detection | 1 + ...unds for Risk-Sensitive Reinforcement Learning | 1 + ...ably Efficient for Decentralized Deep Training | 1 + ...Two Learning Models and Adversarial Robustness | 1 + ...al Networks with Differentiable Contact Models | 1 + ...ion-Aware Local Features by Learning to Deform | 1 + ...tored Multi-Agent Centralised Policy Gradients | 1 + .../FINE Samples for Learning with Noisy Labels | 1 + ...n Federated Learning from a Client Perspective | 1 + .../FLEX: Unifying Evaluation for Few-Shot NLP | 1 + ... Decomposed Near-field and Far-field Attention | 1 + ...ing Structure for Efficient Learning in MOMDPs | 1 + ...Algorithms for Multi-Agent Multi-Armed Bandits | 1 + ... Classification with Adversarial Perturbations | 1 + .../neurips/Fair Clustering Under a Bounded Cost | 1 + .../Fair Exploration via Axiomatic Bargaining | 1 + .../Fair Scheduling for Time-dependent Resources | 1 + ...ial Selection Using Supervised Learning Models | 1 + data/2021/neurips/Fair Sortition Made Transparent | 1 + ...n Invex Relaxation for a Combinatorial Problem | 1 + .../neurips/Fairness in Ranking under Uncertainty | 1 + .../Fairness via Representation Neutralization | 1 + ...g by Similarity-based Consistency Optimization | 1 + ...for Infinite-Horizon Markov Decision Processes | 1 + ...ance Using Concentration of Random Projections | 1 + ...Fast Axiomatic Attribution for Neural Networks | 1 + ...an Cox Processes via Path Integral Formulation | 1 + ...st Certified Robust Training with Short Warmup | 1 + ...artition Function with Weak Mixing Time Bounds | 1 + ...ructured Nonconvex-Nonconcave Minimax Problems | 1 + ...he Presence of Arbitrary Device Unavailability | 1 + ...rial Attacks through Adaptive Norm Constraints | 1 + ...ng for Extreme Multi-label Text Classification | 1 + ... Competitive Games with Entropy Regularization | 1 + ...cations to Sparse Regression in Bioinformatics | 1 + .../neurips/Fast Pure Exploration via Frank-Wolfe | 1 + ...ng in Congestion Games via Exponential Weights | 0 ...Stochastic Compositional Optimization Problems | 1 + ... Lumigraph Representations using Meta Learning | 1 + ...egative Tensors Using Mean-Field Approximation | 1 + ... Differentially Private-SGD via JL Projections | 1 + ... algorithms for low-rank tensor decompositions | 1 + ...ates for prediction with limited expert advice | 1 + ...dit Alignment for Automatic Speech Recognition | 1 + ...Lower Bounds for the Worst-Case Expected Error | 1 + ...ural Networks under Spherically Symmetric Data | 1 + .../neurips/Faster Matchings via Learned Duals | 1 + ...rk Training with Approximate Tensor Operations | 1 + ...n-asymptotic Convergence for Double Q-learning | 1 + ...mization using Jacobi-based eigenvalue methods | 1 + ...nforcement Learning with Theoretical Guarantee | 1 + ...for Nonconvex Federated Composite Optimization | 1 + ...rated Graph Classification over Non-IID Graphs | 1 + ..., Baselines, and Connections to Weight-Sharing | 1 + .../neurips/Federated Linear Contextual Bandits | 1 + ...Task Learning under a Mixture of Distributions | 1 + ...nstruction: Partially Local Federated Learning | 1 + ... Vision Transformer for COVID-19 CXR Diagnosis | 0 ...eterogeneity mitigation and variance reduction | 0 .../Few-Round Learning for Federated Learning | 1 + ...a-Driven Algorithms for Low Rank Approximation | 1 + ...t Detection via Association and DIscrimination | 1 + ... Segmentation via Cycle-Consistent Transformer | 1 + .../Finding Bipartite Components in Hypergraphs | 1 + ...pecific Degradations in Blind Super-Resolution | 1 + ...for Reducing Distortions of Hard-label Attacks | 1 + ...ion-Making via Expected Conditional Covariance | 1 + ...ing Input Features with Predictive Information | 1 + ...ero-Shot Learning with DNA as Side Information | 1 + ...zation Analysis of Inductive Matrix Completion | 0 ...of Average-Reward TD Learning and $Q$-Learning | 1 + ... TD-Learning via Generalized Bellman Operators | 1 + ...ith a differentiable spiking network simulator | 1 + ...ating Improvements in Machine-Learning Systems | 1 + ...der heterogeneous targets with Ordered Dropout | 1 + ... Projection Memory Benefits Continual Learning | 1 + ...vised Learning with Curriculum Pseudo Labeling | 1 + data/2021/neurips/Flexible Option Learning | 1 + ...for Non-Iterative Diverse Candidate Generation | 1 + ...Long-Range Interactions in Vision Transformers | 1 + ...s across covariates instead of across datasets | 1 + ...ess of Neural Networks to Weight Perturbations | 1 + ...ion-Forgetting Trade-off in Continual Learning | 1 + ...composition and Learning Halfspaces with Noise | 1 + ... Symbolic Languages for Model Interpretability | 1 + ...operties of Stochastic Optimization Algorithms | 1 + ... RNN as a kernel method: A neural ODE approach | 1 + ...lysis to Self-supervised Graph Neural Networks | 1 + ...e Re-Sampling Strategies in Stochastic Bandits | 0 ...andom forests and when they are Shapley values | 1 + ...orks for Parametric Image Restoration Problems | 1 + ...orcement Learning via Learned Fourier Features | 1 + ...ference based on Stochastic Process Generators | 1 + ...ledge Transfer for Low-resource Drug Discovery | 1 + .../Fuzzy Clustering with Similarity Queries | 1 + ... Private Aggregation of Teacher Discriminators | 1 + ...t Representations without Iterative Refinement | 1 + ... Network for Multi-agent Trajectory Prediction | 1 + ...ment Reconstruction from Point Cloud Sequences | 1 + data/2021/neurips/Gauge Equivariant Transformer | 1 + ...re Network for Single Image Defocus Deblurring | 1 + ...irectional Graph Neural Networks for Molecules | 1 + ...ization: Geometric Analysis and Sharper Bounds | 1 + ...neral Nonlinearities in SO(2)-Equivariant CNNs | 1 + ... from Observation via Inferring Goal Proximity | 1 + .../Generalizable Multi-linear Attention Network | 1 + ...ta-Learning: An Information-Theoretic Analysis | 1 + ...n Bounds for (Wasserstein) Robust Optimization | 1 + ... Using Negative Sampling: Linear vs Hyperbolic | 0 ...a-Learning via PAC-Bayes and Uniform Stability | 1 + ...e Crossover from the Noiseless to Noisy Regime | 1 + ...ization Guarantee of SGD for Pairwise Learning | 1 + ...earning Algorithms: Recurring and Unseen Tasks | 1 + ...eighting via Class-Level Gradient Manipulation | 1 + ...rsarially Robust and Efficient Neural Networks | 1 + ...Divergence Loss for Learning with Noisy Labels | 1 + ...Linear Bandits with Local Differential Privacy | 1 + ...Proximal Policy Optimization with Sample Reuse | 1 + ...alized Shape Metrics on Neural Representations | 1 + ...bject Detection via SVD-Dictionary Enhancement | 1 + ... Navigation in Partially-Revealed Environments | 1 + ...cy Fields for 3D Surface-Aware Image Synthesis | 1 + ...native: Rethinking The Meta-Continual Learning | 1 + ...eric Neural Architecture Search via Regression | 1 + ...Generation of Molecular 3D Conformer Ensembles | 1 + .../Geometry Processing with Neural Fields | 1 + .../neurips/Glance-and-Gaze Vision Transformer | 1 + ...t for Asymmetric Low-Rank Matrix Factorization | 1 + ...ization for Nonlinear Model Predictive Control | 1 + ...librium in Classes of Nonconvex Zero-Sum Games | 1 + ...lobal Filter Networks for Image Classification | 1 + ...am Search for Neural Abstractive Summarization | 1 + ...Sample Efficient Neural Function Approximation | 1 + ...formers with Recurrent Fast Weight Programmers | 1 + ... Neural Active Learning with Fisher Embeddings | 1 + ...d Classification Measures and How to Find Them | 1 + ...cation Using Gradients for Task Representation | 1 + ...ral Networks for Stable and Efficient Training | 1 + ... Nets: Margin Maximization and Simplicity Bias | 1 + ...tee Fairness in Collaborative Machine Learning | 1 + ...Gradient Inversion with Generative Image Prior | 1 + ...t Image Corruption for Learning-based Steering | 1 + ...amples for Online Task-free Continual Learning | 1 + ...Hyperparameter Optimization Over Long Horizons | 1 + ...daptation without Indexed Intermediate Domains | 1 + .../Grammar-Based Grounded Lexicon Learning | 1 + .../Graph Adversarial Self-Supervised Learning | 1 + ...le Architecture Search with Structure Learning | 1 + .../Graph Neural Networks with Adaptive Residual | 1 + ...ph Neural Networks with Local Graph Parameters | 1 + ...Predictive Uncertainty for Node Classification | 1 + ...s for Representation Learning on Textual Graph | 1 + .../Graphical Models in Heavy-Tailed Markets | 1 + ...ithms for Active Sequential Hypothesis Testing | 1 + ...s with Faster Explicit Superlinear Convergence | 1 + ...ntation Similarity Through Statistical Testing | 0 ...ing Spatio-Temporal Language with Transformers | 1 + ...ages: invariance stems from variations in data | 1 + data/2021/neurips/Group Equivariant Subsampling | 1 + ...nd Temporal Reconstruction of Humans in Motion | 1 + ...bal Parameters for Neural Posterior Estimation | 1 + ...esolution Vision Transformer for Dense Predict | 0 ...antic-Visual Adaptation for Zero-Shot Learning | 1 + ...ing Home Assistants to Rearrange their Habitat | 1 + ...with Non-Newtonian Momentum for Rapid Sampling | 1 + ... Long-tailed Feature Distribution in AdderNets | 1 + ...rd-Attention for Scalable Image Classification | 1 + ...t Latency Prediction for NAS via Meta-Learning | 1 + .../neurips/Hash Layers For Large Sparse Models | 1 + .../Heavy Ball Momentum for Conditional Gradient | 1 + ...vy Ball Neural Ordinary Differential Equations | 1 + ...essibility of Overparametrized Neural Networks | 1 + ...igenspectra of More Realistic Nonlinear Models | 1 + ...ed Bandits: Closing the Gap and Generalization | 1 + .../Heuristic-Guided Reinforcement Learning | 1 + ...: O(1)-Approximation for Well-Clustered Graphs | 1 + ...cal Reinforcement Learning with Timed Subgoals | 1 + .../Hierarchical Skills for Efficient Exploration | 1 + ...ds for Line Search Based on Stochastic Oracles | 1 + ...onvex Stochastic Optimization with Heavy Tails | 1 + ...to Capture Filtrations of Stochastic Processes | 1 + ...g: Experience Replay for Sparse Reward Meta-RL | 1 + ...Transformer for Vision-and-Language Navigation | 1 + ...tive RL and Fragment-based Molecule Generation | 1 + ...ion affects Optimization for Linear Regression | 1 + data/2021/neurips/How Does it Sound? | 0 ...Fine-Tuning Allows for Effective Meta-Learning | 1 + ...ule Networks Be for Systematic Generalization? | 1 + ...ance Predictors in Neural Architecture Search? | 1 + ... Be Fine-Tuned Towards Adversarial Robustness? | 1 + ...ght Can PAC-Bayes be in the Small Data Regime? | 1 + ...pport Causal Understanding of CNN Activations? | 1 + ...n classical multidimensional scaling go wrong? | 1 + ...tecture Impact its Robustness to Noisy Labels? | 1 + ...c reasoning knowledge to learn new algorithms? | 1 + .../Human-Adversarial Visual Question Answering | 1 + ...al Semi-Bandits and Adversarial Linear Bandits | 1 + ... Compact and Expressive Probabilistic Circuits | 1 + ...rbolic Busemann Learning with Ideal Prototypes | 1 + ... Procrustes Analysis Using Riemannian Geometry | 1 + ... and Community Selection for Objects Retrieval | 1 + ...timization Is Deceiving Us, and How to Stop It | 1 + ...yperparameter Tuning is All You Need for LISTA | 1 + ...edge Graph Completion Using Pair-Wise Encoding | 1 + ...Q-Learn: Inverse soft-Q Learning for Imitation | 1 + ...n't: A test case of natural language inference | 1 + ...ntifiability in inverse reinforcement learning | 1 + ...dels for Missing Not at Random Data Imputation | 1 + ...servational Studies with Covariate Information | 1 + ...s for the Non-Gaussian and Heterogeneous Cases | 1 + ...ized Condorcet Winner in Multi-dueling Bandits | 1 + ...ing Natural Out-of-Context Prediction Problems | 1 + .../neurips/Identity testing for Mallows model | 1 + ...Image Generation using Continuous Filter Atoms | 1 + ...l Diffusion for Autoregressive Image Synthesis | 1 + ...ally Elastic Stochastic Differential Equations | 1 + .../neurips/Imitation with Neural Density Models | 1 + ... Networks: a Provable Benefit of Stochasticity | 1 + ...-Based Experimental Design without Likelihoods | 1 + ...ptimal Algorithms for Stochastic Shortest Path | 1 + data/2021/neurips/Implicit Generative Copulas | 1 + ...ough Discrete Exponential Family Distributions | 1 + ...arization in Matrix Sensing via Mirror Descent | 1 + ...Implicit SVD for Graph Representation Learning | 1 + ...sponse Alignment for Partial Domain Adaptation | 1 + ...zation: The Impact of Depth and Early Stopping | 1 + ...ncy Measure for Unsupervised Domain Adaptation | 1 + ...reen Content Image Continuous Super-Resolution | 1 + ...presentation learning with synaptic plasticity | 1 + ...arallel Tree Search with Off-Policy Correction | 1 + ...Algorithms for Power Means in Euclidean Spaces | 1 + ... via new Ordered Contention Resolution Schemes | 1 + ...pe SVM with Sparse Multi-Kernel Representation | 1 + ...Regret Bounds for Tracking Experts with Memory | 1 + ... Robustness for Fine-tuning in Neural Networks | 1 + .../Improved Transformer for High-Resolution GANs | 1 + ...Sets for Linear Bandits and Linear Mixture MDP | 1 + ...scaded Networks and a Temporal-Difference Loss | 1 + ...h the Relationship with Adversarial Robustness | 1 + ...els with Dual-System, Neuro-Symbolic Reasoning | 1 + ...Networks by Decoding Representations to Inputs | 1 + ...l Reinforcement Learning via Stored Embeddings | 1 + ...al Coverage via Orthogonal Quantile Regression | 1 + ...ing on Imbalanced Data via Open-World Sampling | 0 ...g Interpretability by Saliency Guided Training | 1 + ...h Imaginary Tasks from Latent Dynamics Mixture | 1 + .../Improving Robustness using Generated Data | 1 + ...ith Automated Unsupervised Outlier Arbitration | 1 + ...ations via Augmentation-Aware Self-Supervision | 1 + ...s by A Token-based Generator with Transformers | 1 + ... in VAE latent space using decoder uncertainty | 1 + ...cs Organized by Astrocyte-modulated Plasticity | 1 + ...ype Propagation for Zero-Shot Compositionality | 1 + ...Independent mechanism analysis, a new concept? | 1 + ...imum Empirical Divergence for Unimodal Bandits | 1 + ... Privacy Accounting via a R\303\251nyi Filter" | 1 + ...ime Horizon Safety of Bayesian Neural Networks | 1 + ...tterns for Explaining Information Flow in BERT | 1 + ...: Information-Aware Graph Contrastive Learning | 1 + ...ted Reward Learning for Reinforcement Learning | 1 + ...on Directed Sampling for Sparse Linear Bandits | 1 + ...wer: Intrinsic Control via Information Capture | 1 + ...on: can adaptive processing of gradients help? | 1 + ...ation bounds for black-box learning algorithms | 1 + ...al Knowledge Distillation for Object Detection | 1 + data/2021/neurips/Instance-Conditioned GAN | 1 + ...Lipschitz Optimization with Error Certificates | 1 + .../Instance-Dependent Partial Label Learning | 1 + ...noise Learning under a Structural Causal Model | 1 + ...mal Mean Estimation Under Differential Privacy | 1 + ...ral ODEs: Pharmacology and Disease Progression | 1 + ...ee Path in Transformer for Code Representation | 1 + ...Label Cleaning with Example-based Explanations | 1 + ...ious Agent: Learning Task-Agnostic Exploration | 1 + ... Momentum Contrastive Self Supervised Learning | 1 + ...ust generalization even when there is no noise | 1 + ...generic visual processor emerging on the side) | 1 + ... Quality of DNNs for 3D Point Cloud Processing | 1 + ... Inference with Tractable Probabilistic Models | 1 + .../Intriguing Properties of Contrastive Losses | 1 + .../Intriguing Properties of Vision Transformers | 1 + ...Homology and Generalization in Neural Networks | 1 + ...ive Distillation for Robust Question Answering | 1 + ...tleneck for Out-of-Distribution Generalization | 1 + ... Imitation Learning for Generalizable Policies | 1 + ...aracteristics of the Human Sensorimotor System | 1 + ...raging Pre-trained Contrastive Representations | 1 + ... Continuous State Space with Formal Guarantees | 1 + data/2021/neurips/Inverse-Weighted Survival Games | 1 + ...nvertible DenseNets with Concatenated LipSwish | 1 + ...valuation Broken? The Incoherence of Coherence | 1 + ...ing Continuous Control with Bernoulli Policies | 1 + ...s for Convergent Solutions to Inverse Problems | 1 + .../Iterative Amortized Policy Optimization | 1 + ...sence of Latent Confounders and Selection Bias | 1 + ...Connecting Probability Estimation for Networks | 1 + ...hetic Data: Unifying Framework and New Methods | 1 + .../2021/neurips/Iterative Teacher-Aware Learning | 1 + .../neurips/Iterative Teaching by Label Synthesis | 1 + ...is Pursuit with Global Linear Convergence Rate | 1 + ...eural Network Depth and Dropout Regularization | 1 + ...jects and Relations for Scene Graph Generation | 1 + ...akly Supervised RGB-D Salient Object Detection | 1 + ...and input optimization in equilibrium networks | 1 + .../K-Net: Towards Unified Image Segmentation | 1 + ...Reasoning for Zero-Shot Coordination in Hanabi | 1 + ...t Flow for Probabilities with Disjoint Support | 1 + ...er Incomplete Graphs via Graphs Neural Network | 1 + ...ll: Trajectory Attention in Video Transformers | 1 + data/2021/neurips/Kernel Functional Optimisation | 1 + .../Kernel Identification Through Transformers | 1 + .../Kernelized Heterogeneous Risk Minimization | 1 + data/2021/neurips/Knowledge-Adaptation Priors | 1 + ...pired 3D Scene Graph Prediction in Point Cloud | 1 + ...ks via Efficient in-situ Subspace Optimization | 1 + ...tion via Augmentation for Deep Active Learning | 1 + ...al Systems that Generalize Across Environments | 1 + ...ti-purpose Learnt Low-dimensional Binary Codes | 1 + ...ve Hashing Accelerated Simulation and Learning | 1 + ...tition-based Extreme Multilabel Classification | 1 + ...se SGD Provably Prefers Flat Global Minimizers | 1 + ...onsistency in overfitted generalized $k$-means | 1 + ...tive Classification under Overparameterization | 1 + ...etworks for Multi-Node Representation Learning | 1 + ...eration in Hierarchical Reinforcement Learning | 1 + ...gation with Fine-Grained Alignment Supervision | 1 + ...improved power method for tensor decomposition | 1 + ...f the effects of mutations on protein function | 1 + ...lace Redux - Effortless Bayesian Deep Learning | 1 + ...aphs: New Benchmarks and Strong Simple Methods | 1 + ...ith Fourier Features and Tensor Decompositions | 1 + .../Large-Scale Unsupervised Object Discovery | 1 + .../Large-Scale Wasserstein Gradient Flows | 1 + ... for Least-Squares in the Interpolation regime | 1 + ...st-iterate Convergence in Extensive-Form Games | 1 + ...fast computation with arbitrarily slow neurons | 1 + ...ram Synthesis Beyond Domain-Specific Languages | 1 + ...tent Matters: Learning Deep State-Space Models | 1 + .../Lattice partition recovery with dyadic CART | 1 + ...ty of Linear Thresholds from Label Proportions | 1 + ... Multi-dimensional Spatial Positional Encoding | 1 + ...pproach for High-Dimensional Outlier Detection | 1 + ...Correspondence via Canonical Point Autoencoder | 1 + ...nt Learning with Zero Training-time Violations | 1 + ...resentation for Out-of-Distribution Prediction | 1 + ...ive Policies to Solve NP-hard Routing Problems | 1 + ...al Networks using DiscriminAtive Masking (DAM) | 1 + ...ning Conjoint Attentions for Graph Neural Nets | 1 + ...entation via Disentangled Feature Augmentation | 1 + ...gled Representations for Semantic Segmentation | 1 + .../Learning Disentangled Behavior Embeddings | 1 + ...Collaboration Graph for Multi-Agent Perception | 1 + ...Diverse Policies in MOBA Games via Macro-Goals | 1 + ...Representations in Goal-conditioned Block MDPs | 1 + ...rain Connectome with Spatio-Temporal Attention | 1 + ...ibria in Matching Markets from Bandit Feedback | 1 + ...Equivariant Stein Variational Gradient Descent | 1 + .../Learning Fast-Inference Bayesian Networks | 1 + ...omain Approximation for Binary Neural Networks | 1 + ...Models: Precise Asymptotics in High-dimensions | 1 + ...rning Generalized Gumbel-max Causal Mechanisms | 1 + ...rgy-Based Latent Space for Saliency Prediction | 1 + .../2021/neurips/Learning Graph Cellular Automata | 1 + ...ing Graph Models for Retrosynthesis Prediction | 1 + ...zation Problems: A Data Generation Perspective | 1 + ...ject Detection via Kullback-Leibler Divergence | 1 + ... Rule Sets: A Submodular Optimization Approach | 1 + ...aph-based World Models of Textual Environments | 1 + ...hborhood Search Policy for Integer Programming | 1 + ...tial Decision Making by Reinforcement Learning | 0 ...e Abstractions for Deep Reinforcement Learning | 1 + .../Learning Models for Actionable Recourse | 1 + ...etric Volterra Kernels with Gaussian Processes | 1 + .../Learning Optimal Predictive Checklists | 1 + ...nded Constraint Violation for Constrained MDPs | 1 + ...annian metric for disease progression modeling | 1 + ...tterns of Human Brain across Many fMRI Studies | 1 + ...tic Representations to Verify Hardware Designs | 1 + ...ing Signal-Agnostic Manifolds of Neural Fields | 1 + .../Learning Space Partitions for Path Planning | 1 + ...artially Observed or Delayed Dynamical Systems | 1 + ...rom Random Deep Action-conditional Predictions | 1 + ...by Minimizing a PAC-Bayes Generalization Bound | 1 + ...ly Teacher Networks for Knowledge Distillation | 1 + ...xplain Generalisation in Graph Neural Networks | 1 + ...earning Transferable Adversarial Perturbations | 1 + ...Cloud Detection via 3D Contrastive Co-training | 1 + ...s in Panels with General Intervention Patterns | 1 + ...Representation for Deep Reinforcement Learning | 1 + ...Single Neuron with Bias Using Gradient Descent | 1 + .../neurips/Learning and Generalization in RNNs | 1 + ...ealistic datasets with a teacher-student model | 1 + ...ing and Reasoning for Video Question Answering | 1 + ... in Multi-Stage Decentralized Matching Markets | 1 + ...erative Configurable Markov Decision Processes | 1 + ...ly observable Markov games with perfect recall | 0 ...l trajectories via augmented behavioral models | 1 + ...rning latent causal graphs via mixture oracles | 1 + ...ficient for Estimating (Some) Graph Parameters | 1 + ...t attractor structure in decision-making tasks | 1 + ...imal Tikhonov regularizer for inverse problems | 1 + ...ent Domains for Adaptive Semantic Segmentation | 1 + ...ing to Assimilate in Chaotic Dynamical Systems | 1 + ...Example Solutions for Neural Program Synthesis | 1 + .../neurips/Learning to Compose Visual Relations | 1 + ...Draw: Emergent Communication through Sketching | 1 + data/2021/neurips/Learning to Elect | 1 + ...niversal Plan-Conditioned Policies in Robotics | 1 + ...a Pixel-level Noise-aware Adversarial Training | 1 + ...nerate Visual Questions with Noisy Supervision | 1 + ...nd Multi-Agent Communication with Autoencoders | 1 + ...ems with Dual-Aspect Collaborative Transformer | 1 + ...Dense Gaussian Processes for Few-Shot Learning | 1 + .../neurips/Learning to Learn Graph Topologies | 1 + ... Predict Trustworthiness with Steep Slope Loss | 1 + ...ing to Schedule Heuristics in Branch and Bound | 1 + .../neurips/Learning to See by Looking at Noise | 1 + ...enous Events for Marked Temporal Point Process | 1 + ...ms as Interpretable and Generalizable Policies | 1 + ...al Networks Through the Information Bottleneck | 1 + .../neurips/Learning to dehaze with polarization | 1 + ...ng to delegate for large-scale vehicle routing | 1 + ...adient sparsity in meta and continual learning | 1 + ...rithmic Supervision via Continuous Relaxations | 1 + ...rning with Holographic Reduced Representations | 1 + .../Learning with Labeling Induced Abstentions | 1 + ... Noisy Correspondence for Cross-modal Matching | 1 + .../2021/neurips/Learning with User-Level Privacy | 1 + ...with Multiple States via New Ski Rental Bounds | 1 + ...learn non-convex piecewise-Lipschitz functions | 1 + .../Least Square Calibration for Peer Reviews | 1 + ...ath for Cross-Domain Cold-Start Recommendation | 1 + ... Approximate Inference in Combinatorial Spaces | 1 + ...Level Object Pose Estimation from Point Clouds | 1 + ...ral Correlations in Sparsified Mean Estimation | 1 + ...Language Models for Abstract Textual Reasoning | 1 + ...ptation via Consolidated Internal Distribution | 1 + ...presentations with Single-Evaluation Rendering | 1 + ...layer neural networks with mean field training | 1 + ...ling Client Heterogeneity and Sparse Gradients | 1 + ... High-dimensional Vector Autoregressive Models | 1 + ...aming Model: Improved Bounds for Heavy Hitters | 1 + ...babilistic Solution of Boundary Value Problems | 0 ... Synthesis with Visual Context Attentional GAN | 1 + ...t-Decodable Mean Estimation in Nearly-PCA Time | 2 ++ ...lestone Classes are Privately Online Learnable | 1 + ... Regret Minimization in Reinforcement Learning | 1 + ...o-Encoders Using Jacobian $L_1$ Regularization | 1 + ...al Explanation of Dialogue Response Generation | 1 + data/2021/neurips/Local Hyper-Flow Diffusion | 1 + ...ure Learning in Neural Networks Beyond Kernels | 1 + ... using self-supervised contrastive predictions | 1 + ...Local policy search with Bayesian optimization | 1 + data/2021/neurips/Locality Sensitive Teaching | 1 + ...ity in convolutional teacher-student scenarios | 1 + .../neurips/Localization with Sampling-Argmax | 1 + .../Localization, Convexity, and Star Aggregation | 1 + ...ibution Detection using Deep Generative Models | 1 + ... Prediction Intervals for Deep Learning Models | 1 + ...ation of functionals of discrete distributions | 0 .../Locally private online change point detection | 1 + .../Logarithmic Regret from Sublinear Hints | 1 + ...ithmic Regret in Feature-based Dynamic Pricing | 1 + ...t-Term Transformer for Online Action Detection | 1 + ...Efficient Transformers for Language and Vision | 1 + ...rounding of Narrations in Instructional Videos | 1 + ...anations with Sobol-based Sensitivity Analysis | 1 + ...for Contrastive Semantic Segmentation Learning | 1 + ... application to particle variational inference | 1 + .../Lossy Compression for Lossless Prediction | 1 + ... Optimization for Temporal Action Localization | 1 + ...raints for Fast Inference in Structured Models | 1 + ...ooth and Low-Rank Matrix Optimization Problems | 1 + data/2021/neurips/Low-Rank Subspaces in GANs | 1 + ...epresentations is Reflected in Brain Responses | 1 + ...alized Optimization Over Time-Varying Networks | 1 + ...ing Methods for Well-Conditioned Distributions | 1 + ... the Pseudo-Dimension of Tensor Network Models | 1 + .../neurips/Luna: Linear Unified Nested Attention | 1 + ...ree Approximations of Second-Order Information | 1 + ...via Maximizing Deviation from Explored Regions | 1 + ...g with a Team of Reinforcement Learning Agents | 1 + ...ion-Aware Unit for Video Prediction and Beyond | 1 + ...Text and Human Text using Divergence Frontiers | 1 + ...nference via Uncorrected Hamiltonian Annealing | 1 + ...LOT: Multimodal Neural Script Knowledge Models | 1 + ...Economic Sparse Training Framework on the Edge | 1 + ...state similarity for Markov decision processes | 1 + ...mputation via Learning Missing Data Mechanisms | 1 + .../MLP-Mixer: An all-MLP Architecture for Vision | 1 + ...OMA: Multi-Object Multi-Actor Activity Parsing | 1 + ...pervised Transformer for Visual Representation | 1 + ...g for Variance Reduction in Online Experiments | 1 + ...rackets for forecasting irreversible processes | 1 + ...Attention in Deep Reinforcement Learning Tasks | 1 + .../MagNet: A Neural Network for Directed Graphs | 1 + ...ork for Verifying Probabilistic Specifications | 1 + ...terfactual) Difference One Rationale at a Time | 1 + ...online learning for optimal allocation of time | 1 + ...ence: a Framework for Comparing Data Manifolds | 1 + .../Manipulating SGD with Data Ordering Attacks | 1 + ...Online Multiclass Learning via Convex Geometry | 1 + ...alised Gaussian Processes with Nested Sampling | 1 + .../MarioNette: Self-Supervised Sprite Learning | 1 + .../Mastering Atari Games with Limited Data | 1 + ...a Desired Causal State via Shift Interventions | 1 + ...networks for neural combinatorial optimization | 1 + ...on and the interpretation of geodesic distance | 1 + ...ihood Training of Score-Based Diffusion Models | 1 + ...easuring Generalization with Optimal Transport | 1 + ...ng to Identify High-Risk States and Treatments | 1 + ...mory Efficient Meta-Learning with Large Images | 1 + ...ithms for Max-k-Cut and Correlation Clustering | 1 + ...t Patch-based Inference for Tiny Deep Learning | 1 + data/2021/neurips/Meta Internal Learning | 1 + ...Meta Learning Backpropagation And Improving It | 1 + ...Learning Kernels for Testing with Limited Data | 1 + ...ptive Nonlinear Control: Theory and Algorithms | 1 + ...Learning Reliable Priors in the Function Space | 1 + ...earning Sparse Implicit Neural Representations | 1 + ...Learning for Relative Density-Ratio Estimation | 1 + ...ck-Box Random Search Based Adversarial Attacks | 1 + ...-Learning via Learning with Distributed Memory | 0 .../neurips/Meta-learning to Improve Pre-training | 1 + .../Meta-learning with an Adaptive Task Scheduler | 1 + ...ble Clothed Human Models from Few Depth Images | 1 + ...Task Bandits with Bayesian Hierarchical Models | 1 + ...gs Data Augmentation for Graph Neural Networks | 1 + ...poral Generalization in Neural Language Models | 1 + ...ent Slot Set Encoder for Scalable Set Encoding | 1 + ...hods for Stochastic Weakly Convex Optimization | 1 + ...arial Regret via Root-Logarithmic Regularizers | 1 + .../Minimax Regret for Stochastic Shortest Path | 1 + ...ent in Social Networks via Link Recommendation | 1 + ...efits of Two-stage and One-stage HOI Detection | 1 + ...gevin Monte Carlo: the Case Under Isoperimetry | 1 + ...specified Gaussian Process Bandit Optimization | 1 + ...earning via Offline Data With Partial Coverage | 0 ...ine Continual Learning with Neuron Calibration | 1 + ...fer via Distillation of Activated Channel Maps | 1 + ... Forecasting with Microscopic Time Series Data | 1 + ...nt: Fast online multiclass logistic regression | 1 + ...ransferring Mask Prior and Semantic Similarity | 1 + ... Estimation and PU Learning: A Modern Approach | 1 + ...ion for Alpha-Divergence Variational Inference | 1 + ...ased Imitation Learning From Observation Alone | 1 + ...tory Forecasting for Human Mobility Prediction | 1 + .../Modality-Agnostic Topology Aware Localization | 1 + ...pervised Domain Adaptation without Source Data | 1 + .../Model Selection for Bayesian Autoencoders | 1 + ...n of gradient flow in the random feature model | 1 + .../neurips/Model-Based Domain Generalization | 1 + ...pisodic Memory Induces Dynamic Hybrid Controls | 1 + ...t Learning via Imagination with Derived Memory | 1 + ...rchies with Relation-specific Hyperbolic Cones | 1 + .../Modified Frank Wolfe in Probability Space | 1 + ...dular Gaussian Processes for Transfer Learning | 1 + ...nchronous Update for Adaptive Gradient Methods | 1 + ...h With Iteratively Refining State Abstractions | 1 + ...k (MA): A New Potential Risk of Screen Photos" | 0 ...Knowledge Distillation from Out-of-Domain Data | 1 + ...ergence-based Generative Modeling on Manifolds | 1 + ...d Training on Heterogeneous Unreliable Devices | 1 + ...sed Learning for Molecular Property Prediction | 1 + ...Voltage Control on Power Distribution Networks | 1 + ...ement Learning in Stochastic Networked Systems | 1 + ...est-Arm Identification and Regret Minimization | 1 + ...ulti-Facet Clustering Variational Autoencoders | 1 + ...abel Learning with Pairwise Relevance Ordering | 1 + data/2021/neurips/Multi-Objective Meta Learning | 1 + ...ovement with Safety Constraints in Finite MDPs | 1 + ...otion Prediction with Multi-Range Transformers | 1 + ...ulti-Scale Representation Learning on Proteins | 1 + ...ian Optimization with Unknown Evaluation Costs | 1 + ...ation Learning via Total Correlation Objective | 1 + ...-armed Bandit Requiring Monotone Arm Sequences | 1 + ...lti-modal Dependency Tree for Video Captioning | 1 + ...ask Learning of Order-Consistent Causal Graphs | 1 + .../Multi-view Contrastive Graph Clustering | 1 + ...ticlass Boosting and the Cost of Weak Learning | 1 + ...sus Binary Differentially Private PAC Learning | 1 + ...re-training with Universal Dependency Learning | 1 + ... Few-Shot Learning with Frozen Language Models | 1 + .../neurips/Multimodal Virtual Point 3D Detection | 1 + ...ngual Embeddings for Large-Scale Speech Mining | 1 + ... Descent: Design Your Own Generalization Curve | 1 + ...d Operator Learning for Differential Equations | 1 + ...NAS-Bench-x11 and the Power of Learning Curves | 1 + ...ing on the Orbits of a Deterministic Transform | 0 ...tion Problems on Geometric Intersection Graphs | 1 + ...ality Assessment using Non-Matching References | 1 + ...mization using Implicit Neural Representations | 1 + ...o the Best Policy in Markov Decision Processes | 1 + ... for Sparse-view 3D Reconstruction in the Wild | 1 + .../NeRV: Neural Representations for Videos | 1 + .../Near Optimal Policy Optimization via REPS | 1 + ...nvex Optimization For All Orders of Smoothness | 1 + ...erimental Design for Causal Structure Learning | 1 + ...ar-Optimal No-Regret Learning in General Games | 1 + ...rcement Learning via Double Variance Reduction | 1 + ...thms for Learning Non-Linear Dynamical Systems | 1 + ...ly Horizon-Free Offline Reinforcement Learning | 1 + ...mal Reinforcement Learning for Discounted MDPs | 1 + ...blivious Algorithms for Explainable Clustering | 1 + ... causal graphical models with hidden variables | 1 + .../Neighborhood Reconstructing Autoencoders | 1 + ...ware Graph Neural Networks for Link Prediction | 1 + ...ification from Arbitrary Surrogate Experiments | 1 + data/2021/neurips/Nested Graph Neural Networks | 1 + data/2021/neurips/Nested Variational Inference | 1 + ...orcing Occam's Razor to Improve Generalization | 1 + ...Volume Rendering for Multi-view Reconstruction | 1 + ...Index Network For Restless Bandits Via Deep RL | 1 + ...al Active Learning with Performance Guarantees | 1 + ...nterpretable Machine Learning with Neural Nets | 1 + ...al Algorithmic Reasoners are Implicit Planners | 1 + ...ng Speech from Self-Supervised Representations | 1 + ...al Auto-Curricula in Two-Player Zero-Sum Games | 0 ...h Neural Network Framework for Link Prediction | 1 + data/2021/neurips/Neural Bootstrapper | 1 + ... Circuit Synthesis from Specification Patterns | 1 + ...l Distance Embeddings for Biological Sequences | 1 + ...ubber: Dubbing for Videos According to Scripts | 1 + ...h for Uncertainty Estimation and Dataset Shift | 1 + ...al Flows: Efficient Alternative to Neural ODEs | 1 + ...adiance Fields for Human Performance Rendering | 1 + ...With Multiple Modes and Stochastic Transitions | 1 + ...the Role of Stochasticity in Robust Perception | 1 + data/2021/neurips/Neural Production Systems | 1 + ...ural Program Generation Modulo Static Analysis | 1 + ...seudo-Label Optimism for the Bank Loan Problem | 1 + ...ral Taskonomy at Scale in Rodent Visual Cortex | 1 + ...ural Relightable Participating Media Rendering | 1 + data/2021/neurips/Neural Routing by Memory | 1 + ... Machine For Transformer-Based Text Generation | 1 + data/2021/neurips/Neural Scene Flow Prior | 1 + ...tonian Equations on General Coordinate Systems | 1 + ...Neural Tangent Kernel Maximum Mean Discrepancy | 1 + .../neurips/Neural Trees for Learning on Graphs | 1 + ...r Semi-Supervised Few-Shot Learning of 3D Pose | 1 + ...mal feedback control with local learning rules | 1 + ...egrated Lighting for Reflectance Decomposition | 1 + ...tic for Solving the Traveling Salesman Problem | 1 + ...Reliable Route Recommendation on Road Networks | 1 + ...Full Batch (in Stochastic Convex Optimization) | 1 + ...hout Trade-offs for the Sketched Newton Update | 1 + ...ation for Federated Learning with Non-IID Data | 1 + ...ation: Learning to Navigate without Navigating | 1 + .../No Regrets for Learning the Prior in Bandits | 1 + data/2021/neurips/No-Press Diplomacy from Scratch | 1 + ...gret Online Learning over Riemannian Manifolds | 1 + ...nt Local Smoothing for Scalable Graph Learning | 1 + ...rks: meta-learning useful conserved quantities | 1 + ...: Role of Symmetry Breaking in Neural Networks | 1 + ...upervised Image Denoising without Clean Images | 1 + ...03\251vy Flights in Attractor Neural Networks" | 1 + data/2021/neurips/Noisy Recurrent Neural Networks | 1 + ...TDC with General Smooth Function Approximation | 1 + ...ian Gaussian Processes for Few-Shot Regression | 1 + ...hs via Discrete Difference of Convex Algorithm | 1 + ...asymptotic Error Bounds for Bidirectional GANs | 1 + ...r Wasserstein approximation using point clouds | 1 + ...y Robust Optimization: Non-asymptotic Analysis | 1 + ...ion for Self-Adaptive 3D Human Pose Estimation | 1 + ...imation of continuous DPPs with kernel methods | 1 + ...ntiation for Machine-Learning and Optimization | 1 + ... and Log Odds Correction with Rare Events Data | 1 + ...c Transformers for Efficient Image Recognition | 1 + ...ers are Robust in Graph Convolutional Networks | 1 + ...the Constrained Most Probable Explanation Task | 1 + ...g Statistics and Mutual Knowledge Distillation | 1 + ...: A Simple yet Effective Exploration Criterion | 1 + .../Numerical Composition of Differential Privacy | 1 + ...rical influence of ReLU'(0) on backpropagation | 1 + ...on for Natural Language Understanding via ADMM | 1 + ...eep Generative Models for Lossless Compression | 1 + ...GCNN: 3D Object Detection using Dynamic Graphs | 1 + ...ressing Causal Confusion in Imitation Learning | 1 + ...with Generative Spatial-Temporal Factorization | 1 + ...ive Learning for Debiased Scene Representation | 1 + ...Observation-Free Attacks on Stochastic Bandits | 1 + ...ierarchical Implicit Functions for 3D Modeling | 1 + ...f-Policy Risk Assessment in Contextual Bandits | 1 + ... Learning via Pessimistic Dual Value Iteration | 1 + ...enges and Effective Data Collection Strategies | 1 + .../Offline Model-based Adaptable Policy Learning | 1 + .../Offline RL Without Off-Policy Evaluation | 1 + ... Learning as One Big Sequence Modeling Problem | 1 + ... Learning with Reverse Model-based Imagination | 1 + ...untable Multi-Agent Sequential Decision Making | 1 + ...n Calibration and Out-of-Domain Generalization | 1 + ... Interactions in Two-Stage Recommender Systems | 1 + ...astive Representations of Stochastic Processes | 1 + .../On Density Estimation with Diffusion Models | 0 ...heduling of Model-based Reinforcement Learning | 1 + ...imization with Dependent and Heavy-Tailed Data | 1 + ..., Prototypical Networks, and Few-Shot Learning | 1 + ... for Heterogeneous Treatment Effect Estimation | 1 + ...d Corruptions in Natural Corruption Robustness | 1 + ...r Solving Placement and Routing in Chip Design | 1 + ...n Large-Cohort Training for Federated Learning | 1 + ...ns for Transfer Learning with Multiple Sources | 1 + ...of SGD and Input-Smoothness of Neural Networks | 1 + .../On Locality of Local Explanation Models | 1 + ...gin-Based Cluster Recovery with Oracle Queries | 1 + ...zation in Probabilistic Deep Generative Models | 1 + ...led Object Detection and Instance Segmentation | 1 + .../On Optimal Interpolation in Linear Regression | 1 + ...rsarial Corruption in Online Decision Problems | 1 + ...ls: Group Representation and Isotropic Scaling | 1 + ...nforcement Learning from Expert Demonstrations | 1 + ...lly Frozen Weights in Sequential Task Learning | 1 + ...Depth in Training Graph Convolutional Networks | 1 + ...e Matrices with the Bures-Wasserstein Geometry | 1 + ...tational Complexity and Barycenter Computation | 1 + ...A Second Look at Transferable Targeted Attacks | 1 + ...plication to Ranking from Pairwise Comparisons | 1 + data/2021/neurips/On Training Implicit Models | 1 + data/2021/neurips/On UMAP's True Loss Function | 1 + ...rning sparse vectors from mixture of responses | 1 + ...n sensitivity of meta-learning to support data | 1 + ... Algorithmic Stability of Adversarial Training | 1 + ...iance-Cost Tradeoff of Stochastic Optimization | 1 + ...sed Model-Agnostic Meta-Reinforcement Learning | 1 + ...ncy of Variance-Reduced Policy Gradient Method | 1 + ...or-Guided Zeroth-Order Optimization Algorithms | 1 + ...ep Decay Step-Size for Stochastic Optimization | 1 + ...c Hardness of Learning Single Periodic Neurons | 1 + ...ween Neural Network and Support Vector Machine | 1 + .../On the Estimation Bias in Double Q-Learning | 1 + ... Existence of The Adversarial Bayes Classifier | 1 + .../On the Expected Complexity of Maxout Networks | 1 + .../neurips/On the Expressivity of Markov Reward | 1 + .../On the Frequency Bias of Generative Models | 1 + ... the Generative Utility of Cyclic Conditionals | 1 + ...or Detecting Distributional Shifts in the Wild | 1 + ...eneralization of Probabilistic Image Modelling | 1 + ...ning with Batch Normalization and Weight Decay | 1 + ...erentiable Learning versus PAC and SQ Learning | 1 + .../On the Power of Edge Independent Graph Models | 1 + ...le Generalization of Recurrent Neural Networks | 1 + ...e Representation Power of Set Pooling Networks | 1 + ...of Solutions to Elliptic PDEs in Barron Spaces | 1 + ...ation in Double Descent: A Least Squares Study | 1 + ...mplexity of Learning under Geometric Stability | 0 ... of Privately Learning Axis-Aligned Rectangles | 1 + ...onvergence Properties of Random Search Methods | 1 + ...the Stochastic Stability of Deep Markov Models | 1 + ...mality of Thompson Sampling in High Dimensions | 1 + ...cement Learning with Once-per-Episode Feedback | 1 + ...f Graph Neural Networks on Large Random Graphs | 1 + ... with Stochastic Differential Equations (SDEs) | 1 + ...te Gradients in Variational Autoencoder Models | 1 + ...d Function Approximation in Imitation Learning | 1 + ...ce of the Fisher Information for Deep Learning | 1 + ...e and loss function in classification problems | 1 + ...ured Attention Graphs for Image Classification | 1 + ...gle Cosine Similarity based Learning Objective | 1 + ...operative Bandits with Imperfect Communication | 1 + ...ges with Cross-lingual Dense Passage Retrieval | 1 + ... Active Learning with Surrogate Loss Functions | 1 + .../Online Adaptation to Label Distribution Shift | 1 + ...trol of Unknown Time-Varying Dynamical Systems | 1 + ...imization with Continuous Switching Constraint | 1 + .../Online Facility Location with Multiple Advice | 1 + .../Online Knapsack with Frequency Predictions | 1 + ...ral Computations From Sparse Temporal Feedback | 1 + ...f Complex Dynamical Systems from Sensory Input | 1 + .../Online Learning in Periodic Zero-Sum Games | 1 + ... Equilibrium with Application to Fair Division | 1 + ...on-Asymptotic Performances of Greedy Algorithm | 1 + ...ne Multi-Armed Bandits with Adaptive Inference | 1 + ... Reinforcement Learning with Model Uncertainty | 1 + ...Selective Classification with Limited Feedback | 1 + ...f the Number of Errors in Thresholding Bandits | 1 + ...e Variational Filtering and Parameter Learning | 1 + ...ment Learning by Planning with a Learned Model | 1 + ...e control for anomaly detection in time series | 1 + ...ear function approximation and bandit feedback | 1 + ... Neural Network Training And Pruning Framework | 1 + data/2021/neurips/Open Rule Induction | 1 + ...mprove Robustness Against Inherent Label Noise | 1 + ...rning with Open-set Consistency Regularization | 1 + ...s for Stochastic Contextual Preference Bandits | 1 + ... Identification Methods for Tail-Risk Measures | 1 + ...Algorithms for Non-concave Bandit Optimization | 1 + ...der Simple Regret for Gaussian Process Bandits | 1 + .../neurips/Optimal Policies Tend To Seek Power | 1 + ...ity Estimation under Communication Constraints | 1 + ...mal Rates for Random Order Online Optimization | 1 + .../Optimal Sketching for Trace Estimation | 1 + .../Optimal Underdamped Langevin MCMC Method | 1 + ...eneous, Reward-Free and Task-Agnostic Settings | 1 + ...of Markov chains with and without spectral gap | 1 + ... Federated Learning: A Game-theoretic Approach | 1 + ...e for stochasticblock model with missing links | 1 + ...tigrid Coarsening Using Reinforcement Learning | 1 + ...nditional Value-At-Risk of Black-Box Functions | 1 + ...ralization Bound via Anisotropic Noise of SGLD | 0 ...wledge for Continual Learning via Metalearning | 1 + ...Complexity in Nonsmooth Nonconvex Optimization | 1 + ...zation in Factored MDPs with Unknown Structure | 1 + ...stribution Generalization in Kernel Regression | 1 + ...inforcement Learning via Variational Inference | 1 + ...ental Few-Shot Learning by Finding Flat Minima | 1 + ...ercoming the Convex Barrier for Simplex Inputs | 1 + ...ian regularization in semi-supervised learning | 1 + ...reveals image classification model pathologies | 1 + ...bustness to Covariate Shift in High Dimensions | 1 + ...ssage Passing in Rotationally Invariant Models | 1 + ...ks Motivated by Partial Differential Equations | 1 + ...ing generative models with recovery guarantees | 0 ...of Program Learning, Understanding, and Repair | 1 + ...ing via Penalizing Out-of-Distribution Samples | 1 + ...presentations for Effective Probability Models | 1 + ...Conceptual, Relational, and Physical Reasoning | 1 + ...D Scene Reconstruction From a Single RGB Image | 1 + ...l Ridge Regression by Feature Space Partitions | 1 + ...jectives with Expected Hypervolume Improvement | 1 + ...and Efficient Hierarchical k-Median Clustering | 1 + data/2021/neurips/Parallelizing Thompson Sampling | 1 + .../Parameter Inference with Bifurcation Diagrams | 1 + ...meter Prediction for Unseen Deep Architectures | 1 + ...Parameter-free HE-friendly Logistic Regression | 1 + ...e Transfer for Personalized Federated Learning | 1 + ...ds for Approximating PDEs with Neural Networks | 1 + ...ed Quantum Policies for Reinforcement Learning | 1 + ...nted Algorithms for Online Conversion Problems | 1 + ...osing the gap between human and machine vision | 1 + ... Federated Learning via Partial Initialization | 1 + ...essage Passing Generative Adversarial Networks | 1 + ... Network with Global Convergence Rate Analysis | 1 + ...tion and Code: learning how to compress graphs | 1 + ...r Optimization of Trained ReLU Neural Networks | 1 + ...ral networks predicts human visual selectivity | 1 + ... Signal Mid-level Patches in Referential Games | 1 + data/2021/neurips/Pay Attention to MLPs | 1 + ...ultilingual and Multi-Domain Sequence Modeling | 1 + ... is Not All You Need for Semantic Segmentation | 1 + ...terogeneous Agents via Personalized Simulators | 1 + ...What Data Modalities Does Your Model Perceive? | 1 + ...iodic Activation Functions Induce Stationarity | 1 + ...ncoder for Graph-Level Representation Learning | 1 + .../Permuton-induced Chinese Restaurant Process | 1 + ...zed Federated Learning With Gaussian Processes | 1 + ...g and learning in discrete energy-based models | 1 + ...urbation Theory for the Information Bottleneck | 1 + ...dictive Control in Linear Time Varying Systems | 1 + ...ly Efficient Offline Mean-Field Multi-Agent RL | 1 + ...oo: Gym for Multi-Agent Reinforcement Learning | 1 + ...rential Privacy with Direct Feedback Alignment | 1 + ...with Deep Learning for Scalable Flood Modeling | 1 + ...r Robust and Interpretable Generative Modeling | 1 + ...le Learning To Rank via Differentiable Sorting | 1 + .../Pipeline Combinators for Gradual AutoML | 1 + ...ltidimensional Planner for DNN Parallelization | 1 + ...nments with Combinatorially Hard Search Spaces | 1 + ...g Games as Classifying Markov Decision Process | 1 + ...irtual Trajectories for Reinforcement Learning | 1 + ...ion Estimation under Communication Constraints | 1 + ... Detection and Segmentation with Polar Pillars | 0 ...ient Offline and Online Reinforcement Learning | 1 + .../Policy Learning Using Weak Supervision | 1 + ...MDPs: Improved Exploration via Dilated Bonuses | 1 + .../Pooling by Sliced-Wasserstein Embedding | 1 + ...ble and High-Quality Generative Text-to-Speech | 1 + .../2021/neurips/Post-Contextual-Bandit Inference | 1 + ...t-Training Quantization for Vision Transformer | 1 + .../Post-Training Sparsity-Aware Quantization | 1 + .../Post-processing for Individual Fairness | 1 + ...llapse and Latent Variable Non-identifiability | 1 + .../Posterior Meta-Replay for Continual Learning | 1 + ... A sparsity inducing weight reparameterisation | 1 + ... Programming using Primal-Dual Hybrid Gradient | 1 + ...actical Near Neighbor Search via Group Testing | 1 + ...ealizable Setting: The Power of True Believers | 1 + ...pression for Human-in-the-Loop Decision-Making | 1 + ... predictive distribution of deep ReLU networks | 1 + ...r-Parameterized Nonconvex Matrix Factorization | 1 + ...neralization with Perturbation Response Curves | 1 + ... Memorability from Contextual Visual Semantics | 1 + ... Conformation via Dynamic Graph Score Matching | 1 + ... Know Helps: Provable Self-Supervised Learning | 1 + ...with brain-inspired predictive coding dynamics | 1 + ...eferences in Auction Design with Deep Learning | 1 + ...irectional compression in distributed settings | 1 + ...ions for Data-Efficient Reinforcement Learning | 1 + ... Dynamic Auctions for a Value-maximizing Buyer | 1 + ...e Non-smooth ERM and SCO in Subquadratic Steps | 1 + ...on-private Uniformity Testing for Ranking Data | 1 + .../Private learning implies quantum stability | 1 + ...ly Learning Mixtures of Axis-Aligned Gaussians | 1 + data/2021/neurips/Privately Learning Subspaces | 1 + .../Privately Publishable Per-instance Privacy | 1 + ...am-Guided Transformer for Program-Guided Tasks | 1 + ...ilistic Attention for Interactive Segmentation | 1 + ...tion Model for Reasoning over Knowledge Graphs | 1 + ...robabilistic Forecasting: A Level-Set Approach | 1 + ...r Instance Reweighting in Adversarial Training | 1 + ...position of Neural Population Spiking Activity | 1 + ...abilistic Transformer For Time Series Analysis | 1 + ...ths and the Structure of Predictions over Time | 1 + ...Attention for Vision-and-Language Pre-training | 1 + ... Society (PALMS) with Values-Targeted Datasets | 1 + ...i-Objective Stein Variational Gradient Descent | 1 + ...t Learning for Partially Observed Environments | 1 + ...e Transforms for Monocular 3D Object Detection | 1 + ...ure Interaction Search for Deep Sparse Network | 1 + data/2021/neurips/Projected GANs Converge Faster | 1 + data/2021/neurips/Proper Value Equivalence | 1 + ...rks for Few-Shot Molecular Property Prediction | 1 + ...articipatory Budgeting with Additive Utilities | 1 + ... for Multiple Object Tracking and Segmentation | 1 + ...tic Methods for Offline Reinforcement Learning | 1 + ...d Deep Learning with Spectral Contrastive Loss | 1 + ...ng: Shelve Optimism, Embrace Virtual Curvature | 1 + ...or Imitation with Contrastive Fourier Features | 1 + ...isoning Attacks Against Reinforcement Learning | 1 + ...nt Learning with Confounded Observational Data | 1 + ...ion Approximation under Adaptivity Constraints | 1 + ...bly Faster Algorithms for Bilevel Optimization | 1 + ...ation Benefit for Invariance in Kernel Methods | 1 + ...ask reinforcement learning with model transfer | 1 + ... efficient, succinct, and precise explanations | 1 + ...of Neural Networks Trained by Gradient Descent | 1 + ... Normalization while Removing Batch Dependence | 1 + ...d Neural Networks with Iterative Randomization | 1 + .../Pseudo-Spherical Contrastive Divergence | 1 + .../Pure Exploration in Kernel and Neural Bandits | 1 + ...ion with Synthetic Boundary Supporting Samples | 1 + ...n Artifacts for Achieving Adversarial Outcomes | 1 + ...lation with Applications to Federated Learning | 1 + ...oving Transferability in Domain Generalization | 1 + ...-Drop: Regularized Dropout for Neural Networks | 1 + ...Structured Compression of Deep Neural Networks | 1 + ...iple Interacting People under Weak Supervision | 1 + ... Efficient and Robust Semi-Supervised Learning | 1 + ...able Influence-based Active Learning on Graphs | 1 + ...tent MDPs: Regret Guarantees and a Lower Bound | 1 + ...d Reinforcement Learning is a Dataflow Problem | 1 + ... for Cooperative Reinforcement Learning Agents | 1 + ...mory Management for Class-Incremental Learning | 1 + ... Defense Against Query-Based Black-Box Attacks | 1 + ... After Many Epochs on Ill-Conditioned Problems | 1 + ...ecovery: Subgradient Method and Exact Recovery | 1 + data/2021/neurips/Ranking Policy Decisions | 1 + ...e-Optimal Subspace Estimation on Random Graphs | 1 + ...Plug-in Estimators via Barycentric Projections | 1 + ... MLP based Multi-Modal Information Unscrambler | 0 ...ieval and transductive few-shot classification | 1 + ...tribution Detection With Rectified Activations | 1 + .../neurips/ReLU Regression with Massart Noise | 1 + ...elf-Supervised Learning with Weak Augmentation | 1 + ...c evaluation of transductive few-shot learning | 1 + ...Auxiliary Classifier GANs with Stable Training | 1 + ...ounding Bandits for Modeling Satiation Effects | 1 + ...ognizing Vector Graphics without Rasterization | 1 + ...onstruction for Powerful Graph Representations | 1 + ...or for Generalization to Distributional Shifts | 1 + ...iors using the Restricted Eigenvalue Condition | 1 + .../Rectangular Flows for Manifold Learning | 1 + ...t Learning of Background for Few-Shot Learning | 1 + ...ural Networks with Recurrent Layer Aggregation | 1 + ...er Chains for Exact Multi-Label Classification | 1 + ...ular Welfare and Matroid Blocking Semi-Bandits | 1 + ...xt-Free Grammars and Dynamic Bayesian Networks | 1 + ...resence of Latent Variables and Selection Bias | 1 + ...Insights from Multi-particle Dynamical Systems | 1 + ...ed Motion Planning Using Graph Neural Networks | 1 + ...ck for Weakly Supervised Semantic Segmentation | 1 + ...ft by Mirror Samples in Cross Domain Alignment | 1 + ...e-step Approach to Multi-task Visual Grounding | 1 + ...ng Bounds for Kernel and Approximate $k$-Means | 1 + ...anguage Models with Compositional Explanations | 1 + ...hot Action Recognition for Multi-label Actions | 1 + data/2021/neurips/Regime Switching Bandits | 1 + ...Gaussian-Process Optimization in Large Domains | 1 + ...ce Replay in Off-Policy Reinforcement Learning | 1 + ...Regularization in ResNet with Stochastic Depth | 1 + ...Dense CRFs: Generalizing Mean Field and Beyond | 1 + ...egularized Softmax Deep Multi-Agent Q-Learning | 1 + ...gulating algorithmic filtering on social media | 1 + ...on Function Learning for Bayesian Optimization | 1 + ...g Enhanced Explainer for Graph Neural Networks | 1 + ...ease Progression Model for Alzheimer's Disease | 1 + ...: Constant Regret and Representation Selection | 1 + ...forcement Learning in Newcomblike Environments | 1 + .../Reinforcement Learning in Reward-Mixing MDPs | 1 + .../Reinforcement Learning with Latent Flow | 1 + ...iselessly Observable Markov Decision Processes | 1 + ...n of variational quantum circuit architectures | 1 + ...s Missing in Attention for Video Understanding | 1 + .../neurips/Relative Flatness and Generalization | 1 + ...nty Learning for Facial Expression Recognition | 1 + ...eomorphisms indicates performance in deep nets | 1 + ...ncy for Differentially Private Query Answering | 1 + data/2021/neurips/Relaxing Local Robustness | 1 + ...entralized Deep Learning on Heterogeneous Data | 1 + ...h Improved Exact Search and Weaker Assumptions | 1 + .../Reliable Decisions with Threshold Calibration | 1 + ...criminator in Reproducing Kernel Hilbert Space | 1 + ...ations: Modeling Uncertainty in Explainability | 1 + ...rning for Health Using Dataset Shift Detection | 1 + ...t to Forget: Algorithms for Machine Unlearning | 1 + ...y from Functional Data in Systems Neuroscience | 1 + ...bsampled Shuffle Model In Distributed Learning | 1 + ...sed Policy Search via Recursive Classification | 1 + .../Replay-Guided Adversarial Environment Design | 1 + ...of Linear Neural Networks: Analysis and Design | 1 + ...on Learning Beyond Linear Prediction Functions | 1 + ...n Learning for Event-based Visuomotor Policies | 1 + .../Representation Learning on Spatial Networks | 1 + ...on of Deep Neural Networks and Ensemble Models | 1 + ... Space Accurately using Multi-Component Floats | 1 + ...or Graph Neural Networks with Global Attention | 1 + .../neurips/Repulsive Deep Ensembles are Bayesian | 1 + ...Models with Improved Representation Guarantees | 1 + ...n Efficient Transformer for Visual Recognition | 1 + ...thway Priors for Soft Equivariance Constraints | 1 + ...axation for Multi-view Representation Learning | 1 + ...: Debiasing graph embedding with random graphs | 1 + ...l Networks: Do Not Be Afraid of Overconfidence | 1 + ...ing Graph Transformers with Spectral Attention | 1 + ...Rethinking Neural Operations for Diverse Tasks | 1 + ...verage for Efficient Video Object Segmentation | 1 + ...-Label Ranking: Consistency and Generalization | 1 + ...sing geometrically structured latent manifolds | 1 + ...ent sparsification as total error minimization | 1 + ...ning Criteria for Convolutional Neural Network | 1 + ...rpretation of Accelerated Optimization Methods | 1 + ... Adult: New Datasets for Fair Machine Learning | 1 + ...ive Projections over Submodular Base Polytopes | 1 + ... and Protecting Labels in Distributed Training | 1 + ...imization via machine learning with noisy data | 1 + ... optimizers reveals known and novel mechanisms | 1 + ...th Jacobian switching linear dynamical systems | 1 + ...plement Equivariant Networks for DNA Sequences | 1 + ...arning through the Lens of Multi-Task Learning | 1 + ...bject Detection From an Egocentric Perspective | 1 + ...nsupervised Learning of Visual Representations | 1 + ...visiting Deep Learning Models for Tabular Data | 1 + ...r-discriminator Cooperative Compression Scheme | 1 + ...ormation Bottleneck for Adversarial Robustness | 1 + ...el Stitching to Compare Neural Representations | 1 + ...Nets: Improved Training and Scaling Strategies | 1 + .../neurips/Revisiting Smoothed Online Learning | 1 + ...ting the Calibration of Modern Neural Networks | 1 + ...Self-Supervised Visual Representation Learning | 1 + .../2021/neurips/Reward is enough for convex MDPs | 1 + ...nt Learning with Linear Function Approximation | 1 + ...ation for a Smart Predict-then-Optimize Method | 1 + ...Margin Classification on Sub-Gaussian Mixtures | 1 + ... Guarantees for Supervised and Policy Learning | 1 + .../Risk Monotonicity in Statistical Learning | 1 + ...k-Averse Bayes-Adaptive Reinforcement Learning | 1 + ...einforcement Learning using Successor Features | 1 + ...k-averse Heteroscedastic Bayesian Optimization | 1 + ...daptation for Offline Model-based Optimization | 1 + .../Robust Allocations with Diversity Constraints | 1 + ...obust Auction Design in the Auto-bidding World | 1 + ...ressed Sensing MRI with Deep Generative Priors | 1 + ...ing Negative Samples with Diminished Semantics | 1 + ...rfactual Explanations on Graph Neural Networks | 1 + ...einforcement Learning through Adversarial Loss | 1 + ...n Shift via Minimum Discriminating Information | 1 + ...plicit Networks via Non-Euclidean Contractions | 1 + ...nt Learning under Transition Dynamics Mismatch | 1 + .../neurips/Robust Learning of Optimal Auctions | 1 + .../neurips/Robust Online Correlation Clustering | 0 ... Multilingual Translation with Imbalanced Data | 1 + ...rowded Scenes with Direct Pose-Level Inference | 1 + data/2021/neurips/Robust Predictable Control | 1 + ...ed: Acceleration and Improved Estimation Rates | 1 + ...ing via Language Guided Neural Module Networks | 1 + ...mposable Average Precision for Image Retrieval | 1 + ...-and-Bounded Learning (With Outliers) Problems | 1 + ...ust and differentially private mean estimation | 1 + ...of Learning Latent Trees with Vector Variables | 1 + .../Robustness between the worst and average case | 1 + .../Robustness of Graph Neural Networks at Scale | 1 + ...stness via Uncertainty-aware Cycle Consistency | 1 + ...ity by Projection in Knowledge Graph Embedding | 1 + ...inate Frames For Interacting Dynamical Systems | 1 + ...ow-clustering of a Point Process-valued Matrix | 1 + ...Dual Graph Aggregation Network for Text-to-SQL | 1 + ...e Progressive Encoding for Neural Optimization | 1 + ...l Networks via Stochastic Bilevel Optimization | 1 + ...lecular wavefunctions and electronic densities | 1 + ... Learning using Exploration and 3D Consistency | 1 + ...Regularization, Batch-size and Multiple-epochs | 1 + ...bolic Interactive Language Grounding Benchmark | 0 ...s Based Active Learning In Realistic Scenarios | 1 + ...entations via Unsupervised Video Decomposition | 1 + ...s Structure Learning for Graph Neural Networks | 1 + ...erence in High-Dimensional Logistic Regression | 1 + ... Solving Noisy Inverse Problems Stochastically | 1 + ...Transformer for Vision-and-Language Navigation | 1 + ...oftmax-free Transformer with Linear Complexity | 1 + .../SOLQ: Segmenting Objects by Learning Queries | 1 + .../SOPE: Spectrum of Off-Policy Estimators | 1 + ...-scale Approximate Nearest Neighborhood Search | 0 ... by Decoupling Multi-Hop and Logical Reasoning | 1 + ... Learning for Domain Adaptive Object Detection | 1 + .../SSMF: Shifting Seasonal Matrix Factorization | 1 + ... for Exemplar-based Class-Incremental Learning | 1 + ...munication Complexities for Federated Learning | 1 + ...esence of Limited In-Distribution Labeled Data | 1 + ... Recursive Momentum for Nonconvex Optimization | 0 ... and Universal Framework of Adaptive Gradients | 1 + ...: Domain Generalization by Seeking Flat Minima | 1 + ...cal Generalized Linear Function Approximations | 1 + .../Safe Pontryagin Differentiable Programming | 1 + ...orcement Learning by Imagining the Near Future | 1 + ...ent Learning with Natural Language Constraints | 1 + ...arning against Both Stragglers and Adversaries | 1 + ...e Graph Explanations for Commonsense Reasoning | 1 + ...for Active Ranking from Multi-wise Comparisons | 1 + ...earch Configuration: Cutting Planes and Beyond | 15 +++++++++++++++ .../Sample Selection for Fair and Robust Training | 1 + ...of Stackelberg Equilibria in General-Sum Games | 1 + ...nearly Realizable MDPs with Limited Revisiting | 1 + ...rly-Parameterized MDPs with a Generative Model | 1 + ...es Your Winning Ticket Really Win the Jackpot? | 1 + ...evance determination and discrete noise models | 1 + ...del Selection for Accessible Transfer Learning | 1 + ...ching of the Fokker-Planck-Kolmogorov Equation | 1 + ...parsely-changing Gaussian Markov Random Fields | 1 + ...ntervention Target Estimation in Linear Models | 1 + ...rver: A Data Recommender for Transfer Learning | 1 + ...lanning via Reinforcement Learning Fine-Tuning | 1 + ...Inference for Instrumental Variable Regression | 1 + ...tion Learning for Interpretable Classification | 1 + ... Sampling using Sparse Gaussian Process Models | 1 + ...Flexible Classifiers with Fairness Constraints | 1 + ...ne learning, structured like classical physics | 1 + ...rsarial Patches with Sparse Superficial Layers | 1 + ...istillation to Many Classes with Proxy Targets | 1 + ...vative Information Using Variational Inference | 1 + ...gent Kernels via Sketching and Random Features | 1 + ...t Neural Network Compression by ReLU Stability | 1 + .../Scaling Vision with Sparse Mixture of Experts | 1 + ...Markov Chains Helps Resolve Underspecification | 1 + ...Databases to Scalable Differentiable Reasoning | 1 + ...rbrain: Unifying Sparse and Low-rank Attention | 1 + .../Scheduling jobs with stochastic holding costs | 1 + ...core-based Generative Modeling in Latent Space | 1 + ...ral Networks for Large-Scale Optimal Transport | 1 + ...ing Parameterized AP Loss for Object Detection | 1 + ...r Efficient Transformers for Language Modeling | 1 + ...arching the Search Space of Vision Transformer | 1 + .../neurips/Second-Order Neural ODE Optimizer | 1 + ... Consistency Learning for Scene Classification | 1 + ...gn for Semantic Segmentation with Transformers | 1 + ...ve Sampling for Online Best-arm Identification | 1 + ...Point Processes with Nonparametric Time Decays | 1 + ...Individual Input-Output Pairs in Deep Learning | 1 + .../neurips/Self-Consistent Models and Values | 1 + ...ted Samples in Generative Adversarial Networks | 1 + ...ed Recurrent Units with Dynamic Soft Recursion | 1 + ...with Transformation Equivariant Interpretation | 1 + ...ed Medical Image Segmentation with Meta-labels | 1 + .../Self-Supervised Bug Detection and Repair | 1 + .../Self-Supervised GANs with Label Augmentation | 1 + ...g Disentangled Group Representation as Feature | 1 + ...ased Optical Flow with Spiking Neural Networks | 1 + ...entations Provably Isolates Content from Style | 1 + ...d Learning with Kernel Dependence Maximization | 1 + ...i-Object Tracking with Cross-input Consistency | 1 + ...rk Weights for Model Characteristic Prediction | 1 + ...egmentation via Adaptive Equalization Learning | 1 + ...brium Models and Applications to Certification | 1 + ...ernel and Feature-Learning Probability Metrics | 1 + ...-Sequence Learning with Latent Neural Grammars | 1 + ...orithms for Testing Closeness of Distributions | 1 + ...Imitation Learning with Unobserved Confounders | 1 + .../neurips/Set Prediction in the Latent Space | 1 + ...g the Variance of Multi-Agent Policy Gradients | 1 + ...ape As Points: A Differentiable Poisson Solver | 1 + ...Shape Registration in the Time of Transformers | 1 + ...red 3D Shape and Motion of Fast Moving Objects | 1 + ...ization Approach to Deterministic Autoencoders | 1 + ...Transformer using Factorized Reshaped Matrices | 1 + ... activity-context priors from egocentric video | 1 + ...e limits of the Shapley value for explanations | 1 + ...ponent Analysis for Multi-Subject Neuroimaging | 1 + ... Impossibility Results for Hyper-graph Testing | 1 + ...t Invariance Can Reduce Adversarial Robustness | 1 + ...e Limitations of Localized Graph Training data | 1 + ... for Spatio-Temporal Representational Learning | 1 + ...ethods for stochastic variational inequalities | 1 + data/2021/neurips/Sim and Real: Better Together | 1 + ...Training using Gradient Similarity Measurement | 1 + ...and Matching of Neural Network Representations | 1 + ...dient Descent Algorithms for Pairwise Learning | 1 + ...olfe and generalized self-concordant functions | 1 + ...m Likelihood for Out-of-Distribution Detection | 1 + ...SketchGen: Generating Constrained CAD Sketches | 1 + ...ased Piano Transcription With Neural Semi-CRFs | 1 + .../Slice Sampling Reparameterization Gradients | 1 + ...: A Scalable Measure of Statistical Dependence | 1 + ...ilarity Computation via Knowledge Distillation | 1 + ...erparameterized low-rank matrix reconstruction | 1 + ... Bilevel Programming for Sparse Regularization | 1 + data/2021/neurips/Smooth Normalizing Flows | 1 + ... Smoothed Classifiers for Certified Robustness | 1 + ...ession Techniques for Distributed Optimization | 1 + ...onal continuous control via parameter freezing | 1 + ...oft Calibration Objectives for Neural Networks | 1 + ... Games with Tree Search and Imitation Learning | 0 ...h Hidden Structure via Gradient Descent Ascent | 1 + ...via $k$-Sparse Discrete Wasserstein Barycenter | 1 + ...ce-time Mixing Attention for Video Transformer | 1 + ...ework Immune to Local Traps and Miscalibration | 1 + .../Sparse Flows: Pruning Continuous-depth Models | 1 + ...ith Application to Permutation Synchronisation | 1 + data/2021/neurips/Sparse Spiking Gradient Descent | 1 + ...ation and Tracking of Object Poses in 3D Space | 1 + ...ting Pruning Plasticity with Neuroregeneration | 1 + ...ntation in Deep Learning with Inducing Weights | 1 + .../Sparse is Enough in Scaling Transformers | 1 + ...n and Planning in Partially Observable Domains | 1 + ...othing Mechanism for Student-Teacher Framework | 1 + ...ellite Imagery via Conditional Pixel Synthesis | 1 + ...Spatio-Temporal Variational Gaussian Processes | 1 + ...omposition in 3D Convolutional Neural Networks | 1 + ...for dynamic networks with stability guarantees | 1 + ...tion for Accurate Blind Image Super-Resolution | 1 + ...s Fully Recurrent Convolutional Neural Network | 1 + ...ch-T: Transducer for Text to Speech and Beyond | 1 + ...ance Estimation for Neural Architecture Search | 1 + ...ized Neural Network using SGD and Weight Decay | 1 + ...of Topological Changes via Geometric Alignment | 1 + ...suit: Tuning-Free Noisy Robust Matrix Recovery | 1 + ...ral Networks without the Neural Tangent Kernel | 1 + ...mal Risk Bounds with Convergence Rate $O(1 n)$ | 1 + ...vel Programming in Hyperparameter Optimization | 1 + ...nd Vision Transformers under Data Augmentation | 1 + ... Dynamical Systems via Policy Gradient Methods | 1 + ...ints for Defending Against Adversarial Attacks | 1 + ...ed Attention with Relative Positional Encoding | 1 + ...teful ODE-Nets using Basis Function Expansions | 1 + data/2021/neurips/Stateful Strategic Regression | 1 + ...with M-Estimators on Adaptively Collected Data | 1 + ...er Bounds for List-Decodable Linear Regression | 1 + ...tein Autoencoder with Latent Space Consistency | 1 + ...l Models in the Presence of Latent Confounders | 1 + ... Efficient Linear Meta-representation Learning | 1 + ...n Mixing for Nonconvex Stochastic Optimization | 1 + .../Stochastic Bias-Reduced Gradient Methods | 1 + ...vergence Analysis under Expected Co-coercivity | 1 + ...stic Multi-Armed Bandits with Control Variates | 1 + ...ession: the Forward Algorithm to Replace Ridge | 1 + ...cision-Recall Curves with Provable Convergence | 1 + ...Parameter-Free and Towards Horizon-Free Regret | 1 + ...roblems using the Prior Implicit in a Denoiser | 1 + ...Stochastic bandits with groups of similar arms | 1 + ...cay schedules, and high probability guarantees | 0 ...r General Stochastic Automatic Differentiation | 1 + ...liss: Iterative Voting Improves Social Welfare | 1 + ... Identification with Reverse Experience Replay | 2 ++ .../neurips/Stronger NAS with Weaker Predictors | 1 + ...n Neural Networks using Reinforcement Learning | 1 + ... Bregman information, and exponential families | 1 + ...ructure-Aware Random Fourier Kernel for Graphs | 1 + ...sing Diffusion Models in Discrete State-Spaces | 1 + ...ational Inference for Bayesian Neural Networks | 1 + ...ing Latent Alignments in Sequence Transduction | 1 + ...eraging Dropout in RNNs for Efficient Training | 1 + ...logue Generation with Multi-Pass Dual Learning | 1 + ...Sub-Linear Memory: How to Make Performers SLiM | 1 + ...ta for Self-Supervised Representation Learning | 1 + .../Subgame solving without common knowledge | 1 + ...ampling for Off-Policy Evaluation and Learning | 1 + .../Subgoal Search For Complex Reasoning Tasks | 1 + ...ated Learning with Missing Neighbor Generation | 1 + ...lization and Fairness of Graph Neural Networks | 1 + data/2021/neurips/Submodular + Concave | 1 + ...erparameterization for Shallow Neural Networks | 1 + ...orizon Goal-Conditioned Reinforcement Learning | 1 + ...ergy-based Contrastive Representation Transfer | 1 + ...sing the Transfer of Reasoning Patterns in VQA | 1 + ... Signals from a Mixture of Linear Measurements | 1 + ...n coincide with very high-dimensional features | 1 + .../Surrogate Regret Bounds for Polyhedral Losses | 1 + ...eous Treatment Effects from Time-to-Event Data | 1 + ...arnt Hamiltonian Dynamics Inferred from Vision | 1 + ... Learning Enhanced Genetic Programming Seeding | 0 ...act Gradient of Neural ODE with Minimal Memory | 1 + ...t Effect Estimation with Longitudinal Outcomes | 1 + ...to Experimental Design with Synthetic Controls | 1 + ...stematic Generalization with Edge Transformers | 1 + ...r Structured Sparsity and Smoothness on Graphs | 1 + ...y Abstract Actor-Critic for Continuous Control | 1 + ... NAS Predictor with a Self-evolution Framework | 1 + ...pproach towards Few-shot Hypothesis Adaptation | 1 + ...moting Gradient Diversity and Model Smoothness | 1 + ...-Supervised Test-Time Training Fail or Thrive? | 1 + ... from Scratch with Deep Reinforcement Learning | 1 + ... and Pessimism for Deep Reinforcement Learning | 1 + ...ing unsupervised objectives at prediction time | 1 + ...Cooperative Multi-Agent Reinforcement Learning | 1 + .../neurips/Targeted Neural Dynamical Modeling | 1 + ... Network Search with Meta-Contrastive Learning | 1 + ...re Deactivation Using Out-of-Distribution Data | 1 + ...al structure in neural network loss landscapes | 1 + ...Reinforcement Learning via Advice Distillation | 1 + ...ng an Active Learner with Contrastive Examples | 1 + ...the Learning-with-Equivalence-Queries Paradigm | 1 + .../Techniques for Symbol Grounding with SATNet | 1 + ...ariance Pooling Networks for Video Recognition | 1 + .../neurips/Temporally Abstract Partial Models | 1 + ...ensor Normal Training for Deep Learning Models | 1 + ...r correlations by nonlinear Hebbian plasticity | 1 + ...Execution of Imperative Deep Learning Programs | 1 + ...odule for Model-Agnostic Domain Generalization | 1 + ...n with a Transformer for Human Pose Estimation | 1 + data/2021/neurips/Test-time Collective Prediction | 1 + ...labeled Test Instances for Deep Learning Tasks | 1 + data/2021/neurips/Testing Probabilistic Circuits | 1 + ...imator and a Paradox Concerning Logging Policy | 1 + ...ularization from SGD in Least Squares Problems | 1 + ...n: Expressiveness, Learnability, and Inference | 1 + ...etwork Learning: Revisiting the Superstructure | 1 + .../neurips/The Complexity of Sparse Tensor PCA | 1 + ...assive Learning in Deep Reinforcement Learning | 1 + ...on the Generalization of Quadratic Classifiers | 1 + .../neurips/The Elastic Lottery Ticket Hypothesis | 1 + ...s: Learning Zero-shot Segmentation from Videos | 1 + ...Duality of Adaptive Dropout and Regularization | 1 + ... Combinatorial Semi-bandits with Greedy Oracle | 1 + .../The Image Local Autoregressive Transformer | 1 + ...f Minima Stability: A View from Function Space | 1 + .../neurips/The Inductive Bias of Quantum Kernels | 1 + ...orithm is Universal on Strongly Convex Domains | 0 ... Networks: A Deep Gaussian Process Perspective | 1 + .../The Limits of Optimal Pricing in the Dark | 1 + .../neurips/The Many Faces of Adversarial Risk | 1 + ...ch Methods for Feature Importance Explanations | 1 + ...model selection for general Contextual Bandits | 1 + ... Few-Shot Classification and How to Infer Them | 1 + .../The Semi-Random Satisfaction of Voting Axioms | 1 + ...ant Neural Networks for Reinforcement Learning | 1 + ... for Differentially Private Federated Learning | 1 + ...ein Distance: Conic Formulation and Relaxation | 1 + ...Explainable AI in Ad Hoc Human-Machine Teaming | 1 + ...lue of Information When Deciding What to Learn | 1 + ...oice in distance-regularized domain adaptation | 1 + ...ersarial episodic MDPs with unknown transition | 1 + ...bandits and uncertainty to optimism and beyond | 1 + ...y embedding constructed from the $k$-Laplacian | 1 + ...s correlation with automatic evaluation scores | 1 + ...hways with self-supervised predictive learning | 1 + ...finite-depth-and-width limit at initialization | 1 + ...hierarchical structure can guide deep learning | 1 + ...for Reversibility-Aware Reinforcement Learning | 1 + ...mall: Do Language Models Distil Occam's Razor? | 1 + ...chastic Gradients, and Adaptive Learning Rates | 1 + ...d motion correction for Neuropixels recordings | 1 + ...r Stochastic Approximation with Fixed Stepsize | 1 + ...lization Error Bounds via Wasserstein Distance | 1 + ... Action Repetition for Policy Gradient Methods | 1 + ...ization Bounds for SGLD in Non-convex Settings | 1 + ...ime-series Generation by Contrastive Imitation | 1 + ...is a Question of Cooperation for Language GANs | 1 + ...ce-driven monocular 3D category reconstruction | 1 + ...d Alignment for Unsupervised Domain Adaptation | 1 + ...r: Adaptive Space-Time Tokenization for Videos | 1 + ...ocument Graph-based Neural Network Perspective | 1 + ...opicNet: Semantic Graph-Guided Topic Discovery | 1 + .../Topographic VAEs learn Equivariant Capsules | 1 + ...ological Attention for Time Series Forecasting | 1 + ...ological Detection of Trojaned Neural Networks | 1 + .../Topological Relational Learning on Graphs | 1 + ...arning for Semi-Supervised Node Classification | 1 + ...ll-Worlds Online Learning with Feedback Graphs | 0 ...bly Robust Models against Adversarial Examples | 1 + ... Biologically Plausible Convolutional Networks | 1 + ...iled Visual Recognition from Prior Perspective | 1 + ...Context-Agnostic Learning Using Synthetic Data | 1 + ...forcement Learning with Spectral Normalization | 1 + ...s Efficient and Effective Adversarial Training | 1 + ...ards Enabling Meta-Learning from Target Models | 1 + ...imization with Non-convex Followers and Beyond | 1 + ...y Selection for Offline Reinforcement Learning | 1 + ... Offline Reinforcement Learning with Pessimism | 1 + ...er Bounds on the Depth of ReLU Neural Networks | 1 + ...ained Explainability for Graph Neural Networks | 1 + ...polation: An Inductive Graph Learning Approach | 1 + ...g Self-Driving Perception Models in Simulation | 1 + .../Towards Robust Bisimulation Metric Learning | 1 + ...wards Robust and Reliable Algorithmic Recourse | 1 + ...se Retrieval with Sparse and Generative Priors | 1 + ...mple-efficient Overparameterized Meta-learning | 1 + ...Try-On via Patch-Routed Spatially-Adaptive GAN | 1 + ...eneralization Bounds for Structured Prediction | 1 + .../neurips/Towards Stable and Robust AdderNets | 1 + ...tion Lower Bounds for Distributed Optimisation | 1 + ...ulti-Agent Q-Learning with Value Factorization | 1 + ...okahead Generalizes Better Than SGD and Beyond | 1 + ...sity for Open-ended Learning in Zero-sum Games | 1 + ...ramework of Out-of-Distribution Generalization | 1 + ...ew of Adversarial Perturbations and Robustness | 0 ...rmation-Theoretic Framework for Generalization | 1 + ...hical memory for reinforcement learning agents | 1 + ...taining from prediction with OOD test examples | 1 + ...by multi-task learning on monkey visual cortex | 1 + ...standing retrosynthesis by energy-based models | 1 + .../Tracking People with 3D Representations | 1 + ... Without Re-recognition in Humans and Machines | 1 + ...arned Manifolds with Conformal Embedding Flows | 1 + ...table Regularization of Probabilistic Circuits | 1 + ...Networks with Efficient Local Lipschitz Bounds | 1 + ...licit Differentiation on the Equilibrium State | 1 + .../Training Neural Networks is ER-complete | 1 + ...aining Neural Networks with Fixed Sparse Masks | 1 + ...erized Models with Non-decomposable Objectives | 1 + ...nt Interpolation Loss to Generalize Along Time | 1 + ...Can Make One Strong GAN, and That Can Scale Up | 1 + ... Learning for Whole Slide Image Classification | 1 + ...ers for Generalizable Person Re-identification | 1 + ...tworks with Ego-graph Information Maximization | 1 + data/2021/neurips/Transformer in Transformer | 1 + ...ar RGB Scene Reconstruction using Transformers | 1 + ...ts and Can be Extended to Graphs & Hypergraphs | 1 + ...trategy for Single Image Reflection Separation | 1 + ...n Tree: from Decision Trees to Decision Graphs | 1 + ...n-centric Audio-visual Representation Learning | 0 .../True Few-Shot Learning with Language Models | 1 + .../Truncated Marginal Neural Ratio Estimation | 1 + ... Mixture of Normal-inverse Gamma Distributions | 1 + ...Networks via Zero-Shot Hyperparameter Transfer | 0 ... the Fly for Efficient Population Based AutoRL | 1 + ...of Bounded-Precision Recurrent Neural Networks | 1 + ...ivalence between robustness and regularization | 1 + ...gn of Spatial Attention in Vision Transformers | 1 + ...earning Evaluation: In vs. Out of Distribution | 1 + data/2021/neurips/Two steps to risk sensitivity | 1 + ...ided fairness in rankings via Lorenz dominance | 1 + ...esian optimization with inequality constraints | 1 + ...iance Fields for Dynamic Scene View Synthesis" | 1 + ...ms for Multinomial Logistic Regression Bandits | 1 + ...Modal Controls for Conditional Image Synthesis | 1 + ...Stochastic Combinatorial Optimization Problems | 1 + data/2021/neurips/Ultrahyperbolic Neural Networks | 1 + ... Examples: Designing Objects for Robust Vision | 1 + ...rough Non-negative Penalized Linear Regression | 1 + ...gh Bias-Contrastive and Bias-Balanced Learning | 1 + ...ecisions Facilitate Better Preference Learning | 1 + ...libration for Ensemble-Based Debiasing Methods | 1 + .../Uncertainty Quantification and Deep Ensembles | 1 + ...forcement Learning with Diversified Q-Ensemble | 1 + ...-Driven Loss for Single Image Super-Resolution | 1 + ...Integration In Deep Speech Recognition Systems | 1 + .../Understanding Bandits with Graph Feedback | 1 + ...cess in Over-parametrized Tensor Decomposition | 1 + ... Learning Methods as Implicit Parameterization | 1 + ...nding How Encoder-Decoder Architectures Attend | 1 + ... Interpretability of Variational Auto-Encoders | 1 + ...ocking Dynamics of Cooperative Rationalization | 1 + ...native Self-supervised Representation Learning | 1 + ...al Multi-Label Learning via Mutual Information | 1 + ... Early Stopping for Learning with Noisy Labels | 1 + ...Effect of Stochasticity in Policy Optimization | 1 + ...it of Model Invariance from a Data Perspective | 1 + ...upervised Domain Adaptation via Data Poisoning | 1 + ... Under-Coverage Bias in Uncertainty Estimation | 1 + ... Taylor's Approximations for Image Restoration | 1 + ...etraining Framework for Document Understanding | 1 + ...ward a Unified Framework for Robust Clustering | 1 + ...sian Width, Norm Bounds and Benign Overfitting | 1 + .../Uniform Sampling over Episode Difficulty | 1 + ...nt Learning with Linear Function Approximation | 1 + ...inforcement Learning via Off-Policy Evaluation | 1 + ...Methods for Quasi-Self-Concordant Optimization | 1 + ...s on prediction dimension of convex surrogates | 0 ...alification Rate Disparities and Interventions | 1 + ...ique sparse decomposition of low rank matrices | 1 + ...ation Using Well-Conditioned Normalizing Flows | 1 + .../Universal Graph Convolutional Networks | 1 + data/2021/neurips/Universal Off-Policy Evaluation | 1 + ...rception Representations for Lossy Compression | 1 + .../neurips/Universal Semi-Supervised Learning | 1 + .../Unlabeled Principal Component Analysis | 1 + ...al Models via Contrast-Regularized Fine-Tuning | 1 + ...namics-Aware Rewards in Reinforcement Learning | 1 + ...eground Extraction via Deep Region Competition | 1 + ...ised Learning of Compositional Energy Concepts | 1 + ...resentation Learning with Capsule Autoencoders | 1 + ...by Discriminator-Constrained Optimal Transport | 1 + ...odels For 3D Partially Observable Environments | 1 + ...evel Representation Learning from Scene Images | 1 + ...Part Discovery from Contrastive Reconstruction | 1 + ...l Networks: I Believe I Can Distill On-the-Fly | 1 + data/2021/neurips/Unsupervised Speech Recognition | 1 + ...ially Private Learning via Correlated Sampling | 0 ... and Repeated Measures in Deep Neural Networks | 1 + ...on Factorization with Variable Agent Sub-Teams | 1 + ...rvised Learning from Raw Video, Audio and Text | 1 + ...raph Neural Networks using Vector Quantization | 1 + ...icket Hypothesis with Inertial Manifold Theory | 1 + ...Replication Robust Volume-based Data Valuation | 1 + ... Evaluation with Linear Function Approximation | 1 + ...Sparse-Reward Cooperative Multi-Agent Problems | 1 + .../Variational Bayesian Optimistic Sampling | 1 + ...sian Reinforcement Learning with Regret Bounds | 1 + .../Variational Continual Bayesian Meta-Learning | 1 + ...or Continuous-Time Switching Dynamical Systems | 1 + .../neurips/Variational Model Inversion Attacks | 1 + ...Multi-Task Learning with Gumbel-Softmax Priors | 1 + ... Space of Symmetric Positive Definite Matrices | 1 + ...ifolds via Gauge Independent Projected Kernels | 1 + ...ddings for Articulated 3D Shape Reconstruction | 1 + ...Advanced by Exploring Intrinsic Inductive Bias | 1 + ...tanding via Video-Distilled Knowledge Transfer | 1 + ...n using Inter-Frame Communication Transformers | 1 + ...ess for Coordination Detection on Social Media | 1 + ...al Imitation Learning using Variational Models | 1 + ... Nets and Humans Share Similar Inherent Biases | 1 + ...rgence of Intermediate Visual Patterns in DNNs | 1 + .../VoiceMixer: Adversarial Voice Style Mixup | 1 + .../Volume Rendering of Neural Implicit Surfaces | 1 + ...uction of Multiple Objects from a Single Image | 1 + ...sis of Representation Learning in Actor-Critic | 1 + ...grained Classification via Similarity Transfer | 1 + ...for offline model-based reinforcement learning | 1 + ...Weisfeiler and Lehman Go Cellular: CW Networks | 1 + ...ll-tuned Simple Nets Excel on Tabular Datasets | 1 + ...i-Modal Learning Better than Single (Provably) | 1 + ...at Matters for Adversarial Imitation Learning? | 1 + ...al networks actually say about generalization? | 1 + ...aining reveals about neural network complexity | 1 + ...ood imputation to predict with missing values? | 1 + ...When Are Solutions Connected in Deep Networks? | 1 + ... Trainability: Fewer than $n$ Neurons Can Work | 1 + ...mization with Low FPR for Multipartite Ranking | 1 + ...eneralizable Reinforcement Learning Tractable? | 1 + ...When Is Unsupervised Disentanglement Possible? | 1 + ...ial Robustness from Pretraining to Finetuning? | 1 + ...tainty Quantification for Epidemic Forecasting | 1 + ...earning Objectives are Sufficient for Control? | 1 + ...s and Who Follows in Strategic Classification? | 1 + ... Functions Lead to Less Transferable Features? | 1 + ...m Tasks? An Analysis of Head and Prompt Tuning | 1 + ...emic POMDPs and Implicit Partial Observability | 1 + ...of Sample Complexity on Sparse Neural Networks | 0 ...ion Stabilizes GANs: Analysis and Improvements | 1 + ...xplanation and Context-Aware Data Augmentation | 1 + ...s and Heuristics Over the Atari-2600 Benchmark | 1 + ...gregation of Voter Information and Preferences | 1 + ...s Functions for Diachronic Word Representation | 1 + .../XCiT: Cross-Covariance Image Transformers | 1 + ...uble Oracle Algorithm for Extensive-Form Games | 1 + ...wn Papers: An Owner-Assisted Scoring Mechanism | 1 + data/2021/neurips/You Never Cluster Alone | 1 + ...Transformer in Vision through Object Detection | 1 + ...t! Making a lottery ticket claim its ownership | 1 + ...al-driven models of the primate dorsal pathway | 1 + ...ling Predictions in Early Exit Neural Networks | 1 + data/2021/neurips/argmax centroid | 0 ...cient Lossless Compression via a Uniform Coder | 1 + ...gh-dimensional Neural Tangent Kernel Approach" | 1 + ...ass-Contrastive Back-Propagation Explanations" | 1 + ...antitative Study of Scalability with Dimension | 1 + ...domized Smoothing for Decision Stump Ensembles | 1 + ...ine Bipartite Matching with Degree Information | 1 + ...y for Self-training and Hyper-parameter Tuning | 1 + .../neurips/3D Concept Grounding on Neural Fields | 1 + ...Framework for Debugging Computer Vision Models | 1 + ...egular Latent Grids for 3D Generative Modeling | 1 + ...ing Semantic Novelty Detection on Point Clouds | 1 + .../2022/neurips/4D Unsupervised Object Discovery | 1 + ...A Benchmark for Compositional Visual Reasoning | 1 + ...ds Algorithm for Bandits with Delayed Feedback | 1 + .../A Boosting Approach to Reinforcement Learning | 1 + data/2022/neurips/A Causal Analysis of Harm | 1 + ... for Non-Autoregressive Sentence Summarization | 1 + ...pervised Adversarially Robust PAC Learnability | 1 + ...n: Stability, Robustness, and Inductive Biases | 1 + .../neurips/A Closer Look at Offline RL Agents | 0 ...e Classifier for Few-shot Image Classification | 1 + ...ly-Supervised Audio-Visual Source Localization | 1 + ...ersarial Robustness of Deep Equilibrium Models | 1 + ...e on the Optimization of Shallow ReLU Networks | 1 + ...ng Algorithm for Training Deep Neural Networks | 1 + ...ear Convergence for Federated Minimax Learning | 1 + ...le Graph Training: Benchmarking and Rethinking | 1 + ...r Sparse Logistic Regression in High-Dimension | 1 + ...iable Lp Canonical Calibration Error Estimator | 1 + ...for Support Vector Machines via Data Reduction | 0 ...s Time Framework for Discrete Denoising Models | 1 + ...ntrastive Framework for Neural Text Generation | 1 + ...rity for Practical Vertical Federated Learning | 1 + ... Analytical Moments And Sampling-Free Training | 0 ... Development Goal of Safe Working Environments | 1 + ...arning Dataloader with Shared Data Preparation | 1 + ...ement Learning Framework for Column Generation | 1 + ...babilistic Embedding for Cross-Modal Retrieval | 1 + ...e fPTAS for the Minimum Enclosing Ball Problem | 1 + ...ation of AIXI Using Logical State Abstractions | 1 + ...st-Training Pruning Framework for Transformers | 1 + ...-negative Least Squares with Non-negative Data | 1 + ... RL with Resets and Linear Value Approximation | 1 + .../A Fourier Approach to Mixture Learning | 1 + ...diting Differentially Private Machine Learning | 1 + ...metric Perspective on Variational Autoencoders | 1 + ...mputational Linguistics and Political Analysis | 1 + ...istic for Assessing Implicit Generative Models | 1 + ...Lagrangian Duality Approach to Active Learning | 1 + ...e Search Dataset for Unbiased Learning to Rank | 1 + .../A Lower Bound of Hash Codes' Performance | 1 + ...esource Management with Function Approximation | 1 + ...prises for Unsupervised Reinforcement Learning | 1 + ... U-Nets with Applications to Hierarchical VAEs | 1 + ...anguage Understanding and Judgement Prediction | 1 + ...mework for Approximate Nearest Neighbor Search | 1 + ...rithm for Online Learning with Feedback Graphs | 1 + ...al-Dual Method for Off-Policy Learning in CMDP | 1 + ...A Neural Corpus Indexer for Document Retrieval | 1 + ... Learning Algorithm to Reduce Label Complexity | 1 + ...lization Bounds Using Samplewise Evaluated CMI | 1 + ...for High-Dimensional Generalized Linear Models | 1 + ...of Non-parametric Temporal-Difference Learning | 1 + ... Generalization Bound for Equivariant Networks | 1 + ...on Approach for Offline Reinforcement Learning | 1 + .../A Practical, Progressively-Expressive GNN | 1 + ...tic Graph Coupling View of Dimension Reduction | 1 + ...tochastic Multi-level Composition Optimization | 1 + ...trol Variates and Adaptive Importance Sampling | 1 + ...eometric Approach to Neural-Network Smoothness | 1 + ...ary Approach for Debiasing Multiclass Datasets | 1 + ...A Regret-Variance Trade-Off in Online Learning | 1 + ...harpness Measure Based on Information Geometry | 1 + ...r Corruption-Tolerant Gaussian Process Bandits | 1 + ...ution for Hierarchical Representation Learning | 1 + ...n Algorithm for Training Optimal Decision Tree | 0 ...mple Approach to Automated Spectral Clustering | 1 + .../A Simple Decentralized Cross-Entropy Method | 1 + ...Learning with Safety against Heavy-tailed Risk | 1 + ...ynchronous Federated Contextual Linear Bandits | 1 + ... Approximation with Multiple Coupled Sequences | 1 + ... Scalable Learning in Neural ILP Architectures | 1 + .../A Spectral Approach to Item Response Theory | 1 + ... Approach in Averaged Stochastic Approximation | 0 ... Method for Decentralized Bilevel Optimization | 1 + ...ublicly Available US Criminal Justice Datasets | 1 + ...A Theoretical Framework for Inference Learning | 1 + ...heoretical Study on Solving Continual Learning | 1 + ...f Gradient Bias in Meta-Reinforcement Learning | 1 + ...heoretical View on Sparsely Activated Networks | 1 + ... Learnability under Transformation Invariances | 1 + ...ctor with Coarse-Fine Crossing Representations | 1 + ...d Learning with Arbitrary Client Participation | 1 + ...Data Augmentation: A Loss Function Perspective | 1 + ...ce Theorem for Stochastic Optimization Methods | 1 + ... Measure for Multiagent Reinforcement Learning | 0 ...l Backdoor Learning: Frameworks and Benchmarks | 1 + ...ing Offline Model Training and Policy Learning | 1 + ...Unified Framework for Deep Symbolic Regression | 0 ...amework for Solving Geometrically Complex PDEs | 1 + ...nified Model for Multi-class Anomaly Detection | 1 + .../A Unified Sequence Interface for Vision Tasks | 1 + ...Online Optimization with Long-Term Constraints | 1 + ...f Off-Policy General Value Function Evaluation | 1 + ...t Predictions Applied to Online Graph Problems | 1 + ...nt of Anderson Mixing with Minimal Memory Size | 1 + ...l for Supervised Graph Representation Learning | 1 + ... Sparse and Robust Pre-trained Language Models | 1 + ...ady-state simulations on high-resolution grids | 1 + .../A consistently adaptive trust-region method | 1 + .../neurips/A contrastive rule for meta-learning | 1 + ...astic and global variance reduction algorithms | 1 + ...ero-order optimization with two point feedback | 1 + ...Lipschitz functions in high and low dimensions | 1 + ... dataset for multilingual keyphrase generation | 0 .../A permutation-free kernel two-sample test | 1 + ...F result with applications in network modeling | 1 + ...ntinual learning: Repeated Augmented Rehearsal | 1 + ...ry of weight distribution-constrained learning | 0 ...ormation encoding in recurrent neural networks | 1 + ...ted Attacker for Boosting Adversarial Training | 1 + ...h Absolute Memorization and Privacy Protection | 1 + ... Dropout for Robust Language Model Fine-Tuning | 1 + .../neurips/ADBench: Anomaly Detection Benchmark | 1 + ...e Replay for Incremental Semantic Segmentation | 1 + ...hical Learning for Composite Multi-Agent Tasks | 1 + ...hmark for Versatile Medical Image Segmentation | 1 + ...rallel Strategies with Heterogeneity Awareness | 1 + ...tion Network for Click-Through Rate Prediction | 1 + ...chmark for Animal Pose Estimation and Tracking | 1 + ...aptive Skill Priors for Reinforcement Learning | 1 + ...ng CP Tensor Decomposition by Self Supervision | 1 + ...n for Compute-Efficient Hyper-parameter Tuning | 1 + ...anguage Embodied Navigation in 3D Environments | 1 + ...signal uncorrelation on spatio-temporal graphs | 1 + ...place Approximation for Bayesian Deep Learning | 1 + ...e Saddle-Point Problems with Bilinear Coupling | 1 + ...for Sparsity Constrained Optimization Problems | 1 + ...etworks (PINNs) using Meshless Discretizations | 1 + ...ied Robustness Training via Knowledge Transfer | 1 + ...onditioned Huge-Scale Online Matrix Completion | 1 + ...e Convolution with Column Vector-Wise Sparsity | 1 + .../Acceleration in Distributed Sparse Regression | 0 ...ivity arises from distributed control policies | 1 + ...sing Wearable Sensors in a Kitchen Environment | 1 + .../2022/neurips/Active Bayesian Causal Inference | 1 + ...Exploration for Inverse Reinforcement Learning | 1 + ...ctive Labeling: Streaming Stochastic Gradients | 1 + ...elps Pretrained Models Learn the Intended Task | 1 + ...Active Learning Polynomial Threshold Functions | 1 + .../Active Learning Through a Covering Lens | 1 + .../Active Learning for Multiple Target Models | 0 ...ing of Classifiers with Label and Seed Queries | 1 + ...tworks: Insights from Nonparametric Statistics | 1 + .../Active Learning with Safety Constraints | 1 + ...Ranking without Strong Stochastic Transitivity | 1 + ...g Approach to Label-Efficient Model Evaluation | 1 + ...bilities of Deep Learning-based Stereo Methods | 1 + ...daFocal: Calibration-aware Adaptive Focal Loss | 1 + ...verge Without Any Modification On Update Rules | 1 + ...n Transformers for Scalable Visual Recognition | 1 + ...ayesian Inference in Attractor Neural Networks | 1 + ...bing Attention-Conditioned Masking Consistency | 1 + ...to Online Label Shift with Provable Guarantees | 1 + ...ive Data Debiasing through Bounded Exploration | 1 + ...t Learning with Hierarchical Optimal Transport | 1 + ...e Interest for Emphatic Reinforcement Learning | 1 + ...n for Learning Latent Space Energy-based Model | 1 + .../Adaptive Oracle-Efficient Online Learning | 1 + data/2022/neurips/Adaptive Sampling for Discovery | 1 + ...duction for Non-convex Finite-Sum Minimization | 1 + ...ly Exploiting d-Separators with Causal Bandits | 1 + ...table Multiple Instance Learning for Pathology | 1 + ...ddressing Leakage in Concept Bottleneck Models | 1 + ...al Pretraining and Unified-Vocabulary Datasets | 1 + ...Gaussian process driven differential equations | 1 + ...ferable Adversarial Attack on Face Recognition | 1 + ...ancing Model Pruning via Bi-level Optimization | 1 + ...o Mitigate Black-Box Score-Based Query Attacks | 1 + ...resentation Learning Principle Guided Approach | 1 + .../neurips/Adversarial Reprogramming Revisited | 1 + ...arial Robustness is at Odds with Lazy Training | 1 + ...or Domain Generalized Urban-Scene Segmentation | 1 + ...Adversarial Task Up-sampling for Meta-learning | 1 + ...n the Benefit of Gradually Informative Attacks | 1 + ...ducing Confidence Along Adversarial Directions | 1 + ...versarial training for high-stakes reliability | 1 + ...c Minimax Optimal Learner and Characterization | 1 + ...etable Assessment of Implicit Graph Generators | 1 + ...ce of Neural Networks under Distribution Shift | 1 + ...ting Reynolds-Averaged Navier-Stokes Solutions | 1 + ...rning Linear Thresholds from Label Proportions | 1 + ...mate Data Removal: New Results and Limitations | 1 + .../neurips/Algorithms with Prediction Portfolios | 1 + ...lustering with Anchor Matching Correspondences | 1 + ...rains with fused unbalanced Gromov Wasserstein | 1 + ...emporal Attention for Video Action Recognition | 1 + ...ics is Local: Redistricting via Local Fairness | 1 + ...se\" in Deep Topic Models via Policy Gradient" | 1 + ... Attacks on Variational Autoencoders with MCMC | 1 + ...earning by Removing Projection to the Centroid | 1 + ...g Mirror Descent for Constrained Min-Max Games | 1 + ...dgments for Robust Visual Event Classification | 1 + ...rtized Inference for Causal Structure Learning | 1 + ...ce for Heterogeneous Reconstruction in Cryo-EM | 1 + ...tized Mixing Coupling Processes for Clustering | 1 + ...ation for Sliced Wasserstein Generative Models | 1 + data/2022/neurips/Amortized Proximal Optimization | 1 + ...lifying Membership Exposure via Data Poisoning | 1 + ...ary Environments with Piecewise Stable Context | 1 + ...rated Learning of Heterogeneous Causal Effects | 1 + ...or Learning Switched Linear Dynamics from Data | 1 + .../2022/neurips/An Analysis of Ensemble Sampling | 1 + ...urriculum Learning in Teacher-Student Networks | 1 + ...tched Algorithm for the Dueling Bandit Problem | 1 + ... Approach to Semi-Supervised Few-Shot Learning | 1 + ...nglement of Negative-free Contrastive Learning | 1 + ...n In-depth Study of Stochastic Backpropagation | 1 + ...ormation-Theoretic Framework for Deep Learning | 1 + ...to Whitening Loss for Self-supervised Learning | 1 + ... ultra-large combinatorial synthesis libraries | 1 + ... compute-optimal large language model training | 0 ...tric Properties for Graph Contrastive Learning | 1 + ...ypothesis from PAC-Bayesian Theory Perspective | 1 + ...: Progressive Sharpening and Edge of Stability | 1 + ...ent for Multi-Objective Reinforcement Learning | 1 + ...-Aware Face Image Generation for Video Avatars | 1 + ...sual Correspondence from Open Source 3D Movies | 1 + ...d Super-Resolution Models for Animation Videos | 1 + ... of Spurious Minima in Two-Layer ReLU Networks | 1 + ...t Benchmark for Unsupervised Anomaly Detection | 1 + ...ized Histograms in Intermediate Privacy Models | 1 + .../Anonymous Bandits for Multi-User Systems | 1 + ... Performativity by Predicting from Predictions | 1 + ...Based Generative Models for Protein Structures | 1 + ...ime-Valid Inference For Multinomial Count Data | 1 + ...with Application to Gradient-Free Optimization | 1 + ...ths and distances beyond Johnson-Lindenstrauss | 1 + ...ations for the Cubic Regularization Subproblem | 1 + data/2022/neurips/Approximate Value Equivalence | 1 + ...lev Space: with Applications to Classification | 1 + ...s in High Dimensions Under Minimal Assumptions | 1 + ...s Created Equal: A Neural Collapse Perspective | 1 + ...ke Agents Robust to Adversarial Perturbations? | 1 + ...Are Defenses for Graph Neural Networks Robust? | 1 + data/2022/neurips/Are GANs overkill for NLP? | 1 + ...ng Disparate Treatment in Fair Neural Networks | 1 + ...lation for Fingerprinting Deep Neural Networks | 1 + ...ive Sparse Labeling for Video Action Detection | 1 + ...rning to Leverage an Expert for Embodied Tasks | 1 + ...stribution Generalization in Transfer Learning | 1 + ...tive Teaching of Motor Control Tasks to Humans | 1 + ...ir Effects in Video through Coordination Games | 1 + ...Multi-Task Classification with Category Shifts | 1 + ...Scaling Makes Larger Networks Teach Well Again | 1 + ...ic Approximation: A Jump Diffusion Perspective | 1 + ...ies for Bayesian Neural Network in Besov Space | 1 + ...Partial AUC Optimization: Theory and Algorithm | 1 + ...asserstein distances in the small noise regime | 1 + ... \342\204\2232 Regularized Network Embeddings" | 1 + ...-Critic for Multi-Agent Reinforcement Learning | 1 + ...SGD Beats Minibatch SGD Under Arbitrary Delays | 1 + ...sformers via Attentive Class Activation Tokens | 1 + .../Attention-based Neural Cellular Automata | 1 + ...ple Approach for Source-free Domain Adaptation | 1 + ...udio-Driven Co-Speech Gesture Video Generation | 1 + ...ontrastive Learning: Fabricated and Generative | 1 + ...r Adaptive Control of Linear Quadratic Systems | 1 + ...etons and Object Outlines by Linking Keypoints | 1 + data/2022/neurips/AutoML Two-Sample Test | 1 + ... for Novelty Detection with Error Rate Control | 1 + ...k for Automating Efficient Multi-Task Learning | 1 + ...niversal Modeling of Spatio-temporal Sequences | 1 + ...ing Automated Weak Supervision with 100 Labels | 1 + .../Autoformalization with Large Language Models | 1 + ...Uncertainty Aware Inversion of Neural Networks | 1 + ...entiation of Programs with Discrete Randomness | 1 + ...ferentiation of nonsmooth iterative algorithms | 1 + ...utoregressive Perturbations for Data Poisoning | 1 + ... Generating Substrings as Document Identifiers | 1 + ...ralization Using Procedurally Generated Worlds | 1 + .../Average Sensitivity of Euclidean k-Clustering | 1 + ...ex Optimization with Communication Compression | 1 + ...imple and Robust LiDAR-Camera Fusion Framework | 1 + ...t Algorithm for Joint Alignment of Time Series | 0 ...l Architecture Search Benchmark and Algorithms | 1 + ...m Update for Continual Video-Language Modeling | 1 + ...ation Made Easy: A Simple First-Order Approach | 1 + ...ier Node Detection on Static Attributed Graphs | 1 + ...as Reduced Self-Normalized Importance Sampling | 1 + ...xplore: Exploration by Bootstrapped Prediction | 1 + ...er Learning by Self-Sparsified Backpropagation | 1 + ...A Comprehensive Benchmark of Backdoor Learning | 1 + ...Prompt: Backdoor Attacks on Continuous Prompts | 1 + ...ip: A Certified Defense Against Data Poisoning | 1 + ...d Directed Evolution for Sequence Optimization | 1 + ...utations using the Acquisition Weighted Kernel | 1 + ...n via density-ratio estimation with guarantees | 1 + ...delity Active Learning with Budget Constraints | 1 + .../Batch size-invariance for policy optimization | 1 + ...ilistically Triggered Arms or Independent Arms | 1 + ...Learnable Predictive Coding Associative Memory | 1 + ...earning with Fully Bayesian Gaussian Processes | 1 + ... a Mixture of Dynamic Poisson Factor Analyzers | 1 + ...oration for Model-based Reinforcement Learning | 1 + ...ed Spaces via Probabilistic Reparameterization | 1 + .../Bayesian Persuasion for Algorithmic Recourse | 1 + .../Bayesian Risk Markov Decision Processes | 1 + ...Nonlinear Dynamics with Quantified Uncertainty | 1 + ...ayesian inference via sparse Hamiltonian flows | 1 + ...Transformers: Cloning $k$ modes with one stone | 1 + ...onalization for Offline Reinforcement Learning | 1 + ...ct Models through the Lens of Interpretability | 1 + ...an Pose and Shape Estimation Beyond Algorithms | 1 + ...ient and collaborative optimization benchmarks | 1 + ...ise in Composing Classes with Bounded Capacity | 1 + ...Permutation-Equivariance in Auction Mechanisms | 1 + ...ing in Two-layer Convolutional Neural Networks | 1 + ...gn Underfitting of Stochastic Gradient Descent | 1 + ...phic: Toward a Refined Taxonomy of Overfitting | 0 ...nsmission Effects in Multi-Mode Optical Fibres | 1 + .../neurips/Best of Both Worlds Model Selection | 1 + ...Worlds Bounds for Bandits with Switching Costs | 1 + .../Better SGD using Second-order Momentum | 1 + ...ia Proper Scores for Classification and Beyond | 1 + ...ization: Improved Regret Bounds via Smoothness | 1 + ...ti-Class Prediction via Information Projection | 1 + ... decision-making in heterogeneous environments | 1 + ...L1: Faster and Better Sparse Models with skglm | 1 + ...Mahalanobis Distance for Textual OOD Detection | 0 ...nual Learning with Backward Knowledge Transfer | 1 + ...pirical Study of Node Classification with GNNs | 1 + ...tive on Offline Multiagent Behavioral Analysis | 1 + ...tive Representations to Related Subpopulations | 1 + ... via Clairvoyant Multiplicative Weights Update | 1 + ...bio-plausible temporal credit assignment rules | 1 + ...Parameter learning for the deviated components | 1 + ...ws: beating power law scaling via data pruning | 1 + ...role of the topology in decentralized learning | 1 + ...unctional Estimation in Infinite-Armed Bandits | 1 + ...r User-specified Error-measuring Distributions | 1 + ...zier Gaussian Processes for Tall and Wide Data | 1 + ...tillation for Whole Slide Image Classification | 1 + ...chitectures for Vision Multi-Layer Perceptrons | 1 + ...Robustly Binarized Multi-distilled Transformer | 1 + ...ffline Infinite-width Model-based Optimization | 1 + ...Centric Biomedical Natural Language Processing | 1 + ...obabilistic Model for Binaural Audio Synthesis | 1 + ...e Representations of Commuting Transformations | 1 + ...Dynamic Thresholds for Spiking Neural Networks | 1 + ...ons for spiking networks with efficient coding | 1 + ...rks for Blind Separation of Correlated Sources | 1 + ... arbitrary timespans via local neuromodulators | 1 + ... Classification with Optimal Label Permutation | 1 + ...ralization: Stability of Zeroth-Order Learning | 1 + .../Black-box coreset variational inference | 1 + ...Blackbox Attacks via Surrogate Ensemble Search | 1 + ...ave Flatter Landscape Around the True Solution | 1 + data/2022/neurips/Block-Recurrent Transformers | 1 + ...s: A New Perspective on Adversarial Robustness | 1 + ...f-distribution Detection with Typical Features | 1 + ... Network Frameworks with Log-supermodular CRFs | 1 + ... Attacks with Reverse Adversarial Perturbation | 1 + ...Transformer for Offline Reinforcement Learning | 1 + ...rediction Error, Constraints, and Nonlinearity | 1 + ...ersectional Fairness through Marginal Fairness | 1 + data/2022/neurips/Brain Network Transformer | 1 + ...ratively Solvable Problems in Predict+Optimize | 0 ... Dataset for Geometric Fracture and Reassembly | 1 + ...chitecture Spaces via A Cross-Domain Predictor | 1 + ...rential Privacy in Data Acquisition Mechanisms | 1 + ...onvolutional Neural Networks on Small Datasets | 1 + ... Representations for Open-Vocabulary Detection | 1 + ...es in Non-contrastive Self-supervised Learning | 0 ...valuation of Neural Network Binary Classifiers | 1 + ...tially Private Stochastic Minimax Optimization | 1 + ...eo via Frame-Clip Consistency of Object Tokens | 1 + ...mizing Privacy Subject to Accuracy Constraints | 1 + data/2022/neurips/Byzantine Spectral Ranking | 1 + ...Gaussian process regression for streaming data | 1 + ...-Mixup: Improving Generalization in Regression | 1 + ...Networks for Precise Probabilistic Forecasting | 1 + ...ating Multimodal Referring Expression Datasets | 0 ...ouping for 3D Object Detection on Point Clouds | 1 + ...Classification and Regression Diffusion Models | 1 + ...from Simulation to multiple Real-World Domains | 1 + ...tegory-agnostic Skeletal Animal Reconstruction | 1 + ...ext Generation APIs via Conditional Watermarks | 1 + data/2022/neurips/CCCP is Frank-Wolfe in disguise | 1 + ...s of Real-World Concepts on NLP Model Behavior | 1 + ...ons for Optical Chemical Structure Recognition | 1 + ...for Reinforcement Learning with Demonstrations | 1 + ...: Benchmark Tasks for Continual Graph Learning | 1 + ...MLE for Multimodal Conditional Image Synthesis | 1 + ...nerative Counterfactual Explanations on Graphs | 1 + ...ibing Physical and Causal Events the Human Way | 1 + ...wing Synthesis through Language-Image Encoders | 1 + ...opfield Networks with InfoLOOB Outperform CLIP | 1 + ...arning Benchmark for Vision-and-Language Tasks | 1 + ...trained Text Generation with Langevin Dynamics | 1 + ...ey Values for Data Valuation in Classification | 1 + data/2022/neurips/CUP: Critic-Guided Policy Reuse | 1 + ... Resampling for Training Recommender Retriever | 1 + ...d for Generalized 3D Deformation and Animation | 1 + ...rated Adversarial Training with Label Skewness | 1 + ...Constraints with Exact Satisfaction Guarantees | 1 + ...raining Be Manipulated By Non-Robust Features? | 1 + ...etworks Help Solve the Maximum Clique Problem? | 1 + ...enerative Models Fit Multimodal Distributions? | 1 + ...rge Language Models via Human Cognitive Biases | 1 + ...Capturing Graphs with Hypo-Elliptic Diffusions | 1 + ...Training in Extreme Multi-label Classification | 1 + ...Supervised Learning Approach and A New Dataset | 1 + ...ts Under the Sparse Mechanism Shift Hypothesis | 1 + ...t Variable Models Subject to Measurement Error | 1 + ...valence: Calculus, Algorithm, and Completeness | 1 + ...ith Non-IID Data using Linear Graphical Models | 1 + ...n and Classification using Neurochaos Learning | 1 + ...Structure Discovery for Reinforcement Learning | 1 + ...ated multi-shortcut identification and removal | 0 ...tworks for Distribution-Free Survival Analysis | 1 + ...on under Orthogonal Gromov-Wasserstein Threats | 1 + ...onal Fairness with Subpopulation Decomposition | 1 + ...in of Thought Imitation with Procedure Cloning | 1 + ...ing Elicits Reasoning in Large Language Models | 1 + ...n Assumptions in Convex Reinforcement Learning | 1 + ...ry from Spatio-temporal Remote Sensing Imagery | 1 + ...nd Dense Functional Data in General Dimensions | 1 + ... Intrinsic to Neural Network Training with SGD | 1 + ...iled Limits for Deterministic Gradient Descent | 1 + ...wards Rigorous Benchmarking of Language Models | 1 + ...sk for Locally Strongly Convex Population Risk | 1 + ...with Response-Optimized Neural Encoding Models | 1 + ...atasets for UTXO and Account-based Blockchains | 1 + ...ndom Tables: Non-Trigonometric Random Features | 1 + ... Shortcut Learning with Generative Classifiers | 1 + .../Chromatic Correlation Clustering, Revisited | 1 + ...al Transformers for Medical Image Segmentation | 1 + ...Learning with Cycle-Consistency Regularization | 1 + ...on Enabling Robustness on Efficient Inferences | 1 + ...riational Inequalities with Heavy-Tailed Noise | 1 + ...ed Designs for One-Sided Bipartite Experiments | 1 + ...gregate: Face Recognition with Large Probe Set | 1 + ...ve Learning for Imbalanced Node Classification | 1 + .../CoNSoLe: Convex Neural Symbolic Learning | 1 + .../CoNT: Contrastive Neural Text Generation | 1 + ...llaborative Inference via Feature Purification | 1 + ...guage Pre-training with Fusion in the Backbone | 1 + ...trained Models and Deep Reinforcement Learning | 1 + ...ansform for Generalizable Deep Metric Learning | 1 + ...Image Generation via Hierarchical Transformers | 1 + ...ative Decision Making Using Action Suggestions | 1 + ...e Learning by Detecting Collaboration Partners | 1 + ...er Heterogeneity and Communication Constraints | 1 + ...Adversarial Agents: Near-Optimal Regret Bounds | 1 + ...icient Message Passing for 3D Molecular Graphs | 1 + ...glement and Segmentation via Image Composition | 1 + ...MU: Dataset for Combinatorial Music Generation | 1 + ...ear Constraints: Beyond Knapsacks and Fairness | 1 + ...zation for Efficient Learning in Deep Networks | 1 + ...cating Natural Programs to Humans and Machines | 1 + ...ted Primal-Dual Algorithm with an Inexact Prox | 1 + ...ted Learning for Kernelized Contextual Bandits | 1 + ...erated Learning for Generalized Linear Bandits | 1 + ...entralized Learning with $O(1)$ Consensus Rate | 1 + ...nspace estimation with arbitrary node failures | 1 + ...mposite Feature Selection Using Deep Ensembles | 1 + ... Theorems for Interactive Differential Privacy | 1 + ...Study on Disentanglement and Emergent Language | 1 + ...ations in human and artificial neural networks | 1 + ...omposable NeRF via Rank-residual Decomposition | 1 + ...Reinforcement Learning for Linear Mixture MDPs | 1 + ...ata Encoding in Parameterized Quantum Circuits | 1 + ...lized Framework For Concept-Based Explanations | 1 + ...: Beyond the Accuracy-Explainability Trade-Off | 1 + ...: Generalized Score Matching for Discrete Data | 1 + ...ional Diffusion Process for Inverse Halftoning | 1 + ...stic Data and Applications to Causal Discovery | 1 + ...tional Meta-Learning of Linear Representations | 1 + ... Free-Standing Social Interactions in the Wild | 1 + ...ence-based Reliable Learning under Dual Noises | 1 + .../neurips/Confident Adaptive Language Modeling | 1 + ...formal Frequency Estimation with Sketched Data | 1 + ...al Off-Policy Prediction in Contextual Bandits | 1 + ... Prediction with Temporal Quantile Adjustments | 1 + ...Conformalized Fairness via Quantile Regression | 1 + ...ting Image Data Privacy with Causal Confounder | 1 + ...r Efficient Model-Based Reinforcement Learning | 1 + ...ng under Graph Induced Fair Planted Partitions | 1 + ...ting Ensembles via the Manifold-Hilbert Kernel | 1 + ...ng the decision of any classifier or regressor | 0 data/2022/neurips/Constants of motion network | 1 + ...r Zero-Shot Transfer in Reinforcement Learning | 1 + ...rithms with L-mixing External Random Variables | 1 + ...ally Plausible Model of the Cortical Hierarchy | 1 + ... Optimization with State-dependent Markov Data | 1 + ...rojection Approach to Safe Policy Optimization | 1 + ...tems of Linear Ordinary Differential Equations | 1 + .../Contact-aware Human Motion Forecasting | 1 + ...mic Pricing with Partially Linear Demand Model | 0 ... Bandits with Knapsacks for a Conversion Model | 1 + ...Explore-then-UCB Strategy and Improved Regrets | 1 + ...on for Efficient Few-Shot Image Classification | 1 + ...g In Environments With Polynomial Mixing Times | 1 + ...tinual Learning with Evolving Class Ontologies | 1 + ...icient algorithm, and fundamental obstructions | 1 + ...blems: Normalized Advantage Functions Analysis | 1 + ... Homomorphisms and Homomorphic Policy Gradient | 1 + .../neurips/Continuously Tempered PDMP samplers | 1 + ...Adapters for Foundation Model Group Robustness | 1 + ... via Information Bottleneck for Recommendation | 1 + ...guage-Image Pre-Training with Knowledge Graphs | 1 + ...ing as Goal-Conditioned Reinforcement Learning | 1 + .../neurips/Contrastive Neural Ratio Estimation | 1 + ...er Global and Local Spectral Embedding Methods | 1 + ...s with Conditional Generative Occupancy Fields | 1 + ...ext Generation with Neurally-Decomposed Oracle | 1 + ... to Stop Tuning Penalties and Love Constraints | 1 + ...-parameterized regime using Rayleigh quotients | 1 + ...generative modeling with polynomial complexity | 1 + ...ograms in Human and Artificial Neural Networks | 1 + .../neurips/Convexity Certificates from Hessians | 1 + ...Graphs with Chebyshev Approximation, Revisited | 1 + ...ive Distribution Alignment via JSD Upper Bound | 1 + ...e Reduction for Generalized Linear Programming | 1 + ...Prior Helps Implicit Neural 3D representations | 1 + .../2022/neurips/Coreset for Line-Sets Clustering | 1 + ...esets for Relational Data and The Applications | 1 + ...zed Linear Regression and $K$-Means Clustering | 1 + ... Distributionally Robust Optimization Problems | 1 + ...aining for Optimizing Non-Decomposable Metrics | 1 + ...etwork embeddings for tensor-structured inputs | 1 + ...mage Models Extract Universal Representations? | 1 + ...ual Fairness with Partially Known Causal Graph | 1 + ...al Influence of Misinformation on Social Media | 1 + .../Counterfactual Temporal Point Processes | 1 + data/2022/neurips/Counterfactual harm | 1 + ...sk Alignments for Referring Image Segmentation | 1 + ...g for 3D Vision Tasks by Cross-View Completion | 1 + ... Aggregation Transformer for Image Restoration | 1 + ...ross-Image Context for Single Image Inpainting | 1 + ...ing for cross-modality representation learning | 1 + ... for Image-Guided Point Cloud Shape Completion | 1 + ...ncrypted Graph Convolutional Network Inference | 1 + ...ness of Learning Halfspaces with Massart Noise | 1 + ...ld Models Yields Zero-Shot Object Manipulation | 1 + ...ptimal Transport via Gradual Domain Adaptation | 1 + ... Cyclic Contrastive Language-Image Pretraining | 1 + ... and Algorithms for Universal Self-Supervision | 1 + ... a Log-Determinant Acyclicity Characterization | 1 + ...Disentanglement-Augmented Rationale Extraction | 1 + ...del with Diverse Accessories and Rich Textures | 1 + ...ort Constrained Offline Reinforcement Learning | 1 + .../DC-BENCH: Dataset Condensation Benchmark | 1 + ... A New Dataset For Automatic Medical Diagnosis | 1 + .../DENSE: Data-Free One-Shot Federated Learning | 1 + ... Financial Dataset for Graph Anomaly Detection | 1 + ...and Sparse Hierarchical Reinforcement Learning | 1 + ...Solver for Combinatorial Optimization Problems | 1 + ...ersarial Defense with Local Implicit Functions | 1 + ... for learning to locomote with a changing body | 1 + ... Optimization with a Dual Network Architecture | 1 + ...ralized Context in Meta-Reinforcement Learning | 1 + ...ic Exploration for Safe Reinforcement Learning | 1 + ...tically Optimal and Differentially Private PCA | 1 + ...robabilistic Model Sampling in Around 10 Steps | 1 + ...ng for Non-IID Clients via Secret Data Sharing | 1 + ... Guidance for Semi-Supervised Object Detection | 1 + ...ptation for Unsupervised Semantic Segmentation | 1 + ...ning spike timing and reconstructive attention | 1 + ...MC for Bayesian Inference from Privatized Data | 1 + ... Advancing Predictive Models of the Microbiome | 1 + ...e Emergent In-Context Learning in Transformers | 1 + ...for efficient learning from parametric experts | 1 + .../Data-Driven Conditional Robust Optimization | 1 + ...n-Making via Invariant Representation Learning | 1 + ...ient Augmentation for Training Neural Networks | 1 + ...fline Reinforcement Learning with Limited Data | 1 + ...Structured Pruning via Submodular Optimization | 1 + ...ps with heterogeneous outcomes in tabular data | 1 + ...DataMUX: Data Multiplexing for Neural Networks | 1 + ...t Distillation using Neural Feature Regression | 1 + .../Dataset Distillation via Factorization | 1 + .../Dataset Inference for Self-Supervised Models | 1 + ...mable Voxel Radiance Fields for Dynamic Scenes | 1 + ...Effects Estimation with Unmeasured Confounding | 1 + ...without Sample-Splitting for Stable Estimators | 1 + ...sed Self-Training for Semi-Supervised Learning | 1 + ...ommendation through Multi-Visit Clinic Records | 1 + ... via Learning Disentangled Causal Substructure | 1 + ...aches: An Influence Function Based Perspective | 1 + ...level Optimization over Communication Networks | 1 + ...ic Extra-Gradient for Variational Inequalities | 1 + ...oundation Models in Heterogeneous Environments | 1 + ...n-free Learning in Structured Matching Markets | 1 + ...Equivalent Sampling for Reinforcement Learning | 1 + .../Decision Trees with Short Explainable Rules | 1 + ...ng: Learning Locally Optimized Decision Losses | 1 + ...ransformers via Patch-wise Adversarial Removal | 1 + ... with Nearly-Linear Gradient Oracle Complexity | 1 + ...on for Class-Incremental Semantic Segmentation | 1 + ...eRF for Editing via Feature Field Distillation | 1 + ...n Similarity for Comparison of Neural Networks | 1 + ...essing for Context Augmented Language Modeling | 1 + .../Decoupled Self-supervised Learning for Graphs | 1 + ...hot Object Detection and Instance Segmentation | 1 + ...ical Propagation for Video Object Segmentation | 1 + ...orization: Retrieval-augmented Prompt Learning | 1 + ...ctive Learning by Leveraging Training Dynamics | 1 + ...s for Its Convergence: A Fine-Grained Analysis | 1 + ...g for Solving Constraint Optimization Problems | 1 + ...rectional Language-Knowledge Graph Pretraining | 1 + data/2022/neurips/Deep Combinatorial Aggregation | 1 + ... Compression of Pre-trained Transformer Models | 1 + ...timation with Categorical Background Variables | 1 + .../Deep Differentiable Logic Gate Networks | 1 + .../Deep Ensembles Work, But Are They Necessary? | 1 + ...eep Equilibrium Approaches to Diffusion Models | 1 + data/2022/neurips/Deep Fourier Up-Sampling | 1 + .../Deep Generalized Schr\303\266dinger Bridge" | 1 + .../Deep Generative Model for Periodic Graphs | 1 + .../Deep Hierarchical Planning from Pixels | 1 + ...ximal Inference via Maximum Moment Restriction | 1 + data/2022/neurips/Deep Model Reassembly | 1 + ...al Effect Estimation With Unstructured Proxies | 1 + ... Surrogate Assisted Generation of Environments | 1 + ...tworks with differentiable augmentation layers | 1 + ...sh Simulation with Deep Reinforcement Learning | 1 + ...: 3D Object Detection via Modality Interaction | 1 + ...Mediation Analysis with Debiased Deep Learning | 1 + ...ep Threshold-Optimal Policy for MDPs and RMABs | 1 + ... Adversarial Attacks via Neural Dynamic System | 1 + .../Defining and Characterizing Reward Gaming | 0 ...e Transformer for Spectral Compressive Imaging | 1 + ...ging for Domain Adaptive Semantic Segmentation | 1 + ...Detection with Vision-Language Representations | 1 + ...into Sequential Patches for Deepfake Detection | 1 + .../Denoising Diffusion Restoration Models | 1 + .../neurips/Dense Interspecies Face Embedding | 1 + ...gularization for Out-of-distribution Detection | 1 + ...h with Prediction Concatenation in Deep Forest | 1 + ...gy Produce Heterogeneous Graph Neural Networks | 1 + ...ralleled Pre-training for Open-world Detection | 1 + ...Changes in Sequential Pairwise Comparison Data | 1 + ...zation of Changes in Conditional Distributions | 1 + ... with Normalizing Flows for Bayesian Inference | 1 + ...nnections for Locality Preserving Sparse Codes | 1 + ...: Differential Spectral Clustering of Features | 1 + ...tribution shift in real-world medical settings | 1 + ...es are as Effective as Structured State Spaces | 1 + ... Prior Dictionary Knowledge for Text-to-Speech | 1 + ...Quantum Computing for Optimization and Control | 1 + ...te gradient search for spiking neural networks | 1 + .../Differentially Private Covariance Revisited | 1 + ...ly Private Generalized Linear Models Revisited | 1 + ... via Sensitivity-Bounded Personalized PageRank | 1 + ...eeds Hidden State (Or Much Faster Convergence) | 1 + ...tially Private Learning with Margin Guarantees | 1 + ...es: Efficient Implementations and Applications | 1 + .../Differentially Private Model Compression | 1 + ...ally Private Online-to-batch for Smooth Losses | 1 + ...ating Local Curvature in High Dimensional Data | 1 + .../Diffusion Models as Plug-and-Play Priors | 1 + .../Diffusion Visual Counterfactual Explanations | 1 + ...usion-LM Improves Controllable Text Generation | 1 + ...cule Generation with Informative Prior Bridges | 1 + ...ularization for GAN Training with Limited Data | 1 + data/2022/neurips/Direct Advantage Estimation | 1 + data/2022/neurips/Discovered Policy Optimisation | 1 + .../Discovering Design Concepts for CAD Sketches | 1 + ...se-engineered Data-free Knowledge Distillation | 1 + ...iscovery of Single Independent Latent Variable | 1 + ...on for Goal Conditioned Reinforcement Learning | 0 ... for Warm-Starting Algorithms with Predictions | 1 + ...ions in the Presence of Unobserved Confounders | 1 + ...g Transfer in Continual Reinforcement Learning | 1 + ...ep Ensembles through the Neural Tangent Kernel | 1 + ...r Input Attribution in the Deep Neural Network | 1 + ...ations from GAN Generator via Squeeze and Span | 1 + ...g Learning Rules with Brain Machine Interfaces | 1 + ...l variability using warped autoregressive HMMs | 1 + ...Robust Optimization with Non-Convex Objectives | 1 + ...s for Parallel MARL in Large Networked Systems | 1 + ...Reinforcement Learning for Multi-agent Systems | 1 + ...ntiles in the Reproducing Kernel Hilbert Space | 1 + ...onal Inequalities, with Theoretical Guarantees | 1 + ...vex Optimization with Compressed Communication | 1 + ...Dimension Independent Communication Complexity | 1 + ...ural Networks for Domain Adaptation Regression | 0 ... Convergence of the Sliced Wasserstein Process | 1 + ...forcement Learning for Risk-Sensitive Policies | 1 + ...ective Multi-agent Deep Reinforcement Learning | 1 + ...utionally Adaptive Meta Reinforcement Learning | 1 + ...bust Optimization via Ball Oracle Acceleration | 1 + ...ionally Robust Optimization with Data Geometry | 1 + ...butionally robust weighted k-nearest neighbors | 1 + ...BO: Diversity-aware CASH for Ensemble Learning | 1 + ...eraging for Out-of-Distribution Generalization | 1 + ...endations for Agents with Adaptive Preferences | 1 + ...e generalization in one-shot generative models | 1 + ...ert More Attention to Vision-Language Tracking | 1 + ...n Adaptation via Adaptive Contrastive Learning | 1 + ...timization Methods in Deep Learning Even Help? | 1 + ...retize Neural Ordinary Differential Equations? | 1 + ...GNN Pretraining Help Molecular Representation? | 1 + ...the Implicit Regularization on Separable Data? | 1 + ...ly Improve Reinforcement Learning from Pixels? | 1 + ... meets Individual Fairness. And they get along | 1 + .../Domain Adaptation under Open Set Label Shift | 1 + ...Learning and Removing Domain-specific Features | 1 + ...n Generalization without Excess Empirical Risk | 1 + ...emporal Logic for Temporal Action Segmentation | 1 + ...ery Distortion of Matching Problems and Beyond | 0 ...cing Certified Robustness through Transitivity | 1 + ... Bidirectional Offline Model-Based Imagination | 1 + .../Doubly Robust Counterfactual Classification | 1 + ...Making Value Iteration Asynchronous in Actions | 1 + ...mage Generation with Contextual RQ-Transformer | 1 + ...ribution with Neuro-Symbolic Generative Models | 1 + ...edding Table Placement for Recommender Systems | 1 + ...ective Method for Improving Deep Architectures | 1 + ...cer Prognosis Analysis with Whole Slide Images | 1 + ...k for Imbalanced Graph-level Anomaly Detection | 1 + ...lti-Label Recognition with Limited Annotations | 1 + ...ngeons and Data: A Large-Scale NetHack Dataset | 1 + ...Dynamic Fair Division with Partial Information | 1 + ...works Under Spatio-Temporal Distribution Shift | 1 + ...nt Learning for Characterizing Animal Behavior | 1 + .../Dynamic Learning in Large Matching Markets | 1 + ...nstraint under Unknown Parametric Demand Model | 1 + ...ries Classification: Learning What to \"See\"" | 1 + .../neurips/Dynamic Tensor Product Regression | 1 + ...g and assortment under a contextual MNL demand | 1 + ...ive Variants and Convergence to Exact Solution | 1 + ...cement Learning with Parallel Program Guidance | 1 + ...Automatic Reward Shaping in Language-guided RL | 1 + ...biased Compression in Distributed Optimization | 1 + ...nergy-Guided Stochastic Differential Equations | 1 + ...to-SQL Benchmark for Electronic Health Records | 1 + ...ical Reasoning with Adaptive Symbolic Compiler | 1 + ...or Evaluating Language-Augmented Visual Models | 1 + ...ing to Index and Search in Large Output Spaces | 1 + ...on Alignment as a Multi-Agent Intrinsic Reward | 1 + ...For Post-Processing Ensemble Weather Forecasts | 1 + ...mark: VIdeo Segmentations and Object Relations | 1 + ...t Aware Dose Allocation for Precision Medicine | 1 + ...al Representation Learning in Echocardiography | 1 + ...o-Cost Proxies For Neural Architecture Scoring | 1 + ... Training Mildly Parameterized Neural Networks | 1 + ...Time Transformers for Earth System Forecasting | 1 + ...Energy-Saving Attention with Linear Complexity | 1 + ...ask Co-Training for Unified Autonomous Driving | 1 + ... by Exploiting Sensitivity of Poisoned Samples | 1 + ... Dimension in Bandit Problems under Censorship | 1 + ...and Accurate Single-Stage Pedestrian Detection | 1 + ...ffects of Data Geometry in Early Deep Learning | 1 + ...ciency Ordering of Stochastic Gradient Descent | 1 + .../Efficient Active Learning with Abstention | 1 + ...Worst-Case-Aware Robust Reinforcement Learning | 1 + ...d Kernel Tests using Incomplete $U$-statistics | 1 + ...fficient Architecture Search for Diverse Tasks | 1 + ...istillation using Random Feature Approximation | 1 + ...or Generalized Low-Rank Matrix Bandit Problems | 1 + ...rity Computation with Alignment Regularization | 1 + ... Knowledge Distillation from Model Checkpoints | 1 + ... Learning for Preference-based Fast Adaptation | 1 + ...ent Methods for Non-stationary Online Learning | 1 + ...on via Self-supervised Information Aggregation | 1 + ...-Parametric Optimizer Search for Diverse Tasks | 1 + ...Extensive-Form Games via Online Mirror Descent | 1 + .../Efficient Risk-Averse Reinforcement Learning | 1 + ...ling on Riemannian Manifolds via Langevin MCMC | 1 + ...a Augmentation for Deep Reinforcement Learning | 1 + ...ence for Conditional GANs and Diffusion Models | 1 + ...timization under Noise: Local Search is Robust | 1 + ...ient Training of Low-Curvature Neural Networks | 1 + ...Augmentation Strategy for Adversarial Training | 1 + ...rouping via Meta Learning on Task Combinations | 1 + ...Effective Optimal Transport-Based Biclustering | 1 + ...Efficient and Modular Implicit Differentiation | 1 + ...line Learning for Generalized Linear Functions | 1 + ...ent and Stable Fully Dynamic Facility Location | 1 + ...capacity, and the emergence of retinal mosaics | 1 + ...rmative features in simulation-based inference | 1 + ...models with time-series privileged information | 1 + ...Former: Vision Transformers at MobileNet Speed | 1 + ...tants of Neural Networks via Bound Propagation | 1 + ...olean Matrices using Proximal Gradient Descent | 1 + ...Understanding Human Tasks in Egocentric Videos | 1 + .../neurips/Egocentric Video-Language Pretraining | 1 + ...tion for self-supervised multi-view stereopsis | 0 .../Eliciting Thinking Hierarchy without a Prior | 1 + ...ign Space of Diffusion-Based Generative Models | 1 + ...amical systems with uncertainty quantification | 1 + .../Embodied Scene-aware Human Pose Estimation | 1 + ...p: VAEs Perform Independent Mechanism Analysis | 1 + ...e Approach for Spatio-Temporal Video Grounding | 1 + ...ingle Sheet of Self-Organizing Spiking Neurons | 1 + ... Generalization and Overfitting in Lewis Games | 1 + ...cal Conventions in a Visual Communication Game | 1 + ...rical Gateaux Derivatives for Causal Inference | 1 + ...hree-layer Neural Networks with Infinite Width | 1 + ... Evaluation Through Video Dataset Augmentation | 1 + ...t Networks: Extrapolation without Overthinking | 0 ...tochastic Optimization with Energy-based Model | 1 + ...d-to-end Symbolic Regression with Transformers | 1 + ...Contrastive Learning of Visual Representations | 1 + ...presentation via Discrete Adversarial Training | 1 + ...nced Bilevel Optimization via Bregman Distance | 1 + ...l Image Denoising via Alternative Optimization | 1 + ...a Demonstrations in Sparse Reward Environments | 0 ...fe Exploration Using Safety State Augmentation | 1 + ... Boosting Performance in Domain Generalization | 1 + ...Precision Quantization for Deep Network Design | 0 ...orcement Learning Environment Execution Engine | 1 + ...lti-head Neural Network for Invariant Learning | 1 + .../Envy-free Policy Teaching to Multiple Agents | 1 + .../EpiGRAF: Rethinking training of 3D GANs | 1 + ...ivariant Graph Hierarchy-Based Neural Networks | 1 + .../Equivariant Networks for Crystal Structures | 1 + ...quivariant Networks for Zero-Shot Coordination | 1 + ...r Analysis of Tensor-Train Cross Approximation | 1 + .../neurips/Error Correction Code Transformer | 1 + ...ective Generalization on Class-Imbalanced Data | 1 + ...ation Efficient Nonconvex Distributed Learning | 1 + ...lizations in Deep Variational Quantum Circuits | 1 + ...el Correlations for Noisy Multi-Label Learning | 1 + ...formance When Both Covariates and Labels Shift | 1 + ... with applications to single-cell gene network | 1 + ...l ROC Curve and Lower Bounding the Maximal AUC | 1 + ...Constant Space with Improved Sample Complexity | 1 + ...or Meta Learning: Tightness and Expressiveness | 1 + ...ive Models with Contrastively Learned Features | 1 + ...-ML Models under Realistic Distribution Shifts | 1 + ...tion Performance on Document Image Classifiers | 1 + ...o Dataset Shift via Parametric Robustness Sets | 1 + ...rmance: Analyzing Concepts in AlphaZero in Hex | 1 + ...s Improves Robustness of Graph Neural Networks | 1 + ... Kernels under Benign and Adversarial Training | 1 + ... Shape Correspondence via 2D graph convolution | 1 + .../Exact Solutions of a Deep Linear Network | 1 + ...s of deep linear networks with prior knowledge | 1 + ...on for Weakly-Supervised Semantic Segmentation | 1 + ...for Compact Video-and-Language Representations | 1 + .../Expected Improvement for Contextual Bandits | 1 + ...ormer for Dense Prediction without Fine-tuning | 1 + ...nctionals in Reproducing Kernel Hilbert Spaces | 1 + ...g-Term Memory by predicting uncertain outcomes | 1 + .../neurips/Explainability Via Causal Self-Talk | 1 + ...le Reinforcement Learning via Model Transforms | 1 + .../Explaining Preferences with Shapley Values | 1 + data/2022/neurips/Explicable Policy Search | 1 + ...ersarial and Natural Distributional Robustness | 1 + ...rvative Exploitation via Linear Reward Shaping | 1 + ...xploitability Minimization in Games and Beyond | 1 + ...Semantic Relations for Glass Surface Detection | 1 + ...d Cosine Similarity for Attribution Protection | 1 + .../Exploration via Elliptical Episodic Bonuses | 1 + ...g for Information about the Optimal Trajectory | 1 + ...or Reinforcement Learning under Sparse Rewards | 1 + ...loring Example Influence in Continual Learning | 1 + ...ssignment Mechanism in Perceptual Organization | 0 ...Length Generalization in Large Language Models | 1 + ...language models as protein function predictors | 1 + ...tion of AUPRC Optimization with List Stability | 1 + ...ace of Autoencoders with Interventional Assays | 1 + ...ng for Detoxifying Large-Scale Language Models | 1 + ...he Whole Rashomon Set of Sparse Decision Trees | 1 + ... Random Curiosity with General Value Functions | 1 + ...ased Reinforcement Learning via Score Matching | 1 + ...ntial Separations in Symmetric Neural Networks | 1 + ...sfeiler-Lehman Test with Graph Neural Networks | 1 + ...ructures for Fast and Accurate Sparse Training | 1 + ...oise-Adaptive Accelerated Second-Order Methods | 1 + ...echanisms from neural data using low-rank RNNs | 1 + ... with Hadamard Product: a Polynomial Net Study | 1 + ...rk for Fast Training-free Test-time Adaptation | 1 + ...overning Abstractions Behind Integer Sequences | 1 + ...undational Models for Expert Task Applications | 1 + ...d of Words Represented as Non-Linear Functions | 1 + ... Federated Learning Annotated Image Repository | 1 + ...ated Learning in Realistic Healthcare Settings | 1 + ...VR: Neural Volume Rendering for Face Animation | 1 + ...d for Monocular Real-time Human Reconstruction | 1 + .../FP8 Quantization: The Power of the Exponent | 1 + ... Folded Rationalization with a Unified Encoder | 1 + ...tion for Non-Stationary Reinforcement Learning | 1 + ...ionally Robust Policies for Contextual Bandits | 1 + ... Parameter Factorization & Similarity Matching | 0 ...Language Models for Open-Ended Text Generation | 1 + ...es-Optimal Classifiers Under Predictive Parity | 1 + ... Biased Training Data Points Without Refitting | 1 + data/2022/neurips/Fair Rank Aggregation | 1 + .../Fair Ranking with Noisy Protected Attributes | 1 + .../Fair Wrapping for Black-box Predictions | 1 + ...ient Allocations Without Obvious Manipulations | 1 + ...Decision Trees: A Dynamic Programming Approach | 1 + ...ramework with Contrastive Adversarial Learning | 1 + data/2022/neurips/Fairness Reprogramming | 1 + ...rability Subject to Bounded Distribution Shift | 1 + ...rness in Federated Learning via Core-Stability | 1 + ...ut Demographics through Knowledge Distillation | 1 + ...proach for Approximate Nearest Neighbor Search | 1 + ...fore Extrapolation in Causal Effect Estimation | 1 + ...for Packing Proportional Fairness and its Dual | 1 + ...ts via Subsampling and Quasi-Newton Refinement | 1 + ...nt Process Intensity as Function of Covariates | 1 + ...h Bayesian Quadrature via Kernel Recombination | 1 + .../Fast Distance Oracles for Any Symmetric Norm | 1 + .../Fast Instrument Learning with Faster Rates | 1 + ...nt Descent with Normalization and Weight Decay | 1 + ...ural Kernel Embeddings for General Activations | 1 + ...ed Frank-Wolfe Algorithm under Parallelization | 1 + .../Fast Vision Transformers with HiLo Attention | 1 + ...nforcement Learning with Slower Online Network | 1 + .../Faster Linear Algebra for Distance Matrices | 1 + ...orithms for Densest Subgraph and Decomposition | 1 + ...k: Fast and Accurate Interpretable Risk Scores | 1 + data/2022/neurips/Fault-Aware Neural Code Rankers | 1 + .../FeLMi : Few shot Learning with hard Mixup | 1 + ...arized DNNs: Attraction Repulsion and Sparsity | 1 + ...re-Proxy Transformer for Few-Shot Segmentation | 1 + ... Local Updates Lead to Representation Learning | 1 + ...n Approach for Personalised Federated Learning | 1 + ...ted Learning with Rolling Sub-Model Extraction | 1 + ...n Generalization Method for Federated Learning | 1 + ...rained Models: A Contrastive Learning Approach | 1 + ...Audio-Visual Learning of Environment Acoustics | 1 + .../Few-Shot Continual Active Learning by a Robot | 1 + .../Few-Shot Fast-Adaptive Anomaly Detection | 1 + ...etric Learning with Deep Latent Variable Model | 1 + ...is Better and Cheaper than In-Context Learning | 1 + ...eration via Adaptation-Aware Kernel Modulation | 1 + ...on with Hilbert-Schmidt Independence Criterion | 1 + ... Reasoning via Connection Subgraph Pretraining | 1 + ...re Search for Distilling Large Language Models | 1 + ...ep Learning via Feature-wise Linear Modulation | 1 + ...ry Model for Long-term Time Series Forecasting | 1 + ...r Data-Driven Financial Reinforcement Learning | 1 + ...onstrained Markov Game: A Primal-Dual Approach | 1 + ...nvNets Using Counterfactual Simulation Testing | 1 + ...Occurring Physical Backdoors in Image Datasets | 1 + ...ts with Semi-bandit Feedback and Finite Budget | 1 + ...onconvex-Strongly-Concave Minimax Optimization | 1 + ...Finding and Listing Front-door Adjustment Sets | 1 + ...ralization for Modern Meta Learning Algorithms | 1 + ...antically Aligned Vision-Language Pre-Training | 1 + ...fectively by Optimizing Subnetworks Adaptively | 1 + ... using Activation Quantization with Guarantees | 1 + ...greement among humans with diverse preferences | 1 + ...lysis Of Dynamic Regression Parameter Learning | 1 + ...mple Maximum Likelihood Estimation of Location | 1 + ... Difference Learning with Deep Neural Networks | 1 + ...Convergence for Learning in Multi-Player Games | 1 + ...hms for Exponential Family Multi-Armed Bandits | 1 + ...Adaptation via Mutual Information Maximization | 1 + ...enerating Manifold, Graph and Categorical Data | 1 + ...s Better Than Last for Language Data Influence | 1 + ...Min-Max Optimization in Geodesic Metric Spaces | 1 + .../Fixed-Distance Hamiltonian Monte Carlo | 1 + ... a Visual Language Model for Few-Shot Learning | 1 + ...enomenological Nighttime Flare Removal Dataset | 1 + ...ry-Efficient Exact Attention with IO-Awareness | 1 + .../Flexible Diffusion Modeling of Long Videos | 1 + ...ible Neural Image Compression via Code Editing | 1 + ...MM: Flow-based continuous hidden Markov models | 1 + ...lowification: Everything is a normalizing flow | 1 + ...isual navigation using panoramic stereo vision | 1 + data/2022/neurips/Focal Modulation Networks | 1 + ...Markov Decision Processes with Bandit Feedback | 1 + ...sting Future World Events With Neural Networks | 1 + ...orecasting Human Trajectory from Scene History | 1 + ...tency and Coherence of Representation Learning | 1 + ...mulating Robustness Against Unforeseen Attacks | 1 + ... for Hidden Continuous-Time semi-Markov Chains | 1 + ...eriors for Approximate Probabilistic Inference | 1 + ...mer Meets Generalized Fourier Integral Theorem | 1 + ...cal optical encoders for computational imaging | 1 + ...gorithms for Approximating Tyler's M-estimator | 1 + ...omponents for Training GANs under Limited Data | 1 + ...f feed-forward fully connected neural networks | 1 + ...Powerful Defense against Data Poisoning Attack | 1 + ...s to Learning with Stochastic Gradient Descent | 1 + ...tage 3D Object Detection on LiDAR Range Images | 1 + .../2022/neurips/Fully Sparse 3D Object Detection | 1 + ...iable Nonlinear Independent Component Analysis | 1 + .../2022/neurips/Functional Ensemble Distillation | 1 + ... for Better Out-of-distribution Generalization | 1 + ...lternating Least Squares for Tensor Clustering | 1 + data/2022/neurips/Fuzzy Learning Machine | 1 + ...Age-path of Generalized Self-paced Regularizer | 1 + ...ecentralized Multi-Organization Collaborations | 1 + ...ent Learning via Generalizable Logic Synthesis | 1 + ...erative Adversarial Multi-Object Scene Attacks | 1 + ...zed Autoregressive Paraphrase-Identification X | 1 + ...lized Autoregression for Multi-Fidelity Fusion | 1 + ...al Architect for Immersive 3D Scene Generation | 1 + ...synchronous Training for Recommendation Models | 1 + ...ENIE: Higher-Order Denoising Diffusion Solvers | 1 + ...Quality 3D Textured Shapes Learned from Images | 1 + ...te-and-Fire Neuron for Spiking Neural Networks | 1 + ...Localization and Vision-Language Understanding | 1 + ...tudinal Human Behavior Modeling Generalization | 1 + ... based Generative Semantic Segmentation Models | 1 + .../GOOD: A Graph Out-of-Distribution Benchmark | 1 + ...atrix Multiplication for Transformers at Scale | 0 ...etrosynthetic Planning with Goal-driven Policy | 1 + ...ramework for Learning Graph Distance Functions | 1 + ...etworks with Structure-Aware Cooperative Games | 1 + ...Synthesis with Generative Adversarial Networks | 1 + ...rediction-based metric between representations | 1 + data/2022/neurips/Gaussian Copula Embeddings | 1 + ...ing of Generalizable Signed Distance Functions | 1 + ...for Generalizable Out-Of-Domain Text-to-Speech | 0 ...-Propagation-Based Neural Network Verification | 1 + .../Generalised Implicit Neural Representations | 1 + ...tual Information for Discriminative Clustering | 1 + ...Passing Neural Networks on Large Random Graphs | 1 + ...nalysis on Learning with a Concurrent Verifier | 1 + ...mating Causal Effects of Continuous Treatments | 1 + ...ient Methods via Discrete and Continuous Prior | 1 + ...Class via Distributionally Robust Optimization | 1 + ...r Bounds on Deep Learning with Markov Datasets | 1 + .../Generalization Gap in Amortized Inference | 1 + ...AS under Activation and Skip Connection Search | 1 + ...ification with overparameterized linear models | 1 + ... Post-Click Information in Recommender Systems | 1 + data/2022/neurips/Generalized Laplacian Eigenmaps | 1 + ... Adaptation of Generative Adversarial Networks | 1 + ... Gaussian Measures meet Bayesian Deep Learning | 1 + ...Optimization with Decision-theoretic Entropies | 1 + ...jection to be Compatible with Arbitrary Losses | 1 + ...ent Learning with Variational Causal Reasoning | 1 + .../Generating Long Videos of Dynamic Scenes | 1 + ...dels: Towards Zero-Shot Language Understanding | 1 + ...with COmmon Source CoordInated GAN (COSCI-GAN) | 1 + .../Generative Neural Articulated Radiance Fields | 1 + ... Information Decoupling for Image Rain Removal | 1 + ...g with Diffusion, Denoise, and Disentanglement | 1 + ...ional Control of Pre-Trained Generative Models | 1 + ... learning mitigates target-causing confounding | 1 + ...r for physics-informed (and) operator learning | 1 + ...urfaces Learning for Multi-view Reconstruction | 1 + ...ble Geometric Shapes in Deep Image Classifiers | 1 + ... Few-Shot Generalization in Euclidean Geometry | 1 + ...rk for Efficient Graph Representation Learning | 1 + .../Geodesic Self-Attention for 3D Point Clouds | 1 + ...Topology Compression for Graph Neural Networks | 1 + .../Geometric Order Learning for Rank Estimation | 1 + ...e PIFu Representation for Human Reconstruction | 1 + ...ating Sparse Training with Gradient Correction | 1 + ...ance Cheap Operation with Long-Range Attention | 1 + ...ale Kernel Matrix-Vector Multiplication on GPU | 1 + ...ractive Student Programs with Meta-Exploration | 1 + ...Interpretable, Leak-proof Concept-based Models | 1 + ...e and Stability of Stochastic Gradient Descent | 1 + ...nce of Federated Learning for Mixed Regression | 1 + ...gence of IRLS for Non-Smooth Robust Regression | 1 + ...ming Speech Recognition in a Modular Framework | 1 + ...al K-Medoids Clustering of One Million Samples | 1 + ...Convergent Policy Search for Output Estimation | 1 + .../neurips/Globally Gated Deep Linear Networks | 1 + ...g deep learning: How much physics do we need?" | 1 + ...tter Data Permutations than Random Reshuffling | 1 + ...ricted Secant Inequality And Upper Error Bound | 1 + .../Gradient Descent: The Ultimate Optimizer | 1 + ...dient Estimation with Discrete Stein Operators | 1 + ...thods Provably Converge to Non-Robust Networks | 1 + ...networks for square loss and orthogonal inputs | 1 + ...nd Stochastic Nonsmooth Nonconvex Optimization | 1 + ...Assembly and Viral Quasispecies Reconstruction | 1 + ...ng Guarantee and Item Mixture Powered Strategy | 1 + ...ew-shot Learning with Task-specific Structures | 1 + ...g Assisted Multi-Objective Integer Programming | 0 data/2022/neurips/Graph Neural Network Bandits | 1 + .../Graph Neural Networks are Dynamic Programmers | 1 + .../Graph Neural Networks with Adaptive Readouts | 1 + ...ering for Cache-Efficient Near Neighbor Search | 1 + .../Graph Scattering beyond Wavelet Shackles | 1 + ...ed Learning with Accurate Discrepancy Learning | 1 + ...ng and Out-of-Distribution Detection on Graphs | 1 + ...: Quantum Neural Tangent Kernel for Graph Data | 1 + ...omolecular Structures and Interaction Networks | 1 + ...l Vision Transformer for Masked Image Modeling | 1 + ...riddlyJS: A Web IDE for Reinforcement Learning | 1 + ... Learning to Win the Game under Human Commands | 1 + .../neurips/Grounded Video Situation Recognition | 1 + ...ncertainty for Unsupervised Environment Design | 1 + ...tocratic Fairness in Linear Contextual Bandits | 1 + ... Framework for Continuous Categories Discovery | 1 + ...tum for Learning Particle-based Fluid Dynamics | 1 + ...dinal Dataset of Commercial ML API Predictions | 1 + ...ce Reconstruction Using High-Frequency Details | 1 + ...or Modeling Surfaces with Arbitrary Topologies | 1 + ...for 3D Point Clouds by Learning Hyper Surfaces | 1 + ...nditioned Human Motion Generation in 3D Scenes | 1 + ...rchitecture for Accelerated MRI Reconstruction | 1 + ...for Long-Horizon Prediction of Event Sequences | 1 + ... and motion disentanglement in image sequences | 1 + .../Hand-Object Interaction Image Generation | 1 + ...munication in Physical and Social Environments | 1 + .../Handcrafted Backdoors in Deep Neural Networks | 1 + ...ntations for Objects with Strong Spurious Cues | 1 + ...Markov Decision Processes: Theory and Practice | 1 + ... Learning for Two-Hidden-Layer Neural Networks | 1 + ...strategies of deep neural networks with humans | 1 + ...istribution Matching for Human Pose Estimation | 1 + ...Augmentation in Probabilistic Graphical Models | 1 + ...rogeneous Skill Learning for Multi-agent Tasks | 0 ...D Learns Parities Near the Computational Limit | 1 + .../Hiding Images in Deep Probabilistic Models | 1 + ...upervised Representations for Speech Synthesis | 1 + ...ive Graph Clustering in Poly-Logarithmic Depth | 1 + ...Communication-efficient Collaborative Learning | 1 + ... Graph Transformer with Adaptive Node Sampling | 1 + ...e Layer for Partially Monotone Neural Networks | 0 ...lization for Robust Monocular Depth Estimation | 1 + ...al classification at multiple operating points | 1 + ...raph Neural Networks with Tensor Decomposition | 1 + ...ssian Processes under Monotonicity Constraints | 1 + ... One Gradient Step Improves the Representation | 1 + ...r SGD: Effective dynamics and critical scaling | 1 + ...Distillation for Cross-Dimensionality Networks | 1 + ...act Gradients Through Finite Size Oscillations | 1 + data/2022/neurips/Homomorphic Matrix Completion | 1 + ...lization in Competitive Reinforcement Learning | 1 + ...Interactions with Recursive Gated Convolutions | 1 + ...tasets via Capacity-Aware Neuron Steganography | 1 + ...oretical Understandings of Masked Autoencoders | 1 + ...re K-hop Message Passing Graph Neural Networks | 1 + ...s the Robustness of Stochastic Neural Networks | 1 + ...Video Representations Based on Synthetic Data? | 1 + ... Model Human Real-time and Life-long Learning? | 1 + ...eel? Estimating Wellbeing From Video Scenarios | 1 + ... On the Incentives of Users of Learning Agents | 1 + ...earn: Instructions, descriptions, and autonomy | 1 + ...sfer Learning from A Hub of Pre-trained Models | 1 + .../Human-AI Collaborative Bayesian Optimisation | 1 + .../Human-AI Shared Control via Policy Dissection | 1 + ...s Collaborating Agents for Symmetrical Walking | 1 + ... Detector to Model the Manual Labeling Process | 1 + ...ng in Visual and Other Sensory Neuroprostheses | 1 + ...Models: Sampling Unseen Neural Network Weights | 1 + ...Adaptation for Generative Adversarial Networks | 1 + ...opic Taxonomy Mining with Hyperbolic Embedding | 1 + ...erTree Proof Search for Neural Theorem Proving | 1 + ...nference for Structured Multi-Label Prediction | 1 + ... Estimation and Infinite Sampling on Manifolds | 1 + ...nalysis and a Scalable Hyper-Ensemble Solution | 1 + ...g for Differentially Private Linear Regression | 1 + ...t Attention for Zero-Shot Image Classification | 1 + ...2Q: A Fully Decentralized Q-Learning Algorithm | 0 ...KEA-Manual: Seeing Shape Assembly Step by Step | 1 + ... Maximization Loss for Spiking Neural Networks | 1 + ... learning of ergodic Markov decision processes | 1 + ...mplicit Neural Representation for Audio Scenes | 0 ...iple experts in Inverse Reinforcement Learning | 1 + ...enerative models without auxiliary information | 1 + ...ent: A bridge to Gaussian Differential Privacy | 1 + ...ently learn low-degree plus sparse polynomials | 1 + ...ons are the Answer, Then What is the Question? | 1 + ...e Trouble: Revisiting Neural-Collapse Geometry | 1 + ...mitating Past Successes can be Very Suboptimal | 1 + ...minacy for Explanations of Automated Decisions | 1 + ...rized Models: On Equivalence to Mirror Descent | 1 + ... Neural Representations with Levels-of-Experts | 1 + ...ct Risk Trajectories of SGD in High Dimensions | 1 + ...Implicit Warping for Animation with Image Sets | 1 + ...Improved Algorithms for Neural Active Learning | 1 + ...ty for Representing Piecewise Linear Functions | 1 + ... Reduction and its Application to Optimization | 1 + .../Improved Coresets for Euclidean k-Means | 1 + ...l Private Linear Operators on Adaptive Streams | 1 + ...ed Feature Distillation via Projector Ensemble | 1 + ...-Tuning by Better Leveraging Pre-Training Data | 1 + ...vex Regularizers with Global Optima Guarantees | 1 + ...r Bandits and Horizon-Free Linear Mixture MDPs | 1 + ...proved Utility Analysis of Private CountSketch | 1 + ...ved techniques for deterministic l2 robustness | 1 + ... Synthesis with A Geometry-aware Discriminator | 1 + ...criminating Unlabeled Samples with Super-Class | 0 ...ia Statistical Learning with Logical Reasoning | 1 + ...or Inverse Problems using Manifold Constraints | 1 + .../Improving GANs with A Dynamic Discriminator | 1 + ...works via Adversarial Learning in Latent Space | 1 + ...trinsic Exploration with Language Abstractions | 1 + ...ns with Nesterov's Accelerated Gradient Method | 1 + ...by Adversarial Training with Structured Priors | 1 + ...cy Learning via Language Dynamics Distillation | 1 + ...ng by Characterizing Idealized Representations | 1 + ...earning via Adaptive Vicinal Risk Minimization | 1 + ...ansformer with an Admixture of Attention Heads | 1 + ...encoders with Density Gap-based Regularization | 1 + ...earning using Generalized Similarity Functions | 1 + ...ary Scalarization for Deep Multi-Task Learning | 1 + ...Histogram Leakage in Ensemble Private Learning | 1 + ...orks Invariant and How Should We Measure This? | 1 + ...r: Robust Prediction with Causal User Modeling | 1 + ...Incentivizing Combinatorial Bandit Exploration | 1 + data/2022/neurips/Inception Transformer | 1 + ...o Contrastive Loss for Collaborative Filtering | 1 + ...nfidence in Adversarial Robustness Evaluations | 1 + ...tive Bayesian Optimization in Nested Subspaces | 1 + ...ement Learning under Mixed and Delayed Rewards | 1 + ...e Testing for Bounded Degree Bayesian Networks | 1 + ...asurement Error and Linear Non-Gaussian Models | 1 + ...Improving Optimization of Adversarial Examples | 1 + ...ous Design-and-Play Ensures Global Convergence | 1 + ... Classifier at the End of Deep Neural Network? | 1 + ...ve Logical Query Answering in Knowledge Graphs | 1 + .../Inference and Sampling for Archimax Copulas | 1 + ...commendation Networks: A Data-Centric Approach | 1 + ...lity Coregionalization for Physical Simulation | 1 + ... Behavior in Multiagent Reinforcement Learning | 1 + ...gression: relevancy, efficiency and optimality | 1 + ...ompression with Variational Energy-based Model | 1 + ...retic Safe Exploration with Gaussian Processes | 1 + ...ble Reinforcement Learning in Natural Language | 1 + ...al Networks for Predicting Material Properties | 1 + ...rformant Insertion-based Text Generation Model | 1 + ...roposal for Online Video Instance Segmentation | 1 + ... into Pre-training via Simpler Synthetic Tasks | 1 + ...ima in GAN Training with Kernel Discriminators | 1 + ...timation for Gradient-Boosted Regression Trees | 1 + ...on in Linear MDPs via Online Experiment Design | 1 + ...e-based Learning for Knowledge Base Completion | 1 + ...-optimal PAC Algorithms for Contextual Bandits | 1 + .../Integral Probability Metrics PAC-Bayes Bounds | 1 + .../Interaction Modeling with Multiplex Attention | 1 + ...ounded Learning with Action-Inclusive Feedback | 1 + ...Transformer for Few-Shot Semantic Segmentation | 1 + ...olation and Regularization for Causal Learning | 1 + ...rspective from Influence-Directed Explanations | 1 + ...Experimental Design for Causal Models at Scale | 1 + ...gent speech permits zero-shot task acquisition | 1 + ...ensionality estimation using Normalizing Flows | 0 ...tage approach for Inference in Neural Networks | 1 + .../Invariance Learning based on Label Hierarchy | 1 + ...rks with Differentiable Laplace Approximations | 1 + ...riance-Aware Randomized Smoothing Certificates | 1 + ... Representations for Anti-Causal Domain Shifts | 1 + ...re Interactions using Graph Network Simulators | 1 + ...erg Games: the Blessing of Bounded Rationality | 1 + ...tible Monotone Operators for Normalizing Flows | 1 + ... Arithmetic Enough for Deep Learning Training? | 1 + .../Is Out-of-Distribution Detection Learnable? | 1 + .../Is Sortition Both Representative and Fair? | 1 + .../neurips/Is a Modular Architecture Enough? | 1 + ...hmark for noisy and ambiguous label estimation | 1 + ...nd Query Efficient Model Agnostic Explanations | 0 ...oncontrollable Visual Dynamics in World Models | 1 + ... 3D Adversarial Examples in the Physical World | 1 + ...n Generalization with Logarithmic Environments | 1 + .../2022/neurips/Iterative Scene Graph Generation | 1 + ...rative Structural Inference of Directed Graphs | 1 + ...n Joint Architecture And Hyperparameter Search | 1 + ...g Predictive Uncertainty Under Covariate Shift | 1 + ...h For Maximally-Informed Bayesian Optimization | 1 + ...arch for Multi-Objective Bayesian Optimization | 1 + ... 2D-3D Weakly Supervised Semantic Segmentation | 1 + ...apturing High-order Statistics in Transformers | 1 + ...sferable Visual Models with External Knowledge | 1 + ...tonomous Driving in Various Weather Conditions | 1 + ... Positional Embedding for Length Extrapolation | 1 + .../neurips/KSD Aggregated Goodness-of-fit Test | 1 + ...k! Wasserstein GANs are not Optimal Transport? | 1 + .../Kernel Interpolation with Sparse Grids | 1 + ...orks: A Unifying Framework for Memory Modeling | 1 + .../Kernel Multimodal Continuous Attention | 1 + ...rnel similarity matching with Hebbian networks | 1 + ...pplications in Heterogeneous Domain Adaptation | 1 + ...ructure Augmentation for Graph Neural Networks | 1 + ...Knowledge Distillation from A Stronger Teacher | 1 + ...stillation: Bad Models Can Be Good Role Models | 1 + .../Knowledge-Aware Bayesian Deep Topic Model | 1 + ...for training next generation image-text models | 1 + ...Text from Gradients with Language Model Priors | 1 + ...ptimization for Offline Reinforcement Learning | 1 + ...om Sparse Image Ensemble via 3D Part Discovery | 1 + ...for Individual-Level Unbiased Learning to Rank | 1 + ...Cooperative Multi-Agent Reinforcement Learning | 1 + ...sodic Count for Task-Specific Intrinsic Reward | 1 + ... Denoising Network for Video-Language Modeling | 1 + ...Tuning for Non-language Machine Learning Tasks | 1 + ...Point Diffusion Models for 3D Shape Generation | 1 + ...Industrial Physical Simulation benchmark suite | 1 + ...Interpretable Skill Abstractions from Language | 1 + ...ptation for Label-Efficient OOD Generalization | 0 ... Training on Improving l2 Certified Robustness | 1 + ...rameter and Memory Efficient Transfer Learning | 1 + ...able Thresholding Neurons and Moderate Dropout | 1 + ... Novel Perspective to Study Robust Overfitting | 1 + ...lti-Label Learning with Single Positive Labels | 1 + ...ation for Semi-Supervised Graph Classification | 1 + ...oders for Learning Deep Latent Variable Models | 1 + ...ial Relation Reasoning for 3D Object Grounding | 1 + ...rs are Strong Few-Shot Video-Language Learners | 1 + ...coders for Learning Stochastic Representations | 1 + .../Large Language Models are Zero-Shot Reasoners | 1 + ...fferentiable Causal Discovery of Factor Graphs | 1 + ...rge-Scale Retrieval for Reinforcement Learning | 1 + ...dictions: Training Faster R-CNN in 4.2 Minutes | 1 + ...Partial AUC in a Range of False Positive Rates | 1 + ...tive Structure-aware Generative Language Model | 1 + ...t Method for Monotone Variational Inequalities | 1 + .../Latency-aware Spatial-wise Dynamic Networks | 1 + ...usal Structure Discovery with Rank Constraints | 1 + .../Latent Planning via Expansive Tree Search | 1 + ...ces of a Generic Framework for Sparse Training | 1 + ... MAP Inference for Determinantal Point Process | 1 + ... Thought Chains for Science Question Answering | 1 + ...nforcement Learning in Markov Matching Markets | 1 + ...itation learning with task-relevant embeddings | 1 + ...variant and Equivariant Convolutional Networks | 1 + ...arning (Very) Simple Generative Models Is Hard | 1 + ...ning Active Camera for Multi-Object Navigation | 1 + ... Dynamics with Lagrangian Graph Neural Network | 1 + ...Using Scene Graphs for Audio Source Separation | 1 + ...g Best Combination for Efficient N: M Sparsity | 1 + ...te Graphs: Heavy Tails and Multiple Components | 1 + ...r Out-of-Distribution Generalization on Graphs | 1 + ...arning Chaotic Dynamics in Dissipative Systems | 0 ...ncept Credible Models for Mitigating Shortcuts | 1 + ... Functions Progressively from Raw Point Clouds | 1 + ...Contrastive Embedding in Low-Dimensional Space | 1 + ...ning Debiased Classifier with Biased Committee | 1 + .../Learning Deep Input-Output Stable Dynamics | 1 + ...ple Views for Low-shot Category Generalization | 1 + ... and Representative Modes for Image Captioning | 1 + ...etwork Load Balancing as Markov Potential Game | 1 + ...Networks in the Presence of Arbitrary Outliers | 1 + ...egression in Reproducing Kernel Hilbert Spaces | 1 + ...formers via Fine-Grained Manifold Distillation | 1 + ...Networks with Generalized Fenchel-Young Losses | 1 + ... for Tabular Data via Neighborhood Propagation | 0 ...ant Segmentation with Instance-Unique Querying | 1 + ...ations with Mixture of Expert Neural Processes | 1 + ...es in Neural Stochastic Differential Equations | 1 + ...Models in a Handful of Reward-Free Deployments | 1 + ...le Routing Problems via Knowledge Distillation | 1 + ...sed Feature Representation for 3D Point Clouds | 1 + ...r Relational Stochastic Shortest Path Problems | 1 + ...ck-tracing for Object Tracking in Event Clouds | 1 + ...ised Clustering Approach Using Adaptive Fusion | 1 + ...tless Multi-Action Bandits via Index Awareness | 1 + ...ace Conditions in Domain Decomposition Solvers | 1 + ...tations for Out-of-Distribution Generalization | 1 + ...nd Representations for Time Series Forecasting | 1 + ...Term Crop Management Strategies with CyclesGym | 1 + ...ions with Conditional Variational Autoencoders | 1 + ...ed Multinomial Logits with Provable Guarantees | 1 + ...ng Modular Simulations for Homogeneous Systems | 1 + ...h Spectral Attention for Robust Shape Matching | 1 + ...m Graph and Provable Auction-Fitted Q-learning | 1 + data/2022/neurips/Learning Neural Acoustic Fields | 1 + ... Set Functions Under the Optimal Subset Oracle | 1 + ...ing Optical Flow from Continuous Spike Streams | 1 + ... Flows for Non-Equilibrium Importance Sampling | 1 + .../2022/neurips/Learning Options via Compression | 1 + .../Learning Partial Equivariances From Data | 1 + ...mics with Subequivariant Graph Neural Networks | 1 + ...hysics Constrained Dynamics Using Autoencoders | 1 + ...ng Predictions for Algorithms with Predictions | 1 + ...dels from Generator Latent Spaces with Hat EBM | 1 + ...nce Environment to Enhance Prediction Accuracy | 1 + ...avioral Metric for Deep Reinforcement Learning | 1 + ...ust Dynamics through Variational Sparse Gating | 1 + ...for Abstract Reasoning via Internal Inferences | 1 + ...sual Representations from Audible Interactions | 1 + ...erarchical Representation Learning by Chunking | 1 + ... Out-of-Distribution Molecular Representations | 1 + ...erpoint Graph Cut for 3D Instance Segmentation | 1 + .../neurips/Learning Symmetric Rules with SATNet | 1 + ...istic Models from Inconsistent Local Estimates | 1 + ...ction Approximation and Correlated Equilibrium | 0 ...presentations by Recovering Tokens in 3D Space | 1 + ...ory-Efficient Video Class-Incremental Learning | 1 + ...endent Random Variables with Unbounded Support | 1 + ...of deep linear networks with multiple pathways | 1 + ...ual Linear Bandits Without Sharing the Context | 0 ...th Composition and Locality at Multiple Scales | 1 + ...f-Training Framework for Semantic Segmentation | 1 + ...Label Proportions by Learning with Label Noise | 1 + ...arning from Stochastically Revealed Preference | 1 + .../Learning from a Sample in Online Algorithms | 1 + ...rning in Congestion Games with Bandit Feedback | 1 + ...s, without Computationally Intractable Oracles | 1 + ...ical systems with latent Gaussian process ODEs | 1 + ...ble natural features from retina using a U-net | 1 + ...bitrary Graph Topologies via Predictive Coding | 1 + ...nline Learning with Stochastic Feedback Graphs | 1 + ...ngle-index models with shallow neural networks | 1 + ...res can lead to overfitting in neural networks | 1 + ...ge Networked Systems Obeying Conservation Laws | 1 + ...erential Equations via Latent Global Evolution | 1 + ...-based Reinforcement Learning Attack Framework | 1 + .../neurips/Learning to Branch with Tree MDPs | 1 + ...igating Repetitions for Neural Text Generation | 1 + ...in Branch and Bound with Graph Neural Networks | 1 + ...ter Networks with Neural Algorithmic Reasoning | 1 + ... Policy Optimization with Virtual Trust Region | 1 + .../Learning to Discover and Detect Objects | 1 + ...Adversarial Approach to Training Sequence VAEs | 1 + ...ing to Follow Instructions in Text-Based Games | 1 + ...enerate Inversion-Resistant Model Explanations | 1 + ...to Mitigate AI Collusion on Economic Platforms | 1 + ...g to Navigate Wikipedia by Taking Random Walks | 1 + ...ptimal Transport for Imbalanced Classification | 1 + ...neralization, Unseen Data and Boolean Measures | 1 + ...Spatiotemporal Graphs with Sparse Observations | 1 + ...-shot Reasoning over Temporal Knowledge Graphs | 1 + ...ld: Optimizing Model Explanations for Teaching | 1 + ...n Networked Multi-Agent Reinforcement Learning | 0 ...ution and pooling operations in kernel methods | 1 + data/2022/neurips/Learning with little mixing | 1 + ...for Online Linear and Semidefinite Programming | 1 + ... Environments Using GNNs and Temporal Encoding | 1 + ...N-based best-first search for Sokoban Planning | 1 + .../Less-forgetting Multi-lingual Fine-tuning | 1 + ... Cloud Cross-Modal Training for Shape Analysis | 1 + .../Lethal Dose Conjecture on Data Poisoning | 1 + ...t Offline Reinforcement Learning in Healthcare | 1 + ...yer Dependency for Post -Training Quantization | 1 + ...ptive Bidding in Repeated First-Price Auctions | 1 + .../LieGG: Studying Learned Lie Group Generators | 1 + ...earning Cumulatively Online without Forgetting | 1 + ...ting Weak Supervision To Structured Prediction | 1 + ...is of Thompson Sampling for Contextual Bandits | 1 + .../Linear Label Ranking with Bounded Noise | 1 + data/2022/neurips/Linear tree shap | 1 + .../Lipschitz Bandits with Batched Feedback | 1 + .../neurips/List-Decodable Sparse Mean Estimation | 1 + ...n Estimation via Difference-of-Pairs Filtering | 1 + ...c Interpretability for Audio Networks with NMF | 1 + ...hitecture Search for Efficient Language Models | 1 + ... Stationary Distribution Correction Estimation | 1 + ...mization via maximizing probability of descent | 1 + ...ility of Deep ReLU Neural Networks: the Theory | 1 + ...e Bayesian Optimization over Structured Inputs | 1 + ...bspace Optimization via Strict Complementarity | 1 + ... in Contextual Bandits with Continuous Actions | 1 + ... Longitudinally-consistent Neuroimage Analysis | 1 + ...l-Global MCMC kernels: the best of both worlds | 1 + ... Auto-Regressive Modeling for Image Generation | 1 + ...cating and Editing Factual Associations in GPT | 1 + ...l Noise Distributions for Differential Privacy | 1 + ...e Gaussian Processes Using Binary Tree Kernels | 1 + .../neurips/Log-Polar Space Convolution Layers | 1 + ...Logical Reasoning via Adversarial Pre-training | 1 + ...equivalents of Probabilistic Boolean Operators | 1 + data/2022/neurips/Logical Credal Networks | 1 + data/2022/neurips/Long Range Graph Benchmark | 1 + ... with Multimodal Temporal Contrastive Learning | 1 + ...Knowledge Distillation for 3D Visual Grounding | 1 + .../Look More but Care Less in Video Recognition | 1 + ...eneralization in visual Reinforcement Learning | 1 + ...t Multilingual and Multitask Speech Processing | 1 + ...tangled models at combinatorial generalisation | 1 + ...Initializations with Sparse Trainable Networks | 1 + ...ular Reinforcement Learning via Muscle Synergy | 1 + ...sport: Approximation, Statistics and Debiasing | 1 + ...ral networks via matrix differential equations | 1 + ...ibuted Learning with Communication Compression | 1 + ...Preconditioned Lasso via Robust Sparse Designs | 1 + .../Luckiness in Multiscale Online Learning | 1 + .../M2N: Mesh Movement Networks for PDE Solvers | 1 + ...Musical Score Provided Mandarin Singing Corpus | 1 + ...ster Forest Training Using Multi-Armed Bandits | 1 + ...al Networks for Fast and Accurate Force Fields | 1 + ...ual Knowledge for Unpaired Image-text Matching | 1 + ...earning in Distributed Target Coverage Control | 1 + .../MAgNet: Mesh Agnostic Neural PDE Solver | 1 + ... A Manifold Attention Network for EEG Decoding | 1 + .../MBW: Multi-view Bootstrapping in the Wild | 1 + ...works with Multiple Specialized Discriminators | 1 + ...: Masked Convolution Meets Masked Autoencoders | 1 + ... for Prediction, Generation, and Interpolation | 1 + ...ime Robustness via Adaptation and Augmentation | 1 + ... Targeted Sentiment on COVID-19 Related Tweets | 1 + ...odel Extraction Crossover Membership Inference | 1 + ...ale Graph Neural Networks with Implicit Layers | 1 + ... for Multi-Object Multi-Actor Activity Parsing | 1 + ...stness Evaluation with Model Reweighing Attack | 1 + ...ised Movable Object Segmentation and Detection | 1 + ...it String Dataset for Handwriting Verification | 1 + ...tructure Across Multiple Levels of Abstraction | 1 + ...or Real-World Multi-View Object Classification | 1 + ...n Stronger: A Sparsified Perturbation Approach | 1 + ...and Efficient Single-Step Adversarial Training | 1 + ...ro-Shot Learning for Novel Attribute Synthesis | 1 + ...trategies Feasible with Neural Tangent Kernels | 1 + ...lack-box Explanations Using Dependence Measure | 1 + ...timal-Transport Flows for Trajectory Inference | 1 + ...arning with Class-Level Overfitting Mitigation | 1 + ...Variational Inference with Markovian Gradients | 1 + .../neurips/Markovian Interference in Experiments | 1 + ...: Backdoor Attacks with Arbitrary Target Class | 1 + ...Matching Transformer for Few-Shot Segmentation | 1 + ...tent Reconstruction for Reinforcement Learning | 1 + ... via Reinforced Visual Representation Learning | 1 + ...ng Spurious Correlations by Forcing to Explore | 1 + ...Masked Autoencoders As Spatiotemporal Learners | 1 + data/2022/neurips/Masked Autoencoders that Listen | 1 + ...for Scalable and Generalizable Decision Making | 1 + ...etworks are Data-Efficient Generation Learners | 1 + ...d Prediction: A Parameter Identifiability View | 1 + .../Matching in Multi-arm Bandit with Collision | 1 + ...Zero-Sum Games: Conservation Laws & Recurrence | 1 + .../neurips/Matryoshka Representation Learning | 1 + ...orst-Case Robustness to Model Misspecification | 1 + ... under Market Shrinkage and Market Uncertainty | 1 + ... in Multi-armed Bandits with Graph Information | 1 + ...ass Separation as Inductive Bias in One Matrix | 1 + ...Retrieval: Late and Early Interaction Networks | 1 + ...Training of Implicit Nonlinear Diffusion Model | 1 + ...tinal ganglion cells with deep denoiser priors | 1 + ...forcement Learning with Finite-Time Guarantees | 1 + ...ensional Binary Markov Gaussian Mixture Models | 1 + ...th User-level Privacy under Data Heterogeneity | 1 + ...s of Information Reflect Memorization Patterns | 1 + ...on Defenses in Collaborative Inference Systems | 1 + ...te Regression in Structured Prediction for NLP | 1 + ...the Training Dynamics of Large Language Models | 1 + ...al Networks with Minimum Over-parameterization | 1 + ...Efficient Continual Learning with Transformers | 1 + .../Memory safe computations with XLA compiler | 1 + .../Merging Models with Fisher-Weighted Averaging | 1 + .../Mesoscopic modeling of hidden spiking neurons | 1 + ...Training Tasks - a Density Estimation Approach | 1 + ...Meta-Dataset for Few-Shot Image Classification | 1 + ...ving Parametric Partial Differential Equations | 1 + ...emantics of Short Texts in Neural Topic Models | 0 ...t by Meta-Distillation from Mixture-of-Experts | 1 + ...ning Dynamics Forecasting Using Task Inference | 1 + ...a-Learning with Self-Improving Momentum Target | 1 + ...mativeness Dilemma in Open-set Active Learning | 1 + ...orcement Learning with Self-Modifying Networks | 1 + ...ng for Preference-based Reinforcement Learning | 1 + ...ng within randomly initialized neural networks | 1 + ...sional Confounder for Self-Supervised Learning | 1 + ...in Adaptation for Medical Image Classification | 1 + ...f Correlation Exploring in Similarity Learning | 1 + ...h Modeling for Graph Variational Auto-Encoders | 1 + ... Q-Learning for Offline Reinforcement Learning | 1 + ...ntation Framework without Video-based Training | 1 + ...structing complex images from brain activities | 1 + ...ulti-modal Contrastive Representation Learning | 1 + ... Embodied Agents with Internet-Scale Knowledge | 1 + ...Cooperative Multi-Agent Reinforcement Learning | 1 + ...ithms for Fixed-Budget Best Arm Identification | 1 + ...dget Best Arm Identification in Linear Bandits | 1 + ...nline Imitation Learning via Replay Estimation | 1 + .../neurips/Minimax Regret for Cascading Bandits | 1 + ...ent RL in Markov Games With a Generative Model | 1 + ...ulti-Label Samples from Single Positive Labels | 1 + ...A Simple Baseline for Incremental Segmentation | 1 + ...ized Margin and Can Be Implemented Efficiently | 1 + ...re Spaces, with application to Sinkhorn and EM | 1 + ...t Model-Policy Optimization for Model-Based RL | 1 + ...ing Data with Continuous Additive Noise Models | 1 + ...ierarchical Models and Hamiltonian Monte Carlo | 1 + ...ecified Phase Retrieval with Generative Priors | 1 + ...ogy with Data Mixing for Domain Generalization | 1 + .../Mixture-of-Experts with Expert Choice Routing | 1 + ...ti-Task Dataset for Simulated Humanoid Control | 1 + ...: Model-based Counterfactual Data Augmentation | 1 + ... Object Detection with Ground Depth Estimation | 1 + ...zed Vectors for High-Fidelity Image Generation | 1 + ...del Preserving Compression for Neural Networks | 1 + ...f Diverse Populations of Neural Network Models | 1 + ...del-Based Imitation Learning for Urban Driving | 1 + ...rning with Pessimism-Modulated Dynamics Belief | 1 + data/2022/neurips/Model-Based Opponent Modeling | 1 + ...inforcement Learning with Bayesian Exploration | 1 + ...g: Structural Conditions and Sample Complexity | 1 + ...trained Proximal Policy Optimization Algorithm | 1 + ...rough Resource-Rational Reinforcement Learning | 1 + ...Directed Graphs via Binary Code Box Embeddings | 1 + .../Modeling the Machine Learning Multiverse | 1 + ... Fourier Lens on Distribution Shift Robustness | 1 + ...kdoor Defender for Pre-trained Language Models | 1 + ...dular Flows: Differential Molecular Generation | 1 + ...dule-Aware Optimization for Auxiliary Learning | 1 + ...on by Principal Subgraph Mining and Assembling | 1 + ...ributionally Robust Tree Structured Prediction | 1 + ...ion Shifts in Data-Free Knowledge Distillation | 1 + ...omentum Aggregation for Private Non-convex ERM | 1 + ...ues for Neural Implicit Surface Reconstruction | 1 + ...ocular Dynamic View Synthesis: A Reality Check | 1 + ...cement Learning from Suboptimal Demonstrations | 1 + ... Carlo Tree Descent for Black-Box Optimization | 1 + ...ion for High Dimensional Bayesian Optimization | 1 + ... Injecting Morphology in Tensorized Embeddings | 1 + ...ns Can Win the Lottery Without Excessive Depth | 1 + ...ion Localization and Local Movement Refinement | 1 + ...zation with Application to Wind Energy Systems | 1 + ...former for 3D Object Detection on Point Clouds | 1 + ...n for Decentralized Optimization and Averaging | 1 + ...cement Learning is a Sequence Modeling Problem | 1 + .../neurips/Multi-Class $H$-Consistency Bounds | 1 + .../Multi-Fidelity Best-Arm Identification | 0 .../2022/neurips/Multi-Game Decision Transformers | 1 + ...ralized Medical Visual Representation Learning | 1 + ...diction and Out-of-Distribution Generalization | 1 + ...ivil Rights Lawsuits at Multiple Granularities | 1 + ...timodal Pre-training for Cross-modal Retrieval | 0 ... Deep Learning with Adaptive Reference Vectors | 1 + ...i-Sample Training for Neural Image Compression | 1 + ...le Adaptive Network for Single Image Denoising | 1 + .../Multi-agent Dynamic Algorithm Configuration | 1 + ...Greedy Deployment and Consensus Seeking Agents | 1 + ...plications in Multi-task Deep AUC Maximization | 1 + ...timator for Coupled Compositional Optimization | 1 + ... of Transformers for Robust Action Recognition | 1 + ...delity Monte Carlo: a pseudo-marginal approach | 1 + ...te Evolution Under Random Convolutional Design | 1 + ...r Weakly-Supervised Audio-Visual Video Parsing | 1 + ...ta Generation with Correlated Property Control | 1 + ...ew Subspace Clustering on Topological Manifold | 0 ...el Classification against Adversarial Examples | 1 + ...g for 3D environments with articulated objects | 1 + ...ltiagent Q-learning with Sub-Team Coordination | 1 + ...k: Universal Rates and Partial Concept Classes | 1 + ...Comment Detection at Scale for Indic Languages | 1 + ...h LIMoE: the Language-Image Mixture of Experts | 1 + ...A Neural Model for Bilingual Cognitive Reserve | 1 + ...with Temporal Polynomial Graph Neural Networks | 1 + ... Body Reconstruction from Uncalibrated Cameras | 1 + ... Coarse-Grained Attention for Music Generation | 1 + ...nified Metric for Multimodal Generative Models | 1 + ...idge trajectory optimization and deep learning | 1 + ...ask Learning with Model-Accelerator Co-design" | 1 + ...ng Neural Architecture Search on Diverse Tasks | 1 + ... Benchmarking Graph Neural Architecture Search | 1 + ...ro: Accelerating Research on Zero Cost Proxies | 1 + ...rior for Effective Unsupervised Shape Matching | 1 + ...linear Manifold Decoders for Operator Learning | 1 + ...t-time Adaptation Against Temporal Correlation | 1 + .../NS3: Neuro-symbolic Semantic Code Search | 1 + ...bilistic Framework for Satisfiability Problems | 1 + ...ssive Generation for Infinite Visual Synthesis | 1 + ...owards Boosting Black-box Unrestricted Attacks | 1 + ...ables fast sampling in spiking neural networks | 1 + ...iational information bottleneck representation | 1 + ...ematical Proof Generation with Language Models | 1 + ... Pseudo-Task Simulation for Continual Learning | 1 + ...: Neural Motion Fields for Kinematic Animation | 1 + ... Reinforcement Learning for Deterministic MDPs | 1 + ... Kronecker-Structured Random Tensor Embeddings | 1 + ...Near-Optimal Collaborative Learning in Bandits | 1 + ...ar-Optimal Correlation Clustering with Privacy | 1 + ...cement Learning in Non-Stationary Environments | 1 + ...Multi-Agent Learning for Safe Coverage Control | 1 + ...ret Learning Dynamics for General Convex Games | 1 + ...ar-Optimal Private and Scalable $k$-Clustering | 0 ...loration for Tabular Markov Decision Processes | 1 + ... Bounds for Multi-batch Reinforcement Learning | 1 + ...r Adversarial MDP with Delayed Bandit Feedback | 1 + ... Sample Complexity Bounds for Constrained MDPs | 1 + ...ontextual Bandits with Adversarial Corruptions | 1 + ...ithms for Online Learning with Feedback Graphs | 1 + ...ght Bounds for Testing Histogram Distributions | 0 ...d Benchmark for Offline Reinforcement Learning | 1 + ...ameter-Agnostic Nonconvex Minimax Optimization | 1 + ... localisation under local differential privacy | 1 + ... Adaptive Overfitting for Neural Shape Editing | 1 + ...ral Geometry and Physics from Monocular Videos | 1 + ...ur2SP: Neural Two-Stage Stochastic Programming | 1 + ...Enabling Parametric Photonic Device Simulation | 1 + data/2022/neurips/Neural Abstractions | 1 + ...al Approximation of Graph Topological Features | 1 + data/2022/neurips/Neural Attentive Circuits | 1 + .../Neural Basis Models for Interpretability | 1 + ...cuit Architectural Priors for Embodied Control | 1 + ...eometric Analysis over the Riemannian Manifold | 1 + ...nservation Laws: A Divergence-Free Perspective | 1 + ... Neural Nets Through Continuous Learning Rules | 1 + ...pplications to Differentiable Subset Selection | 1 + ...wn Nonlinear Systems with Stability Guarantees | 1 + ...n of Matching Fields for Visual Correspondence | 1 + ...al Network Architecture Beyond Width and Depth | 1 + ...d Stable Payoff Allocations Among Team Members | 1 + ...ing with Discrete Functions in High Dimensions | 1 + data/2022/neurips/Neural Shape Deformation Priors | 1 + ...ctive on Heterophily and Oversmoothing in GNNs | 1 + data/2022/neurips/Neural Stochastic Control | 1 + ...Learning of Continuous Spatiotemporal Dynamics | 1 + ... of Dynamic Scenes with Monocular RGB-D Camera | 1 + ...ion Learning on Continuous-Time Dynamic Graphs | 1 + ...al Topological Ordering for Computation Graphs | 1 + .../neurips/Neural Transmitted Radiance Fields | 1 + ...ntangled Framework for Complex Query Answering | 1 + ...sed Scheduling Method for High-level Synthesis | 1 + ...Steady Response Leads to Better Generalization | 1 + ... for Sequence Data with Relational Constraints | 1 + ...Methods: Staying Intrinsic, Complete and Sound | 1 + ...imation and a Generalized Fingerprinting Lemma | 1 + ...h Models of the Entorhinal-Hippocampal Circuit | 1 + ...es and adaptivity via learning rate separation | 1 + ...ent learning one step closer to the real world | 1 + ...e Learning Transformer for Node Classification | 1 + ...Enhancing Noise Robustness by Gradient Scaling | 1 + ...level Games with Critical Point Selection Maps | 1 + data/2022/neurips/Non-Gaussian Tensor Programs | 1 + data/2022/neurips/Non-Linear Coordination Graphs | 1 + ...or Contrastive Learning of Sentence Embeddings | 1 + ...s via Interpretable Multiple Instance Learning | 1 + ...C-Based Non-Autoregressive Machine Translation | 1 + ...yoffs: Improved Planning with Sublinear Regret | 1 + ...ex online learning via algorithmic equivalence | 1 + data/2022/neurips/Non-deep Networks | 1 + ...isspecification in Models of Molecular Fitness | 1 + ...lization in the Bandits with Knapsacks Problem | 1 + ...d Registration with Neural Deformation Pyramid | 1 + .../neurips/Non-stationary Bandits with Knapsacks | 1 + ...ng the Stationarity in Time Series Forecasting | 1 + .../Nonlinear MCMC for Bayesian Machine Learning | 1 + ...ion Reduction with a Stochastic Neural Network | 1 + ...ive Tensor Completion via Integer Optimization | 1 + ...cation for Single Deterministic Neural Network | 1 + ...nary Dual Averaging and Online Fair Allocation | 1 + ...for Knockoff-free Controlled Feature Selection | 1 + ...: Heterogeneous Precisions via Trainable Noise | 1 + ... theoretical analysis of graph (over)smoothing | 1 + ...gmentation from Rigid Dynamics of Point Clouds | 1 + ... Labels for Investigating Visual Eye Semantics | 1 + ... of Message-Passing GNNs in Larger Test Graphs | 1 + ...thogonal Propagation with Ego-Network Modeling | 1 + ...Data Subset Selection under Distribution Shift | 1 + ...Fake Detection via One-Shot Test-Time Training | 1 + ...owledge Graph Embeddings via Optimal Transport | 1 + ...s Sequences with Class Prompt for Visual Tasks | 1 + ...ement Algorithms with Implicit Differentiation | 1 + .../Object Scene Representation Transformer | 1 + .../Object-Category Aware Reinforcement Learning | 1 + ...lel Data Balanced in Gender within Occupations | 1 + ...r Action-Dependent Non-stationary Environments | 1 + ...Decision Processes under Non-Parametric Models | 1 + ... with Deficient Support Using Side Information | 1 + ...on with Policy-Dependent Optimization Response | 1 + data/2022/neurips/Off-Team Learning | 1 + ...orcement Learning via $f$-Advantage Regression | 1 + ...forcement Learning with Knowledge Distillation | 1 + ...ing Better by Making Statistical Matches Match | 1 + ...radient Flow can Make Vanilla-GCNs Great Again | 1 + ...el for Image-Language and Video-Language Tasks | 1 + .../On A Mallows-type Model For (Ranked) Choices | 1 + ...ties of Diffusion-based Deep Generative Models | 1 + ...Teaching with Sample Complexity Bounded by VCD | 1 + ... Probabilistic Explanations for Decision Trees | 1 + ...ty Invariant Bounds, Non-smoothness and Beyond | 1 + ...n and Estimation of Distributions on Manifolds | 1 + ...ivergence Measures for Bayesian Pseudocoresets | 1 + ...t Online Imitation Learning via Classification | 1 + ...ies for Bandit Fixed-Confidence Identification | 1 + ...or Numerical Features in Tabular Deep Learning | 1 + ...ed Meta-Learning for Rapid Few-Shot Adaptation | 1 + ...rning in the Presence of Spurious Correlations | 1 + ...dent Bounds for Offline Reinforcement Learning | 1 + ... of the Optimal Solutions to Accuracy and Dice | 1 + ...arations Between Simple and Optimal Mechanisms | 1 + ...ernelized Multi-Armed Bandits with Constraints | 1 + ...ng Fairness and Accuracy on Multiple Subgroups | 1 + ...n in Noninteractive Local Differential Privacy | 0 ...ditional Mutual Information For Generalization | 1 + ...argin Maximization in Linear and ReLU Networks | 1 + ...gins and Generalisation for Voting Classifiers | 1 + ...n Measuring Excess Capacity in Neural Networks | 1 + ...n-Linear operators for Geometric Deep Learning | 1 + ...Optimal Learning Under Targeted Data Poisoning | 1 + ...rsonalization in Cross-Silo Federated Learning | 1 + ...anguage Models with no Catastrophic Forgetting | 1 + .../neurips/On Robust Multiclass Learnability | 1 + ...sonalized Collaborative and Federated Learning | 1 + data/2022/neurips/On Scalable Testing of Samplers | 1 + ...na for Randomly Initialized Recurrent Networks | 1 + ...cle-Consistent Generative Adversarial Networks | 1 + ...d Data Augmentation in Bayesian Classification | 1 + ...o infinite width using linear parameterization | 1 + ...e Adversarial Robustness of Mixture of Experts | 1 + ... the Complexity of Adversarial Decision Making | 1 + ...nce Theory for Hessian-Free Bilevel Algorithms | 1 + ...lti-Objective Gradient Manipulation and Beyond | 1 + ... Mean Aggregation Feature Imputation in Graphs | 1 + ...ent of Random Features Models Trained with SGD | 1 + ...ent Modality on Offline Reinforcement Learning | 1 + ...orks: Convergence Guarantees and Implicit Bias | 1 + ...Fine-tuning Versus Meta-reinforcement Learning | 1 + ...pschitz-Driven Rehearsal in Continual Learning | 1 + .../On the Effectiveness of Persistent Homology | 1 + ...uracy Optimality of Profile Maximum Likelihood | 1 + ...he Epistemic Limits of Personalized Prediction | 1 + .../On the Frequency-bias of Coordinate-MLPs | 0 ...lity and Predictability of Recommender Systems | 1 + ...fitted Three-Layer Neural Tangent Kernel Model | 1 + ...oftmax Gradient Play in Markov Potential Games | 1 + ...iability of Nonlinear ICA: Sparsity and Beyond | 1 + ...rtance of Gradient Norm in PAC-Bayesian Bounds | 1 + ...ral Networks Through Model Gradient Similarity | 1 + ... the Learning Mechanisms in Physical Reasoning | 1 + ...itations of Stochastic Pre-processing Defenses | 1 + ... Initialization of Diagonal State Space Models | 1 + ...entation Collapse of Sparse Mixture of Experts | 1 + ...ering Models: Adversarial Attacks and Defenses | 1 + ...aph Neural Diffusion to Topology Perturbations | 1 + ...Scaling Rules for Adaptive Gradient Algorithms | 1 + ...Machine Learning: A Maximum Deviation Approach | 1 + ...Stabilizing LTI Systems on a Single Trajectory | 1 + ...al Neural Tangent and Gaussian Process Kernels | 1 + ... and Scalability of Node Perturbation Learning | 1 + ...cy of Reward-Free Exploration in Non-Linear RL | 1 + ...on Between Model Invariance and Generalization | 1 + ...ning Models and their Internal Representations | 1 + ...f Noise Correlation in Stochastic Optimization | 1 + ...n the Tradeoff Between Robustness and Fairness | 1 + ... Receiver Operating Characteristic (ROC) curve | 1 + ...to Nash equilibria in general stochastic games | 1 + ...es in the likelihood for variational inference | 1 + ...ficulty of learning chaotic dynamics with RNNs | 1 + ...on of learning algorithms that do not converge | 1 + ...ion to optimally learn compositional functions | 1 + ...eep learning: quantifying the cost of symmetry | 1 + ...iational inference and auto-associative memory | 1 + ...ce learning with linear function approximation | 1 + ... Vector Diffusion Maps on the Projective Plane | 1 + ...Learning Optimally from Multiple Distributions | 1 + .../neurips/On-Device Training Under 256KB Memory | 1 + ...n Image Manipulation with Semantic Modulations | 1 + ...ve Multi-Label Learning with Label Enhancement | 1 + ...ic and Preference Learning over Multiple Users | 1 + ...Position Encoding for Point Cloud Registration | 1 + ...ackdoor Erasing via Adversarial Weight Masking | 1 + ...Shot Object Pose Estimation without CAD Models | 1 + .../neurips/Online Agnostic Multiclass Boosting | 1 + .../Online Algorithms for the Santa Claus Problem | 1 + ...d Learning in the Presence of Strategic Agents | 1 + ...-Consistency Tradeoffs for the Two-Stage Model | 1 + ...nts: Towards the Best of Two Worlds and Beyond | 1 + data/2022/neurips/Online Decision Mediation | 1 + ...brium Learning for Regularization by Denoising | 1 + .../Online Frank-Wolfe with Arbitrary Delays | 1 + ...ork Revenue Management with Reusable Resources | 1 + ...ation: Multicalibeating and Other Applications | 1 + ...tion with Hierarchical Dirichlet Point Process | 1 + data/2022/neurips/Online PAC-Bayes Learning | 1 + ...Reinforcement Learning for Mixed Policy Scopes | 1 + ...ining Through Time for Spiking Neural Networks | 1 + ...ion for Ontological Multi-Label Classification | 0 ...Dataset - With Application to Super-Resolution | 1 + ...orcement Learning with Neural Reward Functions | 1 + ...AUC: Towards AUC-Oriented Open-Set Recognition | 1 + ...Benchmark Datasets for Full Waveform Inversion | 1 + ...ize Research Access to Social Media AR Filters | 1 + ...king Generalized Out-of-Distribution Detection | 1 + ...sing intraoperative stimulated Raman histology | 1 + ...a Transparent Evaluation of Model Explanations | 1 + ...ined connectivity of recurrent neural networks | 1 + .../neurips/Operator Splitting Value Iteration | 1 + ...entralized Stochastic Variational Inequalities | 1 + .../Optimal Binary Classification Beyond Accuracy | 1 + ...ccurate Post-Training Quantization and Pruning | 1 + ...r Adaptive Online Learning with Switching Cost | 1 + .../neurips/Optimal Dynamic Regret in LQR Control | 1 + ...fficiency-Envy Trade-Off via Optimal Transport | 1 + ...imal Distributed Optimization Under Similarity | 0 ...Latent Transformation for Contrastive Learning | 1 + ...uery Complexities for Dynamic Trace Estimation | 1 + ...egularized Conditional Mean Embedding Learning | 1 + ... Locally Balanced Proposals in Discrete Spaces | 1 + .../Optimal Transport of Classifiers to Fairness | 1 + ...entity-invariant Facial Expression Recognition | 1 + data/2022/neurips/Optimal Weak to Strong Learning | 1 + ...mal and Adaptive Monteiro-Svaiter Acceleration | 1 + .../neurips/Optimal-er Auctions through Attention | 1 + ...Coarse Correlated Equilibria in Bimatrix Games | 1 + ...Learning with Few Samples and Tight Guarantees | 1 + ...rches for Combinatorial Black-Box Optimization | 0 ...ptimizing Data Collection for Machine Learning | 1 + ...aps of Vision Transformers Improves Robustness | 1 + ...el Selection in Offline Reinforcement Learning | 1 + ...cient Online Learning for Smoothed Adversaries | 0 ...dinality Estimators Are Differentially Private | 1 + .../neurips/Ordered Subgraph Aggregation Networks | 1 + ...Prompts for Language-Guided Ordinal Regression | 1 + ...nsformer Backbone with Token Orthogonalization | 1 + ...sion and Procession of Hippocampal Place Cells | 1 + ...tion via Conditional Kernel Independence Model | 1 + ...kelihood Ratio on Informative Hierarchical VAE | 1 + ...e Limit of Low-bit Transformer Language Models | 1 + ... Sparse Estimation via Non-Convex Optimization | 1 + ...Mean Estimation for Heavy-Tailed Distributions | 1 + ...a Efficient Collaborative Open-Source Sampling | 1 + ...arameterization from Computational Constraints | 1 + ...t Cloud Analysis with Point-to-Pixel Prompting | 1 + ... So Tight That They Can Explain Generalization | 1 + ...dictions in Multi-Agent Reinforcement Learning | 0 .../neurips/PALBERT: Teaching ALBERT to Ponder | 1 + ...ion Loop with Memory for Long-Horizon Planning | 1 + ...sive Benchmark for Scientific Machine Learning | 1 + ...ted Domain Programming, Learning, and Planning | 0 ...k Benchmark for Protein Sequence Understanding | 1 + ... Detectors via Pearson Correlation Coefficient | 1 + ...try Dataset for Machine Learning in Proteomics | 0 ...ompositional Multi-task Reinforcement Learning | 1 + ...e Latent Manifold for Unsupervised Pretraining | 1 + ...sion via Alternating Reverse Filtering Network | 1 + ...arallel Tempering With a Variational Reference | 1 + ...l Transport with semi-dual Brenier formulation | 1 + .../neurips/Parameter-Efficient Masking Networks | 1 + ...ee Dynamic Graph Embedding for Link Prediction | 1 + ...ee Regret in High Probability with Heavy Tails | 1 + ... Overparameterization and Membership Inference | 1 + ...etargetable Decision-Makers Tend To Seek Power | 1 + ...ng Is All You Need for Novel Object Captioning | 1 + ...ing for Expensive Multi-Objective Optimization | 1 + ...atment Effects with Implicit Generative Models | 1 + ...s for 3D Shape Completion on Unseen Categories | 1 + ...pen-vocabulary models by interpolating weights | 1 + ...odels Can Better Exploit Test-Time Computation | 1 + ...misconceptions about Lipschitz neural networks | 1 + .../PeRFception: Perception using Radiance Fields | 1 + .../neurips/Peer Prediction for Learning Agents | 1 + ...ce Image Quality Models with Human-in-the-Loop | 1 + .../Perfect Sampling from Pairwise Comparisons | 1 + ...DouDizhu with Perfect Information Distillation | 1 + data/2022/neurips/Performative Power | 1 + ...rmers for Crystal Material Property Prediction | 1 + data/2022/neurips/Peripheral Vision Transformer | 1 + ...munication Efficiency, Robustness and Fairness | 0 ...nline Federated Learning with Multiple Kernels | 1 + .../Perturbation Learning Based Anomaly Detection | 1 + ...on from Clean Training to Adversarial Training | 1 + ... in high-dimensional two-layer neural networks | 1 + .../Phase transitions in when feedback is useful | 1 + ...fusion Models with Deep Language Understanding | 1 + ...issue Deformation in Image-Guided Neurosurgery | 1 + ...ed Face Rendering for NIR-VIS Face Recognition | 1 + ...ral PDE Solvers with Mixed Boundary Conditions | 1 + ...t Representations of Equilibrium Network Flows | 1 + ...ic Monoculture lead to Outcome Homogenization? | 1 + ... the Law and a 256GB Open-Source Legal Dataset | 1 + ...ainty Quantification through Loss Minimisation | 1 + ...g Model For Model-Based Reinforcement Learning | 1 + ...anning for Sample Efficient Imitation Learning | 1 + ...izon of BAMDPs via Epistemic State Abstraction | 1 + ...nd, and Snow for Optimization Time Integration | 1 + ... Image Completion with Gaussian Mixture Models | 0 ...d Vector Attention and Partition-based Pooling | 1 + ...ders for Hierarchical Point Cloud Pre-training | 1 + ... with Improved Training and Scaling Strategies | 1 + ...l Action Detection with Learnable Query Points | 1 + data/2022/neurips/Poisson Flow Generative Models | 1 + ... Augmentation Technique for LiDAR Point Clouds | 1 + ...cy Gradient With Serial Markov Chain Reasoning | 1 + ...ames: Unified Framework and Faster Convergence | 1 + ...ion for Long-Term Fairness in Decision Systems | 1 + ...ization with Linear Temporal Logic Constraints | 1 + ...t Multi-Task Adaptation for Dense Vision Tasks | 1 + ...lds for Subband Decomposition and Manipulation | 1 + ... time guarantees for the Burer-Monteiro method | 2 ++ ...libria with a Mediator in Extensive-Form Games | 1 + ...ental Design for Optimal Sparse Linear Bandits | 1 + ...Forests via Recursive Greedy Risk Minimization | 1 + ...ely Weighted Kernel Quadrature via Subsampling | 1 + ... estimators for learning to defer to an expert | 1 + ...r-independent Mechanisms with Value Maximizers | 1 + ...ior Collapse of a Linear Latent Variable Model | 1 + .../Posterior Matching for Arbitrary Conditioning | 1 + ... Sample Efficiency in Bayesian Neural Networks | 1 + ...omputational Uncertainty in Gaussian Processes | 1 + ...of single-qubit native quantum neural networks | 1 + ...s on Spatiotemporal Traffic Forecasting Models | 1 + ...al Adversarial Multivalid Conformal Prediction | 1 + ...ical Demonstrations in Multi-Goal Environments | 1 + ...sian Transfer Learning with Informative Priors | 1 + ...or Generalizable Visual Reinforcement Learning | 1 + ...anguage Models for Interactive Decision-Making | 1 + ...ty Evaluation for Small-Data Transfer Learning | 1 + ...tivation Distributions Expose Backdoor Neurons | 1 + .../neurips/Pre-trained Adversarial Perturbations | 1 + ...der Scalings for Dot-product Kernel Regression | 0 ...or Log-loss via a Truncated Bayesian Algorithm | 1 + ...Drug Perturbations at a Single-Cell Resolution | 1 + ...ng Label Distribution from Multi-label Ranking | 1 + ...redictive Coding beyond Gaussian Distributions | 1 + ...ying for Autoregressive Neural Sequence Models | 1 + ...by Not-True Distillation in Federated Learning | 1 + ...on-Computation Gaps and Sparse Mean Estimation | 1 + ...ent: More Iterations without More Privacy Loss | 1 + .../neurips/Private Estimation with Public Data | 1 + ...Path Distance Release with Improved Error Rate | 1 + data/2022/neurips/Private Isotonic Regression | 1 + .../Private Multiparty Perception for Navigation | 1 + ...Set Generation with Discriminative Information | 1 + ...ta for Multitask Learning and Marginal Queries | 1 + ...on-Efficient Algorithms for Entropy Estimation | 1 + ...utation for Mixed Categorical and Ordered Data | 1 + ...tributions for RNA Folding and Molecule Design | 1 + ... Generalization via Quantile Risk Minimization | 1 + ...e Unreliable for Concept Removal and Detection | 1 + ...ral Image Programs for Representation Learning | 1 + ...r Revenue Maximization with Multiple Purchases | 1 + ...ect Segmentation from Real-world Single Images | 1 + ...Randomized Gradient Smoothing and Quantization | 0 ...ustomizable and Composable Learning Algorithms | 1 + ...elf-Explainable Prototypical Variational Model | 1 + ...a Reinforcement Learning Agent via Prototyping | 1 + ...t for Few-Shot 3D Point Cloud Object Detection | 1 + ...resentation Learning in Reinforcement Learning | 1 + ...st Backdoor Policies in Reinforcement Learning | 1 + ...entation Learning in Multitask Bandits and MDP | 1 + ...erparameterized Meta-learning Trained with SGD | 1 + ...e Identification Under Post-Nonlinear Mixtures | 1 + ... of Out-of-Distribution Data (Almost) for Free | 1 + ...strained RL with Linear Function Approximation | 1 + ...Reinforcement Learning via Strategy-wise Bonus | 1 + ...ning in Partially Observable Dynamical Systems | 1 + ...nforcement Learning via Active Reward Learning | 1 + .../Provably expressive temporal graph networks | 1 + ...RL with side information about latent dynamics | 1 + ...rovably tuning the ElasticNet across instances | 1 + ...imal Learning With Opponent-Learning Awareness | 1 + .../neurips/Proximal Point Imitation Learning | 1 + ...ong rat visual cortex and deep neural networks | 1 + ...ts and Convex Geometry: Towards No Assumptions | 1 + ...uning has a disparate impact on model accuracy | 1 + ...hrough the Lens of Training and Regularization | 1 + ...Pseudo-Riemannian Graph Convolutional Networks | 1 + ...er Co-Attention for Social Text Classification | 1 + ... for Pulsative Physiological Signal Imputation | 1 + .../Pure Transformers are Powerful Graph Learners | 1 + ... impossibility: Who's the fairest of them all? | 1 + ...yramid Attention For Source Code Summarization | 0 ...lignment for Vision-language Model Pretraining | 1 + ...toencoders in Python - A Benchmarking Use Case | 1 + ...and Fully Quantized Low-bit Vision Transformer | 1 + ...Controllable Image Generation and Manipulation | 1 + ...ble Text Generation with Reinforced Unlearning | 1 + ... between Dataset Design and Robustness of CLIP | 1 + ...ased Image Segmentation by Selective Inference | 1 + ...ning Framework Constraining Outage Probability | 1 + ...d Training of Gradient Boosting Decision Trees | 1 + ...ributions and Estimating Normalizing Constants | 1 + ...o Logarithmic Regret Stochastic Convex Bandits | 1 + ...Quasi-Newton Methods for Saddle Point Problems | 1 + ... Regression via Spatial-Aware Part-Level Query | 1 + ...Dynamic Capacity Region of Multiplayer Bandits | 1 + ...e Key Towards Long-Term Multi-Object Tracking? | 1 + ...ial Model-Based Offline Reinforcement Learning | 1 + ...s in Knowledge-Based Visual Question Answering | 1 + ...zed Decision Learning with Sensitive Variables | 1 + .../RKHS-SHAP: Shapley Values for Kernel Methods | 1 + ...raining for Human-Object Interaction Detection | 1 + ...Stable Assemblies of Recurrent Neural Networks | 1 + ...nforcement Learning via Conservative Smoothing | 1 + ...ive Augmentations for Self-supervised Learning | 1 + ...al-Time Semantic Segmentation with Transformer | 1 + ...enchmark for Spatial Precipitation Downscaling | 1 + ...malization Aggregation for Adversarial Defense | 1 + ...ly Fair Randomized Facility Location Mechanism | 1 + .../neurips/Random Sharpness-Aware Minimization | 1 + ...ckdoor Attack Detection without Clean Datasets | 1 + ...ray-box Certificates for Graph Neural Networks | 1 + ... Clustering: Fast and Optimal Kernel $k$-Means | 1 + .../Rank Diminishing in Deep Neural Networks | 1 + ...ture Removal for Out-of-distribution Detection | 1 + ... Model Architecture Adaption for Meta-Learning | 1 + ...opolis Algorithms for Model Selection Problems | 1 + ...ems: Finding Lottery Tickets at Initialization | 1 + ... for Predictive Multiplicity in Classification | 1 + ... Generalization Error for Distributed Learning | 1 + ...Convex Optimization in Adaptive Linear Control | 1 + ...Matrix Approximation via Dyson Brownian Motion | 1 + ...Retrieve and Co-segment for Zero-shot Transfer | 1 + ...ased Models from a Message-Passing Perspective | 1 + ... Unsuitable for Complex-Valued Neural Networks | 1 + ...rk Pruning and the Undecayed Pruning Algorithm | 1 + ...eceding Horizon Inverse Reinforcement Learning | 1 + ... General, Powerful, Scalable Graph Transformer | 1 + .../Recommender Forest for Efficient Retrieval | 1 + ...ing Training Data From Trained Neural Networks | 1 + ...nstruction on Trees and Low-Degree Polynomials | 1 + ... Text in Federated Learning of Language Models | 1 + .../Recruitment Strategies That Take a Chance | 1 + ...al Networks Learn Succinct Learning Algorithms | 1 + data/2022/neurips/Recurrent Memory Transformer | 1 + ...n Transformer with Guided Deformable Attention | 1 + ...inimax Games: A Level $k$ Gradient Play Method | 1 + .../2022/neurips/Recursive Reinforcement Learning | 1 + .../RecursiveMix: Mixed Learning with History | 1 + ...intrinsic rewards via constrained optimization | 1 + ...ghts and Activations for AdderNet Quantization | 1 + ... Fields for Effective Non-rigid Shape Matching | 1 + ...ence Diagrams of Networks: CoralTDA and PrunIT | 1 + ...Free Message Passing for Graph Neural Networks | 0 ...ns help generalization in wide neural networks | 1 + ...entanglement of Multilingual Translation Model | 0 ...or Information-Directed Reinforcement Learning | 1 + ...tilabel Classification in Sparse Label Regimes | 1 + ...unds for Risk-Sensitive Reinforcement Learning | 1 + ...nt Ascent for Two-Player Zero-Sum Markov Games | 1 + .../Regularized Molecular Conformation Fields | 1 + ...using Prior Computation to Accelerate Progress | 1 + ...etic Algorithm for Structure-based Drug Design | 1 + ...ss: Breaking the Dependence on the State Space | 1 + ... Learning with Automated Auxiliary Loss Search | 1 + ...ng with Logarithmic Regret and Policy Switches | 1 + ...forcement Learning with Neural Radiance Fields | 1 + ...ment Learning with Non-Exponential Discounting | 1 + .../Reinforcement Learning with a Terminator | 1 + ...ation-Constrained Decoding for Text Generation | 1 + ...t Relationships as Fine-Grained Discriminators | 1 + ...: Provable Efficiency and Applications to MARL | 1 + ...traints with Non-stationary Continuous Filters | 1 + ... into Addressable Memories for Neural Networks | 1 + ...cations to Private and Robust Machine Learning | 1 + ...ural Networks by Leaving the Right Past Behind | 1 + ...esenting Spatial Trajectories as Distributions | 1 + ...Optimization: Theoretical Framework and Limits | 1 + ...ent Reinforcement Learning Value Factorization | 1 + .../neurips/ResT V2: Simpler, Faster and Stronger | 1 + ... Filter Networks for Multiscale Reconstruction | 1 + ...lving the data ambiguity for periodic crystals | 1 + ...ed Learning with All-In-One Neural Composition | 1 + ...pecting Transfer Gap in Knowledge Distillation | 1 + ...Knowledge for Learning with Dynamic Definition | 1 + ...ignment in Video Super-Resolution Transformers | 1 + ...king Generalization in Few-Shot Classification | 1 + ...hinking Image Restoration for Object Detection | 1 + ...Cooperative Multi-Agent Reinforcement Learning | 1 + ...aph Evaluation Under the Open-World Assumption | 1 + ...ied Robustness: A Boolean Function Perspective | 1 + ... in the Context of Efficient Video Recognition | 1 + ...g for Generalization in Reinforcement Learning | 1 + ...Probabilistic Programs with Stochastic Support | 1 + ...apley Value-based Approach in Frequency Domain | 0 ...y Efficient Approach with Group Discrimination | 1 + ...ing the Reverse-engineering of Trojan Triggers | 1 + ...through regularization in the hyperbolic space | 1 + .../neurips/Retrieval-Augmented Diffusion Models | 1 + ...ing Accurate and Faithful Patient Instructions | 1 + ...tive Adversarial Replay for Continual Learning | 1 + ... of mSGD under milder requirement on step size | 1 + ...ting Active Sets for Gaussian Process Decoders | 1 + ...earning from the Perspective of Graph Spectrum | 1 + ...visiting Heterophily For Graph Neural Networks | 1 + ...iting Injective Attacks on Recommender Systems | 1 + ...ing Neural Scaling Laws in Language and Vision | 1 + ...s for Robust and Generalizable Stereo Matching | 1 + ...n-convex Stochastic Decentralized Optimization | 1 + ...nference and Adaptation by Anchored Clustering | 1 + ...n on Images: From Vectorization to Convolution | 1 + ...rse Convolutional Model for Visual Recognition | 1 + data/2022/neurips/Riemannian Diffusion Models | 1 + ...arning Stochastic Representations on Manifolds | 1 + .../Riemannian Score-Based Generative Modelling | 1 + ... for Least Squares in the Interpolation Regime | 1 + .../Risk-Driven Design of Perception Systems | 1 + ...Disabling Shortcuts and Learning New Knowledge | 1 + ... Anytime Learning of Markov Decision Processes | 1 + ...bust Bayesian Regression via Hard Thresholding | 1 + ...odels by Pruning Randomly-initialized Networks | 1 + ...ibration with Multi-domain Temperature Scaling | 1 + ...e-Level Adversaries are Interpretability Tools | 1 + ...d Method of Moments: A Finite Sample Viewpoint | 1 + ...ucture Learning via Multiple Statistical Tests | 1 + ...of a Few Demonstrations with a Backwards Model | 1 + ... Mirror Descent Inverse Reinforcement Learning | 1 + ...Robust Learning against Relational Adversaries | 1 + ... Selection and Nearly-Proper Learning for GMMs | 1 + .../neurips/Robust Models are less Over-Confident | 1 + ...ior Estimation and Statistical Model Criticism | 1 + ...nt Policy Evaluation in Reinforcement Learning | 1 + ...bust Reinforcement Learning using Offline Data | 1 + data/2022/neurips/Robust Rent Division | 1 + ...ised Learning when Not All Classes have Labels | 1 + data/2022/neurips/Robust Streaming PCA | 1 + ...bust Testing in High-Dimensional Sparse Models | 1 + ...dels Against Visual and Language Perturbations | 1 + .../Robustness Disparities in Face Detection | 1 + ...the bad (depth), and the ugly (initialization) | 1 + ...Depends on the Shape of the Noise Distribution | 1 + ...to Unbounded Smoothness of Generalized SignSGD | 1 + ...ures in Microservices through Causal Discovery | 1 + ...elds for Learning a Natural Illumination Prior | 1 + ...on Learning with Skew R\303\251nyi Divergence" | 1 + ...u for Single-view Clothed Human Reconstruction | 1 + ... Occam's Razor for Domain Incremental Learning | 1 + ...Augmentation in Offline Reinforcement Learning | 1 + ...om Shading and Shadow under a Single Viewpoint | 1 + ...3GC: Scalable Self-Supervised Graph Clustering | 1 + ... as Multidimensional Signals with State Spaces | 0 ...tacking Lattice Cryptography with Transformers | 1 + ...trained Real-world Arbitrary Image collections | 1 + ...Aware Point Affiliation for Feature Upsampling | 1 + ... Method for Nonconvex-Concave Minimax Problems | 1 + ...-Aware Pipeline for Data Parallel DNN Training | 1 + ...Object-Centric Learning from Real-World Videos | 1 + ...sformer Pruning via Collaborative Optimization | 1 + ...or Camera Measurement of Physiological Signals | 1 + ...asting with Sample Convolution and Interaction | 1 + ...ly-Supervised Whole-Slide Image Classification | 1 + ...Unknown Environments by Volumetric Integration | 1 + ...ld through Simultaneous Generation and Mapping | 1 + ...apley Value Theory into Multi-Agent Q-Learning | 1 + .../SHINE: SubHypergraph Inductive Neural nEtwork | 1 + ...ions for Detecting Out-of-Distribution Objects | 1 + ...O: Smoothing Inference with Twisted Objectives | 1 + ...rated Gradients Estimation of Neuron Relevance | 1 + ...Flow: Learning Optical Flow with Super Kernels | 1 + ...ring and Process Control Learning Environments | 1 + .../SNAKE: Shape-aware Neural 3D Keypoint Field | 1 + ...twork through Regularized Adversarial Training | 1 + ...pretable unsupervised domain adaptation in EEG | 1 + ...nsupervised Multi-agent Reinforcement Learning | 1 + ...rmer for Dense Point Cloud Semantic Completion | 1 + ...for Learning Single Neurons with Massart Noise | 1 + ...ter-Efficient Image-to-Video Transfer Learning | 1 + ...tion Activity with Spatiotemporal Transformers | 1 + .../STaR: Bootstrapping Reasoning With Reasoning | 1 + .../Safe Opponent-Exploitation Subgame Refinement | 1 + ...m for Safety Evaluation of Autonomous Vehicles | 1 + ...namic Systems via Stochastic Barrier Functions | 1 + ...ageMix: Saliency-Guided Mixup for Point Clouds | 1 + .../Saliency-Aware Neural Architecture Search | 1 + ... Functions for Greedy-Best-First and A* Search | 1 + ...Sample Constrained Treatment Effect Estimation | 1 + ...Benchmark for Practical Molecular Optimization | 1 + ... Correlated Equilibria in Extensive-Form Games | 1 + ... Learning of Partially Observable Markov Games | 1 + ...e-Then-Optimize Batch Neural Thompson Sampling | 1 + ...istributions with Infinity-Distance Guarantees | 1 + ... Orthogonal-Space Variational Gradient Descent | 1 + ...Hamiltonian Monte Carlo in a Constrained Space | 1 + ...aster Rates in Finite-Sum Minimax Optimization | 1 + ... Temporal and Multi-Spectral Satellite Imagery | 1 + ...ass of Non-Convex Optimization with Guarantees | 1 + data/2022/neurips/Scalable Infomin Learning | 1 + .../Scalable Interpretability via Polynomials | 1 + ...ing Option Discovery based on Kronecker Graphs | 1 + ...esentations with Learnable Positional Features | 1 + ...extual Bandits with Constant Regret Guarantees | 1 + ...t Estimates of Continuous-Valued Interventions | 1 + ...cient Non-adaptive Deterministic Group Testing | 0 ...onal Neural Networks with Differential Privacy | 1 + ...sing Discrete Optimization with Graph Coloring | 1 + .../Scale-invariant Learning by Physics Inversion | 1 + ...res: A New Baseline for Efficient Model Tuning | 2 ++ ...ning via Cross-Modality Gradient Harmonization | 1 + ...d Diffusion meets Annealed Importance Sampling | 1 + ...Score-Based Generative Models Detect Manifolds | 1 + ...ng Secretly Minimizes the Wasserstein Distance | 1 + ...poral Alignment in Few-Shot Action Recognition | 1 + ... to Re-Align With Human Values from Text Edits | 1 + ...aussianization protocol for Federated Learning | 1 + ...dual and collective dynamics with transformers | 1 + ...nal Attention Design for Semantic Segmentation | 1 + ...ic Segmentation with Plain Vision Transformers | 1 + ...s via an Object-Centric Layered Representation | 1 + ...biased Learning by Contradicting-pair Sampling | 1 + ...sentations for variable-rate image compression | 1 + .../Self-Aware Personalized Federated Learning | 1 + ...ry of Kernel Evolution in Wide Neural Networks | 1 + .../Self-Explaining Deviations for Coordination | 1 + ...Cooperative Multi-agent Reinforcement Learning | 1 + ...ages as Differentiable Fractal Representations | 1 + ...erts for Test-Agnostic Long-Tailed Recognition | 1 + ...For Time Series via Time-Frequency Consistency | 1 + ...r Representation Learning without Demographics | 1 + ... Image Restoration with Blurry and Noisy Pairs | 1 + ...f-Supervised Learning Through Efference Copies | 1 + ...of Brain Dynamics from Broad Neuroimaging Data | 1 + ...Supervised Learning via Maximum Entropy Coding | 1 + ...ing with an Information Maximization Criterion | 1 + ...vised Pretraining for Large-Scale Point Clouds | 1 + ...Representation Learning with Semantic Grouping | 1 + ...plaining deep models with logic rule reasoning | 1 + ...lf-supervised Amodal Video Object Segmentation | 1 + ...ph Pre-training Based on Structural Clustering | 1 + ...epth estimation with volumetric feature fusion | 1 + ...uided Masking for Learning Masked Autoencoders | 1 + ...ic Diffusion Network for Semantic Segmentation | 1 + ...ge Abstractions and Pretrained Representations | 1 + ...obabilistic Layers for Neuro-Symbolic Learning | 1 + ...ainty intervals for disentangled latent spaces | 1 + ...zing Flows through Differentiable Tessellation | 1 + ... Generative Models for Multiagent Trajectories | 1 + ... Graph Laplacian Tree Alternating Optimization | 1 + ...tic Segmentation via Gentle Teaching Assistant | 1 + ...tion Based on Uncertainty-Guided Pseudo Labels | 0 ...finitely Constrained Markov Decision Processes | 1 + .../Semi-supervised Active Linear Regression | 1 + ...ith Prototype-based Consistency Regularization | 1 + .../Semi-supervised Vision Transformers at Scale | 1 + ... for Unlabeled Clients with Alternate Training | 1 + ...ate Text Generation via Knowledge Distillation | 1 + ...el Imitation Learning with Unobserved Contexts | 1 + .../neurips/Sequence-to-Set Generative Models | 1 + .../Sequencer: Deep LSTM for Image Classification | 1 + ...ation Design: Learning to Persuade in the Dark | 1 + ... Meta-Interpolation for Few-Task Meta-Learning | 1 + ...Bridging Offline and Online Knowledge Transfer | 1 + ... And Structure Preserving Differential Privacy | 1 + ...ages using Monte Carlo Rendering and Denoising | 1 + ...ive Text-Conditioned 3D Shape Generation Model | 1 + ...ge for Meta-learning with Feature Descriptions | 1 + ...on under Global Kurdyka-Lojasiewicz Inequality | 1 + ...ous SGD for Distributed and Federated Learning | 1 + .../neurips/Sharpness-Aware Training for Free | 1 + ...on for Safe Multi-Agent Reinforcement Learning | 1 + ...n Efficient ConvNet for Image Super-Resolution | 1 + .../neurips/SignRFF: Sign Random Fourier Features | 1 + ...Processing for Implicit Neural Representations | 1 + ...cal Perspectives and the Role of Rank Collapse | 1 + ...y with Non-Expansive Generative Network Priors | 1 + ...fare Maximization in Rich Advertising Auctions | 1 + ...c Learning for Complex and Naturalistic Videos | 1 + ...al Greedy Online Contention Resolution Schemes | 1 + .../Simplified Graph Convolution with Heterophily | 1 + ...m Search for Neural Combinatorial Optimization | 1 + ... Imputation and Structure Learning with Groups | 1 + ...an Homotopy Method for Non-convex Optimization | 1 + ...ainty Estimation via Stochastic Data Centering | 1 + ...elationship Learning using Conditional Queries | 1 + ...tion with Instance-sensitive Sample Complexity | 0 ...ase deep learning in cortico-cortical networks | 1 + ...gmentation requires Few-parameters Fine-tuning | 1 + ...ural networks: interpolation and approximation | 1 + ...g Size-Generalization in Graph Neural Networks | 1 + ...al Networks with Sublinear Training Complexity | 1 + ...Boosted Decision Tree for Multioutput Problems | 1 + ... Image Classification with Provable Guarantees | 1 + ... for Multi-task Offline Reinforcement Learning | 1 + ...xperts for fine-grained debugging and analysis | 1 + ...doors for Neural Networks Trained from Scratch | 1 + ...with Perturbed Payoffs and Unknown Transitions | 1 + ...hed Embeddings for Certified Few-Shot Learning | 1 + ...imization Based on Discounted-Normal-Predictor | 1 + ...ayesian Optimization with Pathwise Exploration | 1 + ...n arbitrary length-scales in molecular systems | 1 + ...Refinery for Imbalanced Partial-Label Learning | 1 + ...arning Elliptic Equations via Gradient Descent | 1 + ...cial Contagion Management with Task Migrations | 1 + ... Regret Bounds of Concurrent Thompson Sampling | 0 ...Unsupervised Anomaly Detection with Noisy Data | 1 + ...tative Reasoning Problems with Language Models | 1 + ...erated Learning with Communication Compression | 1 + ...ent Variables Given Local Background Knowledge | 0 ...d Complete Verification of Polynomial Networks | 1 + ...mulation Platform for Visual-Acoustic Learning | 1 + .../SparCL: Sparse Continual Learning on the Edge | 1 + ...rier Backpropagation in Cryo-EM Reconstruction | 1 + ...rocess Hyperparameters: Optimize or Integrate? | 1 + ...Detection Thresholds in Stochastic Block Model | 1 + ...ure Interaction Detection and Sparse Selection | 1 + ...Probabilistic Circuits via Pruning and Growing | 1 + .../Sparse Structure Search for Delta Tuning | 1 + ...g Tickets are Data-Efficient Image Recognizers | 1 + ...to Densify 3D Features for 3D Object Detection | 1 + .../Sparsity in Continuous-Depth Neural Networks | 1 + ...tiable Sparsity via Regularized Transportation | 1 + data/2022/neurips/Spatial Mixture-of-Experts | 1 + ... Convolution for Efficient 3D Object Detection | 1 + ...ing Set for Deep Networks in the Kernel Regime | 1 + ...lization in Image-based Reinforcement Learning | 1 + ...ely: Accelerating MCTS with Virtual Expansions | 1 + ...ical Channels for Modeling Atomic Interactions | 1 + ...zation Layer: Representation Using Only Angles | 1 + ...t-kl Inequalities for Ternary Random Variables | 1 + ...t Transformer for Automatic Speech Recognition | 1 + ... Generalization Bounds of Adversarial Training | 1 + ...f Gradient Methods for Shallow Neural Networks | 1 + ...n for Markov Chain Stochastic Gradient Methods | 1 + ...stering: from Single Kernel to Multiple Kernel | 1 + ...e Under Interference Without Network Knowledge | 1 + ...ttention for Recurrent Processing of Sequences | 1 + ... Sequence Modeling with Partially Labeled Data | 1 + ...ale Graph Building for Clustering and Learning | 0 ...verse Problems: A Stochastic Gradient Approach | 1 + ...al Guarantees for Sliced Wasserstein Distances | 1 + ...pproximating Turing Machines with Transformers | 1 + ...ks: A Social Psychology Perspective of Loafing | 1 + .../Stochastic Adaptive Activation Function | 1 + ...e Reduction for Stochastic Monotone Inclusions | 1 + ...stic Multiple Target Sampling Gradient Descent | 1 + ... Graphs: Finite-Time and Asymptotic Optimality | 1 + ...lexity of SGD for Gradient-Dominated Functions | 1 + ...astic Window Transformer for Image Restoration | 1 + ...reaming Radiance Fields for 3D Video Synthesis | 1 + ...k Dataset for Sub-second Action Identification | 1 + ... the Learnability of Gomory Mixed Integer Cuts | 1 + ... Optimization and Symbolical Optimal Transport | 1 + ...al Knowledge Distillation for Object Detection | 1 + ...ructural Pruning via Latency-Saliency Knapsack | 1 + ...Aware Image Segmentation with Homotopy Warping | 1 + ...D Garment Modeling with Neural Sewing Machines | 1 + .../neurips/Structured Energy Network As a Loss | 0 ...ion for Generative Models with Explaining Away | 1 + ...cturing Representations Using Group Invariants | 1 + ...d Sampling in Stochastic Segmentation Networks | 1 + ...lower bounds for Principal Components Analysis | 1 + .../Subgame Solving in Adversarial Team Games | 1 + ... On Trees: An Empirical Baseline Investigation | 1 + ...blinear Algorithms for Hierarchical Clustering | 1 + .../Submodular Maximization in Clean Linear Time | 1 + ...sion with Applications to Tensor Decomposition | 1 + ...type Alignment for Universal Domain Adaptation | 1 + ...om Heterogeneous Data with Non-isotropic Noise | 1 + ...transitions & Statistical-to-Computational gap | 1 + .../Supervised Training of Conditional Monge Maps | 1 + ...Fidelity Race of Hyperparameter Configurations | 1 + ...rt Recovery in Sparse PCA with Incomplete Data | 1 + ...ptimization for Offline Reinforcement Learning | 1 + ... Assist Blind Navigation in Urban Environments | 1 + ... Multi-Agent Learning with Energy-based Models | 0 ...Online Reinforcement Learning for Auto-bidding | 1 + ...e and Strong Baseline for Transformer Tracking | 1 + ...etricity for Neural Combinatorial Optimization | 1 + ...istillation for Learned TCP Congestion Control | 1 + ...try Teleportation for Accelerated Optimization | 1 + .../Symmetry-induced Disentanglement on Graphs | 1 + ...arning Hamiltonians from Noisy and Sparse Data | 1 + ...per-parameters in Contextual Bandit Algorithms | 1 + ... Collaborate to Improve Adversarial Robustness | 1 + ...ise Approach to Unsupervised Ensemble Learning | 1 + ...of neural network quantum states using Lanczos | 0 ...coding Scheme for Neural Network Architectures | 1 + ...y-Aware Large Scale Mixture-of-Expert Training | 1 + ...bust 3D Stylization via Lighting Decomposition | 1 + ... for Drug-Protein Binding Structure Prediction | 1 + ... A Benchmark for Tracking Any Point in a Video | 1 + ...ning using Bootstrapped Neural Tangent Kernels | 1 + ... Text Generation of Pretrained Language Models | 1 + ...ion Transformer with Noun-Pronoun Distillation | 1 + ...-KNN: K Nearest Neighbor Search at Peak FLOP s | 1 + ...sient Redundancy Elimination-based Convolution | 1 + ... and its Application to Reinforcement Learning | 1 + .../TUSK: Task-Agnostic Unsupervised Keypoints | 1 + .../TVLT: Textless Vision-Language Transformer | 1 + .../TaSIL: Taylor Series Imitation Learning | 1 + ...Neural Architecture Search on Tabular Datasets | 1 + ...taset for Chinese Vision-Language Pre-training | 1 + ...nfinite Variance) Noise in Federated Learning" | 1 + ...ge Objects without Explicit Goal Specification | 1 + ...alignment in truncated kernel ridge regression | 1 + ...g the Tasks that Neural Networks Generalize on | 1 + .../2022/neurips/Task-Agnostic Graph Explanations | 1 + ...rning via Online Discrepancy Distance Learning | 1 + ...ask-level Differentially Private Meta Learning | 1 + ...ndistillable Classes in Knowledge Distillation | 0 ... Recovers Reward Functions for Text Generation | 1 + ...namically Evolving and Newly Emerging Entities | 1 + ...eural Network with Optimal Transport Distances | 1 + ...el Training through Memory Footprint Reduction | 1 + ...Batch Normalization in Spiking Neural Networks | 1 + ...low Processing Mechanisms in Sequence Learning | 1 + ...emporally Disentangled Representation Learning | 1 + .../Temporally-Consistent Survival Analysis | 1 + ...pose Benchmark Dataset for Recommender Systems | 1 + ...ogram Optimization with Probabilistic Programs | 1 + ...position and Its Tensor Completion Application | 1 + ...st Time Adaptation via Conjugate Pseudo-labels | 1 + ...-Shot Generalization in Vision-Language Models | 1 + .../Test-Time Training with Masked Autoencoders | 1 + .../neurips/Text Classification with Born's Rule | 1 + ...al Prototype Matching for Video-Text Retrieval | 1 + ...Corpus: A 1.6TB Composite Multilingual Dataset | 1 + ... can fail even above the Barvinok-Pataki bound | 1 + ...: Rate of Differentiating Through Optimization | 1 + ...aphic and Socioeconomic Diversity of the World | 1 + ...tion and Data Augmentation are Class Dependent | 1 + ...gh-Order Methods in Smooth Convex Optimization | 1 + ...y-Convex-Strongly-Concave Minimax Optimization | 1 + ...onal Trade-offs in High Dimensional Statistics | 1 + .../The Gyro-Structure of Some Matrix Manifolds | 1 + data/2022/neurips/The Hessian Screening Rule | 1 + ...tion in Evaluating Deep Reinforcement Learning | 1 + data/2022/neurips/The Implicit Delta Method | 1 + ...ad in Non-contrastive Self-supervised Learning | 1 + ...moting Collaborative Metric Learning Algorithm | 1 + ...Reciprocal Twin of Invariant Risk Minimization | 1 + ...lti-step Distributional Reinforcement Learning | 1 + ...ite Depth-and-Width Networks at Initialization | 1 + ...e Neural Testbed: Evaluating Joint Predictions | 1 + data/2022/neurips/The Phenomenon of Policy Churn | 1 + ...ls of Regularization in Off-Policy TD Learning | 1 + ...rative Routing Neural Networks for Chip Design | 1 + ...ng for Linear Regression under Covariate Shift | 1 + ...Privacy Onion Effect: Memorization is Relative | 1 + .../neurips/The Query Complexity of Cake Cutting | 2 ++ ...e of Baselines in Policy Gradient Optimization | 1 + ...Complexity of One-Hidden-Layer Neural Networks | 1 + ...Sequence Length Warmup for Training GPT Models | 1 + ...veness of PPO in Cooperative Multi-Agent Games | 1 + ...of Fully-Connected Layers for Low-Data Regimes | 1 + ...ns in Few-shot Prompting for Textual Reasoning | 1 + ...helps select flat minima: A stability analysis | 1 + ...d learning benefits of Daleian neural networks | 1 + ...ol principle for local learning at equilibrium | 1 + ...noise structure in low-rank matrix estimation? | 1 + ...airness in linear bandits with biased feedback | 1 + ...on models : 100GB to 10MB Criteo-tb DLRM model | 0 ...networks for temporally dependent observations | 1 + ... with Smoothness-Aware Quantization Techniques | 1 + ...Theoretically Provable Spiking Neural Networks | 1 + ...anched Optimal Transport with Multiple Sources | 1 + ...rary for Differentiable Nonlinear Optimization | 1 + ...eralized Linear Stochastic Convex Optimization | 1 + ...for sparse graphs with overlapping communities | 1 + ...ZCZE, a comprehensive NLP benchmark for Polish | 1 + ...iciently Learns to Control Diffusion Processes | 1 + ... Language Models and Automated Theorem Provers | 1 + ...in the Face of Uncertainty and Constant Regret | 1 + ...radient Methods For Nonconvex Minimax Problems | 1 + ...rantees for Zero-Shot Learning with Attributes | 1 + ...With Contrastive Fenchel-Legendre Optimization | 1 + ... Transport Robust under Martingale Constraints | 1 + ...nvolution Networks for Time-series Forecasting | 0 ... update? Neurons at equilibrium in deep models | 1 + ...ingerprinting in Computer-Aided Drug Discovery | 1 + ...Token-level Data Augmentation for Transformers | 1 + data/2022/neurips/Top Two Algorithms Revisited | 1 + ...l Diffusion for Molecular Conformer Generation | 1 + ...lf-Portrait Videos of Faces, Hands, and Bodies | 1 + ...Learning from Human-Collected Vision and Touch | 1 + ...ient Descent and Discretization Error Analysis | 1 + ...eural Network Against Adversarial Perturbation | 1 + ...eged Features Distillation in Learning-to-Rank | 1 + ...ing in the brain with self-supervised learning | 1 + ... Better Evaluation for Dynamic Link Prediction | 1 + ...ards Consistency in Adversarial Classification | 1 + ...ntangling Information Paths with Coded ResNeXt | 1 + ...ot Adaption of Generative Adversarial Networks | 1 + ... in Zero-Resource Sounding Object Localization | 1 + ...D Object Detection with Knowledge Distillation | 1 + ...ng Quantization of Pre-trained Language Models | 1 + ...on via 3D-aware Global Correspondence Learning | 1 + ...erous Manipulation with Reinforcement Learning | 1 + ...bration in Object Detection Under Domain Shift | 1 + ...ving Faithfulness in Abstractive Summarization | 1 + ...al Hyperparameter Optimizers with Transformers | 1 + ... Black-Box Attack Against Deep Neural Networks | 0 ...plexity in Distributed Non-Convex Optimization | 1 + ...equential Event Prediction: A Causal Treatment | 1 + ...rol of Singular Values of Convolutional Layers | 1 + ...nsductive Minimum Description Length Inference | 1 + ...ed Graph Structure Attacks via Gradient Debias | 1 + ...nference with Balanced Neural Ratio Estimation | 1 + ...e Restoration with Codebook Lookup Transformer | 1 + ...forcement Learning with a Safety Editor Policy | 1 + ...ly Inspired Neural Initialization Optimization | 1 + ...rs' Reasoning with Deep Reinforcement Learning | 1 + ...An Effective Theory of Representation Learning | 1 + ...nsation of Neural Networks at Initial Training | 1 + ... the Mixture-of-Experts Layer in Deep Learning | 1 + .../neurips/Towards Versatile Embodied Navigation | 1 + ...ual Question Answering: Benchmark and Baseline | 1 + ...mance Evaluation Protocol for Cooperative MARL | 1 + ...Variable Selection with Theoretical Guarantees | 1 + ...mble Bayesian Model for Robust Neural Decoding | 1 + ...iational Inference in Bayesian Neural Networks | 1 + .../Tractable Optimality in Episodic Latent MABs | 1 + ...in Shapley-Fair Collaborative Machine Learning | 1 + ...ff Resource Budgets For Improved Regret Bounds | 1 + ...ry with Regularized Deterministic Autoencoders | 1 + ...ness, and Complexity in Emergent Communication | 1 + ...orks on the Sphere Can Happen in Three Regimes | 1 + ...ral Networks with Event-driven Backpropagation | 1 + ...ing Neural Networks with Local Tandem Learning | 1 + ...Training Subset Selection for Weak Supervision | 1 + ...e Classifiers with Conformalized Deep Learning | 1 + ... Any-Order Autoregressive Models the Right Way | 1 + ...els to follow instructions with human feedback | 1 + ...upralinear networks by dynamics-neutral growth | 1 + ...Injected and Natural Backdoors During Training | 1 + ...nference via Mean-field Langevin in Path Space | 1 + ...lance: Improved credit assignment in GFlowNets | 1 + ... Saturation and Convergence in High Dimensions | 1 + ...tonomous Driving: A Simple yet Strong Baseline | 1 + ...t ImageNet Performance using Deep Transduction | 1 + ...ransferable Tabular Transformers Across Tables | 1 + ...entence Scoring with Sliding Language Modeling | 1 + ...eature Spaces for Treatment Effects Estimation | 1 + ...ion Shifts via Fair Consistency Regularization | 1 + ...entations with Cross-modal Similarity Matching | 1 + ...fficient Operator Learning in Frequency Domain | 1 + ...former Memory as a Differentiable Search Index | 1 + ...ent Reinforcement Learning with Action Parsing | 1 + .../Transformers from an Optimization Perspective | 1 + ...Attention with Data-Adaptive Sparsity and Cost | 1 + ...works with Directed Acyclic Graph Architecture | 1 + ...works with a Continuous Manifold of Attractors | 1 + ...apping Them into an Easy-to-Replace Subnetwork | 1 + ...Metrics and Stability of Graph Neural Networks | 1 + ...th known constraints over mixed-feature spaces | 1 + ...tive Neuron Morphology Representation Learning | 1 + ...ngulation candidates for Bayesian optimization | 1 + ...Estimation for Robust Generalized Linear Model | 0 .../Truly Deterministic Policy Optimization | 1 + ...ower Iteration for Differentiable DAG Learning | 1 + ...ble and hassle-free simulation-based inference | 1 + ...: Duality and Algorithm for Continuous Actions | 1 + data/2022/neurips/Trustworthy Monte Carlo | 0 ...Machine for Solving Contextual Bandit Problems | 1 + ..., CEs and CCEs with Neural Equilibrium Solvers | 1 + ...ed, Dynamic Tabular Datasets for ML Evaluation | 1 + ...End to End Entity Linking Benchmark for Tweets | 1 + ...-22: Towards Graph-Based Twitter Bot Detection | 1 + ... for Sign Language Recognition and Translation | 1 + ...ptimization guarantee in the mean-field regime | 1 + ...ible TinyML Models for Neural Processing Units | 1 + ...Neural Fields for Mix-and-Match Virtual Try-On | 1 + ...ubpopulation Shift via Uncertainty-Aware Mixup | 1 + ... Deep Classifiers trained via Conditional GANs | 1 + ...pervised Learning Benchmark for Classification | 1 + ...Approach for Vision with Learned Guiding Codes | 1 + ...ated Models Can Improve Human-AI Collaboration | 1 + ...el Dynamics for Offline Reinforcement Learning | 1 + ...ew Data: The Power of Seeing the Whole Picture | 1 + ... Incremental Implicitly-Refined Classification | 1 + ...isk-Sensitive Player Evaluation in Sports Game | 1 + ...with O(log T) Swap Regret in Multiplayer Games | 1 + ...uctural Fairness in Graph Contrastive Learning | 1 + ...hoto Critique Dataset for Aesthetic Assessment | 1 + ...gn Overfitting in Gradient-Based Meta Learning | 1 + ...d on Domain Similarity and Few-Shot Difficulty | 1 + ...tive Learning via Coordinate-wise Optimization | 1 + ...al Computing for Parallel Single-Pass Learning | 1 + ...tworks from the Bayesian-Inference Perspective | 1 + ...upervision via Source-aware Influence Function | 1 + ...hrough the Lens of Representation Similarities | 1 + ...ng Overparametrized Neural Network Classifiers | 1 + ...g Subgraph GNNs by Rethinking Their Symmetries | 1 + ...mers through Patch-based Negative Augmentation | 1 + .../neurips/Understanding the Eluder Dimension | 1 + ... Linear Regions in Deep Reinforcement Learning | 1 + ...of Batch Normalization for Transformers in NLP | 1 + ...t of Normalization Layers: Sharpness Reduction | 1 + ...c 2D Prediction for Multi-Stage Classification | 1 + ...Sparse Generalist Models with Conditional MoEs | 1 + ...rk for Contrastive Language-Image Pre-training | 1 + ...ode Collapse in GANs using a Uniform Generator | 1 + ...fied Inference in Sequential Decision Problems | 1 + ...port Framework for Universal Domain Adaptation | 1 + ...ation with Transformer for 3D Object Detection | 1 + ...Based Training-Free Neural Architecture Search | 1 + .../Universal Rates for Interactive Learning | 0 ... Networks Based on Ridgelet Analysis on Groups | 1 + ...nication in Multi-Agent Reinforcement Learning | 1 + ...sarial Learning for Open-Set Domain Adaptation | 1 + ...tersection-Closed Classes and Extremal Classes | 1 + ...its of Reward Engineering on Sample Complexity | 1 + ...ed Graph Neural Networks for Dynamical Systems | 1 + ...rom Repeated Traversals for Autonomous Driving | 1 + ...ised Causal Generative Understanding of Images | 1 + ...Task Generalization via Retrieval Augmentation | 1 + ...Semantic Segmentation using Depth Distribution | 1 + ...anslation with Density Changing Regularization | 0 ...m Incomplete Measurements for Inverse Problems | 1 + ...imization with Principled Objective Relaxation | 1 + ...arning of Equivariant Structure from Sequences | 1 + ...roup Invariant and Equivariant Representations | 1 + ... Shape Programs with Repeatable Implicit Parts | 1 + ...Unsupervised Learning under Latent Label Shift | 1 + ...ntation by Predicting Probable Motion Patterns | 1 + ... Segmentation Using Radiance Field Propagation | 1 + ...Object Priors Generation and Detector Learning | 0 ...Translation and Rotation Group Equivariant VAE | 1 + ...by Generative Adversarial Autoencoding Network | 1 + ...nt Learning with Contrastive Intrinsic Control | 0 ...rom Pre-trained Diffusion Probabilistic Models | 1 + ...d Skill Discovery via Recurrent Skill Training | 1 + ... via Mutual Information Regularized Assignment | 1 + ...less and Stealthy Dataset Copyright Protection | 1 + data/2022/neurips/Uplifting Bandits | 1 + ...rounded Simulations for Explanation Evaluation | 1 + ...stimation of Peer Influence in Social Networks | 1 + ...rove Accuracy & Out-of-Distribution Robustness | 0 ...artial Monotonicity in Submodular Maximization | 1 + ... to instill human inductive biases in machines | 1 + ...toencoders and Probabilistic Logic Programming | 1 + .../neurips/VCT: A Video Compression Transformer | 1 + ...rgence of Navigation in Embodied Rearrangement | 1 + ... Federated Learning, Efficiently and Securely? | 1 + ...: Variational Interpretable Concept Embeddings | 1 + ...f-Supervised Learning of Local Visual Features | 1 + ...ance Segmentation via Object Token Association | 1 + ...Benchmark for Vision-and-Language Manipulation | 1 + ... Pre-Training with Mixture-of-Modality-Experts | 1 + ...amework for Visual Deep Reinforcement Learning | 1 + ...rmer Compression with Low-Frequency Components | 1 + ...tional Inference Based Algorithm for Phylogeny | 1 + ...rative Design of Reinforcement Learning Agents | 1 + ...CPC leads to acoustic unit discovery in speech | 1 + ..., Theory and Application to Federated Learning | 1 + ...Perturbation for Source-Free Domain Adaptation | 1 + ...ional inference via Wasserstein gradient flows | 1 + ...for Rotation Equivariant Geometry Optimization | 1 + ...rk for Authorship Verification on the Dark Web | 1 + ...fication and search algorithms for causal DAGs | 1 + ...raph Neural Network for Circuit Representation | 1 + ...oNS: Visual Search in Natural Scenes Benchmark | 0 ...ransformer Baselines for Human Pose Estimation | 1 + data/2022/neurips/Video Diffusion Models | 1 + ...ing to Act by Watching Unlabeled Online Videos | 1 + ...chmark of learning-based video-quality metrics | 1 + ...ject Interaction Detection from Tubelet Tokens | 1 + ...earners for Self-Supervised Video Pre-Training | 1 + ...f Visual Recognition to Adversarial Viewpoints | 1 + ...Reconstruction with Viscosity and Coarea Grids | 1 + ...ion with Right-for-the-Right-Reason Objectives | 1 + .../Vision GNN: An Image is Worth Graph of Nodes | 1 + ... Transformers provably learn spatial structure | 1 + ...age Foundations for Image Paragraph Captioning | 1 + data/2022/neurips/Visual Concepts Tokenization | 1 + .../neurips/Visual Prompting via Image Inpainting | 1 + ...prove AI robustness and human-AI team accuracy | 1 + ...Adversarial Attacks with Audio-to-Audio Models | 1 + ...-Aware Image Synthesis with Sparse Voxel Grids | 1 + ...indow-based Transformers for Multi-view Stereo | 1 + ...n The (Im)possibility of Fairwashing Detection | 0 ...means for clustering probability distributions | 1 + ...n Iterative Networks for Barycenter Estimation | 1 + ...rstein Logistic Regression with Mixed Features | 1 + ...Watermarking for Out-of-distribution Detection | 1 + ...rror Bounds for Stable Time Series Forecasting | 1 + ...ature Maps Compression for Image-to-Image CNNs | 1 + .../Wavelet Score-Based Generative Modeling | 1 + ...ntic Segmentation via Dual Similarity Transfer | 1 + ...resentation Learning with Sparse Perturbations | 1 + ...akly supervised causal representation learning | 1 + ...ap Learning for Vision-and-Language Navigation | 1 + ... Web Interaction with Grounded Language Agents | 1 + .../Weighted Distillation with Unlabeled Examples | 1 + ...arning with Diversity-Driven Model Compression | 1 + ...d improving Shapley based feature attributions | 1 + ...eman Go Walking: Random Walk Kernels Revisited | 1 + ...ntext? A Case Study of Simple Function Classes | 1 + ...t Kernel Tell Us About Adversarial Robustness? | 1 + ...valuation Framework for Explainability Methods | 1 + ...hat Makes Graph Neural Networks Miscalibrated? | 1 + ...edge Distillation - A Statistical Perspective" | 1 + ...e is What You Classify: Black Box Attributions | 1 + ...eep Learning via Distributional Generalization | 1 + ... Systems? New Perspectives on NLP Benchmarking | 1 + ...pen-World Phrase-Grounding without Text Inputs | 1 + ...c to Study Generalization of Minimax Learners? | 1 + ... the Fraction Negatively Affected by Treatment | 1 + ...formers: Recipes from Training to Architecture | 1 + ...l Thompson Sampling meets Approximation Regret | 1 + .../neurips/When Do Flat Minima Optimizers Work? | 1 + ...rivate Learning Not Suffer in High Dimensions? | 1 + ...ariant Learning Survive Spurious Correlations? | 1 + ...ned Analysis of Differentially Private Bandits | 1 + ... are Local Queries Useful for Robust Learning? | 1 + ...ine Two-Player Zero-Sum Markov Games Solvable? | 1 + ...? Analyzing the remaining mistakes on ImageNet | 1 + ...rning work for offline reinforcement learning? | 1 + ...rventions in Autonomous Reinforcement Learning | 1 + ...imal Intervention Policies for Critical Events | 1 + ...age Models as Accounts of Human Moral Judgment | 1 + ...brid Offline-and-Online Reinforcement Learning | 1 + ...Constrained Model-based Reinforcement Learning | 1 + ...rameter-Space Saliency Maps for Explainability | 1 + ...tion in Sparse Training for Feature Selection? | 1 + ...orative Perception via Spatial Confidence Maps | 1 + ...ective to Characterizing Post Hoc Explanations | 1 + ...gence Rate of Coupling-based Normalizing Flows | 1 + ...lly Generated Data Help Adversarial Robustness | 0 ... is Difficult: Perspective of Expressive Power | 1 + ... Ensembles, and Why Their Independence Matters | 1 + ...trastive Learning? A Gradient-Bias Perspective | 1 + ...perform deep learning on typical tabular data? | 1 + ... The many regularizers of geometric complexity | 1 + ...rk of in-the-Wild Distribution Shift over Time | 1 + .../Will Bilevel Optimizers Benefit from Loops | 1 + ...chmark to Challenge Vision-and-Language Models | 1 + ...ale Chinese Cross-modal Pre-training Benchmark | 1 + ...trained Transformers Made Simple and Efficient | 1 + ...ormers and RvS Fail in Stochastic Environments | 1 + ...ation via Bank-constrained Manifold Projection | 0 ... Live Once: Single-Life Reinforcement Learning | 1 + ...f-Distribution Detection Method is Not Robust! | 1 + ...ansformer May Not be as Powerful as You Expect | 1 + ...er Optimization for Neural Architecture Search | 1 + ...earn Invariance Without Environment Partition? | 1 + ...al Navigation using Multimodal Goal Embeddings | 1 + ...hot 3D Drug Design by Sketching and Generating | 1 + ...ering via Frozen Bidirectional Language Models | 1 + .../neurips/Zero-Sum Stochastic Stackelberg Games | 1 + ...ogeneous Graph via Knowledge Transfer Networks | 1 + ... Recognition and Acquisition at Inference Time | 1 + ...ning Quantization for Large-Scale Transformers | 1 + ...d-Thresholding: Gradient Error vs. Expansivity | 1 + ...ding: Escaping Saddle Points without Gradients | 1 + ...ins for Lagrangian Neural Network Verification | 1 + ...del Zoo for Out-of-Distribution Generalization | 1 + data/2022/neurips/coVariance Neural Networks | 1 + ...aset using mmWave, RGB-D, and Inertial Sensors | 1 + ... Benchmark for Personalized Federated Learning | 1 + ...r training deep networks with unitary matrices | 1 + ...k Deep Learning based Knowledge Tracing Models | 1 + ...g And Zero-Shot Transfer to Unlabeled Modality | 1 + ...ctivity Using Synthetic Aperture Radar Imagery | 1 + ...Scale Embodied AI Using Procedural Generation" | 1 + ...Automatically Terminate Bayesian Optimization" | 1 + ...zation: A unified approach to private training | 1 + ...g with Bidirectional Communication Compression | 1 + ...D molecule generation by denoising voxel grids | 1 + ...ecting the 3D World into Large Language Models | 1 + .../neurips/4D Panoptic Scene Graph Generation | 1 + ...ing Training Data Attribution In Deep Learning | 1 + ...y Estimation for Computerized Adaptive Testing | 1 + ...s-Moment Approach for Causal Effect Estimation | 1 + ...ed Class Incremental Learning for Vision Tasks | 1 + ...reaming Media Performance over HTTP 3 Browsers | 1 + ...iffusion-Model of Joint Interactive Navigation | 1 + ...tem View of Langevin-Based Non-Convex Sampling | 1 + ...ependent Learning in Zero-Sum Stochastic Games | 1 + ...tiparameter Persistent Homology Decompositions | 1 + ... Robust G-Invariance in G-Equivariant Networks | 1 + ...Correct, Incorrect, and Extrinsic Equivariance | 1 + ...t Extraction and Concept Importance Estimation | 1 + ...antic Similarity Dataset of Historical English | 1 + ... Measure-Theoretic Axiomatisation of Causality | 1 + ... Posterior Sampling in Image Recovery Problems | 1 + ...ing the Projection Robust Wasserstein Distance | 1 + ...m for the Euclidean Bipartite Matching Problem | 1 + ...twork for DSIC Affine Maximizer Auction Design | 1 + ...o Find a Causal Order in Additive Noise Models | 1 + ...ty Bounds for Constrained Minimax Optimization | 1 + ...variance Estimation in Relative Frobenius Norm | 1 + ...timization in Linear Markov Decision Processes | 1 + ...of Skills based on Determinantal Point Process | 1 + ...ery in Nonlinear Generative Compressed Sensing | 1 + ...Model and Dimension for Interactive Estimation | 1 + ...lable Framework for Neural Population Decoding | 1 + ...on: Game Dynamics for Multi-Objective Learning | 1 + ... optimize time-space tradeoff for large models | 0 ...aphon-signal analysis of graph neural networks | 1 + ...xpressive 3D equivariant graph neural networks | 1 + ...r learning to represent visual transformations | 1 + ...or information-theoretic generalization bounds | 1 + ...Gym: Design Choices for Deep Anomaly Detection | 1 + ...Gradient Difference for Preconditioning Matrix | 1 + ...Control in High-Dimensional Mediation Analysis | 1 + ...ral Programming with Interactive Decomposition | 1 + ...nsor Networks for Quantum Many-Body Simulation | 1 + ...Benchmarking Tool for Label Quality Assessment | 1 + ...Regressive Diffusion Model for Text Generation | 1 + ...r Advancing Isolated Sign Language Recognition | 1 + ...icient Parallelization of Deep Neural Networks | 1 + ...hrough Memory Efficient Attention Manipulation | 1 + ...wing Instructions with Latent Diffusion Models | 1 + ...mation Seeking with Large Language Model Agent | 1 + ... Claim Verification with Evidence from the Web | 1 + ... generation of in-vitro functioning antibodies | 1 + ...ining with Module-Wise Descending Asynchronism | 0 ...ex Optimization Problem with Infinite Variance | 1 + ...lerating Motion Planning via Optimal Transport | 1 + ...r Dimensions for Unsupervised Word Translation | 1 + ...ization with Multimodal Unified Representation | 1 + ...t Imitation from Observation with World Models | 1 + ...nt Learning under Limited Visual Observability | 1 + ...neral task space with applications in robotics | 1 + ...vity Grammars for Temporal Action Segmentation | 1 + ...ANNS: A Framework for Adaptive Semantic Search | 1 + ...ors for Data-Efficient Complex Query Answering | 1 + ...ent Regression with Applications to Panel Data | 1 + ...vacy Composition for Accuracy-first Mechanisms | 1 + ...st-Time Personalization for Federated Learning | 1 + ...omputation scaling to unseen difficulty levels | 1 + ...d Thin: Diffusion for Temporal Point Processes | 1 + ...dressing Negative Transfer in Diffusion Models | 1 + ...rial Counterfactual Environment Model Learning | 14 ++++++++++++++ ... Networks for Low Dimensional Linear Subspaces | 1 + ...d Generalization for Gradual Domain Adaptation | 1 + ...versarial Training from Mean Field Perspective | 0 ...ount Tracking via Partial Differential Privacy | 1 + ... under Budget Constraint for Online Algorithms | 0 data/2023/neurips/Affinity-Aware Graph Networks | 1 + ...atter Dataset From Delhi For ML based Modeling | 1 + ... Stationary Distribution Correction Estimation | 1 + ...ibution Alignment for Zero-Shot Generalization | 1 + ...nd Hessian for Neural Signed Distance Function | 1 + ...with Human Preferences via a Bayesian Approach | 1 + ...resentations supports robust few-shot learning | 1 + ...lignment for Weakly-supervised 3D Segmentation | 1 + ...f-Experts for Integrated Multimodal Perception | 1 + ...Alternating Updates for Efficient Transformers | 1 + ...makes the adversary weaker in two-player games | 1 + ...red Text Dataset of Historical U.S. Newspapers | 1 + ...Scalable Variational Inference for Latent SDEs | 1 + ...under the Polyak-\305\201ojasiewicz Condition" | 0 ...ugin and Its Application to Continual Learning | 0 ...roach to Best of Both Worlds in Linear Bandits | 1 + ...racle-Efficient Adversarial Contextual Bandits | 1 + ...n Networks: Approximating Equitable Partitions | 1 + ...content of communication between brain regions | 1 + ...n of Neural Networks through Loss Path Kernels | 1 + ...f Self-Supervised Image Reconstruction Methods | 1 + data/2023/neurips/Anchor Data Augmentation | 1 + ...d Copy-Robust Delegations for Liquid Democracy | 1 + .../Anytime Model Selection in Linear Bandits | 1 + ...itive Reinforcement Learning with Policy Prior | 1 + ...ral Causal Bandits with Unobserved Confounders | 1 + ...nference of marginals using the IBIA framework | 1 + ...iffusion Models Vision-And-Language Reasoners? | 1 + ... More Data Hungry Than Newborn Visual Systems? | 1 + ...Blind Omnidirectional Image Quality Assessment | 1 + data/2023/neurips/Auditing for Human Expertise | 1 + ...gmenting Language Models with Long-Term Memory | 1 + ...e Translation for Daily Communication and News | 1 + ...raph Optimization for Neural Network Evolution | 1 + .../Autodecoding Latent 3D Diffusion Models | 1 + ...and Benchmarking Agents that Solve Fuzzy Tasks | 1 + ...ransformer for Cross-data Learning in the Wild | 0 ...or Efficient Neural Combinatorial Optimization | 1 + ...ing Modal Transformation for Data Augmentation | 0 ...d-Bandit Strategy for Automated Phased Release | 0 ...loud Segmentation with Inter-Part Equivariance | 1 + ...t Task Assignment with Unknown Processing Time | 0 .../BanditPAM++: Faster k-medoids Clustering | 1 + ...casting with Learnable and Interpretable Basis | 1 + ...esTune: Bayesian Sparse Deep Model Fine-tuning | 1 + ...usal Discovery with Multi-Fidelity Experiments | 0 .../Bayesian Learning via Q-Exponential Process | 1 + ... Uncertainty Quantification in Image Retrieval | 1 + ...-Averse Q-Learning with Streaming Observations | 1 + ...s for analyzing neural spike train variability | 1 + ...lignment of LLM via a Human-Preference Dataset | 1 + ...tion Models with Language-Model-as-an-Examiner | 1 + ...ssion Through Better Private Feature Selection | 1 + ...ntralized Learning via Finite-time Convergence | 1 + ...abeled Data through Holistic Predictive Trends | 1 + ...luation Processes: An Optimization-Based Model | 1 + ...on Algorithms for the Submodular Cover Problem | 1 + ...nsional Mechanism Design with Side Information | 1 + .../Bifurcations and loss jumps in RNN training | 1 + ...mation using Multi-modal Satellite Time-series | 0 ...Recovery: A Novel Benchmark Dataset and Method | 1 + ...ck-Box Differential Privacy for Interactive ML | 1 + ...ug-and-Play Methods for Blind Inverse Problems | 1 + data/2023/neurips/Block-State Transformers | 1 + ...Transferability by Achieving Flat Local Maxima | 1 + ...ta via Kernel Correction and Affinity Learning | 0 ...tor for Offline Design of Biological Sequences | 1 + ...ng-Free Semantic Control with Diffusion Models | 1 + ...ounding training data reconstruction in DP-SGD | 1 + ... Discovery using Large Scale Generative Models | 1 + ...-Accuracy Tradeoff with f-Differential Privacy | 1 + ... 3D Scene Understanding with Foundation Models | 1 + ... localization from dense multielectrode probes | 1 + ...-Optimal Adversarial Linear Contextual Bandits | 1 + ... Factors under an Inductive Bias of Confounder | 1 + ...uning for Highly-Accurate Sparse Vision Models | 1 + ...ynamics Under Temporal Environmental Variation | 1 + ...in Space and Time for Video Action Recognition | 1 + ...EIL: Generalized Contextual Imitation Learning | 1 + ... channel-adaptive models in microscopy imaging | 1 + ...d Counterfactual Examples for Image-Text Pairs | 1 + ...Benchmark for Continual Reinforcement Learning | 1 + ...ed Deep Offline Reinforcement Learning Library | 1 + ...orcement Learning with a Quantized World Model | 1 + ...ollable, Robust and Secure Image Steganography | 1 + ...er with Continuously Weighted Contrastive Loss | 1 + .../Cal-DETR: Calibrated Detection Transformer | 1 + ... GNN Over Multi-Relational and Temporal Graphs | 1 + ...Matching: Trainable Kernel Calibration Metrics | 1 + ...for Generating Camoflauged Adversarial Patches | 1 + ...ion Frequency in Delayed Feedback Environments | 0 ...terventional data across multiple environments | 1 + ...iven Augmentations for Text OOD Generalization | 1 + ... of Causal Graphs with Small Conditioning Sets | 1 + ...verfitting in Robust Multiclass Classification | 0 ...on Shift: A Model Weight Perturbation Approach | 1 + ...rarchical Comparisons for Image Classification | 1 + ...e Instruction Tuning for Large Language Models | 1 + ... Metrics for Autoregressive Transformer Models | 0 ...ounding Dataset on City-scale Point Cloud Data | 1 + ...itional Conformal Prediction with Many Classes | 1 + ... Circuits: Parallel Environments and Benchmark | 1 + ...hine Learning for Weather and Climate Modeling | 1 + ...omer: Clustering As A Universal Visual Learner | 0 ...etch: Dynamic Compression for Embedding Tables | 1 + ...gnment for Open-vocabulary 3D Object Detection | 1 + ...Alignment for Open-Vocabulary Object Detection | 1 + ... for Communication-Efficient Private Inference | 1 + ...via Long-Range Modulatory Feedback Connections | 1 + .../neurips/Collaborative Alignment of NLP Models | 0 ...ollaborative Learning via Prediction Consensus | 1 + ...ore Distillation for Consistent Visual Editing | 0 ...ing Linear Models with Structured Missing Data | 1 + ...Collapsed Inference for Bayesian Deep Learning | 1 + ...c Video Representations with Dynamic Codebooks | 0 ...s, Structural Models, Graphs, and Abstractions | 1 + ...ing and Self-Training Under Distribution Shift | 1 + ...ing Neural Networks: Smoothness and Degeneracy | 1 + ...Score-Based) Text-Controlled Generative Models | 1 + ...an-Centered Explanations for Model Improvement | 1 + ...fficient Transfer Learning with Fast Inference | 1 + ...gled Representations in Reinforcement Learning | 1 + ...y-Aware Planning with Diffusion Dynamics Model | 1 + ...arning and its Application to Minimax Theorems | 1 + .../Connecting Certified and Adversarial Training | 1 + ... Estimation for Offline Reinforcement Learning | 1 + ...pic Gaussian Diffusion Model for Image Editing | 1 + ...uction for Offline Meta-Reinforcement Learning | 1 + ...fective Topic Modeling in Low-Resource Regimes | 1 + .../neurips/Context-lumpable stochastic bandits | 1 + ... Learning with Preference-Based Active Queries | 1 + .../Contextual Stochastic Bilevel Optimization | 1 + ...Continuous-Time Functional Diffusion Processes | 1 + ...pervised Halfspace Learning in Polynomial Time | 1 + ...ontrastive Sampling Chains in Diffusion Models | 0 ...xt-to-Image Diffusion by Orthogonal Finetuning | 1 + ...-Concave Zero-Sum Stochastic Stackelberg Games | 0 ...ators for robust and accurate learning of PDEs | 1 + ... Models for Long-Range Spatiotemporal Modeling | 1 + ...e-sets for Fair and Diverse Data Summarization | 1 + ...d Deep Neural Networks without Weight Symmetry | 1 + ...rrespondence Priors for Neural Radiance Fields | 1 + ...for Offline Multi-agent Reinforcement Learning | 1 + ... Evaluation of Peer-Review Assignment Policies | 1 + ...cal Surfaces by Diffeomorphic Mesh Deformation | 1 + .../Covariance-adaptive best arm identification | 1 + ...el Skill Hierarchies in Reinforcement Learning | 1 + ...g a Public Repository for Joining Private Data | 0 ...icy Adaptation via Value-Guided Data Filtering | 1 + ...nking the Debiased GNN from a Data Perspective | 1 + ...eural Network via Discrete Denoising Diffusion | 0 ...-DETR: Divide the Attention Layers and Conquer | 0 ...or visual understanding of mixture-of-datasets | 1 + ... and High-quality Speech-to-Speech Translation | 1 + ... for Vector Set Search with Vector Set Queries | 1 + ...e Replay in Multi-Agent Reinforcement Learning | 1 + ...ffusion Solvers for Combinatorial Optimization | 1 + .../DISCS: A Benchmark for Discrete Sampling | 1 + ...al Multi-Dataset Multi-Task Segmentation Model | 1 + .../neurips/Data Quality in Imitation Learning | 1 + ... for Language Models via Importance Resampling | 1 + ... Linear Systems with Unknown Noise Covariances | 1 + ...neration for Pixel-Level Semantic Segmentation | 1 + .../Debiasing Conditional Stochastic Optimization | 1 + ... Multi-armed Bandit with Heterogeneous Rewards | 1 + ...rcement Learning via Modular Generative Models | 1 + ...s and AIs on the Many Facets of Working Memory | 1 + ... Concept Learning For 3D Novel Class Discovery | 0 ...Subtasks in Multi-Agent Reinforcement Learning | 1 + ...eep Contract Design via Discontinuous Networks | 1 + .../neurips/Deep Fractional Fourier Transform | 0 ... Fields for Graph-Structured Dynamical Systems | 1 + ...ights into Noisy Pseudo Labeling on Graph Data | 1 + ...imal for the Deep Unconstrained Features Model | 1 + ...ced Ant Systems for Combinatorial Optimization | 1 + ...izing Sequential Operations in Neural Networks | 1 + ...Hand-Object Interaction via Physics Simulation | 1 + data/2023/neurips/Delegated Classification | 1 + ...n Graph Neural Networks: Can One Size Fit All? | 1 + ...) Promote Compositional Reasoning in VL Models | 1 + ...ric Learning for Monocular 3D Object Detection | 1 + ...ection with FDR control via conformal e-values | 1 + ...ing Object Detection with Flexible Expressions | 1 + ...Diffusion Transformers for 3D Shape Generation | 1 + ...ate Views using Neural Template Regularization | 0 ...o-Audio Synthesis with Latent Diffusion Models | 1 + ...ng Knowledge From Pre-trained Diffusion Models | 1 + ...ainst Diffusion-Based Adversarial Purification | 1 + ...Diffusion-based Generative 3D Shape Completion | 1 + ...allel Random Patch Diffusion in Histopathology | 1 + ... Trajectory with Diffusion Probabilistic Model | 1 + ...ble Clustering with Perturbed Spanning Forests | 1 + ...tiable sorting for censored time-to-event data | 0 ...gh \316\262-Divergence One Posterior Sampling" | 1 + ... Physics-Augmented Generative Diffusion Models | 1 + ...thesizer for Multi-Task Reinforcement Learning | 1 + ...elf-Guidance for Controllable Image Generation | 1 + ...ed Policy Optimization without Reward Modeling | 1 + ...ts can control the sign of synaptic plasticity | 1 + ...entanglement of Diffusion Probabilistic Models | 1 + ...ding for Multi-Instance Partial-Label Learning | 1 + ...Algorithms with Adversarial Environment Design | 1 + ...einforcement Learning via Contrastive Learning | 1 + ...-Temporal Logic Rules to Explain Human Actions | 1 + ...in Autoencoder for T-Cell Receptor Engineering | 1 + ...gnitive Diagnosis with Limited Exercise Labels | 0 ... with Self-Supervision for Speaker Recognition | 1 + ...ustness from Vision-Language Foundation Models | 1 + ...ing of Large Language Models Over The Internet | 1 + ...buted Personalized Empirical Risk Minimization | 1 + ...ets Towards Calibrated Sparse Network Training | 1 + ...tyle Modulated Generative Adversarial Networks | 1 + ...bution Detection via Informative Extrapolation | 1 + ...ts with Automatic Diffusion-based Augmentation | 1 + ...to-Image Alignment with Iterative VQA Feedback | 1 + ... Mixtures Speeds Up Language Model Pretraining | 1 + ...via Environment Augmentation Learn Invariance? | 1 + ...ork training problem always admit an optimum ? | 1 + ...ut Learning due to Gradients and Cross Entropy | 1 + ...gnitude! Your mask topology is a secret weapon | 0 ... Generic Algorithm and Robust Partial Coverage | 1 + ...th Dimension-Independent Convergence Guarantee | 1 + ...lication to High-Dimensional Synthetic Control | 1 + ...ented Transfer for Meta-Reinforcement Learning | 1 + ...ble Multivariate Time Series Anomaly Detection | 1 + ...out Individual Global Max for Cooperative MARL | 1 + ...nce of Gene Regulatory Networks with GFlowNets | 1 + ...Point: Dynamic Neural Point For View Synthesis | 1 + ... and Interpretable Autoregressive Transformers | 1 + ...ed Learning with Adaptive Differential Privacy | 1 + ...amic Regret of Adversarial Linear Mixture MDPs | 1 + ...mic Sparsity Is Channel-Level Sparsity Learner | 1 + ...e Input Views with Monocular Depth Adaptation" | 1 + ...n with Spatio-Temporal Representation Learning | 1 + ...wering Dataset Combined With Electrocardiogram | 1 + ...nt Diffusion for Planning with Embodied Agents | 1 + ...l Waveform Inversion of Geophysical Properties | 1 + ...k for Few-Shot Evaluation of Foundation Models | 1 + .../ELDEN: Exploration via Local Dependencies | 1 + .../neurips/Easy Learning from Label Proportions | 1 + ...scedastic Regression with Deep Neural Networks | 1 + ...Shifts for Models with Different Training Data | 1 + ...on Sets in Hierarchical Reinforcement Learning | 0 ...earning via Robustness-Aware Coreset Selection | 1 + ...lized Linear Bandits with Heavy-tailed Rewards | 1 + ...Extrapolation using Prior-Data Fitted Networks | 1 + ...on Policies For Offline Reinforcement Learning | 1 + ...ear Graph Neural Networks via Node Subsampling | 0 ...icient Model-Free Exploration in Low-Rank MDPs | 1 + ...ning using Inverse Dynamic Bisimulation Metric | 0 ...an Optimization for Arbitrary Uncertain inputs | 1 + ...l Equations with Positive Semi-Definite Models | 1 + ...th Second-Order Degradation and Reconstruction | 1 + ...duction for Over-Parameterized Neural Networks | 1 + ...ions into geometric deep learning force fields | 1 + ...term Egocentric Visual Object Tracking Dataset | 1 + ...a Abnormal Adversarial Examples Regularization | 1 + ...al Neural Networks through Activation Sparsity | 1 + .../Emergent Communication for Rules Reasoning | 1 + ...volutional Neural Nets with MetaSin Activation | 0 ...ls for Inverse Problems in High Energy Physics | 1 + ...ntext Update in Text-to-Image Diffusion Models | 1 + .../Energy-Efficient Scheduling with Predictions | 1 + ...Incremental Learning with Data-free Subnetwork | 1 + ...urring in High-Speed Scenes with Spike Streams | 0 ...ware Optimization Through Variance Suppression | 1 + ...ural Optimal Transport via Diffusion Processes | 1 + ... Methods for Scalable Neural Implicit Samplers | 1 + ...k Learning with Heterogeneous Neural Processes | 1 + ...ual Opportunity of Coverage in Fair Regression | 1 + ...tation Learning from Imbalanced Demonstrations | 0 ... a Combination of Observations and Experiments | 0 ...o provably learn large scale dynamical systems | 1 + ...ric with Noise-Contaminated Intrinsic Distance | 0 ...Planning in Large Language Models with CogEval | 1 + ...ng Neuron Interpretation Methods of NLP Models | 1 + data/2023/neurips/Evaluating Open-QA Evaluation | 1 + ... Graph Neural Networks via Robustness Analysis | 0 ... Models Under Structural Distributional Shifts | 1 + ...rvised Learning for Molecular Graph Embeddings | 1 + ...Augmented Computation-Intensive Math Reasoning | 1 + ...hrough Explanation Invariance and Equivariance | 1 + ...g Small-Scale Datasets with Guided Imagination | 1 + ...erimental Designs for Heteroskedastic Variance | 1 + ...ed Auxiliary Feedback in Parameterized Bandits | 1 + ...nvex games for convergence to Nash equilibrium | 1 + ...d Training Strategy in Spiking Neural Networks | 1 + ...oring Question Decomposition for Zero-Shot VQA | 1 + ...Algorithms for Supervised Matrix Factorization | 1 + ...tion Glitches with Flip-Flop Language Modeling | 1 + ...and their unfair treatment of diffusion models | 1 + ...riant Networks for Spectral Geometric Learning | 1 + .../Expressivity-Preserving GNN Simulation | 0 .../FAMO: Fast Adaptive Multitask Optimization | 1 + ...uation of Open-Domain Text-to-Video Generation | 1 + ...chmark for Evaluating Interpretability Methods | 1 + ... Algorithm for Multinomial Logistic Regression | 0 ... in Federated Learning using Invariant Dropout | 1 + ...ld Model Backbones: RNNs, Transformers, and S4 | 1 + ...ussian Process Optimization with Regret Bounds | 1 + data/2023/neurips/Fair Graph Distillation | 1 + ...nalysis: Statistical and Algorithmic Viewpoint | 1 + ...lexible and Controllable Optimization Approach | 0 ...Scene Understanding in Open-World Environments | 1 + ...te: Limits of Transformers on Compositionality | 1 + ...ry Proportion control for aggregated Knockoffs | 1 + ...milarity Graphs with Kernel Density Estimation | 1 + .../Fast Model DeBias with Machine Unlearning | 1 + ...ision Matrix Estimation under Total Positivity | 1 + ... Best Order Score Search and Grow Shrink Trees | 1 + ...e Convex Optimization via Second-Order Methods | 1 + ...nimization with Predictions: The M-Convex Case | 1 + ...aster approximate subgraph counts with privacy | 1 + ...g for Interpretable, Performant Decision Trees | 1 + ...eralization of Generative Models Using Samples | 1 + ... Selection in the Contrastive Analysis Setting | 1 + ...Learning with Self-Adjusting Gradient Balancer | 1 + ...rated Training of Graph Convolutional Networks | 1 + ...ation with Normalized Annealing Regularization | 1 + ...Linear Bandits with Finite Adversarial Actions | 1 + ...nima Efficiently in Decentralized Optimization | 0 ...fe Zones of Markov Decision Processes Policies | 1 + ...Using a Correlation-Aware Homography Estimator | 1 + data/2023/neurips/Fine-Grained Visual Prompting | 1 + ...tic Guarantees for Treatment Effect Estimation | 1 + .../neurips/Flat Seeking Bayesian Neural Networks | 1 + ...vate Prompt Learning for Large Language Models | 1 + .../Flow Factorized Representation Learning | 1 + ...Infrastructure Cooperative 3D Object Detection | 1 + ...: Per-instance Personalized Federated Learning | 1 + ...us Your Attention when Few-Shot Classification | 0 ...l Mining Transformer for Few-Shot Segmentation | 1 + ...g Contextual Bandits via Post-serving Contexts | 1 + ...ete Probability Flow Through Optimal Transport | 1 + ...ries Forecasting from a Pure Graph Perspective | 1 + ...D Hand Representation Using Fourier Query Flow | 1 + ...orial and Mixed-variable Bayesian Optimization | 1 + .../Frequency Domain-Based Dataset Distillation | 1 + ... Effective Learners in Time Series Forecasting | 1 + ... Language Models to Pre-trained Machine Reader | 1 + ...e Negative Depth to Edge Heterophily in Graphs | 1 + ...Protein Pocket Design via Iterative Refinement | 1 + ...ian Pseudocoreset for Bayesian Neural Networks | 1 + ...ivity of Reducible Hyperbolic Tangent Networks | 1 + ... Abnormality for Out-of-distribution Detection | 1 + ... A Library for Gaussian Processes in Chemistry | 1 + ...r Instantaneous Graph Learning Model Selection | 1 + ...deo Generation via GLOBal Guided Video DecodER | 1 + .../2023/neurips/GMSF: Global Matching Scene Flow | 1 + ...rk For Interpreting Artificial Neural Networks | 1 + ...ining of Spatio-Temporal Graph Neural Networks | 1 + ...nguage Model to Use Tools via Self-instruction | 1 + ...ner Common Information Variational Autoencoder | 1 + .../neurips/Gaussian Membership Inference Privacy | 1 + ...ction and Application to High-dimensional Data | 1 + ...ess Probes (GPP) for Uncertainty-Aware Probing | 1 + ...ale Benchmark for Detecting AI-Generated Image | 1 + ... Surface Reconstruction from Multi-View Images | 1 + ...tific Simulators via Amortized Cost Estimation | 1 + data/2023/neurips/Generalized Belief Transport | 1 + ...ls by Removing Label Bias in Foundation Models | 1 + ...ted Path Consistency for Mastering Atari Games | 1 + ...s between subsampling and ridge regularization | 1 + ...formance in extreme multi-label classification | 1 + ... Diverse Policies with Latent Diffusion Models | 1 + ...ar SDEs with Additive and Multiplicative Noise | 1 + ...erse Evaluation Dataset for Object Recognition | 1 + ...desic Multi-Modal Mixup for Robust Fine-Tuning | 1 + ...ometric Analysis of Matrix Sensing over Graphs | 1 + ...er Neural Network without Overparameterization | 0 ...onary Learning via Matrix Volume Optimization" | 0 ...fusion Process for Low-light Image Enhancement | 1 + ...t-Based Feature Learning under Structured Data | 1 + ...Language Generation with Large Language Models | 1 + ...vised learning of latent temporal dependencies | 1 + ...Datasets and Evaluations for Accident Analysis | 1 + ...ann Manifold Flows for Stable Shape Generation | 1 + ...ation with Grounded Models for Embodied Agents | 1 + data/2023/neurips/Group Fairness in Peer Review | 1 + ... in Federated Learning with Pre-Trained Models | 1 + ...med Representations for Dexterous Manipulation | 1 + ...Comprehensive Assembly Knowledge Understanding | 1 + ...Human Handover Dataset with Large Object Count | 1 + ...ce Properties of Text-Guided Image Classifiers | 1 + ...the power of choices in decision tree learning | 1 + ...rarchical Encoding-based Neural Representation | 1 + ...l Learning: Rethinking Obscured Sub-optimality | 1 + .../Hierarchical Multi-Agent Skill Discovery | 0 ...th Application to Diffusion Model Acceleration | 1 + ...odel Evaluation with Conditional Randomization | 1 + ...ensional Asymptotics of Denoising Autoencoders | 1 + .../Holistic Evaluation of Text-to-Image Models | 1 + ... of NeuralODEs for accurate dynamics discovery | 1 + ... State of Instruction Tuning on Open Resources | 1 + .../How Re-sampling Helps for Long-Tail Learning? | 1 + ...ical abilities in a pre-trained language model | 1 + ...Model Shift and Model Bias Policy Optimization | 1 + data/2023/neurips/How to Scale Your EMA | 1 + ...wledge Graph Embeddings into Generative Models | 1 + ...ting via Hub Generation and Pin-hub Connection | 1 + ...uman-Guided Complexity-Controlled Abstractions | 1 + ...or Deep Stimulus Encoding in Visual Prostheses | 1 + ...nstructing Facial Skin-Spectra from RGB Images | 1 + ...oosts Fine-Grained Learning from Coarse Labels | 1 + ...ersible Backdoor Attacks in Federated Learning | 1 + ...es Inversely Correlated on Real-world Datasets | 1 + ...ntation Method for Training Robust Classifiers | 1 + ...oom and Spatial Biases in Image Classification | 1 + ...Human Preferences for Text-to-Image Generation | 1 + ...hesis with Scene Graph Hallucination Diffusion | 0 ...ion: Theoretical Justifications and Algorithms | 1 + .../Imitation Learning from Vague Feedback | 1 + ...r Logistic Regression at the Edge of Stability | 1 + .../Implicit Manifold Gaussian Process Regression | 1 + ...n in Over-Parameterized Support Vector Machine | 1 + ...onal Inference for High-Dimensional Posteriors | 1 + ...variance regularization in non-contrastive SSL | 1 + ...dits Using Tail Bounds for Martingale Mixtures | 1 + ...Yield Reduced Social Welfare Under Competition | 1 + ...General Regularizers and Multiple Optimal Arms | 1 + ...ipped Gradient Methods with Heavy Tailed Noise | 1 + ...mproving Robustness with Adaptive Weight Decay | 1 + ...presentations using human similarity judgments | 1 + ...or Calibrated and Consistent Learning to Defer | 1 + ...ls Large Language Models' Strengths and Biases | 1 + ...Context Learning Unlocked for Diffusion Models | 1 + .../Individual Arbitrariness and Group Fairness | 0 ...ferring Hybrid Neural Fluid Fields from Videos | 1 + .../Inferring the Future by Imagining the Past | 1 + ...amfer Distance Loss for Point Cloud Completion | 1 + ...ompt Tuning for Natural Language Understanding | 1 + ...eometry of the Retinal Representation Manifold | 1 + ...ine Approach for Partially Observable Problems | 1 + ...exity of Linear Predictors and Neural Networks | 1 + .../Inner Product-based Neural Network Similarity | 1 + ...r: Instruction-driven Physics-based Characters | 1 + ...ng Anybody in Diffusion Models via Celeb Basis | 1 + ...f Language Model with Sparse Human Supervision | 1 + ...Interactive Visual Reasoning under Uncertainty | 1 + ...Scale: Identifying Causal Mechanisms in Alpaca | 1 + ...e Prototype-based Graph Information Bottleneck | 1 + ...ion for Robust Detection of AI-Generated Texts | 1 + ...Probability of Sufficient and Necessary Causes | 1 + ...s Good Representations for Multitask Imitation | 1 + ...ent Learning with the Average Reward Criterion | 1 + ...ity Estimation for Safe Reinforcement Learning | 1 + ... Optimizing the Jaccard Index with Soft Labels | 1 + ...Jailbroken: How Does LLM Safety Training Fail? | 1 + ...ampling Based Conditional Independence Testing | 0 ...wledge Distiller for Any Teacher-Student Pairs | 1 + ...rnel Quadrature with Randomly Pivoted Cholesky | 1 + ... Sound Symbolism in Vision-and-Language Models | 1 + ... Efficient Low-Rank Permutation Representation | 1 + .../neurips/Knowledge Diffusion for Distillation | 1 + ...stillation Performs Partial Variance Reduction | 1 + ... Any-level Descriptions using Diffusion Priors | 0 ... Generalization Bounds and Confidence Boosting | 0 ... Perfect linear concept erasure in closed form | 1 + ...Knowledge Transfer for Lifelong Robot Learning | 1 + ...e Models in Text-to-Image Synthesis Evaluation | 1 + ...edical Imaging via Second-order Graph Matching | 1 + data/2023/neurips/Label Poisoning is All You Need | 1 + ...Model Inversion Attacks via Knowledge Transfer | 1 + ...iffusion Models for Learning from Noisy Labels | 1 + ... Lagrangian Fluid Mechanics Benchmarking Suite | 1 + data/2023/neurips/Langevin Quasi-Monte Carlo | 1 + ...Need: Aligning Perception with Language Models | 1 + ...ful Explanations in Chain-of-Thought Prompting | 1 + ...: Embodied Experiences Enhance Language Models | 1 + ...ers: Towards Unsupervised Text-Image Alignment | 1 + ... Spaces Improve Video Self-Supervised Learning | 1 + ... Semi-Parametric Reinforcement Learning Agents | 1 + ...guage Models are Visual Reasoning Coordinators | 1 + ...Language Models can Implement Policy Iteration | 1 + ...rk Dataset for Large-Scale Traffic Forecasting | 1 + ...dient Primal-Dual Methods for Constrained MDPs | 1 + .../neurips/Latent SDEs on Homogeneous Spaces | 1 + .../Latent exploration for Reinforcement Learning | 1 + ...ling - Defusing Neighborhood Explosion in GNNs | 1 + ...Self-Coding for Generalized Category Discovery | 1 + ...ensity Fields for Clean Cryo-ET Reconstruction | 1 + ...arning Causal Models under Independent Changes | 1 + .../neurips/Learning Cuts via Enumeration Oracles | 1 + .../Learning DAGs from Data with Few Root Causes | 1 + ...Highly-accurate Cross-view Camera Localization | 1 + ...ia Semipermeable Maximum Likelihood Estimation | 1 + .../2023/neurips/Learning Functional Transduction | 1 + ...ecognition Representations Without Real Humans | 1 + ...ecular Representation in Latent Discrete Space | 1 + ...Property Prediction via Graph Segment Training | 1 + ...raphical Models via Bridge-Block Decomposition | 1 + ...Neural Fields via Context Pruned Meta-Learning | 1 + ...l Domain-Invariant Representations for Ranking | 1 + ...ion Refinement for Unsupervised Face Animation | 1 + ...Free Bayesian Inference in Constrained Domains | 1 + ... Regularized Monotone Graphon Mean-Field Games | 1 + ...s Using An Intra-class Correlation Regularizer | 1 + ...ety Constraints from Multi-task Demonstrations | 1 + ...or Individual Neurons from Population Dynamics | 1 + ...ing Trajectories are Generalization Indicators | 1 + ...rsal Policies via Text-Guided Video Generation | 1 + ...rning Visual Prior via Generative Pre-Training | 1 + ...an Involvement through Proxy Value Propagation | 1 + ...rse Locations for Long-tailed Object Detection | 1 + ...vian Decision-Making from State-only Sequences | 1 + ...istributions for Out-of-distribution Detection | 1 + ...rning to Group Auxiliary Datasets for Molecule | 1 + ...Attributes for Open-set Fine-grained Retrieval | 1 + ...p: Intervention-Aware Concept Embedding Models | 1 + .../Learning to Taste: A Multimodal Wine Dataset | 1 + ... Ordering Alignment for Ordinal Classification | 1 + ...ize World Models for Model-based Task Planning | 1 + ... to demonstrate convergence of neural networks | 1 + data/2023/neurips/Lexinvariant Language Models | 1 + ...Linker Co-Design with 3D Equivariant Diffusion | 1 + ...al Lithography for Semiconductor Manufacturing | 1 + .../Lo-Hi: Practical ML Drug Discovery Benchmark | 1 + ...-of-Distribution Detection via Prompt Learning | 1 + ...rated Learning with Isolated Subspace Training | 1 + ...om Stationary Signals with Recovery Guarantees | 1 + data/2023/neurips/Logarithmic Bayes Regret Bounds | 1 + .../Long-Term Fairness with Unknown Dynamics | 1 + ... of Temporal Difference Reinforcement Learning | 1 + ... Compression with Conditional Diffusion Models | 1 + ...t Object Learning with Mutual Exclusivity Bias | 1 + ...Bounds on Adaptive Sensing for Matrix Recovery | 1 + ...e Repainting for Lighting-Realistic Generation | 1 + ...nt Cloud-Based Intra-Patient Lung Registration | 0 ...al of Machine Learning for Materials Discovery | 1 + ...Chinese Historical Document Analysis Benchmark | 1 + ...ngual And Document-Level Large Audited Dataset | 1 + ...on-byte Sequences with Multiscale Transformers | 1 + ... for Compact and Efficient Vision Transformers | 0 ... for Two-Sample Testing Without Data Splitting | 1 + ...pler for MicroMotion-based Gait Classification | 1 + ...achine learning detects terminal singularities | 1 + ...ent by Wire-Mask-Guided Black-Box Optimization | 1 + ...ny-body Approximation for Non-negative Tensors | 1 + ...ibutionally Equivalent Model Extraction Attack | 1 + ...ing for Efficient Dynamic Scene Reconstruction | 1 + ...ave a Role in Mathematical Architecture Design | 1 + data/2023/neurips/Max-Sliced Mutual Information | 1 + ...p Learning with Adversarial Ranking Robustness | 1 + ...Set: Self-Training through Dynamic Programming | 1 + ...ic Pre-Training for 3D Molecular Conformations | 1 + ...al Aggregation on Multi-Scaled Graph Hierarchy | 1 + ...ng Medical Prescriptions and Satellite Imagery | 1 + ... Optimizer with Momentum for Few-Shot Learning | 0 .../Meta-Learning with Neural Bandit Scheduler | 0 ...a-in-context learning in large language models | 1 + ...king networks using simulation-based inference | 1 + ...g and Enhancing In-Network Regular Expressions | 1 + ...Shape-Image-Text Aligned Latent Representation | 1 + ...kernels and neural networks in fixed dimension | 1 + ... Learning Environments for Goal-Oriented Tasks | 1 + ... of Evolving Tasks with Performance Guarantees | 1 + ...ization Guarantees for Representation Learning | 1 + .../Minimum-Risk Recalibration of Classifiers | 1 + ...nsformers via Regularized Nonlocal Functionals | 1 + ...gating Test-Time Bias for Fair Image Retrieval | 1 + ...Incidental Correlations on Part-based Learning | 1 + ... Filtering: A Dimensional Collapse Perspective | 0 ...FormerV2: Efficient Fully Transformer Tracking | 1 + ...arning to Train Transformers with Transformers | 1 + ...l Alignment on Causal and Moral Judgment Tasks | 1 + ...Learning with Meta-Learned Masked Auto-Encoder | 1 + ...et Weakly-Supervised Audio-Visual Event Parser | 1 + ...or Bayesian Neural Networks in Mutual Learning | 1 + ...odel-Based Control with Sparse Neural Dynamics | 1 + ...dient Methods: Theory and Practical Algorithms | 1 + ...erior Sampling via Learning Rate Randomization | 1 + ...ditive Mechanism Shift Variational Autoencoder | 1 + ...stillation for Multimodality Foundation Models | 1 + .../Moment Matching Denoising Gibbs Sampling | 1 + ...ive Video Moment Retrieval from Random to Real | 1 + .../Momentum Provably Improves Error Feedback! | 1 + ...A Simple Sub-Quadratic GEMM-Based Architecture | 1 + ...e Carlo Tree Search with Boltzmann Exploration | 1 + ... 3D Expressive Whole-body Human Motion Dataset | 1 + ...g with Heterogeneous Linear Contextual Bandits | 0 ...Sharper Convergence Rates with Task Similarity | 1 + ...or Multi-Source Unsupervised Domain Adaptation | 1 + ...vised Rigid Segmentation and Motion Estimation | 1 + ...lti-modal Queried Object Detection in the Wild | 1 + .../Multi-scale Diffusion Denoised Smoothing | 1 + ...g: Simple and Intuitive Weak Learning Criteria | 1 + ...ormality on Null Covariates in High-Dimensions | 1 + ...timation of Targeted Average Treatment Effects | 1 + ...ith High-Quality 3D Shape and Pose Annotations | 1 + ...rk for Deep Learning on non-Cartesian Lattices | 1 + ...forcement Learning with Function Approximation | 1 + ...Linear Time Algorithm for the Chamfer Distance | 0 ...timal k-Clustering in the Sliding Window Model | 1 + .../Nearest Neighbour with Bandit Feedback | 1 + .../Nearly Optimal Bounds for Cyclic Forgetting | 0 ...timal Decision Trees using Dynamic Programming | 1 + ...isson Compressed Sensing in the Olfactory Bulb | 1 + ...ontext Pretraining for Neural Spiking Activity | 1 + ...on of Latent Representations on Dynamic Scenes | 1 + data/2023/neurips/Neural Functional Transformers | 1 + ... with Neural Stochastic Differential Equations | 1 + ...eneralization, Robustness, and Spectral Biases | 1 + ...ral Lyapunov Control for Discrete-Time Systems | 1 + ...ed Learning Framework for Improved Reliability | 0 ...Neural Priming for Sample-Efficient Adaptation | 1 + data/2023/neurips/Neural Processes with Stability | 1 + ...rchical Exponential-family Energy-based Models | 1 + ...stimation by Learning Neural Gradient Function | 1 + ...nary Optimizers for Deep Learning Applications | 1 + ...Tuning of Regression Problems Across Instances | 1 + ... Model Change Maximization for Active Learning | 0 ...y Detection by Point Sequential Reconstruction | 1 + ...otic Analysis of a UCB-based Top Two Algorithm | 1 + ...neralization Bounds for Sparse Neural Networks | 1 + ...re All That Sharpness-Aware Minimization Needs | 1 + ...Normalizing flow neural networks by JKO scheme | 1 + ...Analysis and Mitigation of Reasoning Shortcuts | 1 + ...ed Dataset of Interleaved Image-Text Documents | 1 + ...t Model-free Reinforcement Learning for POMDPs | 1 + ...RTS: Towards Open-Vocabulary Part Segmentation | 1 + ...nchmarks from lightweight tracking annotations | 0 ...cy Prediction Benchmark for Autonomous Driving | 1 + .../OceanBench: The Sea Surface Height Edition | 1 + ...ning with Variational Counterfactual Reasoning | 1 + ...Representations for Generalizability in POMDPs | 1 + ...Approximations to the Gaussian Mixture Entropy | 1 + ...mpling from Gaussian and Product Distributions | 1 + ...eneralization Bounds for Projective Clustering | 1 + ...asked Pre-training and the Marginal Likelihood | 1 + ...ning with Experts: Algorithms and Lower Bounds | 0 ... network posteriors: a variational perspective | 1 + ...tions in distillation: does it pay to disobey? | 1 + ...s of Out-of-distribution Generalization Models | 1 + ... Best-Arm Identification with Fixed Confidence | 1 + ...ning Data Diversity and Fine-tuning Robustness | 1 + ...Solution of Shuffling-Type Gradient Algorithms | 1 + ...ratically-Bounded Losses: an Improved Analysis | 1 + ...nd Interpretability of Gaussian Process Models | 1 + ...of Sparse ICA without Assuming Non-Gaussianity | 0 ...e Overlooked Structure of Stochastic Gradients | 1 + ...rge Language Models - A Critical Investigation | 1 + ...bustness of Removal-Based Feature Attributions | 1 + ... Networks: Exponential Gaps for Long Sequences | 1 + ...ization in Adversarially Robust Classification | 1 + ...er-class Diversity for Supervised Pre-training | 1 + ...the spectral bias of two-layer linear networks | 1 + ... on Model-Based Offline Reinforcement Learning | 1 + ...eneous Architectures in Knowledge Distillation | 1 + ...e-step differentiation of iterative algorithms | 1 + ...odels under Concept Drift by Online Ensembling | 1 + ...arning: Provable Guarantees for Generalization | 0 .../neurips/Online Control for Meta-optimization | 0 ...imal Dynamic Regret meets Practical Algorithms | 1 + ...arning under Adversarial Nonlinear Constraints | 1 + ...Planning with Anytime Deterministic Guarantees | 1 + .../Online robust non-stationary estimation | 0 ...t Style Compensation for Semantic Segmentation | 1 + ...elation-Oriented Multimodality Model Prompting | 1 + ...sk3D: Open-Vocabulary 3D Instance Segmentation | 1 + ...nchmark of Spatio-Temporal Predictive Learning | 1 + ...y Stopping for Robustifying Differentiable NAS | 1 + ...al Fields: Tackling PDEs on General Geometries | 1 + ...thms for the Inhomogeneous Spiked Wigner Model | 1 + ...ction for Graph-based Semi-supervised Learning | 1 + ...escent in Discounted Markov Decision Processes | 1 + ...iational Inequalities with Separable Structure | 1 + ...zation Methods Under a Fixed Computation Model | 1 + ...imal Transport for Treatment Effect Estimation | 1 + ...ned test statistics across independent studies | 1 + ...nt Policy Optimization Framework for Online RL | 1 + ...ptimizing Prompts for Text-to-Image Generation | 1 + ...of Dataset Imbalance for Multilingual Learning | 1 + ...ng with Unreliable Out-of-distribution Sources | 1 + ...ficient Zero-Shot TTS through Speech Prompting | 1 + ...rning Linear Thresholds from Label Proportions | 1 + ...Accurate Long Rollouts with Neural PDE Solvers | 1 + ... Paragraph via Latent Language Diffusion Model | 1 + ...Object Search in Partially Unknown Environment | 0 ...odel for Patient-independent Seizure Detection | 1 + ...n for Enhanced Self-Supervised Image Denoising | 1 + ...echnique for Detecting Poisoned Samples in NLP | 1 + .../Parallel Submodular Function Minimization | 1 + ...mentoring for Offline Model-based Optimization | 1 + ...ed text, but retrieval is an effective defense | 1 + ...y Propagation guided Candidate Label Shrinkage | 0 ...articipatory Personalization in Classification | 1 + ...nce with Generalized Wasserstein Gradient Flow | 1 + ...re Data-Efficient Training of Diffusion Models | 1 + ...ransformer for any Aspect Ratio and Resolution | 1 + ... effect of the learning rate, depth, and width | 1 + ...ine Inference of Different Physical Properties | 1 + ...nE: Representation Learning over Planar Graphs | 1 + ... of protein families as sequences-of-sequences | 1 + ...olicy Space Diversity for Non-Transitive Games | 1 + ...Reconstruction via Guided Set Diffusion Models | 1 + ...tain Structured Strong Winning Lottery Tickets | 1 + ...of Language Models Can Improve Language Models | 1 + ...ata for Improving Utility on Selected Measures | 1 + .../PrObeD: Proactive Object Detection Wrapper | 1 + ...tation Nowcasting with Latent Diffusion Models | 1 + ...trix Factorization via Scaled Gradient Descent | 1 + ... Protein's Stability under a Million Mutations | 1 + ...nd Control in Continual Reinforcement Learning | 1 + ...Asymmetric Kernel SVD in Primal Representation | 1 + ...ation Trade-off in Distributed Mean Estimation | 1 + ...econd-Order Stationary Points and Excess Risks | 1 + ...for stochastic block models and mixture models | 1 + ...d Multimodal Dataset for Molecular Biology Lab | 1 + ...obing Privacy Leakage in Large Language Models | 1 + ...nt Learning with Randomized Linear Classifiers | 1 + ...l Optimization with Convex Lower-level Problem | 1 + ...x Optimization via Efficient Newton Iterations | 1 + ...IR: Prompting for All-in-One Image Restoration | 0 ... Knowledge Updates to LMs Through Distillation | 1 + ...arks for Protein Fitness Prediction and Design | 0 ... Folding on Diverse Tasks, Models, and Metrics | 1 + ...ling, Noise-Contrastive Estimation, and beyond | 1 + .../neurips/Provable benefits of score matching | 1 + ...guarantees for black-box variational inference | 1 + ...More) Sample-Efficient Offline RL with Options | 1 + .../Provably Bounding Neural Network Preimages | 1 + ...-Informed Calibration for Deep Neural Networks | 1 + ...An Efficient Low-bit Quantized Diffusion Model | 1 + ...ltonian Prediction Benchmark for QM9 Molecules | 1 + ...Outliers by Helping Attention Heads Do Nothing | 1 + ...n with Explicit Motion for 3D Object Detection | 0 ...eneration via Large Mixture of Diffusion Paths | 1 + ...active Defense Against Model Poisoning Attacks | 1 + ...pling and Denoising for Real-Time Path Tracing | 1 + ...a Benchmark for Radiology Language Evaluations | 1 + ... and Pre-trained Models for Continual Learning | 1 + ...dom Cuts are Optimal for Explainable k-Medians | 1 + ...for Efficient and Accurate 3D Object Detection | 0 ...adient Descent and Small Random Initialization | 1 + ...er Training on Efficient Panoptic Segmentation | 1 + ...lay Atari with the Help of Instruction Manuals | 1 + ...lyline Transformer with Relative Pose Encoding | 1 + ...asting Continual Learning as Sequence Modeling | 1 + ...on-Convex Iteratively Reweighted Least Squares | 1 + .../Recurrent Temporal Revision Graph Networks | 1 + ...ion for Length Generalization with Scalability | 1 + ...p Neural Networks with Feature Synthesis Tools | 1 + ...+: (In)Stability and Fast Convergence in Games | 1 + ...ng for Discounted MDPs with Short Burn-In Time | 1 + ...hearsal Learning for Avoiding Undesired Future | 0 ...ugh Data-Centric AI: A Comprehensive Benchmark | 1 + ...for Fine-tuning Text-to-Image Diffusion Models | 1 + ...cement Learning with Fast and Forgetful Memory | 1 + ...cement Learning via Representation Distinction | 1 + ...ing Perspective to (Unbalanced) Classification | 1 + ...le Off-Policy Learning for Dosage Combinations | 1 + ...al Text Degeneration from the Data Perspective | 1 + .../Replicability in Reinforcement Learning | 1 + ...: a Framework for Alias-free Operator Learning | 1 + ...ce Learning: A Case For Algorithmic Unit Tests | 1 + ...or Image Super-resolution by Residual Shifting | 1 + ...g the Optimizer in Deep RL: An Empirical Study | 1 + ...Uncovering the Mechanisms of Residual Networks | 1 + data/2023/neurips/Resilient Constrained Learning | 1 + ...uling: An LLM-Empowered LLM Inference Pipeline | 1 + .../Responsible AI (RAI) Games and Ensembles | 1 + ...rt Sampling for Improving Generative Processes | 1 + ...eaking the Uniform Global Attractor Assumption | 1 + ...Architectures Make for Fairer Face Recognition | 1 + ...stems: Are Monotone Rewards Always Beneficial? | 1 + ...rking video action recognition under occlusion | 0 ...lation from the Perspective of Robust Fairness | 0 ...aining and Generalization across Threat Models | 1 + ...LP: Benchmarks, Analysis, and LLMs Evaluations | 1 + ...ng the Evaluation of Image Synthesis with GANs | 1 + ... with Sketching for Contextual Batched Bandits | 1 + ...Distribution Estimation and Reward Improvement | 1 + ...polating weights fine-tuned on diverse rewards | 1 + ... for Language-Supervised Semantic Segmentation | 1 + .../neurips/Riemannian Residual Neural Networks | 1 + ...ess-Aware Minimization on Riemannian Manifolds | 0 ...ptimization methods avoid strict saddle points | 1 + ... Timely Outcome Prediction under Cost Pressure | 1 + data/2023/neurips/Robust Bayesian Satisficing | 1 + ...abel Noise via Maximizing Re-labeling Accuracy | 1 + ...ve Data Expansion Against Spurious Correlation | 1 + ...Robust Matrix Sensing in the Semi-Random Model | 0 ...on Without Moments for Symmetric Distributions | 1 + ...easoning and Fitting via Dual Sparsity Pursuit | 0 ...ively Secure Serverless Collaborative Learning | 1 + ...ith missing values and cell-wise contamination | 1 + ...aining via approximate orthonormal constraints | 1 + ...s Solver for Fast Sampling of Diffusion Models | 1 + ...arning attack on LWE with sparse small secrets | 0 ...re-aware Shapley-based Multipiece Explanations | 0 ...c-Aware Normalizing Flow for Anomaly Detection | 1 + ...istration for Robust 6D Object Pose Estimation | 1 + .../SE(3) Equivariant Augmented Coupling Flows | 1 + ...st High-Quality Sampling from Diffusion Models | 1 + ...t between human and machine visual abstraction | 1 + ... Structured Constrained Nonconvex Optimization | 1 + ...ixing for Distillation with Unlabeled Examples | 1 + ...luster for Referring Video Object Segmentation | 1 + ...al Alignment Perspective for Domain Adaptation | 1 + ... Contribution Evaluation in Federated Learning | 1 + ...Spiked Random Model for Reinforcement Learning | 1 + ...ith Spatiotemporal Annotations of Sound Events | 1 + ...rative Model for Text-to-Behavior in Minecraft | 1 + ...lator for Machine Learning in Particle Physics | 1 + ...for Enhanced Molecular Representation Learning | 0 ...ion Learning with Non-Preferred Demonstrations | 1 + ...n of Decision-Tree Policies in Continuous Time | 1 + ...hing: Causal Discovery and Generative Modeling | 1 + ...onditioned Hierarchical Reinforcement Learning | 0 ...bjective Molecular Optimization with GFlowNets | 1 + ... using Remote Sensing and Citizen Science Data | 0 ...ed Language Models Using Declarative Prompting | 1 + .../Scalable 3D Captioning with Pretrained Models | 1 + .../neurips/Scalable Fair Influence Maximization | 1 + ...Model via Scaling Network Long Skip Connection | 1 + .../Scaling Open-Vocabulary Object Detection | 1 + .../neurips/Scaling Riemannian Diffusion Models | 1 + ...ling laws for language encoding models in fMRI | 1 + ...s and Token Composition in 1-layer Transformer | 1 + ...ble Driving Scenario Generation With Diffusion | 1 + ... Step-sizes with Multidimensional Backtracking | 1 + ...n Task Generalization with Energy-Based Models | 1 + ...orcement Learning against Spurious Correlation | 1 + ...ion Refinement with Discrete Diffusion Process | 1 + .../2023/neurips/Segment Anything in High Quality | 1 + ...proach to Forgetting in Deep Generative Models | 1 + ...es Improves Multi-Agent Reinforcement Learning | 1 + ...ainst On-body Displacement of Flexible Sensors | 1 + ... for Video Localization and Question Answering | 1 + ... Optimization through Bayesian Active Learning | 1 + data/2023/neurips/Self-Predictive Universal AI | 1 + ... and more human-aligned visual representations | 0 ...ntial Subset Matching for Dataset Distillation | 1 + ...n Pretrained Language Models through Honeypots | 1 + ...ised Functional Map Regularized Reconstruction | 1 + ...ize Sharpness To Achieve Better Generalization | 1 + ...dent Networks Copy or Average Teacher Weights? | 1 + ...ramework for Multi-modal Domain Generalization | 1 + ...ive Clustering via One-Dimensional Projections | 1 + ...g Transformers for Large-Graph Representations | 1 + ...ies: Improved Analysis under Weaker Conditions | 1 + ...Contexts into Generative Commonsense Reasoning | 1 + ...aptive Regularization with Frequent Directions | 1 + .../Slot-guided Volumetric Object Radiance Fields | 1 + .../Small batch deep reinforcement learning | 1 + ...mmetrization for deep learning on point clouds | 1 + ... Analysis of Sequential Probability Assignment | 1 + ...ust Test-Time Adaptation on Noisy Data Streams | 1 + ...l Motion Prediction with Cognitive Hierarchies | 1 + ...Efficient Training of Attention-based Networks | 0 ...vex Minimax Optimization in Federated Learning | 1 + ... for Time Series Data: Theory and Applications | 1 + ...lar Activation for Efficient Sequence Modeling | 1 + ...meterization for Epitomic Dataset Distillation | 0 ...lly Private Training of Large Embedding Models | 1 + ...annels, shape bias, and adversarial robustness | 1 + ...ology Images via Bi-modal Contrastive Learning | 1 + ...Estimation for Low-Rank Reinforcement Learning | 1 + data/2023/neurips/Spike-driven Transformer | 1 + ...mmetry breaking in generative diffusion models | 1 + ...ilies: A New Class of Tractable Density Models | 1 + ...ation at ImageNet Scale From A New Perspective | 1 + ...tance Functions and Finer Shape Representation | 1 + ...g Societal Representations in Diffusion Models | 0 ...ent Homology using Signed Barcodes as Measures | 1 + ...d Learning for Federated Domain Generalization | 1 + ... Fourier Transform for Representation Learning | 1 + ...ased Representations for Off-Policy Evaluation | 1 + ...approximators with exponential decaying memory | 1 + ... Benefit Agent Learning and User Understanding | 1 + ...Attacks in the Context of Selective Forgetting | 0 ...te Learning Process in Quantum Neural Networks | 1 + ...Knowledge Assessment for Large Language Models | 1 + ...l Trade-off in Multi-Agent Multi-Armed Bandits | 1 + ...ce Assessment through Conditional Permutations | 1 + .../neurips/Stein \316\240-Importance Sampling" | 1 + ...le Free Sampling of Molecular Transition Paths | 1 + ...d Benchmark for Continuous Story Visualization | 1 + data/2023/neurips/Strategic Apple Tasting | 1 + ...nds for Estimating Correlation Clustering Cost | 0 .../2023/neurips/Streaming PCA for Markovian Data | 1 + ...ulation of Human Percepts via Robustified ANNs | 1 + ...Neural Inverse Graphics from a Pile of Objects | 1 + data/2023/neurips/Structure of universal formulas | 1 + ...ks for Density Estimation and Causal Inference | 1 + ...o-End Stability and Output Tracking Guarantees | 1 + ...rediction with Stronger Consistency Guarantees | 1 + ...tyleDrop: Text-to-Image Synthesis of Any Style | 0 ...StyleGAN knows Normal, Depth, Albedo, and More | 1 + ...roportional high-dimensional Linear Regression | 1 + ...unterexample for the Success of Early Stopping | 1 + .../Successor-Predecessor Intrinsic Exploration | 1 + ...braic Decomposition via Reinforcement Learning | 0 ...Survival Analysis with Time-Varying Covariates | 0 ...ng Environments for Sustainable Energy Systems | 1 + ...e Prompt Adaptation for Vision-Language Models | 1 + ...orcement Learning for Adaptive Mesh Refinement | 1 + ...nd Slow Thinking for Complex Interactive Tasks | 1 + ...Molecules, Proteins, and Crystalline Materials | 1 + ... Pose Estimation with Geometric Reconstruction | 1 + ... through Object-Centric Relational Abstraction | 1 + ...n-world Compositional Text-to-image Generation | 1 + ...rastive Loss for Visual Reinforcement Learning | 1 + ...omplex Reasoning over Temporal Knowledge Graph | 1 + data/2023/neurips/TOA: Task-oriented Active VQA | 1 + ...auditing training data for improved regression | 1 + ...x Optimal and Instance-Dependent Regret Bounds | 1 + ...g Self-Attention for Graph via Rooted Subtrees | 1 + ...ects in Graph-based Spatiotemporal Forecasting | 1 + ...atures for Scalable Molecular Machine Learning | 1 + ... Space: Improved Editing of Pre-Trained Models | 1 + ...e Weight Analysis, and Neural Network Training | 1 + ...Prior Compensation for Human Motion Prediction | 1 + ...rse Classifier for Fine-Grained Classification | 1 + ...g for Matching-based Video Object Segmentation | 1 + ... Efficient Unified Model for Massive NLP Tasks | 1 + .../Textually Pretrained Speech Language Models | 1 + ...and Convergence of Local Bayesian Optimization | 1 + ...ching Consensus and Convergence to Equilibrium | 1 + ... Normalization in Sharpness-Aware Minimization | 1 + ...Reinforcement Learning with a Generative Model | 1 + ...stortion of Binomial Voting Defies Expectation | 1 + ... Stability under Regularized Learning in Games | 1 + .../The Gain from Ordering in Online Learning | 0 ...e Portability and Implications for ML Progress | 1 + ... Subgraph Densities to Stochastic Block Models | 1 + .../The Learnability of In-Context Learning | 1 + ...tperforming Curated Corpora with Web Data Only | 0 ...ly-supervised Whole Slide Image Classification | 1 + ...f Emergent In-Context Learning in Transformers | 1 + ...g Data Representations in Deep Neural Networks | 1 + ...sive power of pooling in Graph Neural Networks | 1 + ...level in linear regression with dependent data | 1 + .../The probability flow ODE is provably fast | 1 + ...tability with respect to distributional shifts | 1 + ...nductive Biases in Deep Convolutional Networks | 0 ...ization, Generalization and Conflict-Avoidance | 1 + ... Large Language Models with External Knowledge | 1 + ...ounds for Volumetric Spanners and Applications | 1 + ... Bounds for Gradient Descent on Separable Data | 1 + ...ransformer for Irregularly Sampled Time Series | 1 + ...guage Models Can Teach Themselves to Use Tools | 1 + ...eep Ensemble Works in Selective Classification | 0 ...-supervised Simplicial Representation Learning | 1 + ...verification and retrieval without fine-tuning | 1 + ...Learning: New Architecture and Unified Library | 1 + ...Initialization: What Makes a Good Sparse Mask? | 0 ...e Study on Overparameterized Linear Regression | 1 + ...bution-Agnostic Generalized Category Discovery | 1 + ...: Characterizing Scaling and Transfer Behavior | 1 + ...ds Higher Ranks via Adversarial Weight Pruning | 1 + .../Towards In-context Scene Understanding | 1 + ...cene Understanding by Vision Foundation Models | 1 + ...ng for Group Robustness with Fewer Annotations | 1 + ...nd Chain of Thought: A Theoretical Perspective | 1 + ...ICD Coding via Tree-based Contrastive Learning | 1 + ...door Purification through Feature Shift Tuning | 1 + ...ark for High-Level Synthesis Targeted to FPGAs | 1 + ... of Kernel-based Methods Under Covariate Shift | 1 + ...tive Learning for Disentangled Representations | 1 + ...rons with Clustered Compositional Explanations | 1 + ...n Dataset on Large Tensor Computational Graphs | 1 + ...Fight Easy: Robust Meta Reinforcement Learning | 1 + ...Chain-of-Thought via Latent-Variable Inference | 1 + ...ted Neural Networks is \342\210\203R-Complete" | 1 + ...ges Improves Robustness to Adversarial Attacks | 1 + ...ion for Variable-Sized Text-to-Image Synthesis | 1 + ...of Stability Phenomenon via Bifurcation Theory | 1 + ...bust Generalization for Tensor Neural Networks | 1 + ...for Lidar View Synthesis and 3D Reconstruction | 1 + ...or Continual Knowledge Retention and Promotion | 1 + ... with data-constrained spiking neural networks | 1 + ... Trojan Prompt Attack on Large Language Models | 1 + ...metric Super-Resolution with BLASTNet 2.0 Data | 1 + ... Untuned SGD and the Power of Adaptive Methods | 1 + ...-Stage Learning to Defer with Multiple Experts | 1 + ... Retrieve Any Object via Prompt-based Tracking | 1 + ...A Real-World Dataset for Under-Display Cameras | 0 ...d for Real-Time Rendering of Large-Scale Scene | 1 + ...strained Pose Prior-Free Neural Radiance Field | 1 + ...ommendation Unlearning via Error Decomposition | 0 ...odels with structured discrete representations | 1 + ... Network for Cross-Domain Video-Text Retrieval | 0 ...e Instance Reweighting for Off-Policy Learning | 1 + ...Unconstrained Dynamic Regret via Sparse Coding | 1 + ...r Weakly Open-Vocabulary Semantic Segmentation | 1 + ...-supervised Learning under Distribution Shifts | 1 + ...ives as the ELBO with Simple Data Augmentation | 1 + ...dness and Adaptation Difficulty via Attributes | 0 ...Learning via Stage-wise Relaxed Initialization | 1 + ...earning for Out-of-Distribution Generalization | 1 + ...ar property prediction: Insights and Solutions | 1 + ...ental class-level effects of data augmentation | 1 + ...ramework for Fast Sampling of Diffusion Models | 1 + ...rm Convergence with Square-Root Lipschitz Loss | 1 + .../Universality and Limitations of Prompt Tuning | 1 + ...mization in Auditing Differentially Private ML | 1 + .../Unsupervised Anomaly Detection with Rejection | 1 + ...ture Search with Disentangled Self-Supervision | 1 + ...supervised Image Denoising with Score Function | 1 + ...Representation for CT Metal Artifact Reduction | 1 + ... Science Applications of Large Language Models | 1 + .../neurips/Utilitarian Algorithm Configuration | 1 + ...ext Omni-Modality Foundation Model and Dataset | 1 + ...n Vision-Language Tasks via Pre-trained Models | 1 + .../neurips/VaRT: Variational Regression Trees | 1 + ...aling on Graphs for Combinatorial Optimization | 1 + ...n Models as Rewards for Reinforcement Learning | 1 + ...or Keystep Recognition in Instructional Videos | 1 + ...ent between AI and Humans in Visual Perception | 1 + ...g gender bias in image-text pronoun resolution | 1 + ...n Inversion: Image Editing via Image Prompting | 0 ... for Fast Neural Radiance Field Reconstruction | 1 + ...Quantum Many-Body Schr\303\266dinger Equation" | 1 + ... Supervised Learning with Intermediate Targets | 1 + .../What Can We Learn from Unlearnable Datasets? | 1 + ... Trajectory Prediction for Autonomous Driving? | 1 + ... Evaluation of Zero-Shot Semantic Segmentation | 1 + ...ork for Offline Inverse Reinforcement Learning | 1 + ...Nets Outperform Boosted Trees on Tabular Data? | 1 + ...oes Confidence-Based Cascade Deferral Suffice? | 1 + ...es Optimizing a Proper Loss Yield Calibration? | 1 + ...rom? Origin Attribution of AI-Generated Images | 0 ...icial Visual Cortex for Embodied Intelligence? | 1 + ...Unseen Novel Categories of Articulated Objects | 1 + ...Aware Minimization Generalize Better Than SGD? | 1 + ...soning emerges from the locality of experience | 1 + ...dal time series for wildfire spread prediction | 0 ... Memory Efficient Adaptation of Language Model | 1 + ...ch Implementations: Guarantees and Limitations | 1 + .../XAGen: 3D Expressive Human Avatars Generation | 1 + .../Zero-One Laws of Graph Neural Networks | 1 + ...ce-Aware Structured Pruning of Language Models | 1 + ...rk for Goal-Conditioned RL using f-Divergences | 1 + ...etter Initialization with Differential Privacy | 1 + .../neurips/rPPG-Toolbox: Deep Remote PPG Toolbox | 1 + ...sentation Learner for Single-Cell RNA-Seq Data | 1 + 8074 files changed, 7920 insertions(+) create mode 100644 data/2018/vldb/Declarative Recursive Computation on an RDBMS create mode 100644 data/2020/neurips/(De)Randomized Smoothing for Certifiable Defense against Patch Attacks create mode 100644 data/2020/neurips/3D Multi-bodies: Fitting Sets of Plausible 3D Human Models to Ambiguous Image Data create mode 100644 data/2020/neurips/3D Self-Supervised Methods for Medical Imaging create mode 100644 data/2020/neurips/3D Shape Reconstruction from Vision and Touch create mode 100644 data/2020/neurips/A B Testing in Dense Large-Scale Networks: Design and Inference create mode 100644 data/2020/neurips/A Bandit Learning Algorithm and Applications to Auction Design create mode 100644 data/2020/neurips/A Bayesian Nonparametrics View into Deep Representations create mode 100644 data/2020/neurips/A Bayesian Perspective on Training Speed and Model Selection create mode 100644 data/2020/neurips/A Benchmark for Systematic Generalization in Grounded Language Understanding create mode 100644 data/2020/neurips/A Biologically Plausible Neural Network for Slow Feature Analysis create mode 100644 data/2020/neurips/A Boolean Task Algebra for Reinforcement Learning create mode 100644 data/2020/neurips/A Catalyst Framework for Minimax Optimization create mode 100644 data/2020/neurips/A Causal View on Robustness of Neural Networks create mode 100644 data/2020/neurips/A Class of Algorithms for General Instrumental Variable Models create mode 100644 data/2020/neurips/A Closer Look at Accuracy vs. Robustness create mode 100644 data/2020/neurips/A Closer Look at the Training Strategy for Modern Meta-Learning create mode 100644 data/2020/neurips/A Combinatorial Perspective on Transfer Learning create mode 100644 data/2020/neurips/A Computational Separation between Private Learning and Online Learning create mode 100644 data/2020/neurips/A Continuous-Time Mirror Descent Approach to Sparse Phase Retrieval create mode 100644 data/2020/neurips/A Contour Stochastic Gradient Langevin Dynamics Algorithm for Simulations of Multi-modal Distributions create mode 100644 data/2020/neurips/A Convolutional Auto-Encoder for Haplotype Assembly and Viral Quasispecies Reconstruction create mode 100644 data/2020/neurips/A Decentralized Parallel Algorithm for Training Generative Adversarial Nets create mode 100644 data/2020/neurips/A Dictionary Approach to Domain-Invariant Learning in Deep Networks create mode 100644 data/2020/neurips/A Discrete Variational Recurrent Topic Model without the Reparametrization Trick create mode 100644 data/2020/neurips/A Dynamical Central Limit Theorem for Shallow Neural Networks create mode 100644 data/2020/neurips/A Fair Classifier Using Kernel Density Estimation create mode 100644 data/2020/neurips/A Feasible Level Proximal Point Method for Nonconvex Sparse Constrained Optimization create mode 100644 data/2020/neurips/A Finite-Time Analysis of Two Time-Scale Actor-Critic Methods create mode 100644 data/2020/neurips/A Flexible Framework for Designing Trainable Priors with Adaptive Smoothing and Game Encoding create mode 100644 data/2020/neurips/A Game Theoretic Analysis of Additive Adversarial Attacks and Defenses create mode 100644 data/2020/neurips/A Game-Theoretic Analysis of the Empirical Revenue Maximization Algorithm with Endogenous Sampling create mode 100644 data/2020/neurips/A General Large Neighborhood Search Framework for Solving Integer Linear Programs create mode 100644 data/2020/neurips/A General Method for Robust Learning from Batches create mode 100644 data/2020/neurips/A Generalized Neural Tangent Kernel Analysis for Two-layer Neural Networks create mode 100644 data/2020/neurips/A Group-Theoretic Framework for Data Augmentation create mode 100644 data/2020/neurips/A Limitation of the PAC-Bayes Framework create mode 100644 data/2020/neurips/A Local Temporal Difference Code for Distributional Reinforcement Learning create mode 100644 data/2020/neurips/A Loss Function for Generative Neural Networks Based on Watson's Perceptual Model create mode 100644 data/2020/neurips/A Matrix Chernoff Bound for Markov Chains and Its Application to Co-occurrence Matrices create mode 100644 data/2020/neurips/A Maximum-Entropy Approach to Off-Policy Evaluation in Average-Reward MDPs create mode 100644 data/2020/neurips/A Measure-Theoretic Approach to Kernel Conditional Mean Embeddings create mode 100644 data/2020/neurips/A Non-Asymptotic Analysis for Stein Variational Gradient Descent create mode 100644 data/2020/neurips/A Novel Approach for Constrained Optimization in Graphical Models create mode 100644 data/2020/neurips/A Novel Automated Curriculum Strategy to Solve Hard Sokoban Planning Instances create mode 100644 data/2020/neurips/A Randomized Algorithm to Reduce the Support of Discrete Measures create mode 100644 data/2020/neurips/A Ranking-based, Balanced Loss Function Unifying Classification and Localisation in Object Detection create mode 100644 data/2020/neurips/A Robust Functional EM Algorithm for Incomplete Panel Count Data create mode 100644 data/2020/neurips/A Scalable Approach for Privacy-Preserving Collaborative Machine Learning create mode 100644 data/2020/neurips/A Scalable MIP-based Method for Learning Optimal Multivariate Decision Trees create mode 100644 data/2020/neurips/A Self-Tuning Actor-Critic Algorithm create mode 100644 data/2020/neurips/A Simple Language Model for Task-Oriented Dialogue create mode 100644 data/2020/neurips/A Simple and Efficient Smoothing Method for Faster Optimization and Local Exploration create mode 100644 data/2020/neurips/A Single Recipe for Online Submodular Maximization with Adversarial or Stochastic Constraints create mode 100644 data/2020/neurips/A Single-Loop Smoothed Gradient Descent-Ascent Algorithm for Nonconvex-Concave Min-Max Problems create mode 100644 data/2020/neurips/A Spectral Energy Distance for Parallel Speech Synthesis create mode 100644 data/2020/neurips/A Statistical Framework for Low-bitwidth Training of Deep Neural Networks create mode 100644 data/2020/neurips/A Statistical Mechanics Framework for Task-Agnostic Sample Design in Machine Learning create mode 100644 data/2020/neurips/A Stochastic Path Integral Differential EstimatoR Expectation Maximization Algorithm create mode 100644 data/2020/neurips/A Study on Encodings for Neural Architecture Search create mode 100644 data/2020/neurips/A Theoretical Framework for Target Propagation create mode 100644 data/2020/neurips/A Tight Lower Bound and Efficient Reduction for Swap Regret create mode 100644 data/2020/neurips/A Topological Filter for Learning with Label Noise create mode 100644 data/2020/neurips/A Unified Switching System Perspective and Convergence Analysis of Q-Learning Algorithms create mode 100644 data/2020/neurips/A Unified View of Label Shift Estimation create mode 100644 data/2020/neurips/A Unifying View of Optimism in Episodic Reinforcement Learning create mode 100644 data/2020/neurips/A Universal Approximation Theorem of Deep Neural Networks for Expressing Probability Distributions create mode 100644 data/2020/neurips/A Variational Approach for Learning from Positive and Unlabeled Data create mode 100644 data/2020/neurips/A causal view of compositional zero-shot recognition create mode 100644 data/2020/neurips/A convex optimization formulation for multivariate regression create mode 100644 data/2020/neurips/A game-theoretic analysis of networked system control for common-pool resource management using multi-agent reinforcement learning create mode 100644 data/2020/neurips/A kernel test for quasi-independence create mode 100644 data/2020/neurips/A mathematical model for automatic differentiation in machine learning create mode 100644 data/2020/neurips/A mathematical theory of cooperative communication create mode 100644 data/2020/neurips/A mean-field analysis of two-player zero-sum games create mode 100644 data/2020/neurips/A meta-learning approach to (re)discover plasticity rules that carve a desired function into a neural network create mode 100644 data/2020/neurips/A new convergent variant of Q-learning with linear function approximation create mode 100644 data/2020/neurips/A new inference approach for training shallow and deep generalized linear models of noisy interacting neurons create mode 100644 data/2020/neurips/A novel variational form of the Schatten-$p$ quasi-norm create mode 100644 data/2020/neurips/A polynomial-time algorithm for learning nonparametric causal graphs create mode 100644 data/2020/neurips/A shooting formulation of deep learning create mode 100644 data/2020/neurips/A simple normative network approximates local non-Hebbian learning in the cortex create mode 100644 data/2020/neurips/AI Feynman 2.0: Pareto-optimal symbolic regression exploiting graph modularity create mode 100644 data/2020/neurips/AOT: Appearance Optimal Transport Based Identity Swapping for Forgery Detection create mode 100644 data/2020/neurips/ARMA Nets: Expanding Receptive Field for Dense Prediction create mode 100644 data/2020/neurips/AViD Dataset: Anonymized Videos from Diverse Countries create mode 100644 data/2020/neurips/Accelerating Training of Transformer-Based Language Models with Progressive Layer Dropping create mode 100644 data/2020/neurips/Acceleration with a Ball Optimization Oracle create mode 100644 data/2020/neurips/Achieving Equalized Odds by Resampling Sensitive Attributes create mode 100644 data/2020/neurips/Active Invariant Causal Prediction: Experiment Selection through Stability create mode 100644 data/2020/neurips/Active Structure Learning of Causal DAGs via Directed Clique Trees create mode 100644 data/2020/neurips/AdaBelief Optimizer: Adapting Stepsizes by the Belief in Observed Gradients create mode 100644 data/2020/neurips/AdaShare: Learning What To Share For Efficient Deep Multi-Task Learning create mode 100644 data/2020/neurips/AdaTune: Adaptive Tensor Program Compilation Made Efficient create mode 100644 data/2020/neurips/Adam with Bandit Sampling for Deep Learning create mode 100644 data/2020/neurips/Adaptation Properties Allow Identification of Optimized Neural Codes create mode 100644 data/2020/neurips/Adapting Neural Architectures Between Domains create mode 100644 data/2020/neurips/Adapting to Misspecification in Contextual Bandits create mode 100644 data/2020/neurips/Adaptive Discretization for Model-Based Reinforcement Learning create mode 100644 data/2020/neurips/Adaptive Experimental Design with Temporal Interference: A Maximum Likelihood Approach create mode 100644 data/2020/neurips/Adaptive Gradient Quantization for Data-Parallel SGD create mode 100644 data/2020/neurips/Adaptive Graph Convolutional Recurrent Network for Traffic Forecasting create mode 100644 data/2020/neurips/Adaptive Importance Sampling for Finite-Sum Optimization and Sampling with Decreasing Step-Sizes create mode 100644 data/2020/neurips/Adaptive Learning of Rank-One Models for Efficient Pairwise Sequence Alignment create mode 100644 data/2020/neurips/Adaptive Online Estimation of Piecewise Polynomial Trends create mode 100644 data/2020/neurips/Adaptive Probing Policies for Shortest Path Routing create mode 100644 data/2020/neurips/Adaptive Reduced Rank Regression create mode 100644 data/2020/neurips/Adaptive Sampling for Stochastic Risk-Averse Learning create mode 100644 data/2020/neurips/Adaptive Shrinkage Estimation for Streaming Graphs create mode 100644 data/2020/neurips/AdvFlow: Inconspicuous Black-box Adversarial Attacks using Normalizing Flows create mode 100644 data/2020/neurips/Advances in Black-Box VI: Normalizing Flows, Importance Weighting, and Optimization create mode 100644 data/2020/neurips/Adversarial Attacks on Deep Graph Matching create mode 100644 data/2020/neurips/Adversarial Attacks on Linear Contextual Bandits create mode 100644 data/2020/neurips/Adversarial Bandits with Corruptions: Regret Lower Bound and No-regret Algorithm create mode 100644 data/2020/neurips/Adversarial Blocking Bandits create mode 100644 data/2020/neurips/Adversarial Crowdsourcing Through Robust Rank-One Matrix Completion create mode 100644 data/2020/neurips/Adversarial Distributional Training for Robust Deep Learning create mode 100644 data/2020/neurips/Adversarial Example Games create mode 100644 data/2020/neurips/Adversarial Learning for Robust Deep Clustering create mode 100644 data/2020/neurips/Adversarial Robustness of Supervised Sparse Coding create mode 100644 data/2020/neurips/Adversarial Self-Supervised Contrastive Learning create mode 100644 data/2020/neurips/Adversarial Soft Advantage Fitting: Imitation Learning without Policy Optimization create mode 100644 data/2020/neurips/Adversarial Sparse Transformer for Time Series Forecasting create mode 100644 data/2020/neurips/Adversarial Style Mining for One-Shot Unsupervised Domain Adaptation create mode 100644 data/2020/neurips/Adversarial Training is a Form of Data-dependent Operator Norm Regularization create mode 100644 data/2020/neurips/Adversarial Weight Perturbation Helps Robust Generalization create mode 100644 data/2020/neurips/Adversarial robustness via robust low rank representations create mode 100644 data/2020/neurips/Adversarially Robust Few-Shot Learning: A Meta-Learning Approach create mode 100644 data/2020/neurips/Adversarially Robust Streaming Algorithms via Differential Privacy create mode 100644 data/2020/neurips/Adversarially-learned Inference via an Ensemble of Discrete Undirected Graphical Models create mode 100644 data/2020/neurips/Agnostic $Q$-learning with Function Approximation in Deterministic Systems: Near-Optimal Bounds on Approximation Error and Sample Complexity create mode 100644 data/2020/neurips/Agnostic Learning of a Single Neuron with Gradient Descent create mode 100644 data/2020/neurips/Agnostic Learning with Multiple Objectives create mode 100644 data/2020/neurips/Agree to Disagree: Adaptive Ensemble Knowledge Distillation in Gradient Space create mode 100644 data/2020/neurips/Algorithmic recourse under imperfect causal knowledge: a probabilistic approach create mode 100644 data/2020/neurips/All Word Embeddings from One Embedding create mode 100644 data/2020/neurips/All-or-nothing statistical and computational phase transitions in sparse spiked matrix estimation create mode 100644 data/2020/neurips/Almost Optimal Model-Free Reinforcement Learningvia Reference-Advantage Decomposition create mode 100644 data/2020/neurips/Almost Surely Stable Deep Dynamics create mode 100644 data/2020/neurips/An Analysis of SVD for Deep Rotation Estimation create mode 100644 data/2020/neurips/An Asymptotically Optimal Primal-Dual Incremental Algorithm for Contextual Linear Bandits create mode 100644 data/2020/neurips/An Efficient Adversarial Attack for Tree Ensembles create mode 100644 data/2020/neurips/An Efficient Asynchronous Method for Integrating Evolutionary and Gradient-based Policy Search create mode 100644 data/2020/neurips/An Efficient Framework for Clustered Federated Learning create mode 100644 data/2020/neurips/An Empirical Process Approach to the Union Bound: Practical Algorithms for Combinatorial and Linear Bandits create mode 100644 data/2020/neurips/An Equivalence between Loss Functions and Non-Uniform Sampling in Experience Replay create mode 100644 data/2020/neurips/An Imitation from Observation Approach to Transfer Learning with Dynamics Mismatch create mode 100644 data/2020/neurips/An Improved Analysis of (Variance-Reduced) Policy Gradient and Natural Policy Gradient Methods create mode 100644 data/2020/neurips/An Improved Analysis of Stochastic Gradient Descent with Momentum create mode 100644 data/2020/neurips/An Optimal Elimination Algorithm for Learning a Best Arm create mode 100644 data/2020/neurips/An Unbiased Risk Estimator for Learning with Augmented Classes create mode 100644 data/2020/neurips/An Unsupervised Information-Theoretic Perceptual Quality Metric create mode 100644 data/2020/neurips/An analytic theory of shallow networks dynamics for hinge loss classification create mode 100644 data/2020/neurips/An efficient nonconvex reformulation of stagewise convex optimization problems create mode 100644 data/2020/neurips/An implicit function learning approach for parametric modal regression create mode 100644 data/2020/neurips/An operator view of policy gradient methods create mode 100644 data/2020/neurips/Analysis and Design of Thompson Sampling for Stochastic Partial Monitoring create mode 100644 data/2020/neurips/Analytic Characterization of the Hessian in Shallow ReLU Models: A Tale of Symmetry create mode 100644 data/2020/neurips/Analytical Probability Distributions and Exact Expectation-Maximization for Deep Generative Networks create mode 100644 data/2020/neurips/Applications of Common Entropy for Causal Inference create mode 100644 data/2020/neurips/Approximate Cross-Validation for Structured Models create mode 100644 data/2020/neurips/Approximate Cross-Validation with Low-Rank Data in High Dimensions create mode 100644 data/2020/neurips/Approximate Heavily-Constrained Learning with Lagrange Multiplier Models create mode 100644 data/2020/neurips/Approximation Based Variance Reduction for Reparameterization Gradients create mode 100644 data/2020/neurips/Assessing SATNet's Ability to Solve the Symbol Grounding Problem create mode 100644 data/2020/neurips/Assisted Learning: A Framework for Multi-Organization Learning create mode 100644 data/2020/neurips/Asymmetric Shapley values: incorporating causal knowledge into model-agnostic explainability create mode 100644 data/2020/neurips/Asymptotic Guarantees for Generative Modeling Based on the Smooth Wasserstein Distance create mode 100644 data/2020/neurips/Asymptotic normality and confidence intervals for derivatives of 2-layers neural network in the random features model create mode 100644 data/2020/neurips/Asymptotically Optimal Exact Minibatch Metropolis-Hastings create mode 100644 data/2020/neurips/Attack of the Tails: Yes, You Really Can Backdoor Federated Learning create mode 100644 data/2020/neurips/AttendLight: Universal Attention-Based Reinforcement Learning Model for Traffic Signal Control create mode 100644 data/2020/neurips/Attention-Gated Brain Propagation: How the brain can implement reward-based error backpropagation create mode 100644 data/2020/neurips/Attribute Prototype Network for Zero-Shot Learning create mode 100644 data/2020/neurips/Attribution Preservation in Network Compression for Reliable Network Interpretation create mode 100644 data/2020/neurips/Audeo: Audio Generation for a Silent Performance Video create mode 100644 data/2020/neurips/Auditing Differentially Private Machine Learning: How Private is Private SGD? create mode 100644 data/2020/neurips/Auto Learning Attention create mode 100644 data/2020/neurips/Auto-Panoptic: Cooperative Multi-Component Architecture Search for Panoptic Segmentation create mode 100644 data/2020/neurips/AutoBSS: An Efficient Algorithm for Block Stacking Style Search create mode 100644 data/2020/neurips/AutoPrivacy: Automated Layer-wise Parameter Selection for Secure Neural Network Inference create mode 100644 data/2020/neurips/AutoSync: Learning to Synchronize for Data-Parallel Distributed Deep Learning create mode 100644 data/2020/neurips/Autoencoders that don't overfit towards the Identity create mode 100644 data/2020/neurips/Autofocused oracles for model-based design create mode 100644 data/2020/neurips/Automatic Curriculum Learning through Value Disagreement create mode 100644 data/2020/neurips/Automatic Perturbation Analysis for Scalable Certified Robustness and Beyond create mode 100644 data/2020/neurips/Automatically Learning Compact Quality-aware Surrogates for Optimization Problems create mode 100644 data/2020/neurips/Autoregressive Score Matching create mode 100644 data/2020/neurips/Auxiliary Task Reweighting for Minimum-data Learning create mode 100644 data/2020/neurips/AvE: Assistance via Empowerment create mode 100644 data/2020/neurips/Avoiding Side Effects By Considering Future Tasks create mode 100644 data/2020/neurips/Avoiding Side Effects in Complex Environments create mode 100644 data/2020/neurips/Axioms for Learning from Pairwise Comparisons create mode 100644 data/2020/neurips/BAIL: Best-Action Imitation Learning for Batch Deep Reinforcement Learning create mode 100644 data/2020/neurips/BERT Loses Patience: Fast and Robust Inference with Early Exit create mode 100644 data/2020/neurips/BOSS: Bayesian Optimization over String Spaces create mode 100644 data/2020/neurips/BRP-NAS: Prediction-based NAS using GCNs create mode 100644 data/2020/neurips/Backpropagating Linearly Improves Transferability of Adversarial Examples create mode 100644 data/2020/neurips/Bad Global Minima Exist and SGD Can Reach Them create mode 100644 data/2020/neurips/Balanced Meta-Softmax for Long-Tailed Visual Recognition create mode 100644 data/2020/neurips/Bandit Linear Control create mode 100644 data/2020/neurips/Bandit Samplers for Training Graph Neural Networks create mode 100644 data/2020/neurips/BanditPAM: Almost Linear Time k-Medoids Clustering via Multi-Armed Bandits create mode 100644 data/2020/neurips/Barking up the right tree: an approach to search over molecule synthesis DAGs create mode 100644 data/2020/neurips/Batch normalization provably avoids ranks collapse for randomly initialised deep networks create mode 100644 data/2020/neurips/Batched Coarse Ranking in Multi-Armed Bandits create mode 100644 data/2020/neurips/Baxter Permutation Process create mode 100644 data/2020/neurips/BayReL: Bayesian Relational Learning for Multi-omics Data Integration create mode 100644 data/2020/neurips/Bayes Consistency vs. H-Consistency: The Interplay between Surrogate Loss Functions and the Scoring Function Class create mode 100644 data/2020/neurips/Bayesian Attention Modules create mode 100644 data/2020/neurips/Bayesian Bits: Unifying Quantization and Pruning create mode 100644 data/2020/neurips/Bayesian Causal Structural Learning with Zero-Inflated Poisson Bayesian Networks create mode 100644 data/2020/neurips/Bayesian Deep Ensembles via the Neural Tangent Kernel create mode 100644 data/2020/neurips/Bayesian Deep Learning and a Probabilistic Perspective of Generalization create mode 100644 data/2020/neurips/Bayesian Meta-Learning for the Few-Shot Setting via Deep Kernels create mode 100644 data/2020/neurips/Bayesian Multi-type Mean Field Multi-agent Imitation Learning create mode 100644 data/2020/neurips/Bayesian Optimization for Iterative Learning create mode 100644 data/2020/neurips/Bayesian Optimization of Risk Measures create mode 100644 data/2020/neurips/Bayesian Probabilistic Numerical Integration with Tree-Based Models create mode 100644 data/2020/neurips/Bayesian Pseudocoresets create mode 100644 data/2020/neurips/Bayesian Robust Optimization for Imitation Learning create mode 100644 data/2020/neurips/Bayesian filtering unifies adaptive and non-adaptive neural network optimization methods create mode 100644 data/2020/neurips/Belief Propagation Neural Networks create mode 100644 data/2020/neurips/Belief-Dependent Macro-Action Discovery in POMDPs using the Value of Information create mode 100644 data/2020/neurips/Benchmarking Deep Inverse Models over time, and the Neural-Adjoint method create mode 100644 data/2020/neurips/Benchmarking Deep Learning Interpretability in Time Series Predictions create mode 100644 data/2020/neurips/Beta Embeddings for Multi-Hop Logical Reasoning in Knowledge Graphs create mode 100644 data/2020/neurips/Beta R-CNN: Looking into Pedestrian Detection from Another Perspective create mode 100644 data/2020/neurips/Better Full-Matrix Regret via Parameter-Free Online Learning create mode 100644 data/2020/neurips/Better Set Representations For Relational Reasoning create mode 100644 data/2020/neurips/Beyond Homophily in Graph Neural Networks: Current Limitations and Effective Designs create mode 100644 data/2020/neurips/Beyond Individualized Recourse: Interpretable and Interactive Summaries of Actionable Recourses create mode 100644 data/2020/neurips/Beyond Lazy Training for Over-parameterized Tensor Decomposition create mode 100644 data/2020/neurips/Beyond Perturbations: Learning Guarantees with Arbitrary Adversarial Test Examples create mode 100644 data/2020/neurips/Beyond accuracy: quantifying trial-by-trial behaviour of CNNs and humans by measuring error consistency create mode 100644 data/2020/neurips/Beyond the Mean-Field: Structured Deep Gaussian Processes Improve the Predictive Uncertainties create mode 100644 data/2020/neurips/Bi-level Score Matching for Learning Energy-based Latent Variable Models create mode 100644 data/2020/neurips/Bias no more: high-probability data-dependent regret bounds for adversarial bandits and MDPs create mode 100644 data/2020/neurips/Bidirectional Convolutional Poisson Gamma Dynamical Systems create mode 100644 data/2020/neurips/Big Bird: Transformers for Longer Sequences create mode 100644 data/2020/neurips/Big Self-Supervised Models are Strong Semi-Supervised Learners create mode 100644 data/2020/neurips/Biological credit assignment through dynamic inversion of feedforward networks create mode 100644 data/2020/neurips/Biologically Inspired Mechanisms for Adversarial Robustness create mode 100644 data/2020/neurips/Black-Box Certification with Randomized Smoothing: A Functional Optimization Based Framework create mode 100644 data/2020/neurips/Black-Box Optimization with Local Generative Surrogates create mode 100644 data/2020/neurips/Black-Box Ripper: Copying black-box models using generative evolutionary algorithms create mode 100644 data/2020/neurips/Blind Video Temporal Consistency via Deep Video Prior create mode 100644 data/2020/neurips/BlockGAN: Learning 3D Object-aware Scene Representations from Unlabelled Images create mode 100644 data/2020/neurips/BoTorch: A Framework for Efficient Monte-Carlo Bayesian Optimization create mode 100644 data/2020/neurips/Boosting Adversarial Training with Hypersphere Embedding create mode 100644 data/2020/neurips/Boosting First-Order Methods by Shifting Objective: New Schemes with Faster Worst-Case Rates create mode 100644 data/2020/neurips/Bootstrap Your Own Latent - A New Approach to Self-Supervised Learning create mode 100644 data/2020/neurips/Bootstrapping neural processes create mode 100644 data/2020/neurips/Boundary thickness and robustness in learning models create mode 100644 data/2020/neurips/BoxE: A Box Embedding Model for Knowledge Base Completion create mode 100644 data/2020/neurips/Breaking Reversibility Accelerates Langevin Dynamics for Non-Convex Optimization create mode 100644 data/2020/neurips/Breaking the Communication-Privacy-Accuracy Trilemma create mode 100644 data/2020/neurips/Breaking the Sample Size Barrier in Model-Based Reinforcement Learning with a Generative Model create mode 100644 data/2020/neurips/Bridging Imagination and Reality for Model-Based Deep Reinforcement Learning create mode 100644 data/2020/neurips/Bridging the Gap between Sample-based and One-shot Neural Architecture Search with BONAS create mode 100644 data/2020/neurips/Building powerful and equivariant graph neural networks with structural message-passing create mode 100644 data/2020/neurips/Byzantine Resilient Distributed Multi-Task Learning create mode 100644 data/2020/neurips/CASTLE: Regularization via Auxiliary Causal Graph Discovery create mode 100644 data/2020/neurips/CHIP: A Hawkes Process Model for Continuous-time Networks with Scalable and Consistent Estimation create mode 100644 data/2020/neurips/CLEARER: Multi-Scale Neural Architecture Search for Image Restoration create mode 100644 data/2020/neurips/COBE: Contextualized Object Embeddings from Narrated Instructional Video create mode 100644 data/2020/neurips/COOT: Cooperative Hierarchical Transformer for Video-Text Representation Learning create mode 100644 data/2020/neurips/COPT: Coordinated Optimal Transport on Graphs create mode 100644 data/2020/neurips/COT-GAN: Generating Sequential Data via Causal Optimal Transport create mode 100644 data/2020/neurips/CSER: Communication-efficient SGD with Error Reset create mode 100644 data/2020/neurips/CSI: Novelty Detection via Contrastive Learning on Distributionally Shifted Instances create mode 100644 data/2020/neurips/CaSPR: Learning Canonical Spatiotemporal Point Cloud Representations create mode 100644 data/2020/neurips/Calibrated Reliable Regression using Maximum Mean Discrepancy create mode 100644 data/2020/neurips/Calibrating CNNs for Lifelong Learning create mode 100644 data/2020/neurips/Calibrating Deep Neural Networks using Focal Loss create mode 100644 data/2020/neurips/Calibration of Shared Equilibria in General Sum Partially Observable Markov Games create mode 100644 data/2020/neurips/Can Graph Neural Networks Count Substructures? create mode 100644 data/2020/neurips/Can I Trust My Fairness Metric? Assessing Fairness with Unlabeled Data and Bayesian Inference create mode 100644 data/2020/neurips/Can Implicit Bias Explain Generalization? Stochastic Convex Optimization as a Case Study create mode 100644 data/2020/neurips/Can Q-Learning with Graph Networks Learn a Generalizable Branching Heuristic for a SAT Solver? create mode 100644 data/2020/neurips/Can Temporal-Difference and Q-Learning Learn Representation? A Mean-Field Theory create mode 100644 data/2020/neurips/Can the Brain Do Backpropagation? - Exact Implementation of Backpropagation in Predictive Coding Networks create mode 100644 data/2020/neurips/Canonical 3D Deformer Maps: Unifying parametric and non-parametric methods for dense weakly-supervised category reconstruction create mode 100644 data/2020/neurips/Cascaded Text Generation with Markov Transformers create mode 100644 data/2020/neurips/Causal Discovery from Soft Interventions with Unknown Targets: Characterization and Learning create mode 100644 data/2020/neurips/Causal Discovery in Physical Systems from Videos create mode 100644 data/2020/neurips/Causal Estimation with Functional Confounders create mode 100644 data/2020/neurips/Causal Intervention for Weakly-Supervised Semantic Segmentation create mode 100644 data/2020/neurips/Causal Shapley Values: Exploiting Causal Knowledge to Explain Individual Predictions of Complex Models create mode 100644 data/2020/neurips/Causal analysis of Covid-19 Spread in Germany create mode 100644 data/2020/neurips/Certifiably Adversarially Robust Detection of Out-of-Distribution Data create mode 100644 data/2020/neurips/Certified Defense to Image Transformations via Randomized Smoothing create mode 100644 data/2020/neurips/Certified Monotonic Neural Networks create mode 100644 data/2020/neurips/Certified Robustness of Graph Convolution Networks for Graph Classification under Topological Attacks create mode 100644 data/2020/neurips/Certifying Confidence via Randomized Smoothing create mode 100644 data/2020/neurips/Certifying Strategyproof Auction Networks create mode 100644 data/2020/neurips/Chaos, Extremism and Optimism: Volume Analysis of Learning in Games create mode 100644 data/2020/neurips/Characterizing Optimal Mixed Policies: Where to Intervene and What to Observe create mode 100644 data/2020/neurips/Characterizing emergent representations in a space of candidate learning rules for deep networks create mode 100644 data/2020/neurips/Choice Bandits create mode 100644 data/2020/neurips/CircleGAN: Generative Adversarial Learning across Spherical Circles create mode 100644 data/2020/neurips/Classification Under Misspecification: Halfspaces, Generalized Linear Models, and Evolvability create mode 100644 data/2020/neurips/Classification with Valid and Adaptive Coverage create mode 100644 data/2020/neurips/Closing the Dequantization Gap: PixelCNN as a Single-Layer Flow create mode 100644 data/2020/neurips/Co-Tuning for Transfer Learning create mode 100644 data/2020/neurips/Co-exposure Maximization in Online Social Networks create mode 100644 data/2020/neurips/CoADNet: Collaborative Aggregation-and-Distribution Networks for Co-Salient Object Detection create mode 100644 data/2020/neurips/CoMIR: Contrastive Multimodal Image Representation for Registration create mode 100644 data/2020/neurips/CoSE: Compositional Stroke Embeddings create mode 100644 data/2020/neurips/CodeCMR: Cross-Modal Retrieval For Function-Level Binary Source Code Matching create mode 100644 data/2020/neurips/Coded Sequential Matrix Multiplication For Straggler Mitigation create mode 100644 data/2020/neurips/CogLTX: Applying BERT to Long Texts create mode 100644 data/2020/neurips/CogMol: Target-Specific and Selective Drug Design for COVID-19 Using Deep Generative Models create mode 100644 data/2020/neurips/Coherent Hierarchical Multi-Label Classification Networks create mode 100644 data/2020/neurips/CoinDICE: Off-Policy Confidence Interval Estimation create mode 100644 data/2020/neurips/CoinPress: Practical Private Mean and Covariance Estimation create mode 100644 data/2020/neurips/ColdGANs: Taming Language GANs with Cautious Sampling Strategies create mode 100644 data/2020/neurips/Collapsing Bandits and Their Application to Public Health Intervention create mode 100644 data/2020/neurips/Collegial Ensembles create mode 100644 data/2020/neurips/Color Visual Illusions: A Statistics-based Computational Model create mode 100644 data/2020/neurips/Combining Deep Reinforcement Learning and Search for Imperfect-Information Games create mode 100644 data/2020/neurips/Community detection in sparse time-evolving graphs with a dynamical Bethe-Hessian create mode 100644 "data/2020/neurips/Community detection using fast low-cardinality semidefinite programming\342\200\251" create mode 100644 data/2020/neurips/Compact task representations as a normative model for higher-order brain activity create mode 100644 data/2020/neurips/Comparator-Adaptive Convex Bandits create mode 100644 data/2020/neurips/Complex Dynamics in Simple Neural Networks: Understanding Gradient Flow in Phase Retrieval create mode 100644 data/2020/neurips/Compositional Explanations of Neurons create mode 100644 data/2020/neurips/Compositional Generalization by Learning Analytical Expressions create mode 100644 data/2020/neurips/Compositional Generalization via Neural-Symbolic Stack Machines create mode 100644 data/2020/neurips/Compositional Visual Generation with Energy Based Models create mode 100644 data/2020/neurips/Compositional Zero-Shot Learning via Fine-Grained Dense Feature Composition create mode 100644 data/2020/neurips/Comprehensive Attention Self-Distillation for Weakly-Supervised Object Detection create mode 100644 data/2020/neurips/Compressing Images by Encoding Their Latent Representations with Relative Entropy Coding create mode 100644 data/2020/neurips/Computing Valid p-value for Optimal Changepoint by Selective Inference using Dynamic Programming create mode 100644 data/2020/neurips/Conditioning and Processing: Techniques to Improve Information-Theoretic Generalization Bounds create mode 100644 data/2020/neurips/Confidence sequences for sampling without replacement create mode 100644 data/2020/neurips/Conformal Symplectic and Relativistic Optimization create mode 100644 data/2020/neurips/Confounding-Robust Policy Evaluation in Infinite-Horizon Reinforcement Learning create mode 100644 data/2020/neurips/Conic Descent and its Application to Memory-efficient Optimization over Positive Semidefinite Matrices create mode 100644 data/2020/neurips/Consequences of Misaligned AI create mode 100644 data/2020/neurips/Conservative Q-Learning for Offline Reinforcement Learning create mode 100644 data/2020/neurips/Consistency Regularization for Certified Robustness of Smoothed Classifiers create mode 100644 data/2020/neurips/Consistent Estimation of Identifiable Nonparametric Mixture Models from Grouped Observations create mode 100644 data/2020/neurips/Consistent Plug-in Classifiers for Complex Objectives and Constraints create mode 100644 data/2020/neurips/Consistent Structural Relation Learning for Zero-Shot Segmentation create mode 100644 data/2020/neurips/Consistent feature selection for analytic deep neural networks create mode 100644 data/2020/neurips/Constant-Expansion Suffices for Compressed Sensing with Generative Priors create mode 100644 data/2020/neurips/Constrained episodic reinforcement learning in concave-convex and knapsack settings create mode 100644 data/2020/neurips/Constraining Variational Inference with Geometric Jensen-Shannon Divergence create mode 100644 data/2020/neurips/Content Provider Dynamics and Coordination in Recommendation Ecosystems create mode 100644 data/2020/neurips/Contextual Games: Multi-Agent Learning with Side Information create mode 100644 data/2020/neurips/Contextual Reserve Price Optimization in Auctions via Mixed Integer Programming create mode 100644 data/2020/neurips/Continual Deep Learning by Functional Regularisation of Memorable Past create mode 100644 data/2020/neurips/Continual Learning in Low-rank Orthogonal Subspaces create mode 100644 data/2020/neurips/Continual Learning of Control Primitives : Skill Discovery via Reset-Games create mode 100644 data/2020/neurips/Continual Learning of a Mixed Sequence of Similar and Dissimilar Tasks create mode 100644 data/2020/neurips/Continual Learning with Node-Importance based Adaptive Group Sparse Regularization create mode 100644 data/2020/neurips/Continuous Meta-Learning without Tasks create mode 100644 data/2020/neurips/Continuous Object Representation Networks: Novel View Synthesis without Target View Supervision create mode 100644 data/2020/neurips/Continuous Regularized Wasserstein Barycenters create mode 100644 data/2020/neurips/Continuous Submodular Maximization: Beyond DR-Submodularity create mode 100644 data/2020/neurips/Continuous Surface Embeddings create mode 100644 data/2020/neurips/ContraGAN: Contrastive Learning for Conditional Image Generation create mode 100644 data/2020/neurips/Contrastive Learning with Adversarial Examples create mode 100644 data/2020/neurips/Contrastive learning of global and local features for medical image segmentation with limited annotations create mode 100644 data/2020/neurips/ConvBERT: Improving BERT with Span-based Dynamic Convolution create mode 100644 data/2020/neurips/Convergence and Stability of Graph Convolutional Networks on Large Random Graphs create mode 100644 data/2020/neurips/Convergence of Meta-Learning with Task-Specific Adaptation over Partial Parameters create mode 100644 data/2020/neurips/Convex optimization based on global lower second-order models create mode 100644 data/2020/neurips/Convolutional Generation of Textured 3D Meshes create mode 100644 data/2020/neurips/Convolutional Tensor-Train LSTM for Spatio-Temporal Learning create mode 100644 data/2020/neurips/Cooperative Heterogeneous Deep Reinforcement Learning create mode 100644 data/2020/neurips/Cooperative Multi-player Bandit Optimization create mode 100644 data/2020/neurips/Coresets for Near-Convex Functions create mode 100644 data/2020/neurips/Coresets for Regressions with Panel Data create mode 100644 data/2020/neurips/Coresets for Robust Training of Deep Neural Networks against Noisy Labels create mode 100644 data/2020/neurips/Coresets via Bilevel Optimization for Continual Learning and Streaming create mode 100644 data/2020/neurips/Correlation Robust Influence Maximization create mode 100644 data/2020/neurips/Correspondence learning via linearly-invariant embedding create mode 100644 data/2020/neurips/Counterexample-Guided Learning of Monotonic Neural Networks create mode 100644 data/2020/neurips/Counterfactual Contrastive Learning for Weakly-Supervised Vision-Language Grounding create mode 100644 data/2020/neurips/Counterfactual Data Augmentation using Locally Factored Dynamics create mode 100644 data/2020/neurips/Counterfactual Prediction for Bundle Treatment create mode 100644 data/2020/neurips/Counterfactual Predictions under Runtime Confounding create mode 100644 data/2020/neurips/Counterfactual Vision-and-Language Navigation: Unravelling the Unseen create mode 100644 data/2020/neurips/Coupling-based Invertible Neural Networks Are Universal Diffeomorphism Approximators create mode 100644 data/2020/neurips/Cream of the Crop: Distilling Prioritized Paths For One-Shot Neural Architecture Search create mode 100644 data/2020/neurips/Critic Regularized Regression create mode 100644 data/2020/neurips/Cross-Scale Internal Graph Neural Network for Image Super-Resolution create mode 100644 data/2020/neurips/Cross-lingual Retrieval for Iterative Self-Supervised Training create mode 100644 data/2020/neurips/Cross-validation Confidence Intervals for Test Error create mode 100644 data/2020/neurips/CrossTransformers: spatially-aware few-shot transfer create mode 100644 data/2020/neurips/Crush Optimism with Pessimism: Structured Bandits Beyond Asymptotic Optimality create mode 100644 data/2020/neurips/Curriculum By Smoothing create mode 100644 data/2020/neurips/Curriculum Learning by Dynamic Instance Hardness create mode 100644 data/2020/neurips/Curriculum learning for multilevel budgeted combinatorial problems create mode 100644 data/2020/neurips/Curvature Regularization to Prevent Distortion in Graph Embedding create mode 100644 data/2020/neurips/Cycle-Contrast for Self-Supervised Video Representation Learning create mode 100644 data/2020/neurips/DAGs with No Fears: A Closer Look at Continuous Optimization for Learning Bayesian Networks create mode 100644 data/2020/neurips/DISK: Learning local features with policy gradient create mode 100644 data/2020/neurips/DVERGE: Diversifying Vulnerabilities for Enhanced Robust Generation of Ensembles create mode 100644 data/2020/neurips/Dark Experience for General Continual Learning: a Strong, Simple Baseline create mode 100644 data/2020/neurips/Data Diversification: A Simple Strategy For Neural Machine Translation create mode 100644 data/2020/neurips/De-Anonymizing Text by Fingerprinting Language Generation create mode 100644 data/2020/neurips/Debiased Contrastive Learning create mode 100644 data/2020/neurips/Debiasing Averaged Stochastic Gradient Descent to handle missing values create mode 100644 data/2020/neurips/Debiasing Distributed Second Order Optimization with Surrogate Sketching and Scaled Regularization create mode 100644 data/2020/neurips/Debugging Tests for Model Explanations create mode 100644 data/2020/neurips/Decentralized Accelerated Proximal Gradient Descent create mode 100644 data/2020/neurips/Decentralized Langevin Dynamics for Bayesian Learning create mode 100644 data/2020/neurips/Decentralized TD Tracking with Linear Function Approximation and its Finite-Time Analysis create mode 100644 data/2020/neurips/Decision trees as partitioning machines to characterize their generalization properties create mode 100644 data/2020/neurips/Decision-Making with Auto-Encoding Variational Bayes create mode 100644 data/2020/neurips/Decisions, Counterfactual Explanations and Strategic Behavior create mode 100644 data/2020/neurips/Deep Archimedean Copulas create mode 100644 data/2020/neurips/Deep Automodulators create mode 100644 data/2020/neurips/Deep Diffusion-Invariant Wasserstein Distributional Classification create mode 100644 data/2020/neurips/Deep Direct Likelihood Knockoffs create mode 100644 data/2020/neurips/Deep Energy-based Modeling of Discrete-Time Physics create mode 100644 data/2020/neurips/Deep Evidential Regression create mode 100644 data/2020/neurips/Deep Graph Pose: a semi-supervised deep graphical model for improved animal pose tracking create mode 100644 data/2020/neurips/Deep Imitation Learning for Bimanual Robotic Manipulation create mode 100644 data/2020/neurips/Deep Metric Learning with Spherical Embedding create mode 100644 data/2020/neurips/Deep Multimodal Fusion by Channel Exchanging create mode 100644 data/2020/neurips/Deep Rao-Blackwellised Particle Filters for Time Series Forecasting create mode 100644 data/2020/neurips/Deep Reinforcement Learning with Stacked Hierarchical Attention for Text-based Games create mode 100644 data/2020/neurips/Deep Reinforcement and InfoMax Learning create mode 100644 data/2020/neurips/Deep Relational Topic Modeling via Graph Poisson Gamma Belief Network create mode 100644 data/2020/neurips/Deep Shells: Unsupervised Shape Correspondence with Optimal Transport create mode 100644 data/2020/neurips/Deep Smoothing of the Implied Volatility Surface create mode 100644 data/2020/neurips/Deep Statistical Solvers create mode 100644 data/2020/neurips/Deep Structural Causal Models for Tractable Counterfactual Inference create mode 100644 data/2020/neurips/Deep Subspace Clustering with Data Augmentation create mode 100644 data/2020/neurips/Deep Transformation-Invariant Clustering create mode 100644 data/2020/neurips/Deep Transformers with Latent Depth create mode 100644 data/2020/neurips/Deep Variational Instance Segmentation create mode 100644 data/2020/neurips/Deep Wiener Deconvolution: Wiener Meets Deep Learning for Image Deblurring create mode 100644 data/2020/neurips/Deep active inference agents using Monte-Carlo methods create mode 100644 data/2020/neurips/Deep learning versus kernel learning: an empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel create mode 100644 data/2020/neurips/Deep reconstruction of strange attractors from time series create mode 100644 data/2020/neurips/DeepI2I: Enabling Deep Hierarchical Image-to-Image Translation by Transferring from GANs create mode 100644 data/2020/neurips/DeepSVG: A Hierarchical Generative Network for Vector Graphics Animation create mode 100644 data/2020/neurips/Deeply Learned Spectral Total Variation Decomposition create mode 100644 data/2020/neurips/Delay and Cooperation in Nonstochastic Linear Bandits create mode 100644 data/2020/neurips/Delta-STN: Efficient Bilevel Optimization for Neural Networks using Structured Response Jacobians create mode 100644 data/2020/neurips/Delving into the Cyclic Mechanism in Semi-supervised Video Object Segmentation create mode 100644 data/2020/neurips/Demixed shared component analysis of neural population data from multiple brain areas create mode 100644 data/2020/neurips/Demystifying Orthogonal Monte Carlo and Beyond create mode 100644 data/2020/neurips/Denoised Smoothing: A Provable Defense for Pretrained Classifiers create mode 100644 data/2020/neurips/Denoising Diffusion Probabilistic Models create mode 100644 data/2020/neurips/Dense Correspondences between Human Bodies via Learning Transformation Synchronization on Graphs create mode 100644 data/2020/neurips/Depth Uncertainty in Neural Networks create mode 100644 data/2020/neurips/Design Space for Graph Neural Networks create mode 100644 data/2020/neurips/Detecting Hands and Recognizing Physical Contact in the Wild create mode 100644 data/2020/neurips/Detecting Interactions from Neural Networks via Topological Analysis create mode 100644 data/2020/neurips/Detection as Regression: Certified Object Detection with Median Smoothing create mode 100644 data/2020/neurips/Deterministic Approximation for Submodular Maximization over a Matroid in Nearly Linear Time create mode 100644 data/2020/neurips/Dialog without Dialog Data: Learning Visual Dialog Agents from VQA Data create mode 100644 data/2020/neurips/DiffGCN: Graph Convolutional Networks via Differential Operators and Algebraic Multigrid Pooling create mode 100644 data/2020/neurips/Differentiable Augmentation for Data-Efficient GAN Training create mode 100644 data/2020/neurips/Differentiable Causal Discovery from Interventional Data create mode 100644 data/2020/neurips/Differentiable Expected Hypervolume Improvement for Parallel Multi-Objective Bayesian Optimization create mode 100644 data/2020/neurips/Differentiable Meta-Learning of Bandit Policies create mode 100644 data/2020/neurips/Differentiable Neural Architecture Search in Equivalent Space with Exploration Enhancement create mode 100644 data/2020/neurips/Differentiable Top-k with Optimal Transport create mode 100644 data/2020/neurips/Differentially Private Clustering: Tight Approximation Ratios create mode 100644 data/2020/neurips/Differentially-Private Federated Linear Bandits create mode 100644 data/2020/neurips/Digraph Inception Convolutional Networks create mode 100644 data/2020/neurips/Direct Feedback Alignment Scales to Modern Deep Learning Tasks and Architectures create mode 100644 data/2020/neurips/Direct Policy Gradients: Direct Optimization of Policies in Discrete Action Spaces create mode 100644 data/2020/neurips/Directional Pruning of Deep Neural Networks create mode 100644 data/2020/neurips/Directional convergence and alignment in deep learning create mode 100644 data/2020/neurips/Dirichlet Graph Variational Autoencoder create mode 100644 data/2020/neurips/DisARM: An Antithetic Gradient Estimator for Binary Latent Variables create mode 100644 data/2020/neurips/DisCor: Corrective Feedback in Reinforcement Learning via Distribution Correction create mode 100644 data/2020/neurips/Discovering Reinforcement Learning Algorithms create mode 100644 data/2020/neurips/Discovering Symbolic Models from Deep Learning with Inductive Biases create mode 100644 data/2020/neurips/Discovering conflicting groups in signed networks create mode 100644 data/2020/neurips/Discriminative Sounding Objects Localization via Self-supervised Audiovisual Matching create mode 100644 data/2020/neurips/Disentangling Human Error from Ground Truth in Segmentation of Medical Images create mode 100644 data/2020/neurips/Disentangling by Subspace Diffusion create mode 100644 data/2020/neurips/Displacement-Invariant Matching Cost Learning for Accurate Optical Flow Estimation create mode 100644 data/2020/neurips/Dissecting Neural ODEs create mode 100644 data/2020/neurips/Distance Encoding: Design Provably More Powerful Neural Networks for Graph Representation Learning create mode 100644 data/2020/neurips/Distributed Distillation for On-Device Learning create mode 100644 data/2020/neurips/Distributed Newton Can Communicate Less and Resist Byzantine Workers create mode 100644 data/2020/neurips/Distributed Training with Heterogeneous Data: Bridging Median- and Mean-Based Algorithms create mode 100644 data/2020/neurips/Distribution Aligning Refinery of Pseudo-label for Imbalanced Semi-supervised Learning create mode 100644 data/2020/neurips/Distribution Matching for Crowd Counting create mode 100644 data/2020/neurips/Distribution-free binary classification: prediction sets, confidence intervals and calibration create mode 100644 data/2020/neurips/Distributional Robustness with IPMs and links to Regularization and GANs create mode 100644 data/2020/neurips/Distributionally Robust Federated Averaging create mode 100644 data/2020/neurips/Distributionally Robust Local Non-parametric Conditional Estimation create mode 100644 data/2020/neurips/Distributionally Robust Parametric Maximum Likelihood Estimation create mode 100644 data/2020/neurips/Diverse Image Captioning with Context-Object Split Latent Spaces create mode 100644 data/2020/neurips/Diversity can be Transferred: Output Diversification for White- and Black-box Attacks create mode 100644 data/2020/neurips/Diversity-Guided Multi-Objective Bayesian Optimization With Batch Evaluations create mode 100644 data/2020/neurips/Do Adversarially Robust ImageNet Models Transfer Better? create mode 100644 data/2020/neurips/Does Unsupervised Architecture Representation Learning Help Neural Architecture Search? create mode 100644 data/2020/neurips/Domain Adaptation as a Problem of Inference on Graphical Models create mode 100644 data/2020/neurips/Domain Adaptation with Conditional Distribution Matching and Generalized Label Shift create mode 100644 data/2020/neurips/Domain Generalization for Medical Imaging Classification with Linear-Dependency Regularization create mode 100644 data/2020/neurips/Domain Generalization via Entropy Regularization create mode 100644 data/2020/neurips/Doubly Robust Off-Policy Value and Gradient Estimation for Deterministic Policies create mode 100644 data/2020/neurips/Dual Instrumental Variable Regression create mode 100644 data/2020/neurips/Dual Manifold Adversarial Robustness: Defense against Lp and non-Lp Adversarial Attacks create mode 100644 data/2020/neurips/Dual T: Reducing Estimation Error for Transition Matrix in Label-noise Learning create mode 100644 data/2020/neurips/Dual-Free Stochastic Decentralized Optimization with Variance Reduction create mode 100644 data/2020/neurips/Dual-Resolution Correspondence Networks create mode 100644 data/2020/neurips/Duality-Induced Regularizer for Tensor Factorization Based Knowledge Graph Completion create mode 100644 data/2020/neurips/DynaBERT: Dynamic BERT with Adaptive Width and Depth create mode 100644 data/2020/neurips/Dynamic Fusion of Eye Movement Data and Verbal Narrations in Knowledge-rich Domains create mode 100644 data/2020/neurips/Dynamic Regret of Convex and Smooth Functions create mode 100644 data/2020/neurips/Dynamic Regret of Policy Optimization in Non-Stationary Environments create mode 100644 data/2020/neurips/Dynamic Submodular Maximization create mode 100644 data/2020/neurips/Dynamic allocation of limited memory resources in reinforcement learning create mode 100644 data/2020/neurips/Dynamical mean-field theory for stochastic gradient descent in Gaussian mixture classification create mode 100644 data/2020/neurips/Early-Learning Regularization Prevents Memorization of Noisy Labels create mode 100644 data/2020/neurips/EcoLight: Intersection Control in Developing Regions Under Extreme Budget and Network Constraints create mode 100644 data/2020/neurips/Effective Dimension Adaptive Sketching Methods for Faster Regularized Least-Squares Optimization create mode 100644 data/2020/neurips/Effective Diversity in Population Based Reinforcement Learning create mode 100644 data/2020/neurips/Efficient Algorithms for Device Placement of DNN Graph Operators create mode 100644 data/2020/neurips/Efficient Clustering Based On A Unified View Of $K$-means And Ratio-cut create mode 100644 data/2020/neurips/Efficient Clustering for Stretched Mixtures: Landscape and Optimality create mode 100644 data/2020/neurips/Efficient Contextual Bandits with Continuous Actions create mode 100644 data/2020/neurips/Efficient Distance Approximation for Structured High-Dimensional Distributions via Learning create mode 100644 data/2020/neurips/Efficient Exact Verification of Binarized Neural Networks create mode 100644 data/2020/neurips/Efficient Exploration of Reward Functions in Inverse Reinforcement Learning via Bayesian Optimization create mode 100644 data/2020/neurips/Efficient Generation of Structured Objects with Constrained Adversarial Networks create mode 100644 data/2020/neurips/Efficient Learning of Discrete Graphical Models create mode 100644 data/2020/neurips/Efficient Learning of Generative Models via Finite-Difference Score Matching create mode 100644 data/2020/neurips/Efficient Low Rank Gaussian Variational Inference for Neural Networks create mode 100644 data/2020/neurips/Efficient Marginalization of Discrete and Structured Latent Variables via Sparsity create mode 100644 data/2020/neurips/Efficient Model-Based Reinforcement Learning through Optimistic Policy Search and Planning create mode 100644 data/2020/neurips/Efficient Nonmyopic Bayesian Optimization via One-Shot Multi-Step Trees create mode 100644 data/2020/neurips/Efficient Online Learning of Optimal Rankings: Dimensionality Reduction via Gradient Descent create mode 100644 data/2020/neurips/Efficient Planning in Large MDPs with Weak Linear Function Approximation create mode 100644 data/2020/neurips/Efficient Projection-free Algorithms for Saddle Point Problems create mode 100644 data/2020/neurips/Efficient Variational Inference for Sparse Deep Learning with Theoretical Guarantee create mode 100644 data/2020/neurips/Efficient active learning of sparse halfspaces with arbitrary bounded noise create mode 100644 data/2020/neurips/Efficient estimation of neural tuning during naturalistic behavior create mode 100644 data/2020/neurips/Efficient semidefinite-programming-based inference for binary and multi-class MRFs create mode 100644 data/2020/neurips/Elastic-InfoGAN: Unsupervised Disentangled Representation Learning in Class-Imbalanced Data create mode 100644 data/2020/neurips/Election Coding for Distributed Learning: Protecting SignSGD against Byzantine Attacks create mode 100644 data/2020/neurips/Emergent Complexity and Zero-shot Transfer via Unsupervised Environment Design create mode 100644 data/2020/neurips/Emergent Reciprocity and Team Formation from Randomized Uncertain Social Preferences create mode 100644 data/2020/neurips/Empirical Likelihood for Contextual Bandits create mode 100644 data/2020/neurips/Enabling certification of verification-agnostic networks via memory-efficient semidefinite programming create mode 100644 data/2020/neurips/End-to-End Learning and Intervention in Games create mode 100644 data/2020/neurips/Energy-based Out-of-distribution Detection create mode 100644 data/2020/neurips/Ensemble Distillation for Robust Model Fusion in Federated Learning create mode 100644 data/2020/neurips/Ensembling geophysical models with Bayesian Neural Networks create mode 100644 data/2020/neurips/Ensuring Fairness Beyond the Training Data create mode 100644 data/2020/neurips/Entropic Causal Inference: Identifiability and Finite Sample Results create mode 100644 data/2020/neurips/Entropic Optimal Transport between Unbalanced Gaussian Measures has a Closed Form create mode 100644 data/2020/neurips/Entrywise convergence of iterative methods for eigenproblems create mode 100644 data/2020/neurips/Equivariant Networks for Hierarchical Structures create mode 100644 data/2020/neurips/Erdos Goes Neural: an Unsupervised Learning Framework for Combinatorial Optimization on Graphs create mode 100644 data/2020/neurips/Error Bounds of Imitating Policies and Environments create mode 100644 data/2020/neurips/Escaping Saddle-Point Faster under Interpolation-like Conditions create mode 100644 data/2020/neurips/Escaping the Gravitational Pull of Softmax create mode 100644 data/2020/neurips/Estimating Fluctuations in Neural Representations of Uncertain Environments create mode 100644 data/2020/neurips/Estimating Rank-One Spikes from Heavy-Tailed Noise via Self-Avoiding Walks create mode 100644 data/2020/neurips/Estimating Training Data Influence by Tracing Gradient Descent create mode 100644 data/2020/neurips/Estimating decision tree learnability with polylogarithmic sample complexity create mode 100644 data/2020/neurips/Estimating the Effects of Continuous-valued Interventions using Generative Adversarial Networks create mode 100644 data/2020/neurips/Estimating weighted areas under the ROC curve create mode 100644 data/2020/neurips/Estimation of Skill Distribution from a Tournament create mode 100644 data/2020/neurips/Evaluating Attribution for Graph Neural Networks create mode 100644 data/2020/neurips/Evaluating and Rewarding Teamwork Using Cooperative Game Abstractions create mode 100644 data/2020/neurips/Every View Counts: Cross-View Consistency in 3D Object Detection with Hybrid-Cylindrical-Spherical Voxelization create mode 100644 data/2020/neurips/Evidential Sparsification of Multimodal Latent Spaces in Conditional Variational Autoencoders create mode 100644 data/2020/neurips/EvolveGraph: Multi-Agent Trajectory Prediction with Dynamic Relational Reasoning create mode 100644 data/2020/neurips/Evolving Graphical Planner: Contextual Global Planning for Vision-and-Language Navigation create mode 100644 data/2020/neurips/Evolving Normalization-Activation Layers create mode 100644 data/2020/neurips/Exact Recovery of Mangled Clusters with Same-Cluster Queries create mode 100644 data/2020/neurips/Exact expressions for double descent and implicit regularization via surrogate random design create mode 100644 data/2020/neurips/Exactly Computing the Local Lipschitz Constant of ReLU Networks create mode 100644 data/2020/neurips/Exchangeable Neural ODE for Set Modeling create mode 100644 data/2020/neurips/Exemplar Guided Active Learning create mode 100644 data/2020/neurips/Exemplar VAE: Linking Generative Models, Nearest Neighbor Retrieval, and Data Augmentation create mode 100644 data/2020/neurips/ExpandNets: Linear Over-parameterization to Train Compact Convolutional Networks create mode 100644 data/2020/neurips/Experimental design for MRI by greedy policy search create mode 100644 data/2020/neurips/Expert-Supervised Reinforcement Learning for Offline Policy Learning and Evaluation create mode 100644 data/2020/neurips/Explainable Voting create mode 100644 data/2020/neurips/Explaining Naive Bayes and Other Linear Classifiers with Polynomial Time and Delay create mode 100644 data/2020/neurips/Explicit Regularisation in Gaussian Noise Injections create mode 100644 data/2020/neurips/Exploiting Higher Order Smoothness in Derivative-free Optimization and Continuous Bandits create mode 100644 data/2020/neurips/Exploiting MMD and Sinkhorn Divergences for Fair and Transferable Representation Learning create mode 100644 data/2020/neurips/Exploiting the Surrogate Gap in Online Multiclass Classification create mode 100644 data/2020/neurips/Exploiting weakly supervised visual patterns to learn from partial annotations create mode 100644 data/2020/neurips/Explore Aggressively, Update Conservatively: Stochastic Extragradient Methods with Variable Stepsize Scaling create mode 100644 data/2020/neurips/Extrapolation Towards Imaginary 0-Nearest Neighbour and Its Improved Convergence Rate create mode 100644 data/2020/neurips/FLAMBE: Structural Complexity and Representation Learning of Low Rank MDPs create mode 100644 data/2020/neurips/Factor Graph Grammars create mode 100644 data/2020/neurips/Factor Graph Neural Networks create mode 100644 data/2020/neurips/Factorizable Graph Convolutional Networks create mode 100644 data/2020/neurips/Factorized Neural Processes for Neural Processes: K-Shot Prediction of Neural Responses create mode 100644 data/2020/neurips/Fair Hierarchical Clustering create mode 100644 data/2020/neurips/Fair Multiple Decision Making Through Soft Interventions create mode 100644 data/2020/neurips/Fair Performance Metric Elicitation create mode 100644 data/2020/neurips/Fair regression via plug-in estimator and recalibration with statistical guarantees create mode 100644 data/2020/neurips/Fair regression with Wasserstein barycenters create mode 100644 data/2020/neurips/Fairness constraints can help exact inference in structured prediction create mode 100644 data/2020/neurips/Fairness in Streaming Submodular Maximization: Algorithms and Hardness create mode 100644 data/2020/neurips/Fairness with Overlapping Groups; a Probabilistic Perspective create mode 100644 data/2020/neurips/Fairness without Demographics through Adversarially Reweighted Learning create mode 100644 data/2020/neurips/Faithful Embeddings for Knowledge Base Queries create mode 100644 data/2020/neurips/Falcon: Fast Spectral Inference on Encrypted Data create mode 100644 data/2020/neurips/Fast Adaptive Non-Monotone Submodular Maximization Subject to a Knapsack Constraint create mode 100644 data/2020/neurips/Fast Convergence of Langevin Dynamics on Manifold: Geodesics meet Log-Sobolev create mode 100644 data/2020/neurips/Fast Epigraphical Projection-based Incremental Algorithms for Wasserstein Distributionally Robust Support Vector Machine create mode 100644 data/2020/neurips/Fast Fourier Convolution create mode 100644 data/2020/neurips/Fast Matrix Square Roots with Applications to Gaussian Processes and Bayesian Optimization create mode 100644 data/2020/neurips/Fast Transformers with Clustered Attention create mode 100644 data/2020/neurips/Fast Unbalanced Optimal Transport on a Tree create mode 100644 data/2020/neurips/Fast and Accurate $k$-means++ via Rejection Sampling create mode 100644 data/2020/neurips/Fast and Flexible Temporal Point Processes with Triangular Maps create mode 100644 data/2020/neurips/Fast geometric learning with symbolic matrices create mode 100644 data/2020/neurips/Fast, Accurate, and Simple Models for Tabular Data via Augmented Distillation create mode 100644 data/2020/neurips/Faster DBSCAN via subsampled similarity queries create mode 100644 "data/2020/neurips/Faster Differentially Private Samplers via R\303\251nyi Divergence Analysis of Discretized Langevin MCMC" create mode 100644 data/2020/neurips/Faster Randomized Infeasible Interior Point Methods for Tall Wide Linear Programs create mode 100644 data/2020/neurips/Faster Wasserstein Distance Estimation with the Sinkhorn Divergence create mode 100644 data/2020/neurips/Feature Importance Ranking for Deep Learning create mode 100644 data/2020/neurips/Feature Shift Detection: Localizing Which Features Have Shifted via Conditional Distribution Tests create mode 100644 data/2020/neurips/FedSplit: an algorithmic framework for fast federated optimization create mode 100644 data/2020/neurips/Federated Accelerated Stochastic Gradient Descent create mode 100644 data/2020/neurips/Federated Bayesian Optimization via Thompson Sampling create mode 100644 data/2020/neurips/Federated Principal Component Analysis create mode 100644 data/2020/neurips/Few-Cost Salient Object Detection with Adversarial-Paced Learning create mode 100644 data/2020/neurips/Few-shot Image Generation with Elastic Weight Consolidation create mode 100644 data/2020/neurips/Few-shot Visual Reasoning with Meta-Analogical Contrastive Learning create mode 100644 data/2020/neurips/Fewer is More: A Deep Graph Metric Learning Perspective Using Fewer Proxies create mode 100644 data/2020/neurips/Fictitious Play for Mean Field Games: Continuous Time Analysis and Applications create mode 100644 data/2020/neurips/Field-wise Learning for Multi-field Categorical Data create mode 100644 data/2020/neurips/Fighting Copycat Agents in Behavioral Cloning from Observation Histories create mode 100644 data/2020/neurips/Finding Second-Order Stationary Points Efficiently in Smooth Nonconvex Linearly Constrained Optimization Problems create mode 100644 data/2020/neurips/Finding the Homology of Decision Boundaries with Active Learning create mode 100644 data/2020/neurips/Fine-Grained Dynamic Head for Object Detection create mode 100644 data/2020/neurips/Finer Metagenomic Reconstruction via Biodiversity Optimization create mode 100644 data/2020/neurips/Finite Continuum-Armed Bandits create mode 100644 data/2020/neurips/Finite Versus Infinite Neural Networks: an Empirical Study create mode 100644 data/2020/neurips/Finite-Time Analysis for Double Q-learning create mode 100644 data/2020/neurips/Finite-Time Analysis of Round-Robin Kullback-Leibler Upper Confidence Bounds for Optimal Adaptive Allocation with Multiple Plays and Markovian Rewards create mode 100644 data/2020/neurips/Firefly Neural Architecture Descent: a General Approach for Growing Neural Networks create mode 100644 data/2020/neurips/First Order Constrained Optimization in Policy Space create mode 100644 data/2020/neurips/First-Order Methods for Large-Scale Market Equilibrium Computation create mode 100644 data/2020/neurips/FixMatch: Simplifying Semi-Supervised Learning with Consistency and Confidence create mode 100644 data/2020/neurips/Fixed-Support Wasserstein Barycenters: Computational Hardness and Fast Algorithm create mode 100644 data/2020/neurips/FleXOR: Trainable Fractional Quantization create mode 100644 data/2020/neurips/Flexible mean field variational inference using mixtures of non-overlapping exponential families create mode 100644 data/2020/neurips/Flows for simultaneous manifold learning and density estimation create mode 100644 data/2020/neurips/Focus of Attention Improves Information Transfer in Visual Features create mode 100644 data/2020/neurips/Follow the Perturbed Leader: Optimism and Fast Parallel Algorithms for Smooth Minimax Games create mode 100644 data/2020/neurips/Forethought and Hindsight in Credit Assignment create mode 100644 data/2020/neurips/Forget About the LiDAR: Self-Supervised Depth Estimators with MED Probability Volumes create mode 100644 data/2020/neurips/Fourier Features Let Networks Learn High Frequency Functions in Low Dimensional Domains create mode 100644 data/2020/neurips/Fourier Sparse Leverage Scores and Approximate Kernel Learning create mode 100644 data/2020/neurips/Fourier Spectrum Discrepancies in Deep Network Generated Images create mode 100644 data/2020/neurips/Fourier-transform-based attribution priors improve the interpretability and stability of deep learning models for genomics create mode 100644 data/2020/neurips/FracTrain: Fractionally Squeezing Bit Savings Both Temporally and Spatially for Efficient DNN Training create mode 100644 data/2020/neurips/From Boltzmann Machines to Neural Networks and Back Again create mode 100644 data/2020/neurips/From Predictions to Decisions: Using Lookahead Regularization create mode 100644 data/2020/neurips/From Trees to Continuous Embeddings and Back: Hyperbolic Hierarchical Clustering create mode 100644 data/2020/neurips/FrugalML: How to use ML Prediction APIs more accurately and cheaply create mode 100644 data/2020/neurips/Fully Convolutional Mesh Autoencoder using Efficient Spatially Varying Kernels create mode 100644 data/2020/neurips/Fully Dynamic Algorithm for Constrained Submodular Optimization create mode 100644 data/2020/neurips/Functional Regularization for Representation Learning: A Unified Theoretical Perspective create mode 100644 data/2020/neurips/Funnel-Transformer: Filtering out Sequential Redundancy for Efficient Language Processing create mode 100644 data/2020/neurips/Further Analysis of Outlier Detection with Deep Generative Models create mode 100644 data/2020/neurips/GAIT-prop: A biologically plausible learning rule derived from backpropagation of error create mode 100644 data/2020/neurips/GAN Memory with No Forgetting create mode 100644 data/2020/neurips/GANSpace: Discovering Interpretable GAN Controls create mode 100644 "data/2020/neurips/GCN meets GPU: Decoupling \"When to Sample\" from \"How to Sample\"" create mode 100644 data/2020/neurips/GCOMB: Learning Budget-constrained Combinatorial Algorithms over Billion-sized Graphs create mode 100644 data/2020/neurips/GNNGuard: Defending Graph Neural Networks against Adversarial Attacks create mode 100644 data/2020/neurips/GOCor: Bringing Globally Optimized Correspondence Volumes into Your Neural Network create mode 100644 data/2020/neurips/GPS-Net: Graph-based Photometric Stereo Network create mode 100644 data/2020/neurips/GPU-Accelerated Primal Learning for Extremely Fast Large-Scale Classification create mode 100644 data/2020/neurips/GRAF: Generative Radiance Fields for 3D-Aware Image Synthesis create mode 100644 data/2020/neurips/GS-WGAN: A Gradient-Sanitized Approach for Learning Differentially Private Generators create mode 100644 data/2020/neurips/Gamma-Models: Generative Temporal Difference Learning for Infinite-Horizon Prediction create mode 100644 data/2020/neurips/Gaussian Gated Linear Networks create mode 100644 data/2020/neurips/Gaussian Process Bandit Optimization of the Thermodynamic Variational Objective create mode 100644 data/2020/neurips/General Control Functions for Causal Effect Estimation from IVs create mode 100644 data/2020/neurips/General Transportability of Soft Interventions: Completeness Results create mode 100644 data/2020/neurips/Generalised Bayesian Filtering via Sequential Monte Carlo create mode 100644 data/2020/neurips/Generalization Bound of Gradient Descent for Non-Convex Metric Learning create mode 100644 data/2020/neurips/Generalization bound of globally optimal non-convex neural network training: Transportation map estimation by infinite dimensional Langevin dynamics create mode 100644 data/2020/neurips/Generalization error in high-dimensional perceptrons: Approaching Bayes error with convex optimization create mode 100644 data/2020/neurips/Generalized Boosting create mode 100644 data/2020/neurips/Generalized Focal Loss: Learning Qualified and Distributed Bounding Boxes for Dense Object Detection create mode 100644 data/2020/neurips/Generalized Hindsight for Reinforcement Learning create mode 100644 data/2020/neurips/Generalized Independent Noise Condition for Estimating Latent Variable Causal Graphs create mode 100644 data/2020/neurips/Generalized Leverage Score Sampling for Neural Networks create mode 100644 data/2020/neurips/Generating Adjacency-Constrained Subgoals in Hierarchical Reinforcement Learning create mode 100644 data/2020/neurips/Generating Correct Answers for Progressive Matrices Intelligence Tests create mode 100644 data/2020/neurips/Generative 3D Part Assembly via Dynamic Graph Learning create mode 100644 data/2020/neurips/Generative Neurosymbolic Machines create mode 100644 data/2020/neurips/Generative View Synthesis: From Single-view Semantics to Novel-view Images create mode 100644 data/2020/neurips/Generative causal explanations of black-box classifiers create mode 100644 data/2020/neurips/Geo-PIFu: Geometry and Pixel Aligned Implicit Functions for Single-view Human Reconstruction create mode 100644 data/2020/neurips/Geometric All-way Boolean Tensor Decomposition create mode 100644 data/2020/neurips/Geometric Dataset Distances via Optimal Transport create mode 100644 data/2020/neurips/Geometric Exploration for Online Control create mode 100644 data/2020/neurips/Gibbs Sampling with People create mode 100644 data/2020/neurips/Glance and Focus: a Dynamic Approach to Reducing Spatial Redundancy in Image Classification create mode 100644 data/2020/neurips/Global Convergence and Variance Reduction for a Class of Nonconvex-Nonconcave Minimax Problems create mode 100644 data/2020/neurips/Global Convergence of Deep Networks with One Wide Layer Followed by Pyramidal Topology create mode 100644 data/2020/neurips/Glow-TTS: A Generative Flow for Text-to-Speech via Monotonic Alignment Search create mode 100644 data/2020/neurips/Glyph: Fast and Accurately Training Deep Neural Networks on Encrypted Data create mode 100644 data/2020/neurips/Goal-directed Generation of Discrete Structures with Conditional Generative Models create mode 100644 data/2020/neurips/GradAug: A New Regularization Method for Deep Neural Networks create mode 100644 data/2020/neurips/Gradient Boosted Normalizing Flows create mode 100644 data/2020/neurips/Gradient Estimation with Stochastic Softmax Tricks create mode 100644 data/2020/neurips/Gradient Regularized V-Learning for Dynamic Treatment Regimes create mode 100644 data/2020/neurips/Gradient Surgery for Multi-Task Learning create mode 100644 data/2020/neurips/Gradient-EM Bayesian Meta-Learning create mode 100644 data/2020/neurips/Graduated Assignment for Joint Multi-Graph Matching and Clustering with Application to Unsupervised Graph Matching Network Learning create mode 100644 data/2020/neurips/GramGAN: Deep 3D Texture Synthesis From 2D Exemplars create mode 100644 data/2020/neurips/Graph Cross Networks with Vertex Infomax Pooling create mode 100644 data/2020/neurips/Graph Geometry Interaction Learning create mode 100644 data/2020/neurips/Graph Information Bottleneck create mode 100644 data/2020/neurips/Graph Meta Learning via Local Subgraphs create mode 100644 data/2020/neurips/Graph Policy Network for Transferable Active Learning on Graphs create mode 100644 data/2020/neurips/Graph Random Neural Networks for Semi-Supervised Learning on Graphs create mode 100644 data/2020/neurips/Graph Stochastic Neural Networks for Semi-supervised Learning create mode 100644 data/2020/neurips/Graphon Neural Networks and the Transferability of Graph Neural Networks create mode 100644 data/2020/neurips/Grasp Proposal Networks: An End-to-End Solution for Visual Learning of Robotic Grasps create mode 100644 data/2020/neurips/Greedy Optimization Provably Wins the Lottery: Logarithmic Number of Winning Tickets is Enough create mode 100644 data/2020/neurips/Greedy inference with structure-exploiting lazy maps create mode 100644 data/2020/neurips/GreedyFool: Distortion-Aware Sparse Adversarial Attack create mode 100644 data/2020/neurips/Group Contextual Encoding for 3D Point Clouds create mode 100644 data/2020/neurips/Group Knowledge Transfer: Federated Learning of Large CNNs at the Edge create mode 100644 data/2020/neurips/Group-Fair Online Allocation in Continuous Time create mode 100644 data/2020/neurips/Guided Adversarial Attack for Evaluating and Enhancing Adversarial Defenses create mode 100644 data/2020/neurips/Guiding Deep Molecular Optimization with Genetic Exploration create mode 100644 data/2020/neurips/H-Mem: Harnessing synaptic plasticity with Hebbian Memory Networks create mode 100644 data/2020/neurips/HAWQ-V2: Hessian Aware trace-Weighted Quantization of Neural Networks create mode 100644 data/2020/neurips/HM-ANN: Efficient Billion-Point Nearest Neighbor Search on Heterogeneous Memory create mode 100644 data/2020/neurips/HOI Analysis: Integrating and Decomposing Human-Object Interaction create mode 100644 data/2020/neurips/HRN: A Holistic Approach to One Class Learning create mode 100644 data/2020/neurips/HYDRA: Pruning Adversarially Robust Neural Networks create mode 100644 data/2020/neurips/Hamiltonian Monte Carlo using an adjoint-differentiated Laplace approximation: Bayesian inference for latent Gaussian models and beyond create mode 100644 data/2020/neurips/Handling Missing Data with Graph Representation Learning create mode 100644 data/2020/neurips/Hard Example Generation by Texture Synthesis for Cross-domain Shape Similarity Learning create mode 100644 data/2020/neurips/Hard Negative Mixing for Contrastive Learning create mode 100644 data/2020/neurips/Hard Shape-Constrained Kernel Machines create mode 100644 data/2020/neurips/Hardness of Learning Neural Networks with Natural Weights create mode 100644 data/2020/neurips/Hausdorff Dimension, Heavy Tails, and Generalization in Neural Networks create mode 100644 data/2020/neurips/Heavy-tailed Representations, Text Polarity Classification & Data Augmentation create mode 100644 data/2020/neurips/Hedging in games: Faster convergence of external and swap regrets create mode 100644 data/2020/neurips/Heuristic Domain Adaptation create mode 100644 data/2020/neurips/HiFi-GAN: Generative Adversarial Networks for Efficient and High Fidelity Speech Synthesis create mode 100644 data/2020/neurips/HiPPO: Recurrent Memory with Optimal Polynomial Projections create mode 100644 data/2020/neurips/Hierarchical Gaussian Process Priors for Bayesian Neural Network Weights create mode 100644 data/2020/neurips/Hierarchical Granularity Transfer Learning create mode 100644 data/2020/neurips/Hierarchical Neural Architecture Search for Deep Stereo Matching create mode 100644 data/2020/neurips/Hierarchical Patch VAE-GAN: Generating Diverse Videos from a Single Sample create mode 100644 data/2020/neurips/Hierarchical Poset Decoding for Compositional Generalization in Language create mode 100644 data/2020/neurips/Hierarchical Quantized Autoencoders create mode 100644 data/2020/neurips/Hierarchical nucleation in deep neural networks create mode 100644 data/2020/neurips/Hierarchically Organized Latent Modules for Exploratory Search in Morphogenetic Systems create mode 100644 data/2020/neurips/High-Dimensional Bayesian Optimization via Nested Riemannian Manifolds create mode 100644 data/2020/neurips/High-Dimensional Contextual Policy Search with Unknown Context Rewards using Bayesian Optimization create mode 100644 data/2020/neurips/High-Dimensional Sparse Linear Bandits create mode 100644 data/2020/neurips/High-Fidelity Generative Image Compression create mode 100644 data/2020/neurips/High-Throughput Synchronous Deep RL create mode 100644 "data/2020/neurips/High-contrast \"gaudy\" images improve the training of deep neural network models of visual cortex" create mode 100644 data/2020/neurips/High-recall causal discovery for autocorrelated time series with latent confounders create mode 100644 data/2020/neurips/Higher-Order Certification For Randomized Smoothing create mode 100644 data/2020/neurips/Higher-Order Spectral Clustering of Directed Graphs create mode 100644 data/2020/neurips/Hitting the High Notes: Subset Selection for Maximizing Expected Order Statistics create mode 100644 data/2020/neurips/Hold me tight! Influence of discriminative features on deep network boundaries create mode 100644 data/2020/neurips/How Can I Explain This to You? An Empirical Study of Deep Neural Network Explanation Methods create mode 100644 data/2020/neurips/How Robust are the Estimated Effects of Nonpharmaceutical Interventions against COVID-19? create mode 100644 data/2020/neurips/How do fair decisions fare in long-term qualification? create mode 100644 data/2020/neurips/How does This Interaction Affect Me? Interpretable Attribution for Feature Interactions create mode 100644 data/2020/neurips/How does Weight Correlation Affect Generalisation Ability of Deep Neural Networks? create mode 100644 data/2020/neurips/How hard is to distinguish graphs with graph neural networks? create mode 100644 data/2020/neurips/How many samples is a good initial point worth in Low-rank Matrix Recovery? create mode 100644 data/2020/neurips/How to Characterize The Landscape of Overparameterized Convolutional Neural Networks create mode 100644 data/2020/neurips/How to Learn a Useful Critic? Model-based Action-Gradient-Estimator Policy Optimization create mode 100644 data/2020/neurips/Human Parsing Based Texture Transfer from Single Image to 3D Human via Cross-View Consistency create mode 100644 data/2020/neurips/HyNet: Learning Local Descriptor with Hybrid Similarity Measure and Triplet Loss create mode 100644 data/2020/neurips/Hybrid Models for Learning to Branch create mode 100644 data/2020/neurips/Hybrid Variance-Reduced SGD Algorithms For Minimax Problems with Nonconvex-Linear Function create mode 100644 data/2020/neurips/Hypersolvers: Toward Fast Continuous-Depth Models create mode 100644 data/2020/neurips/ICAM: Interpretable Classification via Disentangled Representations and Feature Attribution Mapping create mode 100644 data/2020/neurips/ICE-BeeM: Identifiable Conditional Energy-Based Deep Models Based on Nonlinear ICA create mode 100644 data/2020/neurips/ICNet: Intra-saliency Correlation Network for Co-Saliency Detection create mode 100644 data/2020/neurips/IDEAL: Inexact DEcentralized Accelerated Augmented Lagrangian Method create mode 100644 data/2020/neurips/ISTA-NAS: Efficient and Consistent Neural Architecture Search by Sparse Coding create mode 100644 data/2020/neurips/Identifying Causal-Effect Inference Failure with Uncertainty-Aware Models create mode 100644 data/2020/neurips/Identifying Learning Rules From Neural Network Observables create mode 100644 data/2020/neurips/Identifying Mislabeled Data using the Area Under the Margin Ranking create mode 100644 data/2020/neurips/Identifying signal and noise structure in neural population activity with Gaussian process factor models create mode 100644 data/2020/neurips/ImpatientCapsAndRuns: Approximately Optimal Algorithm Configuration from an Infinite Pool create mode 100644 data/2020/neurips/Implicit Bias in Deep Linear Classification: Initialization Scale vs Training Accuracy create mode 100644 data/2020/neurips/Implicit Distributional Reinforcement Learning create mode 100644 data/2020/neurips/Implicit Graph Neural Networks create mode 100644 data/2020/neurips/Implicit Neural Representations with Periodic Activation Functions create mode 100644 data/2020/neurips/Implicit Rank-Minimizing Autoencoder create mode 100644 data/2020/neurips/Implicit Regularization in Deep Learning May Not Be Explainable by Norms create mode 100644 data/2020/neurips/Impossibility Results for Grammar-Compressed Linear Algebra create mode 100644 data/2020/neurips/Improved Algorithms for Convex-Concave Minimax Optimization create mode 100644 data/2020/neurips/Improved Algorithms for Online Submodular Maximization via First-order Regret Bounds create mode 100644 data/2020/neurips/Improved Analysis of Clipping Algorithms for Non-convex Optimization create mode 100644 data/2020/neurips/Improved Guarantees for k-means++ and k-means++ Parallel create mode 100644 data/2020/neurips/Improved Sample Complexity for Incremental Autonomous Exploration in MDPs create mode 100644 data/2020/neurips/Improved Schemes for Episodic Memory-based Lifelong Learning create mode 100644 data/2020/neurips/Improved Techniques for Training Score-Based Generative Models create mode 100644 data/2020/neurips/Improved Variational Bayesian Phylogenetic Inference with Normalizing Flows create mode 100644 data/2020/neurips/Improved guarantees and a multiple-descent curve for Column Subset Selection and the Nystrom method create mode 100644 data/2020/neurips/Improving Auto-Augment via Augmentation-Wise Weight Sharing create mode 100644 data/2020/neurips/Improving GAN Training with Probability Ratio Clipping and Sample Reweighting create mode 100644 data/2020/neurips/Improving Generalization in Reinforcement Learning with Mixture Regularization create mode 100644 data/2020/neurips/Improving Inference for Neural Image Compression create mode 100644 data/2020/neurips/Improving Local Identifiability in Probabilistic Box Embeddings create mode 100644 data/2020/neurips/Improving Natural Language Processing Tasks with Human Gaze-Guided Neural Attention create mode 100644 data/2020/neurips/Improving Neural Network Training in Low Dimensional Random Bases create mode 100644 data/2020/neurips/Improving Online Rent-or-Buy Algorithms with Sequential Decision Making and ML Predictions create mode 100644 data/2020/neurips/Improving Policy-Constrained Kidney Exchange via Pre-Screening create mode 100644 data/2020/neurips/Improving Sample Complexity Bounds for (Natural) Actor-Critic Algorithms create mode 100644 data/2020/neurips/Improving Sparse Vector Technique with Renyi Differential Privacy create mode 100644 data/2020/neurips/Improving model calibration with accuracy versus uncertainty optimization create mode 100644 data/2020/neurips/Improving robustness against common corruptions by covariate shift adaptation create mode 100644 data/2020/neurips/In search of robust measures of generalization create mode 100644 data/2020/neurips/Incorporating BERT into Parallel Sequence Decoding with Adapters create mode 100644 data/2020/neurips/Incorporating Interpretable Output Constraints in Bayesian Neural Networks create mode 100644 data/2020/neurips/Incorporating Pragmatic Reasoning Communication into Emergent Language create mode 100644 data/2020/neurips/Independent Policy Gradient Methods for Competitive Reinforcement Learning create mode 100644 data/2020/neurips/Inductive Quantum Embedding create mode 100644 data/2020/neurips/Inference Stage Optimization for Cross-scenario 3D Human Pose Estimation create mode 100644 data/2020/neurips/Inference for Batched Bandits create mode 100644 data/2020/neurips/Inferring learning rules from animal decision-making create mode 100644 data/2020/neurips/Influence-Augmented Online Planning for Complex Environments create mode 100644 data/2020/neurips/Information Maximization for Few-Shot Learning create mode 100644 data/2020/neurips/Information Theoretic Counterfactual Learning from Missing-Not-At-Random Feedback create mode 100644 data/2020/neurips/Information Theoretic Regret Bounds for Online Nonlinear Control create mode 100644 data/2020/neurips/Information theoretic limits of learning a sparse rule create mode 100644 data/2020/neurips/Information-theoretic Task Selection for Meta-Reinforcement Learning create mode 100644 data/2020/neurips/Input-Aware Dynamic Backdoor Attack create mode 100644 data/2020/neurips/Instance Based Approximations to Profile Maximum Likelihood create mode 100644 data/2020/neurips/Instance Selection for GANs create mode 100644 data/2020/neurips/Instance-based Generalization in Reinforcement Learning create mode 100644 data/2020/neurips/Instance-optimality in differential privacy via approximate inverse sensitivity mechanisms create mode 100644 data/2020/neurips/Instance-wise Feature Grouping create mode 100644 data/2020/neurips/Instead of Rewriting Foreign Code for Machine Learning, Automatically Synthesize Fast Gradients create mode 100644 data/2020/neurips/Interferobot: aligning an optical interferometer by a reinforcement learning agent create mode 100644 data/2020/neurips/Interior Point Solving for LP-based prediction+optimisation create mode 100644 data/2020/neurips/Interpolation Technique to Speed Up Gradients Propagation in Neural ODEs create mode 100644 data/2020/neurips/Interpretable Sequence Learning for Covid-19 Forecasting create mode 100644 data/2020/neurips/Interpretable and Personalized Apprenticeship Scheduling: Learning Interpretable Scheduling Policies from Heterogeneous User Demonstrations create mode 100644 data/2020/neurips/Interpretable multi-timescale models for predicting fMRI responses to continuous natural speech create mode 100644 data/2020/neurips/Interstellar: Searching Recurrent Architecture for Knowledge Graph Embedding create mode 100644 data/2020/neurips/Interventional Few-Shot Learning create mode 100644 data/2020/neurips/Intra Order-preserving Functions for Calibration of Multi-Class Neural Networks create mode 100644 data/2020/neurips/Intra-Processing Methods for Debiasing Neural Networks create mode 100644 data/2020/neurips/Introducing Routing Uncertainty in Capsule Networks create mode 100644 data/2020/neurips/Inverse Learning of Symmetries create mode 100644 data/2020/neurips/Inverse Rational Control with Partially Observable Continuous Nonlinear Dynamics create mode 100644 data/2020/neurips/Invertible Gaussian Reparameterization: Revisiting the Gumbel-Softmax create mode 100644 data/2020/neurips/Inverting Gradients - How easy is it to break privacy in federated learning? create mode 100644 data/2020/neurips/Investigating Gender Bias in Language Models Using Causal Mediation Analysis create mode 100644 data/2020/neurips/Is Long Horizon RL More Difficult Than Short Horizon RL? create mode 100644 data/2020/neurips/Is Plug-in Solver Sample-Efficient for Feature-based Reinforcement Learning? create mode 100644 data/2020/neurips/Is normalization indispensable for training deep neural network? create mode 100644 data/2020/neurips/Iterative Deep Graph Learning for Graph Neural Networks: Better and Robust Node Embeddings create mode 100644 data/2020/neurips/JAX MD: A Framework for Differentiable Physics create mode 100644 data/2020/neurips/Joint Contrastive Learning with Infinite Possibilities create mode 100644 data/2020/neurips/Joint Policy Search for Multi-agent Collaboration with Imperfect Information create mode 100644 data/2020/neurips/Joints in Random Forests create mode 100644 data/2020/neurips/Just Pick a Sign: Optimizing Deep Multitask Models with Gradient Sign Dropout create mode 100644 data/2020/neurips/KFC: A Scalable Approximation Algorithm for $k$-center Fair Clustering create mode 100644 data/2020/neurips/Kalman Filtering Attention for User Behavior Modeling in CTR Prediction create mode 100644 data/2020/neurips/Kernel Alignment Risk Estimator: Risk Prediction from Training Data create mode 100644 data/2020/neurips/Kernel Based Progressive Distillation for Adder Neural Networks create mode 100644 data/2020/neurips/Kernel Methods Through the Roof: Handling Billions of Points Efficiently create mode 100644 data/2020/neurips/Kernelized information bottleneck leads to biologically plausible 3-factor Hebbian learning in deep networks create mode 100644 data/2020/neurips/Knowledge Augmented Deep Neural Networks for Joint Facial Expression and Action Unit Recognition create mode 100644 data/2020/neurips/Knowledge Distillation in Wide Neural Networks: Risk Bound, Data Efficiency and Imperfect Teacher create mode 100644 data/2020/neurips/Knowledge Transfer in Multi-Task Deep Reinforcement Learning for Continuous Control create mode 100644 data/2020/neurips/LAPAR: Linearly-Assembled Pixel-Adaptive Regression Network for Single Image Super-resolution and Beyond create mode 100644 data/2020/neurips/Label-Aware Neural Tangent Kernel: Toward Better Generalization and Local Elasticity create mode 100644 data/2020/neurips/Labelling unlabelled videos from scratch with multi-modal self-supervision create mode 100644 data/2020/neurips/Lamina-specific neuronal properties promote robust, stable signal propagation in feedforward networks create mode 100644 data/2020/neurips/Language Models are Few-Shot Learners create mode 100644 data/2020/neurips/Language Through a Prism: A Spectral Approach for Multiscale Language Representations create mode 100644 data/2020/neurips/Language and Visual Entity Relationship Graph for Agent Navigation create mode 100644 data/2020/neurips/Language as a Cognitive Tool to Imagine Goals in Curiosity Driven Exploration create mode 100644 data/2020/neurips/Language-Conditioned Imitation Learning for Robot Manipulation Tasks create mode 100644 data/2020/neurips/Large-Scale Adversarial Training for Vision-and-Language Representation Learning create mode 100644 data/2020/neurips/Large-Scale Methods for Distributionally Robust Optimization create mode 100644 data/2020/neurips/Latent Bandits Revisited create mode 100644 data/2020/neurips/Latent Dynamic Factor Analysis of High-Dimensional Neural Recordings create mode 100644 data/2020/neurips/Latent Template Induction with Gumbel-CRFs create mode 100644 data/2020/neurips/Latent World Models For Intrinsically Motivated Exploration create mode 100644 data/2020/neurips/Leap-Of-Thought: Teaching Pre-Trained Models to Systematically Reason Over Implicit Knowledge create mode 100644 data/2020/neurips/Learnability with Indirect Supervision Signals create mode 100644 data/2020/neurips/Learning About Objects by Learning to Interact with Them create mode 100644 data/2020/neurips/Learning Affordance Landscapes for Interaction Exploration in 3D Environments create mode 100644 data/2020/neurips/Learning Agent Representations for Ice Hockey create mode 100644 data/2020/neurips/Learning Augmented Energy Minimization via Speed Scaling create mode 100644 data/2020/neurips/Learning Bounds for Risk-sensitive Learning create mode 100644 data/2020/neurips/Learning Causal Effects via Weighted Empirical Risk Minimization create mode 100644 data/2020/neurips/Learning Certified Individually Fair Representations create mode 100644 data/2020/neurips/Learning Composable Energy Surrogates for PDE Order Reduction create mode 100644 data/2020/neurips/Learning Compositional Rules via Neural Program Synthesis create mode 100644 data/2020/neurips/Learning Continuous System Dynamics from Irregularly-Sampled Partial Observations create mode 100644 data/2020/neurips/Learning Deep Attribution Priors Based On Prior Knowledge create mode 100644 data/2020/neurips/Learning Deformable Tetrahedral Meshes for 3D Reconstruction create mode 100644 data/2020/neurips/Learning Differentiable Programs with Admissible Neural Heuristics create mode 100644 data/2020/neurips/Learning Differential Equations that are Easy to Solve create mode 100644 data/2020/neurips/Learning Discrete Energy-based Models via Auxiliary-variable Local Exploration create mode 100644 data/2020/neurips/Learning Disentangled Representations and Group Structure of Dynamical Environments create mode 100644 data/2020/neurips/Learning Disentangled Representations of Videos with Missing Data create mode 100644 data/2020/neurips/Learning Diverse and Discriminative Representations via the Principle of Maximal Coding Rate Reduction create mode 100644 data/2020/neurips/Learning Dynamic Belief Graphs to Generalize on Text-Based Games create mode 100644 data/2020/neurips/Learning Feature Sparse Principal Subspace create mode 100644 data/2020/neurips/Learning Global Transparent Models consistent with Local Contrastive Explanations create mode 100644 data/2020/neurips/Learning Graph Structure With A Finite-State Automaton Layer create mode 100644 data/2020/neurips/Learning Guidance Rewards with Trajectory-space Smoothing create mode 100644 data/2020/neurips/Learning Implicit Credit Assignment for Cooperative Multi-Agent Reinforcement Learning create mode 100644 data/2020/neurips/Learning Implicit Functions for Topology-Varying Dense 3D Shape Correspondence create mode 100644 data/2020/neurips/Learning Individually Inferred Communication for Multi-Agent Cooperation create mode 100644 data/2020/neurips/Learning Invariances in Neural Networks from Training Data create mode 100644 data/2020/neurips/Learning Invariants through Soft Unification create mode 100644 data/2020/neurips/Learning Kernel Tests Without Data Splitting create mode 100644 data/2020/neurips/Learning Latent Space Energy-Based Prior Model create mode 100644 data/2020/neurips/Learning Linear Programs from Optimal Decisions create mode 100644 data/2020/neurips/Learning Loss for Test-Time Augmentation create mode 100644 data/2020/neurips/Learning Manifold Implicitly via Explicit Heat-Kernel Learning create mode 100644 data/2020/neurips/Learning Multi-Agent Communication through Structured Attentive Reasoning create mode 100644 data/2020/neurips/Learning Multi-Agent Coordination for Enhancing Target Coverage in Directional Sensor Networks create mode 100644 data/2020/neurips/Learning Mutational Semantics create mode 100644 data/2020/neurips/Learning Object-Centric Representations of Multi-Object Scenes from Multiple Views create mode 100644 data/2020/neurips/Learning Optimal Representations with the Decodable Information Bottleneck create mode 100644 data/2020/neurips/Learning Parities with Neural Networks create mode 100644 data/2020/neurips/Learning Physical Constraints with Neural Projections create mode 100644 data/2020/neurips/Learning Physical Graph Representations from Visual Scenes create mode 100644 data/2020/neurips/Learning Representations from Audio-Visual Spatial Alignment create mode 100644 data/2020/neurips/Learning Restricted Boltzmann Machines with Sparse Latent Variables create mode 100644 data/2020/neurips/Learning Retrospective Knowledge with Reverse Reinforcement Learning create mode 100644 data/2020/neurips/Learning Rich Rankings create mode 100644 data/2020/neurips/Learning Robust Decision Policies from Observational Data create mode 100644 data/2020/neurips/Learning Search Space Partition for Black-box Optimization using Monte Carlo Tree Search create mode 100644 data/2020/neurips/Learning Semantic-aware Normalization for Generative Adversarial Networks create mode 100644 data/2020/neurips/Learning Some Popular Gaussian Graphical Models without Condition Number Bounds create mode 100644 data/2020/neurips/Learning Sparse Prototypes for Text Generation create mode 100644 data/2020/neurips/Learning Strategic Network Emergence Games create mode 100644 data/2020/neurips/Learning Strategy-Aware Linear Classifiers create mode 100644 data/2020/neurips/Learning Structured Distributions From Untrusted Batches: Faster and Simpler create mode 100644 data/2020/neurips/Learning Utilities and Equilibria in Non-Truthful Auctions create mode 100644 data/2020/neurips/Learning abstract structure for drawing by efficient motor program induction create mode 100644 data/2020/neurips/Learning by Minimizing the Sum of Ranked Range create mode 100644 data/2020/neurips/Learning compositional functions via multiplicative weight updates create mode 100644 data/2020/neurips/Learning discrete distributions with infinite support create mode 100644 data/2020/neurips/Learning discrete distributions: user vs item-level privacy create mode 100644 data/2020/neurips/Learning efficient task-dependent representations with synaptic plasticity create mode 100644 data/2020/neurips/Learning from Aggregate Observations create mode 100644 data/2020/neurips/Learning from Failure: De-biasing Classifier from Biased Classifier create mode 100644 data/2020/neurips/Learning from Label Proportions: A Mutual Contamination Framework create mode 100644 data/2020/neurips/Learning from Mixtures of Private and Public Populations create mode 100644 data/2020/neurips/Learning from Positive and Unlabeled Data with Arbitrary Positive Shift create mode 100644 data/2020/neurips/Learning identifiable and interpretable latent models of high-dimensional neural activity using pi-VAE create mode 100644 data/2020/neurips/Learning of Discrete Graphical Models with Neural Networks create mode 100644 data/2020/neurips/Learning outside the Black-Box: The pursuit of interpretable models create mode 100644 data/2020/neurips/Learning sparse codes from compressed representations with biologically plausible local wiring constraints create mode 100644 data/2020/neurips/Learning the Geometry of Wave-Based Imaging create mode 100644 data/2020/neurips/Learning the Linear Quadratic Regulator from Nonlinear Observations create mode 100644 data/2020/neurips/Learning to Adapt to Evolving Domains create mode 100644 data/2020/neurips/Learning to Approximate a Bregman Divergence create mode 100644 data/2020/neurips/Learning to Decode: Reinforcement Learning for Decoding of Sparse Graph-Based Channel Codes create mode 100644 data/2020/neurips/Learning to Dispatch for Job Shop Scheduling via Deep Reinforcement Learning create mode 100644 data/2020/neurips/Learning to Execute Programs with Instruction Pointer Attention Graph Neural Networks create mode 100644 data/2020/neurips/Learning to Extrapolate Knowledge: Transductive Few-shot Out-of-Graph Link Prediction create mode 100644 data/2020/neurips/Learning to Incentivize Other Learning Agents create mode 100644 data/2020/neurips/Learning to Learn Variational Semantic Memory create mode 100644 data/2020/neurips/Learning to Learn with Feedback and Local Plasticity create mode 100644 data/2020/neurips/Learning to Mutate with Hypergradient Guided Population create mode 100644 data/2020/neurips/Learning to Orient Surfaces by Self-supervised Spherical CNNs create mode 100644 data/2020/neurips/Learning to Play No-Press Diplomacy with Best Response Policy Iteration create mode 100644 data/2020/neurips/Learning to Play Sequential Games versus Unknown Opponents create mode 100644 data/2020/neurips/Learning to Prove Theorems by Learning to Generate Theorems create mode 100644 data/2020/neurips/Learning to Select Best Forecast Tasks for Clinical Outcome Prediction create mode 100644 data/2020/neurips/Learning to Utilize Shaping Rewards: A New Approach of Reward Shaping create mode 100644 data/2020/neurips/Learning to search efficiently for causally near-optimal treatments create mode 100644 data/2020/neurips/Learning to solve TV regularised problems with unrolled algorithms create mode 100644 data/2020/neurips/Learning to summarize with human feedback create mode 100644 data/2020/neurips/Learning under Model Misspecification: Applications to Variational and Ensemble methods create mode 100644 data/2020/neurips/Learning with Differentiable Pertubed Optimizers create mode 100644 data/2020/neurips/Learning with Operator-valued Kernels in Reproducing Kernel Krein Spaces create mode 100644 data/2020/neurips/Learning with Optimized Random Features: Exponential Speedup by Quantum Machine Learning without Sparsity and Low-Rank Assumptions create mode 100644 data/2020/neurips/Least Squares Regression with Markovian Data: Fundamental Limits and Algorithms create mode 100644 data/2020/neurips/Leverage the Average: an Analysis of KL Regularization in Reinforcement Learning create mode 100644 data/2020/neurips/Leveraging Predictions in Smoothed Online Convex Optimization via Gradient-based Algorithms create mode 100644 data/2020/neurips/Liberty or Depth: Deep Bayesian Neural Nets Do Not Need Complex Weight Posterior Approximations create mode 100644 data/2020/neurips/Lifelong Policy Gradient Learning of Factored Policies for Faster Training Without Forgetting create mode 100644 data/2020/neurips/Lightweight Generative Adversarial Networks for Text-Guided Image Manipulation create mode 100644 data/2020/neurips/Likelihood Regret: An Out-of-Distribution Detection Score For Variational Auto-encoder create mode 100644 data/2020/neurips/Limits on Testing Structural Changes in Ising Models create mode 100644 data/2020/neurips/Limits to Depth Efficiencies of Self-Attention create mode 100644 data/2020/neurips/Linear Disentangled Representations and Unsupervised Action Estimation create mode 100644 data/2020/neurips/Linear Dynamical Systems as a Core Computational Primitive create mode 100644 data/2020/neurips/Linear Time Sinkhorn Divergences using Positive Features create mode 100644 data/2020/neurips/Linear-Sample Learning of Low-Rank Distributions create mode 100644 data/2020/neurips/Linearly Converging Error Compensated SGD create mode 100644 data/2020/neurips/Lipschitz Bounds and Provably Robust Training by Laplacian Smoothing create mode 100644 data/2020/neurips/Lipschitz-Certifiable Training with a Tight Outer Bound create mode 100644 data/2020/neurips/List-Decodable Mean Estimation via Iterative Multi-Filtering create mode 100644 data/2020/neurips/Listening to Sounds of Silence for Speech Denoising create mode 100644 data/2020/neurips/LoCo: Local Contrastive Representation Learning create mode 100644 data/2020/neurips/Locally Differentially Private (Contextual) Bandits Learning create mode 100644 data/2020/neurips/Locally private non-asymptotic testing of discrete distributions is faster using interactive mechanisms create mode 100644 data/2020/neurips/Locally-Adaptive Nonparametric Online Learning create mode 100644 data/2020/neurips/Log-Likelihood Ratio Minimizing Flows: Towards Robust and Quantifiable Neural Distribution Alignment create mode 100644 data/2020/neurips/Logarithmic Pruning is All You Need create mode 100644 data/2020/neurips/Logarithmic Regret Bound in Partially Observable Linear Dynamical Systems create mode 100644 data/2020/neurips/Long-Horizon Visual Planning with Goal-Conditioned Hierarchical Predictors create mode 100644 data/2020/neurips/Long-Tailed Classification by Keeping the Good and Removing the Bad Momentum Causal Effect create mode 100644 data/2020/neurips/LoopReg: Self-supervised Learning of Implicit Surface Correspondences, Pose and Shape for 3D Human Mesh Registration create mode 100644 data/2020/neurips/Low Distortion Block-Resampling with Spatially Stochastic Networks create mode 100644 data/2020/neurips/MATE: Plugging in Model Awareness to Task Embedding for Meta Learning create mode 100644 data/2020/neurips/MCUNet: Tiny Deep Learning on IoT Devices create mode 100644 data/2020/neurips/MDP Homomorphic Networks: Group Symmetries in Reinforcement Learning create mode 100644 data/2020/neurips/MESA: Boost Ensemble Imbalanced Learning with MEta-SAmpler create mode 100644 data/2020/neurips/MMA Regularization: Decorrelating Weights of Neural Networks by Maximizing the Minimal Angles create mode 100644 data/2020/neurips/MOPO: Model-based Offline Policy Optimization create mode 100644 data/2020/neurips/MOReL: Model-Based Offline Reinforcement Learning create mode 100644 data/2020/neurips/MPNet: Masked and Permuted Pre-training for Language Understanding create mode 100644 data/2020/neurips/MRI Banding Removal via Adversarial Training create mode 100644 data/2020/neurips/Make One-Shot Video Object Segmentation Efficient Again create mode 100644 data/2020/neurips/Making Non-Stochastic Control (Almost) as Easy as Stochastic create mode 100644 data/2020/neurips/Manifold GPLVMs for discovering non-Euclidean latent structure in neural data create mode 100644 data/2020/neurips/Manifold structure in graph embeddings create mode 100644 data/2020/neurips/Marginal Utility for Planning in Continuous or Large Discrete Action Spaces create mode 100644 data/2020/neurips/Margins are Insufficient for Explaining Gradient Boosting create mode 100644 data/2020/neurips/Matrix Completion with Hierarchical Graph Side Information create mode 100644 data/2020/neurips/Matrix Inference and Estimation in Multi-Layer Models create mode 100644 "data/2020/neurips/Mat\303\251rn Gaussian Processes on Riemannian Manifolds" create mode 100644 data/2020/neurips/Maximum-Entropy Adversarial Data Augmentation for Improved Generalization and Robustness create mode 100644 data/2020/neurips/Measuring Robustness to Natural Distribution Shifts in Image Classification create mode 100644 data/2020/neurips/Measuring Systematic Generalization in Neural Proof Generation with Transformers create mode 100644 data/2020/neurips/Memory Based Trajectory-conditioned Policies for Learning from Sparse Rewards create mode 100644 data/2020/neurips/Memory-Efficient Learning of Stable Linear Dynamical Systems for Prediction and Control create mode 100644 data/2020/neurips/MeshSDF: Differentiable Iso-Surface Extraction create mode 100644 data/2020/neurips/Meta-Consolidation for Continual Learning create mode 100644 data/2020/neurips/Meta-Gradient Reinforcement Learning with an Objective Discovered Online create mode 100644 data/2020/neurips/Meta-Learning Requires Meta-Augmentation create mode 100644 data/2020/neurips/Meta-Learning Stationary Stochastic Process Prediction with Convolutional Neural Processes create mode 100644 data/2020/neurips/Meta-Learning through Hebbian Plasticity in Random Networks create mode 100644 data/2020/neurips/Meta-Learning with Adaptive Hyperparameters create mode 100644 data/2020/neurips/Meta-Neighborhoods create mode 100644 data/2020/neurips/Meta-learning from Tasks with Heterogeneous Attribute Spaces create mode 100644 data/2020/neurips/Meta-trained agents implement Bayes-optimal agents create mode 100644 data/2020/neurips/MetaPerturb: Transferable Regularizer for Heterogeneous Tasks and Architectures create mode 100644 data/2020/neurips/MetaPoison: Practical General-purpose Clean-label Data Poisoning create mode 100644 data/2020/neurips/MetaSDF: Meta-Learning Signed Distance Functions create mode 100644 data/2020/neurips/Metric-Free Individual Fairness in Online Learning create mode 100644 data/2020/neurips/MinMax Methods for Optimal Transport and Beyond: Regularization, Approximation and Numerics create mode 100644 data/2020/neurips/MiniLM: Deep Self-Attention Distillation for Task-Agnostic Compression of Pre-Trained Transformers create mode 100644 data/2020/neurips/Minibatch Stochastic Approximate Proximal Point Methods create mode 100644 data/2020/neurips/Minibatch vs Local SGD for Heterogeneous Distributed Learning create mode 100644 data/2020/neurips/Minimax Bounds for Generalized Linear Models create mode 100644 data/2020/neurips/Minimax Classification with 0-1 Loss and Performance Guarantees create mode 100644 data/2020/neurips/Minimax Dynamics of Optimally Balanced Spiking Networks of Excitatory and Inhibitory Neurons create mode 100644 data/2020/neurips/Minimax Estimation of Conditional Moment Models create mode 100644 data/2020/neurips/Minimax Lower Bounds for Transfer Learning with Linear and One-hidden Layer Neural Networks create mode 100644 data/2020/neurips/Minimax Optimal Nonparametric Estimation of Heterogeneous Treatment Effects create mode 100644 data/2020/neurips/Minimax Regret of Switching-Constrained Online Convex Optimization: No Phase Transition create mode 100644 data/2020/neurips/Minimax Value Interval for Off-Policy Evaluation and Policy Optimization create mode 100644 data/2020/neurips/Mitigating Forgetting in Online Continual Learning via Instance-Aware Parameterization create mode 100644 data/2020/neurips/Mitigating Manipulation in Peer Review via Randomized Reviewer Assignments create mode 100644 data/2020/neurips/Mix and Match: An Optimistic Tree-Search Approach for Learning Models from Mixture Distributions create mode 100644 data/2020/neurips/Mixed Hamiltonian Monte Carlo for Mixed Discrete and Continuous Variables create mode 100644 data/2020/neurips/Model Agnostic Multilevel Explanations create mode 100644 data/2020/neurips/Model Class Reliance for Random Forests create mode 100644 data/2020/neurips/Model Fusion via Optimal Transport create mode 100644 data/2020/neurips/Model Rubik's Cube: Twisting Resolution, Depth and Width for TinyNets create mode 100644 data/2020/neurips/Model Selection for Production System via Automated Online Experiments create mode 100644 data/2020/neurips/Model Selection in Contextual Stochastic Bandit Problems create mode 100644 data/2020/neurips/Model-Based Multi-Agent RL in Zero-Sum Markov Games with Near-Optimal Sample Complexity create mode 100644 data/2020/neurips/Model-based Adversarial Meta-Reinforcement Learning create mode 100644 data/2020/neurips/Model-based Policy Optimization with Unsupervised Model Adaptation create mode 100644 data/2020/neurips/Model-based Reinforcement Learning for Semi-Markov Decision Processes with Neural ODEs create mode 100644 data/2020/neurips/Modeling Continuous Stochastic Processes with Dynamic Normalizing Flows create mode 100644 data/2020/neurips/Modeling Noisy Annotations for Crowd Counting create mode 100644 data/2020/neurips/Modeling Shared responses in Neuroimaging Studies through MultiView ICA create mode 100644 data/2020/neurips/Modeling Task Effects on Meaning Representation in the Brain via Zero-Shot MEG Prediction create mode 100644 data/2020/neurips/Modeling and Optimization Trade-off in Meta-learning create mode 100644 data/2020/neurips/Modern Hopfield Networks and Attention for Immune Repertoire Classification create mode 100644 data/2020/neurips/Modular Meta-Learning with Shrinkage create mode 100644 data/2020/neurips/MomentumRNN: Integrating Momentum into Recurrent Neural Networks create mode 100644 data/2020/neurips/Monotone operator equilibrium networks create mode 100644 data/2020/neurips/Movement Pruning: Adaptive Sparsity by Fine-Tuning create mode 100644 data/2020/neurips/MuSCLE: Multi Sweep Compression of LiDAR using Deep Entropy Models create mode 100644 data/2020/neurips/Multi-Fidelity Bayesian Optimization via Deep Neural Networks create mode 100644 data/2020/neurips/Multi-Plane Program Induction with 3D Box Priors create mode 100644 data/2020/neurips/Multi-Robot Collision Avoidance under Uncertainty with Probabilistic Safety Barrier Certificates create mode 100644 data/2020/neurips/Multi-Stage Influence Function create mode 100644 data/2020/neurips/Multi-Task Reinforcement Learning with Soft Modularization create mode 100644 data/2020/neurips/Multi-Task Temporal Shift Attention Networks for On-Device Contactless Vitals Measurement create mode 100644 data/2020/neurips/Multi-agent Trajectory Prediction with Fuzzy Query Attention create mode 100644 data/2020/neurips/Multi-agent active perception with prediction rewards create mode 100644 data/2020/neurips/Multi-label Contrastive Predictive Coding create mode 100644 data/2020/neurips/Multi-label classification: do Hamming loss and subset accuracy really conflict with each other? create mode 100644 data/2020/neurips/Multi-task Additive Models for Robust Estimation and Automatic Structure Discovery create mode 100644 data/2020/neurips/Multi-task Batch Reinforcement Learning with Metric Learning create mode 100644 data/2020/neurips/Multi-task Causal Learning with Gaussian Processes create mode 100644 data/2020/neurips/MultiON: Benchmarking Semantic Map Memory using Multi-Object Navigation create mode 100644 data/2020/neurips/Multifaceted Uncertainty Estimation for Label-Efficient Deep Learning create mode 100644 data/2020/neurips/Multilabel Classification by Hierarchical Partitioning and Data-dependent Grouping create mode 100644 data/2020/neurips/Multimodal Generative Learning Utilizing Jensen-Shannon-Divergence create mode 100644 data/2020/neurips/Multimodal Graph Networks for Compositional Generalization in Visual Question Answering create mode 100644 data/2020/neurips/Multiparameter Persistence Image for Topological Machine Learning create mode 100644 data/2020/neurips/Multipole Graph Neural Operator for Parametric Partial Differential Equations create mode 100644 data/2020/neurips/Multiscale Deep Equilibrium Models create mode 100644 data/2020/neurips/Multiview Neural Surface Reconstruction by Disentangling Geometry and Appearance create mode 100644 data/2020/neurips/Munchausen Reinforcement Learning create mode 100644 data/2020/neurips/Mutual exclusivity as a challenge for deep neural networks create mode 100644 data/2020/neurips/Myersonian Regression create mode 100644 data/2020/neurips/NVAE: A Deep Hierarchical Variational Autoencoder create mode 100644 data/2020/neurips/NanoFlow: Scalable Normalizing Flows with Sublinear Parameter Complexity create mode 100644 data/2020/neurips/Natural Graph Networks create mode 100644 data/2020/neurips/Natural Policy Gradient Primal-Dual Method for Constrained Markov Decision Processes create mode 100644 data/2020/neurips/Near-Optimal Comparison Based Clustering create mode 100644 data/2020/neurips/Near-Optimal Reinforcement Learning with Self-Play create mode 100644 data/2020/neurips/Near-Optimal SQ Lower Bounds for Agnostically Learning Halfspaces and ReLUs under Gaussian Marginals create mode 100644 data/2020/neurips/Network Diffusions via Neural Mean-Field Dynamics create mode 100644 data/2020/neurips/Network size and size of the weights in memorization with two-layers neural networks create mode 100644 data/2020/neurips/Network-to-Network Translation with Conditional Invertible Neural Networks create mode 100644 data/2020/neurips/NeuMiss networks: differentiable programming for supervised learning with missing values create mode 100644 data/2020/neurips/Neural Anisotropy Directions create mode 100644 data/2020/neurips/Neural Architecture Generator Optimization create mode 100644 data/2020/neurips/Neural Bridge Sampling for Evaluating Safety-Critical Autonomous Systems create mode 100644 data/2020/neurips/Neural Complexity Measures create mode 100644 data/2020/neurips/Neural Controlled Differential Equations for Irregular Time Series create mode 100644 data/2020/neurips/Neural Dynamic Policies for End-to-End Sensorimotor Learning create mode 100644 data/2020/neurips/Neural Execution Engines: Learning to Execute Subroutines create mode 100644 data/2020/neurips/Neural FFTs for Universal Texture Image Synthesis create mode 100644 data/2020/neurips/Neural Manifold Ordinary Differential Equations create mode 100644 data/2020/neurips/Neural Mesh Flow: 3D Manifold Mesh Generation via Diffeomorphic Flows create mode 100644 data/2020/neurips/Neural Message Passing for Multi-Relational Ordered and Recursive Hypergraphs create mode 100644 data/2020/neurips/Neural Methods for Point-wise Dependency Estimation create mode 100644 data/2020/neurips/Neural Networks Fail to Learn Periodic Functions and How to Fix It create mode 100644 data/2020/neurips/Neural Networks Learning and Memorization with (almost) no Over-Parameterization create mode 100644 data/2020/neurips/Neural Networks with Recurrent Generative Feedback create mode 100644 data/2020/neurips/Neural Networks with Small Weights and Depth-Separation Barriers create mode 100644 data/2020/neurips/Neural Non-Rigid Tracking create mode 100644 data/2020/neurips/Neural Path Features and Neural Path Kernel : Understanding the role of gates in deep learning create mode 100644 data/2020/neurips/Neural Power Units create mode 100644 data/2020/neurips/Neural Sparse Representation for Image Restoration create mode 100644 data/2020/neurips/Neural Sparse Voxel Fields create mode 100644 data/2020/neurips/Neural Star Domain as Primitive Representation create mode 100644 data/2020/neurips/Neural Topographic Factor Analysis for fMRI Data create mode 100644 data/2020/neurips/Neural Unsigned Distance Fields for Implicit Function Learning create mode 100644 data/2020/neurips/Neural encoding with visual attention create mode 100644 data/2020/neurips/Neuron Merging: Compensating for Pruned Neurons create mode 100644 data/2020/neurips/Neuron Shapley: Discovering the Responsible Neurons create mode 100644 data/2020/neurips/Neuron-level Structured Pruning using Polarization Regularizer create mode 100644 data/2020/neurips/Neuronal Gaussian Process Regression create mode 100644 data/2020/neurips/Neurosymbolic Reinforcement Learning with Formally Verified Exploration create mode 100644 data/2020/neurips/Neurosymbolic Transformers for Multi-Agent Communication create mode 100644 data/2020/neurips/Neutralizing Self-Selection Bias in Sampling for Sortition create mode 100644 data/2020/neurips/Nimble: Lightweight and Parallel GPU Task Scheduling for Deep Learning create mode 100644 data/2020/neurips/No Subclass Left Behind: Fine-Grained Robustness in Coarse-Grained Classification Problems create mode 100644 data/2020/neurips/No-Regret Learning Dynamics for Extensive-Form Correlated Equilibrium create mode 100644 data/2020/neurips/No-Regret Learning and Mixed Nash Equilibria: They Do Not Mix create mode 100644 data/2020/neurips/No-regret Learning in Price Competitions under Consumer Reference Effects create mode 100644 data/2020/neurips/Node Classification on Graphs with Few-Shot Novel Labels via Meta Transformed Network Embedding create mode 100644 data/2020/neurips/Node Embeddings and Exact Low-Rank Representations of Complex Networks create mode 100644 data/2020/neurips/Noise-Contrastive Estimation for Multivariate Point Processes create mode 100644 data/2020/neurips/Noise2Same: Optimizing A Self-Supervised Bound for Image Denoising create mode 100644 data/2020/neurips/Non-Convex SGD Learns Halfspaces with Adversarial Label Noise create mode 100644 data/2020/neurips/Non-Crossing Quantile Regression for Distributional Reinforcement Learning create mode 100644 data/2020/neurips/Non-Euclidean Universal Approximation create mode 100644 data/2020/neurips/Non-Stochastic Control with Bandit Feedback create mode 100644 data/2020/neurips/Non-parametric Models for Non-negative Functions create mode 100644 data/2020/neurips/Non-reversible Gaussian processes for identifying latent dynamical structure in neural data create mode 100644 data/2020/neurips/Nonasymptotic Guarantees for Spiked Matrix Recovery with Generative Priors create mode 100644 data/2020/neurips/Nonconvex Sparse Graph Learning under Laplacian Constrained Graphical Model create mode 100644 data/2020/neurips/Normalizing Kalman Filters for Multivariate Time Series Analysis create mode 100644 data/2020/neurips/Not All Unlabeled Data are Equal: Learning to Weight Data in Semi-supervised Learning create mode 100644 data/2020/neurips/Novelty Search in Representational Space for Sample Efficient Exploration create mode 100644 data/2020/neurips/Numerically Solving Parametric Families of High-Dimensional Kolmogorov Partial Differential Equations via Deep Learning create mode 100644 data/2020/neurips/O(n) Connections are Expressive Enough: Universal Approximability of Sparse Transformers create mode 100644 data/2020/neurips/OOD-MAML: Meta-Learning for Few-Shot Out-of-Distribution Detection and Classification create mode 100644 data/2020/neurips/OTLDA: A Geometry-aware Optimal Transport Approach for Topic Modeling create mode 100644 data/2020/neurips/Object Goal Navigation using Goal-Oriented Semantic Exploration create mode 100644 data/2020/neurips/Object-Centric Learning with Slot Attention create mode 100644 data/2020/neurips/Ode to an ODE create mode 100644 data/2020/neurips/Off-Policy Evaluation and Learning for External Validity under a Covariate Shift create mode 100644 data/2020/neurips/Off-Policy Evaluation via the Regularized Lagrangian create mode 100644 data/2020/neurips/Off-Policy Imitation Learning from Observations create mode 100644 data/2020/neurips/Off-Policy Interval Estimation with Lipschitz Value Iteration create mode 100644 data/2020/neurips/Off-policy Policy Evaluation For Sequential Decisions Under Unobserved Confounding create mode 100644 data/2020/neurips/Offline Imitation Learning with a Misspecified Simulator create mode 100644 data/2020/neurips/On 1 n neural representation and robustness create mode 100644 data/2020/neurips/On Adaptive Attacks to Adversarial Example Defenses create mode 100644 data/2020/neurips/On Adaptive Distance Estimation create mode 100644 data/2020/neurips/On Completeness-aware Concept-Based Explanations in Deep Neural Networks create mode 100644 data/2020/neurips/On Convergence and Generalization of Dropout Training create mode 100644 data/2020/neurips/On Convergence of Nearest Neighbor Classifiers over Feature Transformations create mode 100644 data/2020/neurips/On Correctness of Automatic Differentiation for Non-Differentiable Functions create mode 100644 data/2020/neurips/On Efficiency in Hierarchical Reinforcement Learning create mode 100644 data/2020/neurips/On Infinite-Width Hypernetworks create mode 100644 data/2020/neurips/On Learning Ising Models under Huber's Contamination Model create mode 100644 data/2020/neurips/On Numerosity of Deep Neural Networks create mode 100644 data/2020/neurips/On Power Laws in Deep Ensembles create mode 100644 data/2020/neurips/On Regret with Multiple Best Arms create mode 100644 data/2020/neurips/On Reward-Free Reinforcement Learning with Linear Function Approximation create mode 100644 data/2020/neurips/On Second Order Behaviour in Augmented Neural ODEs create mode 100644 data/2020/neurips/On Testing of Samplers create mode 100644 data/2020/neurips/On Uniform Convergence and Low-Norm Interpolation Learning create mode 100644 data/2020/neurips/On Warm-Starting Neural Network Training create mode 100644 data/2020/neurips/On ranking via sorting by estimated expected utility create mode 100644 data/2020/neurips/On the Almost Sure Convergence of Stochastic Gradient Descent in Non-Convex Problems create mode 100644 data/2020/neurips/On the Convergence of Smooth Regularized Approximate Value Iteration Schemes create mode 100644 data/2020/neurips/On the Equivalence between Online and Private Learnability beyond Binary Classification create mode 100644 data/2020/neurips/On the Ergodicity, Bias and Asymptotic Normality of Randomized Midpoint Sampling Method create mode 100644 data/2020/neurips/On the Error Resistance of Hinge-Loss Minimization create mode 100644 data/2020/neurips/On the Expressiveness of Approximate Inference in Bayesian Neural Networks create mode 100644 data/2020/neurips/On the Loss Landscape of Adversarial Training: Identifying Challenges and How to Overcome Them create mode 100644 data/2020/neurips/On the Modularity of Hypernetworks create mode 100644 data/2020/neurips/On the Power of Louvain in the Stochastic Block Model create mode 100644 data/2020/neurips/On the Role of Sparsity and DAG Constraints for Learning Linear DAGs create mode 100644 data/2020/neurips/On the Similarity between the Laplace and Neural Tangent Kernels create mode 100644 data/2020/neurips/On the Stability and Convergence of Robust Adversarial Reinforcement Learning: A Case Study on Linear Quadratic Systems create mode 100644 data/2020/neurips/On the Theory of Transfer Learning: The Importance of Task Diversity create mode 100644 data/2020/neurips/On the Tightness of Semidefinite Relaxations for Certifying Robustness to Adversarial Examples create mode 100644 data/2020/neurips/On the Trade-off between Adversarial and Backdoor Robustness create mode 100644 data/2020/neurips/On the Value of Out-of-Distribution Testing: An Example of Goodhart's Law create mode 100644 data/2020/neurips/On the distance between two neural networks and the stability of learning create mode 100644 data/2020/neurips/On the equivalence of molecular graph convolution and molecular wave function with poor basis set create mode 100644 data/2020/neurips/On the linearity of large non-linear models: when and why the tangent kernel is constant create mode 100644 data/2020/neurips/Once-for-All Adversarial Training: In-Situ Tradeoff between Robustness and Accuracy for Free create mode 100644 data/2020/neurips/One Ring to Rule Them All: Certifiably Robust Geometric Perception with Outliers create mode 100644 data/2020/neurips/One Solution is Not All You Need: Few-Shot Extrapolation via Structured MaxEnt RL create mode 100644 data/2020/neurips/One-bit Supervision for Image Classification create mode 100644 data/2020/neurips/One-sample Guided Object Representation Disassembling create mode 100644 data/2020/neurips/Online Agnostic Boosting via Regret Minimization create mode 100644 data/2020/neurips/Online Algorithm for Unsupervised Sequential Selection with Contextual Information create mode 100644 data/2020/neurips/Online Algorithms for Multi-shop Ski Rental with Machine Learned Advice create mode 100644 data/2020/neurips/Online Bayesian Goal Inference for Boundedly Rational Planning Agents create mode 100644 data/2020/neurips/Online Bayesian Persuasion create mode 100644 data/2020/neurips/Online Convex Optimization Over Erdos-Renyi Random Networks create mode 100644 data/2020/neurips/Online Decision Based Visual Tracking via Reinforcement Learning create mode 100644 data/2020/neurips/Online Fast Adaptation and Knowledge Accumulation (OSAKA): a New Approach to Continual Learning create mode 100644 data/2020/neurips/Online Influence Maximization under Linear Threshold Model create mode 100644 data/2020/neurips/Online Learning in Contextual Bandits using Gated Linear Networks create mode 100644 data/2020/neurips/Online Learning with Primary and Secondary Losses create mode 100644 data/2020/neurips/Online Linear Optimization with Many Hints create mode 100644 data/2020/neurips/Online MAP Inference of Determinantal Point Processes create mode 100644 data/2020/neurips/Online Matrix Completion with Side Information create mode 100644 data/2020/neurips/Online Meta-Critic Learning for Off-Policy Actor-Critic Methods create mode 100644 data/2020/neurips/Online Multitask Learning with Long-Term Memory create mode 100644 data/2020/neurips/Online Neural Connectivity Estimation with Noisy Group Testing create mode 100644 data/2020/neurips/Online Non-Convex Optimization with Imperfect Feedback create mode 100644 data/2020/neurips/Online Optimization with Memory and Competitive Control create mode 100644 data/2020/neurips/Online Planning with Lookahead Policies create mode 100644 data/2020/neurips/Online Robust Regression via SGD on the l1 loss create mode 100644 data/2020/neurips/Online Sinkhorn: Optimal Transport distances from sample streams create mode 100644 data/2020/neurips/Online Structured Meta-learning create mode 100644 data/2020/neurips/Online learning with dynamics: A minimax perspective create mode 100644 data/2020/neurips/Open Graph Benchmark: Datasets for Machine Learning on Graphs create mode 100644 data/2020/neurips/Optimal Adaptive Electrode Selection to Maximize Simultaneously Recorded Neuron Yield create mode 100644 data/2020/neurips/Optimal Algorithms for Stochastic Multi-Armed Bandits with Heavy Tailed Rewards create mode 100644 data/2020/neurips/Optimal Approximation - Smoothness Tradeoffs for Soft-Max Functions create mode 100644 data/2020/neurips/Optimal Best-arm Identification in Linear Bandits create mode 100644 data/2020/neurips/Optimal Epoch Stochastic Gradient Descent Ascent Methods for Min-Max Optimization create mode 100644 data/2020/neurips/Optimal Iterative Sketching Methods with the Subsampled Randomized Hadamard Transform create mode 100644 data/2020/neurips/Optimal Learning from Verified Training Data create mode 100644 data/2020/neurips/Optimal Lottery Tickets via Subset Sum: Logarithmic Over-Parameterization is Sufficient create mode 100644 data/2020/neurips/Optimal Prediction of the Number of Unseen Species with Multiplicity create mode 100644 data/2020/neurips/Optimal Private Median Estimation under Minimal Distributional Assumptions create mode 100644 data/2020/neurips/Optimal Query Complexity of Secure Stochastic Convex Optimization create mode 100644 data/2020/neurips/Optimal Robustness-Consistency Trade-offs for Learning-Augmented Online Algorithms create mode 100644 data/2020/neurips/Optimal Variance Control of the Score-Function Gradient Estimator for Importance-Weighted Bounds create mode 100644 data/2020/neurips/Optimal and Practical Algorithms for Smooth and Strongly Convex Decentralized Optimization create mode 100644 data/2020/neurips/Optimally Deceiving a Learning Leader in Stackelberg Games create mode 100644 data/2020/neurips/Optimistic Dual Extrapolation for Coherent Non-monotone Variational Inequalities create mode 100644 data/2020/neurips/Optimization and Generalization Analysis of Transduction through Gradient Boosting and Application to Multi-scale Graph Neural Networks create mode 100644 data/2020/neurips/Optimization and Generalization of Shallow Neural Networks with Quadratic Activation Functions create mode 100644 data/2020/neurips/Optimizing Mode Connectivity via Neuron Alignment create mode 100644 data/2020/neurips/Optimizing Neural Networks via Koopman Operator Theory create mode 100644 data/2020/neurips/OrganITE: Optimal transplant donor organ offering using an individual treatment effect create mode 100644 data/2020/neurips/Organizing recurrent network dynamics by task-computation to enable continual learning create mode 100644 data/2020/neurips/Outlier Robust Mean Estimation with Subgaussian Rates via Stability create mode 100644 data/2020/neurips/Over-parameterized Adversarial Training: An Analysis Overcoming the Curse of Dimensionality create mode 100644 data/2020/neurips/Overfitting Can Be Harmless for Basis Pursuit, But Only to a Degree create mode 100644 data/2020/neurips/PAC-Bayes Analysis Beyond the Usual Bounds create mode 100644 data/2020/neurips/PAC-Bayes Learning Bounds for Sample-Dependent Priors create mode 100644 data/2020/neurips/PAC-Bayesian Bound for the Conditional Value at Risk create mode 100644 data/2020/neurips/PC-PG: Policy Cover Directed Exploration for Provable Policy Gradient Learning create mode 100644 data/2020/neurips/PEP: Parameter Ensembling by Perturbation create mode 100644 data/2020/neurips/PGM-Explainer: Probabilistic Graphical Model Explanations for Graph Neural Networks create mode 100644 data/2020/neurips/PIE-NET: Parametric Inference of Point Cloud Edges create mode 100644 data/2020/neurips/PLANS: Neuro-Symbolic Program Learning from Videos create mode 100644 data/2020/neurips/PLLay: Efficient Topological Layer based on Persistent Landscapes create mode 100644 data/2020/neurips/POLY-HOOT: Monte-Carlo Planning in Continuous Space MDPs with Non-Asymptotic Analysis create mode 100644 data/2020/neurips/POMDPs in Continuous Time and Discrete Spaces create mode 100644 data/2020/neurips/POMO: Policy Optimization with Multiple Optima for Reinforcement Learning create mode 100644 data/2020/neurips/PRANK: motion Prediction based on RANKing create mode 100644 data/2020/neurips/Parabolic Approximation Line Search for DNNs create mode 100644 data/2020/neurips/Parameterized Explainer for Graph Neural Network create mode 100644 data/2020/neurips/Parametric Instance Classification for Unsupervised Visual Feature learning create mode 100644 data/2020/neurips/Part-dependent Label Noise: Towards Instance-dependent Label Noise create mode 100644 data/2020/neurips/Partially View-aligned Clustering create mode 100644 "data/2020/neurips/Patch2Self: Denoising Diffusion MRI with Self-Supervised Learning\342\200\213" create mode 100644 data/2020/neurips/Path Integral Based Convolution and Pooling for Graph Neural Networks create mode 100644 data/2020/neurips/Path Sample-Analytic Gradient Estimators for Stochastic Binary Networks create mode 100644 data/2020/neurips/Penalized Langevin dynamics with vanishing penalty for smooth and log-concave targets create mode 100644 data/2020/neurips/Permute-and-Flip: A new mechanism for differentially private selection create mode 100644 data/2020/neurips/Personalized Federated Learning with Moreau Envelopes create mode 100644 data/2020/neurips/Personalized Federated Learning with Theoretical Guarantees: A Model-Agnostic Meta-Learning Approach create mode 100644 data/2020/neurips/Perturbing Across the Feature Hierarchy to Improve Standard and Strict Blackbox Attack Transferability create mode 100644 data/2020/neurips/Phase retrieval in high dimensions: Statistical and computational phase transitions create mode 100644 data/2020/neurips/Pipeline PSRO: A Scalable Approach for Finding Approximate Nash Equilibria in Large Games create mode 100644 data/2020/neurips/Pixel-Level Cycle Association: A New Perspective for Domain Adaptive Semantic Segmentation create mode 100644 data/2020/neurips/PlanGAN: Model-based Planning With Sparse Rewards and Multiple Goals create mode 100644 data/2020/neurips/Planning in Markov Decision Processes with Gap-Dependent Sample Complexity create mode 100644 data/2020/neurips/Planning with General Objective Functions: Going Beyond Total Rewards create mode 100644 data/2020/neurips/Point process models for sequence detection in high-dimensional neural spike trains create mode 100644 data/2020/neurips/Pointer Graph Networks create mode 100644 data/2020/neurips/Policy Improvement via Imitation of Multiple Oracles create mode 100644 data/2020/neurips/Polynomial-Time Computation of Optimal Correlated Equilibria in Two-Player Extensive-Form Games with Public Chance Moves and Beyond create mode 100644 data/2020/neurips/Pontryagin Differentiable Programming: An End-to-End Learning and Control Framework create mode 100644 data/2020/neurips/Position-based Scaled Gradient for Model Quantization and Pruning create mode 100644 data/2020/neurips/Post-training Iterative Hierarchical Data Augmentation for Deep Networks create mode 100644 data/2020/neurips/Posterior Network: Uncertainty Estimation without OOD Samples via Density-Based Pseudo-Counts create mode 100644 data/2020/neurips/Posterior Re-calibration for Imbalanced Datasets create mode 100644 data/2020/neurips/Practical Low-Rank Communication Compression in Decentralized Deep Learning create mode 100644 data/2020/neurips/Practical No-box Adversarial Attacks against DNNs create mode 100644 data/2020/neurips/Practical Quasi-Newton Methods for Training Deep Neural Networks create mode 100644 data/2020/neurips/Pre-training via Paraphrasing create mode 100644 data/2020/neurips/Precise expressions for random projections: Low-rank approximation and randomized Newton create mode 100644 data/2020/neurips/Predicting Training Time Without Training create mode 100644 data/2020/neurips/Prediction with Corrupted Expert Advice create mode 100644 data/2020/neurips/Predictive Information Accelerates Learning in RL create mode 100644 data/2020/neurips/Predictive coding in balanced neural networks with noise, chaos and delays create mode 100644 data/2020/neurips/Predictive inference is free with the jackknife+-after-bootstrap create mode 100644 data/2020/neurips/Preference learning along multiple criteria: A game-theoretic perspective create mode 100644 data/2020/neurips/Preference-based Reinforcement Learning with Finite-Time Guarantees create mode 100644 data/2020/neurips/Primal Dual Interpretation of the Proximal Stochastic Gradient Langevin Algorithm create mode 100644 data/2020/neurips/Primal-Dual Mesh Convolutional Neural Networks create mode 100644 data/2020/neurips/Principal Neighbourhood Aggregation for Graph Nets create mode 100644 data/2020/neurips/Privacy Amplification via Random Check-Ins create mode 100644 data/2020/neurips/Private Identity Testing for High-Dimensional Distributions create mode 100644 data/2020/neurips/Private Learning of Halfspaces: Simplifying the Construction and Reducing the Sample Complexity create mode 100644 data/2020/neurips/Probabilistic Active Meta-Learning create mode 100644 data/2020/neurips/Probabilistic Circuits for Variational Inference in Discrete Graphical Models create mode 100644 data/2020/neurips/Probabilistic Fair Clustering create mode 100644 data/2020/neurips/Probabilistic Inference with Algebraic Constraints: Theoretical Limits and Practical Approximations create mode 100644 data/2020/neurips/Probabilistic Linear Solvers for Machine Learning create mode 100644 data/2020/neurips/Probabilistic Orientation Estimation with Matrix Fisher Distributions create mode 100644 data/2020/neurips/Probabilistic Time Series Forecasting with Shape and Temporal Diversity create mode 100644 data/2020/neurips/Probably Approximately Correct Constrained Learning create mode 100644 data/2020/neurips/Profile Entropy: A Fundamental Measure for the Learnability and Compressibility of Distributions create mode 100644 data/2020/neurips/Program Synthesis with Pragmatic Communication create mode 100644 data/2020/neurips/Projected Stein Variational Gradient Descent create mode 100644 data/2020/neurips/Projection Efficient Subgradient Method and Optimal Nonsmooth Frank-Wolfe Method create mode 100644 data/2020/neurips/Projection Robust Wasserstein Distance and Riemannian Optimization create mode 100644 data/2020/neurips/Promoting Coordination through Policy Regularization in Multi-Agent Deep Reinforcement Learning create mode 100644 data/2020/neurips/Promoting Stochasticity for Expressive Policies via a Simple and Efficient Regularization Method create mode 100644 data/2020/neurips/Prophet Attention: Predicting Attention with Future Attention create mode 100644 data/2020/neurips/Provable Online CP PARAFAC Decomposition of a Structured Tensor via Dictionary Learning create mode 100644 data/2020/neurips/Provable Overlapping Community Detection in Weighted Graphs create mode 100644 data/2020/neurips/Provably Consistent Partial-Label Learning create mode 100644 data/2020/neurips/Provably Efficient Exploration for Reinforcement Learning Using Unsupervised Learning create mode 100644 data/2020/neurips/Provably Efficient Neural Estimation of Structural Equation Models: An Adversarial Approach create mode 100644 data/2020/neurips/Provably Efficient Neural GTD for Off-Policy Learning create mode 100644 data/2020/neurips/Provably Efficient Online Hyperparameter Optimization with Population-Based Bandits create mode 100644 data/2020/neurips/Provably Efficient Reinforcement Learning with Kernel and Neural Function Approximations create mode 100644 data/2020/neurips/Provably Efficient Reward-Agnostic Navigation with Linear Value Iteration create mode 100644 data/2020/neurips/Provably Good Batch Off-Policy Reinforcement Learning Without Great Exploration create mode 100644 data/2020/neurips/Provably Robust Metric Learning create mode 100644 data/2020/neurips/Provably adaptive reinforcement learning in metric spaces create mode 100644 data/2020/neurips/Proximal Mapping for Deep Regularization create mode 100644 data/2020/neurips/Proximity Operator of the Matrix Perspective Function and its Applications create mode 100644 data/2020/neurips/Pruning Filter in Filter create mode 100644 data/2020/neurips/Pruning neural networks without any data by iteratively conserving synaptic flow create mode 100644 data/2020/neurips/Pushing the Limits of Narrow Precision Inferencing at Cloud Scale with Microsoft Floating Point create mode 100644 data/2020/neurips/PyGlove: Symbolic Programming for Automated Machine Learning create mode 100644 data/2020/neurips/Quantifying Learnability and Describability of Visual Concepts Emerging in Representation Learning create mode 100644 data/2020/neurips/Quantifying the Empirical Wasserstein Distance to a Set of Measures: Beating the Curse of Dimensionality create mode 100644 data/2020/neurips/Quantile Propagation for Wasserstein-Approximate Gaussian Processes create mode 100644 data/2020/neurips/Quantitative Propagation of Chaos for SGD in Wide Neural Networks create mode 100644 data/2020/neurips/Quantized Variational Inference create mode 100644 data/2020/neurips/R-learning in actor-critic model offers a biologically relevant mechanism for sequential decision-making create mode 100644 data/2020/neurips/RANet: Region Attention Network for Semantic Segmentation create mode 100644 data/2020/neurips/RATT: Recurrent Attention to Transient Tasks for Continual Image Captioning create mode 100644 data/2020/neurips/RELATE: Physically Plausible Multi-Object Scene Synthesis Using Structured Latent Spaces create mode 100644 data/2020/neurips/RL Unplugged: A Collection of Benchmarks for Offline Reinforcement Learning create mode 100644 data/2020/neurips/RNNPool: Efficient Non-linear Pooling for RAM Constrained Inference create mode 100644 data/2020/neurips/RSKDD-Net: Random Sample-based Keypoint Detector and Descriptor create mode 100644 data/2020/neurips/RandAugment: Practical Automated Data Augmentation with a Reduced Search Space create mode 100644 data/2020/neurips/Random Reshuffling is Not Always Better create mode 100644 data/2020/neurips/Random Reshuffling: Simple Analysis with Vast Improvements create mode 100644 data/2020/neurips/Random Walk Graph Neural Networks create mode 100644 data/2020/neurips/Randomized tests for high-dimensional regression: A more efficient and powerful solution create mode 100644 data/2020/neurips/Rankmax: An Adaptive Projection Alternative to the Softmax Function create mode 100644 data/2020/neurips/Ratio Trace Formulation of Wasserstein Discriminant Analysis create mode 100644 data/2020/neurips/Rational neural networks create mode 100644 data/2020/neurips/Re-Examining Linear Embeddings for High-Dimensional Bayesian Optimization create mode 100644 data/2020/neurips/Real World Games Look Like Spinning Tops create mode 100644 data/2020/neurips/Reasoning about Uncertainties in Discrete-Time Dynamical Systems using Polynomial Forms create mode 100644 data/2020/neurips/Reciprocal Adversarial Learning via Characteristic Functions create mode 100644 data/2020/neurips/Reconciling Modern Deep Learning with Traditional Optimization Analyses: The Intrinsic Learning Rate create mode 100644 data/2020/neurips/Reconsidering Generative Objectives For Counterfactual Reasoning create mode 100644 data/2020/neurips/Reconstructing Perceptive Images from Brain Activity by Shape-Semantic GAN create mode 100644 data/2020/neurips/Recovery of sparse linear classifiers from mixture of responses create mode 100644 data/2020/neurips/Recurrent Quantum Neural Networks create mode 100644 data/2020/neurips/Recurrent Switching Dynamical Systems Models for Multiple Interacting Neural Populations create mode 100644 data/2020/neurips/Recursive Inference for Variational Autoencoders create mode 100644 data/2020/neurips/Reducing Adversarially Robust Learning to Non-Robust PAC Learning create mode 100644 data/2020/neurips/Refactoring Policy for Compositional Generalizability using Self-Supervised Object Proposals create mode 100644 data/2020/neurips/Regression with reject option and application to kNN create mode 100644 data/2020/neurips/Regret Bounds without Lipschitz Continuity: Online Learning with Relative-Lipschitz Losses create mode 100644 data/2020/neurips/Regret in Online Recommendation Systems create mode 100644 data/2020/neurips/Regularized linear autoencoders recover the principal components, eventually create mode 100644 data/2020/neurips/Regularizing Black-box Models for Improved Interpretability create mode 100644 data/2020/neurips/Regularizing Towards Permutation Invariance In Recurrent Models create mode 100644 data/2020/neurips/Reinforced Molecular Optimization with Neighborhood-Controlled Grammars create mode 100644 data/2020/neurips/Reinforcement Learning for Control with Multiple Frequencies create mode 100644 data/2020/neurips/Reinforcement Learning in Factored MDPs: Oracle-Efficient Algorithms and Tighter Regret Bounds for the Non-Episodic Setting create mode 100644 data/2020/neurips/Reinforcement Learning with Augmented Data create mode 100644 data/2020/neurips/Reinforcement Learning with Combinatorial Actions: An Application to Vehicle Routing create mode 100644 data/2020/neurips/Reinforcement Learning with Feedback Graphs create mode 100644 data/2020/neurips/Reinforcement Learning with General Value Function Approximation: Provably Efficient Approach via Bounded Eluder Dimension create mode 100644 data/2020/neurips/Rel3D: A Minimally Contrastive Benchmark for Grounding Spatial Relations in 3D create mode 100644 data/2020/neurips/RelationNet++: Bridging Visual Representations for Object Detection via Transformer Decoder create mode 100644 data/2020/neurips/Relative gradient optimization of the Jacobian term in unsupervised deep learning create mode 100644 data/2020/neurips/Reliable Graph Neural Networks via Robust Aggregation create mode 100644 data/2020/neurips/Removing Bias in Multi-modal Classifiers: Regularization by Maximizing Functional Entropies create mode 100644 data/2020/neurips/RepPoints v2: Verification Meets Regression for Object Detection create mode 100644 data/2020/neurips/Reparameterizing Mirror Descent as Gradient Descent create mode 100644 "data/2020/neurips/Replica-Exchange Nos\303\251-Hoover Dynamics for Bayesian Learning on Large Datasets" create mode 100644 data/2020/neurips/Representation Learning for Integrating Multi-domain Outcomes to Optimize Individualized Treatment create mode 100644 data/2020/neurips/Rescuing neural spike train models from bad MLE create mode 100644 data/2020/neurips/Reservoir Computing meets Recurrent Kernels and Structured Transforms create mode 100644 data/2020/neurips/Residual Distillation: Towards Portable Deep Neural Networks without Shortcuts create mode 100644 data/2020/neurips/Residual Force Control for Agile Human Behavior Imitation and Extended Motion Synthesis create mode 100644 data/2020/neurips/Restless-UCB, an Efficient and Low-complexity Algorithm for Online Restless Bandits create mode 100644 data/2020/neurips/Restoring Negative Information in Few-Shot Object Detection create mode 100644 data/2020/neurips/Rethinking Importance Weighting for Deep Learning under Distribution Shift create mode 100644 data/2020/neurips/Rethinking Learnable Tree Filter for Generic Feature Transform create mode 100644 data/2020/neurips/Rethinking Pre-training and Self-training create mode 100644 data/2020/neurips/Rethinking pooling in graph neural networks create mode 100644 data/2020/neurips/Rethinking the Value of Labels for Improving Class-Imbalanced Learning create mode 100644 data/2020/neurips/Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks create mode 100644 data/2020/neurips/RetroXpert: Decompose Retrosynthesis Prediction Like A Chemist create mode 100644 data/2020/neurips/Reverse-engineering recurrent neural network solutions to a hierarchical inference task for mice create mode 100644 data/2020/neurips/Revisiting Frank-Wolfe for Polytopes: Strict Complementarity and Sparsity create mode 100644 data/2020/neurips/Revisiting Parameter Sharing for Automatic Neural Channel Number Search create mode 100644 data/2020/neurips/Revisiting the Sample Complexity of Sparse Spectrum Approximation of Gaussian Processes create mode 100644 data/2020/neurips/Reward Propagation Using Graph Convolutional Networks create mode 100644 data/2020/neurips/Reward-rational (implicit) choice: A unifying formalism for reward learning create mode 100644 data/2020/neurips/Rewriting History with Inverse RL: Hindsight Inference for Policy Improvement create mode 100644 data/2020/neurips/Ridge Rider: Finding Diverse Solutions by Following Eigenvectors of the Hessian create mode 100644 data/2020/neurips/Riemannian Continuous Normalizing Flows create mode 100644 data/2020/neurips/Risk-Sensitive Reinforcement Learning: Near-Optimal Risk-Sample Tradeoff in Regret create mode 100644 data/2020/neurips/Robust Correction of Sampling Bias using Cumulative Distribution Functions create mode 100644 data/2020/neurips/Robust Deep Reinforcement Learning against Adversarial Perturbations on State Observations create mode 100644 data/2020/neurips/Robust Density Estimation under Besov IPM Losses create mode 100644 data/2020/neurips/Robust Disentanglement of a Few Factors at a Time create mode 100644 data/2020/neurips/Robust Federated Learning: The Case of Affine Distribution Shifts create mode 100644 data/2020/neurips/Robust Gaussian Covariance Estimation in Nearly-Matrix Multiplication Time create mode 100644 data/2020/neurips/Robust Meta-learning for Mixed Linear Regression with Small Batches create mode 100644 data/2020/neurips/Robust Multi-Agent Reinforcement Learning with Model Uncertainty create mode 100644 data/2020/neurips/Robust Multi-Object Matching via Iterative Reweighting of the Graph Connection Laplacian create mode 100644 data/2020/neurips/Robust Optimal Transport with Applications in Generative Modeling and Domain Adaptation create mode 100644 data/2020/neurips/Robust Optimization for Fairness with Noisy Protected Groups create mode 100644 data/2020/neurips/Robust Persistence Diagrams using Reproducing Kernels create mode 100644 data/2020/neurips/Robust Pre-Training by Adversarial Contrastive Learning create mode 100644 data/2020/neurips/Robust Quantization: One Model to Rule Them All create mode 100644 data/2020/neurips/Robust Recovery via Implicit Bias of Discrepant Learning Rates for Double Over-parameterization create mode 100644 data/2020/neurips/Robust Recursive Partitioning for Heterogeneous Treatment Effects with Uncertainty Quantification create mode 100644 data/2020/neurips/Robust Reinforcement Learning via Adversarial training with Langevin Dynamics create mode 100644 data/2020/neurips/Robust Sequence Submodular Maximization create mode 100644 data/2020/neurips/Robust Sub-Gaussian Principal Component Analysis and Width-Independent Schatten Packing create mode 100644 data/2020/neurips/Robust and Heavy-Tailed Mean Estimation Made Simple, via Regret Minimization create mode 100644 data/2020/neurips/Robust compressed sensing using generative models create mode 100644 data/2020/neurips/Robust large-margin learning in hyperbolic space create mode 100644 data/2020/neurips/Robust, Accurate Stochastic Optimization for Variational Inference create mode 100644 data/2020/neurips/Robust-Adaptive Control of Linear Systems: beyond Quadratic Costs create mode 100644 data/2020/neurips/Robustness Analysis of Non-Convex Stochastic Gradient Descent using Biased Expectations create mode 100644 data/2020/neurips/Robustness of Bayesian Neural Networks to Gradient-Based Attacks create mode 100644 data/2020/neurips/Robustness of Community Detection to Random Geometric Perturbations create mode 100644 data/2020/neurips/Rotated Binary Neural Network create mode 100644 data/2020/neurips/Rotation-Invariant Local-to-Global Representation Learning for 3D Point Cloud create mode 100644 data/2020/neurips/SAC: Accelerating and Structuring Self-Attention via Sparse Adaptive Connection create mode 100644 data/2020/neurips/SCOP: Scientific Control for Reliable Neural Network Pruning create mode 100644 data/2020/neurips/SDF-SRN: Learning Signed Distance 3D Object Reconstruction from Static Images create mode 100644 data/2020/neurips/SE(3)-Transformers: 3D Roto-Translation Equivariant Attention Networks create mode 100644 data/2020/neurips/SEVIR : A Storm Event Imagery Dataset for Deep Learning Applications in Radar and Satellite Meteorology create mode 100644 data/2020/neurips/SGD with shuffling: optimal rates without component convexity and large epoch requirements create mode 100644 data/2020/neurips/SIRI: Spatial Relation Induced Network For Spatial Description Resolution create mode 100644 data/2020/neurips/SLIP: Learning to predict in unknown dynamical systems with long-term memory create mode 100644 data/2020/neurips/SMYRF - Efficient Attention using Asymmetric Clustering create mode 100644 data/2020/neurips/SOLOv2: Dynamic and Fast Instance Segmentation create mode 100644 data/2020/neurips/STEER : Simple Temporal Regularization For Neural ODE create mode 100644 data/2020/neurips/STLnet: Signal Temporal Logic Enforced Multivariate Recurrent Neural Networks create mode 100644 data/2020/neurips/SURF: A Simple, Universal, Robust, Fast Distribution Learning Algorithm create mode 100644 data/2020/neurips/SVGD as a kernelized Wasserstein gradient flow of the chi-squared divergence create mode 100644 data/2020/neurips/Safe Reinforcement Learning via Curriculum Induction create mode 100644 data/2020/neurips/Sample Complexity of Asynchronous Q-Learning: Sharper Analysis and Variance Reduction create mode 100644 data/2020/neurips/Sample Complexity of Uniform Convergence for Multicalibration create mode 100644 data/2020/neurips/Sample Efficient Reinforcement Learning via Low-Rank Matrix Estimation create mode 100644 data/2020/neurips/Sample complexity and effective dimension for regression on manifolds create mode 100644 data/2020/neurips/Sample-Efficient Optimization in the Latent Space of Deep Generative Models via Weighted Retraining create mode 100644 data/2020/neurips/Sample-Efficient Reinforcement Learning of Undercomplete POMDPs create mode 100644 data/2020/neurips/Sampling from a k-DPP without looking at all items create mode 100644 data/2020/neurips/Sampling-Decomposable Generative Adversarial Recommender create mode 100644 data/2020/neurips/Sanity-Checking Pruning Methods: Random Tickets can Win the Jackpot create mode 100644 data/2020/neurips/Scalable Belief Propagation via Relaxed Scheduling create mode 100644 data/2020/neurips/Scalable Graph Neural Networks via Bidirectional Propagation create mode 100644 data/2020/neurips/Scalable Multi-Agent Reinforcement Learning for Networked Systems with Average Reward create mode 100644 data/2020/neurips/ScaleCom: Scalable Sparsified Gradient Compression for Communication-Efficient Distributed Training create mode 100644 data/2020/neurips/Scattering GCN: Overcoming Oversmoothness in Graph Convolutional Networks create mode 100644 data/2020/neurips/Searching for Low-Bit Weights in Quantized Neural Networks create mode 100644 data/2020/neurips/Second Order Optimality in Decentralized Non-Convex Optimization via Perturbed Gradient Tracking create mode 100644 data/2020/neurips/Second Order PAC-Bayesian Bounds for the Weighted Majority Vote create mode 100644 data/2020/neurips/Secretary and Online Matching Problems with Machine Learned Advice create mode 100644 data/2020/neurips/Security Analysis of Safe and Seldonian Reinforcement Learning Algorithms create mode 100644 data/2020/neurips/See, Hear, Explore: Curiosity via Audio-Visual Association create mode 100644 data/2020/neurips/Self-Adaptive Training: beyond Empirical Risk Minimization create mode 100644 "data/2020/neurips/Self-Adaptively Learning to Demoir\303\251 from Focused and Defocused Image Pairs" create mode 100644 data/2020/neurips/Self-Distillation Amplifies Regularization in Hilbert Space create mode 100644 data/2020/neurips/Self-Distillation as Instance-Specific Label Smoothing create mode 100644 data/2020/neurips/Self-Imitation Learning via Generalized Lower Bound Q-learning create mode 100644 data/2020/neurips/Self-Learning Transformations for Improving Gaze and Head Redirection create mode 100644 data/2020/neurips/Self-Paced Deep Reinforcement Learning create mode 100644 data/2020/neurips/Self-Supervised Few-Shot Learning on Point Clouds create mode 100644 data/2020/neurips/Self-Supervised Generative Adversarial Compression create mode 100644 data/2020/neurips/Self-Supervised Learning by Cross-Modal Audio-Video Clustering create mode 100644 data/2020/neurips/Self-Supervised MultiModal Versatile Networks create mode 100644 data/2020/neurips/Self-Supervised Relational Reasoning for Representation Learning create mode 100644 data/2020/neurips/Self-Supervised Relationship Probing create mode 100644 data/2020/neurips/Self-Supervised Visual Representation Learning from Hierarchical Grouping create mode 100644 data/2020/neurips/Self-paced Contrastive Learning with Hybrid Memory for Domain Adaptive Object Re-ID create mode 100644 data/2020/neurips/Self-supervised Auxiliary Learning with Meta-paths for Heterogeneous Graphs create mode 100644 data/2020/neurips/Self-supervised Co-Training for Video Representation Learning create mode 100644 data/2020/neurips/Self-supervised learning through the eyes of a child create mode 100644 data/2020/neurips/Self-training Avoids Using Spurious Features Under Domain Shift create mode 100644 data/2020/neurips/Semantic Visual Navigation by Watching YouTube Videos create mode 100644 data/2020/neurips/Semi-Supervised Neural Architecture Search create mode 100644 data/2020/neurips/Semi-Supervised Partial Label Learning via Confidence-Rated Margin Maximization create mode 100644 data/2020/neurips/Semialgebraic Optimization for Lipschitz Constants of ReLU Networks create mode 100644 data/2020/neurips/Sense and Sensitivity Analysis: Simple Post-Hoc Analysis of Bias Due to Unobserved Confounding create mode 100644 data/2020/neurips/Sequential Bayesian Experimental Design with Variable Cost Structure create mode 100644 data/2020/neurips/Set2Graph: Learning Graphs From Sets create mode 100644 data/2020/neurips/ShapeFlow: Learnable Deformation Flows Among 3D Shapes create mode 100644 data/2020/neurips/Shared Experience Actor-Critic for Multi-Agent Reinforcement Learning create mode 100644 data/2020/neurips/Shared Space Transfer Learning for analyzing multi-site fMRI data create mode 100644 data/2020/neurips/Sharp Representation Theorems for ReLU Networks with Precise Dependence on Depth create mode 100644 data/2020/neurips/Sharp uniform convergence bounds through empirical centralization create mode 100644 data/2020/neurips/Sharpened Generalization Bounds based on Conditional Mutual Information and an Application to Noisy, Iterative Algorithms create mode 100644 data/2020/neurips/Sharper Generalization Bounds for Pairwise Learning create mode 100644 data/2020/neurips/ShiftAddNet: A Hardware-Inspired Deep Network create mode 100644 data/2020/neurips/Simple and Fast Algorithm for Binary Integer and Online Linear Programming create mode 100644 data/2020/neurips/Simple and Principled Uncertainty Estimation with Deterministic Deep Learning via Distance Awareness create mode 100644 data/2020/neurips/Simple and Scalable Sparse k-means Clustering via Feature Ranking create mode 100644 data/2020/neurips/Simplify and Robustify Negative Sampling for Implicit Collaborative Filtering create mode 100644 data/2020/neurips/Simulating a Primary Visual Cortex at the Front of CNNs Improves Robustness to Image Perturbations create mode 100644 data/2020/neurips/Simultaneous Preference and Metric Learning from Paired Comparisons create mode 100644 data/2020/neurips/Simultaneously Learning Stochastic and Adversarial Episodic MDPs with Known Transition create mode 100644 data/2020/neurips/Sinkhorn Barycenter via Functional Gradient Descent create mode 100644 data/2020/neurips/Sinkhorn Natural Gradient for Generative Models create mode 100644 data/2020/neurips/Skeleton-bridged Point Completion: From Global Inference to Local Adjustment create mode 100644 data/2020/neurips/Sliding Window Algorithms for k-Clustering Problems create mode 100644 data/2020/neurips/Small Nash Equilibrium Certificates in Very Large Games create mode 100644 data/2020/neurips/Smooth And Consistent Probabilistic Regression Trees create mode 100644 data/2020/neurips/Smoothed Analysis of Online and Differentially Private Learning create mode 100644 data/2020/neurips/Smoothed Geometry for Robust Attribution create mode 100644 data/2020/neurips/Smoothly Bounding User Contributions in Differential Privacy create mode 100644 data/2020/neurips/SnapBoost: A Heterogeneous Boosting Machine create mode 100644 data/2020/neurips/Soft Contrastive Learning for Visual Localization create mode 100644 data/2020/neurips/SoftFlow: Probabilistic Framework for Normalizing Flow on Manifolds create mode 100644 data/2020/neurips/Softmax Deep Double Deterministic Policy Gradients create mode 100644 data/2020/neurips/Solver-in-the-Loop: Learning from Differentiable Physics to Interact with Iterative PDE-Solvers create mode 100644 data/2020/neurips/Space-Time Correspondence as a Contrastive Random Walk create mode 100644 data/2020/neurips/Sparse Graphical Memory for Robust Planning create mode 100644 data/2020/neurips/Sparse Learning with CART create mode 100644 data/2020/neurips/Sparse Spectrum Warped Input Measures for Nonstationary Kernel Learning create mode 100644 data/2020/neurips/Sparse Symplectically Integrated Neural Networks create mode 100644 data/2020/neurips/Sparse Weight Activation Training create mode 100644 data/2020/neurips/Sparse and Continuous Attention Mechanisms create mode 100644 data/2020/neurips/Spectra of the Conjugate Kernel and Neural Tangent Kernel for linear-width neural networks create mode 100644 data/2020/neurips/Spectral Temporal Graph Neural Network for Multivariate Time-series Forecasting create mode 100644 data/2020/neurips/Spike and slab variational Bayes for high dimensional logistic regression create mode 100644 data/2020/neurips/Spin-Weighted Spherical CNNs create mode 100644 data/2020/neurips/Stability of Stochastic Gradient Descent on Nonsmooth Convex Losses create mode 100644 data/2020/neurips/Stable and expressive recurrent vision models create mode 100644 data/2020/neurips/Stage-wise Conservative Linear Bandits create mode 100644 data/2020/neurips/Stateful Posted Pricing with Vanishing Regret via Dynamic Deterministic Markov Decision Processes create mode 100644 data/2020/neurips/Stationary Activations for Uncertainty Calibration in Deep Learning create mode 100644 data/2020/neurips/Statistical Efficiency of Thompson Sampling for Combinatorial Semi-Bandits create mode 100644 data/2020/neurips/Statistical Guarantees of Distributed Nearest Neighbor Classification create mode 100644 data/2020/neurips/Statistical Optimal Transport posed as Learning Kernel Embedding create mode 100644 data/2020/neurips/Statistical and Topological Properties of Sliced Probability Divergences create mode 100644 data/2020/neurips/Statistical control for spatio-temporal MEG EEG source imaging with desparsified mutli-task Lasso create mode 100644 data/2020/neurips/Statistical-Query Lower Bounds via Functional Gradients create mode 100644 data/2020/neurips/Steady State Analysis of Episodic Reinforcement Learning create mode 100644 data/2020/neurips/Steering Distortions to Preserve Classes and Neighbors in Supervised Dimensionality Reduction create mode 100644 data/2020/neurips/Stein Self-Repulsive Dynamics: Benefits From Past Samples create mode 100644 data/2020/neurips/Stochastic Deep Gaussian Processes over Graphs create mode 100644 data/2020/neurips/Stochastic Gradient Descent in Correlated Settings: A Study on Gaussian Processes create mode 100644 data/2020/neurips/Stochastic Latent Actor-Critic: Deep Reinforcement Learning with a Latent Variable Model create mode 100644 data/2020/neurips/Stochastic Normalization create mode 100644 data/2020/neurips/Stochastic Normalizing Flows create mode 100644 data/2020/neurips/Stochastic Optimization for Performative Prediction create mode 100644 data/2020/neurips/Stochastic Optimization with Heavy-Tailed Noise via Accelerated Gradient Clipping create mode 100644 data/2020/neurips/Stochastic Optimization with Laggard Data Pipelines create mode 100644 data/2020/neurips/Stochastic Recursive Gradient Descent Ascent for Stochastic Nonconvex-Strongly-Concave Minimax Problems create mode 100644 data/2020/neurips/Stochastic Segmentation Networks: Modelling Spatially Correlated Aleatoric Uncertainty create mode 100644 data/2020/neurips/Stochastic Stein Discrepancies create mode 100644 data/2020/neurips/Stochasticity of Deterministic Gradient Descent: Large Learning Rate for Multiscale Objective Function create mode 100644 data/2020/neurips/Storage Efficient and Dynamic Flexible Runtime Channel Pruning via Deep Reinforcement Learning create mode 100644 data/2020/neurips/StratLearner: Learning a Strategy for Misinformation Prevention in Social Networks create mode 100644 data/2020/neurips/Strictly Batch Imitation Learning by Energy-based Distribution Matching create mode 100644 data/2020/neurips/Strongly Incremental Constituency Parsing with Graph Neural Networks create mode 100644 data/2020/neurips/Strongly local p-norm-cut algorithms for semi-supervised learning and local graph clustering create mode 100644 data/2020/neurips/Structured Convolutions for Efficient Neural Network Design create mode 100644 data/2020/neurips/Structured Prediction for Conditional Meta-Learning create mode 100644 data/2020/neurips/Sub-linear Regret Bounds for Bayesian Optimisation in Unknown Search Spaces create mode 100644 data/2020/neurips/Sub-sampling for Efficient Non-Parametric Bandit Exploration create mode 100644 data/2020/neurips/Subgraph Neural Networks create mode 100644 data/2020/neurips/Subgroup-based Rank-1 Lattice Quasi-Monte Carlo create mode 100644 data/2020/neurips/Submodular Maximization Through Barrier Functions create mode 100644 data/2020/neurips/Submodular Meta-Learning create mode 100644 data/2020/neurips/Succinct and Robust Multi-Agent Communication With Temporal Message Control create mode 100644 data/2020/neurips/Sufficient dimension reduction for classification using principal optimal transport direction create mode 100644 data/2020/neurips/SuperLoss: A Generic Loss for Robust Curriculum Learning create mode 100644 data/2020/neurips/Supermasks in Superposition create mode 100644 data/2020/neurips/Supervised Contrastive Learning create mode 100644 data/2020/neurips/SurVAE Flows: Surjections to Bridge the Gap between VAEs and Flows create mode 100644 data/2020/neurips/Swapping Autoencoder for Deep Image Manipulation create mode 100644 data/2020/neurips/Synbols: Probing Learning Algorithms with Synthetic Datasets create mode 100644 data/2020/neurips/Synthesize, Execute and Debug: Learning to Repair for Neural Program Synthesis create mode 100644 data/2020/neurips/Synthesizing Tasks for Block-based Programming create mode 100644 data/2020/neurips/Synthetic Data Generators - Sequential and Private create mode 100644 data/2020/neurips/System Identification with Biophysical Constraints: A Circuit Model of the Inner Retina create mode 100644 data/2020/neurips/TSPNet: Hierarchical Feature Learning via Temporal Semantic Pyramid for Sign Language Translation create mode 100644 data/2020/neurips/Tackling the Objective Inconsistency Problem in Heterogeneous Federated Optimization create mode 100644 data/2020/neurips/Taming Discrete Integration via the Boon of Dimensionality create mode 100644 data/2020/neurips/Targeted Adversarial Perturbations for Monocular Depth Prediction create mode 100644 data/2020/neurips/Task-Agnostic Amortized Inference of Gaussian Process Hyperparameters create mode 100644 data/2020/neurips/Task-Agnostic Online Reinforcement Learning with an Infinite Mixture of Gaussian Processes create mode 100644 data/2020/neurips/Task-Oriented Feature Distillation create mode 100644 data/2020/neurips/Task-Robust Model-Agnostic Meta-Learning create mode 100644 data/2020/neurips/Task-agnostic Exploration in Reinforcement Learning create mode 100644 data/2020/neurips/TaylorGAN: Neighbor-Augmented Policy Update Towards Sample-Efficient Natural Language Generation create mode 100644 data/2020/neurips/Teaching a GAN What Not to Learn create mode 100644 data/2020/neurips/Telescoping Density-Ratio Estimation create mode 100644 data/2020/neurips/Temporal Positive-unlabeled Learning for Biomedical Hypothesis Generation via Risk Estimation create mode 100644 data/2020/neurips/Temporal Spike Sequence Learning via Backpropagation for Deep Spiking Neural Networks create mode 100644 data/2020/neurips/Temporal Variability in Implicit Online Learning create mode 100644 data/2020/neurips/Tensor Completion Made Practical create mode 100644 data/2020/neurips/Testing Determinantal Point Processes create mode 100644 data/2020/neurips/Texture Interpolation for Probing Visual Perception create mode 100644 data/2020/neurips/The Adaptive Complexity of Maximizing a Gross Substitutes Valuation create mode 100644 data/2020/neurips/The Advantage of Conditional Meta-Learning for Biased Regularization and Fine Tuning create mode 100644 data/2020/neurips/The All-or-Nothing Phenomenon in Sparse Tensor PCA create mode 100644 data/2020/neurips/The Autoencoding Variational Autoencoder create mode 100644 data/2020/neurips/The Complete Lasso Tradeoff Diagram create mode 100644 data/2020/neurips/The Complexity of Adversarially Robust Proper Learning of Halfspaces with Agnostic Noise create mode 100644 data/2020/neurips/The Cone of Silence: Speech Separation by Localization create mode 100644 data/2020/neurips/The Convex Relaxation Barrier, Revisited: Tightened Single-Neuron Relaxations for Neural Network Verification create mode 100644 data/2020/neurips/The Convolution Exponential and Generalized Sylvester Flows create mode 100644 data/2020/neurips/The Devil is in the Detail: A Framework for Macroscopic Prediction via Microscopic Models create mode 100644 data/2020/neurips/The Dilemma of TriHard Loss and an Element-Weighted TriHard Loss for Person Re-Identification create mode 100644 data/2020/neurips/The Discrete Gaussian for Differential Privacy create mode 100644 data/2020/neurips/The Diversified Ensemble Neural Network create mode 100644 data/2020/neurips/The Flajolet-Martin Sketch Itself Preserves Differential Privacy: Private Counting with Minimal Space create mode 100644 data/2020/neurips/The Generalization-Stability Tradeoff In Neural Network Pruning create mode 100644 data/2020/neurips/The Generalized Lasso with Nonlinear Observations and Generative Priors create mode 100644 data/2020/neurips/The Hateful Memes Challenge: Detecting Hate Speech in Multimodal Memes create mode 100644 data/2020/neurips/The Implications of Local Correlation on Learning Some Deep Functions create mode 100644 data/2020/neurips/The LoCA Regret: A Consistent Metric to Evaluate Model-Based Behavior in Reinforcement Learning create mode 100644 data/2020/neurips/The Lottery Ticket Hypothesis for Pre-trained BERT Networks create mode 100644 data/2020/neurips/The MAGICAL Benchmark for Robust Imitation create mode 100644 data/2020/neurips/The Mean-Squared Error of Double Q-Learning create mode 100644 data/2020/neurips/The NetHack Learning Environment create mode 100644 data/2020/neurips/The Origins and Prevalence of Texture Bias in Convolutional Neural Networks create mode 100644 data/2020/neurips/The Pitfalls of Simplicity Bias in Neural Networks create mode 100644 data/2020/neurips/The Potts-Ising model for discrete multivariate data create mode 100644 data/2020/neurips/The Power of Comparisons for Actively Learning Linear Classifiers create mode 100644 data/2020/neurips/The Power of Predictions in Online Control create mode 100644 data/2020/neurips/The Primal-Dual method for Learning Augmented Algorithms create mode 100644 data/2020/neurips/The Smoothed Possibility of Social Choice create mode 100644 data/2020/neurips/The Statistical Complexity of Early-Stopped Mirror Descent create mode 100644 data/2020/neurips/The Strong Screening Rule for SLOPE create mode 100644 data/2020/neurips/The Surprising Simplicity of the Early-Time Learning Dynamics of Neural Networks create mode 100644 data/2020/neurips/The Value Equivalence Principle for Model-Based Reinforcement Learning create mode 100644 data/2020/neurips/The Wasserstein Proximal Gradient Algorithm create mode 100644 data/2020/neurips/The interplay between randomness and structure during learning in RNNs create mode 100644 data/2020/neurips/The phase diagram of approximation rates for deep neural networks create mode 100644 data/2020/neurips/The route to chaos in routing games: When is price of anarchy too optimistic? create mode 100644 data/2020/neurips/Theoretical Insights Into Multiclass Classification: A High-dimensional Asymptotic View create mode 100644 data/2020/neurips/Theory-Inspired Path-Regularized Differential Network Architecture Search create mode 100644 data/2020/neurips/Throughput-Optimal Topology Design for Cross-Silo Federated Learning create mode 100644 data/2020/neurips/Thunder: a Fast Coordinate Selection Solver for Sparse Learning create mode 100644 data/2020/neurips/Tight First- and Second-Order Regret Bounds for Adversarial Linear Bandits create mode 100644 data/2020/neurips/Tight Nonparametric Convergence Rates for Stochastic Gradient Descent under the Noiseless Linear Model create mode 100644 data/2020/neurips/Tight last-iterate convergence rates for no-regret learning in multi-player games create mode 100644 data/2020/neurips/Time-Reversal Symmetric ODE Network create mode 100644 data/2020/neurips/Timeseries Anomaly Detection using Temporal Hierarchical One-Class Network create mode 100644 data/2020/neurips/TinyTL: Reduce Memory, Not Parameters for Efficient On-Device Learning create mode 100644 data/2020/neurips/Top-KAST: Top-K Always Sparse Training create mode 100644 data/2020/neurips/Top-k Training of GANs: Improving GAN Performance by Throwing Away Bad Samples create mode 100644 data/2020/neurips/TorsionNet: A Reinforcement Learning Approach to Sequential Conformer Search create mode 100644 data/2020/neurips/Toward the Fundamental Limits of Imitation Learning create mode 100644 data/2020/neurips/Towards Better Generalization of Adaptive Gradient Methods create mode 100644 data/2020/neurips/Towards Convergence Rate Analysis of Random Forests for Classification create mode 100644 data/2020/neurips/Towards Crowdsourced Training of Large Neural Networks using Decentralized Mixture-of-Experts create mode 100644 data/2020/neurips/Towards Deeper Graph Neural Networks with Differentiable Group Normalization create mode 100644 data/2020/neurips/Towards Interpretable Natural Language Understanding with Explanations as Latent Variables create mode 100644 data/2020/neurips/Towards Learning Convolutions from Scratch create mode 100644 data/2020/neurips/Towards Maximizing the Representation Gap between In-Domain & Out-of-Distribution Examples create mode 100644 data/2020/neurips/Towards Minimax Optimal Reinforcement Learning in Factored Markov Decision Processes create mode 100644 data/2020/neurips/Towards More Practical Adversarial Attacks on Graph Neural Networks create mode 100644 data/2020/neurips/Towards Neural Programming Interfaces create mode 100644 data/2020/neurips/Towards Playing Full MOBA Games with Deep Reinforcement Learning create mode 100644 data/2020/neurips/Towards Problem-dependent Optimal Learning Rates create mode 100644 data/2020/neurips/Towards Safe Policy Improvement for Non-Stationary MDPs create mode 100644 data/2020/neurips/Towards Scalable Bayesian Learning of Causal DAGs create mode 100644 data/2020/neurips/Towards Scale-Invariant Graph-related Problem Solving by Iterative Homogeneous GNNs create mode 100644 data/2020/neurips/Towards Theoretically Understanding Why Sgd Generalizes Better Than Adam in Deep Learning create mode 100644 data/2020/neurips/Towards Understanding Hierarchical Learning: Benefits of Neural Representations create mode 100644 data/2020/neurips/Towards a Better Global Loss Landscape of GANs create mode 100644 data/2020/neurips/Towards a Combinatorial Characterization of Bounded-Memory Learning create mode 100644 data/2020/neurips/Towards practical differentially private causal graph discovery create mode 100644 data/2020/neurips/Trade-offs and Guarantees of Adversarial Representation Learning for Information Obfuscation create mode 100644 data/2020/neurips/Trading Personalization for Accuracy: Data Debugging in Collaborative Filtering create mode 100644 data/2020/neurips/Train-by-Reconnect: Decoupling Locations of Weights from Their Values create mode 100644 data/2020/neurips/Training Generative Adversarial Networks by Solving Ordinary Differential Equations create mode 100644 data/2020/neurips/Training Generative Adversarial Networks with Limited Data create mode 100644 data/2020/neurips/Training Linear Finite-State Machines create mode 100644 data/2020/neurips/Training Normalizing Flows with the Information Bottleneck for Competitive Generative Classification create mode 100644 data/2020/neurips/Training Stronger Baselines for Learning to Optimize create mode 100644 data/2020/neurips/Trajectory-wise Multiple Choice Learning for Dynamics Generalization in Reinforcement Learning create mode 100644 "data/2020/neurips/Transfer Learning via \342\204\2231 Regularization" create mode 100644 data/2020/neurips/Transferable Calibration with Lower Bias and Variance in Domain Adaptation create mode 100644 data/2020/neurips/Transferable Graph Optimizers for ML Compilers create mode 100644 data/2020/neurips/Tree! I am no Tree! I am a low dimensional Hyperbolic Embedding create mode 100644 data/2020/neurips/Triple descent and the two kinds of overfitting: where & why do they appear? create mode 100644 data/2020/neurips/Truncated Linear Regression in High Dimensions create mode 100644 data/2020/neurips/Trust the Model When It Is Confident: Masked Model-based Actor-Critic create mode 100644 data/2020/neurips/Truthful Data Acquisition via Peer Prediction create mode 100644 data/2020/neurips/UCLID-Net: Single View Reconstruction in Object Space create mode 100644 data/2020/neurips/UCSG-NET- Unsupervised Discovering of Constructive Solid Geometry Tree create mode 100644 data/2020/neurips/UDH: Universal Deep Hiding for Steganography, Watermarking, and Light Field Messaging create mode 100644 data/2020/neurips/UWSOD: Toward Fully-Supervised-Level Capacity Weakly Supervised Object Detection create mode 100644 data/2020/neurips/Ultra-Low Precision 4-bit Training of Deep Neural Networks create mode 100644 data/2020/neurips/Ultrahyperbolic Representation Learning create mode 100644 data/2020/neurips/UnModNet: Learning to Unwrap a Modulo Image for High Dynamic Range Imaging create mode 100644 data/2020/neurips/Unbalanced Sobolev Descent create mode 100644 data/2020/neurips/Uncertainty Aware Semi-Supervised Learning on Graph Data create mode 100644 data/2020/neurips/Uncertainty Quantification for Inferring Hawkes Networks create mode 100644 data/2020/neurips/Uncertainty-Aware Learning for Zero-Shot Semantic Segmentation create mode 100644 data/2020/neurips/Uncertainty-aware Self-training for Few-shot Text Classification create mode 100644 data/2020/neurips/Uncovering the Topology of Time-Varying fMRI Data using Cubical Persistence create mode 100644 data/2020/neurips/Understanding Anomaly Detection with Deep Invertible Networks through Hierarchies of Distributions and Features create mode 100644 data/2020/neurips/Understanding Approximate Fisher Information for Fast Convergence of Natural Gradient Descent in Wide Neural Networks create mode 100644 data/2020/neurips/Understanding Deep Architecture with Reasoning Layer create mode 100644 data/2020/neurips/Understanding Double Descent Requires A Fine-Grained Bias-Variance Decomposition create mode 100644 data/2020/neurips/Understanding Global Feature Contributions With Additive Importance Measures create mode 100644 data/2020/neurips/Understanding Gradient Clipping in Private SGD: A Geometric Perspective create mode 100644 data/2020/neurips/Understanding and Exploring the Network with Stochastic Architectures create mode 100644 data/2020/neurips/Understanding and Improving Fast Adversarial Training create mode 100644 data/2020/neurips/Understanding spiking networks through convex optimization create mode 100644 data/2020/neurips/Understanding the Role of Training Regimes in Continual Learning create mode 100644 data/2020/neurips/Unfolding recurrence by Green's functions for optimized reservoir computing create mode 100644 data/2020/neurips/Unfolding the Alternating Optimization for Blind Super Resolution create mode 100644 data/2020/neurips/Unifying Activation- and Timing-based Learning Rules for Spiking Neural Networks create mode 100644 data/2020/neurips/Universal Domain Adaptation through Self Supervision create mode 100644 data/2020/neurips/Universal Function Approximation on Graphs create mode 100644 data/2020/neurips/Universal guarantees for decision tree induction via a higher-order splitting criterion create mode 100644 data/2020/neurips/Universally Quantized Neural Compression create mode 100644 data/2020/neurips/Unreasonable Effectiveness of Greedy Algorithms in Multi-Armed Bandit with Many Arms create mode 100644 data/2020/neurips/Unsupervised Data Augmentation for Consistency Training create mode 100644 data/2020/neurips/Unsupervised Joint k-node Graph Representations with Compositional Energy-Based Models create mode 100644 data/2020/neurips/Unsupervised Learning of Dense Visual Representations create mode 100644 data/2020/neurips/Unsupervised Learning of Lagrangian Dynamics from Images for Prediction and Control create mode 100644 data/2020/neurips/Unsupervised Learning of Object Landmarks via Self-Training Correspondence create mode 100644 data/2020/neurips/Unsupervised Learning of Visual Features by Contrasting Cluster Assignments create mode 100644 data/2020/neurips/Unsupervised Representation Learning by Invariance Propagation create mode 100644 data/2020/neurips/Unsupervised Semantic Aggregation and Deformable Template Matching for Semi-Supervised Learning create mode 100644 data/2020/neurips/Unsupervised Sound Separation Using Mixture Invariant Training create mode 100644 data/2020/neurips/Unsupervised Text Generation by Learning from Search create mode 100644 data/2020/neurips/Unsupervised Translation of Programming Languages create mode 100644 data/2020/neurips/Unsupervised object-centric video generation and decomposition in 3D create mode 100644 data/2020/neurips/Untangling tradeoffs between recurrence and self-attention in artificial neural networks create mode 100644 data/2020/neurips/Upper Confidence Primal-Dual Reinforcement Learning for CMDP with Adversarial Loss create mode 100644 data/2020/neurips/User-Dependent Neural Sequence Models for Continuous-Time Event Data create mode 100644 data/2020/neurips/Using noise to probe recurrent neural network structure and prune synapses create mode 100644 data/2020/neurips/VAEM: a Deep Generative Model for Heterogeneous Mixed Type Data create mode 100644 data/2020/neurips/VIME: Extending the Success of Self- and Semi-supervised Learning to Tabular Domain create mode 100644 data/2020/neurips/Value-driven Hindsight Modelling create mode 100644 data/2020/neurips/VarGrad: A Low-Variance Gradient Estimator for Variational Inference create mode 100644 data/2020/neurips/Variance Reduction via Accelerated Dual Averaging for Finite-Sum Optimization create mode 100644 data/2020/neurips/Variance reduction for Random Coordinate Descent-Langevin Monte Carlo create mode 100644 data/2020/neurips/Variance-Reduced Off-Policy TDC Learning: Non-Asymptotic Convergence Analysis create mode 100644 data/2020/neurips/Variational Amodal Object Completion create mode 100644 data/2020/neurips/Variational Bayesian Monte Carlo with Noisy Likelihoods create mode 100644 data/2020/neurips/Variational Bayesian Unlearning create mode 100644 data/2020/neurips/Variational Inference for Graph Convolutional Networks in the Absence of Graph Data and Adversarial Settings create mode 100644 data/2020/neurips/Variational Policy Gradient Method for Reinforcement Learning with General Utilities create mode 100644 data/2020/neurips/Video Frame Interpolation without Temporal Priors create mode 100644 data/2020/neurips/Video Object Segmentation with Adaptive Feature Bank and Uncertain-Region Refinement create mode 100644 data/2020/neurips/Walking in the Shadow: A New Perspective on Descent Directions for Constrained Minimization create mode 100644 data/2020/neurips/Walsh-Hadamard Variational Inference for Bayesian Deep Learning create mode 100644 data/2020/neurips/Wasserstein Distances for Stereo Disparity Estimation create mode 100644 data/2020/neurips/Watch out! Motion is Blurring the Vision of Your Deep Neural Networks create mode 100644 data/2020/neurips/Wavelet Flow: Fast Training of High Resolution Normalizing Flows create mode 100644 data/2020/neurips/Weak Form Generalized Hamiltonian Learning create mode 100644 data/2020/neurips/Weakly Supervised Deep Functional Maps for Shape Matching create mode 100644 data/2020/neurips/Weakly-Supervised Reinforcement Learning for Controllable Behavior create mode 100644 data/2020/neurips/Weighted QMIX: Expanding Monotonic Value Function Factorisation for Deep Multi-Agent Reinforcement Learning create mode 100644 data/2020/neurips/Weisfeiler and Leman go sparse: Towards scalable higher-order graph embeddings create mode 100644 data/2020/neurips/Weston-Watkins Hinge Loss and Ordered Partitions create mode 100644 data/2020/neurips/What Did You Think Would Happen? Explaining Agent Behaviour through Intended Outcomes create mode 100644 data/2020/neurips/What Do Neural Networks Learn When Trained With Random Labels? create mode 100644 data/2020/neurips/What Makes for Good Views for Contrastive Learning? create mode 100644 data/2020/neurips/What Neural Networks Memorize and Why: Discovering the Long Tail via Influence Estimation create mode 100644 data/2020/neurips/What if Neural Networks had SVDs? create mode 100644 data/2020/neurips/What is being transferred in transfer learning? create mode 100644 data/2020/neurips/What shapes feature representations? Exploring datasets, architectures, and training create mode 100644 data/2020/neurips/What went wrong and when? Instance-wise feature importance for time-series black-box models create mode 100644 data/2020/neurips/When Counterpoint Meets Chinese Folk Melodies create mode 100644 data/2020/neurips/When Do Neural Networks Outperform Kernel Methods? create mode 100644 data/2020/neurips/When and How to Lift the Lockdown? Global COVID-19 Scenario Analysis and Policy Assessment using Compartmental Gaussian Processes create mode 100644 data/2020/neurips/Why Do Deep Residual Networks Generalize Better than Deep Feedforward Networks? - A Neural Tangent Kernel Perspective create mode 100644 data/2020/neurips/Why Normalizing Flows Fail to Detect Out-of-Distribution Data create mode 100644 data/2020/neurips/Why are Adaptive Methods Good for Attention Models? create mode 100644 data/2020/neurips/Winning the Lottery with Continuous Sparsification create mode 100644 data/2020/neurips/Wisdom of the Ensemble: Improving Consistency of Deep Learning Models create mode 100644 data/2020/neurips/WoodFisher: Efficient Second-Order Approximation for Neural Network Compression create mode 100644 data/2020/neurips/Woodbury Transformations for Deep Generative Flows create mode 100644 data/2020/neurips/Worst-Case Analysis for Randomly Collected Data create mode 100644 data/2020/neurips/X-CAL: Explicit Calibration for Survival Analysis create mode 100644 data/2020/neurips/Your Classifier can Secretly Suffice Multi-Source Domain Adaptation create mode 100644 data/2020/neurips/Your GAN is Secretly an Energy-based Model and You Should Use Discriminator Driven Latent Sampling create mode 100644 data/2020/neurips/Zap Q-Learning With Nonlinear Function Approximation create mode 100644 data/2020/neurips/Zero-Resource Knowledge-Grounded Dialogue Generation create mode 100644 data/2020/neurips/f-Divergence Variational Inference create mode 100644 data/2020/neurips/f-GAIL: Learning f-Divergence for Generative Adversarial Imitation Learning create mode 100644 data/2020/neurips/wav2vec 2.0: A Framework for Self-Supervised Learning of Speech Representations create mode 100644 data/2021/neurips/(Almost) Free Incentivized Exploration from Decentralized Learning Agents create mode 100644 data/2021/neurips/3D Pose Transfer with Correspondence Learning and Mesh Refinement create mode 100644 data/2021/neurips/3D Siamese Voxel-to-BEV Tracker for Sparse Point Clouds create mode 100644 data/2021/neurips/3DP3: 3D Scene Perception via Probabilistic Programming create mode 100644 data/2021/neurips/A 3D Generative Model for Structure-Based Drug Design create mode 100644 data/2021/neurips/A B Testing for Recommender Systems in a Two-sided Marketplace create mode 100644 data/2021/neurips/A B n Testing with Control in the Presence of Subpopulations create mode 100644 data/2021/neurips/A Bayesian-Symbolic Approach to Reasoning and Learning in Intuitive Physics create mode 100644 data/2021/neurips/A Bi-Level Framework for Learning to Solve Combinatorial Optimization on Graphs create mode 100644 data/2021/neurips/A Biased Graph Neural Network Sampler with Near-Optimal Regret create mode 100644 data/2021/neurips/A Causal Lens for Controllable Text Generation create mode 100644 data/2021/neurips/A Central Limit Theorem for Differentially Private Query Answering create mode 100644 data/2021/neurips/A Closer Look at the Worst-case Behavior of Multi-armed Bandit Algorithms create mode 100644 data/2021/neurips/A Compositional Atlas of Tractable Circuit Operations for Probabilistic Inference create mode 100644 data/2021/neurips/A Comprehensively Tight Analysis of Gradient Descent for PCA create mode 100644 data/2021/neurips/A Computationally Efficient Method for Learning Exponential Family Distributions create mode 100644 data/2021/neurips/A Consciousness-Inspired Planning Agent for Model-Based Reinforcement Learning create mode 100644 data/2021/neurips/A Constant Approximation Algorithm for Sequential Random-Order No-Substitution k-Median Clustering create mode 100644 data/2021/neurips/A Continuous Mapping For Augmentation Design create mode 100644 data/2021/neurips/A Contrastive Learning Approach for Training Variational Autoencoder Priors create mode 100644 data/2021/neurips/A Convergence Analysis of Gradient Descent on Graph Neural Networks create mode 100644 data/2021/neurips/A Critical Look at the Consistency of Causal Estimation with Deep Latent Variable Models create mode 100644 data/2021/neurips/A Domain-Shrinking based Bayesian Optimization Algorithm with Order-Optimal Regret Performance create mode 100644 data/2021/neurips/A Faster Decentralized Algorithm for Nonconvex Minimax Problems create mode 100644 data/2021/neurips/A Faster Maximum Cardinality Matching Algorithm with Applications in Machine Learning create mode 100644 data/2021/neurips/A Framework to Learn with Interpretation create mode 100644 data/2021/neurips/A Gang of Adversarial Bandits create mode 100644 data/2021/neurips/A Gaussian Process-Bayesian Bernoulli Mixture Model for Multi-Label Active Learning create mode 100644 data/2021/neurips/A Geometric Analysis of Neural Collapse with Unconstrained Features create mode 100644 data/2021/neurips/A Geometric Perspective towards Neural Calibration via Sensitivity Decomposition create mode 100644 data/2021/neurips/A Geometric Structure of Acceleration and Its Role in Making Gradients Small Fast create mode 100644 data/2021/neurips/A Gradient Method for Multilevel Optimization create mode 100644 data/2021/neurips/A Hierarchical Reinforcement Learning Based Optimization Framework for Large-scale Dynamic Pickup and Delivery Problems create mode 100644 data/2021/neurips/A Highly-Efficient Group Elastic Net Algorithm with an Application to Function-On-Scalar Regression create mode 100644 data/2021/neurips/A Kernel-based Test of Independence for Cluster-correlated Data create mode 100644 data/2021/neurips/A Law of Iterated Logarithm for Multi-Agent Reinforcement Learning create mode 100644 data/2021/neurips/A Little Robustness Goes a Long Way: Leveraging Robust Features for Targeted Transfer Attacks create mode 100644 data/2021/neurips/A Mathematical Framework for Quantifying Transferability in Multi-source Transfer Learning create mode 100644 data/2021/neurips/A Max-Min Entropy Framework for Reinforcement Learning create mode 100644 data/2021/neurips/A Minimalist Approach to Offline Reinforcement Learning create mode 100644 data/2021/neurips/A Multi-Implicit Neural Representation for Fonts create mode 100644 data/2021/neurips/A Near-Optimal Algorithm for Debiasing Trained Machine Learning Models create mode 100644 data/2021/neurips/A Near-Optimal Algorithm for Stochastic Bilevel Optimization via Double-Momentum create mode 100644 data/2021/neurips/A New Theoretical Framework for Fast and Accurate Online Decision-Making create mode 100644 data/2021/neurips/A No-go Theorem for Robust Acceleration in the Hyperbolic Plane create mode 100644 data/2021/neurips/A Non-commutative Extension of Lee-Seung's Algorithm for Positive Semidefinite Factorizations create mode 100644 data/2021/neurips/A Normative and Biologically Plausible Algorithm for Independent Component Analysis create mode 100644 data/2021/neurips/A Note on Sparse Generalized Eigenvalue Problem create mode 100644 data/2021/neurips/A PAC-Bayes Analysis of Adversarial Robustness create mode 100644 data/2021/neurips/A Probabilistic State Space Model for Joint Inference from Differential Equations and Data create mode 100644 data/2021/neurips/A Prototype-Oriented Framework for Unsupervised Domain Adaptation create mode 100644 data/2021/neurips/A Provably Efficient Model-Free Posterior Sampling Method for Episodic Reinforcement Learning create mode 100644 data/2021/neurips/A Provably Efficient Sample Collection Strategy for Reinforcement Learning create mode 100644 data/2021/neurips/A Regression Approach to Learning-Augmented Online Algorithms create mode 100644 data/2021/neurips/A Separation Result Between Data-oblivious and Data-aware Poisoning Attacks create mode 100644 data/2021/neurips/A Shading-Guided Generative Implicit Model for Shape-Accurate 3D-Aware Image Synthesis create mode 100644 data/2021/neurips/A Stochastic Newton Algorithm for Distributed Convex Optimization create mode 100644 data/2021/neurips/A Surrogate Objective Framework for Prediction+Programming with Soft Constraints create mode 100644 data/2021/neurips/A Theoretical Analysis of Fine-tuning with Linear Teachers create mode 100644 data/2021/neurips/A Theory of the Distortion-Perception Tradeoff in Wasserstein Space create mode 100644 data/2021/neurips/A Theory-Driven Self-Labeling Refinement Method for Contrastive Representation Learning create mode 100644 data/2021/neurips/A Topological Perspective on Causal Inference create mode 100644 data/2021/neurips/A Trainable Spectral-Spatial Sparse Coding Model for Hyperspectral Image Restoration create mode 100644 data/2021/neurips/A Unified Approach to Fair Online Learning via Blackwell Approachability create mode 100644 data/2021/neurips/A Unified View of cGANs with and without Classifiers create mode 100644 data/2021/neurips/A Universal Law of Robustness via Isoperimetry create mode 100644 data/2021/neurips/A Variational Perspective on Diffusion-Based Generative Models and Score Matching create mode 100644 data/2021/neurips/A Winning Hand: Compressing Deep Networks Can Improve Out-of-Distribution Robustness create mode 100644 data/2021/neurips/A first-order primal-dual method with adaptivity to local smoothness create mode 100644 data/2021/neurips/A flow-based latent state generative model of neural population responses to natural images create mode 100644 data/2021/neurips/A generative nonparametric Bayesian model for whole genomes create mode 100644 data/2021/neurips/A mechanistic multi-area recurrent network model of decision-making create mode 100644 data/2021/neurips/A nonparametric method for gradual change problems with statistical guarantees create mode 100644 data/2021/neurips/A novel notion of barycenter for probability distributions based on optimal weak mass transport create mode 100644 data/2021/neurips/A sampling-based circuit for optimal decision making create mode 100644 data/2021/neurips/A self consistent theory of Gaussian Processes captures feature learning effects in finite CNNs create mode 100644 data/2021/neurips/A single gradient step finds adversarial examples on random two-layers neural networks create mode 100644 data/2021/neurips/A unified framework for bandit multiple testing create mode 100644 data/2021/neurips/A universal probabilistic spike count model reveals ongoing modulation of neural variability create mode 100644 data/2021/neurips/A variational approximate posterior for the deep Wishart process create mode 100644 data/2021/neurips/A-NeRF: Articulated Neural Radiance Fields for Learning Human Shape, Appearance, and Pose create mode 100644 data/2021/neurips/ABC: Auxiliary Balanced Classifier for Class-imbalanced Semi-supervised Learning create mode 100644 data/2021/neurips/AC DC: Alternating Compressed DeCompressed Training of Deep Neural Networks create mode 100644 data/2021/neurips/AC-GC: Lossy Activation Compression with Guaranteed Convergence create mode 100644 data/2021/neurips/AFEC: Active Forgetting of Negative Transfer in Continual Learning create mode 100644 data/2021/neurips/ASSANet: An Anisotropic Separable Set Abstraction for Efficient Point Cloud Representation Learning create mode 100644 data/2021/neurips/ATISS: Autoregressive Transformers for Indoor Scene Synthesis create mode 100644 data/2021/neurips/Absolute Neighbour Difference based Correlation Test for Detecting Heteroscedastic Relationships create mode 100644 data/2021/neurips/Accelerated Sparse Neural Training: A Provable and Efficient Method to Find N: M Transposable Masks create mode 100644 data/2021/neurips/Accelerating Quadratic Optimization with Reinforcement Learning create mode 100644 data/2021/neurips/Accelerating Robotic Reinforcement Learning via Parameterized Action Primitives create mode 100644 data/2021/neurips/Accommodating Picky Customers: Regret Bound and Exploration Complexity for Multi-Objective Reinforcement Learning create mode 100644 data/2021/neurips/Accumulative Poisoning Attacks on Real-time Data create mode 100644 data/2021/neurips/Accurate Point Cloud Registration with Robust Optimal Transport create mode 100644 data/2021/neurips/Accurately Solving Rod Dynamics with Graph Learning create mode 100644 data/2021/neurips/Achieving Forgetting Prevention and Knowledge Transfer in Continual Learning create mode 100644 data/2021/neurips/Achieving Rotational Invariance with Bessel-Convolutional Neural Networks create mode 100644 data/2021/neurips/Across-animal odor decoding by probabilistic manifold alignment create mode 100644 data/2021/neurips/Action-guided 3D Human Motion Prediction create mode 100644 data/2021/neurips/Activation Sharing with Asymmetric Paths Solves Weight Transport Problem without Bidirectional Connection create mode 100644 data/2021/neurips/Active 3D Shape Reconstruction from Vision and Touch create mode 100644 data/2021/neurips/Active Assessment of Prediction Services as Accuracy Surface Over Attribute Combinations create mode 100644 data/2021/neurips/Active Learning of Convex Halfspaces on Graphs create mode 100644 data/2021/neurips/Active Offline Policy Selection create mode 100644 data/2021/neurips/Active clustering for labeling training data create mode 100644 data/2021/neurips/Actively Identifying Causal Effects with Latent Variables Given Only Response Variable Observable create mode 100644 data/2021/neurips/Adaptable Agent Populations via a Generative Model of Policies create mode 100644 data/2021/neurips/Adapting to function difficulty and growth conditions in private optimization create mode 100644 data/2021/neurips/Adaptive Conformal Inference Under Distribution Shift create mode 100644 data/2021/neurips/Adaptive Data Augmentation on Temporal Graphs create mode 100644 data/2021/neurips/Adaptive Denoising via GainTuning create mode 100644 data/2021/neurips/Adaptive Diffusion in Graph Neural Networks create mode 100644 data/2021/neurips/Adaptive Ensemble Q-learning: Minimizing Estimation Bias via Error Feedback create mode 100644 data/2021/neurips/Adaptive First-Order Methods Revisited: Convex Minimization without Lipschitz Requirements create mode 100644 data/2021/neurips/Adaptive Machine Unlearning create mode 100644 data/2021/neurips/Adaptive Online Packing-guided Search for POMDPs create mode 100644 data/2021/neurips/Adaptive Proximal Gradient Methods for Structured Neural Networks create mode 100644 data/2021/neurips/Adaptive Risk Minimization: Learning to Adapt to Domain Shift create mode 100644 data/2021/neurips/Adaptive Sampling for Minimax Fair Classification create mode 100644 data/2021/neurips/Adaptive wavelet distillation from neural networks through interpretations create mode 100644 data/2021/neurips/Adder Attention for Vision Transformer create mode 100644 data/2021/neurips/Addressing Algorithmic Disparity and Performance Inconsistency in Federated Learning create mode 100644 data/2021/neurips/Adjusting for Autocorrelated Errors in Neural Networks for Time Series create mode 100644 data/2021/neurips/Adversarial Attack Generation Empowered by Min-Max Optimization create mode 100644 data/2021/neurips/Adversarial Attacks on Black Box Video Classifiers: Leveraging the Power of Geometric Transformations create mode 100644 data/2021/neurips/Adversarial Attacks on Graph Classifiers via Bayesian Optimisation create mode 100644 data/2021/neurips/Adversarial Examples Make Strong Poisons create mode 100644 data/2021/neurips/Adversarial Examples for k-Nearest Neighbor Classifiers Based on Higher-Order Voronoi Diagrams create mode 100644 data/2021/neurips/Adversarial Examples in Multi-Layer Random ReLU Networks create mode 100644 data/2021/neurips/Adversarial Feature Desensitization create mode 100644 data/2021/neurips/Adversarial Graph Augmentation to Improve Graph Contrastive Learning create mode 100644 data/2021/neurips/Adversarial Intrinsic Motivation for Reinforcement Learning create mode 100644 data/2021/neurips/Adversarial Neuron Pruning Purifies Backdoored Deep Models create mode 100644 data/2021/neurips/Adversarial Regression with Doubly Non-negative Weighting Matrices create mode 100644 data/2021/neurips/Adversarial Reweighting for Partial Domain Adaptation create mode 100644 data/2021/neurips/Adversarial Robustness with Non-uniform Perturbations create mode 100644 data/2021/neurips/Adversarial Robustness with Semi-Infinite Constrained Learning create mode 100644 data/2021/neurips/Adversarial Robustness without Adversarial Training: A Teacher-Guided Curriculum Learning Approach create mode 100644 data/2021/neurips/Adversarial Teacher-Student Representation Learning for Domain Generalization create mode 100644 data/2021/neurips/Adversarial Training Helps Transfer Learning via Better Representations create mode 100644 data/2021/neurips/Adversarially Robust 3D Point Cloud Recognition Using Self-Supervisions create mode 100644 data/2021/neurips/Adversarially Robust Change Point Detection create mode 100644 data/2021/neurips/Adversarially robust learning for security-constrained optimal power flow create mode 100644 data/2021/neurips/Agent Modelling under Partial Observability for Deep Reinforcement Learning create mode 100644 data/2021/neurips/Agnostic Reinforcement Learning with Low-Rank MDPs and Rich Observations create mode 100644 data/2021/neurips/Algorithmic Instabilities of Accelerated Gradient Descent create mode 100644 data/2021/neurips/Algorithmic stability and generalization of an unsupervised feature selection algorithm create mode 100644 data/2021/neurips/Alias-Free Generative Adversarial Networks create mode 100644 data/2021/neurips/Align before Fuse: Vision and Language Representation Learning with Momentum Distillation create mode 100644 data/2021/neurips/Aligned Structured Sparsity Learning for Efficient Image Super-Resolution create mode 100644 data/2021/neurips/Aligning Pretraining for Detection via Object-Level Contrastive Learning create mode 100644 data/2021/neurips/Aligning Silhouette Topology for Self-Adaptive 3D Human Pose Recovery create mode 100644 data/2021/neurips/Alignment Attention by Matching Key and Query Distributions create mode 100644 data/2021/neurips/All Tokens Matter: Token Labeling for Training Better Vision Transformers create mode 100644 data/2021/neurips/Amortized Synthesis of Constrained Configurations Using a Differentiable Surrogate create mode 100644 data/2021/neurips/Amortized Variational Inference for Simple Hierarchical Models create mode 100644 data/2021/neurips/An Analysis of Constant Step Size SGD in the Non-convex Regime: Asymptotic Normality and Bias create mode 100644 data/2021/neurips/An Axiomatic Theory of Provably-Fair Welfare-Centric Machine Learning create mode 100644 data/2021/neurips/An Efficient Pessimistic-Optimistic Algorithm for Stochastic Linear Bandits with General Constraints create mode 100644 data/2021/neurips/An Efficient Transfer Learning Framework for Multiagent Reinforcement Learning create mode 100644 data/2021/neurips/An Empirical Investigation of Domain Generalization with Empirical Risk Minimizers create mode 100644 data/2021/neurips/An Empirical Study of Adder Neural Networks for Object Detection create mode 100644 data/2021/neurips/An Even More Optimal Stochastic Optimization Algorithm: Minibatching and Interpolation Learning create mode 100644 data/2021/neurips/An Exact Characterization of the Generalization Error for the Gibbs Algorithm create mode 100644 data/2021/neurips/An Exponential Improvement on the Memorization Capacity of Deep Threshold Networks create mode 100644 data/2021/neurips/An Exponential Lower Bound for Linearly Realizable MDP with Constant Suboptimality Gap create mode 100644 data/2021/neurips/An Image is Worth More Than a Thousand Words: Towards Disentanglement in The Wild create mode 100644 data/2021/neurips/An Improved Analysis and Rates for Variance Reduction under Without-replacement Sampling Orders create mode 100644 data/2021/neurips/An Improved Analysis of Gradient Tracking for Decentralized Machine Learning create mode 100644 data/2021/neurips/An Infinite-Feature Extension for Bayesian ReLU Nets That Fixes Their Asymptotic Overconfidence create mode 100644 data/2021/neurips/An Information-theoretic Approach to Distribution Shifts create mode 100644 data/2021/neurips/An Online Method for A Class of Distributionally Robust Optimization with Non-convex Objectives create mode 100644 data/2021/neurips/An Online Riemannian PCA for Stochastic Canonical Correlation Analysis create mode 100644 data/2021/neurips/An Uncertainty Principle is a Price of Privacy-Preserving Microdata create mode 100644 data/2021/neurips/An analysis of Ermakov-Zolotukhin quadrature using kernels create mode 100644 data/2021/neurips/An online passive-aggressive algorithm for difference-of-squares classification create mode 100644 data/2021/neurips/Analogous to Evolutionary Algorithm: Designing a Unified Sequence Model create mode 100644 data/2021/neurips/Analysis of Sensing Spectral for Signal Recovery under a Generalized Linear Model create mode 100644 data/2021/neurips/Analysis of one-hidden-layer neural networks via the resolvent method create mode 100644 data/2021/neurips/Analytic Insights into Structure and Rank of Neural Network Hessian Maps create mode 100644 data/2021/neurips/Analytical Study of Momentum-Based Acceleration Methods in Paradigmatic High-Dimensional Non-Convex Problems create mode 100644 data/2021/neurips/Analyzing the Confidentiality of Undistillable Teachers in Knowledge Distillation create mode 100644 data/2021/neurips/Analyzing the Generalization Capability of SGLD Using Properties of Gaussian Channels create mode 100644 data/2021/neurips/Answering Complex Causal Queries With the Maximum Causal Set Effect create mode 100644 data/2021/neurips/Anti-Backdoor Learning: Training Clean Models on Poisoned Data create mode 100644 data/2021/neurips/Antipodes of Label Differential Privacy: PATE and ALIBI create mode 100644 data/2021/neurips/Approximate Decomposable Submodular Function Minimization for Cardinality-Based Components create mode 100644 data/2021/neurips/Approximate optimization of convex functions with outlier noise create mode 100644 data/2021/neurips/Approximating the Permanent with Deep Rejection Sampling create mode 100644 data/2021/neurips/Arbitrary Conditional Distributions with Energy create mode 100644 data/2021/neurips/Are My Deep Learning Systems Fair? An Empirical Study of Fixed-Seed Training create mode 100644 data/2021/neurips/Are Transformers more robust than CNNs? create mode 100644 data/2021/neurips/Argmax Flows and Multinomial Diffusion: Learning Categorical Distributions create mode 100644 data/2021/neurips/Artistic Style Transfer with Internal-external Learning and Contrastive Learning create mode 100644 data/2021/neurips/Assessing Fairness in the Presence of Missing Data create mode 100644 data/2021/neurips/Associating Objects with Transformers for Video Object Segmentation create mode 100644 data/2021/neurips/Associative Memories via Predictive Coding create mode 100644 data/2021/neurips/Asymptotically Best Causal Effect Identification with Multi-Armed Bandits create mode 100644 data/2021/neurips/Asymptotically Exact Error Characterization of Offline Policy Evaluation with Misspecified Linear Models create mode 100644 data/2021/neurips/Asymptotics of representation learning in finite Bayesian neural networks create mode 100644 data/2021/neurips/Asymptotics of the Bootstrap via Stability with Applications to Inference with Model Selection create mode 100644 data/2021/neurips/Asynchronous Decentralized Online Learning create mode 100644 data/2021/neurips/Asynchronous Decentralized SGD with Quantized and Local Updates create mode 100644 data/2021/neurips/Asynchronous Stochastic Optimization Robust to Arbitrary Delays create mode 100644 data/2021/neurips/Attention Approximates Sparse Distributed Memory create mode 100644 data/2021/neurips/Attention Bottlenecks for Multimodal Fusion create mode 100644 data/2021/neurips/Attention over Learned Object Embeddings Enables Complex Visual Reasoning create mode 100644 data/2021/neurips/Auditing Black-Box Prediction Models for Data Minimization Compliance create mode 100644 data/2021/neurips/AugMax: Adversarial Composition of Random Augmentations for Robust Training create mode 100644 data/2021/neurips/Augmented Shortcuts for Vision Transformers create mode 100644 data/2021/neurips/Auto-Encoding Knowledge Graph for Unsupervised Medical Report Generation create mode 100644 data/2021/neurips/AutoBalance: Optimized Loss Functions for Imbalanced Data create mode 100644 data/2021/neurips/AutoGEL: An Automated Graph Neural Network with Explicit Link Information create mode 100644 data/2021/neurips/Autobahn: Automorphism-based Graph Neural Nets create mode 100644 data/2021/neurips/Autoformer: Decomposition Transformers with Auto-Correlation for Long-Term Series Forecasting create mode 100644 data/2021/neurips/Automated Discovery of Adaptive Attacks on Adversarial Defenses create mode 100644 data/2021/neurips/Automated Dynamic Mechanism Design create mode 100644 data/2021/neurips/Automatic Data Augmentation for Generalization in Reinforcement Learning create mode 100644 data/2021/neurips/Automatic Symmetry Discovery with Lie Algebra Convolutional Network create mode 100644 data/2021/neurips/Automatic Unsupervised Outlier Model Selection create mode 100644 data/2021/neurips/Automatic and Harmless Regularization with Constrained and Lexicographic Optimization: A Dynamic Barrier Approach create mode 100644 data/2021/neurips/Automorphic Equivalence-aware Graph Neural Network create mode 100644 data/2021/neurips/Autonomous Reinforcement Learning via Subgoal Curricula create mode 100644 data/2021/neurips/Average-Reward Learning and Planning with Options create mode 100644 data/2021/neurips/Averaging on the Bures-Wasserstein manifold: dimension-free convergence of gradient descent create mode 100644 data/2021/neurips/BARTScore: Evaluating Generated Text as Text Generation create mode 100644 data/2021/neurips/BAST: Bayesian Additive Regression Spanning Trees for Complex Constrained Domain create mode 100644 data/2021/neurips/BCD Nets: Scalable Variational Approaches for Bayesian Causal Discovery create mode 100644 "data/2021/neurips/BCORLE(\316\273): An Offline Reinforcement Learning and Evaluation Framework for Coupons Allocation in E-commerce Market" create mode 100644 data/2021/neurips/BNS: Building Network Structures Dynamically for Continual Learning create mode 100644 data/2021/neurips/Baby Intuitions Benchmark (BIB): Discerning the goals, preferences, and actions of others create mode 100644 data/2021/neurips/Backdoor Attack with Imperceptible Input and Latent Modification create mode 100644 data/2021/neurips/Backward-Compatible Prediction Updates: A Probabilistic Approach create mode 100644 data/2021/neurips/Balanced Chamfer Distance as a Comprehensive Metric for Point Cloud Completion create mode 100644 data/2021/neurips/Baleen: Robust Multi-Hop Reasoning at Scale via Condensed Retrieval create mode 100644 data/2021/neurips/Bandit Learning with Delayed Impact of Actions create mode 100644 data/2021/neurips/Bandit Phase Retrieval create mode 100644 data/2021/neurips/Bandit Quickest Changepoint Detection create mode 100644 data/2021/neurips/Bandits with Knapsacks beyond the Worst Case create mode 100644 data/2021/neurips/Bandits with many optimal arms create mode 100644 data/2021/neurips/Batch Active Learning at Scale create mode 100644 data/2021/neurips/Batch Multi-Fidelity Bayesian Optimization with Deep Auto-Regressive Networks create mode 100644 data/2021/neurips/Batch Normalization Orthogonalizes Representations in Deep Random Networks create mode 100644 data/2021/neurips/BatchQuant: Quantized-for-all Architecture Search with Robust Quantizer create mode 100644 data/2021/neurips/Batched Thompson Sampling create mode 100644 data/2021/neurips/BayesIMP: Uncertainty Quantification for Causal Data Fusion create mode 100644 data/2021/neurips/Bayesian Adaptation for Covariate Shift create mode 100644 data/2021/neurips/Bayesian Bellman Operators create mode 100644 data/2021/neurips/Bayesian Optimization of Function Networks create mode 100644 data/2021/neurips/Bayesian Optimization with High-Dimensional Outputs create mode 100644 data/2021/neurips/Bayesian decision-making under misspecified priors with applications to meta-learning create mode 100644 data/2021/neurips/Be Confident! Towards Trustworthy Graph Neural Networks via Confidence Calibration create mode 100644 data/2021/neurips/Behavior From the Void: Unsupervised Active Pre-Training create mode 100644 data/2021/neurips/Believe What You See: Implicit Constraint Approach for Offline Multi-Agent Reinforcement Learning create mode 100644 data/2021/neurips/Bellman Eluder Dimension: New Rich Classes of RL Problems, and Sample-Efficient Algorithms create mode 100644 data/2021/neurips/Bellman-consistent Pessimism for Offline Reinforcement Learning create mode 100644 data/2021/neurips/Beltrami Flow and Neural Diffusion on Graphs create mode 100644 data/2021/neurips/Benign Overfitting in Multiclass Classification: All Roads Lead to Interpolation create mode 100644 data/2021/neurips/BernNet: Learning Arbitrary Graph Spectral Filters via Bernstein Approximation create mode 100644 data/2021/neurips/Best Arm Identification in Contaminated Stochastic Bandits create mode 100644 data/2021/neurips/Best of Both Worlds: Practical and Theoretically Optimal Submodular Maximization in Parallel create mode 100644 data/2021/neurips/Best-case lower bounds in online learning create mode 100644 data/2021/neurips/Beta-CROWN: Efficient Bound Propagation with Per-neuron Split Constraints for Neural Network Robustness Verification create mode 100644 data/2021/neurips/Better Algorithms for Individually Fair k-Clustering create mode 100644 data/2021/neurips/Better Safe Than Sorry: Preventing Delusive Adversaries with Adversarial Training create mode 100644 data/2021/neurips/Beware of the Simulated DAG! Causal Discovery Benchmarks May Be Easy to Game create mode 100644 data/2021/neurips/Beyond Bandit Feedback in Online Multiclass Classification create mode 100644 data/2021/neurips/Beyond BatchNorm: Towards a Unified Understanding of Normalization in Deep Learning create mode 100644 data/2021/neurips/Beyond Pinball Loss: Quantile Methods for Calibrated Uncertainty Quantification create mode 100644 data/2021/neurips/Beyond Smoothness: Incorporating Low-Rank Analysis into Nonparametric Density Estimation create mode 100644 data/2021/neurips/Beyond Tikhonov: faster learning with self-concordant losses, via iterative regularization create mode 100644 data/2021/neurips/Beyond Value-Function Gaps: Improved Instance-Dependent Regret Bounds for Episodic Reinforcement Learning create mode 100644 data/2021/neurips/Beyond the Signs: Nonparametric Tensor Completion via Sign Series create mode 100644 data/2021/neurips/Bias Out-of-the-Box: An Empirical Analysis of Intersectional Occupational Biases in Popular Generative Language Models create mode 100644 data/2021/neurips/Bias and variance of the Bayesian-mean decoder create mode 100644 data/2021/neurips/Biological key-value memory networks create mode 100644 data/2021/neurips/Black Box Probabilistic Numerics create mode 100644 data/2021/neurips/BlendGAN: Implicitly GAN Blending for Arbitrary Stylized Face Generation create mode 100644 data/2021/neurips/Blending Anti-Aliasing into Vision Transformer create mode 100644 data/2021/neurips/BooVAE: Boosting Approach for Continual Learning of VAE create mode 100644 data/2021/neurips/BooVI: Provably Efficient Bootstrapped Value Iteration create mode 100644 data/2021/neurips/Boost Neural Networks by Checkpoints create mode 100644 data/2021/neurips/Boosted CVaR Classification create mode 100644 data/2021/neurips/Boosting with Multiple Sources create mode 100644 data/2021/neurips/Bootstrap Your Object Detector via Mixed Training create mode 100644 data/2021/neurips/Bootstrapping the Error of Oja's Algorithm create mode 100644 data/2021/neurips/Bounds all around: training energy-based models with bidirectional bounds create mode 100644 data/2021/neurips/Breaking the Dilemma of Medical Image-to-image Translation create mode 100644 data/2021/neurips/Breaking the Linear Iteration Cost Barrier for Some Well-known Conditional Gradient Methods Using MaxIP Data-structures create mode 100644 data/2021/neurips/Breaking the Moments Condition Barrier: No-Regret Algorithm for Bandits with Super Heavy-Tailed Payoffs create mode 100644 data/2021/neurips/Breaking the Sample Complexity Barrier to Regret-Optimal Model-Free Reinforcement Learning create mode 100644 data/2021/neurips/Breaking the centralized barrier for cross-device federated learning create mode 100644 data/2021/neurips/Brick-by-Brick: Combinatorial Construction with Deep Reinforcement Learning create mode 100644 data/2021/neurips/Bridging Explicit and Implicit Deep Generative Models via Neural Stein Estimators create mode 100644 data/2021/neurips/Bridging Non Co-occurrence with Unlabeled In-the-wild Data for Incremental Object Detection create mode 100644 data/2021/neurips/Bridging Offline Reinforcement Learning and Imitation Learning: A Tale of Pessimism create mode 100644 data/2021/neurips/Bridging the Gap Between Practice and PAC-Bayes Theory in Few-Shot Meta-Learning create mode 100644 data/2021/neurips/Bridging the Imitation Gap by Adaptive Insubordination create mode 100644 data/2021/neurips/Bubblewrap: Online tiling and real-time flow prediction on neural manifolds create mode 100644 data/2021/neurips/BulletTrain: Accelerating Robust Neural Network Training via Boundary Example Mining create mode 100644 data/2021/neurips/ByPE-VAE: Bayesian Pseudocoresets Exemplar VAE create mode 100644 data/2021/neurips/CAM-GAN: Continual Adaptation Modules for Generative Adversarial Networks create mode 100644 data/2021/neurips/CANITA: Faster Rates for Distributed Convex Optimization with Communication Compression create mode 100644 data/2021/neurips/CAPE: Encoding Relative Positions with Continuous Augmented Positional Embeddings create mode 100644 data/2021/neurips/CARMS: Categorical-Antithetic-REINFORCE Multi-Sample Gradient Estimator create mode 100644 data/2021/neurips/CATs: Cost Aggregation Transformers for Visual Correspondence create mode 100644 data/2021/neurips/CBP: backpropagation with constraint on weight precision using a pseudo-Lagrange multiplier method create mode 100644 data/2021/neurips/CCVS: Context-aware Controllable Video Synthesis create mode 100644 data/2021/neurips/CHIP: CHannel Independence-based Pruning for Compact Neural Networks create mode 100644 data/2021/neurips/CLDA: Contrastive Learning for Semi-Supervised Domain Adaptation create mode 100644 data/2021/neurips/CLIP-It! Language-Guided Video Summarization create mode 100644 data/2021/neurips/CO-PILOT: COllaborative Planning and reInforcement Learning On sub-Task curriculum create mode 100644 data/2021/neurips/COCO-LM: Correcting and Contrasting Text Sequences for Language Model Pretraining create mode 100644 data/2021/neurips/COHESIV: Contrastive Object and Hand Embedding Segmentation In Video create mode 100644 data/2021/neurips/COMBO: Conservative Offline Model-Based Policy Optimization create mode 100644 data/2021/neurips/CROCS: Clustering and Retrieval of Cardiac Signals Based on Patient Disease Class, Sex, and Age create mode 100644 data/2021/neurips/CSDI: Conditional Score-based Diffusion Models for Probabilistic Time Series Imputation create mode 100644 data/2021/neurips/Calibrating Predictions to Decisions: A Novel Approach to Multi-Class Calibration create mode 100644 data/2021/neurips/Calibration and Consistency of Adversarial Surrogate Losses create mode 100644 data/2021/neurips/Can Information Flows Suggest Targets for Interventions in Neural Circuits? create mode 100644 data/2021/neurips/Can Less be More? When Increasing-to-Balancing Label Noise Rates Considered Beneficial create mode 100644 data/2021/neurips/Can You Learn an Algorithm? Generalizing from Easy to Hard Problems with Recurrent Networks create mode 100644 data/2021/neurips/Can contrastive learning avoid shortcut solutions? create mode 100644 data/2021/neurips/Can fMRI reveal the representation of syntactic structure in the brain? create mode 100644 data/2021/neurips/Can multi-label classification networks know what they don't know? create mode 100644 data/2021/neurips/Can we globally optimize cross-validation loss? Quasiconvexity in ridge regression create mode 100644 data/2021/neurips/Can we have it all? On the Trade-off between Spatial and Adversarial Robustness of Neural Networks create mode 100644 data/2021/neurips/Canonical Capsules: Self-Supervised Capsules in Canonical Pose create mode 100644 data/2021/neurips/Capacity and Bias of Learned Geometric Embeddings for Directed Graphs create mode 100644 data/2021/neurips/Capturing implicit hierarchical structure in 3D biomedical images with self-supervised hyperbolic representations create mode 100644 data/2021/neurips/Cardinality constrained submodular maximization for random streams create mode 100644 data/2021/neurips/Cardinality-Regularized Hawkes-Granger Model create mode 100644 data/2021/neurips/Catalytic Role Of Noise And Necessity Of Inductive Biases In The Emergence Of Compositional Communication create mode 100644 data/2021/neurips/Catastrophic Data Leakage in Vertical Federated Learning create mode 100644 data/2021/neurips/Catch-A-Waveform: Learning to Generate Audio from a Single Short Example create mode 100644 data/2021/neurips/Causal Abstractions of Neural Networks create mode 100644 data/2021/neurips/Causal Bandits with Unknown Graph Structure create mode 100644 data/2021/neurips/Causal Effect Inference for Structured Treatments create mode 100644 data/2021/neurips/Causal Identification with Matrix Equations create mode 100644 data/2021/neurips/Causal Inference for Event Pairs in Multivariate Point Processes create mode 100644 data/2021/neurips/Causal Influence Detection for Improving Efficiency in Reinforcement Learning create mode 100644 data/2021/neurips/Causal Navigation by Continuous-time Neural Networks create mode 100644 data/2021/neurips/Causal-BALD: Deep Bayesian Active Learning of Outcomes to Infer Treatment-Effects from Observational Data create mode 100644 data/2021/neurips/Celebrating Diversity in Shared Multi-Agent Reinforcement Learning create mode 100644 data/2021/neurips/Center Smoothing: Certified Robustness for Networks with Structured Outputs create mode 100644 data/2021/neurips/CentripetalText: An Efficient Text Instance Representation for Scene Text Detection create mode 100644 data/2021/neurips/Certifying Robustness to Programmable Data Bias in Decision Trees create mode 100644 data/2021/neurips/Challenges and Opportunities in High Dimensional Variational Inference create mode 100644 data/2021/neurips/Change Point Detection via Multivariate Singular Spectrum Analysis create mode 100644 data/2021/neurips/Channel Permutations for N: M Sparsity create mode 100644 data/2021/neurips/Characterizing Generalization under Out-Of-Distribution Shifts in Deep Metric Learning create mode 100644 data/2021/neurips/Characterizing possible failure modes in physics-informed neural networks create mode 100644 data/2021/neurips/Characterizing the risk of fairwashing create mode 100644 data/2021/neurips/Charting and Navigating the Space of Solutions for Recurrent Neural Networks create mode 100644 data/2021/neurips/Chasing Sparsity in Vision Transformers: An End-to-End Exploration create mode 100644 data/2021/neurips/Chebyshev-Cantelli PAC-Bayes-Bennett Inequality for the Weighted Majority Vote create mode 100644 data/2021/neurips/Choose a Transformer: Fourier or Galerkin create mode 100644 data/2021/neurips/Circa: Stochastic ReLUs for Private Deep Learning create mode 100644 data/2021/neurips/Class-Disentanglement and Applications in Adversarial Detection and Defense create mode 100644 data/2021/neurips/Class-Incremental Learning via Dual Augmentation create mode 100644 data/2021/neurips/Class-agnostic Reconstruction of Dynamic Objects from Videos create mode 100644 data/2021/neurips/Clockwork Variational Autoencoders create mode 100644 data/2021/neurips/Closing the Gap: Tighter Analysis of Alternating Stochastic Gradient Methods for Bilevel Problems create mode 100644 data/2021/neurips/Closing the loop in medical decision support by understanding clinical decision-making: A case study on organ transplantation create mode 100644 data/2021/neurips/Clustering Effect of Adversarial Robust Models create mode 100644 data/2021/neurips/Co-Adaptation of Algorithmic and Implementational Innovations in Inference-based Deep Reinforcement Learning create mode 100644 data/2021/neurips/Co-evolution Transformer for Protein Contact Prediction create mode 100644 data/2021/neurips/CoAtNet: Marrying Convolution and Attention for All Data Sizes create mode 100644 data/2021/neurips/CoFiNet: Reliable Coarse-to-fine Correspondences for Robust PointCloud Registration create mode 100644 data/2021/neurips/CoFrNets: Interpretable Neural Architecture Inspired by Continued Fractions create mode 100644 data/2021/neurips/Coarse-to-fine Animal Pose and Shape Estimation create mode 100644 data/2021/neurips/Cockpit: A Practical Debugging Tool for the Training of Deep Neural Networks create mode 100644 data/2021/neurips/CogView: Mastering Text-to-Image Generation via Transformers create mode 100644 data/2021/neurips/Collaborating with Humans without Human Data create mode 100644 data/2021/neurips/Collaborative Causal Discovery with Atomic Interventions create mode 100644 data/2021/neurips/Collaborative Learning in the Jungle (Decentralized, Byzantine, Heterogeneous, Asynchronous and Nonconvex Learning) create mode 100644 data/2021/neurips/Collaborative Uncertainty in Multi-Agent Trajectory Forecasting create mode 100644 data/2021/neurips/Collapsed Variational Bounds for Bayesian Neural Networks create mode 100644 data/2021/neurips/Combating Noise: Semi-supervised Learning by Region Uncertainty Quantification create mode 100644 data/2021/neurips/Combinatorial Optimization for Panoptic Segmentation: A Fully Differentiable Approach create mode 100644 data/2021/neurips/Combinatorial Pure Exploration with Bottleneck Reward Function create mode 100644 data/2021/neurips/Combiner: Full Attention Transformer with Sparse Computation Cost create mode 100644 data/2021/neurips/Combining Human Predictions with Model Probabilities via Confusion Matrices and Calibration create mode 100644 data/2021/neurips/Combining Latent Space and Structured Kernels for Bayesian Optimization over Combinatorial Spaces create mode 100644 data/2021/neurips/Combining Recurrent, Convolutional, and Continuous-time Models with Linear State Space Layers create mode 100644 data/2021/neurips/Communication-efficient SGD: From Local SGD to One-Shot Averaging create mode 100644 data/2021/neurips/Compacter: Efficient Low-Rank Hypercomplex Adapter Layers create mode 100644 data/2021/neurips/Complexity Lower Bounds for Nonconvex-Strongly-Concave Min-Max Optimization create mode 100644 data/2021/neurips/Compositional Modeling of Nonlinear Dynamical Systems with ODE-based Random Features create mode 100644 data/2021/neurips/Compositional Reinforcement Learning from Logical Specifications create mode 100644 data/2021/neurips/Compositional Transformers for Scene Generation create mode 100644 data/2021/neurips/Comprehensive Knowledge Distillation with Causal Intervention create mode 100644 data/2021/neurips/Compressed Video Contrastive Learning create mode 100644 data/2021/neurips/Compressing Neural Networks: Towards Determining the Optimal Layer-wise Decomposition create mode 100644 data/2021/neurips/Compressive Visual Representations create mode 100644 data/2021/neurips/Computer-Aided Design as Language create mode 100644 data/2021/neurips/ConE: Cone Embeddings for Multi-Hop Reasoning over Knowledge Graphs create mode 100644 data/2021/neurips/Concentration inequalities under sub-Gaussian and sub-exponential conditions create mode 100644 data/2021/neurips/Conditional Generation Using Polynomial Expansions create mode 100644 data/2021/neurips/Conditionally Parameterized, Discretization-Aware Neural Networks for Mesh-Based Modeling of Physical Systems create mode 100644 data/2021/neurips/Conditioning Sparse Variational Gaussian Processes for Online Decision-making create mode 100644 data/2021/neurips/Confidence-Aware Imitation Learning from Demonstrations with Varying Optimality create mode 100644 data/2021/neurips/Confident Anchor-Induced Multi-Source Free Domain Adaptation create mode 100644 data/2021/neurips/Conflict-Averse Gradient Descent for Multi-task learning create mode 100644 data/2021/neurips/Conformal Bayesian Computation create mode 100644 data/2021/neurips/Conformal Prediction using Conditional Histograms create mode 100644 data/2021/neurips/Conformal Time-series Forecasting create mode 100644 data/2021/neurips/Conic Blackwell Algorithm: Parameter-Free Convex-Concave Saddle-Point Solving create mode 100644 data/2021/neurips/Conservative Data Sharing for Multi-Task Offline Reinforcement Learning create mode 100644 data/2021/neurips/Conservative Offline Distributional Reinforcement Learning create mode 100644 data/2021/neurips/Consistency Regularization for Variational Auto-Encoders create mode 100644 data/2021/neurips/Consistent Estimation for PCA and Sparse Regression with Oblivious Outliers create mode 100644 data/2021/neurips/Consistent Non-Parametric Methods for Maximizing Robustness create mode 100644 data/2021/neurips/Constrained Optimization to Train Neural Networks on Critical and Under-Represented Classes create mode 100644 data/2021/neurips/Constrained Robust Submodular Partitioning create mode 100644 data/2021/neurips/Container: Context Aggregation Networks create mode 100644 data/2021/neurips/Contextual Recommendations and Low-Regret Cutting-Plane Algorithms create mode 100644 data/2021/neurips/Contextual Similarity Aggregation with Self-attention for Visual Re-ranking create mode 100644 data/2021/neurips/Continual Auxiliary Task Learning create mode 100644 data/2021/neurips/Continual Learning via Local Module Composition create mode 100644 data/2021/neurips/Continual World: A Robotic Benchmark For Continual Reinforcement Learning create mode 100644 data/2021/neurips/Continuized Accelerations of Deterministic and Stochastic Gradient Descents, and of Gossip Algorithms create mode 100644 data/2021/neurips/Continuous Doubly Constrained Batch Reinforcement Learning create mode 100644 data/2021/neurips/Continuous Latent Process Flows create mode 100644 data/2021/neurips/Continuous Mean-Covariance Bandits create mode 100644 data/2021/neurips/Continuous vs. Discrete Optimization of Deep Neural Networks create mode 100644 data/2021/neurips/Continuous-time edge modelling using non-parametric point processes create mode 100644 data/2021/neurips/Contrast and Mix: Temporal Contrastive Video Domain Adaptation with Background Mixing create mode 100644 data/2021/neurips/Contrastive Active Inference create mode 100644 data/2021/neurips/Contrastive Graph Poisson Networks: Semi-Supervised Learning with Extremely Limited Labels create mode 100644 data/2021/neurips/Contrastive Laplacian Eigenmaps create mode 100644 data/2021/neurips/Contrastive Learning for Neural Topic Model create mode 100644 data/2021/neurips/Contrastive Learning of Global and Local Video Representations create mode 100644 data/2021/neurips/Contrastive Reinforcement Learning of Symbolic Reasoning Domains create mode 100644 data/2021/neurips/Contrastively Disentangled Sequential Variational Autoencoder create mode 100644 data/2021/neurips/Control Variates for Slate Off-Policy Evaluation create mode 100644 data/2021/neurips/Controlled Text Generation as Continuous Optimization with Multiple Constraints create mode 100644 data/2021/neurips/Controlling Neural Networks with Rule Representations create mode 100644 data/2021/neurips/Convergence Rates of Stochastic Gradient Descent under Infinite Noise Variance create mode 100644 data/2021/neurips/Convergence of adaptive algorithms for constrained weakly convex optimization create mode 100644 data/2021/neurips/Convex Polytope Trees and its Application to VAE create mode 100644 data/2021/neurips/Convex-Concave Min-Max Stackelberg Games create mode 100644 data/2021/neurips/Convolutional Normalization: Improving Deep Convolutional Network Robustness and Training create mode 100644 data/2021/neurips/Cooperative Stochastic Bandits with Asynchronous Agents and Constrained Feedback create mode 100644 data/2021/neurips/Coordinated Proximal Policy Optimization create mode 100644 data/2021/neurips/Coresets for Classification - Simplified and Strengthened create mode 100644 data/2021/neurips/Coresets for Clustering with Missing Values create mode 100644 data/2021/neurips/Coresets for Decision Trees of Signals create mode 100644 data/2021/neurips/Coresets for Time Series Clustering create mode 100644 data/2021/neurips/Correlated Stochastic Block Models: Exact Graph Matching with Applications to Recovering Communities create mode 100644 data/2021/neurips/Corruption Robust Active Learning create mode 100644 data/2021/neurips/CorticalFlow: A Diffeomorphic Mesh Transformer Network for Cortical Surface Reconstruction create mode 100644 data/2021/neurips/Cortico-cerebellar networks as decoupling neural interfaces create mode 100644 data/2021/neurips/Counterbalancing Learning and Strategic Incentives in Allocation Markets create mode 100644 data/2021/neurips/Counterexample Guided RL Policy Refinement Using Bayesian Optimization create mode 100644 data/2021/neurips/Counterfactual Explanations Can Be Manipulated create mode 100644 data/2021/neurips/Counterfactual Explanations in Sequential Decision Making Under Uncertainty create mode 100644 data/2021/neurips/Counterfactual Invariance to Spurious Correlations in Text Classification create mode 100644 data/2021/neurips/Counterfactual Maximum Likelihood Estimation for Training Deep Networks create mode 100644 data/2021/neurips/Coupled Gradient Estimators for Discrete Latent Variables create mode 100644 data/2021/neurips/Coupled Segmentation and Edge Learning via Dynamic Graph Propagation create mode 100644 data/2021/neurips/Covariance-Aware Private Mean Estimation Without Private Covariance Estimation create mode 100644 data/2021/neurips/Credal Self-Supervised Learning create mode 100644 data/2021/neurips/Credit Assignment Through Broadcasting a Global Error Vector create mode 100644 data/2021/neurips/Credit Assignment in Neural Networks through Deep Feedback Control create mode 100644 data/2021/neurips/Cross-modal Domain Adaptation for Cost-Efficient Visual Reinforcement Learning create mode 100644 data/2021/neurips/Cross-view Geo-localization with Layer-to-Layer Transformer create mode 100644 data/2021/neurips/CrypTen: Secure Multi-Party Computation Meets Machine Learning create mode 100644 data/2021/neurips/Curriculum Design for Teaching via Demonstrations: Theory and Applications create mode 100644 data/2021/neurips/Curriculum Disentangled Recommendation with Noisy Multi-feedback create mode 100644 data/2021/neurips/Curriculum Learning for Vision-and-Language Navigation create mode 100644 data/2021/neurips/Curriculum Offline Imitating Learning create mode 100644 data/2021/neurips/Cycle Self-Training for Domain Adaptation create mode 100644 data/2021/neurips/D2C: Diffusion-Decoding Models for Few-Shot Conditional Generation create mode 100644 data/2021/neurips/DECAF: Generating Fair Synthetic Data Using Causally-Aware Generative Networks create mode 100644 data/2021/neurips/DIB-R++: Learning to Predict Lighting and Material with a Hybrid Differentiable Renderer create mode 100644 data/2021/neurips/DNN-based Topology Optimisation: Spatial Invariance and Neural Tangent Kernel create mode 100644 data/2021/neurips/DOBF: A Deobfuscation Pre-Training Objective for Programming Languages create mode 100644 data/2021/neurips/DOCTOR: A Simple Method for Detecting Misclassification Errors create mode 100644 data/2021/neurips/DP-SSL: Towards Robust Semi-supervised Learning with A Few Labeled Samples create mode 100644 data/2021/neurips/DRIVE: One-bit Distributed Mean Estimation create mode 100644 data/2021/neurips/DROID-SLAM: Deep Visual SLAM for Monocular, Stereo, and RGB-D Cameras create mode 100644 data/2021/neurips/DRONE: Data-aware Low-rank Compression for Large NLP Models create mode 100644 data/2021/neurips/DSelect-k: Differentiable Selection in the Mixture of Experts with Applications to Multi-Task Learning create mode 100644 data/2021/neurips/Damped Anderson Mixing for Deep Reinforcement Learning: Acceleration, Convergence, and Stabilization create mode 100644 data/2021/neurips/Dangers of Bayesian Model Averaging under Covariate Shift create mode 100644 data/2021/neurips/Data Augmentation Can Improve Robustness create mode 100644 data/2021/neurips/Data Sharing and Compression for Cooperative Networked Control create mode 100644 data/2021/neurips/Data driven semi-supervised learning create mode 100644 data/2021/neurips/Data-Efficient GAN Training Beyond (Just) Augmentations: A Lottery Ticket Perspective create mode 100644 data/2021/neurips/Data-Efficient Instance Generation from Instance Discrimination create mode 100644 data/2021/neurips/Dataset Distillation with Infinitely Wide Convolutional Networks create mode 100644 data/2021/neurips/De-randomizing MCMC dynamics with the diffusion Stein operator create mode 100644 data/2021/neurips/Dealing With Misspecification In Fixed-Confidence Linear Top-m Identification create mode 100644 data/2021/neurips/Debiased Visual Question Answering from Feature and Sample Perspectives create mode 100644 data/2021/neurips/Deceive D: Adaptive Pseudo Augmentation for GAN Training with Limited Data create mode 100644 data/2021/neurips/Decentralized Learning in Online Queuing Systems create mode 100644 data/2021/neurips/Decentralized Q-learning in Zero-sum Markov Games create mode 100644 data/2021/neurips/Decision Transformer: Reinforcement Learning via Sequence Modeling create mode 100644 data/2021/neurips/Deconditional Downscaling with Gaussian Processes create mode 100644 data/2021/neurips/Deconvolutional Networks on Graph Data create mode 100644 data/2021/neurips/Decoupling the Depth and Scope of Graph Neural Networks create mode 100644 data/2021/neurips/Decrypting Cryptic Crosswords: Semantically Complex Wordplay Puzzles as a Target for NLP create mode 100644 data/2021/neurips/Deep Bandits Show-Off: Simple and Efficient Exploration with Deep Networks create mode 100644 data/2021/neurips/Deep Conditional Gaussian Mixture Model for Constrained Clustering create mode 100644 data/2021/neurips/Deep Contextual Video Compression create mode 100644 data/2021/neurips/Deep Explicit Duration Switching Models for Time Series create mode 100644 data/2021/neurips/Deep Extended Hazard Models for Survival Analysis create mode 100644 data/2021/neurips/Deep Extrapolation for Attribute-Enhanced Generation create mode 100644 data/2021/neurips/Deep Jump Learning for Off-Policy Evaluation in Continuous Treatment Settings create mode 100644 data/2021/neurips/Deep Learning Through the Lens of Example Difficulty create mode 100644 data/2021/neurips/Deep Learning on a Data Diet: Finding Important Examples Early in Training create mode 100644 data/2021/neurips/Deep Learning with Label Differential Privacy create mode 100644 data/2021/neurips/Deep Marching Tetrahedra: a Hybrid Representation for High-Resolution 3D Shape Synthesis create mode 100644 data/2021/neurips/Deep Markov Factor Analysis: Towards Concurrent Temporal and Spatial Analysis of fMRI Data create mode 100644 data/2021/neurips/Deep Molecular Representation Learning via Fusing Physical and Chemical Information create mode 100644 data/2021/neurips/Deep Networks Provably Classify Data on Curves create mode 100644 data/2021/neurips/Deep Neural Networks as Point Estimates for Deep Gaussian Processes create mode 100644 data/2021/neurips/Deep Proxy Causal Learning and its Application to Confounded Bandit Policy Evaluation create mode 100644 data/2021/neurips/Deep Reinforcement Learning at the Edge of the Statistical Precipice create mode 100644 data/2021/neurips/Deep Residual Learning in Spiking Neural Networks create mode 100644 data/2021/neurips/Deep Self-Dissimilarities as Powerful Visual Fingerprints create mode 100644 data/2021/neurips/Deep Synoptic Monte-Carlo Planning in Reconnaissance Blind Chess create mode 100644 data/2021/neurips/Deep inference of latent dynamics with spatio-temporal super-resolution using selective backpropagation through time create mode 100644 data/2021/neurips/Deep learning is adaptive to intrinsic dimensionality of model smoothness in anisotropic Besov space create mode 100644 data/2021/neurips/DeepGEM: Generalized Expectation-Maximization for Blind Inversion create mode 100644 data/2021/neurips/DeepReduce: A Sparse-tensor Communication Framework for Federated Deep Learning create mode 100644 data/2021/neurips/DeepSITH: Efficient Learning via Decomposition of What and When Across Time Scales create mode 100644 data/2021/neurips/Deeply Shared Filter Bases for Parameter-Efficient Convolutional Neural Networks create mode 100644 data/2021/neurips/Deformable Butterfly: A Highly Structured and Sparse Linear Transform create mode 100644 data/2021/neurips/Delayed Gradient Averaging: Tolerate the Communication Latency for Federated Learning create mode 100644 data/2021/neurips/Delayed Propagation Transformer: A Universal Computation Engine towards Practical Control in Cyber-Physical Systems create mode 100644 data/2021/neurips/Demystifying and Generalizing BinaryConnect create mode 100644 data/2021/neurips/Denoising Normalizing Flow create mode 100644 data/2021/neurips/Dense Keypoints via Multiview Supervision create mode 100644 data/2021/neurips/Dense Unsupervised Learning for Video Segmentation create mode 100644 data/2021/neurips/Densely connected normalizing flows create mode 100644 data/2021/neurips/Derivative-Free Policy Optimization for Linear Risk-Sensitive and Robust Control Design: Implicit Regularization and Sample Complexity create mode 100644 data/2021/neurips/Design of Experiments for Stochastic Contextual Linear Bandits create mode 100644 data/2021/neurips/Designing Counterfactual Generators using Deep Model Inversion create mode 100644 data/2021/neurips/Detecting Anomalous Event Sequences with Temporal Point Processes create mode 100644 data/2021/neurips/Detecting Errors and Estimating Accuracy on Unlabeled Data with Self-training Ensembles create mode 100644 data/2021/neurips/Detecting Individual Decision-Making Style: Exploring Behavioral Stylometry in Chess create mode 100644 data/2021/neurips/Detecting Moments and Highlights in Videos via Natural Language Queries create mode 100644 data/2021/neurips/Detecting and Adapting to Irregular Distribution Shifts in Bayesian Online Learning create mode 100644 data/2021/neurips/Determinantal point processes based on orthogonal polynomials for sampling minibatches in SGD create mode 100644 data/2021/neurips/DiBS: Differentiable Bayesian Structure Learning create mode 100644 data/2021/neurips/Differentiable Annealed Importance Sampling and the Perils of Gradient Noise create mode 100644 data/2021/neurips/Differentiable Equilibrium Computation with Decision Diagrams for Stackelberg Models of Combinatorial Congestion Games create mode 100644 data/2021/neurips/Differentiable Learning Under Triage create mode 100644 data/2021/neurips/Differentiable Multiple Shooting Layers create mode 100644 data/2021/neurips/Differentiable Optimization of Generalized Nondecomposable Functions using Linear Programs create mode 100644 data/2021/neurips/Differentiable Quality Diversity create mode 100644 data/2021/neurips/Differentiable Simulation of Soft Multi-body Systems create mode 100644 data/2021/neurips/Differentiable Spike: Rethinking Gradient-Descent for Training Spiking Neural Networks create mode 100644 data/2021/neurips/Differentiable Spline Approximations create mode 100644 data/2021/neurips/Differentiable Synthesis of Program Architectures create mode 100644 data/2021/neurips/Differentiable Unsupervised Feature Selection based on a Gated Laplacian create mode 100644 data/2021/neurips/Differentiable rendering with perturbed optimizers create mode 100644 data/2021/neurips/Differential Privacy Dynamics of Langevin Diffusion and Noisy Gradient Descent create mode 100644 data/2021/neurips/Differential Privacy Over Riemannian Manifolds create mode 100644 data/2021/neurips/Differentially Private Empirical Risk Minimization under the Fairness Lens create mode 100644 data/2021/neurips/Differentially Private Federated Bayesian Optimization with Distributed Exploration create mode 100644 data/2021/neurips/Differentially Private Learning with Adaptive Clipping create mode 100644 data/2021/neurips/Differentially Private Model Personalization create mode 100644 data/2021/neurips/Differentially Private Multi-Armed Bandits in the Shuffle Model create mode 100644 data/2021/neurips/Differentially Private Sampling from Distributions create mode 100644 data/2021/neurips/Differentially Private Stochastic Optimization: New Results in Convex and Non-Convex Settings create mode 100644 data/2021/neurips/Differentially Private n-gram Extraction create mode 100644 data/2021/neurips/Diffusion Models Beat GANs on Image Synthesis create mode 100644 data/2021/neurips/Diffusion Normalizing Flow create mode 100644 "data/2021/neurips/Diffusion Schr\303\266dinger Bridge with Applications to Score-Based Generative Modeling" create mode 100644 data/2021/neurips/Dimension-free empirical entropy estimation create mode 100644 data/2021/neurips/Dimensionality Reduction for Wasserstein Barycenter create mode 100644 data/2021/neurips/Direct Multi-view Multi-person 3D Pose Estimation create mode 100644 data/2021/neurips/Directed Graph Contrastive Learning create mode 100644 data/2021/neurips/Directed Probabilistic Watershed create mode 100644 data/2021/neurips/Directed Spectrum Measures Improve Latent Network Models Of Neural Populations create mode 100644 data/2021/neurips/Directional Message Passing on Molecular Graphs via Synthetic Coordinates create mode 100644 data/2021/neurips/Dirichlet Energy Constrained Learning for Deep Graph Neural Networks create mode 100644 data/2021/neurips/Discerning Decision-Making Process of Deep Neural Networks with Hierarchical Voting Transformation create mode 100644 data/2021/neurips/Discovering Dynamic Salient Regions for Spatio-Temporal Graph Neural Networks create mode 100644 data/2021/neurips/Discovering and Achieving Goals via World Models create mode 100644 data/2021/neurips/Discovery of Options via Meta-Learned Subgoals create mode 100644 data/2021/neurips/Discrete-Valued Neural Communication create mode 100644 data/2021/neurips/Disentangled Contrastive Learning on Graphs create mode 100644 data/2021/neurips/Disentangling Identifiable Features from Noisy Data with Structured Nonlinear ICA create mode 100644 data/2021/neurips/Disentangling the Roles of Curation, Data-Augmentation and the Prior in the Cold Posterior Effect create mode 100644 data/2021/neurips/Disrupting Deep Uncertainty Estimation Without Harming Accuracy create mode 100644 data/2021/neurips/Dissecting the Diffusion Process in Linear Graph Convolutional Networks create mode 100644 data/2021/neurips/Distilling Image Classifiers in Object Detectors create mode 100644 data/2021/neurips/Distilling Meta Knowledge on Heterogeneous Graph for Illicit Drug Trafficker Detection on Social Media create mode 100644 data/2021/neurips/Distilling Object Detectors with Feature Richness create mode 100644 data/2021/neurips/Distilling Robust and Non-Robust Features in Adversarial Examples by Information Bottleneck create mode 100644 data/2021/neurips/Distributed Deep Learning In Open Collaborations create mode 100644 data/2021/neurips/Distributed Estimation with Multiple Samples per User: Sharp Rates and Phase Transition create mode 100644 data/2021/neurips/Distributed Machine Learning with Sparse Heterogeneous Data create mode 100644 data/2021/neurips/Distributed Principal Component Analysis with Limited Communication create mode 100644 data/2021/neurips/Distributed Saddle-Point Problems Under Data Similarity create mode 100644 data/2021/neurips/Distributed Zero-Order Optimization under Adversarial Noise create mode 100644 data/2021/neurips/Distribution-free inference for regression: discrete, continuous, and in between create mode 100644 data/2021/neurips/Distributional Gradient Matching for Learning Uncertain Neural Dynamics Models create mode 100644 data/2021/neurips/Distributional Reinforcement Learning for Multi-Dimensional Reward Functions create mode 100644 data/2021/neurips/Distributionally Robust Imitation Learning create mode 100644 data/2021/neurips/Divergence Frontiers for Generative Models: Sample Complexity, Quantization Effects, and Frontier Integrals create mode 100644 data/2021/neurips/Diverse Message Passing for Attribute with Heterophily create mode 100644 data/2021/neurips/Diversity Enhanced Active Learning with Strictly Proper Scoring Rules create mode 100644 data/2021/neurips/Diversity Matters When Learning From Ensembles create mode 100644 data/2021/neurips/Do Different Tracking Tasks Require Different Appearance Models? create mode 100644 data/2021/neurips/Do Input Gradients Highlight Discriminative Features? create mode 100644 data/2021/neurips/Do Neural Optimal Transport Solvers Work? A Continuous Wasserstein-2 Benchmark create mode 100644 data/2021/neurips/Do Transformers Really Perform Badly for Graph Representation? create mode 100644 data/2021/neurips/Do Vision Transformers See Like Convolutional Neural Networks? create mode 100644 data/2021/neurips/Do Wider Neural Networks Really Help Adversarial Robustness? create mode 100644 data/2021/neurips/Does Knowledge Distillation Really Work? create mode 100644 data/2021/neurips/Does Preprocessing Help Training Over-parameterized Neural Networks? create mode 100644 data/2021/neurips/Does enforcing fairness mitigate biases caused by subpopulation shift? create mode 100644 data/2021/neurips/Domain Adaptation with Invariant Representation Learning: What Transformations to Learn? create mode 100644 data/2021/neurips/Domain Invariant Representation Learning with Domain Density Transformations create mode 100644 data/2021/neurips/DominoSearch: Find layer-wise fine-grained N: M sparse schemes from dense neural networks create mode 100644 data/2021/neurips/Don't Generate Me: Training Differentially Private Generative Models with Sinkhorn Divergence create mode 100644 data/2021/neurips/Double Debiased Machine Learning for Dynamic Treatment Effects create mode 100644 data/2021/neurips/Double Machine Learning Density Estimation for Local Treatment Effects with Instruments create mode 100644 data/2021/neurips/Doubly Robust Thompson Sampling with Linear Payoffs create mode 100644 data/2021/neurips/Dr Jekyll & Mr Hyde: the strange case of off-policy policy updates create mode 100644 data/2021/neurips/Drawing Robust Scratch Tickets: Subnetworks with Inborn Robustness Are Found within Randomly Initialized Networks create mode 100644 data/2021/neurips/Drop, Swap, and Generate: A Self-Supervised Approach for Generating Neural Activity create mode 100644 data/2021/neurips/Drop-DTW: Aligning Common Signal Between Sequences While Dropping Outliers create mode 100644 data/2021/neurips/DropGNN: Random Dropouts Increase the Expressiveness of Graph Neural Networks create mode 100644 data/2021/neurips/Dual Adaptivity: A Universal Algorithm for Minimizing the Adaptive Regret of Convex Functions create mode 100644 data/2021/neurips/Dual Parameterization of Sparse Variational Gaussian Processes create mode 100644 data/2021/neurips/Dual Progressive Prototype Network for Generalized Zero-Shot Learning create mode 100644 data/2021/neurips/Dual-stream Network for Visual Recognition create mode 100644 data/2021/neurips/DualNet: Continual Learning, Fast and Slow create mode 100644 data/2021/neurips/Dueling Bandits with Adversarial Sleeping create mode 100644 data/2021/neurips/Dueling Bandits with Team Comparisons create mode 100644 data/2021/neurips/Duplex Sequence-to-Sequence Learning for Reversible Machine Translation create mode 100644 data/2021/neurips/Dynaboard: An Evaluation-As-A-Service Platform for Holistic Next-Generation Benchmarking create mode 100644 data/2021/neurips/Dynamic Analysis of Higher-Order Coordination in Neuronal Assemblies via De-Sparsified Orthogonal Matching Pursuit create mode 100644 data/2021/neurips/Dynamic Bottleneck for Robust Self-Supervised Exploration create mode 100644 data/2021/neurips/Dynamic COVID risk assessment accounting for community virus exposure from a spatial-temporal transmission model create mode 100644 data/2021/neurips/Dynamic Causal Bayesian Optimization create mode 100644 data/2021/neurips/Dynamic Distillation Network for Cross-Domain Few-Shot Recognition with Unlabeled Data create mode 100644 data/2021/neurips/Dynamic Grained Encoder for Vision Transformers create mode 100644 data/2021/neurips/Dynamic Inference with Neural Interpreters create mode 100644 data/2021/neurips/Dynamic Neural Representational Decoders for High-Resolution Semantic Segmentation create mode 100644 data/2021/neurips/Dynamic Normalization and Relay for Video Action Recognition create mode 100644 data/2021/neurips/Dynamic Resolution Network create mode 100644 data/2021/neurips/Dynamic Sasvi: Strong Safe Screening for Norm-Regularized Least Squares create mode 100644 data/2021/neurips/Dynamic Trace Estimation create mode 100644 data/2021/neurips/Dynamic Visual Reasoning by Learning Differentiable Physics Models from Video and Language create mode 100644 data/2021/neurips/Dynamic influence maximization create mode 100644 data/2021/neurips/Dynamic population-based meta-learning for multi-agent communication with natural language create mode 100644 data/2021/neurips/DynamicViT: Efficient Vision Transformers with Dynamic Token Sparsification create mode 100644 data/2021/neurips/Dynamical Wasserstein Barycenters for Time-series Modeling create mode 100644 data/2021/neurips/Dynamics of Stochastic Momentum Methods on Large-scale, Quadratic Models create mode 100644 data/2021/neurips/Dynamics-regulated kinematic policy for egocentric pose estimation create mode 100644 data/2021/neurips/E(n) Equivariant Normalizing Flows create mode 100644 data/2021/neurips/EDGE: Explaining Deep Reinforcement Learning Policies create mode 100644 data/2021/neurips/EF21: A New, Simpler, Theoretically Better, and Practically Faster Error Feedback create mode 100644 data/2021/neurips/EIGNN: Efficient Infinite-Depth Graph Neural Networks create mode 100644 data/2021/neurips/ELLA: Exploration through Learned Language Abstraction create mode 100644 data/2021/neurips/Early Convolutions Help Transformers See Better create mode 100644 data/2021/neurips/Early-stopped neural networks are consistent create mode 100644 data/2021/neurips/Edge Representation Learning with Hypergraphs create mode 100644 data/2021/neurips/EditGAN: High-Precision Semantic Image Editing create mode 100644 data/2021/neurips/Editing a classifier by rewriting its prediction rules create mode 100644 data/2021/neurips/Effective Meta-Regularization by Kernelized Proximal Regularization create mode 100644 data/2021/neurips/Efficient Active Learning for Gaussian Process Classification by Error Reduction create mode 100644 data/2021/neurips/Efficient Algorithms for Learning Depth-2 Neural Networks with General ReLU Activations create mode 100644 data/2021/neurips/Efficient Bayesian network structure learning via local Markov boundary search create mode 100644 data/2021/neurips/Efficient Combination of Rematerialization and Offloading for Training DNNs create mode 100644 data/2021/neurips/Efficient Equivariant Network create mode 100644 data/2021/neurips/Efficient First-Order Contextual Bandits: Prediction, Allocation, and Triangular Discrimination create mode 100644 data/2021/neurips/Efficient Generalization with Distributionally Robust Learning create mode 100644 data/2021/neurips/Efficient Learning of Discrete-Continuous Computation Graphs create mode 100644 data/2021/neurips/Efficient Mirror Descent Ascent Methods for Nonsmooth Minimax Problems create mode 100644 data/2021/neurips/Efficient Neural Network Training via Forward and Backward Propagation Sparsification create mode 100644 data/2021/neurips/Efficient Online Estimation of Causal Effects by Deciding What to Observe create mode 100644 data/2021/neurips/Efficient Statistical Assessment of Neural Network Corruption Robustness create mode 100644 data/2021/neurips/Efficient Training of Retrieval Models using Negative Cache create mode 100644 data/2021/neurips/Efficient Training of Visual Transformers with Small Datasets create mode 100644 data/2021/neurips/Efficient Truncated Linear Regression with Unknown Noise Variance create mode 100644 data/2021/neurips/Efficient and Accurate Gradients for Neural SDEs create mode 100644 data/2021/neurips/Efficient and Local Parallel Random Walks create mode 100644 data/2021/neurips/Efficient constrained sampling via the mirror-Langevin algorithm create mode 100644 data/2021/neurips/Efficient hierarchical Bayesian inference for spatio-temporal regression models in neuroimaging create mode 100644 data/2021/neurips/Efficient methods for Gaussian Markov random fields under sparse linear constraints create mode 100644 data/2021/neurips/Efficiently Identifying Task Groupings for Multi-Task Learning create mode 100644 data/2021/neurips/Efficiently Learning One Hidden Layer ReLU Networks From Queries create mode 100644 data/2021/neurips/Embedding Principle of Loss Landscape of Deep Neural Networks create mode 100644 data/2021/neurips/Emergent Communication of Generalizations create mode 100644 data/2021/neurips/Emergent Communication under Varying Sizes and Connectivities create mode 100644 data/2021/neurips/Emergent Discrete Communication in Semantic Spaces create mode 100644 data/2021/neurips/Enabling Fast Differentially Private SGD via Just-in-Time Compilation and Vectorization create mode 100644 data/2021/neurips/Encoding Robustness to Image Style via Adversarial Feature Perturbations create mode 100644 data/2021/neurips/Encoding Spatial Distribution of Convolutional Features for Texture Representation create mode 100644 data/2021/neurips/End-to-End Training of Multi-Document Reader and Retriever for Open-Domain Question Answering create mode 100644 data/2021/neurips/End-to-End Weak Supervision create mode 100644 data/2021/neurips/End-to-end Multi-modal Video Temporal Grounding create mode 100644 data/2021/neurips/End-to-end reconstruction meets data-driven regularization for inverse problems create mode 100644 data/2021/neurips/Ensembling Graph Predictions for AMR Parsing create mode 100644 data/2021/neurips/Entropic Desired Dynamics for Intrinsic Control create mode 100644 data/2021/neurips/Entropy-based adaptive Hamiltonian Monte Carlo create mode 100644 data/2021/neurips/Environment Generation for Zero-Shot Compositional Reinforcement Learning create mode 100644 data/2021/neurips/Episodic Multi-agent Reinforcement Learning with Curiosity-driven Exploration create mode 100644 data/2021/neurips/Equilibrium Refinement for the Age of Machines: The One-Sided Quasi-Perfect Equilibrium create mode 100644 data/2021/neurips/Equilibrium and non-Equilibrium regimes in the learning of Restricted Boltzmann Machines create mode 100644 data/2021/neurips/Equivariant Manifold Flows create mode 100644 data/2021/neurips/Error Compensated Distributed SGD Can Be Accelerated create mode 100644 data/2021/neurips/ErrorCompensatedX: error compensation for variance reduced algorithms create mode 100644 data/2021/neurips/Escape saddle points by a simple gradient-descent based algorithm create mode 100644 data/2021/neurips/Escaping Saddle Points with Compressed SGD create mode 100644 data/2021/neurips/Estimating High Order Gradients of the Data Distribution by Denoising create mode 100644 data/2021/neurips/Estimating Multi-cause Treatment Effects via Single-cause Perturbation create mode 100644 data/2021/neurips/Estimating the Long-Term Effects of Novel Treatments create mode 100644 data/2021/neurips/Estimating the Unique Information of Continuous Variables create mode 100644 data/2021/neurips/Evaluating Efficient Performance Estimators of Neural Architectures create mode 100644 data/2021/neurips/Evaluating Gradient Inversion Attacks and Defenses in Federated Learning create mode 100644 data/2021/neurips/Evaluating State-of-the-Art Classification Models Against Bayes Optimality create mode 100644 data/2021/neurips/Evaluating model performance under worst-case subpopulations create mode 100644 data/2021/neurips/Evaluation of Human-AI Teams for Learned and Rule-Based Agents in Hanabi create mode 100644 data/2021/neurips/Even your Teacher Needs Guidance: Ground-Truth Targets Dampen Regularization Imposed by Self-Distillation create mode 100644 data/2021/neurips/Evidential Softmax for Sparse Multimodal Distributions in Deep Generative Models create mode 100644 data/2021/neurips/EvoGrad: Efficient Gradient-Based Meta-Learning and Hyperparameter Optimization create mode 100644 data/2021/neurips/Evolution Gym: A Large-Scale Benchmark for Evolving Soft Robots create mode 100644 data/2021/neurips/Exact Privacy Guarantees for Markov Chain Implementations of the Exponential Mechanism with Artificial Atoms create mode 100644 data/2021/neurips/Exact marginal prior distributions of finite Bayesian neural networks create mode 100644 data/2021/neurips/Excess Capacity and Backdoor Poisoning create mode 100644 data/2021/neurips/Explainable Semantic Space by Grounding Language to Vision with Cross-Modal Contrastive Learning create mode 100644 data/2021/neurips/Explaining Hyperparameter Optimization via Partial Dependence Plots create mode 100644 data/2021/neurips/Explaining Latent Representations with a Corpus of Examples create mode 100644 data/2021/neurips/Explaining heterogeneity in medial entorhinal cortex with task-driven neural networks create mode 100644 data/2021/neurips/Explanation-based Data Augmentation for Image Classification create mode 100644 data/2021/neurips/Explicable Reward Design for Reinforcement Learning Agents create mode 100644 data/2021/neurips/Explicit loss asymptotics in the gradient descent training of neural networks create mode 100644 data/2021/neurips/Exploiting Chain Rule and Bayes' Theorem to Compare Probability Distributions create mode 100644 data/2021/neurips/Exploiting Data Sparsity in Secure Cross-Platform Social Recommendation create mode 100644 data/2021/neurips/Exploiting Domain-Specific Features to Enhance Domain Generalization create mode 100644 data/2021/neurips/Exploiting Local Convergence of Quasi-Newton Methods Globally: Adaptive Sample Size Approach create mode 100644 data/2021/neurips/Exploiting Opponents Under Utility Constraints in Sequential Games create mode 100644 data/2021/neurips/Exploiting a Zoo of Checkpoints for Unseen Tasks create mode 100644 data/2021/neurips/Exploiting the Intrinsic Neighborhood Structure for Source-free Domain Adaptation create mode 100644 data/2021/neurips/Exploration-Exploitation in Multi-Agent Competition: Convergence with Bounded Rationality create mode 100644 data/2021/neurips/Exploring Architectural Ingredients of Adversarially Robust Deep Neural Networks create mode 100644 data/2021/neurips/Exploring Cross-Video and Cross-Modality Signals for Weakly-Supervised Audio-Visual Video Parsing create mode 100644 data/2021/neurips/Exploring Forensic Dental Identification with Deep Learning create mode 100644 data/2021/neurips/Exploring Social Posterior Collapse in Variational Autoencoder for Interaction Modeling create mode 100644 data/2021/neurips/Exploring the Limits of Out-of-Distribution Detection create mode 100644 data/2021/neurips/Exponential Bellman Equation and Improved Regret Bounds for Risk-Sensitive Reinforcement Learning create mode 100644 data/2021/neurips/Exponential Graph is Provably Efficient for Decentralized Deep Training create mode 100644 data/2021/neurips/Exponential Separation between Two Learning Models and Adversarial Robustness create mode 100644 data/2021/neurips/Extending Lagrangian and Hamiltonian Neural Networks with Differentiable Contact Models create mode 100644 data/2021/neurips/Extracting Deformation-Aware Local Features by Learning to Deform create mode 100644 data/2021/neurips/FACMAC: Factored Multi-Agent Centralised Policy Gradients create mode 100644 data/2021/neurips/FINE Samples for Learning with Noisy Labels create mode 100644 data/2021/neurips/FL-WBC: Enhancing Robustness against Model Poisoning Attacks in Federated Learning from a Client Perspective create mode 100644 data/2021/neurips/FLEX: Unifying Evaluation for Few-Shot NLP create mode 100644 data/2021/neurips/FMMformer: Efficient and Flexible Transformer via Decomposed Near-field and Far-field Attention create mode 100644 data/2021/neurips/Factored Policy Gradients: Leveraging Structure for Efficient Learning in MOMDPs create mode 100644 data/2021/neurips/Fair Algorithms for Multi-Agent Multi-Armed Bandits create mode 100644 data/2021/neurips/Fair Classification with Adversarial Perturbations create mode 100644 data/2021/neurips/Fair Clustering Under a Bounded Cost create mode 100644 data/2021/neurips/Fair Exploration via Axiomatic Bargaining create mode 100644 data/2021/neurips/Fair Scheduling for Time-dependent Resources create mode 100644 data/2021/neurips/Fair Sequential Selection Using Supervised Learning Models create mode 100644 data/2021/neurips/Fair Sortition Made Transparent create mode 100644 data/2021/neurips/Fair Sparse Regression with Clustering: An Invex Relaxation for a Combinatorial Problem create mode 100644 data/2021/neurips/Fairness in Ranking under Uncertainty create mode 100644 data/2021/neurips/Fairness via Representation Neutralization create mode 100644 data/2021/neurips/Fast Abductive Learning by Similarity-based Consistency Optimization create mode 100644 data/2021/neurips/Fast Approximate Dynamic Programming for Infinite-Horizon Markov Decision Processes create mode 100644 data/2021/neurips/Fast Approximation of the Sliced-Wasserstein Distance Using Concentration of Random Projections create mode 100644 data/2021/neurips/Fast Axiomatic Attribution for Neural Networks create mode 100644 data/2021/neurips/Fast Bayesian Inference for Gaussian Cox Processes via Path Integral Formulation create mode 100644 data/2021/neurips/Fast Certified Robust Training with Short Warmup create mode 100644 data/2021/neurips/Fast Doubly-Adaptive MCMC to Estimate the Gibbs Partition Function with Weak Mixing Time Bounds create mode 100644 data/2021/neurips/Fast Extra Gradient Methods for Smooth Structured Nonconvex-Nonconcave Minimax Problems create mode 100644 data/2021/neurips/Fast Federated Learning in the Presence of Arbitrary Device Unavailability create mode 100644 data/2021/neurips/Fast Minimum-norm Adversarial Attacks through Adaptive Norm Constraints create mode 100644 data/2021/neurips/Fast Multi-Resolution Transformer Fine-tuning for Extreme Multi-label Text Classification create mode 100644 data/2021/neurips/Fast Policy Extragradient Methods for Competitive Games with Entropy Regularization create mode 100644 data/2021/neurips/Fast Projection onto the Capped Simplex with Applications to Sparse Regression in Bioinformatics create mode 100644 data/2021/neurips/Fast Pure Exploration via Frank-Wolfe create mode 100644 data/2021/neurips/Fast Routing under Uncertainty: Adaptive Learning in Congestion Games via Exponential Weights create mode 100644 data/2021/neurips/Fast Training Method for Stochastic Compositional Optimization Problems create mode 100644 data/2021/neurips/Fast Training of Neural Lumigraph Representations using Meta Learning create mode 100644 data/2021/neurips/Fast Tucker Rank Reduction for Non-Negative Tensors Using Mean-Field Approximation create mode 100644 data/2021/neurips/Fast and Memory Efficient Differentially Private-SGD via JL Projections create mode 100644 data/2021/neurips/Fast and accurate randomized algorithms for low-rank tensor decompositions create mode 100644 data/2021/neurips/Fast rates for prediction with limited expert advice create mode 100644 data/2021/neurips/FastCorrect: Fast Error Correction with Edit Alignment for Automatic Speech Recognition create mode 100644 data/2021/neurips/Faster Algorithms and Constant Lower Bounds for the Worst-Case Expected Error create mode 100644 data/2021/neurips/Faster Directional Convergence of Linear Neural Networks under Spherically Symmetric Data create mode 100644 data/2021/neurips/Faster Matchings via Learned Duals create mode 100644 data/2021/neurips/Faster Neural Network Training with Approximate Tensor Operations create mode 100644 data/2021/neurips/Faster Non-asymptotic Convergence for Double Q-learning create mode 100644 data/2021/neurips/Faster proximal algorithms for matrix optimization using Jacobi-based eigenvalue methods create mode 100644 data/2021/neurips/Fault-Tolerant Federated Reinforcement Learning with Theoretical Guarantee create mode 100644 data/2021/neurips/FedDR - Randomized Douglas-Rachford Splitting Algorithms for Nonconvex Federated Composite Optimization create mode 100644 data/2021/neurips/Federated Graph Classification over Non-IID Graphs create mode 100644 data/2021/neurips/Federated Hyperparameter Tuning: Challenges, Baselines, and Connections to Weight-Sharing create mode 100644 data/2021/neurips/Federated Linear Contextual Bandits create mode 100644 data/2021/neurips/Federated Multi-Task Learning under a Mixture of Distributions create mode 100644 data/2021/neurips/Federated Reconstruction: Partially Local Federated Learning create mode 100644 data/2021/neurips/Federated Split Task-Agnostic Vision Transformer for COVID-19 CXR Diagnosis create mode 100644 data/2021/neurips/Federated-EM with heterogeneity mitigation and variance reduction create mode 100644 data/2021/neurips/Few-Round Learning for Federated Learning create mode 100644 data/2021/neurips/Few-Shot Data-Driven Algorithms for Low Rank Approximation create mode 100644 data/2021/neurips/Few-Shot Object Detection via Association and DIscrimination create mode 100644 data/2021/neurips/Few-Shot Segmentation via Cycle-Consistent Transformer create mode 100644 data/2021/neurips/Finding Bipartite Components in Hypergraphs create mode 100644 data/2021/neurips/Finding Discriminative Filters for Specific Degradations in Blind Super-Resolution create mode 100644 data/2021/neurips/Finding Optimal Tangent Points for Reducing Distortions of Hard-label Attacks create mode 100644 data/2021/neurips/Finding Regions of Heterogeneity in Decision-Making via Expected Conditional Covariance create mode 100644 data/2021/neurips/Fine-Grained Neural Network Explanation by Identifying Input Features with Predictive Information create mode 100644 data/2021/neurips/Fine-Grained Zero-Shot Learning with DNA as Side Information create mode 100644 data/2021/neurips/Fine-grained Generalization Analysis of Inductive Matrix Completion create mode 100644 data/2021/neurips/Finite Sample Analysis of Average-Reward TD Learning and $Q$-Learning create mode 100644 data/2021/neurips/Finite-Sample Analysis of Off-Policy TD-Learning via Generalized Bellman Operators create mode 100644 data/2021/neurips/Fitting summary statistics of neural data with a differentiable spiking network simulator create mode 100644 data/2021/neurips/Fixes That Fail: Self-Defeating Improvements in Machine-Learning Systems create mode 100644 data/2021/neurips/FjORD: Fair and Accurate Federated Learning under heterogeneous targets with Ordered Dropout create mode 100644 data/2021/neurips/Flattening Sharpness for Dynamic Gradient Projection Memory Benefits Continual Learning create mode 100644 data/2021/neurips/FlexMatch: Boosting Semi-Supervised Learning with Curriculum Pseudo Labeling create mode 100644 data/2021/neurips/Flexible Option Learning create mode 100644 data/2021/neurips/Flow Network based Generative Models for Non-Iterative Diverse Candidate Generation create mode 100644 data/2021/neurips/Focal Attention for Long-Range Interactions in Vision Transformers create mode 100644 data/2021/neurips/For high-dimensional hierarchical models, consider exchangeability of effects across covariates instead of across datasets create mode 100644 data/2021/neurips/Formalizing Generalization and Adversarial Robustness of Neural Networks to Weight Perturbations create mode 100644 data/2021/neurips/Formalizing the Generalization-Forgetting Trade-off in Continual Learning create mode 100644 data/2021/neurips/Forster Decomposition and Learning Halfspaces with Noise create mode 100644 data/2021/neurips/Foundations of Symbolic Languages for Model Interpretability create mode 100644 data/2021/neurips/Fractal Structure and Generalization Properties of Stochastic Optimization Algorithms create mode 100644 data/2021/neurips/Framing RNN as a kernel method: A neural ODE approach create mode 100644 data/2021/neurips/From Canonical Correlation Analysis to Self-supervised Graph Neural Networks create mode 100644 data/2021/neurips/From Optimality to Robustness: Adaptive Re-Sampling Strategies in Stochastic Bandits create mode 100644 data/2021/neurips/From global to local MDI variable importances for random forests and when they are Shapley values create mode 100644 data/2021/neurips/Functional Neural Networks for Parametric Image Restoration Problems create mode 100644 data/2021/neurips/Functional Regularization for Reinforcement Learning via Learned Fourier Features create mode 100644 data/2021/neurips/Functional Variational Inference based on Stochastic Process Generators create mode 100644 data/2021/neurips/Functionally Regionalized Knowledge Transfer for Low-resource Drug Discovery create mode 100644 data/2021/neurips/Fuzzy Clustering with Similarity Queries create mode 100644 data/2021/neurips/G-PATE: Scalable Differentially Private Data Generator via Private Aggregation of Teacher Discriminators create mode 100644 data/2021/neurips/GENESIS-V2: Inferring Unordered Object Representations without Iterative Refinement create mode 100644 data/2021/neurips/GRIN: Generative Relation and Intention Network for Multi-agent Trajectory Prediction create mode 100644 data/2021/neurips/Garment4D: Garment Reconstruction from Point Cloud Sequences create mode 100644 data/2021/neurips/Gauge Equivariant Transformer create mode 100644 data/2021/neurips/Gaussian Kernel Mixture Network for Single Image Defocus Deblurring create mode 100644 data/2021/neurips/GemNet: Universal Directional Graph Neural Networks for Molecules create mode 100644 data/2021/neurips/General Low-rank Matrix Optimization: Geometric Analysis and Sharper Bounds create mode 100644 data/2021/neurips/General Nonlinearities in SO(2)-Equivariant CNNs create mode 100644 data/2021/neurips/Generalizable Imitation Learning from Observation via Inferring Goal Proximity create mode 100644 data/2021/neurips/Generalizable Multi-linear Attention Network create mode 100644 data/2021/neurips/Generalization Bounds For Meta-Learning: An Information-Theoretic Analysis create mode 100644 data/2021/neurips/Generalization Bounds for (Wasserstein) Robust Optimization create mode 100644 data/2021/neurips/Generalization Bounds for Graph Embedding Using Negative Sampling: Linear vs Hyperbolic create mode 100644 data/2021/neurips/Generalization Bounds for Meta-Learning via PAC-Bayes and Uniform Stability create mode 100644 data/2021/neurips/Generalization Error Rates in Kernel Regression: The Crossover from the Noiseless to Noisy Regime create mode 100644 data/2021/neurips/Generalization Guarantee of SGD for Pairwise Learning create mode 100644 data/2021/neurips/Generalization of Model-Agnostic Meta-Learning Algorithms: Recurring and Unseen Tasks create mode 100644 data/2021/neurips/Generalized DataWeighting via Class-Level Gradient Manipulation create mode 100644 data/2021/neurips/Generalized Depthwise-Separable Convolutions for Adversarially Robust and Efficient Neural Networks create mode 100644 data/2021/neurips/Generalized Jensen-Shannon Divergence Loss for Learning with Noisy Labels create mode 100644 data/2021/neurips/Generalized Linear Bandits with Local Differential Privacy create mode 100644 data/2021/neurips/Generalized Proximal Policy Optimization with Sample Reuse create mode 100644 data/2021/neurips/Generalized Shape Metrics on Neural Representations create mode 100644 data/2021/neurips/Generalized and Discriminative Few-Shot Object Detection via SVD-Dictionary Enhancement create mode 100644 data/2021/neurips/Generating High-Quality Explanations for Navigation in Partially-Revealed Environments create mode 100644 data/2021/neurips/Generative Occupancy Fields for 3D Surface-Aware Image Synthesis create mode 100644 data/2021/neurips/Generative vs. Discriminative: Rethinking The Meta-Continual Learning create mode 100644 data/2021/neurips/Generic Neural Architecture Search via Regression create mode 100644 data/2021/neurips/GeoMol: Torsional Geometric Generation of Molecular 3D Conformer Ensembles create mode 100644 data/2021/neurips/Geometry Processing with Neural Fields create mode 100644 data/2021/neurips/Glance-and-Gaze Vision Transformer create mode 100644 data/2021/neurips/Global Convergence of Gradient Descent for Asymmetric Low-Rank Matrix Factorization create mode 100644 data/2021/neurips/Global Convergence of Online Optimization for Nonlinear Model Predictive Control create mode 100644 data/2021/neurips/Global Convergence to Local Minmax Equilibrium in Classes of Nonconvex Zero-Sum Games create mode 100644 data/2021/neurips/Global Filter Networks for Image Classification create mode 100644 data/2021/neurips/Global-aware Beam Search for Neural Abstractive Summarization create mode 100644 data/2021/neurips/Going Beyond Linear RL: Sample Efficient Neural Function Approximation create mode 100644 data/2021/neurips/Going Beyond Linear Transformers with Recurrent Fast Weight Programmers create mode 100644 data/2021/neurips/Gone Fishing: Neural Active Learning with Fisher Embeddings create mode 100644 data/2021/neurips/Good Classification Measures and How to Find Them create mode 100644 data/2021/neurips/Grad2Task: Improved Few-shot Text Classification Using Gradients for Task Representation create mode 100644 data/2021/neurips/GradInit: Learning to Initialize Neural Networks for Stable and Efficient Training create mode 100644 data/2021/neurips/Gradient Descent on Two-layer Nets: Margin Maximization and Simplicity Bias create mode 100644 data/2021/neurips/Gradient Driven Rewards to Guarantee Fairness in Collaborative Machine Learning create mode 100644 data/2021/neurips/Gradient Inversion with Generative Image Prior create mode 100644 data/2021/neurips/Gradient-Free Adversarial Training Against Image Corruption for Learning-based Steering create mode 100644 data/2021/neurips/Gradient-based Editing of Memory Examples for Online Task-free Continual Learning create mode 100644 data/2021/neurips/Gradient-based Hyperparameter Optimization Over Long Horizons create mode 100644 data/2021/neurips/Gradual Domain Adaptation without Indexed Intermediate Domains create mode 100644 data/2021/neurips/Grammar-Based Grounded Lexicon Learning create mode 100644 data/2021/neurips/Graph Adversarial Self-Supervised Learning create mode 100644 data/2021/neurips/Graph Differentiable Architecture Search with Structure Learning create mode 100644 data/2021/neurips/Graph Neural Networks with Adaptive Residual create mode 100644 data/2021/neurips/Graph Neural Networks with Local Graph Parameters create mode 100644 data/2021/neurips/Graph Posterior Network: Bayesian Predictive Uncertainty for Node Classification create mode 100644 data/2021/neurips/GraphFormers: GNN-nested Transformers for Representation Learning on Textual Graph create mode 100644 data/2021/neurips/Graphical Models in Heavy-Tailed Markets create mode 100644 data/2021/neurips/Greedy Approximation Algorithms for Active Sequential Hypothesis Testing create mode 100644 data/2021/neurips/Greedy and Random Quasi-Newton Methods with Faster Explicit Superlinear Convergence create mode 100644 data/2021/neurips/Grounding Representation Similarity Through Statistical Testing create mode 100644 data/2021/neurips/Grounding Spatio-Temporal Language with Transformers create mode 100644 data/2021/neurips/Grounding inductive biases in natural images: invariance stems from variations in data create mode 100644 data/2021/neurips/Group Equivariant Subsampling create mode 100644 data/2021/neurips/H-NeRF: Neural Radiance Fields for Rendering and Temporal Reconstruction of Humans in Motion create mode 100644 data/2021/neurips/HNPE: Leveraging Global Parameters for Neural Posterior Estimation create mode 100644 data/2021/neurips/HRFormer: High-Resolution Vision Transformer for Dense Predict create mode 100644 data/2021/neurips/HSVA: Hierarchical Semantic-Visual Adaptation for Zero-Shot Learning create mode 100644 data/2021/neurips/Habitat 2.0: Training Home Assistants to Rearrange their Habitat create mode 100644 data/2021/neurips/Hamiltonian Dynamics with Non-Newtonian Momentum for Rapid Sampling create mode 100644 data/2021/neurips/Handling Long-tailed Feature Distribution in AdderNets create mode 100644 data/2021/neurips/Hard-Attention for Scalable Image Classification create mode 100644 data/2021/neurips/Hardware-adaptive Efficient Latency Prediction for NAS via Meta-Learning create mode 100644 data/2021/neurips/Hash Layers For Large Sparse Models create mode 100644 data/2021/neurips/Heavy Ball Momentum for Conditional Gradient create mode 100644 data/2021/neurips/Heavy Ball Neural Ordinary Differential Equations create mode 100644 data/2021/neurips/Heavy Tails in SGD and Compressibility of Overparametrized Neural Networks create mode 100644 data/2021/neurips/Hessian Eigenspectra of More Realistic Nonlinear Models create mode 100644 data/2021/neurips/Heterogeneous Multi-player Multi-armed Bandits: Closing the Gap and Generalization create mode 100644 data/2021/neurips/Heuristic-Guided Reinforcement Learning create mode 100644 data/2021/neurips/Hierarchical Clustering: O(1)-Approximation for Well-Clustered Graphs create mode 100644 data/2021/neurips/Hierarchical Reinforcement Learning with Timed Subgoals create mode 100644 data/2021/neurips/Hierarchical Skills for Efficient Exploration create mode 100644 data/2021/neurips/High Probability Complexity Bounds for Line Search Based on Stochastic Oracles create mode 100644 data/2021/neurips/High-probability Bounds for Non-Convex Stochastic Optimization with Heavy Tails create mode 100644 data/2021/neurips/Higher Order Kernel Mean Embeddings to Capture Filtrations of Stochastic Processes create mode 100644 data/2021/neurips/Hindsight Task Relabelling: Experience Replay for Sparse Reward Meta-RL create mode 100644 data/2021/neurips/History Aware Multimodal Transformer for Vision-and-Language Navigation create mode 100644 data/2021/neurips/Hit and Lead Discovery with Explorative RL and Fragment-based Molecule Generation create mode 100644 data/2021/neurips/How Data Augmentation affects Optimization for Linear Regression create mode 100644 data/2021/neurips/How Does it Sound? create mode 100644 data/2021/neurips/How Fine-Tuning Allows for Effective Meta-Learning create mode 100644 data/2021/neurips/How Modular should Neural Module Networks Be for Systematic Generalization? create mode 100644 data/2021/neurips/How Powerful are Performance Predictors in Neural Architecture Search? create mode 100644 data/2021/neurips/How Should Pre-Trained Language Models Be Fine-Tuned Towards Adversarial Robustness? create mode 100644 data/2021/neurips/How Tight Can PAC-Bayes be in the Small Data Regime? create mode 100644 data/2021/neurips/How Well do Feature Visualizations Support Causal Understanding of CNN Activations? create mode 100644 data/2021/neurips/How can classical multidimensional scaling go wrong? create mode 100644 data/2021/neurips/How does a Neural Network's Architecture Impact its Robustness to Noisy Labels? create mode 100644 data/2021/neurips/How to transfer algorithmic reasoning knowledge to learn new algorithms? create mode 100644 data/2021/neurips/Human-Adversarial Visual Question Answering create mode 100644 data/2021/neurips/Hybrid Regret Bounds for Combinatorial Semi-Bandits and Adversarial Linear Bandits create mode 100644 data/2021/neurips/HyperSPNs: Compact and Expressive Probabilistic Circuits create mode 100644 data/2021/neurips/Hyperbolic Busemann Learning with Ideal Prototypes create mode 100644 data/2021/neurips/Hyperbolic Procrustes Analysis Using Riemannian Geometry create mode 100644 data/2021/neurips/Hypergraph Propagation and Community Selection for Objects Retrieval create mode 100644 data/2021/neurips/Hyperparameter Optimization Is Deceiving Us, and How to Stop It create mode 100644 data/2021/neurips/Hyperparameter Tuning is All You Need for LISTA create mode 100644 data/2021/neurips/INDIGO: GNN-Based Inductive Knowledge Graph Completion Using Pair-Wise Encoding create mode 100644 data/2021/neurips/IQ-Learn: Inverse soft-Q Learning for Imitation create mode 100644 data/2021/neurips/IRM - when it works and when it doesn't: A test case of natural language inference create mode 100644 data/2021/neurips/Identifiability in inverse reinforcement learning create mode 100644 data/2021/neurips/Identifiable Generative models for Missing Not at Random Data Imputation create mode 100644 data/2021/neurips/Identification and Estimation of Joint Probabilities of Potential Outcomes in Observational Studies with Covariate Information create mode 100644 data/2021/neurips/Identification of Partially Observed Linear Causal Models: Graphical Conditions for the Non-Gaussian and Heterogeneous Cases create mode 100644 data/2021/neurips/Identification of the Generalized Condorcet Winner in Multi-dueling Bandits create mode 100644 data/2021/neurips/Identifying and Benchmarking Natural Out-of-Context Prediction Problems create mode 100644 data/2021/neurips/Identity testing for Mallows model create mode 100644 data/2021/neurips/Image Generation using Continuous Filter Atoms create mode 100644 data/2021/neurips/ImageBART: Bidirectional Context with Multinomial Diffusion for Autoregressive Image Synthesis create mode 100644 data/2021/neurips/Imitating Deep Learning Dynamics via Locally Elastic Stochastic Differential Equations create mode 100644 data/2021/neurips/Imitation with Neural Density Models create mode 100644 data/2021/neurips/Implicit Bias of SGD for Diagonal Linear Networks: a Provable Benefit of Stochasticity create mode 100644 data/2021/neurips/Implicit Deep Adaptive Design: Policy-Based Experimental Design without Likelihoods create mode 100644 data/2021/neurips/Implicit Finite-Horizon Approximation and Efficient Optimal Algorithms for Stochastic Shortest Path create mode 100644 data/2021/neurips/Implicit Generative Copulas create mode 100644 data/2021/neurips/Implicit MLE: Backpropagating Through Discrete Exponential Family Distributions create mode 100644 data/2021/neurips/Implicit Regularization in Matrix Sensing via Mirror Descent create mode 100644 data/2021/neurips/Implicit SVD for Graph Representation Learning create mode 100644 data/2021/neurips/Implicit Semantic Response Alignment for Partial Domain Adaptation create mode 100644 data/2021/neurips/Implicit Sparse Regularization: The Impact of Depth and Early Stopping create mode 100644 data/2021/neurips/Implicit Task-Driven Probability Discrepancy Measure for Unsupervised Domain Adaptation create mode 100644 data/2021/neurips/Implicit Transformer Network for Screen Content Image Continuous Super-Resolution create mode 100644 data/2021/neurips/Impression learning: Online representation learning with synaptic plasticity create mode 100644 data/2021/neurips/Improve Agents without Retraining: Parallel Tree Search with Off-Policy Correction create mode 100644 data/2021/neurips/Improved Coresets and Sublinear Algorithms for Power Means in Euclidean Spaces create mode 100644 data/2021/neurips/Improved Guarantees for Offline Stochastic Matching via new Ordered Contention Resolution Schemes create mode 100644 data/2021/neurips/Improved Learning Rates of a Functional Lasso-type SVM with Sparse Multi-Kernel Representation create mode 100644 data/2021/neurips/Improved Regret Bounds for Tracking Experts with Memory create mode 100644 data/2021/neurips/Improved Regularization and Robustness for Fine-tuning in Neural Networks create mode 100644 data/2021/neurips/Improved Transformer for High-Resolution GANs create mode 100644 data/2021/neurips/Improved Variance-Aware Confidence Sets for Linear Bandits and Linear Mixture MDP create mode 100644 data/2021/neurips/Improving Anytime Prediction with Parallel Cascaded Networks and a Temporal-Difference Loss create mode 100644 data/2021/neurips/Improving Calibration through the Relationship with Adversarial Robustness create mode 100644 data/2021/neurips/Improving Coherence and Consistency in Neural Sequence Models with Dual-System, Neuro-Symbolic Reasoning create mode 100644 data/2021/neurips/Improving Compositionality of Neural Networks by Decoding Representations to Inputs create mode 100644 data/2021/neurips/Improving Computational Efficiency in Visual Reinforcement Learning via Stored Embeddings create mode 100644 data/2021/neurips/Improving Conditional Coverage via Orthogonal Quantile Regression create mode 100644 data/2021/neurips/Improving Contrastive Learning on Imbalanced Data via Open-World Sampling create mode 100644 data/2021/neurips/Improving Deep Learning Interpretability by Saliency Guided Training create mode 100644 data/2021/neurips/Improving Generalization in Meta-RL with Imaginary Tasks from Latent Dynamics Mixture create mode 100644 data/2021/neurips/Improving Robustness using Generated Data create mode 100644 data/2021/neurips/Improving Self-supervised Learning with Automated Unsupervised Outlier Arbitration create mode 100644 data/2021/neurips/Improving Transferability of Representations via Augmentation-Aware Self-Supervision create mode 100644 data/2021/neurips/Improving Visual Quality of Image Synthesis by A Token-based Generator with Transformers create mode 100644 data/2021/neurips/Improving black-box optimization in VAE latent space using decoder uncertainty create mode 100644 data/2021/neurips/Increasing Liquid State Machine Performance with Edge-of-Chaos Dynamics Organized by Astrocyte-modulated Plasticity create mode 100644 data/2021/neurips/Independent Prototype Propagation for Zero-Shot Compositionality create mode 100644 data/2021/neurips/Independent mechanism analysis, a new concept? create mode 100644 data/2021/neurips/Indexed Minimum Empirical Divergence for Unimodal Bandits create mode 100644 "data/2021/neurips/Individual Privacy Accounting via a R\303\251nyi Filter" create mode 100644 data/2021/neurips/Infinite Time Horizon Safety of Bayesian Neural Networks create mode 100644 data/2021/neurips/Influence Patterns for Explaining Information Flow in BERT create mode 100644 data/2021/neurips/InfoGCL: Information-Aware Graph Contrastive Learning create mode 100644 data/2021/neurips/Information Directed Reward Learning for Reinforcement Learning create mode 100644 data/2021/neurips/Information Directed Sampling for Sparse Linear Bandits create mode 100644 data/2021/neurips/Information is Power: Intrinsic Control via Information Capture create mode 100644 data/2021/neurips/Information-constrained optimization: can adaptive processing of gradients help? create mode 100644 data/2021/neurips/Information-theoretic generalization bounds for black-box learning algorithms create mode 100644 data/2021/neurips/Instance-Conditional Knowledge Distillation for Object Detection create mode 100644 data/2021/neurips/Instance-Conditioned GAN create mode 100644 data/2021/neurips/Instance-Dependent Bounds for Zeroth-order Lipschitz Optimization with Error Certificates create mode 100644 data/2021/neurips/Instance-Dependent Partial Label Learning create mode 100644 data/2021/neurips/Instance-dependent Label-noise Learning under a Structural Causal Model create mode 100644 data/2021/neurips/Instance-optimal Mean Estimation Under Differential Privacy create mode 100644 data/2021/neurips/Integrating Expert ODEs into Neural ODEs: Pharmacology and Disease Progression create mode 100644 data/2021/neurips/Integrating Tree Path in Transformer for Code Representation create mode 100644 data/2021/neurips/Interactive Label Cleaning with Example-based Explanations create mode 100644 data/2021/neurips/Interesting Object, Curious Agent: Learning Task-Agnostic Exploration create mode 100644 data/2021/neurips/Intermediate Layers Matter in Momentum Contrastive Self Supervised Learning create mode 100644 data/2021/neurips/Interpolation can hurt robust generalization even when there is no noise create mode 100644 data/2021/neurips/Interpretable agent communication from scratch (with a generic visual processor emerging on the side) create mode 100644 data/2021/neurips/Interpreting Representation Quality of DNNs for 3D Point Cloud Processing create mode 100644 data/2021/neurips/Interventional Sum-Product Networks: Causal Inference with Tractable Probabilistic Models create mode 100644 data/2021/neurips/Intriguing Properties of Contrastive Losses create mode 100644 data/2021/neurips/Intriguing Properties of Vision Transformers create mode 100644 data/2021/neurips/Intrinsic Dimension, Persistent Homology and Generalization in Neural Networks create mode 100644 data/2021/neurips/Introspective Distillation for Robust Question Answering create mode 100644 data/2021/neurips/Invariance Principle Meets Information Bottleneck for Out-of-Distribution Generalization create mode 100644 data/2021/neurips/Invariant Causal Imitation Learning for Generalizable Policies create mode 100644 data/2021/neurips/Inverse Optimal Control Adapted to the Noise Characteristics of the Human Sensorimotor System create mode 100644 data/2021/neurips/Inverse Problems Leveraging Pre-trained Contrastive Representations create mode 100644 data/2021/neurips/Inverse Reinforcement Learning in a Continuous State Space with Formal Guarantees create mode 100644 data/2021/neurips/Inverse-Weighted Survival Games create mode 100644 data/2021/neurips/Invertible DenseNets with Concatenated LipSwish create mode 100644 data/2021/neurips/Is Automated Topic Model Evaluation Broken? The Incoherence of Coherence create mode 100644 data/2021/neurips/Is Bang-Bang Control All You Need? Solving Continuous Control with Bernoulli Policies create mode 100644 data/2021/neurips/It Has Potential: Gradient-Driven Denoisers for Convergent Solutions to Inverse Problems create mode 100644 data/2021/neurips/Iterative Amortized Policy Optimization create mode 100644 data/2021/neurips/Iterative Causal Discovery in the Possible Presence of Latent Confounders and Selection Bias create mode 100644 data/2021/neurips/Iterative Connecting Probability Estimation for Networks create mode 100644 data/2021/neurips/Iterative Methods for Private Synthetic Data: Unifying Framework and New Methods create mode 100644 data/2021/neurips/Iterative Teacher-Aware Learning create mode 100644 data/2021/neurips/Iterative Teaching by Label Synthesis create mode 100644 data/2021/neurips/Iteratively Reweighted Least Squares for Basis Pursuit with Global Linear Convergence Rate create mode 100644 data/2021/neurips/Joint Inference for Neural Network Depth and Dropout Regularization create mode 100644 data/2021/neurips/Joint Modeling of Visual Objects and Relations for Scene Graph Generation create mode 100644 data/2021/neurips/Joint Semantic Mining for Weakly Supervised RGB-D Salient Object Detection create mode 100644 data/2021/neurips/Joint inference and input optimization in equilibrium networks create mode 100644 data/2021/neurips/K-Net: Towards Unified Image Segmentation create mode 100644 data/2021/neurips/K-level Reasoning for Zero-Shot Coordination in Hanabi create mode 100644 data/2021/neurips/KALE Flow: A Relaxed KL Gradient Flow for Probabilities with Disjoint Support create mode 100644 data/2021/neurips/KS-GNN: Keywords Search over Incomplete Graphs via Graphs Neural Network create mode 100644 data/2021/neurips/Keeping Your Eye on the Ball: Trajectory Attention in Video Transformers create mode 100644 data/2021/neurips/Kernel Functional Optimisation create mode 100644 data/2021/neurips/Kernel Identification Through Transformers create mode 100644 data/2021/neurips/Kernelized Heterogeneous Risk Minimization create mode 100644 data/2021/neurips/Knowledge-Adaptation Priors create mode 100644 data/2021/neurips/Knowledge-inspired 3D Scene Graph Prediction in Point Cloud create mode 100644 data/2021/neurips/L2ight: Enabling On-Chip Learning for Optical Neural Networks via Efficient in-situ Subspace Optimization create mode 100644 data/2021/neurips/LADA: Look-Ahead Data Acquisition via Augmentation for Deep Active Learning create mode 100644 data/2021/neurips/LEADS: Learning Dynamical Systems that Generalize Across Environments create mode 100644 data/2021/neurips/LLC: Accurate, Multi-purpose Learnt Low-dimensional Binary Codes create mode 100644 data/2021/neurips/LSH-SMILE: Locality Sensitive Hashing Accelerated Simulation and Learning create mode 100644 data/2021/neurips/Label Disentanglement in Partition-based Extreme Multilabel Classification create mode 100644 data/2021/neurips/Label Noise SGD Provably Prefers Flat Global Minimizers create mode 100644 data/2021/neurips/Label consistency in overfitted generalized $k$-means create mode 100644 data/2021/neurips/Label-Imbalanced and Group-Sensitive Classification under Overparameterization create mode 100644 data/2021/neurips/Labeling Trick: A Theory of Using Graph Neural Networks for Multi-Node Representation Learning create mode 100644 data/2021/neurips/Landmark-Guided Subgoal Generation in Hierarchical Reinforcement Learning create mode 100644 data/2021/neurips/Landmark-RxR: Solving Vision-and-Language Navigation with Fine-Grained Alignment Supervision create mode 100644 data/2021/neurips/Landscape analysis of an improved power method for tensor decomposition create mode 100644 data/2021/neurips/Language models enable zero-shot prediction of the effects of mutations on protein function create mode 100644 data/2021/neurips/Laplace Redux - Effortless Bayesian Deep Learning create mode 100644 data/2021/neurips/Large Scale Learning on Non-Homophilous Graphs: New Benchmarks and Strong Simple Methods create mode 100644 data/2021/neurips/Large-Scale Learning with Fourier Features and Tensor Decompositions create mode 100644 data/2021/neurips/Large-Scale Unsupervised Object Discovery create mode 100644 data/2021/neurips/Large-Scale Wasserstein Gradient Flows create mode 100644 data/2021/neurips/Last iterate convergence of SGD for Least-Squares in the Interpolation regime create mode 100644 data/2021/neurips/Last-iterate Convergence in Extensive-Form Games create mode 100644 data/2021/neurips/Latent Equilibrium: Arbitrarily fast computation with arbitrarily slow neurons create mode 100644 data/2021/neurips/Latent Execution for Neural Program Synthesis Beyond Domain-Specific Languages create mode 100644 data/2021/neurips/Latent Matters: Learning Deep State-Space Models create mode 100644 data/2021/neurips/Lattice partition recovery with dyadic CART create mode 100644 data/2021/neurips/Learnability of Linear Thresholds from Label Proportions create mode 100644 data/2021/neurips/Learnable Fourier Features for Multi-dimensional Spatial Positional Encoding create mode 100644 data/2021/neurips/Learned Robust PCA: A Scalable Deep Unfolding Approach for High-Dimensional Outlier Detection create mode 100644 data/2021/neurips/Learning 3D Dense Correspondence via Canonical Point Autoencoder create mode 100644 data/2021/neurips/Learning Barrier Certificates: Towards Safe Reinforcement Learning with Zero Training-time Violations create mode 100644 data/2021/neurips/Learning Causal Semantic Representation for Out-of-Distribution Prediction create mode 100644 data/2021/neurips/Learning Collaborative Policies to Solve NP-hard Routing Problems create mode 100644 data/2021/neurips/Learning Compact Representations of Neural Networks using DiscriminAtive Masking (DAM) create mode 100644 data/2021/neurips/Learning Conjoint Attentions for Graph Neural Nets create mode 100644 data/2021/neurips/Learning Debiased Representation via Disentangled Feature Augmentation create mode 100644 data/2021/neurips/Learning Debiased and Disentangled Representations for Semantic Segmentation create mode 100644 data/2021/neurips/Learning Disentangled Behavior Embeddings create mode 100644 data/2021/neurips/Learning Distilled Collaboration Graph for Multi-Agent Perception create mode 100644 data/2021/neurips/Learning Diverse Policies in MOBA Games via Macro-Goals create mode 100644 data/2021/neurips/Learning Domain Invariant Representations in Goal-conditioned Block MDPs create mode 100644 data/2021/neurips/Learning Dynamic Graph Representation of Brain Connectome with Spatio-Temporal Attention create mode 100644 data/2021/neurips/Learning Equilibria in Matching Markets from Bandit Feedback create mode 100644 data/2021/neurips/Learning Equivariant Energy Based Models with Equivariant Stein Variational Gradient Descent create mode 100644 data/2021/neurips/Learning Fast-Inference Bayesian Networks create mode 100644 data/2021/neurips/Learning Frequency Domain Approximation for Binary Neural Networks create mode 100644 data/2021/neurips/Learning Gaussian Mixtures with Generalized Linear Models: Precise Asymptotics in High-dimensions create mode 100644 data/2021/neurips/Learning Generalized Gumbel-max Causal Mechanisms create mode 100644 data/2021/neurips/Learning Generative Vision Transformer with Energy-Based Latent Space for Saliency Prediction create mode 100644 data/2021/neurips/Learning Graph Cellular Automata create mode 100644 data/2021/neurips/Learning Graph Models for Retrosynthesis Prediction create mode 100644 data/2021/neurips/Learning Hard Optimization Problems: A Data Generation Perspective create mode 100644 data/2021/neurips/Learning High-Precision Bounding Box for Rotated Object Detection via Kullback-Leibler Divergence create mode 100644 data/2021/neurips/Learning Interpretable Decision Rule Sets: A Submodular Optimization Approach create mode 100644 data/2021/neurips/Learning Knowledge Graph-based World Models of Textual Environments create mode 100644 data/2021/neurips/Learning Large Neighborhood Search Policy for Integer Programming create mode 100644 data/2021/neurips/Learning MDPs from Features: Predict-Then-Optimize for Sequential Decision Making by Reinforcement Learning create mode 100644 data/2021/neurips/Learning Markov State Abstractions for Deep Reinforcement Learning create mode 100644 data/2021/neurips/Learning Models for Actionable Recourse create mode 100644 data/2021/neurips/Learning Nonparametric Volterra Kernels with Gaussian Processes create mode 100644 data/2021/neurips/Learning Optimal Predictive Checklists create mode 100644 data/2021/neurips/Learning Policies with Zero or Bounded Constraint Violation for Constrained MDPs create mode 100644 data/2021/neurips/Learning Riemannian metric for disease progression modeling create mode 100644 data/2021/neurips/Learning Robust Hierarchical Patterns of Human Brain across Many fMRI Studies create mode 100644 data/2021/neurips/Learning Semantic Representations to Verify Hardware Designs create mode 100644 data/2021/neurips/Learning Signal-Agnostic Manifolds of Neural Fields create mode 100644 data/2021/neurips/Learning Space Partitions for Path Planning create mode 100644 data/2021/neurips/Learning Stable Deep Dynamics Models for Partially Observed or Delayed Dynamical Systems create mode 100644 data/2021/neurips/Learning State Representations from Random Deep Action-conditional Predictions create mode 100644 data/2021/neurips/Learning Stochastic Majority Votes by Minimizing a PAC-Bayes Generalization Bound create mode 100644 data/2021/neurips/Learning Student-Friendly Teacher Networks for Knowledge Distillation create mode 100644 data/2021/neurips/Learning Theory Can (Sometimes) Explain Generalisation in Graph Neural Networks create mode 100644 data/2021/neurips/Learning Transferable Adversarial Perturbations create mode 100644 data/2021/neurips/Learning Transferable Features for Point Cloud Detection via 3D Contrastive Co-training create mode 100644 data/2021/neurips/Learning Treatment Effects in Panels with General Intervention Patterns create mode 100644 data/2021/neurips/Learning Tree Interpretation from Object Representation for Deep Reinforcement Learning create mode 100644 data/2021/neurips/Learning a Single Neuron with Bias Using Gradient Descent create mode 100644 data/2021/neurips/Learning and Generalization in RNNs create mode 100644 data/2021/neurips/Learning curves of generic features maps for realistic datasets with a teacher-student model create mode 100644 data/2021/neurips/Learning from Inside: Self-driven Siamese Sampling and Reasoning for Video Question Answering create mode 100644 data/2021/neurips/Learning in Multi-Stage Decentralized Matching Markets create mode 100644 data/2021/neurips/Learning in Non-Cooperative Configurable Markov Decision Processes create mode 100644 data/2021/neurips/Learning in two-player zero-sum partially observable Markov games with perfect recall create mode 100644 data/2021/neurips/Learning interaction rules from multi-animal trajectories via augmented behavioral models create mode 100644 data/2021/neurips/Learning latent causal graphs via mixture oracles create mode 100644 data/2021/neurips/Learning on Random Balls is Sufficient for Estimating (Some) Graph Parameters create mode 100644 data/2021/neurips/Learning rule influences recurrent network representations but not attractor structure in decision-making tasks create mode 100644 data/2021/neurips/Learning the optimal Tikhonov regularizer for inverse problems create mode 100644 data/2021/neurips/Learning to Adapt via Latent Domains for Adaptive Semantic Segmentation create mode 100644 data/2021/neurips/Learning to Assimilate in Chaotic Dynamical Systems create mode 100644 data/2021/neurips/Learning to Combine Per-Example Solutions for Neural Program Synthesis create mode 100644 data/2021/neurips/Learning to Compose Visual Relations create mode 100644 data/2021/neurips/Learning to Draw: Emergent Communication through Sketching create mode 100644 data/2021/neurips/Learning to Elect create mode 100644 data/2021/neurips/Learning to Execute: Efficient Learning of Universal Plan-Conditioned Policies in Robotics create mode 100644 data/2021/neurips/Learning to Generate Realistic Noisy Images via Pixel-level Noise-aware Adversarial Training create mode 100644 data/2021/neurips/Learning to Generate Visual Questions with Noisy Supervision create mode 100644 data/2021/neurips/Learning to Ground Multi-Agent Communication with Autoencoders create mode 100644 data/2021/neurips/Learning to Iteratively Solve Routing Problems with Dual-Aspect Collaborative Transformer create mode 100644 data/2021/neurips/Learning to Learn Dense Gaussian Processes for Few-Shot Learning create mode 100644 data/2021/neurips/Learning to Learn Graph Topologies create mode 100644 data/2021/neurips/Learning to Predict Trustworthiness with Steep Slope Loss create mode 100644 data/2021/neurips/Learning to Schedule Heuristics in Branch and Bound create mode 100644 data/2021/neurips/Learning to See by Looking at Noise create mode 100644 data/2021/neurips/Learning to Select Exogenous Events for Marked Temporal Point Process create mode 100644 data/2021/neurips/Learning to Synthesize Programs as Interpretable and Generalizable Policies create mode 100644 data/2021/neurips/Learning to Time-Decode in Spiking Neural Networks Through the Information Bottleneck create mode 100644 data/2021/neurips/Learning to dehaze with polarization create mode 100644 data/2021/neurips/Learning to delegate for large-scale vehicle routing create mode 100644 data/2021/neurips/Learning where to learn: Gradient sparsity in meta and continual learning create mode 100644 data/2021/neurips/Learning with Algorithmic Supervision via Continuous Relaxations create mode 100644 data/2021/neurips/Learning with Holographic Reduced Representations create mode 100644 data/2021/neurips/Learning with Labeling Induced Abstentions create mode 100644 data/2021/neurips/Learning with Noisy Correspondence for Cross-modal Matching create mode 100644 data/2021/neurips/Learning with User-Level Privacy create mode 100644 data/2021/neurips/Learning-Augmented Dynamic Power Management with Multiple States via New Ski Rental Bounds create mode 100644 data/2021/neurips/Learning-to-learn non-convex piecewise-Lipschitz functions create mode 100644 data/2021/neurips/Least Square Calibration for Peer Reviews create mode 100644 data/2021/neurips/Leveraging Distribution Alignment via Stein Path for Cross-Domain Cold-Start Recommendation create mode 100644 data/2021/neurips/Leveraging Recursive Gumbel-Max Trick for Approximate Inference in Combinatorial Spaces create mode 100644 data/2021/neurips/Leveraging SE(3) Equivariance for Self-supervised Category-Level Object Pose Estimation from Point Clouds create mode 100644 data/2021/neurips/Leveraging Spatial and Temporal Correlations in Sparsified Mean Estimation create mode 100644 data/2021/neurips/Leveraging the Inductive Bias of Large Language Models for Abstract Textual Reasoning create mode 100644 data/2021/neurips/Lifelong Domain Adaptation via Consolidated Internal Distribution create mode 100644 data/2021/neurips/Light Field Networks: Neural Scene Representations with Single-Evaluation Rendering create mode 100644 data/2021/neurips/Limiting fluctuation and trajectorial stability of multilayer neural networks with mean field training create mode 100644 data/2021/neurips/Linear Convergence in Federated Learning: Tackling Client Heterogeneity and Sparse Gradients create mode 100644 data/2021/neurips/Linear Convergence of Gradient Methods for Estimating Structured Transition Matrices in High-dimensional Vector Autoregressive Models create mode 100644 data/2021/neurips/Linear and Kernel Classification in the Streaming Model: Improved Bounds for Heavy Hitters create mode 100644 data/2021/neurips/Linear-Time Probabilistic Solution of Boundary Value Problems create mode 100644 data/2021/neurips/Lip to Speech Synthesis with Visual Context Attentional GAN create mode 100644 data/2021/neurips/List-Decodable Mean Estimation in Nearly-PCA Time create mode 100644 data/2021/neurips/Littlestone Classes are Privately Online Learnable create mode 100644 data/2021/neurips/Local Differential Privacy for Regret Minimization in Reinforcement Learning create mode 100644 data/2021/neurips/Local Disentanglement in Variational Auto-Encoders Using Jacobian $L_1$ Regularization create mode 100644 data/2021/neurips/Local Explanation of Dialogue Response Generation create mode 100644 data/2021/neurips/Local Hyper-Flow Diffusion create mode 100644 data/2021/neurips/Local Signal Adaptivity: Provable Feature Learning in Neural Networks Beyond Kernels create mode 100644 data/2021/neurips/Local plasticity rules can learn deep representations using self-supervised contrastive predictions create mode 100644 data/2021/neurips/Local policy search with Bayesian optimization create mode 100644 data/2021/neurips/Locality Sensitive Teaching create mode 100644 data/2021/neurips/Locality defeats the curse of dimensionality in convolutional teacher-student scenarios create mode 100644 data/2021/neurips/Localization with Sampling-Argmax create mode 100644 data/2021/neurips/Localization, Convexity, and Star Aggregation create mode 100644 data/2021/neurips/Locally Most Powerful Bayesian Test for Out-of-Distribution Detection using Deep Generative Models create mode 100644 data/2021/neurips/Locally Valid and Discriminative Prediction Intervals for Deep Learning Models create mode 100644 data/2021/neurips/Locally differentially private estimation of functionals of discrete distributions create mode 100644 data/2021/neurips/Locally private online change point detection create mode 100644 data/2021/neurips/Logarithmic Regret from Sublinear Hints create mode 100644 data/2021/neurips/Logarithmic Regret in Feature-based Dynamic Pricing create mode 100644 data/2021/neurips/Long Short-Term Transformer for Online Action Detection create mode 100644 data/2021/neurips/Long-Short Transformer: Efficient Transformers for Language and Vision create mode 100644 data/2021/neurips/Look at What I'm Doing: Self-Supervised Spatial Grounding of Narrations in Instructional Videos create mode 100644 data/2021/neurips/Look at the Variance! Efficient Black-box Explanations with Sobol-based Sensitivity Analysis create mode 100644 data/2021/neurips/Looking Beyond Single Images for Contrastive Semantic Segmentation Learning create mode 100644 data/2021/neurips/Loss function based second-order Jensen inequality and its application to particle variational inference create mode 100644 data/2021/neurips/Lossy Compression for Lossless Prediction create mode 100644 data/2021/neurips/Low-Fidelity Video Encoder Optimization for Temporal Action Localization create mode 100644 data/2021/neurips/Low-Rank Constraints for Fast Inference in Structured Models create mode 100644 data/2021/neurips/Low-Rank Extragradient Method for Nonsmooth and Low-Rank Matrix Optimization Problems create mode 100644 data/2021/neurips/Low-Rank Subspaces in GANs create mode 100644 data/2021/neurips/Low-dimensional Structure in the Space of Language Representations is Reflected in Brain Responses create mode 100644 data/2021/neurips/Lower Bounds and Optimal Algorithms for Smooth and Strongly Convex Decentralized Optimization Over Time-Varying Networks create mode 100644 data/2021/neurips/Lower Bounds on Metropolized Sampling Methods for Well-Conditioned Distributions create mode 100644 data/2021/neurips/Lower and Upper Bounds on the Pseudo-Dimension of Tensor Network Models create mode 100644 data/2021/neurips/Luna: Linear Unified Nested Attention create mode 100644 data/2021/neurips/M-FAC: Efficient Matrix-Free Approximations of Second-Order Information create mode 100644 data/2021/neurips/MADE: Exploration via Maximizing Deviation from Explored Regions create mode 100644 data/2021/neurips/MAP Propagation Algorithm: Faster Learning with a Team of Reinforcement Learning Agents create mode 100644 data/2021/neurips/MAU: A Motion-Aware Unit for Video Prediction and Beyond create mode 100644 data/2021/neurips/MAUVE: Measuring the Gap Between Neural Text and Human Text using Divergence Frontiers create mode 100644 data/2021/neurips/MCMC Variational Inference via Uncorrected Hamiltonian Annealing create mode 100644 data/2021/neurips/MERLOT: Multimodal Neural Script Knowledge Models create mode 100644 data/2021/neurips/MEST: Accurate and Fast Memory-Economic Sparse Training Framework on the Edge create mode 100644 data/2021/neurips/MICo: Improved representations via sampling-based state similarity for Markov decision processes create mode 100644 data/2021/neurips/MIRACLE: Causally-Aware Imputation via Learning Missing Data Mechanisms create mode 100644 data/2021/neurips/MLP-Mixer: An all-MLP Architecture for Vision create mode 100644 data/2021/neurips/MOMA: Multi-Object Multi-Actor Activity Parsing create mode 100644 data/2021/neurips/MST: Masked Self-Supervised Transformer for Visual Representation create mode 100644 data/2021/neurips/Machine Learning for Variance Reduction in Online Experiments create mode 100644 data/2021/neurips/Machine learning structure preserving brackets for forecasting irreversible processes create mode 100644 data/2021/neurips/Machine versus Human Attention in Deep Reinforcement Learning Tasks create mode 100644 data/2021/neurips/MagNet: A Neural Network for Directed Graphs create mode 100644 data/2021/neurips/Make Sure You're Unsure: A Framework for Verifying Probabilistic Specifications create mode 100644 data/2021/neurips/Making a (Counterfactual) Difference One Rationale at a Time create mode 100644 data/2021/neurips/Making the most of your day: online learning for optimal allocation of time create mode 100644 data/2021/neurips/Manifold Topology Divergence: a Framework for Comparing Data Manifolds create mode 100644 data/2021/neurips/Manipulating SGD with Data Ordering Attacks create mode 100644 data/2021/neurips/Margin-Independent Online Multiclass Learning via Convex Geometry create mode 100644 data/2021/neurips/Marginalised Gaussian Processes with Nested Sampling create mode 100644 data/2021/neurips/MarioNette: Self-Supervised Sprite Learning create mode 100644 data/2021/neurips/Mastering Atari Games with Limited Data create mode 100644 data/2021/neurips/Matching a Desired Causal State via Shift Interventions create mode 100644 data/2021/neurips/Matrix encoding networks for neural combinatorial optimization create mode 100644 data/2021/neurips/Matrix factorisation and the interpretation of geodesic distance create mode 100644 data/2021/neurips/Maximum Likelihood Training of Score-Based Diffusion Models create mode 100644 data/2021/neurips/Measuring Generalization with Optimal Transport create mode 100644 data/2021/neurips/Medical Dead-ends and Learning to Identify High-Risk States and Treatments create mode 100644 data/2021/neurips/Memory Efficient Meta-Learning with Large Images create mode 100644 data/2021/neurips/Memory-Efficient Approximation Algorithms for Max-k-Cut and Correlation Clustering create mode 100644 data/2021/neurips/Memory-efficient Patch-based Inference for Tiny Deep Learning create mode 100644 data/2021/neurips/Meta Internal Learning create mode 100644 data/2021/neurips/Meta Learning Backpropagation And Improving It create mode 100644 data/2021/neurips/Meta Two-Sample Testing: Learning Kernels for Testing with Limited Data create mode 100644 data/2021/neurips/Meta-Adaptive Nonlinear Control: Theory and Algorithms create mode 100644 data/2021/neurips/Meta-Learning Reliable Priors in the Function Space create mode 100644 data/2021/neurips/Meta-Learning Sparse Implicit Neural Representations create mode 100644 data/2021/neurips/Meta-Learning for Relative Density-Ratio Estimation create mode 100644 data/2021/neurips/Meta-Learning the Search Distribution of Black-Box Random Search Based Adversarial Attacks create mode 100644 data/2021/neurips/Meta-Learning via Learning with Distributed Memory create mode 100644 data/2021/neurips/Meta-learning to Improve Pre-training create mode 100644 data/2021/neurips/Meta-learning with an Adaptive Task Scheduler create mode 100644 data/2021/neurips/MetaAvatar: Learning Animatable Clothed Human Models from Few Depth Images create mode 100644 data/2021/neurips/Metadata-based Multi-Task Bandits with Bayesian Hierarchical Models create mode 100644 data/2021/neurips/Metropolis-Hastings Data Augmentation for Graph Neural Networks create mode 100644 data/2021/neurips/Mind the Gap: Assessing Temporal Generalization in Neural Language Models create mode 100644 data/2021/neurips/Mini-Batch Consistent Slot Set Encoder for Scalable Set Encoding create mode 100644 data/2021/neurips/Minibatch and Momentum Model-based Methods for Stochastic Weakly Convex Optimization create mode 100644 data/2021/neurips/Minimax Optimal Quantile and Semi-Adversarial Regret via Root-Logarithmic Regularizers create mode 100644 data/2021/neurips/Minimax Regret for Stochastic Shortest Path create mode 100644 data/2021/neurips/Minimizing Polarization and Disagreement in Social Networks via Link Recommendation create mode 100644 data/2021/neurips/Mining the Benefits of Two-stage and One-stage HOI Detection create mode 100644 data/2021/neurips/Mirror Langevin Monte Carlo: the Case Under Isoperimetry create mode 100644 data/2021/neurips/Misspecified Gaussian Process Bandit Optimization create mode 100644 data/2021/neurips/Mitigating Covariate Shift in Imitation Learning via Offline Data With Partial Coverage create mode 100644 data/2021/neurips/Mitigating Forgetting in Online Continual Learning with Neuron Calibration create mode 100644 data/2021/neurips/MixACM: Mixup-Based Robustness Transfer via Distillation of Activated Channel Maps create mode 100644 data/2021/neurips/MixSeq: Connecting Macroscopic Time Series Forecasting with Microscopic Time Series Data create mode 100644 data/2021/neurips/Mixability made efficient: Fast online multiclass logistic regression create mode 100644 data/2021/neurips/Mixed Supervised Object Detection by Transferring Mask Prior and Semantic Similarity create mode 100644 data/2021/neurips/Mixture Proportion Estimation and PU Learning: A Modern Approach create mode 100644 data/2021/neurips/Mixture weights optimisation for Alpha-Divergence Variational Inference create mode 100644 data/2021/neurips/MobILE: Model-Based Imitation Learning From Observation Alone create mode 100644 data/2021/neurips/MobTCast: Leveraging Auxiliary Trajectory Forecasting for Human Mobility Prediction create mode 100644 data/2021/neurips/Modality-Agnostic Topology Aware Localization create mode 100644 data/2021/neurips/Model Adaptation: Historical Contrastive Learning for Unsupervised Domain Adaptation without Source Data create mode 100644 data/2021/neurips/Model Selection for Bayesian Autoencoders create mode 100644 data/2021/neurips/Model, sample, and epoch-wise descents: exact solution of gradient flow in the random feature model create mode 100644 data/2021/neurips/Model-Based Domain Generalization create mode 100644 data/2021/neurips/Model-Based Episodic Memory Induces Dynamic Hybrid Controls create mode 100644 data/2021/neurips/Model-Based Reinforcement Learning via Imagination with Derived Memory create mode 100644 data/2021/neurips/Modeling Heterogeneous Hierarchies with Relation-specific Hyperbolic Cones create mode 100644 data/2021/neurips/Modified Frank Wolfe in Probability Space create mode 100644 data/2021/neurips/Modular Gaussian Processes for Transfer Learning create mode 100644 data/2021/neurips/Momentum Centering and Asynchronous Update for Adaptive Gradient Methods create mode 100644 data/2021/neurips/Monte Carlo Tree Search With Iteratively Refining State Abstractions create mode 100644 "data/2021/neurips/Mori\303\251 Attack (MA): A New Potential Risk of Screen Photos" create mode 100644 data/2021/neurips/Mosaicking to Distill: Knowledge Distillation from Out-of-Domain Data create mode 100644 data/2021/neurips/Moser Flow: Divergence-based Generative Modeling on Manifolds create mode 100644 data/2021/neurips/Moshpit SGD: Communication-Efficient Decentralized Training on Heterogeneous Unreliable Devices create mode 100644 data/2021/neurips/Motif-based Graph Self-Supervised Learning for Molecular Property Prediction create mode 100644 data/2021/neurips/Multi-Agent Reinforcement Learning for Active Voltage Control on Power Distribution Networks create mode 100644 data/2021/neurips/Multi-Agent Reinforcement Learning in Stochastic Networked Systems create mode 100644 data/2021/neurips/Multi-Armed Bandits with Bounded Arm-Memory: Near-Optimal Guarantees for Best-Arm Identification and Regret Minimization create mode 100644 data/2021/neurips/Multi-Facet Clustering Variational Autoencoders create mode 100644 data/2021/neurips/Multi-Label Learning with Pairwise Relevance Ordering create mode 100644 data/2021/neurips/Multi-Objective Meta Learning create mode 100644 data/2021/neurips/Multi-Objective SPIBB: Seldonian Offline Policy Improvement with Safety Constraints in Finite MDPs create mode 100644 data/2021/neurips/Multi-Person 3D Motion Prediction with Multi-Range Transformers create mode 100644 data/2021/neurips/Multi-Scale Representation Learning on Proteins create mode 100644 data/2021/neurips/Multi-Step Budgeted Bayesian Optimization with Unknown Evaluation Costs create mode 100644 data/2021/neurips/Multi-View Representation Learning via Total Correlation Objective create mode 100644 data/2021/neurips/Multi-armed Bandit Requiring Monotone Arm Sequences create mode 100644 data/2021/neurips/Multi-modal Dependency Tree for Video Captioning create mode 100644 data/2021/neurips/Multi-task Learning of Order-Consistent Causal Graphs create mode 100644 data/2021/neurips/Multi-view Contrastive Graph Clustering create mode 100644 data/2021/neurips/Multiclass Boosting and the Cost of Weak Learning create mode 100644 data/2021/neurips/Multiclass versus Binary Differentially Private PAC Learning create mode 100644 data/2021/neurips/Multilingual Pre-training with Universal Dependency Learning create mode 100644 data/2021/neurips/Multimodal Few-Shot Learning with Frozen Language Models create mode 100644 data/2021/neurips/Multimodal Virtual Point 3D Detection create mode 100644 data/2021/neurips/Multimodal and Multilingual Embeddings for Large-Scale Speech Mining create mode 100644 data/2021/neurips/Multiple Descent: Design Your Own Generalization Curve create mode 100644 data/2021/neurips/Multiwavelet-based Operator Learning for Differential Equations create mode 100644 data/2021/neurips/NAS-Bench-x11 and the Power of Learning Curves create mode 100644 data/2021/neurips/NEO: Non Equilibrium Sampling on the Orbits of a Deterministic Transform create mode 100644 data/2021/neurips/NN-Baker: A Neural-network Infused Algorithmic Framework for Optimization Problems on Geometric Intersection Graphs create mode 100644 data/2021/neurips/NORESQA: A Framework for Speech Quality Assessment using Non-Matching References create mode 100644 data/2021/neurips/NTopo: Mesh-free Topology Optimization using Implicit Neural Representations create mode 100644 data/2021/neurips/Navigating to the Best Policy in Markov Decision Processes create mode 100644 data/2021/neurips/NeRS: Neural Reflectance Surfaces for Sparse-view 3D Reconstruction in the Wild create mode 100644 data/2021/neurips/NeRV: Neural Representations for Videos create mode 100644 data/2021/neurips/Near Optimal Policy Optimization via REPS create mode 100644 data/2021/neurips/Near-Optimal Lower Bounds For Convex Optimization For All Orders of Smoothness create mode 100644 data/2021/neurips/Near-Optimal Multi-Perturbation Experimental Design for Causal Structure Learning create mode 100644 data/2021/neurips/Near-Optimal No-Regret Learning in General Games create mode 100644 data/2021/neurips/Near-Optimal Offline Reinforcement Learning via Double Variance Reduction create mode 100644 data/2021/neurips/Near-optimal Offline and Streaming Algorithms for Learning Non-Linear Dynamical Systems create mode 100644 data/2021/neurips/Nearly Horizon-Free Offline Reinforcement Learning create mode 100644 data/2021/neurips/Nearly Minimax Optimal Reinforcement Learning for Discounted MDPs create mode 100644 data/2021/neurips/Nearly-Tight and Oblivious Algorithms for Explainable Clustering create mode 100644 data/2021/neurips/Necessary and sufficient graphical conditions for optimal adjustment sets in causal graphical models with hidden variables create mode 100644 data/2021/neurips/Neighborhood Reconstructing Autoencoders create mode 100644 data/2021/neurips/Neo-GNNs: Neighborhood Overlap-aware Graph Neural Networks for Link Prediction create mode 100644 data/2021/neurips/Nested Counterfactual Identification from Arbitrary Surrogate Experiments create mode 100644 data/2021/neurips/Nested Graph Neural Networks create mode 100644 data/2021/neurips/Nested Variational Inference create mode 100644 data/2021/neurips/Network-to-Network Regularization: Enforcing Occam's Razor to Improve Generalization create mode 100644 data/2021/neurips/NeuS: Learning Neural Implicit Surfaces by Volume Rendering for Multi-view Reconstruction create mode 100644 data/2021/neurips/NeurWIN: Neural Whittle Index Network For Restless Bandits Via Deep RL create mode 100644 data/2021/neurips/Neural Active Learning with Performance Guarantees create mode 100644 data/2021/neurips/Neural Additive Models: Interpretable Machine Learning with Neural Nets create mode 100644 data/2021/neurips/Neural Algorithmic Reasoners are Implicit Planners create mode 100644 data/2021/neurips/Neural Analysis and Synthesis: Reconstructing Speech from Self-Supervised Representations create mode 100644 data/2021/neurips/Neural Auto-Curricula in Two-Player Zero-Sum Games create mode 100644 data/2021/neurips/Neural Bellman-Ford Networks: A General Graph Neural Network Framework for Link Prediction create mode 100644 data/2021/neurips/Neural Bootstrapper create mode 100644 data/2021/neurips/Neural Circuit Synthesis from Specification Patterns create mode 100644 data/2021/neurips/Neural Distance Embeddings for Biological Sequences create mode 100644 data/2021/neurips/Neural Dubber: Dubbing for Videos According to Scripts create mode 100644 data/2021/neurips/Neural Ensemble Search for Uncertainty Estimation and Dataset Shift create mode 100644 data/2021/neurips/Neural Flows: Efficient Alternative to Neural ODEs create mode 100644 data/2021/neurips/Neural Human Performer: Learning Generalizable Radiance Fields for Human Performance Rendering create mode 100644 data/2021/neurips/Neural Hybrid Automata: Learning Dynamics With Multiple Modes and Stochastic Transitions create mode 100644 data/2021/neurips/Neural Population Geometry Reveals the Role of Stochasticity in Robust Perception create mode 100644 data/2021/neurips/Neural Production Systems create mode 100644 data/2021/neurips/Neural Program Generation Modulo Static Analysis create mode 100644 data/2021/neurips/Neural Pseudo-Label Optimism for the Bank Loan Problem create mode 100644 data/2021/neurips/Neural Regression, Representational Similarity, Model Zoology & Neural Taskonomy at Scale in Rodent Visual Cortex create mode 100644 data/2021/neurips/Neural Relightable Participating Media Rendering create mode 100644 data/2021/neurips/Neural Routing by Memory create mode 100644 data/2021/neurips/Neural Rule-Execution Tracking Machine For Transformer-Based Text Generation create mode 100644 data/2021/neurips/Neural Scene Flow Prior create mode 100644 data/2021/neurips/Neural Symplectic Form: Learning Hamiltonian Equations on General Coordinate Systems create mode 100644 data/2021/neurips/Neural Tangent Kernel Maximum Mean Discrepancy create mode 100644 data/2021/neurips/Neural Trees for Learning on Graphs create mode 100644 data/2021/neurips/Neural View Synthesis and Matching for Semi-Supervised Few-Shot Learning of 3D Pose create mode 100644 data/2021/neurips/Neural optimal feedback control with local learning rules create mode 100644 data/2021/neurips/Neural-PIL: Neural Pre-Integrated Lighting for Reflectance Decomposition create mode 100644 data/2021/neurips/NeuroLKH: Combining Deep Learning Model with Lin-Kernighan-Helsgaun Heuristic for Solving the Traveling Salesman Problem create mode 100644 data/2021/neurips/NeuroMLR: Robust & Reliable Route Recommendation on Road Networks create mode 100644 data/2021/neurips/Never Go Full Batch (in Stochastic Convex Optimization) create mode 100644 data/2021/neurips/Newton-LESS: Sparsification without Trade-offs for the Sketched Newton Update create mode 100644 data/2021/neurips/No Fear of Heterogeneity: Classifier Calibration for Federated Learning with Non-IID Data create mode 100644 data/2021/neurips/No RL, No Simulation: Learning to Navigate without Navigating create mode 100644 data/2021/neurips/No Regrets for Learning the Prior in Bandits create mode 100644 data/2021/neurips/No-Press Diplomacy from Scratch create mode 100644 data/2021/neurips/No-regret Online Learning over Riemannian Manifolds create mode 100644 data/2021/neurips/Node Dependent Local Smoothing for Scalable Graph Learning create mode 100644 data/2021/neurips/Noether Networks: meta-learning useful conserved quantities create mode 100644 data/2021/neurips/Noether's Learning Dynamics: Role of Symmetry Breaking in Neural Networks create mode 100644 data/2021/neurips/Noise2Score: Tweedie's Approach to Self-Supervised Image Denoising without Clean Images create mode 100644 "data/2021/neurips/Noisy Adaptation Generates L\303\251vy Flights in Attractor Neural Networks" create mode 100644 data/2021/neurips/Noisy Recurrent Neural Networks create mode 100644 data/2021/neurips/Non-Asymptotic Analysis for Two Time-scale TDC with General Smooth Function Approximation create mode 100644 data/2021/neurips/Non-Gaussian Gaussian Processes for Few-Shot Regression create mode 100644 data/2021/neurips/Non-approximate Inference for Collective Graphical Models on Path Graphs via Discrete Difference of Convex Algorithm create mode 100644 data/2021/neurips/Non-asymptotic Error Bounds for Bidirectional GANs create mode 100644 data/2021/neurips/Non-asymptotic convergence bounds for Wasserstein approximation using point clouds create mode 100644 data/2021/neurips/Non-convex Distributionally Robust Optimization: Non-asymptotic Analysis create mode 100644 data/2021/neurips/Non-local Latent Relation Distillation for Self-Adaptive 3D Human Pose Estimation create mode 100644 data/2021/neurips/Nonparametric estimation of continuous DPPs with kernel methods create mode 100644 data/2021/neurips/Nonsmooth Implicit Differentiation for Machine-Learning and Optimization create mode 100644 data/2021/neurips/Nonuniform Negative Sampling and Log Odds Correction with Rare Events Data create mode 100644 data/2021/neurips/Not All Images are Worth 16x16 Words: Dynamic Transformers for Efficient Image Recognition create mode 100644 data/2021/neurips/Not All Low-Pass Filters are Robust in Graph Convolutional Networks create mode 100644 data/2021/neurips/Novel Upper Bounds for the Constrained Most Probable Explanation Task create mode 100644 data/2021/neurips/Novel Visual Category Discovery with Dual Ranking Statistics and Mutual Knowledge Distillation create mode 100644 data/2021/neurips/NovelD: A Simple yet Effective Exploration Criterion create mode 100644 data/2021/neurips/Numerical Composition of Differential Privacy create mode 100644 data/2021/neurips/Numerical influence of ReLU'(0) on backpropagation create mode 100644 data/2021/neurips/NxMTransformer: Semi-Structured Sparsification for Natural Language Understanding via ADMM create mode 100644 data/2021/neurips/OSOA: One-Shot Online Adaptation of Deep Generative Models for Lossless Compression create mode 100644 data/2021/neurips/Object DGCNN: 3D Object Detection using Dynamic Graphs create mode 100644 data/2021/neurips/Object-Aware Regularization for Addressing Causal Confusion in Imitation Learning create mode 100644 data/2021/neurips/Object-Centric Representation Learning with Generative Spatial-Temporal Factorization create mode 100644 data/2021/neurips/Object-aware Contrastive Learning for Debiased Scene Representation create mode 100644 data/2021/neurips/Observation-Free Attacks on Stochastic Bandits create mode 100644 data/2021/neurips/OctField: Hierarchical Implicit Functions for 3D Modeling create mode 100644 data/2021/neurips/Off-Policy Risk Assessment in Contextual Bandits create mode 100644 data/2021/neurips/Offline Constrained Multi-Objective Reinforcement Learning via Pessimistic Dual Value Iteration create mode 100644 data/2021/neurips/Offline Meta Reinforcement Learning - Identifiability Challenges and Effective Data Collection Strategies create mode 100644 data/2021/neurips/Offline Model-based Adaptable Policy Learning create mode 100644 data/2021/neurips/Offline RL Without Off-Policy Evaluation create mode 100644 data/2021/neurips/Offline Reinforcement Learning as One Big Sequence Modeling Problem create mode 100644 data/2021/neurips/Offline Reinforcement Learning with Reverse Model-based Imagination create mode 100644 data/2021/neurips/On Blame Attribution for Accountable Multi-Agent Sequential Decision Making create mode 100644 data/2021/neurips/On Calibration and Out-of-Domain Generalization create mode 100644 data/2021/neurips/On Component Interactions in Two-Stage Recommender Systems create mode 100644 data/2021/neurips/On Contrastive Representations of Stochastic Processes create mode 100644 data/2021/neurips/On Density Estimation with Diffusion Models create mode 100644 data/2021/neurips/On Effective Scheduling of Model-based Reinforcement Learning create mode 100644 data/2021/neurips/On Empirical Risk Minimization with Dependent and Heavy-Tailed Data create mode 100644 data/2021/neurips/On Episodes, Prototypical Networks, and Few-Shot Learning create mode 100644 data/2021/neurips/On Inductive Biases for Heterogeneous Treatment Effect Estimation create mode 100644 data/2021/neurips/On Interaction Between Augmentations and Corruptions in Natural Corruption Robustness create mode 100644 data/2021/neurips/On Joint Learning for Solving Placement and Routing in Chip Design create mode 100644 data/2021/neurips/On Large-Cohort Training for Federated Learning create mode 100644 data/2021/neurips/On Learning Domain-Invariant Representations for Transfer Learning with Multiple Sources create mode 100644 data/2021/neurips/On Linear Stability of SGD and Input-Smoothness of Neural Networks create mode 100644 data/2021/neurips/On Locality of Local Explanation Models create mode 100644 data/2021/neurips/On Margin-Based Cluster Recovery with Oracle Queries create mode 100644 data/2021/neurips/On Memorization in Probabilistic Deep Generative Models create mode 100644 data/2021/neurips/On Model Calibration for Long-Tailed Object Detection and Instance Segmentation create mode 100644 data/2021/neurips/On Optimal Interpolation in Linear Regression create mode 100644 data/2021/neurips/On Optimal Robustness to Adversarial Corruption in Online Decision Problems create mode 100644 data/2021/neurips/On Path Integration of Grid Cells: Group Representation and Isotropic Scaling create mode 100644 data/2021/neurips/On Pathologies in KL-Regularized Reinforcement Learning from Expert Demonstrations create mode 100644 data/2021/neurips/On Plasticity, Invariance, and Mutually Frozen Weights in Sequential Task Learning create mode 100644 data/2021/neurips/On Provable Benefits of Depth in Training Graph Convolutional Networks create mode 100644 data/2021/neurips/On Riemannian Optimization over Positive Definite Matrices with the Bures-Wasserstein Geometry create mode 100644 data/2021/neurips/On Robust Optimal Transport: Computational Complexity and Barycenter Computation create mode 100644 data/2021/neurips/On Success and Simplicity: A Second Look at Transferable Targeted Attacks create mode 100644 data/2021/neurips/On The Structure of Parametric Tournaments with Application to Ranking from Pairwise Comparisons create mode 100644 data/2021/neurips/On Training Implicit Models create mode 100644 data/2021/neurips/On UMAP's True Loss Function create mode 100644 data/2021/neurips/On learning sparse vectors from mixture of responses create mode 100644 data/2021/neurips/On sensitivity of meta-learning to support data create mode 100644 data/2021/neurips/On the Algorithmic Stability of Adversarial Training create mode 100644 data/2021/neurips/On the Bias-Variance-Cost Tradeoff of Stochastic Optimization create mode 100644 data/2021/neurips/On the Convergence Theory of Debiased Model-Agnostic Meta-Reinforcement Learning create mode 100644 data/2021/neurips/On the Convergence and Sample Efficiency of Variance-Reduced Policy Gradient Method create mode 100644 data/2021/neurips/On the Convergence of Prior-Guided Zeroth-Order Optimization Algorithms create mode 100644 data/2021/neurips/On the Convergence of Step Decay Step-Size for Stochastic Optimization create mode 100644 data/2021/neurips/On the Cryptographic Hardness of Learning Single Periodic Neurons create mode 100644 data/2021/neurips/On the Equivalence between Neural Network and Support Vector Machine create mode 100644 data/2021/neurips/On the Estimation Bias in Double Q-Learning create mode 100644 data/2021/neurips/On the Existence of The Adversarial Bayes Classifier create mode 100644 data/2021/neurips/On the Expected Complexity of Maxout Networks create mode 100644 data/2021/neurips/On the Expressivity of Markov Reward create mode 100644 data/2021/neurips/On the Frequency Bias of Generative Models create mode 100644 data/2021/neurips/On the Generative Utility of Cyclic Conditionals create mode 100644 data/2021/neurips/On the Importance of Gradients for Detecting Distributional Shifts in the Wild create mode 100644 data/2021/neurips/On the Out-of-distribution Generalization of Probabilistic Image Modelling create mode 100644 data/2021/neurips/On the Periodic Behavior of Neural Network Training with Batch Normalization and Weight Decay create mode 100644 data/2021/neurips/On the Power of Differentiable Learning versus PAC and SQ Learning create mode 100644 data/2021/neurips/On the Power of Edge Independent Graph Models create mode 100644 data/2021/neurips/On the Provable Generalization of Recurrent Neural Networks create mode 100644 data/2021/neurips/On the Representation Power of Set Pooling Networks create mode 100644 data/2021/neurips/On the Representation of Solutions to Elliptic PDEs in Barron Spaces create mode 100644 data/2021/neurips/On the Role of Optimization in Double Descent: A Least Squares Study create mode 100644 data/2021/neurips/On the Sample Complexity of Learning under Geometric Stability create mode 100644 data/2021/neurips/On the Sample Complexity of Privately Learning Axis-Aligned Rectangles create mode 100644 data/2021/neurips/On the Second-order Convergence Properties of Random Search Methods create mode 100644 data/2021/neurips/On the Stochastic Stability of Deep Markov Models create mode 100644 data/2021/neurips/On the Suboptimality of Thompson Sampling in High Dimensions create mode 100644 data/2021/neurips/On the Theory of Reinforcement Learning with Once-per-Episode Feedback create mode 100644 data/2021/neurips/On the Universality of Graph Neural Networks on Large Random Graphs create mode 100644 data/2021/neurips/On the Validity of Modeling SGD with Stochastic Differential Equations (SDEs) create mode 100644 data/2021/neurips/On the Value of Infinite Gradients in Variational Autoencoder Models create mode 100644 data/2021/neurips/On the Value of Interaction and Function Approximation in Imitation Learning create mode 100644 data/2021/neurips/On the Variance of the Fisher Information for Deep Learning create mode 100644 data/2021/neurips/On the interplay between data structure and loss function in classification problems create mode 100644 data/2021/neurips/One Explanation is Not Enough: Structured Attention Graphs for Image Classification create mode 100644 data/2021/neurips/One Loss for All: Deep Hashing with a Single Cosine Similarity based Learning Objective create mode 100644 data/2021/neurips/One More Step Towards Reality: Cooperative Bandits with Imperfect Communication create mode 100644 data/2021/neurips/One Question Answering Model for Many Languages with Cross-lingual Dense Passage Retrieval create mode 100644 data/2021/neurips/Online Active Learning with Surrogate Loss Functions create mode 100644 data/2021/neurips/Online Adaptation to Label Distribution Shift create mode 100644 data/2021/neurips/Online Control of Unknown Time-Varying Dynamical Systems create mode 100644 data/2021/neurips/Online Convex Optimization with Continuous Switching Constraint create mode 100644 data/2021/neurips/Online Facility Location with Multiple Advice create mode 100644 data/2021/neurips/Online Knapsack with Frequency Predictions create mode 100644 data/2021/neurips/Online Learning Of Neural Computations From Sparse Temporal Feedback create mode 100644 data/2021/neurips/Online Learning and Control of Complex Dynamical Systems from Sensory Input create mode 100644 data/2021/neurips/Online Learning in Periodic Zero-Sum Games create mode 100644 data/2021/neurips/Online Market Equilibrium with Application to Fair Division create mode 100644 data/2021/neurips/Online Matching in Sparse Random Graphs: Non-Asymptotic Performances of Greedy Algorithm create mode 100644 data/2021/neurips/Online Multi-Armed Bandits with Adaptive Inference create mode 100644 data/2021/neurips/Online Robust Reinforcement Learning with Model Uncertainty create mode 100644 data/2021/neurips/Online Selective Classification with Limited Feedback create mode 100644 data/2021/neurips/Online Sign Identification: Minimization of the Number of Errors in Thresholding Bandits create mode 100644 data/2021/neurips/Online Variational Filtering and Parameter Learning create mode 100644 data/2021/neurips/Online and Offline Reinforcement Learning by Planning with a Learned Model create mode 100644 data/2021/neurips/Online false discovery rate control for anomaly detection in time series create mode 100644 data/2021/neurips/Online learning in MDPs with linear function approximation and bandit feedback create mode 100644 data/2021/neurips/Only Train Once: A One-Shot Neural Network Training And Pruning Framework create mode 100644 data/2021/neurips/Open Rule Induction create mode 100644 data/2021/neurips/Open-set Label Noise Can Improve Robustness Against Inherent Label Noise create mode 100644 data/2021/neurips/OpenMatch: Open-Set Semi-supervised Learning with Open-set Consistency Regularization create mode 100644 data/2021/neurips/Optimal Algorithms for Stochastic Contextual Preference Bandits create mode 100644 data/2021/neurips/Optimal Best-Arm Identification Methods for Tail-Risk Measures create mode 100644 data/2021/neurips/Optimal Gradient-based Algorithms for Non-concave Bandit Optimization create mode 100644 data/2021/neurips/Optimal Order Simple Regret for Gaussian Process Bandits create mode 100644 data/2021/neurips/Optimal Policies Tend To Seek Power create mode 100644 data/2021/neurips/Optimal Rates for Nonparametric Density Estimation under Communication Constraints create mode 100644 data/2021/neurips/Optimal Rates for Random Order Online Optimization create mode 100644 data/2021/neurips/Optimal Sketching for Trace Estimation create mode 100644 data/2021/neurips/Optimal Underdamped Langevin MCMC Method create mode 100644 data/2021/neurips/Optimal Uniform OPE and Model-based Offline Reinforcement Learning in Time-Homogeneous, Reward-Free and Task-Agnostic Settings create mode 100644 data/2021/neurips/Optimal prediction of Markov chains with and without spectral gap create mode 100644 data/2021/neurips/Optimality and Stability in Federated Learning: A Game-theoretic Approach create mode 100644 data/2021/neurips/Optimality of variational inference for stochasticblock model with missing links create mode 100644 data/2021/neurips/Optimization-Based Algebraic Multigrid Coarsening Using Reinforcement Learning create mode 100644 data/2021/neurips/Optimizing Conditional Value-At-Risk of Black-Box Functions create mode 100644 data/2021/neurips/Optimizing Information-theoretical Generalization Bound via Anisotropic Noise of SGLD create mode 100644 data/2021/neurips/Optimizing Reusable Knowledge for Continual Learning via Metalearning create mode 100644 data/2021/neurips/Oracle Complexity in Nonsmooth Nonconvex Optimization create mode 100644 data/2021/neurips/Oracle-Efficient Regret Minimization in Factored MDPs with Unknown Structure create mode 100644 data/2021/neurips/Out-of-Distribution Generalization in Kernel Regression create mode 100644 data/2021/neurips/Outcome-Driven Reinforcement Learning via Variational Inference create mode 100644 data/2021/neurips/Overcoming Catastrophic Forgetting in Incremental Few-Shot Learning by Finding Flat Minima create mode 100644 data/2021/neurips/Overcoming the Convex Barrier for Simplex Inputs create mode 100644 data/2021/neurips/Overcoming the curse of dimensionality with Laplacian regularization in semi-supervised learning create mode 100644 data/2021/neurips/Overinterpretation reveals image classification model pathologies create mode 100644 data/2021/neurips/Overparameterization Improves Robustness to Covariate Shift in High Dimensions create mode 100644 data/2021/neurips/PCA Initialization for Approximate Message Passing in Rotationally Invariant Models create mode 100644 data/2021/neurips/PDE-GCN: Novel Architectures for Graph Neural Networks Motivated by Partial Differential Equations create mode 100644 data/2021/neurips/PLUGIn: A simple algorithm for inverting generative models with recovery guarantees create mode 100644 data/2021/neurips/PLUR: A Unifying, Graph-Based View of Program Learning, Understanding, and Repair create mode 100644 data/2021/neurips/POODLE: Improving Few-shot Learning via Penalizing Out-of-Distribution Samples create mode 100644 data/2021/neurips/PSD Representations for Effective Probability Models create mode 100644 data/2021/neurips/PTR: A Benchmark for Part-based Conceptual, Relational, and Physical Reasoning create mode 100644 data/2021/neurips/Panoptic 3D Scene Reconstruction From a Single RGB Image create mode 100644 data/2021/neurips/ParK: Sound and Efficient Kernel Ridge Regression by Feature Space Partitions create mode 100644 data/2021/neurips/Parallel Bayesian Optimization of Multiple Noisy Objectives with Expected Hypervolume Improvement create mode 100644 data/2021/neurips/Parallel and Efficient Hierarchical k-Median Clustering create mode 100644 data/2021/neurips/Parallelizing Thompson Sampling create mode 100644 data/2021/neurips/Parameter Inference with Bifurcation Diagrams create mode 100644 data/2021/neurips/Parameter Prediction for Unseen Deep Architectures create mode 100644 data/2021/neurips/Parameter-free HE-friendly Logistic Regression create mode 100644 data/2021/neurips/Parameterized Knowledge Transfer for Personalized Federated Learning create mode 100644 data/2021/neurips/Parametric Complexity Bounds for Approximating PDEs with Neural Networks create mode 100644 data/2021/neurips/Parametrized Quantum Policies for Reinforcement Learning create mode 100644 data/2021/neurips/Pareto-Optimal Learning-Augmented Algorithms for Online Conversion Problems create mode 100644 data/2021/neurips/Partial success in closing the gap between human and machine vision create mode 100644 data/2021/neurips/PartialFed: Cross-Domain Personalized Federated Learning via Partial Initialization create mode 100644 data/2021/neurips/Particle Cloud Generation with Message Passing Generative Adversarial Networks create mode 100644 data/2021/neurips/Particle Dual Averaging: Optimization of Mean Field Neural Network with Global Convergence Rate Analysis create mode 100644 data/2021/neurips/Partition and Code: learning how to compress graphs create mode 100644 data/2021/neurips/Partition-Based Formulations for Mixed-Integer Optimization of Trained ReLU Neural Networks create mode 100644 data/2021/neurips/Passive attention in artificial neural networks predicts human visual selectivity create mode 100644 data/2021/neurips/PatchGame: Learning to Signal Mid-level Patches in Referential Games create mode 100644 data/2021/neurips/Pay Attention to MLPs create mode 100644 data/2021/neurips/Pay Better Attention to Attention: Head Selection in Multilingual and Multi-Domain Sequence Modeling create mode 100644 data/2021/neurips/Per-Pixel Classification is Not All You Need for Semantic Segmentation create mode 100644 data/2021/neurips/PerSim: Data-Efficient Offline Reinforcement Learning with Heterogeneous Agents via Personalized Simulators create mode 100644 data/2021/neurips/Perceptual Score: What Data Modalities Does Your Model Perceive? create mode 100644 data/2021/neurips/Periodic Activation Functions Induce Stationarity create mode 100644 data/2021/neurips/Permutation-Invariant Variational Autoencoder for Graph-Level Representation Learning create mode 100644 data/2021/neurips/Permuton-induced Chinese Restaurant Process create mode 100644 data/2021/neurips/Personalized Federated Learning With Gaussian Processes create mode 100644 data/2021/neurips/Perturb-and-max-product: Sampling and learning in discrete energy-based models create mode 100644 data/2021/neurips/Perturbation Theory for the Information Bottleneck create mode 100644 data/2021/neurips/Perturbation-based Regret Analysis of Predictive Control in Linear Time Varying Systems create mode 100644 data/2021/neurips/Pessimism Meets Invariance: Provably Efficient Offline Mean-Field Multi-Agent RL create mode 100644 data/2021/neurips/PettingZoo: Gym for Multi-Agent Reinforcement Learning create mode 100644 data/2021/neurips/Photonic Differential Privacy with Direct Feedback Alignment create mode 100644 data/2021/neurips/Physics-Aware Downsampling with Deep Learning for Scalable Flood Modeling create mode 100644 data/2021/neurips/Physics-Integrated Variational Autoencoders for Robust and Interpretable Generative Modeling create mode 100644 data/2021/neurips/PiRank: Scalable Learning To Rank via Differentiable Sorting create mode 100644 data/2021/neurips/Pipeline Combinators for Gradual AutoML create mode 100644 data/2021/neurips/Piper: Multidimensional Planner for DNN Parallelization create mode 100644 data/2021/neurips/Planning from Pixels in Environments with Combinatorially Hard Search Spaces create mode 100644 data/2021/neurips/Play to Grade: Testing Coding Games as Classifying Markov Decision Process create mode 100644 data/2021/neurips/PlayVirtual: Augmenting Cycle-Consistent Virtual Trajectories for Reinforcement Learning create mode 100644 data/2021/neurips/Pointwise Bounds for Distribution Estimation under Communication Constraints create mode 100644 data/2021/neurips/PolarStream: Streaming Object Detection and Segmentation with Polar Pillars create mode 100644 data/2021/neurips/Policy Finetuning: Bridging Sample-Efficient Offline and Online Reinforcement Learning create mode 100644 data/2021/neurips/Policy Learning Using Weak Supervision create mode 100644 data/2021/neurips/Policy Optimization in Adversarial MDPs: Improved Exploration via Dilated Bonuses create mode 100644 data/2021/neurips/Pooling by Sliced-Wasserstein Embedding create mode 100644 data/2021/neurips/PortaSpeech: Portable and High-Quality Generative Text-to-Speech create mode 100644 data/2021/neurips/Post-Contextual-Bandit Inference create mode 100644 data/2021/neurips/Post-Training Quantization for Vision Transformer create mode 100644 data/2021/neurips/Post-Training Sparsity-Aware Quantization create mode 100644 data/2021/neurips/Post-processing for Individual Fairness create mode 100644 data/2021/neurips/Posterior Collapse and Latent Variable Non-identifiability create mode 100644 data/2021/neurips/Posterior Meta-Replay for Continual Learning create mode 100644 data/2021/neurips/Powerpropagation: A sparsity inducing weight reparameterisation create mode 100644 data/2021/neurips/Practical Large-Scale Linear Programming using Primal-Dual Hybrid Gradient create mode 100644 data/2021/neurips/Practical Near Neighbor Search via Group Testing create mode 100644 data/2021/neurips/Practical, Provably-Correct Interactive Learning in the Realizable Setting: The Power of True Believers create mode 100644 data/2021/neurips/Pragmatic Image Compression for Human-in-the-Loop Decision-Making create mode 100644 data/2021/neurips/Precise characterization of the prior predictive distribution of deep ReLU networks create mode 100644 data/2021/neurips/Preconditioned Gradient Descent for Over-Parameterized Nonconvex Matrix Factorization create mode 100644 data/2021/neurips/Predicting Deep Neural Network Generalization with Perturbation Response Curves create mode 100644 data/2021/neurips/Predicting Event Memorability from Contextual Visual Semantics create mode 100644 data/2021/neurips/Predicting Molecular Conformation via Dynamic Graph Score Matching create mode 100644 data/2021/neurips/Predicting What You Already Know Helps: Provable Self-Supervised Learning create mode 100644 data/2021/neurips/Predify: Augmenting deep neural networks with brain-inspired predictive coding dynamics create mode 100644 data/2021/neurips/PreferenceNet: Encoding Human Preferences in Auction Design with Deep Learning create mode 100644 data/2021/neurips/Preserved central model for faster bidirectional compression in distributed settings create mode 100644 data/2021/neurips/Pretraining Representations for Data-Efficient Reinforcement Learning create mode 100644 data/2021/neurips/Prior-independent Dynamic Auctions for a Value-maximizing Buyer create mode 100644 data/2021/neurips/Private Non-smooth ERM and SCO in Subquadratic Steps create mode 100644 data/2021/neurips/Private and Non-private Uniformity Testing for Ranking Data create mode 100644 data/2021/neurips/Private learning implies quantum stability create mode 100644 data/2021/neurips/Privately Learning Mixtures of Axis-Aligned Gaussians create mode 100644 data/2021/neurips/Privately Learning Subspaces create mode 100644 data/2021/neurips/Privately Publishable Per-instance Privacy create mode 100644 data/2021/neurips/ProTo: Program-Guided Transformer for Program-Guided Tasks create mode 100644 data/2021/neurips/Probabilistic Attention for Interactive Segmentation create mode 100644 data/2021/neurips/Probabilistic Entity Representation Model for Reasoning over Knowledge Graphs create mode 100644 data/2021/neurips/Probabilistic Forecasting: A Level-Set Approach create mode 100644 data/2021/neurips/Probabilistic Margins for Instance Reweighting in Adversarial Training create mode 100644 data/2021/neurips/Probabilistic Tensor Decomposition of Neural Population Spiking Activity create mode 100644 data/2021/neurips/Probabilistic Transformer For Time Series Analysis create mode 100644 data/2021/neurips/Probability Paths and the Structure of Predictions over Time create mode 100644 data/2021/neurips/Probing Inter-modality: Visual Parsing with Self-Attention for Vision-and-Language Pre-training create mode 100644 data/2021/neurips/Process for Adapting Language Models to Society (PALMS) with Values-Targeted Datasets create mode 100644 data/2021/neurips/Profiling Pareto Front With Multi-Objective Stein Variational Gradient Descent create mode 100644 data/2021/neurips/Program Synthesis Guided Reinforcement Learning for Partially Observed Environments create mode 100644 data/2021/neurips/Progressive Coordinate Transforms for Monocular 3D Object Detection create mode 100644 data/2021/neurips/Progressive Feature Interaction Search for Deep Sparse Network create mode 100644 data/2021/neurips/Projected GANs Converge Faster create mode 100644 data/2021/neurips/Proper Value Equivalence create mode 100644 data/2021/neurips/Property-Aware Relation Networks for Few-Shot Molecular Property Prediction create mode 100644 data/2021/neurips/Proportional Participatory Budgeting with Additive Utilities create mode 100644 data/2021/neurips/Prototypical Cross-Attention Networks for Multiple Object Tracking and Segmentation create mode 100644 data/2021/neurips/Provable Benefits of Actor-Critic Methods for Offline Reinforcement Learning create mode 100644 data/2021/neurips/Provable Guarantees for Self-Supervised Deep Learning with Spectral Contrastive Loss create mode 100644 data/2021/neurips/Provable Model-based Nonlinear Bandit and Reinforcement Learning: Shelve Optimism, Embrace Virtual Curvature create mode 100644 data/2021/neurips/Provable Representation Learning for Imitation with Contrastive Fourier Features create mode 100644 data/2021/neurips/Provably Efficient Black-Box Action Poisoning Attacks Against Reinforcement Learning create mode 100644 data/2021/neurips/Provably Efficient Causal Reinforcement Learning with Confounded Observational Data create mode 100644 data/2021/neurips/Provably Efficient Reinforcement Learning with Linear Function Approximation under Adaptivity Constraints create mode 100644 data/2021/neurips/Provably Faster Algorithms for Bilevel Optimization create mode 100644 data/2021/neurips/Provably Strict Generalisation Benefit for Invariance in Kernel Methods create mode 100644 data/2021/neurips/Provably efficient multi-task reinforcement learning with model transfer create mode 100644 data/2021/neurips/Provably efficient, succinct, and precise explanations create mode 100644 data/2021/neurips/Proxy Convexity: A Unified Framework for the Analysis of Neural Networks Trained by Gradient Descent create mode 100644 data/2021/neurips/Proxy-Normalizing Activations to Match Batch Normalization while Removing Batch Dependence create mode 100644 data/2021/neurips/Pruning Randomly Initialized Neural Networks with Iterative Randomization create mode 100644 data/2021/neurips/Pseudo-Spherical Contrastive Divergence create mode 100644 data/2021/neurips/Pure Exploration in Kernel and Neural Bandits create mode 100644 data/2021/neurips/Qimera: Data-free Quantization with Synthetic Boundary Supporting Samples create mode 100644 data/2021/neurips/Qu-ANTI-zation: Exploiting Quantization Artifacts for Achieving Adversarial Outcomes create mode 100644 data/2021/neurips/QuPeD: Quantized Personalization via Distillation with Applications to Federated Learning create mode 100644 data/2021/neurips/Quantifying and Improving Transferability in Domain Generalization create mode 100644 data/2021/neurips/R-Drop: Regularized Dropout for Neural Networks create mode 100644 data/2021/neurips/RED : Looking for Redundancies for Data-FreeStructured Compression of Deep Neural Networks create mode 100644 data/2021/neurips/REMIPS: Physically Consistent 3D Reconstruction of Multiple Interacting People under Weak Supervision create mode 100644 data/2021/neurips/RETRIEVE: Coreset Selection for Efficient and Robust Semi-Supervised Learning create mode 100644 data/2021/neurips/RIM: Reliable Influence-based Active Learning on Graphs create mode 100644 data/2021/neurips/RL for Latent MDPs: Regret Guarantees and a Lower Bound create mode 100644 data/2021/neurips/RLlib Flow: Distributed Reinforcement Learning is a Dataflow Problem create mode 100644 data/2021/neurips/RMIX: Learning Risk-Sensitive Policies for Cooperative Reinforcement Learning Agents create mode 100644 data/2021/neurips/RMM: Reinforced Memory Management for Class-Incremental Learning create mode 100644 data/2021/neurips/Random Noise Defense Against Query-Based Black-Box Attacks create mode 100644 data/2021/neurips/Random Shuffling Beats SGD Only After Many Epochs on Ill-Conditioned Problems create mode 100644 data/2021/neurips/Rank Overspecified Robust Matrix Recovery: Subgradient Method and Exact Recovery create mode 100644 data/2021/neurips/Ranking Policy Decisions create mode 100644 data/2021/neurips/Rate-Optimal Subspace Estimation on Random Graphs create mode 100644 data/2021/neurips/Rates of Estimation of Optimal Transport Maps using Plug-in Estimators via Barycentric Projections create mode 100644 data/2021/neurips/Raw Nav-merge Seismic Data to Subsurface Properties with MLP based Multi-Modal Information Unscrambler create mode 100644 data/2021/neurips/Re-ranking for image retrieval and transductive few-shot classification create mode 100644 data/2021/neurips/ReAct: Out-of-distribution Detection With Rectified Activations create mode 100644 data/2021/neurips/ReLU Regression with Massart Noise create mode 100644 data/2021/neurips/ReSSL: Relational Self-Supervised Learning with Weak Augmentation create mode 100644 data/2021/neurips/Realistic evaluation of transductive few-shot learning create mode 100644 data/2021/neurips/Rebooting ACGAN: Auxiliary Classifier GANs with Stable Training create mode 100644 data/2021/neurips/Rebounding Bandits for Modeling Satiation Effects create mode 100644 data/2021/neurips/Recognizing Vector Graphics without Rasterization create mode 100644 data/2021/neurips/Reconstruction for Powerful Graph Representations create mode 100644 data/2021/neurips/Recovering Latent Causal Factor for Generalization to Distributional Shifts create mode 100644 data/2021/neurips/Recovery Analysis for Plug-and-Play Priors using the Restricted Eigenvalue Condition create mode 100644 data/2021/neurips/Rectangular Flows for Manifold Learning create mode 100644 data/2021/neurips/Rectifying the Shortcut Learning of Background for Few-Shot Learning create mode 100644 data/2021/neurips/Recurrence along Depth: Deep Convolutional Neural Networks with Recurrent Layer Aggregation create mode 100644 data/2021/neurips/Recurrent Bayesian Classifier Chains for Exact Multi-Label Classification create mode 100644 data/2021/neurips/Recurrent Submodular Welfare and Matroid Blocking Semi-Bandits create mode 100644 data/2021/neurips/Recursive Bayesian Networks: Generalising and Unifying Probabilistic Context-Free Grammars and Dynamic Bayesian Networks create mode 100644 data/2021/neurips/Recursive Causal Structure Learning in the Presence of Latent Variables and Selection Bias create mode 100644 data/2021/neurips/Redesigning the Transformer Architecture with Insights from Multi-particle Dynamical Systems create mode 100644 data/2021/neurips/Reducing Collision Checking for Sampling-Based Motion Planning Using Graph Neural Networks create mode 100644 data/2021/neurips/Reducing Information Bottleneck for Weakly Supervised Semantic Segmentation create mode 100644 data/2021/neurips/Reducing the Covariate Shift by Mirror Samples in Cross Domain Alignment create mode 100644 data/2021/neurips/Referring Transformer: A One-step Approach to Multi-task Visual Grounding create mode 100644 data/2021/neurips/Refined Learning Bounds for Kernel and Approximate $k$-Means create mode 100644 data/2021/neurips/Refining Language Models with Compositional Explanations create mode 100644 data/2021/neurips/Reformulating Zero-shot Action Recognition for Multi-label Actions create mode 100644 data/2021/neurips/Regime Switching Bandits create mode 100644 data/2021/neurips/Regret Bounds for Gaussian-Process Optimization in Large Domains create mode 100644 data/2021/neurips/Regret Minimization Experience Replay in Off-Policy Reinforcement Learning create mode 100644 data/2021/neurips/Regularization in ResNet with Stochastic Depth create mode 100644 data/2021/neurips/Regularized Frank-Wolfe for Dense CRFs: Generalizing Mean Field and Beyond create mode 100644 data/2021/neurips/Regularized Softmax Deep Multi-Agent Q-Learning create mode 100644 data/2021/neurips/Regulating algorithmic filtering on social media create mode 100644 data/2021/neurips/Reinforced Few-Shot Acquisition Function Learning for Bayesian Optimization create mode 100644 data/2021/neurips/Reinforcement Learning Enhanced Explainer for Graph Neural Networks create mode 100644 data/2021/neurips/Reinforcement Learning based Disease Progression Model for Alzheimer's Disease create mode 100644 data/2021/neurips/Reinforcement Learning in Linear MDPs: Constant Regret and Representation Selection create mode 100644 data/2021/neurips/Reinforcement Learning in Newcomblike Environments create mode 100644 data/2021/neurips/Reinforcement Learning in Reward-Mixing MDPs create mode 100644 data/2021/neurips/Reinforcement Learning with Latent Flow create mode 100644 data/2021/neurips/Reinforcement Learning with State Observation Costs in Action-Contingent Noiselessly Observable Markov Decision Processes create mode 100644 data/2021/neurips/Reinforcement learning for optimization of variational quantum circuit architectures create mode 100644 data/2021/neurips/Relational Self-Attention: What's Missing in Attention for Video Understanding create mode 100644 data/2021/neurips/Relative Flatness and Generalization create mode 100644 data/2021/neurips/Relative Uncertainty Learning for Facial Expression Recognition create mode 100644 data/2021/neurips/Relative stability toward diffeomorphisms indicates performance in deep nets create mode 100644 data/2021/neurips/Relaxed Marginal Consistency for Differentially Private Query Answering create mode 100644 data/2021/neurips/Relaxing Local Robustness create mode 100644 data/2021/neurips/RelaySum for Decentralized Deep Learning on Heterogeneous Data create mode 100644 data/2021/neurips/Reliable Causal Discovery with Improved Exact Search and Weaker Assumptions create mode 100644 data/2021/neurips/Reliable Decisions with Threshold Calibration create mode 100644 data/2021/neurips/Reliable Estimation of KL Divergence using a Discriminator in Reproducing Kernel Hilbert Space create mode 100644 data/2021/neurips/Reliable Post hoc Explanations: Modeling Uncertainty in Explainability create mode 100644 data/2021/neurips/Reliable and Trustworthy Machine Learning for Health Using Dataset Shift Detection create mode 100644 data/2021/neurips/Remember What You Want to Forget: Algorithms for Machine Unlearning create mode 100644 data/2021/neurips/Removing Inter-Experimental Variability from Functional Data in Systems Neuroscience create mode 100644 data/2021/neurips/Renyi Differential Privacy of The Subsampled Shuffle Model In Distributed Learning create mode 100644 data/2021/neurips/Replacing Rewards with Examples: Example-Based Policy Search via Recursive Classification create mode 100644 data/2021/neurips/Replay-Guided Adversarial Environment Design create mode 100644 data/2021/neurips/Representation Costs of Linear Neural Networks: Analysis and Design create mode 100644 data/2021/neurips/Representation Learning Beyond Linear Prediction Functions create mode 100644 data/2021/neurips/Representation Learning for Event-based Visuomotor Policies create mode 100644 data/2021/neurips/Representation Learning on Spatial Networks create mode 100644 data/2021/neurips/Representer Point Selection via Local Jacobian Expansion for Post-hoc Classifier Explanation of Deep Neural Networks and Ensemble Models create mode 100644 data/2021/neurips/Representing Hyperbolic Space Accurately using Multi-Component Floats create mode 100644 data/2021/neurips/Representing Long-Range Context for Graph Neural Networks with Global Attention create mode 100644 data/2021/neurips/Repulsive Deep Ensembles are Bayesian create mode 100644 data/2021/neurips/ResNEsts and DenseNEsts: Block-based DNN Models with Improved Representation Guarantees create mode 100644 data/2021/neurips/ResT: An Efficient Transformer for Visual Recognition create mode 100644 data/2021/neurips/Residual Pathway Priors for Soft Equivariance Constraints create mode 100644 data/2021/neurips/Residual Relaxation for Multi-view Representation Learning create mode 100644 data/2021/neurips/Residual2Vec: Debiasing graph embedding with random graphs create mode 100644 data/2021/neurips/Rethinking Calibration of Deep Neural Networks: Do Not Be Afraid of Overconfidence create mode 100644 data/2021/neurips/Rethinking Graph Transformers with Spectral Attention create mode 100644 data/2021/neurips/Rethinking Neural Operations for Diverse Tasks create mode 100644 data/2021/neurips/Rethinking Space-Time Networks with Improved Memory Coverage for Efficient Video Object Segmentation create mode 100644 data/2021/neurips/Rethinking and Reweighting the Univariate Losses for Multi-Label Ranking: Consistency and Generalization create mode 100644 data/2021/neurips/Rethinking conditional GAN training: An approach using geometrically structured latent manifolds create mode 100644 data/2021/neurips/Rethinking gradient sparsification as total error minimization create mode 100644 data/2021/neurips/Rethinking the Pruning Criteria for Convolutional Neural Network create mode 100644 data/2021/neurips/Rethinking the Variational Interpretation of Accelerated Optimization Methods create mode 100644 data/2021/neurips/Retiring Adult: New Datasets for Fair Machine Learning create mode 100644 data/2021/neurips/Reusing Combinatorial Structure: Faster Iterative Projections over Submodular Base Polytopes create mode 100644 data/2021/neurips/Revealing and Protecting Labels in Distributed Training create mode 100644 data/2021/neurips/Revenue maximization via machine learning with noisy data create mode 100644 data/2021/neurips/Reverse engineering learned optimizers reveals known and novel mechanisms create mode 100644 data/2021/neurips/Reverse engineering recurrent neural networks with Jacobian switching linear dynamical systems create mode 100644 data/2021/neurips/Reverse-Complement Equivariant Networks for DNA Sequences create mode 100644 data/2021/neurips/Revisit Multimodal Meta-Learning through the Lens of Multi-Task Learning create mode 100644 data/2021/neurips/Revisiting 3D Object Detection From an Egocentric Perspective create mode 100644 data/2021/neurips/Revisiting Contrastive Methods for Unsupervised Learning of Visual Representations create mode 100644 data/2021/neurips/Revisiting Deep Learning Models for Tabular Data create mode 100644 data/2021/neurips/Revisiting Discriminator in GAN Compression: A Generator-discriminator Cooperative Compression Scheme create mode 100644 data/2021/neurips/Revisiting Hilbert-Schmidt Information Bottleneck for Adversarial Robustness create mode 100644 data/2021/neurips/Revisiting Model Stitching to Compare Neural Representations create mode 100644 data/2021/neurips/Revisiting ResNets: Improved Training and Scaling Strategies create mode 100644 data/2021/neurips/Revisiting Smoothed Online Learning create mode 100644 data/2021/neurips/Revisiting the Calibration of Modern Neural Networks create mode 100644 data/2021/neurips/Revitalizing CNN Attention via Transformers in Self-Supervised Visual Representation Learning create mode 100644 data/2021/neurips/Reward is enough for convex MDPs create mode 100644 data/2021/neurips/Reward-Free Model-Based Reinforcement Learning with Linear Function Approximation create mode 100644 data/2021/neurips/Risk Bounds and Calibration for a Smart Predict-then-Optimize Method create mode 100644 data/2021/neurips/Risk Bounds for Over-parameterized Maximum Margin Classification on Sub-Gaussian Mixtures create mode 100644 data/2021/neurips/Risk Minimization from Adaptively Collected Data: Guarantees for Supervised and Policy Learning create mode 100644 data/2021/neurips/Risk Monotonicity in Statistical Learning create mode 100644 data/2021/neurips/Risk-Averse Bayes-Adaptive Reinforcement Learning create mode 100644 data/2021/neurips/Risk-Aware Transfer in Reinforcement Learning using Successor Features create mode 100644 data/2021/neurips/Risk-averse Heteroscedastic Bayesian Optimization create mode 100644 data/2021/neurips/RoMA: Robust Model Adaptation for Offline Model-based Optimization create mode 100644 data/2021/neurips/Robust Allocations with Diversity Constraints create mode 100644 data/2021/neurips/Robust Auction Design in the Auto-bidding World create mode 100644 data/2021/neurips/Robust Compressed Sensing MRI with Deep Generative Priors create mode 100644 data/2021/neurips/Robust Contrastive Learning Using Negative Samples with Diminished Semantics create mode 100644 data/2021/neurips/Robust Counterfactual Explanations on Graph Neural Networks create mode 100644 data/2021/neurips/Robust Deep Reinforcement Learning through Adversarial Loss create mode 100644 data/2021/neurips/Robust Generalization despite Distribution Shift via Minimum Discriminating Information create mode 100644 data/2021/neurips/Robust Implicit Networks via Non-Euclidean Contractions create mode 100644 data/2021/neurips/Robust Inverse Reinforcement Learning under Transition Dynamics Mismatch create mode 100644 data/2021/neurips/Robust Learning of Optimal Auctions create mode 100644 data/2021/neurips/Robust Online Correlation Clustering create mode 100644 data/2021/neurips/Robust Optimization for Multilingual Translation with Imbalanced Data create mode 100644 data/2021/neurips/Robust Pose Estimation in Crowded Scenes with Direct Pose-Level Inference create mode 100644 data/2021/neurips/Robust Predictable Control create mode 100644 data/2021/neurips/Robust Regression Revisited: Acceleration and Improved Estimation Rates create mode 100644 data/2021/neurips/Robust Visual Reasoning via Language Guided Neural Module Networks create mode 100644 data/2021/neurips/Robust and Decomposable Average Precision for Image Retrieval create mode 100644 data/2021/neurips/Robust and Fully-Dynamic Coreset for Continuous-and-Bounded Learning (With Outliers) Problems create mode 100644 data/2021/neurips/Robust and differentially private mean estimation create mode 100644 data/2021/neurips/Robustifying Algorithms of Learning Latent Trees with Vector Variables create mode 100644 data/2021/neurips/Robustness between the worst and average case create mode 100644 data/2021/neurips/Robustness of Graph Neural Networks at Scale create mode 100644 data/2021/neurips/Robustness via Uncertainty-aware Cycle Consistency create mode 100644 data/2021/neurips/Rot-Pro: Modeling Transitivity by Projection in Knowledge Graph Embedding create mode 100644 data/2021/neurips/Roto-translated Local Coordinate Frames For Interacting Dynamical Systems create mode 100644 data/2021/neurips/Row-clustering of a Point Process-valued Matrix create mode 100644 data/2021/neurips/SADGA: Structure-Aware Dual Graph Aggregation Network for Text-to-SQL create mode 100644 data/2021/neurips/SAPE: Spatially-Adaptive Progressive Encoding for Neural Optimization create mode 100644 data/2021/neurips/SBO-RNN: Reformulating Recurrent Neural Networks via Stochastic Bilevel Optimization create mode 100644 data/2021/neurips/SE(3)-equivariant prediction of molecular wavefunctions and electronic densities create mode 100644 data/2021/neurips/SEAL: Self-supervised Embodied Active Learning using Exploration and 3D Consistency create mode 100644 data/2021/neurips/SGD: The Role of Implicit Regularization, Batch-size and Multiple-epochs create mode 100644 data/2021/neurips/SILG: The Multi-domain Symbolic Interactive Language Grounding Benchmark create mode 100644 data/2021/neurips/SIMILAR: Submodular Information Measures Based Active Learning In Realistic Scenarios create mode 100644 data/2021/neurips/SIMONe: View-Invariant, Temporally-Abstracted Object Representations via Unsupervised Video Decomposition create mode 100644 data/2021/neurips/SLAPS: Self-Supervision Improves Structure Learning for Graph Neural Networks create mode 100644 data/2021/neurips/SLOE: A Faster Method for Statistical Inference in High-Dimensional Logistic Regression create mode 100644 data/2021/neurips/SNIPS: Solving Noisy Inverse Problems Stochastically create mode 100644 data/2021/neurips/SOAT: A Scene- and Object-Aware Transformer for Vision-and-Language Navigation create mode 100644 data/2021/neurips/SOFT: Softmax-free Transformer with Linear Complexity create mode 100644 data/2021/neurips/SOLQ: Segmenting Objects by Learning Queries create mode 100644 data/2021/neurips/SOPE: Spectrum of Off-Policy Estimators create mode 100644 data/2021/neurips/SPANN: Highly-efficient Billion-scale Approximate Nearest Neighborhood Search create mode 100644 data/2021/neurips/SQALER: Scaling Question Answering by Decoupling Multi-Hop and Logical Reasoning create mode 100644 data/2021/neurips/SSAL: Synergizing between Self-Training and Adversarial Learning for Domain Adaptive Object Detection create mode 100644 data/2021/neurips/SSMF: Shifting Seasonal Matrix Factorization create mode 100644 data/2021/neurips/SSUL: Semantic Segmentation with Unknown Label for Exemplar-based Class-Incremental Learning create mode 100644 data/2021/neurips/STEM: A Stochastic Two-Sided Momentum Algorithm Achieving Near-Optimal Sample and Communication Complexities for Federated Learning create mode 100644 data/2021/neurips/STEP: Out-of-Distribution Detection in the Presence of Limited In-Distribution Labeled Data create mode 100644 data/2021/neurips/STORM+: Fully Adaptive SGD with Recursive Momentum for Nonconvex Optimization create mode 100644 data/2021/neurips/SUPER-ADAM: Faster and Universal Framework of Adaptive Gradients create mode 100644 data/2021/neurips/SWAD: Domain Generalization by Seeking Flat Minima create mode 100644 data/2021/neurips/Safe Policy Optimization with Local Generalized Linear Function Approximations create mode 100644 data/2021/neurips/Safe Pontryagin Differentiable Programming create mode 100644 data/2021/neurips/Safe Reinforcement Learning by Imagining the Near Future create mode 100644 data/2021/neurips/Safe Reinforcement Learning with Natural Language Constraints create mode 100644 data/2021/neurips/Sageflow: Robust Federated Learning against Both Stragglers and Adversaries create mode 100644 data/2021/neurips/SalKG: Learning From Knowledge Graph Explanations for Commonsense Reasoning create mode 100644 data/2021/neurips/Sample Complexity Bounds for Active Ranking from Multi-wise Comparisons create mode 100644 data/2021/neurips/Sample Complexity of Tree Search Configuration: Cutting Planes and Beyond create mode 100644 data/2021/neurips/Sample Selection for Fair and Robust Training create mode 100644 data/2021/neurips/Sample-Efficient Learning of Stackelberg Equilibria in General-Sum Games create mode 100644 data/2021/neurips/Sample-Efficient Reinforcement Learning Is Feasible for Linearly Realizable MDPs with Limited Revisiting create mode 100644 data/2021/neurips/Sample-Efficient Reinforcement Learning for Linearly-Parameterized MDPs with a Generative Model create mode 100644 data/2021/neurips/Sanity Checks for Lottery Tickets: Does Your Winning Ticket Really Win the Jackpot? create mode 100644 data/2021/neurips/Scalable Bayesian GPFA with automatic relevance determination and discrete noise models create mode 100644 data/2021/neurips/Scalable Diverse Model Selection for Accessible Transfer Learning create mode 100644 data/2021/neurips/Scalable Inference in SDEs by Direct Matching of the Fokker-Planck-Kolmogorov Equation create mode 100644 data/2021/neurips/Scalable Inference of Sparsely-changing Gaussian Markov Random Fields create mode 100644 data/2021/neurips/Scalable Intervention Target Estimation in Linear Models create mode 100644 data/2021/neurips/Scalable Neural Data Server: A Data Recommender for Transfer Learning create mode 100644 data/2021/neurips/Scalable Online Planning via Reinforcement Learning Fine-Tuning create mode 100644 data/2021/neurips/Scalable Quasi-Bayesian Inference for Instrumental Variable Regression create mode 100644 data/2021/neurips/Scalable Rule-Based Representation Learning for Interpretable Classification create mode 100644 data/2021/neurips/Scalable Thompson Sampling using Sparse Gaussian Process Models create mode 100644 data/2021/neurips/Scalable and Stable Surrogates for Flexible Classifiers with Fairness Constraints create mode 100644 data/2021/neurips/Scalars are universal: Equivariant machine learning, structured like classical physics create mode 100644 data/2021/neurips/ScaleCert: Scalable Certified Defense against Adversarial Patches with Sparse Superficial Layers create mode 100644 data/2021/neurips/Scaling Ensemble Distribution Distillation to Many Classes with Proxy Targets create mode 100644 data/2021/neurips/Scaling Gaussian Processes with Derivative Information Using Variational Inference create mode 100644 data/2021/neurips/Scaling Neural Tangent Kernels via Sketching and Random Features create mode 100644 data/2021/neurips/Scaling Up Exact Neural Network Compression by ReLU Stability create mode 100644 data/2021/neurips/Scaling Vision with Sparse Mixture of Experts create mode 100644 data/2021/neurips/Scaling up Continuous-Time Markov Chains Helps Resolve Underspecification create mode 100644 data/2021/neurips/Scallop: From Probabilistic Deductive Databases to Scalable Differentiable Reasoning create mode 100644 data/2021/neurips/Scatterbrain: Unifying Sparse and Low-rank Attention create mode 100644 data/2021/neurips/Scheduling jobs with stochastic holding costs create mode 100644 data/2021/neurips/Score-based Generative Modeling in Latent Space create mode 100644 data/2021/neurips/Score-based Generative Neural Networks for Large-Scale Optimal Transport create mode 100644 data/2021/neurips/Searching Parameterized AP Loss for Object Detection create mode 100644 data/2021/neurips/Searching for Efficient Transformers for Language Modeling create mode 100644 data/2021/neurips/Searching the Search Space of Vision Transformer create mode 100644 data/2021/neurips/Second-Order Neural ODE Optimizer create mode 100644 data/2021/neurips/See More for Scene: Pairwise Consistency Learning for Scene Classification create mode 100644 data/2021/neurips/SegFormer: Simple and Efficient Design for Semantic Segmentation with Transformers create mode 100644 data/2021/neurips/Selective Sampling for Online Best-arm Identification create mode 100644 data/2021/neurips/Self-Adaptable Point Processes with Nonparametric Time Decays create mode 100644 data/2021/neurips/Self-Attention Between Datapoints: Going Beyond Individual Input-Output Pairs in Deep Learning create mode 100644 data/2021/neurips/Self-Consistent Models and Values create mode 100644 data/2021/neurips/Self-Diagnosing GAN: Diagnosing Underrepresented Samples in Generative Adversarial Networks create mode 100644 data/2021/neurips/Self-Instantiated Recurrent Units with Dynamic Soft Recursion create mode 100644 data/2021/neurips/Self-Interpretable Model with Transformation Equivariant Interpretation create mode 100644 data/2021/neurips/Self-Paced Contrastive Learning for Semi-supervised Medical Image Segmentation with Meta-labels create mode 100644 data/2021/neurips/Self-Supervised Bug Detection and Repair create mode 100644 data/2021/neurips/Self-Supervised GANs with Label Augmentation create mode 100644 data/2021/neurips/Self-Supervised Learning Disentangled Group Representation as Feature create mode 100644 data/2021/neurips/Self-Supervised Learning of Event-Based Optical Flow with Spiking Neural Networks create mode 100644 data/2021/neurips/Self-Supervised Learning with Data Augmentations Provably Isolates Content from Style create mode 100644 data/2021/neurips/Self-Supervised Learning with Kernel Dependence Maximization create mode 100644 data/2021/neurips/Self-Supervised Multi-Object Tracking with Cross-input Consistency create mode 100644 data/2021/neurips/Self-Supervised Representation Learning on Neural Network Weights for Model Characteristic Prediction create mode 100644 data/2021/neurips/Semi-Supervised Semantic Segmentation via Adaptive Equalization Learning create mode 100644 data/2021/neurips/Semialgebraic Representation of Monotone Deep Equilibrium Models and Applications to Certification create mode 100644 data/2021/neurips/Separation Results between Fixed-Kernel and Feature-Learning Probability Metrics create mode 100644 data/2021/neurips/Sequence-to-Sequence Learning with Latent Neural Grammars create mode 100644 data/2021/neurips/Sequential Algorithms for Testing Closeness of Distributions create mode 100644 data/2021/neurips/Sequential Causal Imitation Learning with Unobserved Confounders create mode 100644 data/2021/neurips/Set Prediction in the Latent Space create mode 100644 data/2021/neurips/Settling the Variance of Multi-Agent Policy Gradients create mode 100644 data/2021/neurips/Shape As Points: A Differentiable Poisson Solver create mode 100644 data/2021/neurips/Shape Registration in the Time of Transformers create mode 100644 data/2021/neurips/Shape from Blur: Recovering Textured 3D Shape and Motion of Fast Moving Objects create mode 100644 data/2021/neurips/Shape your Space: A Gaussian Mixture Regularization Approach to Deterministic Autoencoders create mode 100644 data/2021/neurips/Shapeshifter: a Parameter-efficient Transformer using Factorized Reshaped Matrices create mode 100644 data/2021/neurips/Shaping embodied agent behavior with activity-context priors from egocentric video create mode 100644 data/2021/neurips/Shapley Residuals: Quantifying the limits of the Shapley value for explanations create mode 100644 data/2021/neurips/Shared Independent Component Analysis for Multi-Subject Neuroimaging create mode 100644 data/2021/neurips/Sharp Impossibility Results for Hyper-graph Testing create mode 100644 data/2021/neurips/Shift Invariance Can Reduce Adversarial Robustness create mode 100644 data/2021/neurips/Shift-Robust GNNs: Overcoming the Limitations of Localized Graph Training data create mode 100644 data/2021/neurips/Shifted Chunk Transformer for Spatio-Temporal Representational Learning create mode 100644 data/2021/neurips/Sifting through the noise: Universal first-order methods for stochastic variational inequalities create mode 100644 data/2021/neurips/Sim and Real: Better Together create mode 100644 data/2021/neurips/SimiGrad: Fine-Grained Adaptive Batching for Large Scale Training using Gradient Similarity Measurement create mode 100644 data/2021/neurips/Similarity and Matching of Neural Network Representations create mode 100644 data/2021/neurips/Simple Stochastic and Online Gradient Descent Algorithms for Pairwise Learning create mode 100644 data/2021/neurips/Simple steps are all you need: Frank-Wolfe and generalized self-concordant functions create mode 100644 data/2021/neurips/Single Layer Predictive Normalized Maximum Likelihood for Out-of-Distribution Detection create mode 100644 data/2021/neurips/SketchGen: Generating Constrained CAD Sketches create mode 100644 data/2021/neurips/Skipping the Frame-Level: Event-Based Piano Transcription With Neural Semi-CRFs create mode 100644 data/2021/neurips/Slice Sampling Reparameterization Gradients create mode 100644 data/2021/neurips/Sliced Mutual Information: A Scalable Measure of Statistical Dependence create mode 100644 data/2021/neurips/Slow Learning and Fast Inference: Efficient Graph Similarity Computation via Knowledge Distillation create mode 100644 data/2021/neurips/Small random initialization is akin to spectral learning: Optimization and generalization guarantees for overparameterized low-rank matrix reconstruction create mode 100644 data/2021/neurips/Smooth Bilevel Programming for Sparse Regularization create mode 100644 data/2021/neurips/Smooth Normalizing Flows create mode 100644 data/2021/neurips/SmoothMix: Training Confidence-calibrated Smoothed Classifiers for Certified Robustness create mode 100644 data/2021/neurips/Smoothness Matrices Beat Smoothness Constants: Better Communication Compression Techniques for Distributed Optimization create mode 100644 data/2021/neurips/Snowflake: Scaling GNNs to high-dimensional continuous control via parameter freezing create mode 100644 data/2021/neurips/Soft Calibration Objectives for Neural Networks create mode 100644 data/2021/neurips/Solving Graph-based Public Goods Games with Tree Search and Imitation Learning create mode 100644 data/2021/neurips/Solving Min-Max Optimization with Hidden Structure via Gradient Descent Ascent create mode 100644 data/2021/neurips/Solving Soft Clustering Ensemble via $k$-Sparse Discrete Wasserstein Barycenter create mode 100644 data/2021/neurips/Space-time Mixing Attention for Video Transformer create mode 100644 data/2021/neurips/Sparse Deep Learning: A New Framework Immune to Local Traps and Miscalibration create mode 100644 data/2021/neurips/Sparse Flows: Pruning Continuous-depth Models create mode 100644 data/2021/neurips/Sparse Quadratic Optimisation over the Stiefel Manifold with Application to Permutation Synchronisation create mode 100644 data/2021/neurips/Sparse Spiking Gradient Descent create mode 100644 data/2021/neurips/Sparse Steerable Convolutions: An Efficient Learning of SE(3)-Equivariant Features for Estimation and Tracking of Object Poses in 3D Space create mode 100644 data/2021/neurips/Sparse Training via Boosting Pruning Plasticity with Neuroregeneration create mode 100644 data/2021/neurips/Sparse Uncertainty Representation in Deep Learning with Inducing Weights create mode 100644 data/2021/neurips/Sparse is Enough in Scaling Transformers create mode 100644 data/2021/neurips/Sparsely Changing Latent States for Prediction and Planning in Partially Observable Domains create mode 100644 data/2021/neurips/Spatial Ensemble: a Novel Model Smoothing Mechanism for Student-Teacher Framework create mode 100644 data/2021/neurips/Spatial-Temporal Super-Resolution of Satellite Imagery via Conditional Pixel Synthesis create mode 100644 data/2021/neurips/Spatio-Temporal Variational Gaussian Processes create mode 100644 data/2021/neurips/Spatiotemporal Joint Filter Decomposition in 3D Convolutional Neural Networks create mode 100644 data/2021/neurips/Spectral embedding for dynamic networks with stability guarantees create mode 100644 data/2021/neurips/Spectrum-to-Kernel Translation for Accurate Blind Image Super-Resolution create mode 100644 data/2021/neurips/Speech Separation Using an Asynchronous Fully Recurrent Convolutional Neural Network create mode 100644 data/2021/neurips/Speech-T: Transducer for Text to Speech and Beyond create mode 100644 data/2021/neurips/Speedy Performance Estimation for Neural Architecture Search create mode 100644 data/2021/neurips/Spherical Motion Dynamics: Learning Dynamics of Normalized Neural Network using SGD and Weight Decay create mode 100644 data/2021/neurips/Spot the Difference: Detection of Topological Changes via Geometric Alignment create mode 100644 data/2021/neurips/Square Root Principal Component Pursuit: Tuning-Free Noisy Robust Matrix Recovery create mode 100644 data/2021/neurips/Stability & Generalisation of Gradient Descent for Shallow Neural Networks without the Neural Tangent Kernel create mode 100644 data/2021/neurips/Stability and Deviation Optimal Risk Bounds with Convergence Rate $O(1 n)$ create mode 100644 data/2021/neurips/Stability and Generalization of Bilevel Programming in Hyperparameter Optimization create mode 100644 data/2021/neurips/Stabilizing Deep Q-Learning with ConvNets and Vision Transformers under Data Augmentation create mode 100644 data/2021/neurips/Stabilizing Dynamical Systems via Policy Gradient Methods create mode 100644 data/2021/neurips/Stable Neural ODE with Lyapunov-Stable Equilibrium Points for Defending Against Adversarial Attacks create mode 100644 data/2021/neurips/Stable, Fast and Accurate: Kernelized Attention with Relative Positional Encoding create mode 100644 data/2021/neurips/Stateful ODE-Nets using Basis Function Expansions create mode 100644 data/2021/neurips/Stateful Strategic Regression create mode 100644 data/2021/neurips/Statistical Inference with M-Estimators on Adaptively Collected Data create mode 100644 data/2021/neurips/Statistical Query Lower Bounds for List-Decodable Linear Regression create mode 100644 data/2021/neurips/Statistical Regeneration Guarantees of the Wasserstein Autoencoder with Latent Space Consistency create mode 100644 data/2021/neurips/Statistical Undecidability in Linear, Non-Gaussian Causal Models in the Presence of Latent Confounders create mode 100644 data/2021/neurips/Statistically and Computationally Efficient Linear Meta-representation Learning create mode 100644 data/2021/neurips/Stochastic Anderson Mixing for Nonconvex Stochastic Optimization create mode 100644 data/2021/neurips/Stochastic Bias-Reduced Gradient Methods create mode 100644 data/2021/neurips/Stochastic Gradient Descent-Ascent and Consensus Optimization for Smooth Games: Convergence Analysis under Expected Co-coercivity create mode 100644 data/2021/neurips/Stochastic Multi-Armed Bandits with Control Variates create mode 100644 data/2021/neurips/Stochastic Online Linear Regression: the Forward Algorithm to Replace Ridge create mode 100644 data/2021/neurips/Stochastic Optimization of Areas Under Precision-Recall Curves with Provable Convergence create mode 100644 data/2021/neurips/Stochastic Shortest Path: Minimax, Parameter-Free and Towards Horizon-Free Regret create mode 100644 data/2021/neurips/Stochastic Solutions for Linear Inverse Problems using the Prior Implicit in a Denoiser create mode 100644 data/2021/neurips/Stochastic bandits with groups of similar arms create mode 100644 data/2021/neurips/Stochastic optimization under time drift: iterate averaging, step-decay schedules, and high probability guarantees create mode 100644 data/2021/neurips/Storchastic: A Framework for General Stochastic Automatic Differentiation create mode 100644 data/2021/neurips/Strategic Behavior is Bliss: Iterative Voting Improves Social Welfare create mode 100644 data/2021/neurips/Streaming Linear System Identification with Reverse Experience Replay create mode 100644 data/2021/neurips/Stronger NAS with Weaker Predictors create mode 100644 data/2021/neurips/Structural Credit Assignment in Neural Networks using Reinforcement Learning create mode 100644 data/2021/neurips/Structure learning in polynomial time: Greedy algorithms, Bregman information, and exponential families create mode 100644 data/2021/neurips/Structure-Aware Random Fourier Kernel for Graphs create mode 100644 data/2021/neurips/Structured Denoising Diffusion Models in Discrete State-Spaces create mode 100644 data/2021/neurips/Structured Dropout Variational Inference for Bayesian Neural Networks create mode 100644 data/2021/neurips/Structured Reordering for Modeling Latent Alignments in Sequence Transduction create mode 100644 data/2021/neurips/Structured in Space, Randomized in Time: Leveraging Dropout in RNNs for Efficient Training create mode 100644 data/2021/neurips/Stylized Dialogue Generation with Multi-Pass Dual Learning create mode 100644 data/2021/neurips/Sub-Linear Memory: How to Make Performers SLiM create mode 100644 data/2021/neurips/SubTab: Subsetting Features of Tabular Data for Self-Supervised Representation Learning create mode 100644 data/2021/neurips/Subgame solving without common knowledge create mode 100644 data/2021/neurips/Subgaussian and Differentiable Importance Sampling for Off-Policy Evaluation and Learning create mode 100644 data/2021/neurips/Subgoal Search For Complex Reasoning Tasks create mode 100644 data/2021/neurips/Subgraph Federated Learning with Missing Neighbor Generation create mode 100644 data/2021/neurips/Subgroup Generalization and Fairness of Graph Neural Networks create mode 100644 data/2021/neurips/Submodular + Concave create mode 100644 data/2021/neurips/Subquadratic Overparameterization for Shallow Neural Networks create mode 100644 data/2021/neurips/Successor Feature Landmarks for Long-Horizon Goal-Conditioned Reinforcement Learning create mode 100644 data/2021/neurips/Supercharging Imbalanced Data Learning With Energy-based Contrastive Representation Transfer create mode 100644 data/2021/neurips/Supervising the Transfer of Reasoning Patterns in VQA create mode 100644 data/2021/neurips/Support Recovery of Sparse Signals from a Mixture of Linear Measurements create mode 100644 data/2021/neurips/Support vector machines and linear regression coincide with very high-dimensional features create mode 100644 data/2021/neurips/Surrogate Regret Bounds for Polyhedral Losses create mode 100644 data/2021/neurips/SurvITE: Learning Heterogeneous Treatment Effects from Time-to-Event Data create mode 100644 data/2021/neurips/SyMetric: Measuring the Quality of Learnt Hamiltonian Dynamics Inferred from Vision create mode 100644 data/2021/neurips/Symbolic Regression via Deep Reinforcement Learning Enhanced Genetic Programming Seeding create mode 100644 data/2021/neurips/Symplectic Adjoint Method for Exact Gradient of Neural ODE with Minimal Memory create mode 100644 data/2021/neurips/SyncTwin: Treatment Effect Estimation with Longitudinal Outcomes create mode 100644 data/2021/neurips/Synthetic Design: An Optimization Approach to Experimental Design with Synthetic Controls create mode 100644 data/2021/neurips/Systematic Generalization with Edge Transformers create mode 100644 data/2021/neurips/T-LoHo: A Bayesian Regularization Model for Structured Sparsity and Smoothness on Graphs create mode 100644 data/2021/neurips/TAAC: Temporally Abstract Actor-Critic for Continuous Control create mode 100644 data/2021/neurips/TNASP: A Transformer-based NAS Predictor with a Self-evolution Framework create mode 100644 data/2021/neurips/TOHAN: A One-step Approach towards Few-shot Hypothesis Adaptation create mode 100644 data/2021/neurips/TRS: Transferability Reduced Ensemble via Promoting Gradient Diversity and Model Smoothness create mode 100644 data/2021/neurips/TTT++: When Does Self-Supervised Test-Time Training Fail or Thrive? create mode 100644 data/2021/neurips/TacticZero: Learning to Prove Theorems from Scratch with Deep Reinforcement Learning create mode 100644 data/2021/neurips/Tactical Optimism and Pessimism for Deep Reinforcement Learning create mode 100644 data/2021/neurips/Tailoring: encoding inductive biases by optimizing unsupervised objectives at prediction time create mode 100644 data/2021/neurips/Taming Communication and Sample Complexities in Decentralized Policy Evaluation for Cooperative Multi-Agent Reinforcement Learning create mode 100644 data/2021/neurips/Targeted Neural Dynamical Modeling create mode 100644 data/2021/neurips/Task-Adaptive Neural Network Search with Meta-Contrastive Learning create mode 100644 data/2021/neurips/Task-Agnostic Undesirable Feature Deactivation Using Out-of-Distribution Data create mode 100644 data/2021/neurips/Taxonomizing local versus global structure in neural network loss landscapes create mode 100644 data/2021/neurips/Teachable Reinforcement Learning via Advice Distillation create mode 100644 data/2021/neurips/Teaching an Active Learner with Contrastive Examples create mode 100644 data/2021/neurips/Teaching via Best-Case Counterexamples in the Learning-with-Equivalence-Queries Paradigm create mode 100644 data/2021/neurips/Techniques for Symbol Grounding with SATNet create mode 100644 data/2021/neurips/Temporal-attentive Covariance Pooling Networks for Video Recognition create mode 100644 data/2021/neurips/Temporally Abstract Partial Models create mode 100644 data/2021/neurips/Tensor Normal Training for Deep Learning Models create mode 100644 data/2021/neurips/Tensor decompositions of higher-order correlations by nonlinear Hebbian plasticity create mode 100644 data/2021/neurips/Terra: Imperative-Symbolic Co-Execution of Imperative Deep Learning Programs create mode 100644 data/2021/neurips/Test-Time Classifier Adjustment Module for Model-Agnostic Domain Generalization create mode 100644 data/2021/neurips/Test-Time Personalization with a Transformer for Human Pose Estimation create mode 100644 data/2021/neurips/Test-time Collective Prediction create mode 100644 data/2021/neurips/TestRank: Bringing Order into Unlabeled Test Instances for Deep Learning Tasks create mode 100644 data/2021/neurips/Testing Probabilistic Circuits create mode 100644 data/2021/neurips/The Adaptive Doubly Robust Estimator and a Paradox Concerning Logging Policy create mode 100644 data/2021/neurips/The Benefits of Implicit Regularization from SGD in Least Squares Problems create mode 100644 data/2021/neurips/The Causal-Neural Connection: Expressiveness, Learnability, and Inference create mode 100644 data/2021/neurips/The Complexity of Bayesian Network Learning: Revisiting the Superstructure create mode 100644 data/2021/neurips/The Complexity of Sparse Tensor PCA create mode 100644 data/2021/neurips/The Difficulty of Passive Learning in Deep Reinforcement Learning create mode 100644 data/2021/neurips/The Effect of the Intrinsic Dimension on the Generalization of Quadratic Classifiers create mode 100644 data/2021/neurips/The Elastic Lottery Ticket Hypothesis create mode 100644 data/2021/neurips/The Emergence of Objectness: Learning Zero-shot Segmentation from Videos create mode 100644 data/2021/neurips/The Flip Side of the Reweighted Coin: Duality of Adaptive Dropout and Regularization create mode 100644 data/2021/neurips/The Hardness Analysis of Thompson Sampling for Combinatorial Semi-bandits with Greedy Oracle create mode 100644 data/2021/neurips/The Image Local Autoregressive Transformer create mode 100644 data/2021/neurips/The Implicit Bias of Minima Stability: A View from Function Space create mode 100644 data/2021/neurips/The Inductive Bias of Quantum Kernels create mode 100644 data/2021/neurips/The Lazy Online Subgradient Algorithm is Universal on Strongly Convex Domains create mode 100644 data/2021/neurips/The Limitations of Large Width in Neural Networks: A Deep Gaussian Process Perspective create mode 100644 data/2021/neurips/The Limits of Optimal Pricing in the Dark create mode 100644 data/2021/neurips/The Many Faces of Adversarial Risk create mode 100644 data/2021/neurips/The Out-of-Distribution Problem in Explainability and Search Methods for Feature Importance Explanations create mode 100644 data/2021/neurips/The Pareto Frontier of model selection for general Contextual Bandits create mode 100644 data/2021/neurips/The Role of Global Labels in Few-Shot Classification and How to Infer Them create mode 100644 data/2021/neurips/The Semi-Random Satisfaction of Voting Axioms create mode 100644 data/2021/neurips/The Sensory Neuron as a Transformer: Permutation-Invariant Neural Networks for Reinforcement Learning create mode 100644 data/2021/neurips/The Skellam Mechanism for Differentially Private Federated Learning create mode 100644 data/2021/neurips/The Unbalanced Gromov Wasserstein Distance: Conic Formulation and Relaxation create mode 100644 data/2021/neurips/The Utility of Explainable AI in Ad Hoc Human-Machine Teaming create mode 100644 data/2021/neurips/The Value of Information When Deciding What to Learn create mode 100644 data/2021/neurips/The balancing principle for parameter choice in distance-regularized domain adaptation create mode 100644 data/2021/neurips/The best of both worlds: stochastic and adversarial episodic MDPs with unknown transition create mode 100644 data/2021/neurips/The convergence rate of regularized learning in games: From bandits and uncertainty to optimism and beyond create mode 100644 data/2021/neurips/The decomposition of the higher-order homology embedding constructed from the $k$-Laplacian create mode 100644 data/2021/neurips/The effectiveness of feature attribution methods and its correlation with automatic evaluation scores create mode 100644 data/2021/neurips/The functional specialization of visual cortex emerges from training parallel pathways with self-supervised predictive learning create mode 100644 data/2021/neurips/The future is log-Gaussian: ResNets and their infinite-depth-and-width limit at initialization create mode 100644 data/2021/neurips/The staircase property: How hierarchical structure can guide deep learning create mode 100644 data/2021/neurips/There Is No Turning Back: A Self-Supervised Approach for Reversibility-Aware Reinforcement Learning create mode 100644 data/2021/neurips/Think Big, Teach Small: Do Language Models Distil Occam's Razor? create mode 100644 data/2021/neurips/Three Operator Splitting with Subgradients, Stochastic Gradients, and Adaptive Learning Rates create mode 100644 data/2021/neurips/Three-dimensional spike localization and improved motion correction for Neuropixels recordings create mode 100644 data/2021/neurips/Tight High Probability Bounds for Linear Stochastic Approximation with Fixed Stepsize create mode 100644 data/2021/neurips/Tighter Expected Generalization Error Bounds via Wasserstein Distance create mode 100644 data/2021/neurips/Time Discretization-Invariant Safe Action Repetition for Policy Gradient Methods create mode 100644 data/2021/neurips/Time-independent Generalization Bounds for SGLD in Non-convex Settings create mode 100644 data/2021/neurips/Time-series Generation by Contrastive Imitation create mode 100644 data/2021/neurips/To Beam Or Not To Beam: That is a Question of Cooperation for Language GANs create mode 100644 data/2021/neurips/To The Point: Correspondence-driven monocular 3D category reconstruction create mode 100644 data/2021/neurips/ToAlign: Task-Oriented Alignment for Unsupervised Domain Adaptation create mode 100644 data/2021/neurips/TokenLearner: Adaptive Space-Time Tokenization for Videos create mode 100644 data/2021/neurips/Topic Modeling Revisited: A Document Graph-based Neural Network Perspective create mode 100644 data/2021/neurips/TopicNet: Semantic Graph-Guided Topic Discovery create mode 100644 data/2021/neurips/Topographic VAEs learn Equivariant Capsules create mode 100644 data/2021/neurips/Topological Attention for Time Series Forecasting create mode 100644 data/2021/neurips/Topological Detection of Trojaned Neural Networks create mode 100644 data/2021/neurips/Topological Relational Learning on Graphs create mode 100644 data/2021/neurips/Topology-Imbalance Learning for Semi-Supervised Node Classification create mode 100644 data/2021/neurips/Towards Best-of-All-Worlds Online Learning with Feedback Graphs create mode 100644 data/2021/neurips/Towards Better Understanding of Training Certifiably Robust Models against Adversarial Examples create mode 100644 data/2021/neurips/Towards Biologically Plausible Convolutional Networks create mode 100644 data/2021/neurips/Towards Calibrated Model for Long-Tailed Visual Recognition from Prior Perspective create mode 100644 data/2021/neurips/Towards Context-Agnostic Learning Using Synthetic Data create mode 100644 data/2021/neurips/Towards Deeper Deep Reinforcement Learning with Spectral Normalization create mode 100644 data/2021/neurips/Towards Efficient and Effective Adversarial Training create mode 100644 data/2021/neurips/Towards Enabling Meta-Learning from Target Models create mode 100644 data/2021/neurips/Towards Gradient-based Bilevel Optimization with Non-convex Followers and Beyond create mode 100644 data/2021/neurips/Towards Hyperparameter-free Policy Selection for Offline Reinforcement Learning create mode 100644 data/2021/neurips/Towards Instance-Optimal Offline Reinforcement Learning with Pessimism create mode 100644 data/2021/neurips/Towards Lower Bounds on the Depth of ReLU Neural Networks create mode 100644 data/2021/neurips/Towards Multi-Grained Explainability for Graph Neural Networks create mode 100644 data/2021/neurips/Towards Open-World Feature Extrapolation: An Inductive Graph Learning Approach create mode 100644 data/2021/neurips/Towards Optimal Strategies for Training Self-Driving Perception Models in Simulation create mode 100644 data/2021/neurips/Towards Robust Bisimulation Metric Learning create mode 100644 data/2021/neurips/Towards Robust and Reliable Algorithmic Recourse create mode 100644 data/2021/neurips/Towards Sample-Optimal Compressive Phase Retrieval with Sparse and Generative Priors create mode 100644 data/2021/neurips/Towards Sample-efficient Overparameterized Meta-learning create mode 100644 data/2021/neurips/Towards Scalable Unpaired Virtual Try-On via Patch-Routed Spatially-Adaptive GAN create mode 100644 data/2021/neurips/Towards Sharper Generalization Bounds for Structured Prediction create mode 100644 data/2021/neurips/Towards Stable and Robust AdderNets create mode 100644 data/2021/neurips/Towards Tight Communication Lower Bounds for Distributed Optimisation create mode 100644 data/2021/neurips/Towards Understanding Cooperative Multi-Agent Q-Learning with Value Factorization create mode 100644 data/2021/neurips/Towards Understanding Why Lookahead Generalizes Better Than SGD and Beyond create mode 100644 data/2021/neurips/Towards Unifying Behavioral and Response Diversity for Open-ended Learning in Zero-sum Games create mode 100644 data/2021/neurips/Towards a Theoretical Framework of Out-of-Distribution Generalization create mode 100644 data/2021/neurips/Towards a Unified Game-Theoretic View of Adversarial Perturbations and Robustness create mode 100644 data/2021/neurips/Towards a Unified Information-Theoretic Framework for Generalization create mode 100644 data/2021/neurips/Towards mental time travel: a hierarchical memory for reinforcement learning agents create mode 100644 data/2021/neurips/Towards optimally abstaining from prediction with OOD test examples create mode 100644 data/2021/neurips/Towards robust vision by multi-task learning on monkey visual cortex create mode 100644 data/2021/neurips/Towards understanding retrosynthesis by energy-based models create mode 100644 data/2021/neurips/Tracking People with 3D Representations create mode 100644 data/2021/neurips/Tracking Without Re-recognition in Humans and Machines create mode 100644 data/2021/neurips/Tractable Density Estimation on Learned Manifolds with Conformal Embedding Flows create mode 100644 data/2021/neurips/Tractable Regularization of Probabilistic Circuits create mode 100644 data/2021/neurips/Training Certifiably Robust Neural Networks with Efficient Local Lipschitz Bounds create mode 100644 data/2021/neurips/Training Feedback Spiking Neural Networks by Implicit Differentiation on the Equilibrium State create mode 100644 data/2021/neurips/Training Neural Networks is ER-complete create mode 100644 data/2021/neurips/Training Neural Networks with Fixed Sparse Masks create mode 100644 data/2021/neurips/Training Over-parameterized Models with Non-decomposable Objectives create mode 100644 data/2021/neurips/Training for the Future: A Simple Gradient Interpolation Loss to Generalize Along Time create mode 100644 data/2021/neurips/TransGAN: Two Pure Transformers Can Make One Strong GAN, and That Can Scale Up create mode 100644 data/2021/neurips/TransMIL: Transformer based Correlated Multiple Instance Learning for Whole Slide Image Classification create mode 100644 data/2021/neurips/TransMatcher: Deep Image Matching Through Transformers for Generalizable Person Re-identification create mode 100644 data/2021/neurips/Transfer Learning of Graph Neural Networks with Ego-graph Information Maximization create mode 100644 data/2021/neurips/Transformer in Transformer create mode 100644 data/2021/neurips/TransformerFusion: Monocular RGB Scene Reconstruction using Transformers create mode 100644 data/2021/neurips/Transformers Generalize DeepSets and Can be Extended to Graphs & Hypergraphs create mode 100644 data/2021/neurips/Trash or Treasure? An Interactive Dual-Stream Strategy for Single Image Reflection Separation create mode 100644 data/2021/neurips/Tree in Tree: from Decision Trees to Decision Graphs create mode 100644 data/2021/neurips/TriBERT: Human-centric Audio-visual Representation Learning create mode 100644 data/2021/neurips/True Few-Shot Learning with Language Models create mode 100644 data/2021/neurips/Truncated Marginal Neural Ratio Estimation create mode 100644 data/2021/neurips/Trustworthy Multimodal Regression with Mixture of Normal-inverse Gamma Distributions create mode 100644 data/2021/neurips/Tuning Large Neural Networks via Zero-Shot Hyperparameter Transfer create mode 100644 data/2021/neurips/Tuning Mixed Input Hyperparameters on the Fly for Efficient Population Based AutoRL create mode 100644 data/2021/neurips/Turing Completeness of Bounded-Precision Recurrent Neural Networks create mode 100644 data/2021/neurips/Twice regularized MDPs and the equivalence between robustness and regularization create mode 100644 data/2021/neurips/Twins: Revisiting the Design of Spatial Attention in Vision Transformers create mode 100644 data/2021/neurips/Two Sides of Meta-Learning Evaluation: In vs. Out of Distribution create mode 100644 data/2021/neurips/Two steps to risk sensitivity create mode 100644 data/2021/neurips/Two-sided fairness in rankings via Lorenz dominance create mode 100644 data/2021/neurips/Two-step lookahead Bayesian optimization with inequality constraints create mode 100644 "data/2021/neurips/T\303\266RF: Time-of-Flight Radiance Fields for Dynamic Scene View Synthesis" create mode 100644 data/2021/neurips/UCB-based Algorithms for Multinomial Logistic Regression Bandits create mode 100644 data/2021/neurips/UFC-BERT: Unifying Multi-Modal Controls for Conditional Image Synthesis create mode 100644 data/2021/neurips/USCO-Solver: Solving Undetermined Stochastic Combinatorial Optimization Problems create mode 100644 data/2021/neurips/Ultrahyperbolic Neural Networks create mode 100644 data/2021/neurips/Unadversarial Examples: Designing Objects for Robust Vision create mode 100644 data/2021/neurips/Unbalanced Optimal Transport through Non-negative Penalized Linear Regression create mode 100644 data/2021/neurips/Unbiased Classification through Bias-Contrastive and Bias-Balanced Learning create mode 100644 data/2021/neurips/Uncertain Decisions Facilitate Better Preference Learning create mode 100644 data/2021/neurips/Uncertainty Calibration for Ensemble-Based Debiasing Methods create mode 100644 data/2021/neurips/Uncertainty Quantification and Deep Ensembles create mode 100644 data/2021/neurips/Uncertainty-Based Offline Reinforcement Learning with Diversified Q-Ensemble create mode 100644 data/2021/neurips/Uncertainty-Driven Loss for Single Image Super-Resolution create mode 100644 data/2021/neurips/Understanding Adaptive, Multiscale Temporal Integration In Deep Speech Recognition Systems create mode 100644 data/2021/neurips/Understanding Bandits with Graph Feedback create mode 100644 data/2021/neurips/Understanding Deflation Process in Over-parametrized Tensor Decomposition create mode 100644 data/2021/neurips/Understanding End-to-End Model-Based Reinforcement Learning Methods as Implicit Parameterization create mode 100644 data/2021/neurips/Understanding How Encoder-Decoder Architectures Attend create mode 100644 data/2021/neurips/Understanding Instance-based Interpretability of Variational Auto-Encoders create mode 100644 data/2021/neurips/Understanding Interlocking Dynamics of Cooperative Rationalization create mode 100644 data/2021/neurips/Understanding Negative Samples in Instance Discriminative Self-supervised Representation Learning create mode 100644 data/2021/neurips/Understanding Partial Multi-Label Learning via Mutual Information create mode 100644 data/2021/neurips/Understanding and Improving Early Stopping for Learning with Noisy Labels create mode 100644 data/2021/neurips/Understanding the Effect of Stochasticity in Policy Optimization create mode 100644 data/2021/neurips/Understanding the Generalization Benefit of Model Invariance from a Data Perspective create mode 100644 data/2021/neurips/Understanding the Limits of Unsupervised Domain Adaptation via Data Poisoning create mode 100644 data/2021/neurips/Understanding the Under-Coverage Bias in Uncertainty Estimation create mode 100644 data/2021/neurips/Unfolding Taylor's Approximations for Image Restoration create mode 100644 data/2021/neurips/UniDoc: Unified Pretraining Framework for Document Understanding create mode 100644 data/2021/neurips/Uniform Concentration Bounds toward a Unified Framework for Robust Clustering create mode 100644 data/2021/neurips/Uniform Convergence of Interpolators: Gaussian Width, Norm Bounds and Benign Overfitting create mode 100644 data/2021/neurips/Uniform Sampling over Episode Difficulty create mode 100644 data/2021/neurips/Uniform-PAC Bounds for Reinforcement Learning with Linear Function Approximation create mode 100644 data/2021/neurips/Unifying Gradient Estimators for Meta-Reinforcement Learning via Off-Policy Evaluation create mode 100644 data/2021/neurips/Unifying Width-Reduced Methods for Quasi-Self-Concordant Optimization create mode 100644 data/2021/neurips/Unifying lower bounds on prediction dimension of convex surrogates create mode 100644 data/2021/neurips/Unintended Selection: Persistent Qualification Rate Disparities and Interventions create mode 100644 data/2021/neurips/Unique sparse decomposition of low rank matrices create mode 100644 data/2021/neurips/Universal Approximation Using Well-Conditioned Normalizing Flows create mode 100644 data/2021/neurips/Universal Graph Convolutional Networks create mode 100644 data/2021/neurips/Universal Off-Policy Evaluation create mode 100644 data/2021/neurips/Universal Rate-Distortion-Perception Representations for Lossy Compression create mode 100644 data/2021/neurips/Universal Semi-Supervised Learning create mode 100644 data/2021/neurips/Unlabeled Principal Component Analysis create mode 100644 data/2021/neurips/Unleashing the Power of Contrastive Self-Supervised Visual Models via Contrast-Regularized Fine-Tuning create mode 100644 data/2021/neurips/Unsupervised Domain Adaptation with Dynamics-Aware Rewards in Reinforcement Learning create mode 100644 data/2021/neurips/Unsupervised Foreground Extraction via Deep Region Competition create mode 100644 data/2021/neurips/Unsupervised Learning of Compositional Energy Concepts create mode 100644 data/2021/neurips/Unsupervised Motion Representation Learning with Capsule Autoencoders create mode 100644 data/2021/neurips/Unsupervised Noise Adaptive Speech Enhancement by Discriminator-Constrained Optimal Transport create mode 100644 data/2021/neurips/Unsupervised Object-Based Transition Models For 3D Partially Observable Environments create mode 100644 data/2021/neurips/Unsupervised Object-Level Representation Learning from Scene Images create mode 100644 data/2021/neurips/Unsupervised Part Discovery from Contrastive Reconstruction create mode 100644 data/2021/neurips/Unsupervised Representation Transfer for Small Networks: I Believe I Can Distill On-the-Fly create mode 100644 data/2021/neurips/Unsupervised Speech Recognition create mode 100644 data/2021/neurips/User-Level Differentially Private Learning via Correlated Sampling create mode 100644 data/2021/neurips/Using Random Effects to Account for High-Cardinality Categorical Features and Repeated Measures in Deep Neural Networks create mode 100644 data/2021/neurips/VAST: Value Function Factorization with Variable Agent Sub-Teams create mode 100644 data/2021/neurips/VATT: Transformers for Multimodal Self-Supervised Learning from Raw Video, Audio and Text create mode 100644 data/2021/neurips/VQ-GNN: A Universal Framework to Scale up Graph Neural Networks using Vector Quantization create mode 100644 data/2021/neurips/Validating the Lottery Ticket Hypothesis with Inertial Manifold Theory create mode 100644 data/2021/neurips/Validation Free and Replication Robust Volume-based Data Valuation create mode 100644 data/2021/neurips/Variance-Aware Off-Policy Evaluation with Linear Function Approximation create mode 100644 data/2021/neurips/Variational Automatic Curriculum Learning for Sparse-Reward Cooperative Multi-Agent Problems create mode 100644 data/2021/neurips/Variational Bayesian Optimistic Sampling create mode 100644 data/2021/neurips/Variational Bayesian Reinforcement Learning with Regret Bounds create mode 100644 data/2021/neurips/Variational Continual Bayesian Meta-Learning create mode 100644 data/2021/neurips/Variational Inference for Continuous-Time Switching Dynamical Systems create mode 100644 data/2021/neurips/Variational Model Inversion Attacks create mode 100644 data/2021/neurips/Variational Multi-Task Learning with Gumbel-Softmax Priors create mode 100644 data/2021/neurips/Vector-valued Distance and Gyrocalculus on the Space of Symmetric Positive Definite Matrices create mode 100644 data/2021/neurips/Vector-valued Gaussian Processes on Riemannian Manifolds via Gauge Independent Projected Kernels create mode 100644 data/2021/neurips/ViSER: Video-Specific Surface Embeddings for Articulated 3D Shape Reconstruction create mode 100644 data/2021/neurips/ViTAE: Vision Transformer Advanced by Exploring Intrinsic Inductive Bias create mode 100644 data/2021/neurips/VidLanKD: Improving Language Understanding via Video-Distilled Knowledge Transfer create mode 100644 data/2021/neurips/Video Instance Segmentation using Inter-Frame Communication Transformers create mode 100644 data/2021/neurips/VigDet: Knowledge Informed Neural Temporal Point Process for Coordination Detection on Social Media create mode 100644 data/2021/neurips/Visual Adversarial Imitation Learning using Variational Models create mode 100644 data/2021/neurips/Visual Search Asymmetry: Deep Nets and Humans Share Similar Inherent Biases create mode 100644 data/2021/neurips/Visualizing the Emergence of Intermediate Visual Patterns in DNNs create mode 100644 data/2021/neurips/VoiceMixer: Adversarial Voice Style Mixup create mode 100644 data/2021/neurips/Volume Rendering of Neural Implicit Surfaces create mode 100644 data/2021/neurips/Voxel-based 3D Detection and Reconstruction of Multiple Objects from a Single Image create mode 100644 data/2021/neurips/Wasserstein Flow Meets Replicator Dynamics: A Mean-Field Analysis of Representation Learning in Actor-Critic create mode 100644 data/2021/neurips/Weak-shot Fine-grained Classification via Similarity Transfer create mode 100644 data/2021/neurips/Weighted model estimation for offline model-based reinforcement learning create mode 100644 data/2021/neurips/Weisfeiler and Lehman Go Cellular: CW Networks create mode 100644 data/2021/neurips/Well-tuned Simple Nets Excel on Tabular Datasets create mode 100644 data/2021/neurips/What Makes Multi-Modal Learning Better than Single (Provably) create mode 100644 data/2021/neurips/What Matters for Adversarial Imitation Learning? create mode 100644 data/2021/neurips/What can linearized neural networks actually say about generalization? create mode 100644 data/2021/neurips/What training reveals about neural network complexity create mode 100644 data/2021/neurips/What's a good imputation to predict with missing values? create mode 100644 data/2021/neurips/When Are Solutions Connected in Deep Networks? create mode 100644 data/2021/neurips/When Expressivity Meets Trainability: Fewer than $n$ Neurons Can Work create mode 100644 data/2021/neurips/When False Positive is Intolerant: End-to-End Optimization with Low FPR for Multipartite Ranking create mode 100644 data/2021/neurips/When Is Generalizable Reinforcement Learning Tractable? create mode 100644 data/2021/neurips/When Is Unsupervised Disentanglement Possible? create mode 100644 data/2021/neurips/When does Contrastive Learning Preserve Adversarial Robustness from Pretraining to Finetuning? create mode 100644 data/2021/neurips/When in Doubt: Neural Non-Parametric Uncertainty Quantification for Epidemic Forecasting create mode 100644 data/2021/neurips/Which Mutual-Information Representation Learning Objectives are Sufficient for Control? create mode 100644 data/2021/neurips/Who Leads and Who Follows in Strategic Classification? create mode 100644 data/2021/neurips/Why Do Better Loss Functions Lead to Less Transferable Features? create mode 100644 data/2021/neurips/Why Do Pretrained Language Models Help in Downstream Tasks? An Analysis of Head and Prompt Tuning create mode 100644 data/2021/neurips/Why Generalization in RL is Difficult: Epistemic POMDPs and Implicit Partial Observability create mode 100644 data/2021/neurips/Why Lottery Ticket Wins? A Theoretical Perspective of Sample Complexity on Sparse Neural Networks create mode 100644 data/2021/neurips/Why Spectral Normalization Stabilizes GANs: Analysis and Improvements create mode 100644 data/2021/neurips/Widening the Pipeline in Human-Guided Reinforcement Learning with Explanation and Context-Aware Data Augmentation create mode 100644 data/2021/neurips/Width-based Lookaheads with Learnt Base Policies and Heuristics Over the Atari-2600 Benchmark create mode 100644 data/2021/neurips/Wisdom of the Crowd Voting: Truthful Aggregation of Voter Information and Preferences create mode 100644 data/2021/neurips/Word2Fun: Modelling Words as Functions for Diachronic Word Representation create mode 100644 data/2021/neurips/XCiT: Cross-Covariance Image Transformers create mode 100644 data/2021/neurips/XDO: A Double Oracle Algorithm for Extensive-Form Games create mode 100644 data/2021/neurips/You Are the Best Reviewer of Your Own Papers: An Owner-Assisted Scoring Mechanism create mode 100644 data/2021/neurips/You Never Cluster Alone create mode 100644 data/2021/neurips/You Only Look at One Sequence: Rethinking Transformer in Vision through Object Detection create mode 100644 data/2021/neurips/You are caught stealing my winning lottery ticket! Making a lottery ticket claim its ownership create mode 100644 data/2021/neurips/Your head is there to move you around: Goal-driven models of the primate dorsal pathway create mode 100644 data/2021/neurips/Zero Time Waste: Recycling Predictions in Early Exit Neural Networks create mode 100644 data/2021/neurips/argmax centroid create mode 100644 data/2021/neurips/iFlow: Numerically Invertible Flows for Efficient Lossless Compression via a Uniform Coder create mode 100644 "data/2022/neurips/\"Lossless\" Compression of Deep Neural Networks: A High-dimensional Neural Tangent Kernel Approach" create mode 100644 "data/2022/neurips/\"Why Not Other Classes?\": Towards Class-Contrastive Back-Propagation Explanations" create mode 100644 data/2022/neurips/$k$-Sliced Mutual Information: A Quantitative Study of Scalability with Dimension create mode 100644 data/2022/neurips/(De-)Randomized Smoothing for Decision Stump Ensembles create mode 100644 data/2022/neurips/(Optimal) Online Bipartite Matching with Degree Information create mode 100644 data/2022/neurips/360-MLC: Multi-view Layout Consistency for Self-training and Hyper-parameter Tuning create mode 100644 data/2022/neurips/3D Concept Grounding on Neural Fields create mode 100644 data/2022/neurips/3DB: A Framework for Debugging Computer Vision Models create mode 100644 data/2022/neurips/3DILG: Irregular Latent Grids for 3D Generative Modeling create mode 100644 data/2022/neurips/3DOS: Towards 3D Open Set Learning - Benchmarking and Understanding Semantic Novelty Detection on Point Clouds create mode 100644 data/2022/neurips/4D Unsupervised Object Discovery create mode 100644 data/2022/neurips/A Benchmark for Compositional Visual Reasoning create mode 100644 data/2022/neurips/A Best-of-Both-Worlds Algorithm for Bandits with Delayed Feedback create mode 100644 data/2022/neurips/A Boosting Approach to Reinforcement Learning create mode 100644 data/2022/neurips/A Causal Analysis of Harm create mode 100644 data/2022/neurips/A Character-Level Length-Control Algorithm for Non-Autoregressive Sentence Summarization create mode 100644 data/2022/neurips/A Characterization of Semi-Supervised Adversarially Robust PAC Learnability create mode 100644 data/2022/neurips/A Closer Look at Learned Optimization: Stability, Robustness, and Inductive Biases create mode 100644 data/2022/neurips/A Closer Look at Offline RL Agents create mode 100644 data/2022/neurips/A Closer Look at Prototype Classifier for Few-shot Image Classification create mode 100644 data/2022/neurips/A Closer Look at Weakly-Supervised Audio-Visual Source Localization create mode 100644 data/2022/neurips/A Closer Look at the Adversarial Robustness of Deep Equilibrium Models create mode 100644 data/2022/neurips/A Combinatorial Perspective on the Optimization of Shallow ReLU Networks create mode 100644 data/2022/neurips/A Communication-Efficient Distributed Gradient Clipping Algorithm for Training Deep Neural Networks create mode 100644 data/2022/neurips/A Communication-efficient Algorithm with Linear Convergence for Federated Minimax Learning create mode 100644 data/2022/neurips/A Comprehensive Study on Large-Scale Graph Training: Benchmarking and Rethinking create mode 100644 data/2022/neurips/A Conditional Randomization Test for Sparse Logistic Regression in High-Dimension create mode 100644 data/2022/neurips/A Consistent and Differentiable Lp Canonical Calibration Error Estimator create mode 100644 data/2022/neurips/A Consolidated Cross-Validation Algorithm for Support Vector Machines via Data Reduction create mode 100644 data/2022/neurips/A Continuous Time Framework for Discrete Denoising Models create mode 100644 data/2022/neurips/A Contrastive Framework for Neural Text Generation create mode 100644 data/2022/neurips/A Coupled Design of Exploiting Record Similarity for Practical Vertical Federated Learning create mode 100644 data/2022/neurips/A Data-Augmentation Is Worth A Thousand Samples: Analytical Moments And Sampling-Free Training create mode 100644 data/2022/neurips/A Dataset for Efforts Towards Achieving the Sustainable Development Goal of Safe Working Environments create mode 100644 data/2022/neurips/A Deep Learning Dataloader with Shared Data Preparation create mode 100644 data/2022/neurips/A Deep Reinforcement Learning Framework for Column Generation create mode 100644 data/2022/neurips/A Differentiable Semantic Metric Approximation in Probabilistic Embedding for Cross-Modal Retrieval create mode 100644 data/2022/neurips/A Differentially Private Linear-Time fPTAS for the Minimum Enclosing Ball Problem create mode 100644 data/2022/neurips/A Direct Approximation of AIXI Using Logical State Abstractions create mode 100644 data/2022/neurips/A Fast Post-Training Pruning Framework for Transformers create mode 100644 data/2022/neurips/A Fast Scale-Invariant Algorithm for Non-negative Least Squares with Non-negative Data create mode 100644 data/2022/neurips/A Few Expert Queries Suffices for Sample-Efficient RL with Resets and Linear Value Approximation create mode 100644 data/2022/neurips/A Fourier Approach to Mixture Learning create mode 100644 data/2022/neurips/A General Framework for Auditing Differentially Private Machine Learning create mode 100644 data/2022/neurips/A Geometric Perspective on Variational Autoencoders create mode 100644 data/2022/neurips/A Greek Parliament Proceedings Dataset for Computational Linguistics and Political Analysis create mode 100644 data/2022/neurips/A Kernelised Stein Statistic for Assessing Implicit Generative Models create mode 100644 data/2022/neurips/A Lagrangian Duality Approach to Active Learning create mode 100644 data/2022/neurips/A Large Scale Search Dataset for Unbiased Learning to Rank create mode 100644 data/2022/neurips/A Lower Bound of Hash Codes' Performance create mode 100644 data/2022/neurips/A Mean-Field Game Approach to Cloud Resource Management with Function Approximation create mode 100644 data/2022/neurips/A Mixture Of Surprises for Unsupervised Reinforcement Learning create mode 100644 data/2022/neurips/A Multi-Resolution Framework for U-Nets with Applications to Hierarchical VAEs create mode 100644 data/2022/neurips/A Multi-Task Benchmark for Korean Legal Language Understanding and Judgement Prediction create mode 100644 data/2022/neurips/A Multilabel Classification Framework for Approximate Nearest Neighbor Search create mode 100644 data/2022/neurips/A Near-Optimal Best-of-Both-Worlds Algorithm for Online Learning with Feedback Graphs create mode 100644 data/2022/neurips/A Near-Optimal Primal-Dual Method for Off-Policy Learning in CMDP create mode 100644 data/2022/neurips/A Neural Corpus Indexer for Document Retrieval create mode 100644 data/2022/neurips/A Neural Pre-Conditioning Active Learning Algorithm to Reduce Label Complexity create mode 100644 data/2022/neurips/A New Family of Generalization Bounds Using Samplewise Evaluated CMI create mode 100644 data/2022/neurips/A Non-Asymptotic Moreau Envelope Theory for High-Dimensional Generalized Linear Models create mode 100644 data/2022/neurips/A Non-asymptotic Analysis of Non-parametric Temporal-Difference Learning create mode 100644 data/2022/neurips/A PAC-Bayesian Generalization Bound for Equivariant Networks create mode 100644 data/2022/neurips/A Policy-Guided Imitation Approach for Offline Reinforcement Learning create mode 100644 data/2022/neurips/A Practical, Progressively-Expressive GNN create mode 100644 data/2022/neurips/A Probabilistic Graph Coupling View of Dimension Reduction create mode 100644 data/2022/neurips/A Projection-free Algorithm for Constrained Stochastic Multi-level Composition Optimization create mode 100644 data/2022/neurips/A Quadrature Rule combining Control Variates and Adaptive Importance Sampling create mode 100644 data/2022/neurips/A Quantitative Geometric Approach to Neural-Network Smoothness create mode 100644 data/2022/neurips/A Reduction to Binary Approach for Debiasing Multiclass Datasets create mode 100644 data/2022/neurips/A Regret-Variance Trade-Off in Online Learning create mode 100644 data/2022/neurips/A Reparametrization-Invariant Sharpness Measure Based on Information Geometry create mode 100644 data/2022/neurips/A Robust Phased Elimination Algorithm for Corruption-Tolerant Gaussian Process Bandits create mode 100644 data/2022/neurips/A Rotated Hyperbolic Wrapped Normal Distribution for Hierarchical Representation Learning create mode 100644 data/2022/neurips/A Scalable Deterministic Global Optimization Algorithm for Training Optimal Decision Tree create mode 100644 data/2022/neurips/A Simple Approach to Automated Spectral Clustering create mode 100644 data/2022/neurips/A Simple Decentralized Cross-Entropy Method create mode 100644 data/2022/neurips/A Simple and Optimal Policy Design for Online Learning with Safety against Heavy-tailed Risk create mode 100644 data/2022/neurips/A Simple and Provably Efficient Algorithm for Asynchronous Federated Contextual Linear Bandits create mode 100644 data/2022/neurips/A Single-timescale Analysis for Stochastic Approximation with Multiple Coupled Sequences create mode 100644 data/2022/neurips/A Solver-free Framework for Scalable Learning in Neural ILP Architectures create mode 100644 data/2022/neurips/A Spectral Approach to Item Response Theory create mode 100644 data/2022/neurips/A Statistical Online Inference Approach in Averaged Stochastic Approximation create mode 100644 data/2022/neurips/A Stochastic Linearized Augmented Lagrangian Method for Decentralized Bilevel Optimization create mode 100644 data/2022/neurips/A Survey and Datasheet Repository of Publicly Available US Criminal Justice Datasets create mode 100644 data/2022/neurips/A Theoretical Framework for Inference Learning create mode 100644 data/2022/neurips/A Theoretical Study on Solving Continual Learning create mode 100644 data/2022/neurips/A Theoretical Understanding of Gradient Bias in Meta-Reinforcement Learning create mode 100644 data/2022/neurips/A Theoretical View on Sparsely Activated Networks create mode 100644 data/2022/neurips/A Theory of PAC Learnability under Transformation Invariances create mode 100644 data/2022/neurips/A Transformer-Based Object Detector with Coarse-Fine Crossing Representations create mode 100644 data/2022/neurips/A Unified Analysis of Federated Learning with Arbitrary Client Participation create mode 100644 data/2022/neurips/A Unified Analysis of Mixed Sample Data Augmentation: A Loss Function Perspective create mode 100644 data/2022/neurips/A Unified Convergence Theorem for Stochastic Optimization Methods create mode 100644 data/2022/neurips/A Unified Diversity Measure for Multiagent Reinforcement Learning create mode 100644 data/2022/neurips/A Unified Evaluation of Textual Backdoor Learning: Frameworks and Benchmarks create mode 100644 data/2022/neurips/A Unified Framework for Alternating Offline Model Training and Policy Learning create mode 100644 data/2022/neurips/A Unified Framework for Deep Symbolic Regression create mode 100644 data/2022/neurips/A Unified Hard-Constraint Framework for Solving Geometrically Complex PDEs create mode 100644 data/2022/neurips/A Unified Model for Multi-class Anomaly Detection create mode 100644 data/2022/neurips/A Unified Sequence Interface for Vision Tasks create mode 100644 data/2022/neurips/A Unifying Framework for Online Optimization with Long-Term Constraints create mode 100644 data/2022/neurips/A Unifying Framework of Off-Policy General Value Function Evaluation create mode 100644 data/2022/neurips/A Universal Error Measure for Input Predictions Applied to Online Graph Problems create mode 100644 data/2022/neurips/A Variant of Anderson Mixing with Minimal Memory Size create mode 100644 data/2022/neurips/A Variational Edge Partition Model for Supervised Graph Representation Learning create mode 100644 data/2022/neurips/A Win-win Deal: Towards Sparse and Robust Pre-trained Language Models create mode 100644 data/2022/neurips/A composable machine-learning approach for steady-state simulations on high-resolution grids create mode 100644 data/2022/neurips/A consistently adaptive trust-region method create mode 100644 data/2022/neurips/A contrastive rule for meta-learning create mode 100644 data/2022/neurips/A framework for bilevel optimization that enables stochastic and global variance reduction algorithms create mode 100644 data/2022/neurips/A gradient estimator via L1-randomization for online zero-order optimization with two point feedback create mode 100644 data/2022/neurips/A gradient sampling method with complexity guarantees for Lipschitz functions in high and low dimensions create mode 100644 data/2022/neurips/A new dataset for multilingual keyphrase generation create mode 100644 data/2022/neurips/A permutation-free kernel two-sample test create mode 100644 data/2022/neurips/A sharp NMF result with applications in network modeling create mode 100644 data/2022/neurips/A simple but strong baseline for online continual learning: Repeated Augmented Rehearsal create mode 100644 data/2022/neurips/A theory of weight distribution-constrained learning create mode 100644 data/2022/neurips/A time-resolved theory of information encoding in recurrent neural networks create mode 100644 data/2022/neurips/A2: Efficient Automated Attacker for Boosting Adversarial Training create mode 100644 data/2022/neurips/ACIL: Analytic Class-Incremental Learning with Absolute Memorization and Privacy Protection create mode 100644 data/2022/neurips/AD-DROP: Attribution-Driven Dropout for Robust Language Model Fine-Tuning create mode 100644 data/2022/neurips/ADBench: Anomaly Detection Benchmark create mode 100644 data/2022/neurips/ALIFE: Adaptive Logit Regularizer and Feature Replay for Incremental Semantic Segmentation create mode 100644 data/2022/neurips/ALMA: Hierarchical Learning for Composite Multi-Agent Tasks create mode 100644 data/2022/neurips/AMOS: A Large-Scale Abdominal Multi-Organ Benchmark for Versatile Medical Image Segmentation create mode 100644 data/2022/neurips/AMP: Automatically Finding Model Parallel Strategies with Heterogeneity Awareness create mode 100644 data/2022/neurips/APG: Adaptive Parameter Generation Network for Click-Through Rate Prediction create mode 100644 data/2022/neurips/APT-36K: A Large-scale Benchmark for Animal Pose Estimation and Tracking create mode 100644 data/2022/neurips/ASPiRe: Adaptive Skill Priors for Reinforcement Learning create mode 100644 data/2022/neurips/ATD: Augmenting CP Tensor Decomposition by Self Supervision create mode 100644 data/2022/neurips/AUTOMATA: Gradient Based Data Subset Selection for Compute-Efficient Hyper-parameter Tuning create mode 100644 data/2022/neurips/AVLEN: Audio-Visual-Language Embodied Navigation in 3D Environments create mode 100644 data/2022/neurips/AZ-whiteness test: a test for signal uncorrelation on spatio-temporal graphs create mode 100644 data/2022/neurips/Accelerated Linearized Laplace Approximation for Bayesian Deep Learning create mode 100644 data/2022/neurips/Accelerated Primal-Dual Gradient Method for Smooth and Convex-Concave Saddle-Point Problems with Bilinear Coupling create mode 100644 data/2022/neurips/Accelerated Projected Gradient Algorithms for Sparsity Constrained Optimization Problems create mode 100644 data/2022/neurips/Accelerated Training of Physics-Informed Neural Networks (PINNs) using Meshless Discretizations create mode 100644 data/2022/neurips/Accelerating Certified Robustness Training via Knowledge Transfer create mode 100644 data/2022/neurips/Accelerating SGD for Highly Ill-Conditioned Huge-Scale Online Matrix Completion create mode 100644 data/2022/neurips/Accelerating Sparse Convolution with Column Vector-Wise Sparsity create mode 100644 data/2022/neurips/Acceleration in Distributed Sparse Regression create mode 100644 data/2022/neurips/Action-modulated midbrain dopamine activity arises from distributed control policies create mode 100644 data/2022/neurips/ActionSense: A Multimodal Dataset and Recording Framework for Human Activities Using Wearable Sensors in a Kitchen Environment create mode 100644 data/2022/neurips/Active Bayesian Causal Inference create mode 100644 data/2022/neurips/Active Exploration for Inverse Reinforcement Learning create mode 100644 data/2022/neurips/Active Labeling: Streaming Stochastic Gradients create mode 100644 data/2022/neurips/Active Learning Helps Pretrained Models Learn the Intended Task create mode 100644 data/2022/neurips/Active Learning Polynomial Threshold Functions create mode 100644 data/2022/neurips/Active Learning Through a Covering Lens create mode 100644 data/2022/neurips/Active Learning for Multiple Target Models create mode 100644 data/2022/neurips/Active Learning of Classifiers with Label and Seed Queries create mode 100644 data/2022/neurips/Active Learning with Neural Networks: Insights from Nonparametric Statistics create mode 100644 data/2022/neurips/Active Learning with Safety Constraints create mode 100644 data/2022/neurips/Active Ranking without Strong Stochastic Transitivity create mode 100644 data/2022/neurips/Active Surrogate Estimators: An Active Learning Approach to Label-Efficient Model Evaluation create mode 100644 data/2022/neurips/Active-Passive SimStereo - Benchmarking the Cross-Generalization Capabilities of Deep Learning-based Stereo Methods create mode 100644 data/2022/neurips/AdaFocal: Calibration-aware Adaptive Focal Loss create mode 100644 data/2022/neurips/Adam Can Converge Without Any Modification On Update Rules create mode 100644 data/2022/neurips/AdaptFormer: Adapting Vision Transformers for Scalable Visual Recognition create mode 100644 data/2022/neurips/Adaptation Accelerating Sampling-based Bayesian Inference in Attractor Neural Networks create mode 100644 data/2022/neurips/Adapting Self-Supervised Vision Transformers by Probing Attention-Conditioned Masking Consistency create mode 100644 data/2022/neurips/Adapting to Online Label Shift with Provable Guarantees create mode 100644 data/2022/neurips/Adaptive Data Debiasing through Bounded Exploration create mode 100644 data/2022/neurips/Adaptive Distribution Calibration for Few-Shot Learning with Hierarchical Optimal Transport create mode 100644 data/2022/neurips/Adaptive Interest for Emphatic Reinforcement Learning create mode 100644 data/2022/neurips/Adaptive Multi-stage Density Ratio Estimation for Learning Latent Space Energy-based Model create mode 100644 data/2022/neurips/Adaptive Oracle-Efficient Online Learning create mode 100644 data/2022/neurips/Adaptive Sampling for Discovery create mode 100644 data/2022/neurips/Adaptive Stochastic Variance Reduction for Non-convex Finite-Sum Minimization create mode 100644 data/2022/neurips/Adaptively Exploiting d-Separators with Causal Bandits create mode 100644 data/2022/neurips/Additive MIL: Intrinsically Interpretable Multiple Instance Learning for Pathology create mode 100644 data/2022/neurips/Addressing Leakage in Concept Bottleneck Models create mode 100644 data/2022/neurips/Addressing Resource Scarcity across Sign Languages with Multilingual Pretraining and Unified-Vocabulary Datasets create mode 100644 data/2022/neurips/Adjoint-aided inference of Gaussian process driven differential equations create mode 100644 data/2022/neurips/Adv-Attribute: Inconspicuous and Transferable Adversarial Attack on Face Recognition create mode 100644 data/2022/neurips/Advancing Model Pruning via Bi-level Optimization create mode 100644 data/2022/neurips/Adversarial Attack on Attackers: Post-Process to Mitigate Black-Box Score-Based Query Attacks create mode 100644 data/2022/neurips/Adversarial Auto-Augment with Label Preservation: A Representation Learning Principle Guided Approach create mode 100644 data/2022/neurips/Adversarial Reprogramming Revisited create mode 100644 data/2022/neurips/Adversarial Robustness is at Odds with Lazy Training create mode 100644 data/2022/neurips/Adversarial Style Augmentation for Domain Generalized Urban-Scene Segmentation create mode 100644 data/2022/neurips/Adversarial Task Up-sampling for Meta-learning create mode 100644 data/2022/neurips/Adversarial Training with Complementary Labels: On the Benefit of Gradually Informative Attacks create mode 100644 data/2022/neurips/Adversarial Unlearning: Reducing Confidence Along Adversarial Directions create mode 100644 data/2022/neurips/Adversarial training for high-stakes reliability create mode 100644 data/2022/neurips/Adversarially Robust Learning: A Generic Minimax Optimal Learner and Characterization create mode 100644 data/2022/neurips/AgraSSt: Approximate Graph Stein Statistics for Interpretable Assessment of Implicit Graph Generators create mode 100644 data/2022/neurips/Agreement-on-the-line: Predicting the Performance of Neural Networks under Distribution Shift create mode 100644 data/2022/neurips/AirfRANS: High Fidelity Computational Fluid Dynamics Dataset for Approximating Reynolds-Averaged Navier-Stokes Solutions create mode 100644 data/2022/neurips/Algorithms and Hardness for Learning Linear Thresholds from Label Proportions create mode 100644 data/2022/neurips/Algorithms that Approximate Data Removal: New Results and Limitations create mode 100644 data/2022/neurips/Algorithms with Prediction Portfolios create mode 100644 data/2022/neurips/Align then Fusion: Generalized Large-scale Multi-view Clustering with Anchor Matching Correspondences create mode 100644 data/2022/neurips/Aligning individual brains with fused unbalanced Gromov Wasserstein create mode 100644 data/2022/neurips/Alignment-guided Temporal Attention for Video Action Recognition create mode 100644 data/2022/neurips/All Politics is Local: Redistricting via Local Fairness create mode 100644 "data/2022/neurips/Alleviating \"Posterior Collapse\" in Deep Topic Models via Policy Gradient" create mode 100644 data/2022/neurips/Alleviating Adversarial Attacks on Variational Autoencoders with MCMC create mode 100644 data/2022/neurips/Alleviating the Sample Selection Bias in Few-shot Learning by Removing Projection to the Centroid create mode 100644 data/2022/neurips/Alternating Mirror Descent for Constrained Min-Max Games create mode 100644 data/2022/neurips/Ambiguous Images With Human Judgments for Robust Visual Event Classification create mode 100644 data/2022/neurips/Amortized Inference for Causal Structure Learning create mode 100644 data/2022/neurips/Amortized Inference for Heterogeneous Reconstruction in Cryo-EM create mode 100644 data/2022/neurips/Amortized Mixing Coupling Processes for Clustering create mode 100644 data/2022/neurips/Amortized Projection Optimization for Sliced Wasserstein Generative Models create mode 100644 data/2022/neurips/Amortized Proximal Optimization create mode 100644 data/2022/neurips/Amplifying Membership Exposure via Data Poisoning create mode 100644 data/2022/neurips/An Adaptive Deep RL Method for Non-Stationary Environments with Piecewise Stable Context create mode 100644 data/2022/neurips/An Adaptive Kernel Approach to Federated Learning of Heterogeneous Causal Effects create mode 100644 data/2022/neurips/An Algorithm for Learning Switched Linear Dynamics from Data create mode 100644 data/2022/neurips/An Analysis of Ensemble Sampling create mode 100644 data/2022/neurips/An Analytical Theory of Curriculum Learning in Teacher-Student Networks create mode 100644 data/2022/neurips/An Asymptotically Optimal Batched Algorithm for the Dueling Bandit Problem create mode 100644 data/2022/neurips/An Embarrassingly Simple Approach to Semi-Supervised Few-Shot Learning create mode 100644 data/2022/neurips/An Empirical Study on Disentanglement of Negative-free Contrastive Learning create mode 100644 data/2022/neurips/An In-depth Study of Stochastic Backpropagation create mode 100644 data/2022/neurips/An Information-Theoretic Framework for Deep Learning create mode 100644 data/2022/neurips/An Investigation into Whitening Loss for Self-supervised Learning create mode 100644 data/2022/neurips/An efficient graph generative model for navigating ultra-large combinatorial synthesis libraries create mode 100644 data/2022/neurips/An empirical analysis of compute-optimal large language model training create mode 100644 data/2022/neurips/Analyzing Data-Centric Properties for Graph Contrastive Learning create mode 100644 data/2022/neurips/Analyzing Lottery Ticket Hypothesis from PAC-Bayesian Theory Perspective create mode 100644 data/2022/neurips/Analyzing Sharpness along GD Trajectory: Progressive Sharpening and Edge of Stability create mode 100644 data/2022/neurips/Anchor-Changing Regularized Natural Policy Gradient for Multi-Objective Reinforcement Learning create mode 100644 data/2022/neurips/AniFaceGAN: Animatable 3D-Aware Face Image Generation for Video Avatars create mode 100644 data/2022/neurips/AnimeRun: 2D Animation Visual Correspondence from Open Source 3D Movies create mode 100644 data/2022/neurips/AnimeSR: Learning Real-World Super-Resolution Models for Animation Videos create mode 100644 data/2022/neurips/Annihilation of Spurious Minima in Two-Layer ReLU Networks create mode 100644 data/2022/neurips/AnoShift: A Distribution Shift Benchmark for Unsupervised Anomaly Detection create mode 100644 data/2022/neurips/Anonymized Histograms in Intermediate Privacy Models create mode 100644 data/2022/neurips/Anonymous Bandits for Multi-User Systems create mode 100644 data/2022/neurips/Anticipating Performativity by Predicting from Predictions create mode 100644 data/2022/neurips/Antigen-Specific Antibody Design and Optimization with Diffusion-Based Generative Models for Protein Structures create mode 100644 data/2022/neurips/Anytime-Valid Inference For Multinomial Count Data create mode 100644 data/2022/neurips/Approaching Quartic Convergence Rates for Quasi-Stochastic Approximation with Application to Gradient-Free Optimization create mode 100644 data/2022/neurips/Approximate Euclidean lengths and distances beyond Johnson-Lindenstrauss create mode 100644 data/2022/neurips/Approximate Secular Equations for the Cubic Regularization Subproblem create mode 100644 data/2022/neurips/Approximate Value Equivalence create mode 100644 data/2022/neurips/Approximation with CNNs in Sobolev Space: with Applications to Classification create mode 100644 data/2022/neurips/Archimedes Meets Privacy: On Privately Estimating Quantiles in High Dimensions Under Minimal Assumptions create mode 100644 data/2022/neurips/Are All Losses Created Equal: A Neural Collapse Perspective create mode 100644 data/2022/neurips/Are AlphaZero-like Agents Robust to Adversarial Perturbations? create mode 100644 data/2022/neurips/Are Defenses for Graph Neural Networks Robust? create mode 100644 data/2022/neurips/Are GANs overkill for NLP? create mode 100644 data/2022/neurips/Are Two Heads the Same as One? Identifying Disparate Treatment in Fair Neural Networks create mode 100644 data/2022/neurips/Are You Stealing My Model? Sample Correlation for Fingerprinting Deep Neural Networks create mode 100644 data/2022/neurips/Are all Frames Equal? Active Sparse Labeling for Video Action Detection create mode 100644 data/2022/neurips/Ask4Help: Learning to Leverage an Expert for Embodied Tasks create mode 100644 data/2022/neurips/Assaying Out-Of-Distribution Generalization in Transfer Learning create mode 100644 data/2022/neurips/Assistive Teaching of Motor Control Tasks to Humans create mode 100644 data/2022/neurips/Associating Objects and Their Effects in Video through Coordination Games create mode 100644 data/2022/neurips/Association Graph Learning for Multi-Task Classification with Category Shifts create mode 100644 data/2022/neurips/Asymmetric Temperature Scaling Makes Larger Networks Teach Well Again create mode 100644 data/2022/neurips/Asymptotic Behaviors of Projected Stochastic Approximation: A Jump Diffusion Perspective create mode 100644 data/2022/neurips/Asymptotic Properties for Bayesian Neural Network in Besov Space create mode 100644 data/2022/neurips/Asymptotically Unbiased Instance-wise Regularized Partial AUC Optimization: Theory and Algorithm create mode 100644 data/2022/neurips/Asymptotics of smoothed Wasserstein distances in the small noise regime create mode 100644 "data/2022/neurips/Asymptotics of \342\204\2232 Regularized Network Embeddings" create mode 100644 data/2022/neurips/Asynchronous Actor-Critic for Multi-Agent Reinforcement Learning create mode 100644 data/2022/neurips/Asynchronous SGD Beats Minibatch SGD Under Arbitrary Delays create mode 100644 data/2022/neurips/AttCAT: Explaining Transformers via Attentive Class Activation Tokens create mode 100644 data/2022/neurips/Attention-based Neural Cellular Automata create mode 100644 data/2022/neurips/Attracting and Dispersing: A Simple Approach for Source-free Domain Adaptation create mode 100644 data/2022/neurips/Audio-Driven Co-Speech Gesture Video Generation create mode 100644 data/2022/neurips/Augmentations in Hypergraph Contrastive Learning: Fabricated and Generative create mode 100644 data/2022/neurips/Augmented RBMLE-UCB Approach for Adaptive Control of Linear Quadratic Systems create mode 100644 data/2022/neurips/AutoLink: Self-supervised Learning of Human Skeletons and Object Outlines by Linking Keypoints create mode 100644 data/2022/neurips/AutoML Two-Sample Test create mode 100644 data/2022/neurips/AutoMS: Automatic Model Selection for Novelty Detection with Error Rate Control create mode 100644 data/2022/neurips/AutoMTL: A Programming Framework for Automating Efficient Multi-Task Learning create mode 100644 data/2022/neurips/AutoST: Towards the Universal Modeling of Spatio-temporal Sequences create mode 100644 data/2022/neurips/AutoWS-Bench-101: Benchmarking Automated Weak Supervision with 100 Labels create mode 100644 data/2022/neurips/Autoformalization with Large Language Models create mode 100644 data/2022/neurips/Autoinverse: Uncertainty Aware Inversion of Neural Networks create mode 100644 data/2022/neurips/Automatic Differentiation of Programs with Discrete Randomness create mode 100644 data/2022/neurips/Automatic differentiation of nonsmooth iterative algorithms create mode 100644 data/2022/neurips/Autoregressive Perturbations for Data Poisoning create mode 100644 data/2022/neurips/Autoregressive Search Engines: Generating Substrings as Document Identifiers create mode 100644 data/2022/neurips/Avalon: A Benchmark for RL Generalization Using Procedurally Generated Worlds create mode 100644 data/2022/neurips/Average Sensitivity of Euclidean k-Clustering create mode 100644 data/2022/neurips/BEER: Fast $O(1 T)$ Rate for Decentralized Nonconvex Optimization with Communication Compression create mode 100644 data/2022/neurips/BEVFusion: A Simple and Robust LiDAR-Camera Fusion Framework create mode 100644 data/2022/neurips/BILCO: An Efficient Algorithm for Joint Alignment of Time Series create mode 100644 data/2022/neurips/BLOX: Macro Neural Architecture Search Benchmark and Algorithms create mode 100644 data/2022/neurips/BMU-MoCo: Bidirectional Momentum Update for Continual Video-Language Modeling create mode 100644 data/2022/neurips/BOME! Bilevel Optimization Made Easy: A Simple First-Order Approach create mode 100644 data/2022/neurips/BOND: Benchmarking Unsupervised Outlier Node Detection on Static Attributed Graphs create mode 100644 data/2022/neurips/BR-SNIS: Bias Reduced Self-Normalized Importance Sampling create mode 100644 data/2022/neurips/BYOL-Explore: Exploration by Bootstrapped Prediction create mode 100644 data/2022/neurips/Back Razor: Memory-Efficient Transfer Learning by Self-Sparsified Backpropagation create mode 100644 data/2022/neurips/BackdoorBench: A Comprehensive Benchmark of Backdoor Learning create mode 100644 data/2022/neurips/BadPrompt: Backdoor Attacks on Continuous Prompts create mode 100644 data/2022/neurips/BagFlip: A Certified Defense Against Data Poisoning create mode 100644 data/2022/neurips/Bandit Theory and Thompson Sampling-Guided Directed Evolution for Sequence Optimization create mode 100644 data/2022/neurips/Batch Bayesian Optimization on Permutations using the Acquisition Weighted Kernel create mode 100644 data/2022/neurips/Batch Bayesian optimisation via density-ratio estimation with guarantees create mode 100644 data/2022/neurips/Batch Multi-Fidelity Active Learning with Budget Constraints create mode 100644 data/2022/neurips/Batch size-invariance for policy optimization create mode 100644 data/2022/neurips/Batch-Size Independent Regret Bounds for Combinatorial Semi-Bandits with Probabilistically Triggered Arms or Independent Arms create mode 100644 data/2022/neurips/BayesPCN: A Continually Learnable Predictive Coding Associative Memory create mode 100644 data/2022/neurips/Bayesian Active Learning with Fully Bayesian Gaussian Processes create mode 100644 data/2022/neurips/Bayesian Clustering of Neural Spiking Activity Using a Mixture of Dynamic Poisson Factor Analyzers create mode 100644 data/2022/neurips/Bayesian Optimistic Optimization: Optimistic Exploration for Model-based Reinforcement Learning create mode 100644 data/2022/neurips/Bayesian Optimization over Discrete and Mixed Spaces via Probabilistic Reparameterization create mode 100644 data/2022/neurips/Bayesian Persuasion for Algorithmic Recourse create mode 100644 data/2022/neurips/Bayesian Risk Markov Decision Processes create mode 100644 data/2022/neurips/Bayesian Spline Learning for Equation Discovery of Nonlinear Dynamics with Quantified Uncertainty create mode 100644 data/2022/neurips/Bayesian inference via sparse Hamiltonian flows create mode 100644 data/2022/neurips/Behavior Transformers: Cloning $k$ modes with one stone create mode 100644 data/2022/neurips/Bellman Residual Orthogonalization for Offline Reinforcement Learning create mode 100644 data/2022/neurips/Benchmarking Heterogeneous Treatment Effect Models through the Lens of Interpretability create mode 100644 data/2022/neurips/Benchmarking and Analyzing 3D Human Pose and Shape Estimation Beyond Algorithms create mode 100644 data/2022/neurips/Benchopt: Reproducible, efficient and collaborative optimization benchmarks create mode 100644 data/2022/neurips/Benefits of Additive Noise in Composing Classes with Bounded Capacity create mode 100644 data/2022/neurips/Benefits of Permutation-Equivariance in Auction Mechanisms create mode 100644 data/2022/neurips/Benign Overfitting in Two-layer Convolutional Neural Networks create mode 100644 data/2022/neurips/Benign Underfitting of Stochastic Gradient Descent create mode 100644 data/2022/neurips/Benign, Tempered, or Catastrophic: Toward a Refined Taxonomy of Overfitting create mode 100644 data/2022/neurips/Bessel Equivariant Networks for Inversion of Transmission Effects in Multi-Mode Optical Fibres create mode 100644 data/2022/neurips/Best of Both Worlds Model Selection create mode 100644 data/2022/neurips/Better Best of Both Worlds Bounds for Bandits with Switching Costs create mode 100644 data/2022/neurips/Better SGD using Second-order Momentum create mode 100644 data/2022/neurips/Better Uncertainty Calibration via Proper Scores for Classification and Beyond create mode 100644 data/2022/neurips/Between Stochastic and Adversarial Online Convex Optimization: Improved Regret Bounds via Smoothness create mode 100644 data/2022/neurips/Beyond Adult and COMPAS: Fair Multi-Class Prediction via Information Projection create mode 100644 data/2022/neurips/Beyond IID: data-driven decision-making in heterogeneous environments create mode 100644 data/2022/neurips/Beyond L1: Faster and Better Sparse Models with skglm create mode 100644 data/2022/neurips/Beyond Mahalanobis Distance for Textual OOD Detection create mode 100644 data/2022/neurips/Beyond Not-Forgetting: Continual Learning with Backward Knowledge Transfer create mode 100644 data/2022/neurips/Beyond Real-world Benchmark Datasets: An Empirical Study of Node Classification with GNNs create mode 100644 data/2022/neurips/Beyond Rewards: a Hierarchical Perspective on Offline Multiagent Behavioral Analysis create mode 100644 data/2022/neurips/Beyond Separability: Analyzing the Linear Transferability of Contrastive Representations to Related Subpopulations create mode 100644 data/2022/neurips/Beyond Time-Average Convergence: Near-Optimal Uncoupled Online Learning via Clairvoyant Multiplicative Weights Update create mode 100644 data/2022/neurips/Beyond accuracy: generalization properties of bio-plausible temporal credit assignment rules create mode 100644 data/2022/neurips/Beyond black box densities: Parameter learning for the deviated components create mode 100644 data/2022/neurips/Beyond neural scaling laws: beating power law scaling via data pruning create mode 100644 data/2022/neurips/Beyond spectral gap: the role of the topology in decentralized learning create mode 100644 data/2022/neurips/Beyond the Best: Distribution Functional Estimation in Infinite-Armed Bandits create mode 100644 data/2022/neurips/Beyond the Return: Off-policy Function Estimation under User-specified Error-measuring Distributions create mode 100644 data/2022/neurips/Bezier Gaussian Processes for Tall and Wide Data create mode 100644 data/2022/neurips/Bi-directional Weakly Supervised Knowledge Distillation for Whole Slide Image Classification create mode 100644 data/2022/neurips/BiMLP: Compact Binary Architectures for Vision Multi-Layer Perceptrons create mode 100644 data/2022/neurips/BiT: Robustly Binarized Multi-distilled Transformer create mode 100644 data/2022/neurips/Bidirectional Learning for Offline Infinite-width Model-based Optimization create mode 100644 data/2022/neurips/BigBio: A Framework for Data-Centric Biomedical Natural Language Processing create mode 100644 data/2022/neurips/BinauralGrad: A Two-Stage Conditional Diffusion Probabilistic Model for Binaural Audio Synthesis create mode 100644 data/2022/neurips/Biological Learning of Irreducible Representations of Commuting Transformations create mode 100644 data/2022/neurips/Biologically Inspired Dynamic Thresholds for Spiking Neural Networks create mode 100644 data/2022/neurips/Biologically plausible solutions for spiking networks with efficient coding create mode 100644 data/2022/neurips/Biologically-Plausible Determinant Maximization Neural Networks for Blind Separation of Correlated Sources create mode 100644 data/2022/neurips/Biologically-plausible backpropagation through arbitrary timespans via local neuromodulators create mode 100644 data/2022/neurips/Bivariate Causal Discovery for Categorical Data via Classification with Optimal Label Permutation create mode 100644 data/2022/neurips/Black-Box Generalization: Stability of Zeroth-Order Learning create mode 100644 data/2022/neurips/Black-box coreset variational inference create mode 100644 data/2022/neurips/Blackbox Attacks via Surrogate Ensemble Search create mode 100644 data/2022/neurips/Blessing of Depth in Linear Regression: Deeper Models Have Flatter Landscape Around the True Solution create mode 100644 data/2022/neurips/Block-Recurrent Transformers create mode 100644 data/2022/neurips/Boosting Barely Robust Learners: A New Perspective on Adversarial Robustness create mode 100644 data/2022/neurips/Boosting Out-of-distribution Detection with Typical Features create mode 100644 data/2022/neurips/Boosting the Performance of Generic Deep Neural Network Frameworks with Log-supermodular CRFs create mode 100644 data/2022/neurips/Boosting the Transferability of Adversarial Attacks with Reverse Adversarial Perturbation create mode 100644 data/2022/neurips/Bootstrapped Transformer for Offline Reinforcement Learning create mode 100644 data/2022/neurips/Bounded-Regret MPC via Perturbation Analysis: Prediction Error, Constraints, and Nonlinearity create mode 100644 data/2022/neurips/Bounding and Approximating Intersectional Fairness through Marginal Fairness create mode 100644 data/2022/neurips/Brain Network Transformer create mode 100644 data/2022/neurips/Branch & Learn for Recursively and Iteratively Solvable Problems in Predict+Optimize create mode 100644 data/2022/neurips/Breaking Bad: A Dataset for Geometric Fracture and Reassembly create mode 100644 data/2022/neurips/Bridge the Gap Between Architecture Spaces via A Cross-Domain Predictor create mode 100644 data/2022/neurips/Bridging Central and Local Differential Privacy in Data Acquisition Mechanisms create mode 100644 data/2022/neurips/Bridging the Gap Between Vision Transformers and Convolutional Neural Networks on Small Datasets create mode 100644 data/2022/neurips/Bridging the Gap between Object and Image-level Representations for Open-Vocabulary Detection create mode 100644 data/2022/neurips/Bridging the Gap from Asymmetry Tricks to Decorrelation Principles in Non-contrastive Self-supervised Learning create mode 100644 data/2022/neurips/Bridging the Gap: Unifying the Training and Evaluation of Neural Network Binary Classifiers create mode 100644 data/2022/neurips/Bring Your Own Algorithm for Optimal Differentially Private Stochastic Minimax Optimization create mode 100644 data/2022/neurips/Bringing Image Scene Structure to Video via Frame-Clip Consistency of Object Tokens create mode 100644 data/2022/neurips/Brownian Noise Reduction: Maximizing Privacy Subject to Accuracy Constraints create mode 100644 data/2022/neurips/Byzantine Spectral Ranking create mode 100644 data/2022/neurips/Byzantine-tolerant federated Gaussian process regression for streaming data create mode 100644 data/2022/neurips/C-Mixup: Improving Generalization in Regression create mode 100644 data/2022/neurips/C2FAR: Coarse-to-Fine Autoregressive Networks for Precise Probabilistic Forecasting create mode 100644 data/2022/neurips/CAESAR: An Embodied Simulator for Generating Multimodal Referring Expression Datasets create mode 100644 data/2022/neurips/CAGroup3D: Class-Aware Grouping for 3D Object Detection on Point Clouds create mode 100644 data/2022/neurips/CARD: Classification and Regression Diffusion Models create mode 100644 data/2022/neurips/CARLANE: A Lane Detection Benchmark for Unsupervised Domain Adaptation from Simulation to multiple Real-World Domains create mode 100644 data/2022/neurips/CASA: Category-agnostic Skeletal Animal Reconstruction create mode 100644 data/2022/neurips/CATER: Intellectual Property Protection on Text Generation APIs via Conditional Watermarks create mode 100644 data/2022/neurips/CCCP is Frank-Wolfe in disguise create mode 100644 data/2022/neurips/CEBaB: Estimating the Causal Effects of Real-World Concepts on NLP Model Behavior create mode 100644 data/2022/neurips/CEDe: A collection of expert-curated datasets with atom-level entity annotations for Optical Chemical Structure Recognition create mode 100644 data/2022/neurips/CEIP: Combining Explicit and Implicit Priors for Reinforcement Learning with Demonstrations create mode 100644 data/2022/neurips/CGLB: Benchmark Tasks for Continual Graph Learning create mode 100644 data/2022/neurips/CHIMLE: Conditional Hierarchical IMLE for Multimodal Conditional Image Synthesis create mode 100644 data/2022/neurips/CLEAR: Generative Counterfactual Explanations on Graphs create mode 100644 data/2022/neurips/CLEVRER-Humans: Describing Physical and Causal Events the Human Way create mode 100644 data/2022/neurips/CLIPDraw: Exploring Text-to-Drawing Synthesis through Language-Image Encoders create mode 100644 data/2022/neurips/CLOOB: Modern Hopfield Networks with InfoLOOB Outperform CLIP create mode 100644 data/2022/neurips/CLiMB: A Continual Learning Benchmark for Vision-and-Language Tasks create mode 100644 data/2022/neurips/COLD Decoding: Energy-based Constrained Text Generation with Langevin Dynamics create mode 100644 data/2022/neurips/CS-Shapley: Class-wise Shapley Values for Data Valuation in Classification create mode 100644 data/2022/neurips/CUP: Critic-Guided Policy Reuse create mode 100644 data/2022/neurips/Cache-Augmented Inbatch Importance Resampling for Training Recommender Retriever create mode 100644 data/2022/neurips/CageNeRF: Cage-based Neural Radiance Field for Generalized 3D Deformation and Animation create mode 100644 data/2022/neurips/CalFAT: Calibrated Federated Adversarial Training with Label Skewness create mode 100644 data/2022/neurips/Calibrated Data-Dependent Constraints with Exact Satisfaction Guarantees create mode 100644 data/2022/neurips/Can Adversarial Training Be Manipulated By Non-Robust Features? create mode 100644 data/2022/neurips/Can Hybrid Geometric Scattering Networks Help Solve the Maximum Clique Problem? create mode 100644 data/2022/neurips/Can Push-forward Generative Models Fit Multimodal Distributions? create mode 100644 data/2022/neurips/Capturing Failures of Large Language Models via Human Cognitive Biases create mode 100644 data/2022/neurips/Capturing Graphs with Hypo-Elliptic Diffusions create mode 100644 data/2022/neurips/CascadeXML: Rethinking Transformers for End-to-end Multi-resolution Training in Extreme Multi-label Classification create mode 100644 data/2022/neurips/Category-Level 6D Object Pose Estimation in the Wild: A Semi-Supervised Learning Approach and A New Dataset create mode 100644 data/2022/neurips/Causal Discovery in Heterogeneous Environments Under the Sparse Mechanism Shift Hypothesis create mode 100644 data/2022/neurips/Causal Discovery in Linear Latent Variable Models Subject to Measurement Error create mode 100644 data/2022/neurips/Causal Identification under Markov equivalence: Calculus, Algorithm, and Completeness create mode 100644 data/2022/neurips/Causal Inference with Non-IID Data using Linear Graphical Models create mode 100644 data/2022/neurips/Causality Preserving Chaotic Transformation and Classification using Neurochaos Learning create mode 100644 data/2022/neurips/Causality-driven Hierarchical Structure Discovery for Reinforcement Learning create mode 100644 data/2022/neurips/Causally motivated multi-shortcut identification and removal create mode 100644 data/2022/neurips/Censored Quantile Regression Neural Networks for Distribution-Free Survival Analysis create mode 100644 data/2022/neurips/Certifying Robust Graph Classification under Orthogonal Gromov-Wasserstein Threats create mode 100644 data/2022/neurips/Certifying Some Distributional Fairness with Subpopulation Decomposition create mode 100644 data/2022/neurips/Chain of Thought Imitation with Procedure Cloning create mode 100644 data/2022/neurips/Chain-of-Thought Prompting Elicits Reasoning in Large Language Models create mode 100644 data/2022/neurips/Challenging Common Assumptions in Convex Reinforcement Learning create mode 100644 data/2022/neurips/Change Event Dataset for Discovery from Spatio-temporal Remote Sensing Imagery create mode 100644 data/2022/neurips/Change-point Detection for Sparse and Dense Functional Data in General Dimensions create mode 100644 data/2022/neurips/Chaotic Dynamics are Intrinsic to Neural Network Training with SGD create mode 100644 data/2022/neurips/Chaotic Regularization and Heavy-Tailed Limits for Deterministic Gradient Descent create mode 100644 data/2022/neurips/Characteristics of Harmful Text: Towards Rigorous Benchmarking of Language Models create mode 100644 data/2022/neurips/Characterization of Excess Risk for Locally Strongly Convex Population Risk create mode 100644 data/2022/neurips/Characterizing the Ventral Visual Stream with Response-Optimized Neural Encoding Models create mode 100644 data/2022/neurips/Chartalist: Labeled Graph Datasets for UTXO and Account-based Blockchains create mode 100644 data/2022/neurips/Chefs' Random Tables: Non-Trigonometric Random Features create mode 100644 data/2022/neurips/Chroma-VAE: Mitigating Shortcut Learning with Generative Classifiers create mode 100644 data/2022/neurips/Chromatic Correlation Clustering, Revisited create mode 100644 data/2022/neurips/Class-Aware Adversarial Transformers for Medical Image Segmentation create mode 100644 data/2022/neurips/Class-Dependent Label-Noise Learning with Cycle-Consistency Regularization create mode 100644 data/2022/neurips/ClimbQ: Class Imbalanced Quantization Enabling Robustness on Efficient Inferences create mode 100644 data/2022/neurips/Clipped Stochastic Methods for Variational Inequalities with Heavy-Tailed Noise create mode 100644 data/2022/neurips/Cluster Randomized Designs for One-Sided Bipartite Experiments create mode 100644 data/2022/neurips/Cluster and Aggregate: Face Recognition with Large Probe Set create mode 100644 data/2022/neurips/Co-Modality Graph Contrastive Learning for Imbalanced Node Classification create mode 100644 data/2022/neurips/CoNSoLe: Convex Neural Symbolic Learning create mode 100644 data/2022/neurips/CoNT: Contrastive Neural Text Generation create mode 100644 data/2022/neurips/CoPur: Certifiably Robust Collaborative Inference via Feature Purification create mode 100644 data/2022/neurips/Coarse-to-Fine Vision-Language Pre-training with Fusion in the Backbone create mode 100644 data/2022/neurips/CodeRL: Mastering Code Generation through Pretrained Models and Deep Reinforcement Learning create mode 100644 data/2022/neurips/Coded Residual Transform for Generalizable Deep Metric Learning create mode 100644 data/2022/neurips/CogView2: Faster and Better Text-to-Image Generation via Hierarchical Transformers create mode 100644 data/2022/neurips/Collaborative Decision Making Using Action Suggestions create mode 100644 data/2022/neurips/Collaborative Learning by Detecting Collaboration Partners create mode 100644 data/2022/neurips/Collaborative Learning of Discrete Distributions under Heterogeneity and Communication Constraints create mode 100644 data/2022/neurips/Collaborative Linear Bandits with Adversarial Agents: Near-Optimal Regret Bounds create mode 100644 data/2022/neurips/ComENet: Towards Complete and Efficient Message Passing for 3D Molecular Graphs create mode 100644 data/2022/neurips/ComGAN: Unsupervised Disentanglement and Segmentation via Image Composition create mode 100644 data/2022/neurips/ComMU: Dataset for Combinatorial Music Generation create mode 100644 data/2022/neurips/Combinatorial Bandits with Linear Constraints: Beyond Knapsacks and Fairness create mode 100644 data/2022/neurips/Combining Explicit and Implicit Regularization for Efficient Learning in Deep Networks create mode 100644 data/2022/neurips/Communicating Natural Programs to Humans and Machines create mode 100644 data/2022/neurips/Communication Acceleration of Local Gradient Methods via an Accelerated Primal-Dual Algorithm with an Inexact Prox create mode 100644 data/2022/neurips/Communication Efficient Distributed Learning for Kernelized Contextual Bandits create mode 100644 data/2022/neurips/Communication Efficient Federated Learning for Generalized Linear Bandits create mode 100644 data/2022/neurips/Communication-Efficient Topologies for Decentralized Learning with $O(1)$ Consensus Rate create mode 100644 data/2022/neurips/Communication-efficient distributed eigenspace estimation with arbitrary node failures create mode 100644 data/2022/neurips/Composite Feature Selection Using Deep Ensembles create mode 100644 data/2022/neurips/Composition Theorems for Interactive Differential Privacy create mode 100644 data/2022/neurips/Compositional Generalization in Unsupervised Compositional Representation Learning: A Study on Disentanglement and Emergent Language create mode 100644 data/2022/neurips/Compositional generalization through abstract representations in human and artificial neural networks create mode 100644 data/2022/neurips/Compressible-composable NeRF via Rank-residual Decomposition create mode 100644 data/2022/neurips/Computationally Efficient Horizon-Free Reinforcement Learning for Linear Mixture MDPs create mode 100644 data/2022/neurips/Concentration of Data Encoding in Parameterized Quantum Circuits create mode 100644 data/2022/neurips/Concept Activation Regions: A Generalized Framework For Concept-Based Explanations create mode 100644 data/2022/neurips/Concept Embedding Models: Beyond the Accuracy-Explainability Trade-Off create mode 100644 data/2022/neurips/Concrete Score Matching: Generalized Score Matching for Discrete Data create mode 100644 data/2022/neurips/Conditional Diffusion Process for Inverse Halftoning create mode 100644 data/2022/neurips/Conditional Independence Testing with Heteroskedastic Data and Applications to Causal Discovery create mode 100644 data/2022/neurips/Conditional Meta-Learning of Linear Representations create mode 100644 data/2022/neurips/ConfLab: A Data Collection Concept, Dataset, and Benchmark for Machine Analysis of Free-Standing Social Interactions in the Wild create mode 100644 data/2022/neurips/Confidence-based Reliable Learning under Dual Noises create mode 100644 data/2022/neurips/Confident Adaptive Language Modeling create mode 100644 data/2022/neurips/Conformal Frequency Estimation with Sketched Data create mode 100644 data/2022/neurips/Conformal Off-Policy Prediction in Contextual Bandits create mode 100644 data/2022/neurips/Conformal Prediction with Temporal Quantile Adjustments create mode 100644 data/2022/neurips/Conformalized Fairness via Quantile Regression create mode 100644 data/2022/neurips/ConfounderGAN: Protecting Image Data Privacy with Causal Confounder create mode 100644 data/2022/neurips/Conservative Dual Policy Optimization for Efficient Model-Based Reinforcement Learning create mode 100644 data/2022/neurips/Consistency of Constrained Spectral Clustering under Graph Induced Fair Planted Partitions create mode 100644 data/2022/neurips/Consistent Interpolating Ensembles via the Manifold-Hilbert Kernel create mode 100644 data/2022/neurips/Consistent Sufficient Explanations and Minimal Local Rules for explaining the decision of any classifier or regressor create mode 100644 data/2022/neurips/Constants of motion network create mode 100644 data/2022/neurips/Constrained GPI for Zero-Shot Transfer in Reinforcement Learning create mode 100644 data/2022/neurips/Constrained Langevin Algorithms with L-mixing External Random Variables create mode 100644 data/2022/neurips/Constrained Predictive Coding as a Biologically Plausible Model of the Cortical Hierarchy create mode 100644 data/2022/neurips/Constrained Stochastic Nonconvex Optimization with State-dependent Markov Data create mode 100644 data/2022/neurips/Constrained Update Projection Approach to Safe Policy Optimization create mode 100644 data/2022/neurips/Constraining Gaussian Processes to Systems of Linear Ordinary Differential Equations create mode 100644 data/2022/neurips/Contact-aware Human Motion Forecasting create mode 100644 data/2022/neurips/Context-Based Dynamic Pricing with Partially Linear Demand Model create mode 100644 data/2022/neurips/Contextual Bandits with Knapsacks for a Conversion Model create mode 100644 data/2022/neurips/Contextual Dynamic Pricing with Unknown Noise: Explore-then-UCB Strategy and Improved Regrets create mode 100644 data/2022/neurips/Contextual Squeeze-and-Excitation for Efficient Few-Shot Image Classification create mode 100644 data/2022/neurips/Continual Learning In Environments With Polynomial Mixing Times create mode 100644 data/2022/neurips/Continual Learning with Evolving Class Ontologies create mode 100644 data/2022/neurips/Continual learning: a feature extraction formalization, an efficient algorithm, and fundamental obstructions create mode 100644 data/2022/neurips/Continuous Deep Q-Learning in Optimal Control Problems: Normalized Advantage Functions Analysis create mode 100644 data/2022/neurips/Continuous MDP Homomorphisms and Homomorphic Policy Gradient create mode 100644 data/2022/neurips/Continuously Tempered PDMP samplers create mode 100644 data/2022/neurips/Contrastive Adapters for Foundation Model Group Robustness create mode 100644 data/2022/neurips/Contrastive Graph Structure Learning via Information Bottleneck for Recommendation create mode 100644 data/2022/neurips/Contrastive Language-Image Pre-Training with Knowledge Graphs create mode 100644 data/2022/neurips/Contrastive Learning as Goal-Conditioned Reinforcement Learning create mode 100644 data/2022/neurips/Contrastive Neural Ratio Estimation create mode 100644 data/2022/neurips/Contrastive and Non-Contrastive Self-Supervised Learning Recover Global and Local Spectral Embedding Methods create mode 100644 data/2022/neurips/Controllable 3D Face Synthesis with Conditional Generative Occupancy Fields create mode 100644 data/2022/neurips/Controllable Text Generation with Neurally-Decomposed Oracle create mode 100644 data/2022/neurips/Controlled Sparsity via Constrained Optimization or: How I Learned to Stop Tuning Penalties and Love Constraints create mode 100644 data/2022/neurips/Convergence beyond the over-parameterized regime using Rayleigh quotients create mode 100644 data/2022/neurips/Convergence for score-based generative modeling with polynomial complexity create mode 100644 data/2022/neurips/Convergent Representations of Computer Programs in Human and Artificial Neural Networks create mode 100644 data/2022/neurips/Convexity Certificates from Hessians create mode 100644 data/2022/neurips/Convolutional Neural Networks on Graphs with Chebyshev Approximation, Revisited create mode 100644 data/2022/neurips/Cooperative Distribution Alignment via JSD Upper Bound create mode 100644 data/2022/neurips/Coordinate Linear Variance Reduction for Generalized Linear Programming create mode 100644 data/2022/neurips/Coordinates Are NOT Lonely - Codebook Prior Helps Implicit Neural 3D representations create mode 100644 data/2022/neurips/Coreset for Line-Sets Clustering create mode 100644 data/2022/neurips/Coresets for Relational Data and The Applications create mode 100644 data/2022/neurips/Coresets for Vertical Federated Learning: Regularized Linear Regression and $K$-Means Clustering create mode 100644 data/2022/neurips/Coresets for Wasserstein Distributionally Robust Optimization Problems create mode 100644 data/2022/neurips/Cost-Sensitive Self-Training for Optimizing Non-Decomposable Metrics create mode 100644 data/2022/neurips/Cost-efficient Gaussian tensor network embeddings for tensor-structured inputs create mode 100644 data/2022/neurips/Could Giant Pre-trained Image Models Extract Universal Representations? create mode 100644 data/2022/neurips/Counterfactual Fairness with Partially Known Causal Graph create mode 100644 data/2022/neurips/Counterfactual Neural Temporal Point Process for Estimating Causal Influence of Misinformation on Social Media create mode 100644 data/2022/neurips/Counterfactual Temporal Point Processes create mode 100644 data/2022/neurips/Counterfactual harm create mode 100644 data/2022/neurips/CoupAlign: Coupling Word-Pixel with Sentence-Mask Alignments for Referring Image Segmentation create mode 100644 data/2022/neurips/CroCo: Self-Supervised Pre-training for 3D Vision Tasks by Cross-View Completion create mode 100644 data/2022/neurips/Cross Aggregation Transformer for Image Restoration create mode 100644 data/2022/neurips/Cross-Image Context for Single Image Inpainting create mode 100644 data/2022/neurips/Cross-Linked Unified Embedding for cross-modality representation learning create mode 100644 data/2022/neurips/Cross-modal Learning for Image-Guided Point Cloud Shape Completion create mode 100644 data/2022/neurips/CryptoGCN: Fast and Scalable Homomorphically Encrypted Graph Convolutional Network Inference create mode 100644 data/2022/neurips/Cryptographic Hardness of Learning Halfspaces with Massart Noise create mode 100644 data/2022/neurips/Curious Exploration via Structured World Models Yields Zero-Shot Object Manipulation create mode 100644 data/2022/neurips/Curriculum Reinforcement Learning using Optimal Transport via Gradual Domain Adaptation create mode 100644 data/2022/neurips/CyCLIP: Cyclic Contrastive Language-Image Pretraining create mode 100644 data/2022/neurips/DABS 2.0: Improved Datasets and Algorithms for Universal Self-Supervision create mode 100644 data/2022/neurips/DAGMA: Learning DAGs via M-matrices and a Log-Determinant Acyclicity Characterization create mode 100644 data/2022/neurips/DARE: Disentanglement-Augmented Rationale Extraction create mode 100644 data/2022/neurips/DART: Articulated Hand Model with Diverse Accessories and Rich Textures create mode 100644 data/2022/neurips/DASCO: Dual-Generator Adversarial Support Constrained Offline Reinforcement Learning create mode 100644 data/2022/neurips/DC-BENCH: Dataset Condensation Benchmark create mode 100644 data/2022/neurips/DDXPlus: A New Dataset For Automatic Medical Diagnosis create mode 100644 data/2022/neurips/DENSE: Data-Free One-Shot Federated Learning create mode 100644 data/2022/neurips/DGraph: A Large-Scale Financial Dataset for Graph Anomaly Detection create mode 100644 data/2022/neurips/DHRL: A Graph-Based Approach for Long-Horizon and Sparse Hierarchical Reinforcement Learning create mode 100644 data/2022/neurips/DIMES: A Differentiable Meta Solver for Combinatorial Optimization Problems create mode 100644 data/2022/neurips/DISCO: Adversarial Defense with Local Implicit Functions create mode 100644 data/2022/neurips/DMAP: a Distributed Morphological Attention Policy for learning to locomote with a changing body create mode 100644 data/2022/neurips/DNA: Proximal Policy Optimization with a Dual Network Architecture create mode 100644 data/2022/neurips/DOMINO: Decomposed Mutual Information Optimization for Generalized Context in Meta-Reinforcement Learning create mode 100644 data/2022/neurips/DOPE: Doubly Optimistic and Pessimistic Exploration for Safe Reinforcement Learning create mode 100644 data/2022/neurips/DP-PCA: Statistically Optimal and Differentially Private PCA create mode 100644 data/2022/neurips/DPM-Solver: A Fast ODE Solver for Diffusion Probabilistic Model Sampling in Around 10 Steps create mode 100644 data/2022/neurips/DReS-FL: Dropout-Resilient Secure Federated Learning for Non-IID Clients via Secret Data Sharing create mode 100644 data/2022/neurips/DTG-SSOD: Dense Teacher Guidance for Semi-Supervised Object Detection create mode 100644 data/2022/neurips/DaDA: Distortion-aware Domain Adaptation for Unsupervised Semantic Segmentation create mode 100644 data/2022/neurips/Dance of SNN and ANN: Solving binding problem by combining spike timing and reconstructive attention create mode 100644 data/2022/neurips/Data Augmentation MCMC for Bayesian Inference from Privatized Data create mode 100644 data/2022/neurips/Data Augmentation for Compositional Data: Advancing Predictive Models of the Microbiome create mode 100644 data/2022/neurips/Data Distributional Properties Drive Emergent In-Context Learning in Transformers create mode 100644 data/2022/neurips/Data augmentation for efficient learning from parametric experts create mode 100644 data/2022/neurips/Data-Driven Conditional Robust Optimization create mode 100644 data/2022/neurips/Data-Driven Offline Decision-Making via Invariant Representation Learning create mode 100644 data/2022/neurips/Data-Efficient Augmentation for Training Neural Networks create mode 100644 data/2022/neurips/Data-Efficient Pipeline for Offline Reinforcement Learning with Limited Data create mode 100644 data/2022/neurips/Data-Efficient Structured Pruning via Submodular Optimization create mode 100644 data/2022/neurips/Data-IQ: Characterizing subgroups with heterogeneous outcomes in tabular data create mode 100644 data/2022/neurips/DataMUX: Data Multiplexing for Neural Networks create mode 100644 data/2022/neurips/Dataset Distillation using Neural Feature Regression create mode 100644 data/2022/neurips/Dataset Distillation via Factorization create mode 100644 data/2022/neurips/Dataset Inference for Self-Supervised Models create mode 100644 data/2022/neurips/DeVRF: Fast Deformable Voxel Radiance Fields for Dynamic Scenes create mode 100644 data/2022/neurips/Debiased Causal Tree: Heterogeneous Treatment Effects Estimation with Unmeasured Confounding create mode 100644 data/2022/neurips/Debiased Machine Learning without Sample-Splitting for Stable Estimators create mode 100644 data/2022/neurips/Debiased Self-Training for Semi-Supervised Learning create mode 100644 data/2022/neurips/Debiased, Longitudinal and Coordinated Drug Recommendation through Multi-Visit Clinic Records create mode 100644 data/2022/neurips/Debiasing Graph Neural Networks via Learning Disentangled Causal Substructure create mode 100644 data/2022/neurips/Debugging and Explaining Metric Learning Approaches: An Influence Function Based Perspective create mode 100644 data/2022/neurips/Decentralized Gossip-Based Stochastic Bilevel Optimization over Communication Networks create mode 100644 data/2022/neurips/Decentralized Local Stochastic Extra-Gradient for Variational Inequalities create mode 100644 data/2022/neurips/Decentralized Training of Foundation Models in Heterogeneous Environments create mode 100644 data/2022/neurips/Decentralized, Communication- and Coordination-free Learning in Structured Matching Markets create mode 100644 data/2022/neurips/Deciding What to Model: Value-Equivalent Sampling for Reinforcement Learning create mode 100644 data/2022/neurips/Decision Trees with Short Explainable Rules create mode 100644 data/2022/neurips/Decision-Focused Learning without Decision-Making: Learning Locally Optimized Decision Losses create mode 100644 data/2022/neurips/Decision-based Black-box Attack Against Vision Transformers via Patch-wise Adversarial Removal create mode 100644 data/2022/neurips/Decomposable Non-Smooth Convex Optimization with Nearly-Linear Gradient Oracle Complexity create mode 100644 data/2022/neurips/Decomposed Knowledge Distillation for Class-Incremental Semantic Segmentation create mode 100644 data/2022/neurips/Decomposing NeRF for Editing via Feature Field Distillation create mode 100644 data/2022/neurips/Deconfounded Representation Similarity for Comparison of Neural Networks create mode 100644 data/2022/neurips/Decoupled Context Processing for Context Augmented Language Modeling create mode 100644 data/2022/neurips/Decoupled Self-supervised Learning for Graphs create mode 100644 data/2022/neurips/Decoupling Classifier for Boosting Few-shot Object Detection and Instance Segmentation create mode 100644 data/2022/neurips/Decoupling Features in Hierarchical Propagation for Video Object Segmentation create mode 100644 data/2022/neurips/Decoupling Knowledge from Memorization: Retrieval-augmented Prompt Learning create mode 100644 data/2022/neurips/Deep Active Learning by Leveraging Training Dynamics create mode 100644 data/2022/neurips/Deep Architecture Connectivity Matters for Its Convergence: A Fine-Grained Analysis create mode 100644 data/2022/neurips/Deep Attentive Belief Propagation: Integrating Reasoning and Learning for Solving Constraint Optimization Problems create mode 100644 data/2022/neurips/Deep Bidirectional Language-Knowledge Graph Pretraining create mode 100644 data/2022/neurips/Deep Combinatorial Aggregation create mode 100644 data/2022/neurips/Deep Compression of Pre-trained Transformer Models create mode 100644 data/2022/neurips/Deep Counterfactual Estimation with Categorical Background Variables create mode 100644 data/2022/neurips/Deep Differentiable Logic Gate Networks create mode 100644 data/2022/neurips/Deep Ensembles Work, But Are They Necessary? create mode 100644 data/2022/neurips/Deep Equilibrium Approaches to Diffusion Models create mode 100644 data/2022/neurips/Deep Fourier Up-Sampling create mode 100644 "data/2022/neurips/Deep Generalized Schr\303\266dinger Bridge" create mode 100644 data/2022/neurips/Deep Generative Model for Periodic Graphs create mode 100644 data/2022/neurips/Deep Hierarchical Planning from Pixels create mode 100644 data/2022/neurips/Deep Learning Methods for Proximal Inference via Maximum Moment Restriction create mode 100644 data/2022/neurips/Deep Model Reassembly create mode 100644 data/2022/neurips/Deep Multi-Modal Structural Equations For Causal Effect Estimation With Unstructured Proxies create mode 100644 data/2022/neurips/Deep Surrogate Assisted Generation of Environments create mode 100644 data/2022/neurips/Deep invariant networks with differentiable augmentation layers create mode 100644 data/2022/neurips/DeepFoids: Adaptive Bio-Inspired Fish Simulation with Deep Reinforcement Learning create mode 100644 data/2022/neurips/DeepInteraction: 3D Object Detection via Modality Interaction create mode 100644 data/2022/neurips/DeepMed: Semiparametric Causal Mediation Analysis with Debiased Deep Learning create mode 100644 data/2022/neurips/DeepTOP: Deep Threshold-Optimal Policy for MDPs and RMABs create mode 100644 data/2022/neurips/Defending Against Adversarial Attacks via Neural Dynamic System create mode 100644 data/2022/neurips/Defining and Characterizing Reward Gaming create mode 100644 data/2022/neurips/Degradation-Aware Unfolding Half-Shuffle Transformer for Spectral Compressive Imaging create mode 100644 data/2022/neurips/Deliberated Domain Bridging for Domain Adaptive Semantic Segmentation create mode 100644 data/2022/neurips/Delving into Out-of-Distribution Detection with Vision-Language Representations create mode 100644 data/2022/neurips/Delving into Sequential Patches for Deepfake Detection create mode 100644 data/2022/neurips/Denoising Diffusion Restoration Models create mode 100644 data/2022/neurips/Dense Interspecies Face Embedding create mode 100644 data/2022/neurips/Density-driven Regularization for Out-of-distribution Detection create mode 100644 data/2022/neurips/Depth is More Powerful than Width with Prediction Concatenation in Deep Forest create mode 100644 data/2022/neurips/Descent Steps of a Relation-Aware Energy Produce Heterogeneous Graph Neural Networks create mode 100644 data/2022/neurips/DetCLIP: Dictionary-Enriched Visual-Concept Paralleled Pre-training for Open-world Detection create mode 100644 data/2022/neurips/Detecting Abrupt Changes in Sequential Pairwise Comparison Data create mode 100644 data/2022/neurips/Detection and Localization of Changes in Conditional Distributions create mode 100644 data/2022/neurips/Deterministic Langevin Monte Carlo with Normalizing Flows for Bayesian Inference create mode 100644 data/2022/neurips/DevFly: Bio-Inspired Development of Binary Connections for Locality Preserving Sparse Codes create mode 100644 data/2022/neurips/DiSC: Differential Spectral Clustering of Features create mode 100644 data/2022/neurips/Diagnosing failures of fairness transfer across distribution shift in real-world medical settings create mode 100644 data/2022/neurips/Diagonal State Spaces are as Effective as Structured State Spaces create mode 100644 data/2022/neurips/Dict-TTS: Learning to Pronounce with Prior Dictionary Knowledge for Text-to-Speech create mode 100644 data/2022/neurips/Differentiable Analog Quantum Computing for Optimization and Control create mode 100644 data/2022/neurips/Differentiable hierarchical and surrogate gradient search for spiking neural networks create mode 100644 data/2022/neurips/Differentially Private Covariance Revisited create mode 100644 data/2022/neurips/Differentially Private Generalized Linear Models Revisited create mode 100644 data/2022/neurips/Differentially Private Graph Learning via Sensitivity-Bounded Personalized PageRank create mode 100644 data/2022/neurips/Differentially Private Learning Needs Hidden State (Or Much Faster Convergence) create mode 100644 data/2022/neurips/Differentially Private Learning with Margin Guarantees create mode 100644 data/2022/neurips/Differentially Private Linear Sketches: Efficient Implementations and Applications create mode 100644 data/2022/neurips/Differentially Private Model Compression create mode 100644 data/2022/neurips/Differentially Private Online-to-batch for Smooth Losses create mode 100644 data/2022/neurips/Diffusion Curvature for Estimating Local Curvature in High Dimensional Data create mode 100644 data/2022/neurips/Diffusion Models as Plug-and-Play Priors create mode 100644 data/2022/neurips/Diffusion Visual Counterfactual Explanations create mode 100644 data/2022/neurips/Diffusion-LM Improves Controllable Text Generation create mode 100644 data/2022/neurips/Diffusion-based Molecule Generation with Informative Prior Bridges create mode 100644 data/2022/neurips/DigGAN: Discriminator gradIent Gap Regularization for GAN Training with Limited Data create mode 100644 data/2022/neurips/Direct Advantage Estimation create mode 100644 data/2022/neurips/Discovered Policy Optimisation create mode 100644 data/2022/neurips/Discovering Design Concepts for CAD Sketches create mode 100644 data/2022/neurips/Discovering and Overcoming Limitations of Noise-engineered Data-free Knowledge Distillation create mode 100644 data/2022/neurips/Discovery of Single Independent Latent Variable create mode 100644 data/2022/neurips/Discrete Compositional Representations as an Abstraction for Goal Conditioned Reinforcement Learning create mode 100644 data/2022/neurips/Discrete-Convex-Analysis-Based Framework for Warm-Starting Algorithms with Predictions create mode 100644 data/2022/neurips/Disentangling Causal Effects from Sets of Interventions in the Presence of Unobserved Confounders create mode 100644 data/2022/neurips/Disentangling Transfer in Continual Reinforcement Learning create mode 100644 data/2022/neurips/Disentangling the Predictive Variance of Deep Ensembles through the Neural Tangent Kernel create mode 100644 data/2022/neurips/Distilled Gradient Aggregation: Purify Features for Input Attribution in the Deep Neural Network create mode 100644 data/2022/neurips/Distilling Representations from GAN Generator via Squeeze and Span create mode 100644 data/2022/neurips/Distinguishing Learning Rules with Brain Machine Interfaces create mode 100644 data/2022/neurips/Distinguishing discrete and continuous behavioral variability using warped autoregressive HMMs create mode 100644 data/2022/neurips/Distributed Distributionally Robust Optimization with Non-Convex Objectives create mode 100644 data/2022/neurips/Distributed Influence-Augmented Local Simulators for Parallel MARL in Large Networked Systems create mode 100644 data/2022/neurips/Distributed Inverse Constrained Reinforcement Learning for Multi-agent Systems create mode 100644 data/2022/neurips/Distributed Learning of Conditional Quantiles in the Reproducing Kernel Hilbert Space create mode 100644 data/2022/neurips/Distributed Methods with Compressed Communication for Solving Variational Inequalities, with Theoretical Guarantees create mode 100644 data/2022/neurips/Distributed Online Convex Optimization with Compressed Communication create mode 100644 data/2022/neurips/Distributed Optimization for Overparameterized Problems: Achieving Optimal Dimension Independent Communication Complexity create mode 100644 data/2022/neurips/Distribution-Informed Neural Networks for Domain Adaptation Regression create mode 100644 data/2022/neurips/Distributional Convergence of the Sliced Wasserstein Process create mode 100644 data/2022/neurips/Distributional Reinforcement Learning for Risk-Sensitive Policies create mode 100644 data/2022/neurips/Distributional Reward Estimation for Effective Multi-agent Deep Reinforcement Learning create mode 100644 data/2022/neurips/Distributionally Adaptive Meta Reinforcement Learning create mode 100644 data/2022/neurips/Distributionally Robust Optimization via Ball Oracle Acceleration create mode 100644 data/2022/neurips/Distributionally Robust Optimization with Data Geometry create mode 100644 data/2022/neurips/Distributionally robust weighted k-nearest neighbors create mode 100644 data/2022/neurips/DivBO: Diversity-aware CASH for Ensemble Learning create mode 100644 data/2022/neurips/Diverse Weight Averaging for Out-of-Distribution Generalization create mode 100644 data/2022/neurips/Diversified Recommendations for Agents with Adaptive Preferences create mode 100644 data/2022/neurips/Diversity vs. Recognizability: Human-like generalization in one-shot generative models create mode 100644 data/2022/neurips/Divert More Attention to Vision-Language Tracking create mode 100644 data/2022/neurips/Divide and Contrast: Source-free Domain Adaptation via Adaptive Contrastive Learning create mode 100644 data/2022/neurips/Do Current Multi-Task Optimization Methods in Deep Learning Even Help? create mode 100644 data/2022/neurips/Do Residual Neural Networks discretize Neural Ordinary Differential Equations? create mode 100644 data/2022/neurips/Does GNN Pretraining Help Molecular Representation? create mode 100644 data/2022/neurips/Does Momentum Change the Implicit Regularization on Separable Data? create mode 100644 data/2022/neurips/Does Self-supervised Learning Really Improve Reinforcement Learning from Pixels? create mode 100644 data/2022/neurips/Domain Adaptation meets Individual Fairness. And they get along create mode 100644 data/2022/neurips/Domain Adaptation under Open Set Label Shift create mode 100644 data/2022/neurips/Domain Generalization by Learning and Removing Domain-specific Features create mode 100644 data/2022/neurips/Domain Generalization without Excess Empirical Risk create mode 100644 data/2022/neurips/Don't Pour Cereal into Coffee: Differentiable Temporal Logic for Temporal Action Segmentation create mode 100644 data/2022/neurips/Don't Roll the Dice, Ask Twice: The Two-Query Distortion of Matching Problems and Beyond create mode 100644 data/2022/neurips/Double Bubble, Toil and Trouble: Enhancing Certified Robustness through Transitivity create mode 100644 data/2022/neurips/Double Check Your State Before Trusting It: Confidence-Aware Bidirectional Offline Model-Based Imagination create mode 100644 data/2022/neurips/Doubly Robust Counterfactual Classification create mode 100644 data/2022/neurips/Doubly-Asynchronous Value Iteration: Making Value Iteration Asynchronous in Actions create mode 100644 data/2022/neurips/Draft-and-Revise: Effective Image Generation with Contextual RQ-Transformer create mode 100644 data/2022/neurips/Drawing out of Distribution with Neuro-Symbolic Generative Models create mode 100644 data/2022/neurips/DreamShard: Generalizable Embedding Table Placement for Recommender Systems create mode 100644 data/2022/neurips/DropCov: A Simple yet Effective Method for Improving Deep Architectures create mode 100644 data/2022/neurips/Dual-Curriculum Contrastive Multi-Instance Learning for Cancer Prognosis Analysis with Whole Slide Images create mode 100644 data/2022/neurips/Dual-discriminative Graph Neural Network for Imbalanced Graph-level Anomaly Detection create mode 100644 data/2022/neurips/DualCoOp: Fast Adaptation to Multi-Label Recognition with Limited Annotations create mode 100644 data/2022/neurips/Dungeons and Data: A Large-Scale NetHack Dataset create mode 100644 data/2022/neurips/Dynamic Fair Division with Partial Information create mode 100644 data/2022/neurips/Dynamic Graph Neural Networks Under Spatio-Temporal Distribution Shift create mode 100644 data/2022/neurips/Dynamic Inverse Reinforcement Learning for Characterizing Animal Behavior create mode 100644 data/2022/neurips/Dynamic Learning in Large Matching Markets create mode 100644 data/2022/neurips/Dynamic Pricing with Monotonicity Constraint under Unknown Parametric Demand Model create mode 100644 "data/2022/neurips/Dynamic Sparse Network for Time Series Classification: Learning What to \"See\"" create mode 100644 data/2022/neurips/Dynamic Tensor Product Regression create mode 100644 data/2022/neurips/Dynamic pricing and assortment under a contextual MNL demand create mode 100644 data/2022/neurips/Dynamics of SGD with Stochastic Polyak Stepsizes: Truly Adaptive Variants and Convergence to Exact Solution create mode 100644 data/2022/neurips/E-MAPP: Efficient Multi-Agent Reinforcement Learning with Parallel Program Guidance create mode 100644 data/2022/neurips/EAGER: Asking and Answering Questions for Automatic Reward Shaping in Language-guided RL create mode 100644 data/2022/neurips/EF-BV: A Unified Theory of Error Feedback and Variance Reduction Mechanisms for Biased and Unbiased Compression in Distributed Optimization create mode 100644 data/2022/neurips/EGSDE: Unpaired Image-to-Image Translation via Energy-Guided Stochastic Differential Equations create mode 100644 data/2022/neurips/EHRSQL: A Practical Text-to-SQL Benchmark for Electronic Health Records create mode 100644 data/2022/neurips/ELASTIC: Numerical Reasoning with Adaptive Symbolic Compiler create mode 100644 data/2022/neurips/ELEVATER: A Benchmark and Toolkit for Evaluating Language-Augmented Visual Models create mode 100644 data/2022/neurips/ELIAS: End-to-End Learning to Index and Search in Large Output Spaces create mode 100644 data/2022/neurips/ELIGN: Expectation Alignment as a Multi-Agent Intrinsic Reward create mode 100644 data/2022/neurips/ENS-10: A Dataset For Post-Processing Ensemble Weather Forecasts create mode 100644 data/2022/neurips/EPIC-KITCHENS VISOR Benchmark: VIdeo Segmentations and Object Relations create mode 100644 data/2022/neurips/ESCADA: Efficient Safety and Context Aware Dose Allocation for Precision Medicine create mode 100644 data/2022/neurips/ETAB: A Benchmark Suite for Visual Representation Learning in Echocardiography create mode 100644 data/2022/neurips/EZNAS: Evolving Zero-Cost Proxies For Neural Architecture Scoring create mode 100644 data/2022/neurips/Early Stage Convergence and Global Convergence of Training Mildly Parameterized Neural Networks create mode 100644 data/2022/neurips/Earthformer: Exploring Space-Time Transformers for Earth System Forecasting create mode 100644 data/2022/neurips/EcoFormer: Energy-Saving Attention with Linear Complexity create mode 100644 data/2022/neurips/Effective Adaptation in Multi-Task Co-Training for Unified Autonomous Driving create mode 100644 data/2022/neurips/Effective Backdoor Defense by Exploiting Sensitivity of Poisoned Samples create mode 100644 data/2022/neurips/Effective Dimension in Bandit Problems under Censorship create mode 100644 data/2022/neurips/Effectiveness of Vision Transformer for Fast and Accurate Single-Stage Pedestrian Detection create mode 100644 data/2022/neurips/Effects of Data Geometry in Early Deep Learning create mode 100644 data/2022/neurips/Efficiency Ordering of Stochastic Gradient Descent create mode 100644 data/2022/neurips/Efficient Active Learning with Abstention create mode 100644 data/2022/neurips/Efficient Adversarial Training without Attacking: Worst-Case-Aware Robust Reinforcement Learning create mode 100644 data/2022/neurips/Efficient Aggregated Kernel Tests using Incomplete $U$-statistics create mode 100644 data/2022/neurips/Efficient Architecture Search for Diverse Tasks create mode 100644 data/2022/neurips/Efficient Dataset Distillation using Random Feature Approximation create mode 100644 data/2022/neurips/Efficient Frameworks for Generalized Low-Rank Matrix Bandit Problems create mode 100644 data/2022/neurips/Efficient Graph Similarity Computation with Alignment Regularization create mode 100644 data/2022/neurips/Efficient Knowledge Distillation from Model Checkpoints create mode 100644 data/2022/neurips/Efficient Meta Reinforcement Learning for Preference-based Fast Adaptation create mode 100644 data/2022/neurips/Efficient Methods for Non-stationary Online Learning create mode 100644 data/2022/neurips/Efficient Multi-agent Communication via Self-supervised Information Aggregation create mode 100644 data/2022/neurips/Efficient Non-Parametric Optimizer Search for Diverse Tasks create mode 100644 data/2022/neurips/Efficient Phi-Regret Minimization in Extensive-Form Games via Online Mirror Descent create mode 100644 data/2022/neurips/Efficient Risk-Averse Reinforcement Learning create mode 100644 data/2022/neurips/Efficient Sampling on Riemannian Manifolds via Langevin MCMC create mode 100644 data/2022/neurips/Efficient Scheduling of Data Augmentation for Deep Reinforcement Learning create mode 100644 data/2022/neurips/Efficient Spatially Sparse Inference for Conditional GANs and Diffusion Models create mode 100644 data/2022/neurips/Efficient Submodular Optimization under Noise: Local Search is Robust create mode 100644 data/2022/neurips/Efficient Training of Low-Curvature Neural Networks create mode 100644 data/2022/neurips/Efficient and Effective Augmentation Strategy for Adversarial Training create mode 100644 data/2022/neurips/Efficient and Effective Multi-task Grouping via Meta Learning on Task Combinations create mode 100644 data/2022/neurips/Efficient and Effective Optimal Transport-Based Biclustering create mode 100644 data/2022/neurips/Efficient and Modular Implicit Differentiation create mode 100644 data/2022/neurips/Efficient and Near-Optimal Smoothed Online Learning for Generalized Linear Functions create mode 100644 data/2022/neurips/Efficient and Stable Fully Dynamic Facility Location create mode 100644 data/2022/neurips/Efficient coding, channel capacity, and the emergence of retinal mosaics create mode 100644 data/2022/neurips/Efficient identification of informative features in simulation-based inference create mode 100644 data/2022/neurips/Efficient learning of nonlinear prediction models with time-series privileged information create mode 100644 data/2022/neurips/EfficientFormer: Vision Transformers at MobileNet Speed create mode 100644 data/2022/neurips/Efficiently Computing Local Lipschitz Constants of Neural Networks via Bound Propagation create mode 100644 data/2022/neurips/Efficiently Factorizing Boolean Matrices using Proximal Gradient Descent create mode 100644 data/2022/neurips/EgoTaskQA: Understanding Human Tasks in Egocentric Videos create mode 100644 data/2022/neurips/Egocentric Video-Language Pretraining create mode 100644 data/2022/neurips/ElasticMVS: Learning elastic part representation for self-supervised multi-view stereopsis create mode 100644 data/2022/neurips/Eliciting Thinking Hierarchy without a Prior create mode 100644 data/2022/neurips/Elucidating the Design Space of Diffusion-Based Generative Models create mode 100644 data/2022/neurips/Embed and Emulate: Learning to estimate parameters of dynamical systems with uncertainty quantification create mode 100644 data/2022/neurips/Embodied Scene-aware Human Pose Estimation create mode 100644 data/2022/neurips/Embrace the Gap: VAEs Perform Independent Mechanism Analysis create mode 100644 data/2022/neurips/Embracing Consistency: A One-Stage Approach for Spatio-Temporal Video Grounding create mode 100644 data/2022/neurips/Emergence of Hierarchical Layers in a Single Sheet of Self-Organizing Spiking Neurons create mode 100644 data/2022/neurips/Emergent Communication: Generalization and Overfitting in Lewis Games create mode 100644 data/2022/neurips/Emergent Graphical Conventions in a Visual Communication Game create mode 100644 data/2022/neurips/Empirical Gateaux Derivatives for Causal Inference create mode 100644 data/2022/neurips/Empirical Phase Diagram for Three-layer Neural Networks with Infinite Width create mode 100644 data/2022/neurips/Enabling Detailed Action Recognition Evaluation Through Video Dataset Augmentation create mode 100644 data/2022/neurips/End-to-end Algorithm Synthesis with Recurrent Networks: Extrapolation without Overthinking create mode 100644 data/2022/neurips/End-to-end Stochastic Optimization with Energy-based Model create mode 100644 data/2022/neurips/End-to-end Symbolic Regression with Transformers create mode 100644 data/2022/neurips/Energy-Based Contrastive Learning of Visual Representations create mode 100644 data/2022/neurips/Enhance the Visual Representation via Discrete Adversarial Training create mode 100644 data/2022/neurips/Enhanced Bilevel Optimization via Bregman Distance create mode 100644 data/2022/neurips/Enhanced Latent Space Blind Model for Real Image Denoising via Alternative Optimization create mode 100644 data/2022/neurips/Enhanced Meta Reinforcement Learning via Demonstrations in Sparse Reward Environments create mode 100644 data/2022/neurips/Enhancing Safe Exploration Using Safety State Augmentation create mode 100644 data/2022/neurips/Ensemble of Averages: Improving Model Selection and Boosting Performance in Domain Generalization create mode 100644 data/2022/neurips/Entropy-Driven Mixed-Precision Quantization for Deep Network Design create mode 100644 data/2022/neurips/EnvPool: A Highly Parallel Reinforcement Learning Environment Execution Engine create mode 100644 data/2022/neurips/Environment Diversification with Multi-head Neural Network for Invariant Learning create mode 100644 data/2022/neurips/Envy-free Policy Teaching to Multiple Agents create mode 100644 data/2022/neurips/EpiGRAF: Rethinking training of 3D GANs create mode 100644 data/2022/neurips/Equivariant Graph Hierarchy-Based Neural Networks create mode 100644 data/2022/neurips/Equivariant Networks for Crystal Structures create mode 100644 data/2022/neurips/Equivariant Networks for Zero-Shot Coordination create mode 100644 data/2022/neurips/Error Analysis of Tensor-Train Cross Approximation create mode 100644 data/2022/neurips/Error Correction Code Transformer create mode 100644 data/2022/neurips/Escaping Saddle Points for Effective Generalization on Class-Imbalanced Data create mode 100644 data/2022/neurips/Escaping Saddle Points with Bias-Variance Reduced Local Perturbed SGD for Communication Efficient Nonconvex Distributed Learning create mode 100644 data/2022/neurips/Escaping from the Barren Plateau via Gaussian Initializations in Deep Variational Quantum Circuits create mode 100644 data/2022/neurips/Estimating Noise Transition Matrix with Label Correlations for Noisy Multi-Label Learning create mode 100644 data/2022/neurips/Estimating and Explaining Model Performance When Both Covariates and Labels Shift create mode 100644 data/2022/neurips/Estimating graphical models for count data with applications to single-cell gene network create mode 100644 data/2022/neurips/Estimating the Arc Length of the Optimal ROC Curve and Lower Bounding the Maximal AUC create mode 100644 data/2022/neurips/Estimation of Entropy in Constant Space with Improved Sample Complexity create mode 100644 data/2022/neurips/Evaluated CMI Bounds for Meta Learning: Tightness and Expressiveness create mode 100644 data/2022/neurips/Evaluating Graph Generative Models with Contrastively Learned Features create mode 100644 data/2022/neurips/Evaluating Latent Space Robustness and Uncertainty of EEG-ML Models under Realistic Distribution Shifts create mode 100644 data/2022/neurips/Evaluating Out-of-Distribution Performance on Document Image Classifiers create mode 100644 data/2022/neurips/Evaluating Robustness to Dataset Shift via Parametric Robustness Sets create mode 100644 data/2022/neurips/Evaluation beyond Task Performance: Analyzing Concepts in AlphaZero in Hex create mode 100644 data/2022/neurips/EvenNet: Ignoring Odd-Hop Neighbors Improves Robustness of Graph Neural Networks create mode 100644 data/2022/neurips/Evolution of Neural Tangent Kernels under Benign and Adversarial Training create mode 100644 data/2022/neurips/Exact Shape Correspondence via 2D graph convolution create mode 100644 data/2022/neurips/Exact Solutions of a Deep Linear Network create mode 100644 data/2022/neurips/Exact learning dynamics of deep linear networks with prior knowledge create mode 100644 data/2022/neurips/Expansion and Shrinkage of Localization for Weakly-Supervised Semantic Segmentation create mode 100644 data/2022/neurips/Expectation-Maximization Contrastive Learning for Compact Video-and-Language Representations create mode 100644 data/2022/neurips/Expected Improvement for Contextual Bandits create mode 100644 data/2022/neurips/Expediting Large-Scale Vision Transformer for Dense Prediction without Fine-tuning create mode 100644 data/2022/neurips/Experimental Design for Linear Functionals in Reproducing Kernel Hilbert Spaces create mode 100644 data/2022/neurips/Explain My Surprise: Learning Efficient Long-Term Memory by predicting uncertain outcomes create mode 100644 data/2022/neurips/Explainability Via Causal Self-Talk create mode 100644 data/2022/neurips/Explainable Reinforcement Learning via Model Transforms create mode 100644 data/2022/neurips/Explaining Preferences with Shapley Values create mode 100644 data/2022/neurips/Explicable Policy Search create mode 100644 data/2022/neurips/Explicit Tradeoffs between Adversarial and Natural Distributional Robustness create mode 100644 data/2022/neurips/Exploit Reward Shifting in Value-Based Deep-RL: Optimistic Curiosity-Based Exploration and Conservative Exploitation via Linear Reward Shaping create mode 100644 data/2022/neurips/Exploitability Minimization in Games and Beyond create mode 100644 data/2022/neurips/Exploiting Semantic Relations for Glass Surface Detection create mode 100644 data/2022/neurips/Exploiting the Relationship Between Kendall's Rank Correlation and Cosine Similarity for Attribution Protection create mode 100644 data/2022/neurips/Exploration via Elliptical Episodic Bonuses create mode 100644 data/2022/neurips/Exploration via Planning for Information about the Optimal Trajectory create mode 100644 data/2022/neurips/Exploration-Guided Reward Shaping for Reinforcement Learning under Sparse Rewards create mode 100644 data/2022/neurips/Exploring Example Influence in Continual Learning create mode 100644 data/2022/neurips/Exploring Figure-Ground Assignment Mechanism in Perceptual Organization create mode 100644 data/2022/neurips/Exploring Length Generalization in Large Language Models create mode 100644 data/2022/neurips/Exploring evolution-aware & -free protein language models as protein function predictors create mode 100644 data/2022/neurips/Exploring the Algorithm-Dependent Generalization of AUPRC Optimization with List Stability create mode 100644 data/2022/neurips/Exploring the Latent Space of Autoencoders with Interventional Assays create mode 100644 data/2022/neurips/Exploring the Limits of Domain-Adaptive Training for Detoxifying Large-Scale Language Models create mode 100644 data/2022/neurips/Exploring the Whole Rashomon Set of Sparse Decision Trees create mode 100644 data/2022/neurips/Exploring through Random Curiosity with General Value Functions create mode 100644 data/2022/neurips/Exponential Family Model-Based Reinforcement Learning via Score Matching create mode 100644 data/2022/neurips/Exponential Separations in Symmetric Neural Networks create mode 100644 data/2022/neurips/Exponentially Improving the Complexity of Simulating the Weisfeiler-Lehman Test with Graph Neural Networks create mode 100644 data/2022/neurips/Exposing and Exploiting Fine-Grained Block Structures for Fast and Accurate Sparse Training create mode 100644 data/2022/neurips/Extra-Newton: A First Approach to Noise-Adaptive Accelerated Second-Order Methods create mode 100644 data/2022/neurips/Extracting computational mechanisms from neural data using low-rank RNNs create mode 100644 data/2022/neurips/Extrapolation and Spectral Bias of Neural Nets with Hadamard Product: a Polynomial Net Study create mode 100644 data/2022/neurips/Extrapolative Continuous-time Bayesian Neural Network for Fast Training-free Test-time Adaptation create mode 100644 data/2022/neurips/FACT: Learning Governing Abstractions Behind Integer Sequences create mode 100644 data/2022/neurips/FETA: Towards Specializing Foundational Models for Expert Task Applications create mode 100644 data/2022/neurips/FIRE: Semantic Field of Words Represented as Non-Linear Functions create mode 100644 data/2022/neurips/FLAIR: Federated Learning Annotated Image Repository create mode 100644 data/2022/neurips/FLamby: Datasets and Benchmarks for Cross-Silo Federated Learning in Realistic Healthcare Settings create mode 100644 data/2022/neurips/FNeVR: Neural Volume Rendering for Face Animation create mode 100644 data/2022/neurips/FOF: Learning Fourier Occupancy Field for Monocular Real-time Human Reconstruction create mode 100644 data/2022/neurips/FP8 Quantization: The Power of the Exponent create mode 100644 data/2022/neurips/FR: Folded Rationalization with a Unified Encoder create mode 100644 data/2022/neurips/Factored Adaptation for Non-Stationary Reinforcement Learning create mode 100644 data/2022/neurips/Factored DRO: Factored Distributionally Robust Policies for Contextual Bandits create mode 100644 data/2022/neurips/Factorized-FL: Personalized Federated Learning with Parameter Factorization & Similarity Matching create mode 100644 data/2022/neurips/Factuality Enhanced Language Models for Open-Ended Text Generation create mode 100644 data/2022/neurips/Fair Bayes-Optimal Classifiers Under Predictive Parity create mode 100644 data/2022/neurips/Fair Infinitesimal Jackknife: Mitigating the Influence of Biased Training Data Points Without Refitting create mode 100644 data/2022/neurips/Fair Rank Aggregation create mode 100644 data/2022/neurips/Fair Ranking with Noisy Protected Attributes create mode 100644 data/2022/neurips/Fair Wrapping for Black-box Predictions create mode 100644 data/2022/neurips/Fair and Efficient Allocations Without Obvious Manipulations create mode 100644 data/2022/neurips/Fair and Optimal Decision Trees: A Dynamic Programming Approach create mode 100644 data/2022/neurips/FairVFL: A Fair Vertical Federated Learning Framework with Contrastive Adversarial Learning create mode 100644 data/2022/neurips/Fairness Reprogramming create mode 100644 data/2022/neurips/Fairness Transferability Subject to Bounded Distribution Shift create mode 100644 data/2022/neurips/Fairness in Federated Learning via Core-Stability create mode 100644 data/2022/neurips/Fairness without Demographics through Knowledge Distillation create mode 100644 data/2022/neurips/Falconn++: A Locality-sensitive Filtering Approach for Approximate Nearest Neighbor Search create mode 100644 data/2022/neurips/Falsification before Extrapolation in Causal Effect Estimation create mode 100644 data/2022/neurips/Fast Algorithms for Packing Proportional Fairness and its Dual create mode 100644 data/2022/neurips/Fast Bayesian Coresets via Subsampling and Quasi-Newton Refinement create mode 100644 data/2022/neurips/Fast Bayesian Estimation of Point Process Intensity as Function of Covariates create mode 100644 data/2022/neurips/Fast Bayesian Inference with Batch Bayesian Quadrature via Kernel Recombination create mode 100644 data/2022/neurips/Fast Distance Oracles for Any Symmetric Norm create mode 100644 data/2022/neurips/Fast Instrument Learning with Faster Rates create mode 100644 data/2022/neurips/Fast Mixing of Stochastic Gradient Descent with Normalization and Weight Decay create mode 100644 data/2022/neurips/Fast Neural Kernel Embeddings for General Activations create mode 100644 data/2022/neurips/Fast Stochastic Composite Minimization and an Accelerated Frank-Wolfe Algorithm under Parallelization create mode 100644 data/2022/neurips/Fast Vision Transformers with HiLo Attention create mode 100644 data/2022/neurips/Faster Deep Reinforcement Learning with Slower Online Network create mode 100644 data/2022/neurips/Faster Linear Algebra for Distance Matrices create mode 100644 data/2022/neurips/Faster and Scalable Algorithms for Densest Subgraph and Decomposition create mode 100644 data/2022/neurips/FasterRisk: Fast and Accurate Interpretable Risk Scores create mode 100644 data/2022/neurips/Fault-Aware Neural Code Rankers create mode 100644 data/2022/neurips/FeLMi : Few shot Learning with hard Mixup create mode 100644 data/2022/neurips/Feature Learning in $L_2$-regularized DNNs: Attraction Repulsion and Sparsity create mode 100644 data/2022/neurips/Feature-Proxy Transformer for Few-Shot Segmentation create mode 100644 data/2022/neurips/FedAvg with Fine Tuning: Local Updates Lead to Representation Learning create mode 100644 data/2022/neurips/FedPop: A Bayesian Approach for Personalised Federated Learning create mode 100644 data/2022/neurips/FedRolex: Model-Heterogeneous Federated Learning with Rolling Sub-Model Extraction create mode 100644 data/2022/neurips/FedSR: A Simple and Effective Domain Generalization Method for Federated Learning create mode 100644 data/2022/neurips/Federated Learning from Pre-Trained Models: A Contrastive Learning Approach create mode 100644 data/2022/neurips/Few-Shot Audio-Visual Learning of Environment Acoustics create mode 100644 data/2022/neurips/Few-Shot Continual Active Learning by a Robot create mode 100644 data/2022/neurips/Few-Shot Fast-Adaptive Anomaly Detection create mode 100644 data/2022/neurips/Few-Shot Non-Parametric Learning with Deep Latent Variable Model create mode 100644 data/2022/neurips/Few-Shot Parameter-Efficient Fine-Tuning is Better and Cheaper than In-Context Learning create mode 100644 data/2022/neurips/Few-shot Image Generation via Adaptation-Aware Kernel Modulation create mode 100644 data/2022/neurips/Few-shot Learning for Feature Selection with Hilbert-Schmidt Independence Criterion create mode 100644 data/2022/neurips/Few-shot Relational Reasoning via Connection Subgraph Pretraining create mode 100644 data/2022/neurips/Few-shot Task-agnostic Neural Architecture Search for Distilling Large Language Models create mode 100644 data/2022/neurips/FiLM-Ensemble: Probabilistic Deep Learning via Feature-wise Linear Modulation create mode 100644 data/2022/neurips/FiLM: Frequency improved Legendre Memory Model for Long-term Time Series Forecasting create mode 100644 data/2022/neurips/FinRL-Meta: Market Environments and Benchmarks for Data-Driven Financial Reinforcement Learning create mode 100644 data/2022/neurips/Finding Correlated Equilibrium of Constrained Markov Game: A Primal-Dual Approach create mode 100644 data/2022/neurips/Finding Differences Between Transformers and ConvNets Using Counterfactual Simulation Testing create mode 100644 data/2022/neurips/Finding Naturally Occurring Physical Backdoors in Image Datasets create mode 100644 data/2022/neurips/Finding Optimal Arms in Non-stochastic Combinatorial Bandits with Semi-bandit Feedback and Finite Budget create mode 100644 data/2022/neurips/Finding Second-Order Stationary Points in Nonconvex-Strongly-Concave Minimax Optimization create mode 100644 data/2022/neurips/Finding and Listing Front-door Adjustment Sets create mode 100644 data/2022/neurips/Fine-Grained Analysis of Stability and Generalization for Modern Meta Learning Algorithms create mode 100644 data/2022/neurips/Fine-Grained Semantically Aligned Vision-Language Pre-Training create mode 100644 data/2022/neurips/Fine-Tuning Pre-Trained Language Models Effectively by Optimizing Subnetworks Adaptively create mode 100644 data/2022/neurips/Fine-tuning Language Models over Slow Networks using Activation Quantization with Guarantees create mode 100644 data/2022/neurips/Fine-tuning language models to find agreement among humans with diverse preferences create mode 100644 data/2022/neurips/Finite Sample Analysis Of Dynamic Regression Parameter Learning create mode 100644 data/2022/neurips/Finite-Sample Maximum Likelihood Estimation of Location create mode 100644 data/2022/neurips/Finite-Time Analysis of Adaptive Temporal Difference Learning with Deep Neural Networks create mode 100644 data/2022/neurips/Finite-Time Last-Iterate Convergence for Learning in Multi-Player Games create mode 100644 data/2022/neurips/Finite-Time Regret of Thompson Sampling Algorithms for Exponential Family Multi-Armed Bandits create mode 100644 data/2022/neurips/First Contact: Unsupervised Human-Machine Co-Adaptation via Mutual Information Maximization create mode 100644 data/2022/neurips/First Hitting Diffusion Models for Generating Manifold, Graph and Categorical Data create mode 100644 data/2022/neurips/First is Better Than Last for Language Data Influence create mode 100644 data/2022/neurips/First-Order Algorithms for Min-Max Optimization in Geodesic Metric Spaces create mode 100644 data/2022/neurips/Fixed-Distance Hamiltonian Monte Carlo create mode 100644 data/2022/neurips/Flamingo: a Visual Language Model for Few-Shot Learning create mode 100644 data/2022/neurips/Flare7K: A Phenomenological Nighttime Flare Removal Dataset create mode 100644 data/2022/neurips/FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness create mode 100644 data/2022/neurips/Flexible Diffusion Modeling of Long Videos create mode 100644 data/2022/neurips/Flexible Neural Image Compression via Code Editing create mode 100644 data/2022/neurips/FlowHMM: Flow-based continuous hidden Markov models create mode 100644 data/2022/neurips/Flowification: Everything is a normalizing flow create mode 100644 data/2022/neurips/FlyView: a bio-informed optical flow truth dataset for visual navigation using panoramic stereo vision create mode 100644 data/2022/neurips/Focal Modulation Networks create mode 100644 data/2022/neurips/Follow-the-Perturbed-Leader for Adversarial Markov Decision Processes with Bandit Feedback create mode 100644 data/2022/neurips/Forecasting Future World Events With Neural Networks create mode 100644 data/2022/neurips/Forecasting Human Trajectory from Scene History create mode 100644 data/2022/neurips/Formalizing Consistency and Coherence of Representation Learning create mode 100644 data/2022/neurips/Formulating Robustness Against Unforeseen Attacks create mode 100644 data/2022/neurips/Forward-Backward Latent State Inference for Hidden Continuous-Time semi-Markov Chains create mode 100644 data/2022/neurips/Foundation Posteriors for Approximate Probabilistic Inference create mode 100644 data/2022/neurips/FourierFormer: Transformer Meets Generalized Fourier Integral Theorem create mode 100644 data/2022/neurips/FourierNets enable the design of highly non-local optical encoders for computational imaging create mode 100644 data/2022/neurips/Frank-Wolfe-based Algorithms for Approximating Tyler's M-estimator create mode 100644 data/2022/neurips/FreGAN: Exploiting Frequency Components for Training GANs under Limited Data create mode 100644 data/2022/neurips/Free Probability for predicting the performance of feed-forward fully connected neural networks create mode 100644 data/2022/neurips/Friendly Noise against Adversarial Noise: A Powerful Defense against Data Poisoning Attack create mode 100644 data/2022/neurips/From Gradient Flow on Population Loss to Learning with Stochastic Gradient Descent create mode 100644 data/2022/neurips/Fully Convolutional One-Stage 3D Object Detection on LiDAR Range Images create mode 100644 data/2022/neurips/Fully Sparse 3D Object Detection create mode 100644 data/2022/neurips/Function Classes for Identifiable Nonlinear Independent Component Analysis create mode 100644 data/2022/neurips/Functional Ensemble Distillation create mode 100644 data/2022/neurips/Functional Indirection Neural Estimator for Better Out-of-distribution Generalization create mode 100644 data/2022/neurips/Fused Orthogonal Alternating Least Squares for Tensor Clustering create mode 100644 data/2022/neurips/Fuzzy Learning Machine create mode 100644 data/2022/neurips/GAGA: Deciphering Age-path of Generalized Self-paced Regularizer create mode 100644 data/2022/neurips/GAL: Gradient Assisted Learning for Decentralized Multi-Organization Collaborations create mode 100644 data/2022/neurips/GALOIS: Boosting Deep Reinforcement Learning via Generalizable Logic Synthesis create mode 100644 data/2022/neurips/GAMA: Generative Adversarial Multi-Object Scene Attacks create mode 100644 data/2022/neurips/GAPX: Generalized Autoregressive Paraphrase-Identification X create mode 100644 data/2022/neurips/GAR: Generalized Autoregression for Multi-Fidelity Fusion create mode 100644 data/2022/neurips/GAUDI: A Neural Architect for Immersive 3D Scene Generation create mode 100644 data/2022/neurips/GBA: A Tuning-free Approach to Switch between Synchronous and Asynchronous Training for Recommendation Models create mode 100644 data/2022/neurips/GENIE: Higher-Order Denoising Diffusion Solvers create mode 100644 data/2022/neurips/GET3D: A Generative Model of High Quality 3D Textured Shapes Learned from Images create mode 100644 data/2022/neurips/GLIF: A Unified Gated Leaky Integrate-and-Fire Neuron for Spiking Neural Networks create mode 100644 data/2022/neurips/GLIPv2: Unifying Localization and Vision-Language Understanding create mode 100644 data/2022/neurips/GLOBEM Dataset: Multi-Year Datasets for Longitudinal Human Behavior Modeling Generalization create mode 100644 data/2022/neurips/GMMSeg: Gaussian Mixture based Generative Semantic Segmentation Models create mode 100644 data/2022/neurips/GOOD: A Graph Out-of-Distribution Benchmark create mode 100644 data/2022/neurips/GPT3.int8(): 8-bit Matrix Multiplication for Transformers at Scale create mode 100644 data/2022/neurips/GRASP: Navigating Retrosynthetic Planning with Goal-driven Policy create mode 100644 data/2022/neurips/GREED: A Neural Framework for Learning Graph Distance Functions create mode 100644 data/2022/neurips/GStarX: Explaining Graph Neural Networks with Structure-Aware Cooperative Games create mode 100644 data/2022/neurips/GT-GAN: General Purpose Time Series Synthesis with Generative Adversarial Networks create mode 100644 data/2022/neurips/GULP: a prediction-based metric between representations create mode 100644 data/2022/neurips/Gaussian Copula Embeddings create mode 100644 data/2022/neurips/GenSDF: Two-Stage Learning of Generalizable Signed Distance Functions create mode 100644 data/2022/neurips/GenerSpeech: Towards Style Transfer for Generalizable Out-Of-Domain Text-to-Speech create mode 100644 data/2022/neurips/General Cutting Planes for Bound-Propagation-Based Neural Network Verification create mode 100644 data/2022/neurips/Generalised Implicit Neural Representations create mode 100644 data/2022/neurips/Generalised Mutual Information for Discriminative Clustering create mode 100644 data/2022/neurips/Generalization Analysis of Message Passing Neural Networks on Large Random Graphs create mode 100644 data/2022/neurips/Generalization Analysis on Learning with a Concurrent Verifier create mode 100644 data/2022/neurips/Generalization Bounds for Estimating Causal Effects of Continuous Treatments create mode 100644 data/2022/neurips/Generalization Bounds for Gradient Methods via Discrete and Continuous Prior create mode 100644 data/2022/neurips/Generalization Bounds with Minimal Dependency on Hypothesis Class via Distributionally Robust Optimization create mode 100644 data/2022/neurips/Generalization Error Bounds on Deep Learning with Markov Datasets create mode 100644 data/2022/neurips/Generalization Gap in Amortized Inference create mode 100644 data/2022/neurips/Generalization Properties of NAS under Activation and Skip Connection Search create mode 100644 data/2022/neurips/Generalization for multiclass classification with overparameterized linear models create mode 100644 data/2022/neurips/Generalized Delayed Feedback Model with Post-Click Information in Recommender Systems create mode 100644 data/2022/neurips/Generalized Laplacian Eigenmaps create mode 100644 data/2022/neurips/Generalized One-shot Domain Adaptation of Generative Adversarial Networks create mode 100644 data/2022/neurips/Generalized Variational Inference in Function Spaces: Gaussian Measures meet Bayesian Deep Learning create mode 100644 data/2022/neurips/Generalizing Bayesian Optimization with Decision-theoretic Entropies create mode 100644 data/2022/neurips/Generalizing Consistent Multi-Class Classification with Rejection to be Compatible with Arbitrary Losses create mode 100644 data/2022/neurips/Generalizing Goal-Conditioned Reinforcement Learning with Variational Causal Reasoning create mode 100644 data/2022/neurips/Generating Long Videos of Dynamic Scenes create mode 100644 data/2022/neurips/Generating Training Data with Language Models: Towards Zero-Shot Language Understanding create mode 100644 data/2022/neurips/Generating multivariate time series with COmmon Source CoordInated GAN (COSCI-GAN) create mode 100644 data/2022/neurips/Generative Neural Articulated Radiance Fields create mode 100644 data/2022/neurips/Generative Status Estimation and Information Decoupling for Image Rain Removal create mode 100644 data/2022/neurips/Generative Time Series Forecasting with Diffusion, Denoise, and Disentanglement create mode 100644 data/2022/neurips/Generative Visual Prompt: Unifying Distributional Control of Pre-Trained Generative Models create mode 100644 data/2022/neurips/Generative multitask learning mitigates target-causing confounding create mode 100644 data/2022/neurips/Generic bounds on the approximation error for physics-informed (and) operator learning create mode 100644 data/2022/neurips/Geo-Neus: Geometry-Consistent Neural Implicit Surfaces Learning for Multi-view Reconstruction create mode 100644 data/2022/neurips/Geo-SIC: Learning Deformable Geometric Shapes in Deep Image Classifiers create mode 100644 data/2022/neurips/Geoclidean: Few-Shot Generalization in Euclidean Geometry create mode 100644 data/2022/neurips/Geodesic Graph Neural Network for Efficient Graph Representation Learning create mode 100644 data/2022/neurips/Geodesic Self-Attention for 3D Point Clouds create mode 100644 data/2022/neurips/Geometric Knowledge Distillation: Topology Compression for Graph Neural Networks create mode 100644 data/2022/neurips/Geometric Order Learning for Rank Estimation create mode 100644 data/2022/neurips/Geometry-aware Two-scale PIFu Representation for Human Reconstruction create mode 100644 data/2022/neurips/Get More at Once: Alternating Sparse Training with Gradient Correction create mode 100644 data/2022/neurips/GhostNetV2: Enhance Cheap Operation with Long-Range Attention create mode 100644 data/2022/neurips/Giga-scale Kernel Matrix-Vector Multiplication on GPU create mode 100644 data/2022/neurips/Giving Feedback on Interactive Student Programs with Meta-Exploration create mode 100644 data/2022/neurips/GlanceNets: Interpretable, Leak-proof Concept-based Models create mode 100644 data/2022/neurips/Global Convergence and Stability of Stochastic Gradient Descent create mode 100644 data/2022/neurips/Global Convergence of Federated Learning for Mixed Regression create mode 100644 data/2022/neurips/Global Linear and Local Superlinear Convergence of IRLS for Non-Smooth Robust Regression create mode 100644 data/2022/neurips/Global Normalization for Streaming Speech Recognition in a Modular Framework create mode 100644 data/2022/neurips/Global Optimal K-Medoids Clustering of One Million Samples create mode 100644 data/2022/neurips/Globally Convergent Policy Search for Output Estimation create mode 100644 data/2022/neurips/Globally Gated Deep Linear Networks create mode 100644 "data/2022/neurips/Gold-standard solutions to the Schr\303\266dinger equation using deep learning: How much physics do we need?" create mode 100644 data/2022/neurips/GraB: Finding Provably Better Data Permutations than Random Reshuffling create mode 100644 data/2022/neurips/Gradient Descent Is Optimal Under Lower Restricted Secant Inequality And Upper Error Bound create mode 100644 data/2022/neurips/Gradient Descent: The Ultimate Optimizer create mode 100644 data/2022/neurips/Gradient Estimation with Discrete Stein Operators create mode 100644 data/2022/neurips/Gradient Methods Provably Converge to Non-Robust Networks create mode 100644 data/2022/neurips/Gradient flow dynamics of shallow ReLU networks for square loss and orthogonal inputs create mode 100644 data/2022/neurips/Gradient-Free Methods for Deterministic and Stochastic Nonsmooth Nonconvex Optimization create mode 100644 data/2022/neurips/Graph Coloring via Neural Networks for Haplotype Assembly and Viral Quasispecies Reconstruction create mode 100644 data/2022/neurips/Graph Convolution Network based Recommender Systems: Learning Guarantee and Item Mixture Powered Strategy create mode 100644 data/2022/neurips/Graph Few-shot Learning with Task-specific Structures create mode 100644 data/2022/neurips/Graph Learning Assisted Multi-Objective Integer Programming create mode 100644 data/2022/neurips/Graph Neural Network Bandits create mode 100644 data/2022/neurips/Graph Neural Networks are Dynamic Programmers create mode 100644 data/2022/neurips/Graph Neural Networks with Adaptive Readouts create mode 100644 data/2022/neurips/Graph Reordering for Cache-Efficient Near Neighbor Search create mode 100644 data/2022/neurips/Graph Scattering beyond Wavelet Shackles create mode 100644 data/2022/neurips/Graph Self-supervised Learning with Accurate Discrepancy Learning create mode 100644 data/2022/neurips/GraphDE: A Generative Framework for Debiased Learning and Out-of-Distribution Detection on Graphs create mode 100644 data/2022/neurips/GraphQNTK: Quantum Neural Tangent Kernel for Graph Data create mode 100644 data/2022/neurips/Graphein - a Python Library for Geometric Deep Learning and Network Analysis on Biomolecular Structures and Interaction Networks create mode 100644 data/2022/neurips/Green Hierarchical Vision Transformer for Masked Image Modeling create mode 100644 data/2022/neurips/GriddlyJS: A Web IDE for Reinforcement Learning create mode 100644 data/2022/neurips/Grounded Reinforcement Learning: Learning to Win the Game under Human Commands create mode 100644 data/2022/neurips/Grounded Video Situation Recognition create mode 100644 data/2022/neurips/Grounding Aleatoric Uncertainty for Unsupervised Environment Design create mode 100644 data/2022/neurips/Group Meritocratic Fairness in Linear Contextual Bandits create mode 100644 data/2022/neurips/Grow and Merge: A Unified Framework for Continuous Categories Discovery create mode 100644 data/2022/neurips/Guaranteed Conservation of Momentum for Learning Particle-based Fluid Dynamics create mode 100644 data/2022/neurips/HAPI: A Large-scale Longitudinal Dataset of Commercial ML API Predictions create mode 100644 data/2022/neurips/HF-NeuS: Improved Surface Reconstruction Using High-Frequency Details create mode 100644 data/2022/neurips/HSDF: Hybrid Sign and Distance Field for Modeling Surfaces with Arbitrary Topologies create mode 100644 data/2022/neurips/HSurf-Net: Normal Estimation for 3D Point Clouds by Learning Hyper Surfaces create mode 100644 data/2022/neurips/HUMANISE: Language-conditioned Human Motion Generation in 3D Scenes create mode 100644 data/2022/neurips/HUMUS-Net: Hybrid Unrolled Multi-scale Network Architecture for Accelerated MRI Reconstruction create mode 100644 data/2022/neurips/HYPRO: A Hybridly Normalized Probabilistic Model for Long-Horizon Prediction of Event Sequences create mode 100644 data/2022/neurips/Hamiltonian Latent Operators for content and motion disentanglement in image sequences create mode 100644 data/2022/neurips/Hand-Object Interaction Image Generation create mode 100644 data/2022/neurips/HandMeThat: Human-Robot Communication in Physical and Social Environments create mode 100644 data/2022/neurips/Handcrafted Backdoors in Deep Neural Networks create mode 100644 data/2022/neurips/Hard ImageNet: Segmentations for Objects with Strong Spurious Cues create mode 100644 data/2022/neurips/Hardness in Markov Decision Processes: Theory and Practice create mode 100644 data/2022/neurips/Hardness of Noise-Free Learning for Two-Hidden-Layer Neural Networks create mode 100644 data/2022/neurips/Harmonizing the object recognition strategies of deep neural networks with humans create mode 100644 data/2022/neurips/Heatmap Distribution Matching for Human Pose Estimation create mode 100644 data/2022/neurips/Hedging as Reward Augmentation in Probabilistic Graphical Models create mode 100644 data/2022/neurips/Heterogeneous Skill Learning for Multi-agent Tasks create mode 100644 data/2022/neurips/Hidden Progress in Deep Learning: SGD Learns Parities Near the Computational Limit create mode 100644 data/2022/neurips/Hiding Images in Deep Probabilistic Models create mode 100644 data/2022/neurips/HierSpeech: Bridging the Gap between Text and Speech by Hierarchical Variational Inference using Self-supervised Representations for Speech Synthesis create mode 100644 data/2022/neurips/Hierarchical Agglomerative Graph Clustering in Poly-Logarithmic Depth create mode 100644 data/2022/neurips/Hierarchical Channel-spatial Encoding for Communication-efficient Collaborative Learning create mode 100644 data/2022/neurips/Hierarchical Graph Transformer with Adaptive Node Sampling create mode 100644 data/2022/neurips/Hierarchical Lattice Layer for Partially Monotone Neural Networks create mode 100644 data/2022/neurips/Hierarchical Normalization for Robust Monocular Depth Estimation create mode 100644 data/2022/neurips/Hierarchical classification at multiple operating points create mode 100644 data/2022/neurips/High-Order Pooling for Graph Neural Networks with Tensor Decomposition create mode 100644 data/2022/neurips/High-dimensional Additive Gaussian Processes under Monotonicity Constraints create mode 100644 data/2022/neurips/High-dimensional Asymptotics of Feature Learning: How One Gradient Step Improves the Representation create mode 100644 data/2022/neurips/High-dimensional limit theorems for SGD: Effective dynamics and critical scaling create mode 100644 data/2022/neurips/Hilbert Distillation for Cross-Dimensionality Networks create mode 100644 data/2022/neurips/Holomorphic Equilibrium Propagation Computes Exact Gradients Through Finite Size Oscillations create mode 100644 data/2022/neurips/Homomorphic Matrix Completion create mode 100644 data/2022/neurips/Honor of Kings Arena: an Environment for Generalization in Competitive Reinforcement Learning create mode 100644 data/2022/neurips/HorNet: Efficient High-Order Spatial Interactions with Recursive Gated Convolutions create mode 100644 data/2022/neurips/House of Cans: Covert Transmission of Internal Datasets via Capacity-Aware Neuron Steganography create mode 100644 data/2022/neurips/How Mask Matters: Towards Theoretical Understandings of Masked Autoencoders create mode 100644 data/2022/neurips/How Powerful are K-hop Message Passing Graph Neural Networks create mode 100644 data/2022/neurips/How Sampling Impacts the Robustness of Stochastic Neural Networks create mode 100644 data/2022/neurips/How Transferable are Video Representations Based on Synthetic Data? create mode 100644 data/2022/neurips/How Well Do Unsupervised Learning Algorithms Model Human Real-time and Life-long Learning? create mode 100644 data/2022/neurips/How Would The Viewer Feel? Estimating Wellbeing From Video Scenarios create mode 100644 data/2022/neurips/How and Why to Manipulate Your Own Agent: On the Incentives of Users of Learning Agents create mode 100644 data/2022/neurips/How to talk so AI will learn: Instructions, descriptions, and autonomy create mode 100644 data/2022/neurips/Hub-Pathway: Transfer Learning from A Hub of Pre-trained Models create mode 100644 data/2022/neurips/Human-AI Collaborative Bayesian Optimisation create mode 100644 data/2022/neurips/Human-AI Shared Control via Policy Dissection create mode 100644 data/2022/neurips/Human-Robotic Prosthesis as Collaborating Agents for Symmetrical Walking create mode 100644 data/2022/neurips/HumanLiker: A Human-like Object Detector to Model the Manual Labeling Process create mode 100644 data/2022/neurips/Hybrid Neural Autoencoders for Stimulus Encoding in Visual and Other Sensory Neuroprostheses create mode 100644 data/2022/neurips/Hyper-Representations as Generative Models: Sampling Unseen Neural Network Weights create mode 100644 data/2022/neurips/HyperDomainNet: Universal Domain Adaptation for Generative Adversarial Networks create mode 100644 data/2022/neurips/HyperMiner: Topic Taxonomy Mining with Hyperbolic Embedding create mode 100644 data/2022/neurips/HyperTree Proof Search for Neural Theorem Proving create mode 100644 data/2022/neurips/Hyperbolic Embedding Inference for Structured Multi-Label Prediction create mode 100644 data/2022/neurips/Hyperbolic Feature Augmentation via Distribution Estimation and Infinite Sampling on Manifolds create mode 100644 data/2022/neurips/Hyperparameter Sensitivity in Deep Outlier Detection: Analysis and a Scalable Hyper-Ensemble Solution create mode 100644 data/2022/neurips/Hypothesis Testing for Differentially Private Linear Regression create mode 100644 data/2022/neurips/I2DFormer: Learning Image to Document Attention for Zero-Shot Image Classification create mode 100644 data/2022/neurips/I2Q: A Fully Decentralized Q-Learning Algorithm create mode 100644 data/2022/neurips/IKEA-Manual: Seeing Shape Assembly Step by Step create mode 100644 data/2022/neurips/IM-Loss: Information Maximization Loss for Spiking Neural Networks create mode 100644 data/2022/neurips/IMED-RL: Regret optimal learning of ergodic Markov decision processes create mode 100644 data/2022/neurips/INRAS: Implicit Neural Representation for Audio Scenes create mode 100644 data/2022/neurips/Identifiability and generalizability from multiple experts in Inverse Reinforcement Learning create mode 100644 data/2022/neurips/Identifiability of deep generative models without auxiliary information create mode 100644 data/2022/neurips/Identification, Amplification and Measurement: A bridge to Gaussian Differential Privacy create mode 100644 data/2022/neurips/Identifying good directions to escape the NTK regime and efficiently learn low-degree plus sparse polynomials create mode 100644 data/2022/neurips/If Influence Functions are the Answer, Then What is the Question? create mode 100644 data/2022/neurips/Imbalance Trouble: Revisiting Neural-Collapse Geometry create mode 100644 data/2022/neurips/Imitating Past Successes can be Very Suboptimal create mode 100644 data/2022/neurips/Implications of Model Indeterminacy for Explanations of Automated Decisions create mode 100644 data/2022/neurips/Implicit Bias of Gradient Descent on Reparametrized Models: On Equivalence to Mirror Descent create mode 100644 data/2022/neurips/Implicit Neural Representations with Levels-of-Experts create mode 100644 data/2022/neurips/Implicit Regularization or Implicit Conditioning? Exact Risk Trajectories of SGD in High Dimensions create mode 100644 data/2022/neurips/Implicit Warping for Animation with Image Sets create mode 100644 data/2022/neurips/Improved Algorithms for Neural Active Learning create mode 100644 data/2022/neurips/Improved Bounds on Neural Complexity for Representing Piecewise Linear Functions create mode 100644 data/2022/neurips/Improved Convergence Rate of Stochastic Gradient Langevin Dynamics with Variance Reduction and its Application to Optimization create mode 100644 data/2022/neurips/Improved Coresets for Euclidean k-Means create mode 100644 data/2022/neurips/Improved Differential Privacy for SGD via Optimal Private Linear Operators on Adaptive Streams create mode 100644 data/2022/neurips/Improved Feature Distillation via Projector Ensemble create mode 100644 data/2022/neurips/Improved Fine-Tuning by Better Leveraging Pre-Training Data create mode 100644 data/2022/neurips/Improved Imaging by Invex Regularizers with Global Optima Guarantees create mode 100644 data/2022/neurips/Improved Regret Analysis for Variance-Adaptive Linear Bandits and Horizon-Free Linear Mixture MDPs create mode 100644 data/2022/neurips/Improved Utility Analysis of Private CountSketch create mode 100644 data/2022/neurips/Improved techniques for deterministic l2 robustness create mode 100644 data/2022/neurips/Improving 3D-aware Image Synthesis with A Geometry-aware Discriminator create mode 100644 data/2022/neurips/Improving Barely Supervised Learning by Discriminating Unlabeled Samples with Super-Class create mode 100644 data/2022/neurips/Improving Certified Robustness via Statistical Learning with Logical Reasoning create mode 100644 data/2022/neurips/Improving Diffusion Models for Inverse Problems using Manifold Constraints create mode 100644 data/2022/neurips/Improving GANs with A Dynamic Discriminator create mode 100644 data/2022/neurips/Improving Generative Adversarial Networks via Adversarial Learning in Latent Space create mode 100644 data/2022/neurips/Improving Intrinsic Exploration with Language Abstractions create mode 100644 data/2022/neurips/Improving Neural Ordinary Differential Equations with Nesterov's Accelerated Gradient Method create mode 100644 data/2022/neurips/Improving Out-of-Distribution Generalization by Adversarial Training with Structured Priors create mode 100644 data/2022/neurips/Improving Policy Learning via Language Dynamics Distillation create mode 100644 data/2022/neurips/Improving Self-Supervised Learning by Characterizing Idealized Representations create mode 100644 data/2022/neurips/Improving Task-Specific Generalization in Few-Shot Learning via Adaptive Vicinal Risk Minimization create mode 100644 data/2022/neurips/Improving Transformer with an Admixture of Attention Heads create mode 100644 data/2022/neurips/Improving Variational Autoencoders with Density Gap-based Regularization create mode 100644 data/2022/neurips/Improving Zero-Shot Generalization in Offline Reinforcement Learning using Generalized Similarity Functions create mode 100644 data/2022/neurips/In Defense of the Unitary Scalarization for Deep Multi-Task Learning create mode 100644 data/2022/neurips/In Differential Privacy, There is Truth: on Vote-Histogram Leakage in Ensemble Private Learning create mode 100644 data/2022/neurips/In What Ways Are Deep Neural Networks Invariant and How Should We Measure This? create mode 100644 data/2022/neurips/In the Eye of the Beholder: Robust Prediction with Causal User Modeling create mode 100644 data/2022/neurips/Incentivizing Combinatorial Bandit Exploration create mode 100644 data/2022/neurips/Inception Transformer create mode 100644 data/2022/neurips/Incorporating Bias-aware Margins into Contrastive Loss for Collaborative Filtering create mode 100644 data/2022/neurips/Increasing Confidence in Adversarial Robustness Evaluations create mode 100644 data/2022/neurips/Increasing the Scope as You Learn: Adaptive Bayesian Optimization in Nested Subspaces create mode 100644 data/2022/neurips/Incrementality Bidding via Reinforcement Learning under Mixed and Delayed Rewards create mode 100644 data/2022/neurips/Independence Testing for Bounded Degree Bayesian Networks create mode 100644 data/2022/neurips/Independence Testing-Based Approach to Causal Discovery under Measurement Error and Linear Non-Gaussian Models create mode 100644 data/2022/neurips/Indicators of Attack Failure: Debugging and Improving Optimization of Adversarial Examples create mode 100644 data/2022/neurips/Inducing Equilibria via Incentives: Simultaneous Design-and-Play Ensures Global Convergence create mode 100644 data/2022/neurips/Inducing Neural Collapse in Imbalanced Learning: Do We Really Need a Learnable Classifier at the End of Deep Neural Network? create mode 100644 data/2022/neurips/Inductive Logical Query Answering in Knowledge Graphs create mode 100644 data/2022/neurips/Inference and Sampling for Archimax Copulas create mode 100644 data/2022/neurips/Infinite Recommendation Networks: A Data-Centric Approach create mode 100644 data/2022/neurips/Infinite-Fidelity Coregionalization for Physical Simulation create mode 100644 data/2022/neurips/Influencing Long-Term Behavior in Multiagent Reinforcement Learning create mode 100644 data/2022/neurips/Information bottleneck theory of high-dimensional regression: relevancy, efficiency and optimality create mode 100644 data/2022/neurips/Information-Theoretic GAN Compression with Variational Energy-based Model create mode 100644 data/2022/neurips/Information-Theoretic Safe Exploration with Gaussian Processes create mode 100644 data/2022/neurips/Inherently Explainable Reinforcement Learning in Natural Language create mode 100644 data/2022/neurips/Injecting Domain Knowledge from Empirical Interatomic Potentials to Neural Networks for Predicting Material Properties create mode 100644 data/2022/neurips/InsNet: An Efficient, Flexible, and Performant Insertion-based Text Generation Model create mode 100644 data/2022/neurips/InsPro: Propagating Instance Query and Proposal for Online Video Instance Segmentation create mode 100644 data/2022/neurips/Insights into Pre-training via Simpler Synthetic Tasks create mode 100644 data/2022/neurips/Instability and Local Minima in GAN Training with Kernel Discriminators create mode 100644 data/2022/neurips/Instance-Based Uncertainty Estimation for Gradient-Boosted Regression Trees create mode 100644 data/2022/neurips/Instance-Dependent Near-Optimal Policy Identification in Linear MDPs via Online Experiment Design create mode 100644 data/2022/neurips/Instance-based Learning for Knowledge Base Completion create mode 100644 data/2022/neurips/Instance-optimal PAC Algorithms for Contextual Bandits create mode 100644 data/2022/neurips/Integral Probability Metrics PAC-Bayes Bounds create mode 100644 data/2022/neurips/Interaction Modeling with Multiplex Attention create mode 100644 data/2022/neurips/Interaction-Grounded Learning with Action-Inclusive Feedback create mode 100644 data/2022/neurips/Intermediate Prototype Mining Transformer for Few-Shot Semantic Segmentation create mode 100644 data/2022/neurips/Interpolation and Regularization for Causal Learning create mode 100644 data/2022/neurips/Interpreting Operation Selection in Differentiable Architecture Search: A Perspective from Influence-Directed Explanations create mode 100644 data/2022/neurips/Interventions, Where and How? Experimental Design for Causal Models at Scale create mode 100644 data/2022/neurips/Intra-agent speech permits zero-shot task acquisition create mode 100644 data/2022/neurips/Intrinsic dimensionality estimation using Normalizing Flows create mode 100644 data/2022/neurips/Introspective Learning : A Two-Stage approach for Inference in Neural Networks create mode 100644 data/2022/neurips/Invariance Learning based on Label Hierarchy create mode 100644 data/2022/neurips/Invariance Learning in Deep Neural Networks with Differentiable Laplace Approximations create mode 100644 data/2022/neurips/Invariance-Aware Randomized Smoothing Certificates create mode 100644 data/2022/neurips/Invariant and Transportable Representations for Anti-Causal Domain Shifts create mode 100644 data/2022/neurips/Inverse Design for Fluid-Structure Interactions using Graph Network Simulators create mode 100644 data/2022/neurips/Inverse Game Theory for Stackelberg Games: the Blessing of Bounded Rationality create mode 100644 data/2022/neurips/Invertible Monotone Operators for Normalizing Flows create mode 100644 data/2022/neurips/Is Integer Arithmetic Enough for Deep Learning Training? create mode 100644 data/2022/neurips/Is Out-of-Distribution Detection Learnable? create mode 100644 data/2022/neurips/Is Sortition Both Representative and Fair? create mode 100644 data/2022/neurips/Is a Modular Architecture Enough? create mode 100644 data/2022/neurips/Is one annotation enough? - A data-centric image classification benchmark for noisy and ambiguous label estimation create mode 100644 data/2022/neurips/Is this the Right Neighborhood? Accurate and Query Efficient Model Agnostic Explanations create mode 100644 data/2022/neurips/Iso-Dream: Isolating and Leveraging Noncontrollable Visual Dynamics in World Models create mode 100644 data/2022/neurips/Isometric 3D Adversarial Examples in the Physical World create mode 100644 data/2022/neurips/Iterative Feature Matching: Toward Provable Domain Generalization with Logarithmic Environments create mode 100644 data/2022/neurips/Iterative Scene Graph Generation create mode 100644 data/2022/neurips/Iterative Structural Inference of Directed Graphs create mode 100644 data/2022/neurips/JAHS-Bench-201: A Foundation For Research On Joint Architecture And Hyperparameter Search create mode 100644 data/2022/neurips/JAWS: Auditing Predictive Uncertainty Under Covariate Shift create mode 100644 data/2022/neurips/Joint Entropy Search For Maximally-Informed Bayesian Optimization create mode 100644 data/2022/neurips/Joint Entropy Search for Multi-Objective Bayesian Optimization create mode 100644 data/2022/neurips/Joint Learning of 2D-3D Weakly Supervised Semantic Segmentation create mode 100644 data/2022/neurips/Jump Self-attention: Capturing High-order Statistics in Transformers create mode 100644 data/2022/neurips/K-LITE: Learning Transferable Visual Models with External Knowledge create mode 100644 data/2022/neurips/K-Radar: 4D Radar Object Detection for Autonomous Driving in Various Weather Conditions create mode 100644 data/2022/neurips/KERPLE: Kernelized Relative Positional Embedding for Length Extrapolation create mode 100644 data/2022/neurips/KSD Aggregated Goodness-of-fit Test create mode 100644 data/2022/neurips/Kantorovich Strikes Back! Wasserstein GANs are not Optimal Transport? create mode 100644 data/2022/neurips/Kernel Interpolation with Sparse Grids create mode 100644 data/2022/neurips/Kernel Memory Networks: A Unifying Framework for Memory Modeling create mode 100644 data/2022/neurips/Kernel Multimodal Continuous Attention create mode 100644 data/2022/neurips/Kernel similarity matching with Hebbian networks create mode 100644 data/2022/neurips/Keypoint-Guided Optimal Transport with Applications in Heterogeneous Domain Adaptation create mode 100644 data/2022/neurips/Knowledge Distillation Improves Graph Structure Augmentation for Graph Neural Networks create mode 100644 data/2022/neurips/Knowledge Distillation from A Stronger Teacher create mode 100644 data/2022/neurips/Knowledge Distillation: Bad Models Can Be Good Role Models create mode 100644 data/2022/neurips/Knowledge-Aware Bayesian Deep Topic Model create mode 100644 data/2022/neurips/LAION-5B: An open large-scale dataset for training next generation image-text models create mode 100644 data/2022/neurips/LAMP: Extracting Text from Gradients with Language Model Priors create mode 100644 data/2022/neurips/LAPO: Latent-Variable Advantage-Weighted Policy Optimization for Offline Reinforcement Learning create mode 100644 data/2022/neurips/LASSIE: Learning Articulated Shapes from Sparse Image Ensemble via 3D Part Discovery create mode 100644 data/2022/neurips/LBD: Decouple Relevance and Observation for Individual-Level Unbiased Learning to Rank create mode 100644 data/2022/neurips/LDSA: Learning Dynamic Subtask Assignment in Cooperative Multi-Agent Reinforcement Learning create mode 100644 data/2022/neurips/LECO: Learnable Episodic Count for Task-Specific Intrinsic Reward create mode 100644 data/2022/neurips/LGDN: Language-Guided Denoising Network for Video-Language Modeling create mode 100644 data/2022/neurips/LIFT: Language-Interfaced Fine-Tuning for Non-language Machine Learning Tasks create mode 100644 data/2022/neurips/LION: Latent Point Diffusion Models for 3D Shape Generation create mode 100644 data/2022/neurips/LIPS - Learning Industrial Physical Simulation benchmark suite create mode 100644 data/2022/neurips/LISA: Learning Interpretable Skill Abstractions from Language create mode 100644 data/2022/neurips/LOG: Active Model Adaptation for Label-Efficient OOD Generalization create mode 100644 data/2022/neurips/LOT: Layer-wise Orthogonal Training on Improving l2 Certified Robustness create mode 100644 data/2022/neurips/LST: Ladder Side-Tuning for Parameter and Memory Efficient Transfer Learning create mode 100644 data/2022/neurips/LTMD: Learning Improvement of Spiking Neural Networks with Learnable Thresholding Neurons and Moderate Dropout create mode 100644 data/2022/neurips/Label Noise in Adversarial Training: A Novel Perspective to Study Robust Overfitting create mode 100644 data/2022/neurips/Label-Aware Global Consistency for Multi-Label Learning with Single Positive Labels create mode 100644 data/2022/neurips/Label-invariant Augmentation for Semi-Supervised Graph Classification create mode 100644 data/2022/neurips/Langevin Autoencoders for Learning Deep Latent Variable Models create mode 100644 data/2022/neurips/Language Conditioned Spatial Relation Reasoning for 3D Object Grounding create mode 100644 data/2022/neurips/Language Models with Image Descriptors are Strong Few-Shot Video-Language Learners create mode 100644 data/2022/neurips/Laplacian Autoencoders for Learning Stochastic Representations create mode 100644 data/2022/neurips/Large Language Models are Zero-Shot Reasoners create mode 100644 data/2022/neurips/Large-Scale Differentiable Causal Discovery of Factor Graphs create mode 100644 data/2022/neurips/Large-Scale Retrieval for Reinforcement Learning create mode 100644 data/2022/neurips/Large-batch Optimization for Dense Visual Predictions: Training Faster R-CNN in 4.2 Minutes create mode 100644 data/2022/neurips/Large-scale Optimization of Partial AUC in a Range of False Positive Rates create mode 100644 data/2022/neurips/LasUIE: Unifying Information Extraction with Latent Adaptive Structure-aware Generative Language Model create mode 100644 data/2022/neurips/Last-Iterate Convergence of Optimistic Gradient Method for Monotone Variational Inequalities create mode 100644 data/2022/neurips/Latency-aware Spatial-wise Dynamic Networks create mode 100644 data/2022/neurips/Latent Hierarchical Causal Structure Discovery with Rank Constraints create mode 100644 data/2022/neurips/Latent Planning via Expansive Tree Search create mode 100644 data/2022/neurips/Layer Freezing & Data Sieving: Missing Pieces of a Generic Framework for Sparse Training create mode 100644 data/2022/neurips/Lazy and Fast Greedy MAP Inference for Determinantal Point Process create mode 100644 data/2022/neurips/Learn to Explain: Multimodal Reasoning via Thought Chains for Science Question Answering create mode 100644 data/2022/neurips/Learn to Match with No Regret: Reinforcement Learning in Markov Matching Markets create mode 100644 data/2022/neurips/Learn what matters: cross-domain imitation learning with task-relevant embeddings create mode 100644 data/2022/neurips/Learnable Polyphase Sampling for Shift Invariant and Equivariant Convolutional Networks create mode 100644 data/2022/neurips/Learning (Very) Simple Generative Models Is Hard create mode 100644 data/2022/neurips/Learning Active Camera for Multi-Object Navigation create mode 100644 data/2022/neurips/Learning Articulated Rigid Body Dynamics with Lagrangian Graph Neural Network create mode 100644 data/2022/neurips/Learning Audio-Visual Dynamics Using Scene Graphs for Audio Source Separation create mode 100644 data/2022/neurips/Learning Best Combination for Efficient N: M Sparsity create mode 100644 data/2022/neurips/Learning Bipartite Graphs: Heavy Tails and Multiple Components create mode 100644 data/2022/neurips/Learning Causally Invariant Representations for Out-of-Distribution Generalization on Graphs create mode 100644 data/2022/neurips/Learning Chaotic Dynamics in Dissipative Systems create mode 100644 data/2022/neurips/Learning Concept Credible Models for Mitigating Shortcuts create mode 100644 data/2022/neurips/Learning Consistency-Aware Unsigned Distance Functions Progressively from Raw Point Clouds create mode 100644 data/2022/neurips/Learning Contrastive Embedding in Low-Dimensional Space create mode 100644 data/2022/neurips/Learning Debiased Classifier with Biased Committee create mode 100644 data/2022/neurips/Learning Deep Input-Output Stable Dynamics create mode 100644 data/2022/neurips/Learning Dense Object Descriptors from Multiple Views for Low-shot Category Generalization create mode 100644 data/2022/neurips/Learning Distinct and Representative Modes for Image Captioning create mode 100644 data/2022/neurips/Learning Distributed and Fair Policies for Network Load Balancing as Markov Potential Game create mode 100644 data/2022/neurips/Learning Distributions Generated by Single-Layer ReLU Networks in the Presence of Arbitrary Outliers create mode 100644 data/2022/neurips/Learning Dynamical Systems via Koopman Operator Regression in Reproducing Kernel Hilbert Spaces create mode 100644 data/2022/neurips/Learning Efficient Vision Transformers via Fine-Grained Manifold Distillation create mode 100644 data/2022/neurips/Learning Energy Networks with Generalized Fenchel-Young Losses create mode 100644 data/2022/neurips/Learning Enhanced Representation for Tabular Data via Neighborhood Propagation create mode 100644 data/2022/neurips/Learning Equivariant Segmentation with Instance-Unique Querying create mode 100644 data/2022/neurips/Learning Expressive Meta-Representations with Mixture of Expert Neural Processes create mode 100644 data/2022/neurips/Learning Fractional White Noises in Neural Stochastic Differential Equations create mode 100644 data/2022/neurips/Learning General World Models in a Handful of Reward-Free Deployments create mode 100644 data/2022/neurips/Learning Generalizable Models for Vehicle Routing Problems via Knowledge Distillation create mode 100644 data/2022/neurips/Learning Generalizable Part-based Feature Representation for 3D Point Clouds create mode 100644 data/2022/neurips/Learning Generalized Policy Automata for Relational Stochastic Shortest Path Problems create mode 100644 data/2022/neurips/Learning Graph-embedded Key-event Back-tracing for Object Tracking in Event Clouds create mode 100644 data/2022/neurips/Learning Individualized Treatment Rules with Many Treatments: A Supervised Clustering Approach Using Adaptive Fusion create mode 100644 data/2022/neurips/Learning Infinite-Horizon Average-Reward Restless Multi-Action Bandits via Index Awareness create mode 100644 data/2022/neurips/Learning Interface Conditions in Domain Decomposition Solvers create mode 100644 data/2022/neurips/Learning Invariant Graph Representations for Out-of-Distribution Generalization create mode 100644 data/2022/neurips/Learning Latent Seasonal-Trend Representations for Time Series Forecasting create mode 100644 data/2022/neurips/Learning Long-Term Crop Management Strategies with CyclesGym create mode 100644 data/2022/neurips/Learning Manifold Dimensions with Conditional Variational Autoencoders create mode 100644 data/2022/neurips/Learning Mixed Multinomial Logits with Provable Guarantees create mode 100644 data/2022/neurips/Learning Modular Simulations for Homogeneous Systems create mode 100644 data/2022/neurips/Learning Multi-resolution Functional Maps with Spectral Attention for Robust Shape Matching create mode 100644 data/2022/neurips/Learning NP-Hard Multi-Agent Assignment Planning using GNN: Inference on a Random Graph and Provable Auction-Fitted Q-learning create mode 100644 data/2022/neurips/Learning Neural Acoustic Fields create mode 100644 data/2022/neurips/Learning Neural Set Functions Under the Optimal Subset Oracle create mode 100644 data/2022/neurips/Learning Optical Flow from Continuous Spike Streams create mode 100644 data/2022/neurips/Learning Optimal Flows for Non-Equilibrium Importance Sampling create mode 100644 data/2022/neurips/Learning Options via Compression create mode 100644 data/2022/neurips/Learning Partial Equivariances From Data create mode 100644 data/2022/neurips/Learning Physical Dynamics with Subequivariant Graph Neural Networks create mode 100644 data/2022/neurips/Learning Physics Constrained Dynamics Using Autoencoders create mode 100644 data/2022/neurips/Learning Predictions for Algorithms with Predictions create mode 100644 data/2022/neurips/Learning Probabilistic Models from Generator Latent Spaces with Hat EBM create mode 100644 data/2022/neurips/Learning Recourse on Instance Environment to Enhance Prediction Accuracy create mode 100644 data/2022/neurips/Learning Representations via a Robust Behavioral Metric for Deep Reinforcement Learning create mode 100644 data/2022/neurips/Learning Robust Dynamics through Variational Sparse Gating create mode 100644 data/2022/neurips/Learning Robust Rule Representations for Abstract Reasoning via Internal Inferences create mode 100644 data/2022/neurips/Learning State-Aware Visual Representations from Audible Interactions create mode 100644 data/2022/neurips/Learning Structure from the Ground up - Hierarchical Representation Learning by Chunking create mode 100644 data/2022/neurips/Learning Substructure Invariance for Out-of-Distribution Molecular Representations create mode 100644 data/2022/neurips/Learning Superpoint Graph Cut for 3D Instance Segmentation create mode 100644 data/2022/neurips/Learning Symmetric Rules with SATNet create mode 100644 data/2022/neurips/Learning Tractable Probabilistic Models from Inconsistent Local Estimates create mode 100644 data/2022/neurips/Learning Two-Player Markov Games: Neural Function Approximation and Correlated Equilibrium create mode 100644 data/2022/neurips/Learning Viewpoint-Agnostic Visual Representations by Recovering Tokens in 3D Space create mode 100644 data/2022/neurips/Learning a Condensed Frame for Memory-Efficient Video Class-Incremental Learning create mode 100644 data/2022/neurips/Learning and Covering Sums of Independent Random Variables with Unbounded Support create mode 100644 data/2022/neurips/Learning dynamics of deep linear networks with multiple pathways create mode 100644 data/2022/neurips/Learning from Distributed Users in Contextual Linear Bandits Without Sharing the Context create mode 100644 data/2022/neurips/Learning from Few Samples: Transformation-Invariant SVMs with Composition and Locality at Multiple Scales create mode 100644 data/2022/neurips/Learning from Future: A Novel Self-Training Framework for Semantic Segmentation create mode 100644 data/2022/neurips/Learning from Label Proportions by Learning with Label Noise create mode 100644 data/2022/neurips/Learning from Stochastically Revealed Preference create mode 100644 data/2022/neurips/Learning from a Sample in Online Algorithms create mode 100644 data/2022/neurips/Learning in Congestion Games with Bandit Feedback create mode 100644 data/2022/neurips/Learning in Observable POMDPs, without Computationally Intractable Oracles create mode 100644 data/2022/neurips/Learning interacting dynamical systems with latent Gaussian process ODEs create mode 100644 data/2022/neurips/Learning low-dimensional generalizable natural features from retina using a U-net create mode 100644 data/2022/neurips/Learning on Arbitrary Graph Topologies via Predictive Coding create mode 100644 data/2022/neurips/Learning on the Edge: Online Learning with Stochastic Feedback Graphs create mode 100644 data/2022/neurips/Learning single-index models with shallow neural networks create mode 100644 data/2022/neurips/Learning sparse features can lead to overfitting in neural networks create mode 100644 data/2022/neurips/Learning the Structure of Large Networked Systems Obeying Conservation Laws create mode 100644 data/2022/neurips/Learning to Accelerate Partial Differential Equations via Latent Global Evolution create mode 100644 data/2022/neurips/Learning to Attack Federated Learning: A Model-based Reinforcement Learning Attack Framework create mode 100644 data/2022/neurips/Learning to Branch with Tree MDPs create mode 100644 data/2022/neurips/Learning to Break the Loop: Analyzing and Mitigating Repetitions for Neural Text Generation create mode 100644 data/2022/neurips/Learning to Compare Nodes in Branch and Bound with Graph Neural Networks create mode 100644 data/2022/neurips/Learning to Configure Computer Networks with Neural Algorithmic Reasoning create mode 100644 data/2022/neurips/Learning to Constrain Policy Optimization with Virtual Trust Region create mode 100644 data/2022/neurips/Learning to Discover and Detect Objects create mode 100644 data/2022/neurips/Learning to Drop Out: An Adversarial Approach to Training Sequence VAEs create mode 100644 data/2022/neurips/Learning to Follow Instructions in Text-Based Games create mode 100644 data/2022/neurips/Learning to Generate Inversion-Resistant Model Explanations create mode 100644 data/2022/neurips/Learning to Mitigate AI Collusion on Economic Platforms create mode 100644 data/2022/neurips/Learning to Navigate Wikipedia by Taking Random Walks create mode 100644 data/2022/neurips/Learning to Re-weight Examples with Optimal Transport for Imbalanced Classification create mode 100644 data/2022/neurips/Learning to Reason with Neural Networks: Generalization, Unseen Data and Boolean Measures create mode 100644 data/2022/neurips/Learning to Reconstruct Missing Data from Spatiotemporal Graphs with Sparse Observations create mode 100644 data/2022/neurips/Learning to Sample and Aggregate: Few-shot Reasoning over Temporal Knowledge Graphs create mode 100644 data/2022/neurips/Learning to Scaffold: Optimizing Model Explanations for Teaching create mode 100644 data/2022/neurips/Learning to Share in Networked Multi-Agent Reinforcement Learning create mode 100644 data/2022/neurips/Learning with convolution and pooling operations in kernel methods create mode 100644 data/2022/neurips/Learning with little mixing create mode 100644 data/2022/neurips/Learning-Augmented Algorithms for Online Linear and Semidefinite Programming create mode 100644 data/2022/neurips/Learning-based Motion Planning in Dynamic Environments Using GNNs and Temporal Encoding create mode 100644 data/2022/neurips/Left Heavy Tails and the Effectiveness of the Policy and Value Networks in DNN-based best-first search for Sokoban Planning create mode 100644 data/2022/neurips/Less-forgetting Multi-lingual Fine-tuning create mode 100644 data/2022/neurips/Let Images Give You More: Point Cloud Cross-Modal Training for Shape Analysis create mode 100644 data/2022/neurips/Lethal Dose Conjecture on Data Poisoning create mode 100644 data/2022/neurips/Leveraging Factored Action Spaces for Efficient Offline Reinforcement Learning in Healthcare create mode 100644 data/2022/neurips/Leveraging Inter-Layer Dependency for Post -Training Quantization create mode 100644 data/2022/neurips/Leveraging the Hints: Adaptive Bidding in Repeated First-Price Auctions create mode 100644 data/2022/neurips/LieGG: Studying Learned Lie Group Generators create mode 100644 data/2022/neurips/Lifelong Neural Predictive Coding: Learning Cumulatively Online without Forgetting create mode 100644 data/2022/neurips/Lifting Weak Supervision To Structured Prediction create mode 100644 data/2022/neurips/Lifting the Information Ratio: An Information-Theoretic Analysis of Thompson Sampling for Contextual Bandits create mode 100644 data/2022/neurips/Linear Label Ranking with Bounded Noise create mode 100644 data/2022/neurips/Linear tree shap create mode 100644 data/2022/neurips/Lipschitz Bandits with Batched Feedback create mode 100644 data/2022/neurips/List-Decodable Sparse Mean Estimation create mode 100644 data/2022/neurips/List-Decodable Sparse Mean Estimation via Difference-of-Pairs Filtering create mode 100644 data/2022/neurips/Listen to Interpret: Post-hoc Interpretability for Audio Networks with NMF create mode 100644 data/2022/neurips/LiteTransformerSearch: Training-free Neural Architecture Search for Efficient Language Models create mode 100644 data/2022/neurips/LobsDICE: Offline Learning from Observation via Stationary Distribution Correction Estimation create mode 100644 data/2022/neurips/Local Bayesian optimization via maximizing probability of descent create mode 100644 data/2022/neurips/Local Identifiability of Deep ReLU Neural Networks: the Theory create mode 100644 data/2022/neurips/Local Latent Space Bayesian Optimization over Structured Inputs create mode 100644 data/2022/neurips/Local Linear Convergence of Gradient Methods for Subspace Optimization via Strict Complementarity create mode 100644 data/2022/neurips/Local Metric Learning for Off-Policy Evaluation in Contextual Bandits with Continuous Actions create mode 100644 data/2022/neurips/Local Spatiotemporal Representation Learning for Longitudinally-consistent Neuroimage Analysis create mode 100644 data/2022/neurips/Local-Global MCMC kernels: the best of both worlds create mode 100644 data/2022/neurips/Locally Hierarchical Auto-Regressive Modeling for Image Generation create mode 100644 data/2022/neurips/Locating and Editing Factual Associations in GPT create mode 100644 data/2022/neurips/Log-Concave and Multivariate Canonical Noise Distributions for Differential Privacy create mode 100644 data/2022/neurips/Log-Linear-Time Gaussian Processes Using Binary Tree Kernels create mode 100644 data/2022/neurips/Log-Polar Space Convolution Layers create mode 100644 data/2022/neurips/LogiGAN: Learning Logical Reasoning via Adversarial Pre-training create mode 100644 data/2022/neurips/Logical Activation Functions: Logit-space equivalents of Probabilistic Boolean Operators create mode 100644 data/2022/neurips/Logical Credal Networks create mode 100644 data/2022/neurips/Long Range Graph Benchmark create mode 100644 data/2022/neurips/Long-Form Video-Language Pre-Training with Multimodal Temporal Contrastive Learning create mode 100644 data/2022/neurips/Look Around and Refer: 2D Synthetic Semantics Knowledge Distillation for 3D Visual Grounding create mode 100644 data/2022/neurips/Look More but Care Less in Video Recognition create mode 100644 data/2022/neurips/Look where you look! Saliency-guided Q-networks for generalization in visual Reinforcement Learning create mode 100644 data/2022/neurips/Losses Can Be Blessings: Routing Self-Supervised Speech Representations Towards Efficient Multilingual and Multitask Speech Processing create mode 100644 data/2022/neurips/Lost in Latent Space: Examining failures of disentangled models at combinatorial generalisation create mode 100644 data/2022/neurips/Lottery Tickets on a Data Diet: Finding Initializations with Sparse Trainable Networks create mode 100644 data/2022/neurips/Low-Rank Modular Reinforcement Learning via Muscle Synergy create mode 100644 data/2022/neurips/Low-rank Optimal Transport: Approximation, Statistics and Debiasing create mode 100644 data/2022/neurips/Low-rank lottery tickets: finding efficient low-rank neural networks via matrix differential equations create mode 100644 data/2022/neurips/Lower Bounds and Nearly Optimal Algorithms in Distributed Learning with Communication Compression create mode 100644 data/2022/neurips/Lower Bounds on Randomly Preconditioned Lasso via Robust Sparse Designs create mode 100644 data/2022/neurips/Luckiness in Multiscale Online Learning create mode 100644 data/2022/neurips/M2N: Mesh Movement Networks for PDE Solvers create mode 100644 data/2022/neurips/M4Singer: A Multi-Style, Multi-Singer and Musical Score Provided Mandarin Singing Corpus create mode 100644 data/2022/neurips/MABSplit: Faster Forest Training Using Multi-Armed Bandits create mode 100644 data/2022/neurips/MACE: Higher Order Equivariant Message Passing Neural Networks for Fast and Accurate Force Fields create mode 100644 data/2022/neurips/MACK: Multimodal Aligned Conceptual Knowledge for Unpaired Image-text Matching create mode 100644 data/2022/neurips/MATE: Benchmarking Multi-Agent Reinforcement Learning in Distributed Target Coverage Control create mode 100644 data/2022/neurips/MAgNet: Mesh Agnostic Neural PDE Solver create mode 100644 data/2022/neurips/MAtt: A Manifold Attention Network for EEG Decoding create mode 100644 data/2022/neurips/MBW: Multi-view Bootstrapping in the Wild create mode 100644 data/2022/neurips/MCL-GAN: Generative Adversarial Networks with Multiple Specialized Discriminators create mode 100644 data/2022/neurips/MCMAE: Masked Convolution Meets Masked Autoencoders create mode 100644 data/2022/neurips/MCVD - Masked Conditional Video Diffusion for Prediction, Generation, and Interpolation create mode 100644 data/2022/neurips/MEMO: Test Time Robustness via Adaptation and Augmentation create mode 100644 data/2022/neurips/METS-CoV: A Dataset of Medical Entity and Targeted Sentiment on COVID-19 Related Tweets create mode 100644 data/2022/neurips/MExMI: Pool-based Active Model Extraction Crossover Membership Inference create mode 100644 data/2022/neurips/MGNNI: Multiscale Graph Neural Networks with Implicit Layers create mode 100644 data/2022/neurips/MOMA-LRG: Language-Refined Graphs for Multi-Object Multi-Actor Activity Parsing create mode 100644 data/2022/neurips/MORA: Improving Ensemble Robustness Evaluation with Model Reweighing Attack create mode 100644 data/2022/neurips/MOVE: Unsupervised Movable Object Segmentation and Detection create mode 100644 data/2022/neurips/MSDS: A Large-Scale Chinese Signature and Token Digit String Dataset for Handwriting Verification create mode 100644 data/2022/neurips/MTNeuro: A Benchmark for Evaluating Representations of Brain Structure Across Multiple Levels of Abstraction create mode 100644 data/2022/neurips/MVP-N: A Dataset and Benchmark for Real-World Multi-View Object Classification create mode 100644 data/2022/neurips/Make Sharpness-Aware Minimization Stronger: A Sparsified Perturbation Approach create mode 100644 data/2022/neurips/Make Some Noise: Reliable and Efficient Single-Step Adversarial Training create mode 100644 data/2022/neurips/Make an Omelette with Breaking Eggs: Zero-Shot Learning for Novel Attribute Synthesis create mode 100644 data/2022/neurips/Making Look-Ahead Active Learning Strategies Feasible with Neural Tangent Kernels create mode 100644 data/2022/neurips/Making Sense of Dependence: Efficient Black-box Explanations Using Dependence Measure create mode 100644 data/2022/neurips/Manifold Interpolating Optimal-Transport Flows for Trajectory Inference create mode 100644 data/2022/neurips/Margin-Based Few-Shot Class-Incremental Learning with Class-Level Overfitting Mitigation create mode 100644 data/2022/neurips/Markov Chain Score Ascent: A Unifying Framework of Variational Inference with Markovian Gradients create mode 100644 data/2022/neurips/Markovian Interference in Experiments create mode 100644 data/2022/neurips/Marksman Backdoor: Backdoor Attacks with Arbitrary Target Class create mode 100644 data/2022/neurips/Mask Matching Transformer for Few-Shot Segmentation create mode 100644 data/2022/neurips/Mask-based Latent Reconstruction for Reinforcement Learning create mode 100644 data/2022/neurips/MaskPlace: Fast Chip Placement via Reinforced Visual Representation Learning create mode 100644 data/2022/neurips/MaskTune: Mitigating Spurious Correlations by Forcing to Explore create mode 100644 data/2022/neurips/Masked Autoencoders As Spatiotemporal Learners create mode 100644 data/2022/neurips/Masked Autoencoders that Listen create mode 100644 data/2022/neurips/Masked Autoencoding for Scalable and Generalizable Decision Making create mode 100644 data/2022/neurips/Masked Generative Adversarial Networks are Data-Efficient Generation Learners create mode 100644 data/2022/neurips/Masked Prediction: A Parameter Identifiability View create mode 100644 data/2022/neurips/Matching in Multi-arm Bandit with Collision create mode 100644 data/2022/neurips/Matrix Multiplicative Weights Updates in Quantum Zero-Sum Games: Conservation Laws & Recurrence create mode 100644 data/2022/neurips/Matryoshka Representation Learning create mode 100644 data/2022/neurips/Max-Min Off-Policy Actor-Critic Method Focusing on Worst-Case Robustness to Model Misspecification create mode 100644 data/2022/neurips/Maximizing Revenue under Market Shrinkage and Market Uncertainty create mode 100644 data/2022/neurips/Maximizing and Satisficing in Multi-armed Bandits with Graph Information create mode 100644 data/2022/neurips/Maximum Class Separation as Inductive Bias in One Matrix create mode 100644 data/2022/neurips/Maximum Common Subgraph Guided Graph Retrieval: Late and Early Interaction Networks create mode 100644 data/2022/neurips/Maximum Likelihood Training of Implicit Nonlinear Diffusion Model create mode 100644 data/2022/neurips/Maximum a posteriori natural scene reconstruction from retinal ganglion cells with deep denoiser priors create mode 100644 data/2022/neurips/Maximum-Likelihood Inverse Reinforcement Learning with Finite-Time Guarantees create mode 100644 data/2022/neurips/Mean Estimation in High-Dimensional Binary Markov Gaussian Mixture Models create mode 100644 data/2022/neurips/Mean Estimation with User-level Privacy under Data Heterogeneity create mode 100644 data/2022/neurips/Measures of Information Reflect Memorization Patterns create mode 100644 data/2022/neurips/Measuring Data Reconstruction Defenses in Collaborative Inference Systems create mode 100644 data/2022/neurips/Measuring and Reducing Model Update Regression in Structured Prediction for NLP create mode 100644 data/2022/neurips/Memorization Without Overfitting: Analyzing the Training Dynamics of Large Language Models create mode 100644 data/2022/neurips/Memorization and Optimization in Deep Neural Networks with Minimum Over-parameterization create mode 100644 data/2022/neurips/Memory Efficient Continual Learning with Transformers create mode 100644 data/2022/neurips/Memory safe computations with XLA compiler create mode 100644 data/2022/neurips/Merging Models with Fisher-Weighted Averaging create mode 100644 data/2022/neurips/Mesoscopic modeling of hidden spiking neurons create mode 100644 data/2022/neurips/Meta Reinforcement Learning with Finite Training Tasks - a Density Estimation Approach create mode 100644 data/2022/neurips/Meta-Album: Multi-domain Meta-Dataset for Few-Shot Image Classification create mode 100644 data/2022/neurips/Meta-Auto-Decoder for Solving Parametric Partial Differential Equations create mode 100644 data/2022/neurips/Meta-Complementing the Semantics of Short Texts in Neural Topic Models create mode 100644 data/2022/neurips/Meta-DMoE: Adapting to Domain Shift by Meta-Distillation from Mixture-of-Experts create mode 100644 data/2022/neurips/Meta-Learning Dynamics Forecasting Using Task Inference create mode 100644 data/2022/neurips/Meta-Learning with Self-Improving Momentum Target create mode 100644 data/2022/neurips/Meta-Query-Net: Resolving Purity-Informativeness Dilemma in Open-set Active Learning create mode 100644 data/2022/neurips/Meta-Reinforcement Learning with Self-Modifying Networks create mode 100644 data/2022/neurips/Meta-Reward-Net: Implicitly Differentiable Reward Learning for Preference-based Reinforcement Learning create mode 100644 data/2022/neurips/Meta-ticket: Finding optimal subnetworks for few-shot learning within randomly initialized neural networks create mode 100644 data/2022/neurips/MetaMask: Revisiting Dimensional Confounder for Self-Supervised Learning create mode 100644 data/2022/neurips/MetaTeacher: Coordinating Multi-Model Domain Adaptation for Medical Image Classification create mode 100644 data/2022/neurips/MetricFormer: A Unified Perspective of Correlation Exploring in Similarity Learning create mode 100644 data/2022/neurips/Micro and Macro Level Graph Modeling for Graph Variational Auto-Encoders create mode 100644 data/2022/neurips/Mildly Conservative Q-Learning for Offline Reinforcement Learning create mode 100644 data/2022/neurips/MinVIS: A Minimal Video Instance Segmentation Framework without Video-based Training create mode 100644 data/2022/neurips/Mind Reader: Reconstructing complex images from brain activities create mode 100644 data/2022/neurips/Mind the Gap: Understanding the Modality Gap in Multi-modal Contrastive Representation Learning create mode 100644 data/2022/neurips/MineDojo: Building Open-Ended Embodied Agents with Internet-Scale Knowledge create mode 100644 data/2022/neurips/Mingling Foresight with Imagination: Model-Based Cooperative Multi-Agent Reinforcement Learning create mode 100644 data/2022/neurips/Minimax Optimal Algorithms for Fixed-Budget Best Arm Identification create mode 100644 data/2022/neurips/Minimax Optimal Fixed-Budget Best Arm Identification in Linear Bandits create mode 100644 data/2022/neurips/Minimax Optimal Online Imitation Learning via Replay Estimation create mode 100644 data/2022/neurips/Minimax Regret for Cascading Bandits create mode 100644 data/2022/neurips/Minimax-Optimal Multi-Agent RL in Markov Games With a Generative Model create mode 100644 data/2022/neurips/Mining Multi-Label Samples from Single Positive Labels create mode 100644 data/2022/neurips/Mining Unseen Classes via Regional Objectness: A Simple Baseline for Incremental Segmentation create mode 100644 data/2022/neurips/Mirror Descent Maximizes Generalized Margin and Can Be Implemented Efficiently create mode 100644 data/2022/neurips/Mirror Descent with Relative Smoothness in Measure Spaces, with application to Sinkhorn and EM create mode 100644 data/2022/neurips/Mismatched No More: Joint Model-Policy Optimization for Model-Based RL create mode 100644 data/2022/neurips/MissDAG: Causal Discovery in the Presence of Missing Data with Continuous Additive Noise Models create mode 100644 data/2022/neurips/Missing Data Imputation and Acquisition with Deep Hierarchical Models and Hamiltonian Monte Carlo create mode 100644 data/2022/neurips/Misspecified Phase Retrieval with Generative Priors create mode 100644 data/2022/neurips/Mix and Reason: Reasoning over Semantic Topology with Data Mixing for Domain Generalization create mode 100644 data/2022/neurips/Mixture-of-Experts with Expert Choice Routing create mode 100644 data/2022/neurips/MoCapAct: A Multi-Task Dataset for Simulated Humanoid Control create mode 100644 data/2022/neurips/MoCoDA: Model-based Counterfactual Data Augmentation create mode 100644 data/2022/neurips/MoGDE: Boosting Mobile Monocular 3D Object Detection with Ground Depth Estimation create mode 100644 data/2022/neurips/MoVQ: Modulating Quantized Vectors for High-Fidelity Image Generation create mode 100644 data/2022/neurips/Model Preserving Compression for Neural Networks create mode 100644 data/2022/neurips/Model Zoos: A Dataset of Diverse Populations of Neural Network Models create mode 100644 data/2022/neurips/Model-Based Imitation Learning for Urban Driving create mode 100644 data/2022/neurips/Model-Based Offline Reinforcement Learning with Pessimism-Modulated Dynamics Belief create mode 100644 data/2022/neurips/Model-Based Opponent Modeling create mode 100644 data/2022/neurips/Model-based Lifelong Reinforcement Learning with Bayesian Exploration create mode 100644 data/2022/neurips/Model-based RL with Optimistic Posterior Sampling: Structural Conditions and Sample Complexity create mode 100644 data/2022/neurips/Model-based Safe Deep Reinforcement Learning via a Constrained Proximal Policy Optimization Algorithm create mode 100644 data/2022/neurips/Modeling Human Exploration Through Resource-Rational Reinforcement Learning create mode 100644 data/2022/neurips/Modeling Transitivity and Cyclicity in Directed Graphs via Binary Code Box Embeddings create mode 100644 data/2022/neurips/Modeling the Machine Learning Multiverse create mode 100644 data/2022/neurips/Models Out of Line: A Fourier Lens on Distribution Shift Robustness create mode 100644 data/2022/neurips/Moderate-fitting as a Natural Backdoor Defender for Pre-trained Language Models create mode 100644 data/2022/neurips/Modular Flows: Differential Molecular Generation create mode 100644 data/2022/neurips/Module-Aware Optimization for Auxiliary Learning create mode 100644 data/2022/neurips/Molecule Generation by Principal Subgraph Mining and Assembling create mode 100644 data/2022/neurips/Moment Distributionally Robust Tree Structured Prediction create mode 100644 data/2022/neurips/Momentum Adversarial Distillation: Handling Large Distribution Shifts in Data-Free Knowledge Distillation create mode 100644 data/2022/neurips/Momentum Aggregation for Private Non-convex ERM create mode 100644 data/2022/neurips/MonoSDF: Exploring Monocular Geometric Cues for Neural Implicit Surface Reconstruction create mode 100644 data/2022/neurips/Monocular Dynamic View Synthesis: A Reality Check create mode 100644 data/2022/neurips/Monte Carlo Augmented Actor-Critic for Sparse Reward Deep Reinforcement Learning from Suboptimal Demonstrations create mode 100644 data/2022/neurips/Monte Carlo Tree Descent for Black-Box Optimization create mode 100644 data/2022/neurips/Monte Carlo Tree Search based Variable Selection for High Dimensional Bayesian Optimization create mode 100644 data/2022/neurips/MorphTE: Injecting Morphology in Tensorized Embeddings create mode 100644 data/2022/neurips/Most Activation Functions Can Win the Lottery Without Excessive Depth create mode 100644 data/2022/neurips/Motion Transformer with Global Intention Localization and Local Movement Refinement create mode 100644 data/2022/neurips/Movement Penalized Bayesian Optimization with Application to Wind Energy Systems create mode 100644 data/2022/neurips/MsSVT: Mixed-scale Sparse Voxel Transformer for 3D Object Detection on Point Clouds create mode 100644 data/2022/neurips/Muffliato: Peer-to-Peer Privacy Amplification for Decentralized Optimization and Averaging create mode 100644 data/2022/neurips/Multi-Agent Reinforcement Learning is a Sequence Modeling Problem create mode 100644 data/2022/neurips/Multi-Class $H$-Consistency Bounds create mode 100644 data/2022/neurips/Multi-Fidelity Best-Arm Identification create mode 100644 data/2022/neurips/Multi-Game Decision Transformers create mode 100644 data/2022/neurips/Multi-Granularity Cross-modal Alignment for Generalized Medical Visual Representation Learning create mode 100644 data/2022/neurips/Multi-Instance Causal Representation Learning for Instance Label Prediction and Out-of-Distribution Generalization create mode 100644 data/2022/neurips/Multi-LexSum: Real-world Summaries of Civil Rights Lawsuits at Multiple Granularities create mode 100644 data/2022/neurips/Multi-Lingual Acquisition on Multimodal Pre-training for Cross-modal Retrieval create mode 100644 data/2022/neurips/Multi-Objective Deep Learning with Adaptive Reference Vectors create mode 100644 data/2022/neurips/Multi-Sample Training for Neural Image Compression create mode 100644 data/2022/neurips/Multi-Scale Adaptive Network for Single Image Denoising create mode 100644 data/2022/neurips/Multi-agent Dynamic Algorithm Configuration create mode 100644 data/2022/neurips/Multi-agent Performative Prediction with Greedy Deployment and Consensus Seeking Agents create mode 100644 data/2022/neurips/Multi-block Min-max Bilevel Optimization with Applications in Multi-task Deep AUC Maximization create mode 100644 data/2022/neurips/Multi-block-Single-probe Variance Reduced Estimator for Coupled Compositional Optimization create mode 100644 data/2022/neurips/Multi-dataset Training of Transformers for Robust Action Recognition create mode 100644 data/2022/neurips/Multi-fidelity Monte Carlo: a pseudo-marginal approach create mode 100644 data/2022/neurips/Multi-layer State Evolution Under Random Convolutional Design create mode 100644 data/2022/neurips/Multi-modal Grouping Network for Weakly-Supervised Audio-Visual Video Parsing create mode 100644 data/2022/neurips/Multi-objective Deep Data Generation with Correlated Property Control create mode 100644 data/2022/neurips/Multi-view Subspace Clustering on Topological Manifold create mode 100644 data/2022/neurips/MultiGuard: Provably Robust Multi-label Classification against Adversarial Examples create mode 100644 data/2022/neurips/MultiScan: Scalable RGBD scanning for 3D environments with articulated objects create mode 100644 data/2022/neurips/Multiagent Q-learning with Sub-Team Coordination create mode 100644 data/2022/neurips/Multiclass Learnability Beyond the PAC Framework: Universal Rates and Partial Concept Classes create mode 100644 data/2022/neurips/Multilingual Abusive Comment Detection at Scale for Indic Languages create mode 100644 data/2022/neurips/Multimodal Contrastive Learning with LIMoE: the Language-Image Mixture of Experts create mode 100644 data/2022/neurips/Multitasking Models are Robust to Structural Failure: A Neural Model for Bilingual Cognitive Reserve create mode 100644 data/2022/neurips/Multivariate Time-Series Forecasting with Temporal Polynomial Graph Neural Networks create mode 100644 data/2022/neurips/Multiview Human Body Reconstruction from Uncalibrated Cameras create mode 100644 data/2022/neurips/Museformer: Transformer with Fine- and Coarse-Grained Attention for Music Generation create mode 100644 data/2022/neurips/Mutual Information Divergence: A Unified Metric for Multimodal Generative Models create mode 100644 data/2022/neurips/Myriad: a real-world testbed to bridge trajectory optimization and deep learning create mode 100644 "data/2022/neurips/M\302\263ViT: Mixture-of-Experts Vision Transformer for Efficient Multi-task Learning with Model-Accelerator Co-design" create mode 100644 data/2022/neurips/NAS-Bench-360: Benchmarking Neural Architecture Search on Diverse Tasks create mode 100644 data/2022/neurips/NAS-Bench-Graph: Benchmarking Graph Neural Architecture Search create mode 100644 data/2022/neurips/NAS-Bench-Suite-Zero: Accelerating Research on Zero Cost Proxies create mode 100644 data/2022/neurips/NCP: Neural Correspondence Prior for Effective Unsupervised Shape Matching create mode 100644 data/2022/neurips/NOMAD: Nonlinear Manifold Decoders for Operator Learning create mode 100644 data/2022/neurips/NOTE: Robust Continual Test-time Adaptation Against Temporal Correlation create mode 100644 data/2022/neurips/NS3: Neuro-symbolic Semantic Code Search create mode 100644 data/2022/neurips/NSNet: A General Neural Probabilistic Framework for Satisfiability Problems create mode 100644 data/2022/neurips/NUWA-Infinity: Autoregressive over Autoregressive Generation for Infinite Visual Synthesis create mode 100644 data/2022/neurips/Natural Color Fool: Towards Boosting Black-box Unrestricted Attacks create mode 100644 data/2022/neurips/Natural gradient enables fast sampling in spiking neural networks create mode 100644 data/2022/neurips/Natural image synthesis for the retina with variational information bottleneck representation create mode 100644 data/2022/neurips/NaturalProver: Grounded Mathematical Proof Generation with Language Models create mode 100644 data/2022/neurips/Navigating Memory Construction by Global Pseudo-Task Simulation for Continual Learning create mode 100644 data/2022/neurips/NeMF: Neural Motion Fields for Kinematic Animation create mode 100644 data/2022/neurips/Near Instance-Optimal PAC Reinforcement Learning for Deterministic MDPs create mode 100644 data/2022/neurips/Near-Isometric Properties of Kronecker-Structured Random Tensor Embeddings create mode 100644 data/2022/neurips/Near-Optimal Collaborative Learning in Bandits create mode 100644 data/2022/neurips/Near-Optimal Correlation Clustering with Privacy create mode 100644 data/2022/neurips/Near-Optimal Goal-Oriented Reinforcement Learning in Non-Stationary Environments create mode 100644 data/2022/neurips/Near-Optimal Multi-Agent Learning for Safe Coverage Control create mode 100644 data/2022/neurips/Near-Optimal No-Regret Learning Dynamics for General Convex Games create mode 100644 data/2022/neurips/Near-Optimal Private and Scalable $k$-Clustering create mode 100644 data/2022/neurips/Near-Optimal Randomized Exploration for Tabular Markov Decision Processes create mode 100644 data/2022/neurips/Near-Optimal Regret Bounds for Multi-batch Reinforcement Learning create mode 100644 data/2022/neurips/Near-Optimal Regret for Adversarial MDP with Delayed Bandit Feedback create mode 100644 data/2022/neurips/Near-Optimal Sample Complexity Bounds for Constrained MDPs create mode 100644 data/2022/neurips/Nearly Optimal Algorithms for Linear Contextual Bandits with Adversarial Corruptions create mode 100644 data/2022/neurips/Nearly Optimal Best-of-Both-Worlds Algorithms for Online Learning with Feedback Graphs create mode 100644 data/2022/neurips/Nearly-Tight Bounds for Testing Histogram Distributions create mode 100644 data/2022/neurips/NeoRL: A Near Real-World Benchmark for Offline Reinforcement Learning create mode 100644 data/2022/neurips/Nest Your Adaptive Algorithm for Parameter-Agnostic Nonconvex Minimax Optimization create mode 100644 data/2022/neurips/Network change point localisation under local differential privacy create mode 100644 data/2022/neurips/NeuForm: Adaptive Overfitting for Neural Shape Editing create mode 100644 data/2022/neurips/NeuPhysics: Editable Neural Geometry and Physics from Monocular Videos create mode 100644 data/2022/neurips/Neur2SP: Neural Two-Stage Stochastic Programming create mode 100644 data/2022/neurips/NeurOLight: A Physics-Agnostic Neural Operator Enabling Parametric Photonic Device Simulation create mode 100644 data/2022/neurips/Neural Abstractions create mode 100644 data/2022/neurips/Neural Approximation of Graph Topological Features create mode 100644 data/2022/neurips/Neural Attentive Circuits create mode 100644 data/2022/neurips/Neural Basis Models for Interpretability create mode 100644 data/2022/neurips/Neural Circuit Architectural Priors for Embodied Control create mode 100644 data/2022/neurips/Neural Collapse with Normalized Features: A Geometric Analysis over the Riemannian Manifold create mode 100644 data/2022/neurips/Neural Conservation Laws: A Divergence-Free Perspective create mode 100644 data/2022/neurips/Neural Differential Equations for Learning to Program Neural Nets Through Continuous Learning Rules create mode 100644 data/2022/neurips/Neural Estimation of Submodular Functions with Applications to Differentiable Subset Selection create mode 100644 data/2022/neurips/Neural Lyapunov Control of Unknown Nonlinear Systems with Stability Guarantees create mode 100644 data/2022/neurips/Neural Matching Fields: Implicit Representation of Matching Fields for Visual Correspondence create mode 100644 data/2022/neurips/Neural Network Architecture Beyond Width and Depth create mode 100644 data/2022/neurips/Neural Payoff Machines: Predicting Fair and Stable Payoff Allocations Among Team Members create mode 100644 data/2022/neurips/Neural Set Function Extensions: Learning with Discrete Functions in High Dimensions create mode 100644 data/2022/neurips/Neural Shape Deformation Priors create mode 100644 data/2022/neurips/Neural Sheaf Diffusion: A Topological Perspective on Heterophily and Oversmoothing in GNNs create mode 100644 data/2022/neurips/Neural Stochastic Control create mode 100644 data/2022/neurips/Neural Stochastic PDEs: Resolution-Invariant Learning of Continuous Spatiotemporal Dynamics create mode 100644 data/2022/neurips/Neural Surface Reconstruction of Dynamic Scenes with Monocular RGB-D Camera create mode 100644 data/2022/neurips/Neural Temporal Walks: Motif-Aware Representation Learning on Continuous-Time Dynamic Graphs create mode 100644 data/2022/neurips/Neural Topological Ordering for Computation Graphs create mode 100644 data/2022/neurips/Neural Transmitted Radiance Fields create mode 100644 data/2022/neurips/Neural-Symbolic Entangled Framework for Complex Query Answering create mode 100644 data/2022/neurips/NeuroSchedule: A Novel Effective GNN-based Scheduling Method for High-level Synthesis create mode 100644 data/2022/neurips/Neuron with Steady Response Leads to Better Generalization create mode 100644 data/2022/neurips/Neurosymbolic Deep Generative Models for Sequence Data with Relational Constraints create mode 100644 data/2022/neurips/New Definitions and Evaluations for Saliency Methods: Staying Intrinsic, Complete and Sound create mode 100644 data/2022/neurips/New Lower Bounds for Private Estimation and a Generalized Fingerprinting Lemma create mode 100644 data/2022/neurips/No Free Lunch from Deep Learning in Neuroscience: A Case Study through Models of the Entorhinal-Hippocampal Circuit create mode 100644 data/2022/neurips/No-regret learning in games with noisy feedback: Faster rates and adaptivity via learning rate separation create mode 100644 data/2022/neurips/Nocturne: a scalable driving benchmark for bringing multi-agent learning one step closer to the real world create mode 100644 data/2022/neurips/NodeFormer: A Scalable Graph Structure Learning Transformer for Node Classification create mode 100644 data/2022/neurips/Noise Attention Learning: Enhancing Noise Robustness by Gradient Scaling create mode 100644 data/2022/neurips/Non-Convex Bilevel Games with Critical Point Selection Maps create mode 100644 data/2022/neurips/Non-Gaussian Tensor Programs create mode 100644 data/2022/neurips/Non-Linear Coordination Graphs create mode 100644 data/2022/neurips/Non-Linguistic Supervision for Contrastive Learning of Sentence Embeddings create mode 100644 data/2022/neurips/Non-Markovian Reward Modelling from Trajectory Labels via Interpretable Multiple Instance Learning create mode 100644 data/2022/neurips/Non-Monotonic Latent Alignments for CTC-Based Non-Autoregressive Machine Translation create mode 100644 data/2022/neurips/Non-Stationary Bandits under Recharging Payoffs: Improved Planning with Sublinear Regret create mode 100644 data/2022/neurips/Non-convex online learning via algorithmic equivalence create mode 100644 data/2022/neurips/Non-deep Networks create mode 100644 data/2022/neurips/Non-identifiability and the Blessings of Misspecification in Models of Molecular Fitness create mode 100644 data/2022/neurips/Non-monotonic Resource Utilization in the Bandits with Knapsacks Problem create mode 100644 data/2022/neurips/Non-rigid Point Cloud Registration with Neural Deformation Pyramid create mode 100644 data/2022/neurips/Non-stationary Bandits with Knapsacks create mode 100644 data/2022/neurips/Non-stationary Transformers: Exploring the Stationarity in Time Series Forecasting create mode 100644 data/2022/neurips/Nonlinear MCMC for Bayesian Machine Learning create mode 100644 data/2022/neurips/Nonlinear Sufficient Dimension Reduction with a Stochastic Neural Network create mode 100644 data/2022/neurips/Nonnegative Tensor Completion via Integer Optimization create mode 100644 data/2022/neurips/Nonparametric Uncertainty Quantification for Single Deterministic Neural Network create mode 100644 data/2022/neurips/Nonstationary Dual Averaging and Online Fair Allocation create mode 100644 data/2022/neurips/Normalizing Flows for Knockoff-free Controlled Feature Selection create mode 100644 data/2022/neurips/Not All Bits have Equal Value: Heterogeneous Precisions via Trainable Noise create mode 100644 data/2022/neurips/Not too little, not too much: a theoretical analysis of graph (over)smoothing create mode 100644 data/2022/neurips/OGC: Unsupervised 3D Object Segmentation from Rigid Dynamics of Point Clouds create mode 100644 data/2022/neurips/OLIVES Dataset: Ophthalmic Labels for Investigating Visual Eye Semantics create mode 100644 data/2022/neurips/OOD Link Prediction Generalization Capabilities of Message-Passing GNNs in Larger Test Graphs create mode 100644 data/2022/neurips/OPEN: Orthogonal Propagation with Ego-Network Modeling create mode 100644 data/2022/neurips/ORIENT: Submodular Mutual Information Measures for Data Subset Selection under Distribution Shift create mode 100644 data/2022/neurips/OST: Improving Generalization of DeepFake Detection via One-Shot Test-Time Training create mode 100644 data/2022/neurips/OTKGE: Multi-modal Knowledge Graph Embeddings via Optimal Transport create mode 100644 data/2022/neurips/Obj2Seq: Formatting Objects as Sequences with Class Prompt for Visual Tasks create mode 100644 data/2022/neurips/Object Representations as Fixed Points: Training Iterative Refinement Algorithms with Implicit Differentiation create mode 100644 data/2022/neurips/Object Scene Representation Transformer create mode 100644 data/2022/neurips/Object-Category Aware Reinforcement Learning create mode 100644 data/2022/neurips/OccGen: Selection of Real-world Multilingual Parallel Data Balanced in Gender within Occupations create mode 100644 data/2022/neurips/Off-Policy Evaluation for Action-Dependent Non-stationary Environments create mode 100644 data/2022/neurips/Off-Policy Evaluation for Episodic Partially Observable Markov Decision Processes under Non-Parametric Models create mode 100644 data/2022/neurips/Off-Policy Evaluation with Deficient Support Using Side Information create mode 100644 data/2022/neurips/Off-Policy Evaluation with Policy-Dependent Optimization Response create mode 100644 data/2022/neurips/Off-Team Learning create mode 100644 data/2022/neurips/Offline Goal-Conditioned Reinforcement Learning via $f$-Advantage Regression create mode 100644 data/2022/neurips/Offline Multi-Agent Reinforcement Learning with Knowledge Distillation create mode 100644 data/2022/neurips/Okapi: Generalising Better by Making Statistical Matches Match create mode 100644 data/2022/neurips/Old can be Gold: Better Gradient Flow can Make Vanilla-GCNs Great Again create mode 100644 data/2022/neurips/OmniVL: One Foundation Model for Image-Language and Video-Language Tasks create mode 100644 data/2022/neurips/On A Mallows-type Model For (Ranked) Choices create mode 100644 data/2022/neurips/On Analyzing Generative and Denoising Capabilities of Diffusion-based Deep Generative Models create mode 100644 data/2022/neurips/On Batch Teaching with Sample Complexity Bounded by VCD create mode 100644 data/2022/neurips/On Computing Probabilistic Explanations for Decision Trees create mode 100644 data/2022/neurips/On Convergence of FedProx: Local Dissimilarity Invariant Bounds, Non-smoothness and Beyond create mode 100644 data/2022/neurips/On Deep Generative Models for Approximation and Estimation of Distributions on Manifolds create mode 100644 data/2022/neurips/On Divergence Measures for Bayesian Pseudocoresets create mode 100644 data/2022/neurips/On Efficient Online Imitation Learning via Classification create mode 100644 data/2022/neurips/On Elimination Strategies for Bandit Fixed-Confidence Identification create mode 100644 data/2022/neurips/On Embeddings for Numerical Features in Tabular Deep Learning create mode 100644 data/2022/neurips/On Enforcing Better Conditioned Meta-Learning for Rapid Few-Shot Adaptation create mode 100644 data/2022/neurips/On Feature Learning in the Presence of Spurious Correlations create mode 100644 data/2022/neurips/On Gap-dependent Bounds for Offline Reinforcement Learning create mode 100644 data/2022/neurips/On Image Segmentation With Noisy Labels: Characterization and Volume Properties of the Optimal Solutions to Accuracy and Dice create mode 100644 data/2022/neurips/On Infinite Separations Between Simple and Optimal Mechanisms create mode 100644 data/2022/neurips/On Kernelized Multi-Armed Bandits with Constraints create mode 100644 data/2022/neurips/On Learning Fairness and Accuracy on Multiple Subgroups create mode 100644 data/2022/neurips/On Learning and Refutation in Noninteractive Local Differential Privacy create mode 100644 data/2022/neurips/On Leave-One-Out Conditional Mutual Information For Generalization create mode 100644 data/2022/neurips/On Margin Maximization in Linear and ReLU Networks create mode 100644 data/2022/neurips/On Margins and Generalisation for Voting Classifiers create mode 100644 data/2022/neurips/On Measuring Excess Capacity in Neural Networks create mode 100644 data/2022/neurips/On Non-Linear operators for Geometric Deep Learning create mode 100644 data/2022/neurips/On Optimal Learning Under Targeted Data Poisoning create mode 100644 data/2022/neurips/On Privacy and Personalization in Cross-Silo Federated Learning create mode 100644 data/2022/neurips/On Reinforcement Learning and Distribution Matching for Fine-Tuning Language Models with no Catastrophic Forgetting create mode 100644 data/2022/neurips/On Robust Multiclass Learnability create mode 100644 data/2022/neurips/On Sample Optimality in Personalized Collaborative and Federated Learning create mode 100644 data/2022/neurips/On Scalable Testing of Samplers create mode 100644 data/2022/neurips/On Scrambling Phenomena for Randomly Initialized Recurrent Networks create mode 100644 data/2022/neurips/On Translation and Reconstruction Guarantees of the Cycle-Consistent Generative Adversarial Networks create mode 100644 data/2022/neurips/On Uncertainty, Tempering, and Data Augmentation in Bayesian Classification create mode 100644 data/2022/neurips/On global convergence of ResNets: From finite to infinite width using linear parameterization create mode 100644 data/2022/neurips/On the Adversarial Robustness of Mixture of Experts create mode 100644 data/2022/neurips/On the Complexity of Adversarial Decision Making create mode 100644 data/2022/neurips/On the Convergence Theory for Hessian-Free Bilevel Algorithms create mode 100644 data/2022/neurips/On the Convergence of Stochastic Multi-Objective Gradient Manipulation and Beyond create mode 100644 data/2022/neurips/On the Discrimination Risk of Mean Aggregation Feature Imputation in Graphs create mode 100644 data/2022/neurips/On the Double Descent of Random Features Models Trained with SGD create mode 100644 data/2022/neurips/On the Effect of Pre-training for Transformer in Different Modality on Offline Reinforcement Learning create mode 100644 data/2022/neurips/On the Effective Number of Linear Regions in Shallow Univariate ReLU Networks: Convergence Guarantees and Implicit Bias create mode 100644 data/2022/neurips/On the Effectiveness of Fine-tuning Versus Meta-reinforcement Learning create mode 100644 data/2022/neurips/On the Effectiveness of Lipschitz-Driven Rehearsal in Continual Learning create mode 100644 data/2022/neurips/On the Effectiveness of Persistent Homology create mode 100644 data/2022/neurips/On the Efficient Implementation of High Accuracy Optimality of Profile Maximum Likelihood create mode 100644 data/2022/neurips/On the Epistemic Limits of Personalized Prediction create mode 100644 data/2022/neurips/On the Frequency-bias of Coordinate-MLPs create mode 100644 data/2022/neurips/On the Generalizability and Predictability of Recommender Systems create mode 100644 data/2022/neurips/On the Generalization Power of the Overfitted Three-Layer Neural Tangent Kernel Model create mode 100644 data/2022/neurips/On the Global Convergence Rates of Decentralized Softmax Gradient Play in Markov Potential Games create mode 100644 data/2022/neurips/On the Identifiability of Nonlinear ICA: Sparsity and Beyond create mode 100644 data/2022/neurips/On the Importance of Gradient Norm in PAC-Bayesian Bounds create mode 100644 data/2022/neurips/On the Interpretability of Regularisation for Neural Networks Through Model Gradient Similarity create mode 100644 data/2022/neurips/On the Learning Mechanisms in Physical Reasoning create mode 100644 data/2022/neurips/On the Limitations of Stochastic Pre-processing Defenses create mode 100644 data/2022/neurips/On the Parameterization and Initialization of Diagonal State Space Models create mode 100644 data/2022/neurips/On the Representation Collapse of Sparse Mixture of Experts create mode 100644 data/2022/neurips/On the Robustness of Deep Clustering Models: Adversarial Attacks and Defenses create mode 100644 data/2022/neurips/On the Robustness of Graph Neural Diffusion to Topology Perturbations create mode 100644 data/2022/neurips/On the SDEs and Scaling Rules for Adaptive Gradient Algorithms create mode 100644 data/2022/neurips/On the Safety of Interpretable Machine Learning: A Maximum Deviation Approach create mode 100644 data/2022/neurips/On the Sample Complexity of Stabilizing LTI Systems on a Single Trajectory create mode 100644 data/2022/neurips/On the Spectral Bias of Convolutional Neural Tangent and Gaussian Process Kernels create mode 100644 data/2022/neurips/On the Stability and Scalability of Node Perturbation Learning create mode 100644 data/2022/neurips/On the Statistical Efficiency of Reward-Free Exploration in Non-Linear RL create mode 100644 data/2022/neurips/On the Strong Correlation Between Model Invariance and Generalization create mode 100644 data/2022/neurips/On the Symmetries of Deep Learning Models and their Internal Representations create mode 100644 data/2022/neurips/On the Theoretical Properties of Noise Correlation in Stochastic Optimization create mode 100644 data/2022/neurips/On the Tradeoff Between Robustness and Fairness create mode 100644 data/2022/neurips/On the consistent estimation of optimal Receiver Operating Characteristic (ROC) curve create mode 100644 data/2022/neurips/On the convergence of policy gradient methods to Nash equilibria in general stochastic games create mode 100644 data/2022/neurips/On the detrimental effect of invariances in the likelihood for variational inference create mode 100644 data/2022/neurips/On the difficulty of learning chaotic dynamics with RNNs create mode 100644 data/2022/neurips/On the generalization of learning algorithms that do not converge create mode 100644 data/2022/neurips/On the inability of Gaussian process regression to optimally learn compositional functions create mode 100644 data/2022/neurips/On the non-universality of deep learning: quantifying the cost of symmetry create mode 100644 data/2022/neurips/On the relationship between variational inference and auto-associative memory create mode 100644 data/2022/neurips/On the role of overparameterization in off-policy Temporal Difference learning with linear function approximation create mode 100644 data/2022/neurips/On the symmetries of the synchronization problem in Cryo-EM: Multi-Frequency Vector Diffusion Maps on the Projective Plane create mode 100644 data/2022/neurips/On-Demand Sampling: Learning Optimally from Multiple Distributions create mode 100644 data/2022/neurips/On-Device Training Under 256KB Memory create mode 100644 data/2022/neurips/One Model to Edit Them All: Free-Form Text-Driven Image Manipulation with Semantic Modulations create mode 100644 data/2022/neurips/One Positive Label is Sufficient: Single-Positive Multi-Label Learning with Label Enhancement create mode 100644 data/2022/neurips/One for All: Simultaneous Metric and Preference Learning over Multiple Users create mode 100644 data/2022/neurips/One-Inlier is First: Towards Efficient Position Encoding for Point Cloud Registration create mode 100644 data/2022/neurips/One-shot Neural Backdoor Erasing via Adversarial Weight Masking create mode 100644 data/2022/neurips/OnePose++: Keypoint-Free One-Shot Object Pose Estimation without CAD Models create mode 100644 data/2022/neurips/Online Agnostic Multiclass Boosting create mode 100644 data/2022/neurips/Online Algorithms for the Santa Claus Problem create mode 100644 data/2022/neurips/Online Allocation and Learning in the Presence of Strategic Agents create mode 100644 data/2022/neurips/Online Bipartite Matching with Advice: Tight Robustness-Consistency Tradeoffs for the Two-Stage Model create mode 100644 data/2022/neurips/Online Convex Optimization with Hard Constraints: Towards the Best of Two Worlds and Beyond create mode 100644 data/2022/neurips/Online Decision Mediation create mode 100644 data/2022/neurips/Online Deep Equilibrium Learning for Regularization by Denoising create mode 100644 data/2022/neurips/Online Frank-Wolfe with Arbitrary Delays create mode 100644 data/2022/neurips/Online Learning and Pricing for Network Revenue Management with Reusable Resources create mode 100644 data/2022/neurips/Online Minimax Multiobjective Optimization: Multicalibeating and Other Applications create mode 100644 data/2022/neurips/Online Neural Sequence Detection with Hierarchical Dirichlet Point Process create mode 100644 data/2022/neurips/Online PAC-Bayes Learning create mode 100644 data/2022/neurips/Online Reinforcement Learning for Mixed Policy Scopes create mode 100644 data/2022/neurips/Online Training Through Time for Spiking Neural Networks create mode 100644 data/2022/neurips/Ontologue: Declarative Benchmark Construction for Ontological Multi-Label Classification create mode 100644 data/2022/neurips/Open High-Resolution Satellite Imagery: The WorldStrat Dataset - With Application to Super-Resolution create mode 100644 data/2022/neurips/Open-Ended Reinforcement Learning with Neural Reward Functions create mode 100644 data/2022/neurips/OpenAUC: Towards AUC-Oriented Open-Set Recognition create mode 100644 data/2022/neurips/OpenFWI: Large-scale Multi-structural Benchmark Datasets for Full Waveform Inversion create mode 100644 data/2022/neurips/OpenFilter: A Framework to Democratize Research Access to Social Media AR Filters create mode 100644 data/2022/neurips/OpenOOD: Benchmarking Generalized Out-of-Distribution Detection create mode 100644 data/2022/neurips/OpenSRH: optimizing brain tumor surgery using intraoperative stimulated Raman histology create mode 100644 data/2022/neurips/OpenXAI: Towards a Transparent Evaluation of Model Explanations create mode 100644 data/2022/neurips/Operative dimensions in unconstrained connectivity of recurrent neural networks create mode 100644 data/2022/neurips/Operator Splitting Value Iteration create mode 100644 data/2022/neurips/Optimal Algorithms for Decentralized Stochastic Variational Inequalities create mode 100644 data/2022/neurips/Optimal Binary Classification Beyond Accuracy create mode 100644 data/2022/neurips/Optimal Brain Compression: A Framework for Accurate Post-Training Quantization and Pruning create mode 100644 data/2022/neurips/Optimal Comparator Adaptive Online Learning with Switching Cost create mode 100644 data/2022/neurips/Optimal Dynamic Regret in LQR Control create mode 100644 data/2022/neurips/Optimal Efficiency-Envy Trade-Off via Optimal Transport create mode 100644 data/2022/neurips/Optimal Gradient Sliding and its Application to Optimal Distributed Optimization Under Similarity create mode 100644 data/2022/neurips/Optimal Positive Generation via Latent Transformation for Contrastive Learning create mode 100644 data/2022/neurips/Optimal Query Complexities for Dynamic Trace Estimation create mode 100644 data/2022/neurips/Optimal Rates for Regularized Conditional Mean Embedding Learning create mode 100644 data/2022/neurips/Optimal Scaling for Locally Balanced Proposals in Discrete Spaces create mode 100644 data/2022/neurips/Optimal Transport of Classifiers to Fairness create mode 100644 data/2022/neurips/Optimal Transport-based Identity Matching for Identity-invariant Facial Expression Recognition create mode 100644 data/2022/neurips/Optimal Weak to Strong Learning create mode 100644 data/2022/neurips/Optimal and Adaptive Monteiro-Svaiter Acceleration create mode 100644 data/2022/neurips/Optimal-er Auctions through Attention create mode 100644 data/2022/neurips/Optimistic Mirror Descent Either Converges to Nash or to Strong Coarse Correlated Equilibria in Bimatrix Games create mode 100644 data/2022/neurips/Optimistic Posterior Sampling for Reinforcement Learning with Few Samples and Tight Guarantees create mode 100644 data/2022/neurips/Optimistic Tree Searches for Combinatorial Black-Box Optimization create mode 100644 data/2022/neurips/Optimizing Data Collection for Machine Learning create mode 100644 data/2022/neurips/Optimizing Relevance Maps of Vision Transformers Improves Robustness create mode 100644 data/2022/neurips/Oracle Inequalities for Model Selection in Offline Reinforcement Learning create mode 100644 data/2022/neurips/Oracle-Efficient Online Learning for Smoothed Adversaries create mode 100644 data/2022/neurips/Order-Invariant Cardinality Estimators Are Differentially Private create mode 100644 data/2022/neurips/Ordered Subgraph Aggregation Networks create mode 100644 data/2022/neurips/OrdinalCLIP: Learning Rank Prompts for Language-Guided Ordinal Regression create mode 100644 data/2022/neurips/Orthogonal Transformer: An Efficient Vision Transformer Backbone with Token Orthogonalization create mode 100644 data/2022/neurips/Oscillatory Tracking of Continuous Attractor Neural Networks Account for Phase Precession and Procession of Hippocampal Place Cells create mode 100644 data/2022/neurips/Out-of-Distribution Detection via Conditional Kernel Independence Model create mode 100644 data/2022/neurips/Out-of-Distribution Detection with An Adaptive Likelihood Ratio on Informative Hierarchical VAE create mode 100644 data/2022/neurips/Outlier Suppression: Pushing the Limit of Low-bit Transformer Language Models create mode 100644 data/2022/neurips/Outlier-Robust Sparse Estimation via Non-Convex Optimization create mode 100644 data/2022/neurips/Outlier-Robust Sparse Mean Estimation for Heavy-Tailed Distributions create mode 100644 data/2022/neurips/Outsourcing Training without Uploading Data via Efficient Collaborative Open-Source Sampling create mode 100644 data/2022/neurips/Overparameterization from Computational Constraints create mode 100644 data/2022/neurips/P2P: Tuning Pre-trained Image Models for Point Cloud Analysis with Point-to-Pixel Prompting create mode 100644 data/2022/neurips/PAC-Bayes Compression Bounds So Tight That They Can Explain Generalization create mode 100644 data/2022/neurips/PAC: Assisted Value Factorization with Counterfactual Predictions in Multi-Agent Reinforcement Learning create mode 100644 data/2022/neurips/PALBERT: Teaching ALBERT to Ponder create mode 100644 data/2022/neurips/PALMER: Perception - Action Loop with Memory for Long-Horizon Planning create mode 100644 data/2022/neurips/PDEBench: An Extensive Benchmark for Scientific Machine Learning create mode 100644 data/2022/neurips/PDSketch: Integrated Domain Programming, Learning, and Planning create mode 100644 data/2022/neurips/PEER: A Comprehensive and Multi-Task Benchmark for Protein Sequence Understanding create mode 100644 data/2022/neurips/PKD: General Distillation Framework for Object Detectors via Pearson Correlation Coefficient create mode 100644 data/2022/neurips/PROSPECT: Labeled Tandem Mass Spectrometry Dataset for Machine Learning in Proteomics create mode 100644 data/2022/neurips/PaCo: Parameter-Compositional Multi-task Reinforcement Learning create mode 100644 data/2022/neurips/Palm up: Playing in the Latent Manifold for Unsupervised Pretraining create mode 100644 data/2022/neurips/Panchromatic and Multispectral Image Fusion via Alternating Reverse Filtering Network create mode 100644 data/2022/neurips/Parallel Tempering With a Variational Reference create mode 100644 data/2022/neurips/Parameter tuning and model selection in Optimal Transport with semi-dual Brenier formulation create mode 100644 data/2022/neurips/Parameter-Efficient Masking Networks create mode 100644 data/2022/neurips/Parameter-free Dynamic Graph Embedding for Link Prediction create mode 100644 data/2022/neurips/Parameter-free Regret in High Probability with Heavy Tails create mode 100644 data/2022/neurips/Parameters or Privacy: A Provable Tradeoff Between Overparameterization and Membership Inference create mode 100644 data/2022/neurips/Parametrically Retargetable Decision-Makers Tend To Seek Power create mode 100644 data/2022/neurips/Paraphrasing Is All You Need for Novel Object Captioning create mode 100644 data/2022/neurips/Pareto Set Learning for Expensive Multi-Objective Optimization create mode 100644 data/2022/neurips/Partial Identification of Treatment Effects with Implicit Generative Models create mode 100644 data/2022/neurips/PatchComplete: Learning Multi-Resolution Patch Priors for 3D Shape Completion on Unseen Categories create mode 100644 data/2022/neurips/Patching open-vocabulary models by interpolating weights create mode 100644 data/2022/neurips/Path Independent Equilibrium Models Can Better Exploit Test-Time Computation create mode 100644 data/2022/neurips/Pay attention to your loss : understanding misconceptions about Lipschitz neural networks create mode 100644 data/2022/neurips/PeRFception: Perception using Radiance Fields create mode 100644 data/2022/neurips/Peer Prediction for Learning Agents create mode 100644 data/2022/neurips/Perceptual Attacks of No-Reference Image Quality Models with Human-in-the-Loop create mode 100644 data/2022/neurips/Perfect Sampling from Pairwise Comparisons create mode 100644 data/2022/neurips/PerfectDou: Dominating DouDizhu with Perfect Information Distillation create mode 100644 data/2022/neurips/Performative Power create mode 100644 data/2022/neurips/Periodic Graph Transformers for Crystal Material Property Prediction create mode 100644 data/2022/neurips/Peripheral Vision Transformer create mode 100644 data/2022/neurips/Personalized Federated Learning towards Communication Efficiency, Robustness and Fairness create mode 100644 data/2022/neurips/Personalized Online Federated Learning with Multiple Kernels create mode 100644 data/2022/neurips/Perturbation Learning Based Anomaly Detection create mode 100644 data/2022/neurips/Phase Transition from Clean Training to Adversarial Training create mode 100644 data/2022/neurips/Phase diagram of Stochastic Gradient Descent in high-dimensional two-layer neural networks create mode 100644 data/2022/neurips/Phase transitions in when feedback is useful create mode 100644 data/2022/neurips/Photorealistic Text-to-Image Diffusion Models with Deep Language Understanding create mode 100644 data/2022/neurips/PhysGNN: A Physics-Driven Graph Neural Network Based Model for Predicting Soft Tissue Deformation in Image-Guided Neurosurgery create mode 100644 data/2022/neurips/Physically-Based Face Rendering for NIR-VIS Face Recognition create mode 100644 data/2022/neurips/Physics-Embedded Neural Networks: Graph Neural PDE Solvers with Mixed Boundary Conditions create mode 100644 data/2022/neurips/Physics-Informed Implicit Representations of Equilibrium Network Flows create mode 100644 data/2022/neurips/Picking on the Same Person: Does Algorithmic Monoculture lead to Outcome Homogenization? create mode 100644 data/2022/neurips/Pile of Law: Learning Responsible Data Filtering from the Law and a 256GB Open-Source Legal Dataset create mode 100644 data/2022/neurips/Pitfalls of Epistemic Uncertainty Quantification through Loss Minimisation create mode 100644 data/2022/neurips/Plan To Predict: Learning an Uncertainty-Foreseeing Model For Model-Based Reinforcement Learning create mode 100644 data/2022/neurips/Planning for Sample Efficient Imitation Learning create mode 100644 data/2022/neurips/Planning to the Information Horizon of BAMDPs via Epistemic State Abstraction create mode 100644 data/2022/neurips/PlasticityNet: Learning to Simulate Metal, Sand, and Snow for Optimization Time Integration create mode 100644 data/2022/neurips/Pluralistic Image Completion with Gaussian Mixture Models create mode 100644 data/2022/neurips/Point Transformer V2: Grouped Vector Attention and Partition-based Pooling create mode 100644 data/2022/neurips/Point-M2AE: Multi-scale Masked Autoencoders for Hierarchical Point Cloud Pre-training create mode 100644 data/2022/neurips/PointNeXt: Revisiting PointNet++ with Improved Training and Scaling Strategies create mode 100644 data/2022/neurips/PointTAD: Multi-Label Temporal Action Detection with Learnable Query Points create mode 100644 data/2022/neurips/Poisson Flow Generative Models create mode 100644 data/2022/neurips/PolarMix: A General Data Augmentation Technique for LiDAR Point Clouds create mode 100644 data/2022/neurips/Policy Gradient With Serial Markov Chain Reasoning create mode 100644 data/2022/neurips/Policy Optimization for Markov Games: Unified Framework and Faster Convergence create mode 100644 data/2022/neurips/Policy Optimization with Advantage Regularization for Long-Term Fairness in Decision Systems create mode 100644 data/2022/neurips/Policy Optimization with Linear Temporal Logic Constraints create mode 100644 data/2022/neurips/Polyhistor: Parameter-Efficient Multi-Task Adaptation for Dense Vision Tasks create mode 100644 data/2022/neurips/Polynomial Neural Fields for Subband Decomposition and Manipulation create mode 100644 data/2022/neurips/Polynomial time guarantees for the Burer-Monteiro method create mode 100644 data/2022/neurips/Polynomial-Time Optimal Equilibria with a Mediator in Extensive-Form Games create mode 100644 data/2022/neurips/PopArt: Efficient Sparse Regression and Experimental Design for Optimal Sparse Linear Bandits create mode 100644 data/2022/neurips/Positive-Unlabeled Learning using Random Forests via Recursive Greedy Risk Minimization create mode 100644 data/2022/neurips/Positively Weighted Kernel Quadrature via Subsampling create mode 100644 data/2022/neurips/Post-hoc estimators for learning to defer to an expert create mode 100644 data/2022/neurips/Posted Pricing and Dynamic Prior-independent Mechanisms with Value Maximizers create mode 100644 data/2022/neurips/Posterior Collapse of a Linear Latent Variable Model create mode 100644 data/2022/neurips/Posterior Matching for Arbitrary Conditioning create mode 100644 data/2022/neurips/Posterior Refinement Improves Sample Efficiency in Bayesian Neural Networks create mode 100644 data/2022/neurips/Posterior and Computational Uncertainty in Gaussian Processes create mode 100644 data/2022/neurips/Power and limitations of single-qubit native quantum neural networks create mode 100644 data/2022/neurips/Practical Adversarial Attacks on Spatiotemporal Traffic Forecasting Models create mode 100644 data/2022/neurips/Practical Adversarial Multivalid Conformal Prediction create mode 100644 data/2022/neurips/Pragmatically Learning from Pedagogical Demonstrations in Multi-Goal Environments create mode 100644 data/2022/neurips/Pre-Train Your Loss: Easy Bayesian Transfer Learning with Informative Priors create mode 100644 data/2022/neurips/Pre-Trained Image Encoder for Generalizable Visual Reinforcement Learning create mode 100644 data/2022/neurips/Pre-Trained Language Models for Interactive Decision-Making create mode 100644 data/2022/neurips/Pre-Trained Model Reusability Evaluation for Small-Data Transfer Learning create mode 100644 data/2022/neurips/Pre-activation Distributions Expose Backdoor Neurons create mode 100644 data/2022/neurips/Pre-trained Adversarial Perturbations create mode 100644 data/2022/neurips/Precise Learning Curves and Higher-Order Scalings for Dot-product Kernel Regression create mode 100644 data/2022/neurips/Precise Regret Bounds for Log-loss via a Truncated Bayesian Algorithm create mode 100644 data/2022/neurips/Predicting Cellular Responses to Novel Drug Perturbations at a Single-Cell Resolution create mode 100644 data/2022/neurips/Predicting Label Distribution from Multi-label Ranking create mode 100644 data/2022/neurips/Predictive Coding beyond Gaussian Distributions create mode 100644 data/2022/neurips/Predictive Querying for Autoregressive Neural Sequence Models create mode 100644 data/2022/neurips/Preservation of the Global Knowledge by Not-True Distillation in Federated Learning create mode 100644 data/2022/neurips/Privacy Induces Robustness: Information-Computation Gaps and Sparse Mean Estimation create mode 100644 data/2022/neurips/Privacy of Noisy Stochastic Gradient Descent: More Iterations without More Privacy Loss create mode 100644 data/2022/neurips/Private Estimation with Public Data create mode 100644 data/2022/neurips/Private Graph All-Pairwise-Shortest-Path Distance Release with Improved Error Rate create mode 100644 data/2022/neurips/Private Isotonic Regression create mode 100644 data/2022/neurips/Private Multiparty Perception for Navigation create mode 100644 data/2022/neurips/Private Set Generation with Discriminative Information create mode 100644 data/2022/neurips/Private Synthetic Data for Multitask Learning and Marginal Queries create mode 100644 data/2022/neurips/Private and Communication-Efficient Algorithms for Entropy Estimation create mode 100644 data/2022/neurips/Probabilistic Missing Value Imputation for Mixed Categorical and Ordered Data create mode 100644 data/2022/neurips/Probabilistic Transformer: Modelling Ambiguities and Distributions for RNA Folding and Molecule Design create mode 100644 data/2022/neurips/Probable Domain Generalization via Quantile Risk Minimization create mode 100644 data/2022/neurips/Probing Classifiers are Unreliable for Concept Removal and Detection create mode 100644 data/2022/neurips/Procedural Image Programs for Representation Learning create mode 100644 data/2022/neurips/Product Ranking for Revenue Maximization with Multiple Purchases create mode 100644 data/2022/neurips/Promising or Elusive? Unsupervised Object Segmentation from Real-world Single Images create mode 100644 data/2022/neurips/Prompt Certified Machine Unlearning with Randomized Gradient Smoothing and Quantization create mode 100644 data/2022/neurips/Proppo: a Message Passing Framework for Customizable and Composable Learning Algorithms create mode 100644 data/2022/neurips/ProtoVAE: A Trustworthy Self-Explainable Prototypical Variational Model create mode 100644 data/2022/neurips/ProtoX: Explaining a Reinforcement Learning Agent via Prototyping create mode 100644 data/2022/neurips/Prototypical VoteNet for Few-Shot 3D Point Cloud Object Detection create mode 100644 data/2022/neurips/Provable Benefit of Multitask Representation Learning in Reinforcement Learning create mode 100644 data/2022/neurips/Provable Defense against Backdoor Policies in Reinforcement Learning create mode 100644 data/2022/neurips/Provable General Function Class Representation Learning in Multitask Bandits and MDP create mode 100644 data/2022/neurips/Provable Generalization of Overparameterized Meta-learning Trained with SGD create mode 100644 data/2022/neurips/Provable Subspace Identification Under Post-Nonlinear Mixtures create mode 100644 data/2022/neurips/Provably Adversarially Robust Detection of Out-of-Distribution Data (Almost) for Free create mode 100644 data/2022/neurips/Provably Efficient Model-Free Constrained RL with Linear Function Approximation create mode 100644 data/2022/neurips/Provably Efficient Offline Multi-agent Reinforcement Learning via Strategy-wise Bonus create mode 100644 data/2022/neurips/Provably Efficient Reinforcement Learning in Partially Observable Dynamical Systems create mode 100644 data/2022/neurips/Provably Feedback-Efficient Reinforcement Learning via Active Reward Learning create mode 100644 data/2022/neurips/Provably expressive temporal graph networks create mode 100644 data/2022/neurips/Provably sample-efficient RL with side information about latent dynamics create mode 100644 data/2022/neurips/Provably tuning the ElasticNet across instances create mode 100644 data/2022/neurips/Proximal Learning With Opponent-Learning Awareness create mode 100644 data/2022/neurips/Proximal Point Imitation Learning create mode 100644 data/2022/neurips/Prune and distill: similar reformatting of image information along rat visual cortex and deep neural networks create mode 100644 data/2022/neurips/Pruning Neural Networks via Coresets and Convex Geometry: Towards No Assumptions create mode 100644 data/2022/neurips/Pruning has a disparate impact on model accuracy create mode 100644 data/2022/neurips/Pruning's Effect on Generalization Through the Lens of Training and Regularization create mode 100644 data/2022/neurips/Pseudo-Riemannian Graph Convolutional Networks create mode 100644 data/2022/neurips/Public Wisdom Matters! Discourse-Aware Hyperbolic Fourier Co-Attention for Social Text Classification create mode 100644 data/2022/neurips/PulseImpute: A Novel Benchmark Task for Pulsative Physiological Signal Imputation create mode 100644 data/2022/neurips/Pure Transformers are Powerful Graph Learners create mode 100644 data/2022/neurips/Pushing the limits of fairness impossibility: Who's the fairest of them all? create mode 100644 data/2022/neurips/Pyramid Attention For Source Code Summarization create mode 100644 data/2022/neurips/PyramidCLIP: Hierarchical Feature Alignment for Vision-language Model Pretraining create mode 100644 data/2022/neurips/Pythae: Unifying Generative Autoencoders in Python - A Benchmarking Use Case create mode 100644 data/2022/neurips/Q-ViT: Accurate and Fully Quantized Low-bit Vision Transformer create mode 100644 data/2022/neurips/QC-StyleGAN - Quality Controllable Image Generation and Manipulation create mode 100644 data/2022/neurips/QUARK: Controllable Text Generation with Reinforced Unlearning create mode 100644 data/2022/neurips/Quality Not Quantity: On the Interaction between Dataset Design and Robustness of CLIP create mode 100644 data/2022/neurips/Quantifying Statistical Significance of Neural Network-based Image Segmentation by Selective Inference create mode 100644 data/2022/neurips/Quantile Constrained Reinforcement Learning: A Reinforcement Learning Framework Constraining Outage Probability create mode 100644 data/2022/neurips/Quantized Training of Gradient Boosting Decision Trees create mode 100644 data/2022/neurips/Quantum Algorithms for Sampling Log-Concave Distributions and Estimating Normalizing Constants create mode 100644 data/2022/neurips/Quantum Speedups of Optimizing Approximately Convex Functions with Applications to Logarithmic Regret Stochastic Convex Bandits create mode 100644 data/2022/neurips/Quasi-Newton Methods for Saddle Point Problems create mode 100644 data/2022/neurips/QueryPose: Sparse Multi-Person Pose Regression via Spatial-Aware Part-Level Query create mode 100644 data/2022/neurips/Queue Up Your Regrets: Achieving the Dynamic Capacity Region of Multiplayer Bandits create mode 100644 data/2022/neurips/Quo Vadis: Is Trajectory Forecasting the Key Towards Long-Term Multi-Object Tracking? create mode 100644 data/2022/neurips/RAMBO-RL: Robust Adversarial Model-Based Offline Reinforcement Learning create mode 100644 data/2022/neurips/REVIVE: Regional Visual Representation Matters in Knowledge-Based Visual Question Answering create mode 100644 data/2022/neurips/RISE: Robust Individualized Decision Learning with Sensitive Variables create mode 100644 data/2022/neurips/RKHS-SHAP: Shapley Values for Kernel Methods create mode 100644 data/2022/neurips/RLIP: Relational Language-Image Pre-training for Human-Object Interaction Detection create mode 100644 data/2022/neurips/RNNs of RNNs: Recursive Construction of Stable Assemblies of Recurrent Neural Networks create mode 100644 data/2022/neurips/RORL: Robust Offline Reinforcement Learning via Conservative Smoothing create mode 100644 data/2022/neurips/RSA: Reducing Semantic Shift from Aggressive Augmentations for Self-supervised Learning create mode 100644 data/2022/neurips/RTFormer: Efficient Design for Real-Time Semantic Segmentation with Transformer create mode 100644 data/2022/neurips/RainNet: A Large-Scale Imagery Dataset and Benchmark for Spatial Precipitation Downscaling create mode 100644 data/2022/neurips/Random Normalization Aggregation for Adversarial Defense create mode 100644 data/2022/neurips/Random Rank: The One and Only Strategyproof and Proportionally Fair Randomized Facility Location Mechanism create mode 100644 data/2022/neurips/Random Sharpness-Aware Minimization create mode 100644 data/2022/neurips/Randomized Channel Shuffling: Minimal-Overhead Backdoor Attack Detection without Clean Datasets create mode 100644 data/2022/neurips/Randomized Message-Interception Smoothing: Gray-box Certificates for Graph Neural Networks create mode 100644 data/2022/neurips/Randomized Sketches for Clustering: Fast and Optimal Kernel $k$-Means create mode 100644 data/2022/neurips/Rank Diminishing in Deep Neural Networks create mode 100644 data/2022/neurips/RankFeat: Rank-1 Feature Removal for Out-of-distribution Detection create mode 100644 data/2022/neurips/Rapid Model Architecture Adaption for Meta-Learning create mode 100644 data/2022/neurips/Rapidly Mixing Multiple-try Metropolis Algorithms for Model Selection Problems create mode 100644 data/2022/neurips/Rare Gems: Finding Lottery Tickets at Initialization create mode 100644 data/2022/neurips/Rashomon Capacity: A Metric for Predictive Multiplicity in Classification create mode 100644 data/2022/neurips/Rate-Distortion Theoretic Bounds on Generalization Error for Distributed Learning create mode 100644 data/2022/neurips/Rate-Optimal Online Convex Optimization in Adaptive Linear Control create mode 100644 data/2022/neurips/Re-Analyze Gauss: Bounds for Private Matrix Approximation via Dyson Brownian Motion create mode 100644 data/2022/neurips/ReCo: Retrieve and Co-segment for Zero-shot Transfer create mode 100644 data/2022/neurips/ReFactor GNNs: Revisiting Factorisation-based Models from a Message-Passing Perspective create mode 100644 data/2022/neurips/Real-Valued Backpropagation is Unsuitable for Complex-Valued Neural Networks create mode 100644 data/2022/neurips/Recall Distortion in Neural Network Pruning and the Undecayed Pruning Algorithm create mode 100644 data/2022/neurips/Receding Horizon Inverse Reinforcement Learning create mode 100644 data/2022/neurips/Recipe for a General, Powerful, Scalable Graph Transformer create mode 100644 data/2022/neurips/Recommender Forest for Efficient Retrieval create mode 100644 data/2022/neurips/Reconstructing Training Data From Trained Neural Networks create mode 100644 data/2022/neurips/Reconstruction on Trees and Low-Degree Polynomials create mode 100644 data/2022/neurips/Recovering Private Text in Federated Learning of Language Models create mode 100644 data/2022/neurips/Recruitment Strategies That Take a Chance create mode 100644 data/2022/neurips/Recurrent Convolutional Neural Networks Learn Succinct Learning Algorithms create mode 100644 data/2022/neurips/Recurrent Memory Transformer create mode 100644 data/2022/neurips/Recurrent Video Restoration Transformer with Guided Deformable Attention create mode 100644 data/2022/neurips/Recursive Reasoning in Minimax Games: A Level $k$ Gradient Play Method create mode 100644 data/2022/neurips/Recursive Reinforcement Learning create mode 100644 data/2022/neurips/RecursiveMix: Mixed Learning with History create mode 100644 data/2022/neurips/Redeeming intrinsic rewards via constrained optimization create mode 100644 data/2022/neurips/Redistribution of Weights and Activations for AdderNet Quantization create mode 100644 data/2022/neurips/Reduced Representation of Deformation Fields for Effective Non-rigid Shape Matching create mode 100644 data/2022/neurips/Reduction Algorithms for Persistence Diagrams of Networks: CoralTDA and PrunIT create mode 100644 data/2022/neurips/Redundancy-Free Message Passing for Graph Neural Networks create mode 100644 data/2022/neurips/Redundant representations help generalization in wide neural networks create mode 100644 data/2022/neurips/Refining Low-Resource Unsupervised Translation by Language Disentanglement of Multilingual Translation Model create mode 100644 data/2022/neurips/Regret Bounds for Information-Directed Reinforcement Learning create mode 100644 data/2022/neurips/Regret Bounds for Multilabel Classification in Sparse Label Regimes create mode 100644 data/2022/neurips/Regret Bounds for Risk-Sensitive Reinforcement Learning create mode 100644 data/2022/neurips/Regularized Gradient Descent Ascent for Two-Player Zero-Sum Markov Games create mode 100644 data/2022/neurips/Regularized Molecular Conformation Fields create mode 100644 data/2022/neurips/Reincarnating Reinforcement Learning: Reusing Prior Computation to Accelerate Progress create mode 100644 data/2022/neurips/Reinforced Genetic Algorithm for Structure-based Drug Design create mode 100644 data/2022/neurips/Reinforcement Learning in a Birth and Death Process: Breaking the Dependence on the State Space create mode 100644 data/2022/neurips/Reinforcement Learning with Automated Auxiliary Loss Search create mode 100644 data/2022/neurips/Reinforcement Learning with Logarithmic Regret and Policy Switches create mode 100644 data/2022/neurips/Reinforcement Learning with Neural Radiance Fields create mode 100644 data/2022/neurips/Reinforcement Learning with Non-Exponential Discounting create mode 100644 data/2022/neurips/Reinforcement Learning with a Terminator create mode 100644 data/2022/neurips/Relation-Constrained Decoding for Text Generation create mode 100644 data/2022/neurips/Relational Proxies: Emergent Relationships as Fine-Grained Discriminators create mode 100644 data/2022/neurips/Relational Reasoning via Set Transformers: Provable Efficiency and Applications to MARL create mode 100644 data/2022/neurips/Relaxing Equivariance Constraints with Non-stationary Continuous Filters create mode 100644 data/2022/neurips/Remember the Past: Distilling Datasets into Addressable Memories for Neural Networks create mode 100644 data/2022/neurips/Renyi Differential Privacy of Propose-Test-Release and Applications to Private and Robust Machine Learning create mode 100644 data/2022/neurips/Repairing Neural Networks by Leaving the Right Past Behind create mode 100644 data/2022/neurips/Representing Spatial Trajectories as Distributions create mode 100644 data/2022/neurips/Reproducibility in Optimization: Theoretical Framework and Limits create mode 100644 data/2022/neurips/ResQ: A Residual Q Function-based Approach for Multi-Agent Reinforcement Learning Value Factorization create mode 100644 data/2022/neurips/ResT V2: Simpler, Faster and Stronger create mode 100644 data/2022/neurips/Residual Multiplicative Filter Networks for Multiscale Reconstruction create mode 100644 data/2022/neurips/Resolving the data ambiguity for periodic crystals create mode 100644 data/2022/neurips/Resource-Adaptive Federated Learning with All-In-One Neural Composition create mode 100644 data/2022/neurips/Respecting Transfer Gap in Knowledge Distillation create mode 100644 data/2022/neurips/Retaining Knowledge for Learning with Dynamic Definition create mode 100644 data/2022/neurips/Rethinking Alignment in Video Super-Resolution Transformers create mode 100644 data/2022/neurips/Rethinking Generalization in Few-Shot Classification create mode 100644 data/2022/neurips/Rethinking Image Restoration for Object Detection create mode 100644 data/2022/neurips/Rethinking Individual Global Max in Cooperative Multi-Agent Reinforcement Learning create mode 100644 data/2022/neurips/Rethinking Knowledge Graph Evaluation Under the Open-World Assumption create mode 100644 data/2022/neurips/Rethinking Lipschitz Neural Networks and Certified Robustness: A Boolean Function Perspective create mode 100644 data/2022/neurips/Rethinking Resolution in the Context of Efficient Video Recognition create mode 100644 data/2022/neurips/Rethinking Value Function Learning for Generalization in Reinforcement Learning create mode 100644 data/2022/neurips/Rethinking Variational Inference for Probabilistic Programs with Stochastic Support create mode 100644 data/2022/neurips/Rethinking and Improving Robustness of Convolutional Neural Networks: a Shapley Value-based Approach in Frequency Domain create mode 100644 data/2022/neurips/Rethinking and Scaling Up Graph Contrastive Learning: An Extremely Efficient Approach with Group Discrimination create mode 100644 data/2022/neurips/Rethinking the Reverse-engineering of Trojan Triggers create mode 100644 data/2022/neurips/Rethinking the compositionality of point clouds through regularization in the hyperbolic space create mode 100644 data/2022/neurips/Retrieval-Augmented Diffusion Models create mode 100644 data/2022/neurips/Retrieve, Reason, and Refine: Generating Accurate and Faithful Patient Instructions create mode 100644 data/2022/neurips/Retrospective Adversarial Replay for Continual Learning create mode 100644 data/2022/neurips/Revisit last-iterate convergence of mSGD under milder requirement on step size create mode 100644 data/2022/neurips/Revisiting Active Sets for Gaussian Process Decoders create mode 100644 data/2022/neurips/Revisiting Graph Contrastive Learning from the Perspective of Graph Spectrum create mode 100644 data/2022/neurips/Revisiting Heterophily For Graph Neural Networks create mode 100644 data/2022/neurips/Revisiting Injective Attacks on Recommender Systems create mode 100644 data/2022/neurips/Revisiting Neural Scaling Laws in Language and Vision create mode 100644 data/2022/neurips/Revisiting Non-Parametric Matching Cost Volumes for Robust and Generalizable Stereo Matching create mode 100644 data/2022/neurips/Revisiting Optimal Convergence Rate for Smooth and Non-convex Stochastic Decentralized Optimization create mode 100644 data/2022/neurips/Revisiting Realistic Test-Time Training: Sequential Inference and Adaptation by Anchored Clustering create mode 100644 data/2022/neurips/Revisiting Sliced Wasserstein on Images: From Vectorization to Convolution create mode 100644 data/2022/neurips/Revisiting Sparse Convolutional Model for Visual Recognition create mode 100644 data/2022/neurips/Riemannian Diffusion Models create mode 100644 data/2022/neurips/Riemannian Neural SDE: Learning Stochastic Representations on Manifolds create mode 100644 data/2022/neurips/Riemannian Score-Based Generative Modelling create mode 100644 data/2022/neurips/Risk Bounds of Multi-Pass SGD for Least Squares in the Interpolation Regime create mode 100644 data/2022/neurips/Risk-Driven Design of Perception Systems create mode 100644 data/2022/neurips/Roadblocks for Temporarily Disabling Shortcuts and Learning New Knowledge create mode 100644 data/2022/neurips/Robust Anytime Learning of Markov Decision Processes create mode 100644 data/2022/neurips/Robust Bayesian Regression via Hard Thresholding create mode 100644 data/2022/neurips/Robust Binary Models by Pruning Randomly-initialized Networks create mode 100644 data/2022/neurips/Robust Calibration with Multi-domain Temperature Scaling create mode 100644 data/2022/neurips/Robust Feature-Level Adversaries are Interpretability Tools create mode 100644 data/2022/neurips/Robust Generalized Method of Moments: A Finite Sample Viewpoint create mode 100644 data/2022/neurips/Robust Graph Structure Learning via Multiple Statistical Tests create mode 100644 data/2022/neurips/Robust Imitation of a Few Demonstrations with a Backwards Model create mode 100644 data/2022/neurips/Robust Imitation via Mirror Descent Inverse Reinforcement Learning create mode 100644 data/2022/neurips/Robust Learning against Relational Adversaries create mode 100644 data/2022/neurips/Robust Model Selection and Nearly-Proper Learning for GMMs create mode 100644 data/2022/neurips/Robust Models are less Over-Confident create mode 100644 data/2022/neurips/Robust Neural Posterior Estimation and Statistical Model Criticism create mode 100644 data/2022/neurips/Robust On-Policy Sampling for Data-Efficient Policy Evaluation in Reinforcement Learning create mode 100644 data/2022/neurips/Robust Reinforcement Learning using Offline Data create mode 100644 data/2022/neurips/Robust Rent Division create mode 100644 data/2022/neurips/Robust Semi-Supervised Learning when Not All Classes have Labels create mode 100644 data/2022/neurips/Robust Streaming PCA create mode 100644 data/2022/neurips/Robust Testing in High-Dimensional Sparse Models create mode 100644 data/2022/neurips/Robustness Analysis of Video-Language Models Against Visual and Language Perturbations create mode 100644 data/2022/neurips/Robustness Disparities in Face Detection create mode 100644 data/2022/neurips/Robustness in deep learning: The good (width), the bad (depth), and the ugly (initialization) create mode 100644 data/2022/neurips/Robustness to Label Noise Depends on the Shape of the Noise Distribution create mode 100644 data/2022/neurips/Robustness to Unbounded Smoothness of Generalized SignSGD create mode 100644 data/2022/neurips/Root Cause Analysis of Failures in Microservices through Causal Discovery create mode 100644 data/2022/neurips/Rotation-Equivariant Conditional Spherical Neural Fields for Learning a Natural Illumination Prior create mode 100644 "data/2022/neurips/R\303\251nyiCL: Contrastive Representation Learning with Skew R\303\251nyi Divergence" create mode 100644 data/2022/neurips/S-PIFu: Integrating Parametric Human Models with PIFu for Single-view Clothed Human Reconstruction create mode 100644 data/2022/neurips/S-Prompts Learning with Pre-trained Transformers: An Occam's Razor for Domain Incremental Learning create mode 100644 data/2022/neurips/S2P: State-conditioned Image Synthesis for Data Augmentation in Offline Reinforcement Learning create mode 100644 data/2022/neurips/S3-NeRF: Neural Reflectance Field from Shading and Shadow under a Single Viewpoint create mode 100644 data/2022/neurips/S3GC: Scalable Self-Supervised Graph Clustering create mode 100644 data/2022/neurips/S4ND: Modeling Images and Videos as Multidimensional Signals with State Spaces create mode 100644 data/2022/neurips/SALSA: Attacking Lattice Cryptography with Transformers create mode 100644 data/2022/neurips/SAMURAI: Shape And Material from Unconstrained Real-world Arbitrary Image collections create mode 100644 data/2022/neurips/SAPA: Similarity-Aware Point Affiliation for Feature Upsampling create mode 100644 data/2022/neurips/SAPD+: An Accelerated Stochastic Method for Nonconvex-Concave Minimax Problems create mode 100644 data/2022/neurips/SAPipe: Staleness-Aware Pipeline for Data Parallel DNN Training create mode 100644 data/2022/neurips/SAVi++: Towards End-to-End Object-Centric Learning from Real-World Videos create mode 100644 data/2022/neurips/SAViT: Structure-Aware Vision Transformer Pruning via Collaborative Optimization create mode 100644 data/2022/neurips/SCAMPS: Synthetics for Camera Measurement of Physiological Signals create mode 100644 data/2022/neurips/SCINet: Time Series Modeling and Forecasting with Sample Convolution and Interaction create mode 100644 data/2022/neurips/SCL-WC: Cross-Slide Contrastive Learning for Weakly-Supervised Whole-Slide Image Classification create mode 100644 data/2022/neurips/SCONE: Surface Coverage Optimization in Unknown Environments by Volumetric Integration create mode 100644 data/2022/neurips/SGAM: Building a Virtual 3D World through Simultaneous Generation and Mapping create mode 100644 data/2022/neurips/SHAQ: Incorporating Shapley Value Theory into Multi-Agent Q-Learning create mode 100644 data/2022/neurips/SHINE: SubHypergraph Inductive Neural nEtwork create mode 100644 data/2022/neurips/SIREN: Shaping Representations for Detecting Out-of-Distribution Objects create mode 100644 data/2022/neurips/SIXO: Smoothing Inference with Twisted Objectives create mode 100644 data/2022/neurips/SInGE: Sparsity via Integrated Gradients Estimation of Neuron Relevance create mode 100644 data/2022/neurips/SKFlow: Learning Optical Flow with Super Kernels create mode 100644 data/2022/neurips/SMPL: Simulated Industrial Manufacturing and Process Control Learning Environments create mode 100644 data/2022/neurips/SNAKE: Shape-aware Neural 3D Keypoint Field create mode 100644 data/2022/neurips/SNN-RAT: Robustness-enhanced Spiking Neural Network through Regularized Adversarial Training create mode 100644 data/2022/neurips/SPD domain-specific batch normalization to crack interpretable unsupervised domain adaptation in EEG create mode 100644 data/2022/neurips/SPD: Synergy Pattern Diversifying Oriented Unsupervised Multi-agent Reinforcement Learning create mode 100644 data/2022/neurips/SPoVT: Semantic-Prototype Variational Transformer for Dense Point Cloud Semantic Completion create mode 100644 data/2022/neurips/SQ Lower Bounds for Learning Single Neurons with Massart Noise create mode 100644 data/2022/neurips/ST-Adapter: Parameter-Efficient Image-to-Video Transfer Learning create mode 100644 data/2022/neurips/STNDT: Modeling Neural Population Activity with Spatiotemporal Transformers create mode 100644 data/2022/neurips/STaR: Bootstrapping Reasoning With Reasoning create mode 100644 data/2022/neurips/Safe Opponent-Exploitation Subgame Refinement create mode 100644 data/2022/neurips/SafeBench: A Benchmarking Platform for Safety Evaluation of Autonomous Vehicles create mode 100644 data/2022/neurips/Safety Guarantees for Neural Network Dynamic Systems via Stochastic Barrier Functions create mode 100644 data/2022/neurips/SageMix: Saliency-Guided Mixup for Point Clouds create mode 100644 data/2022/neurips/Saliency-Aware Neural Architecture Search create mode 100644 data/2022/neurips/Sample Complexity of Learning Heuristic Functions for Greedy-Best-First and A* Search create mode 100644 data/2022/neurips/Sample Constrained Treatment Effect Estimation create mode 100644 data/2022/neurips/Sample Efficiency Matters: A Benchmark for Practical Molecular Optimization create mode 100644 data/2022/neurips/Sample-Efficient Learning of Correlated Equilibria in Extensive-Form Games create mode 100644 data/2022/neurips/Sample-Efficient Reinforcement Learning of Partially Observable Markov Games create mode 100644 data/2022/neurips/Sample-Then-Optimize Batch Neural Thompson Sampling create mode 100644 data/2022/neurips/Sampling from Log-Concave Distributions with Infinity-Distance Guarantees create mode 100644 data/2022/neurips/Sampling in Constrained Domains with Orthogonal-Space Variational Gradient Descent create mode 100644 data/2022/neurips/Sampling with Riemannian Hamiltonian Monte Carlo in a Constrained Space create mode 100644 data/2022/neurips/Sampling without Replacement Leads to Faster Rates in Finite-Sum Minimax Optimization create mode 100644 data/2022/neurips/SatMAE: Pre-training Transformers for Temporal and Multi-Spectral Satellite Imagery create mode 100644 data/2022/neurips/Scalable Distributional Robustness in a Class of Non-Convex Optimization with Guarantees create mode 100644 data/2022/neurips/Scalable Infomin Learning create mode 100644 data/2022/neurips/Scalable Interpretability via Polynomials create mode 100644 data/2022/neurips/Scalable Multi-agent Covering Option Discovery based on Kronecker Graphs create mode 100644 data/2022/neurips/Scalable Neural Video Representations with Learnable Positional Features create mode 100644 data/2022/neurips/Scalable Representation Learning in Linear Contextual Bandits with Constant Regret Guarantees create mode 100644 data/2022/neurips/Scalable Sensitivity and Uncertainty Analyses for Causal-Effect Estimates of Continuous-Valued Interventions create mode 100644 data/2022/neurips/Scalable and Efficient Non-adaptive Deterministic Group Testing create mode 100644 data/2022/neurips/Scalable and Efficient Training of Large Convolutional Neural Networks with Differential Privacy create mode 100644 data/2022/neurips/Scalable design of Error-Correcting Output Codes using Discrete Optimization with Graph Coloring create mode 100644 data/2022/neurips/Scale-invariant Learning by Physics Inversion create mode 100644 data/2022/neurips/Scaling & Shifting Your Features: A New Baseline for Efficient Model Tuning create mode 100644 data/2022/neurips/Scaling Multimodal Pre-Training via Cross-Modality Gradient Harmonization create mode 100644 data/2022/neurips/Score-Based Diffusion meets Annealed Importance Sampling create mode 100644 data/2022/neurips/Score-Based Generative Models Detect Manifolds create mode 100644 data/2022/neurips/Score-based Generative Modeling Secretly Minimizes the Wasserstein Distance create mode 100644 data/2022/neurips/Searching for Better Spatio-temporal Alignment in Few-Shot Action Recognition create mode 100644 data/2022/neurips/Second Thoughts are Best: Learning to Re-Align With Human Values from Text Edits create mode 100644 data/2022/neurips/SecureFedYJ: a safe feature Gaussianization protocol for Federated Learning create mode 100644 data/2022/neurips/Seeing the forest and the tree: Building representations of both individual and collective dynamics with transformers create mode 100644 data/2022/neurips/SegNeXt: Rethinking Convolutional Attention Design for Semantic Segmentation create mode 100644 data/2022/neurips/SegViT: Semantic Segmentation with Plain Vision Transformers create mode 100644 data/2022/neurips/Segmenting Moving Objects via an Object-Centric Layered Representation create mode 100644 data/2022/neurips/SelecMix: Debiased Learning by Contradicting-pair Sampling create mode 100644 data/2022/neurips/Selective compression learning of latent representations for variable-rate image compression create mode 100644 data/2022/neurips/Self-Aware Personalized Federated Learning create mode 100644 data/2022/neurips/Self-Consistent Dynamical Field Theory of Kernel Evolution in Wide Neural Networks create mode 100644 data/2022/neurips/Self-Explaining Deviations for Coordination create mode 100644 data/2022/neurips/Self-Organized Group for Cooperative Multi-agent Reinforcement Learning create mode 100644 data/2022/neurips/Self-Similarity Priors: Neural Collages as Differentiable Fractal Representations create mode 100644 data/2022/neurips/Self-Supervised Aggregation of Diverse Experts for Test-Agnostic Long-Tailed Recognition create mode 100644 data/2022/neurips/Self-Supervised Contrastive Pre-Training For Time Series via Time-Frequency Consistency create mode 100644 data/2022/neurips/Self-Supervised Fair Representation Learning without Demographics create mode 100644 data/2022/neurips/Self-Supervised Image Restoration with Blurry and Noisy Pairs create mode 100644 data/2022/neurips/Self-Supervised Learning Through Efference Copies create mode 100644 data/2022/neurips/Self-Supervised Learning of Brain Dynamics from Broad Neuroimaging Data create mode 100644 data/2022/neurips/Self-Supervised Learning via Maximum Entropy Coding create mode 100644 data/2022/neurips/Self-Supervised Learning with an Information Maximization Criterion create mode 100644 data/2022/neurips/Self-Supervised Pretraining for Large-Scale Point Clouds create mode 100644 data/2022/neurips/Self-Supervised Visual Representation Learning with Semantic Grouping create mode 100644 data/2022/neurips/Self-explaining deep models with logic rule reasoning create mode 100644 data/2022/neurips/Self-supervised Amodal Video Object Segmentation create mode 100644 data/2022/neurips/Self-supervised Heterogeneous Graph Pre-training Based on Structural Clustering create mode 100644 data/2022/neurips/Self-supervised surround-view depth estimation with volumetric feature fusion create mode 100644 data/2022/neurips/SemMAE: Semantic-Guided Masking for Learning Masked Autoencoders create mode 100644 data/2022/neurips/Semantic Diffusion Network for Semantic Segmentation create mode 100644 data/2022/neurips/Semantic Exploration from Language Abstractions and Pretrained Representations create mode 100644 data/2022/neurips/Semantic Probabilistic Layers for Neuro-Symbolic Learning create mode 100644 data/2022/neurips/Semantic uncertainty intervals for disentangled latent spaces create mode 100644 data/2022/neurips/Semi-Discrete Normalizing Flows through Differentiable Tessellation create mode 100644 data/2022/neurips/Semi-Supervised Generative Models for Multiagent Trajectories create mode 100644 data/2022/neurips/Semi-Supervised Learning with Decision Trees: Graph Laplacian Tree Alternating Optimization create mode 100644 data/2022/neurips/Semi-Supervised Semantic Segmentation via Gentle Teaching Assistant create mode 100644 data/2022/neurips/Semi-Supervised Video Salient Object Detection Based on Uncertainty-Guided Pseudo Labels create mode 100644 data/2022/neurips/Semi-infinitely Constrained Markov Decision Processes create mode 100644 data/2022/neurips/Semi-supervised Active Linear Regression create mode 100644 data/2022/neurips/Semi-supervised Semantic Segmentation with Prototype-based Consistency Regularization create mode 100644 data/2022/neurips/Semi-supervised Vision Transformers at Scale create mode 100644 data/2022/neurips/SemiFL: Semi-Supervised Federated Learning for Unlabeled Clients with Alternate Training create mode 100644 data/2022/neurips/SeqPATE: Differentially Private Text Generation via Knowledge Distillation create mode 100644 data/2022/neurips/Sequence Model Imitation Learning with Unobserved Contexts create mode 100644 data/2022/neurips/Sequence-to-Set Generative Models create mode 100644 data/2022/neurips/Sequencer: Deep LSTM for Image Classification create mode 100644 data/2022/neurips/Sequential Information Design: Learning to Persuade in the Dark create mode 100644 data/2022/neurips/Set-based Meta-Interpolation for Few-Task Meta-Learning create mode 100644 data/2022/neurips/Shadow Knowledge Distillation: Bridging Offline and Online Knowledge Transfer create mode 100644 data/2022/neurips/Shape And Structure Preserving Differential Privacy create mode 100644 data/2022/neurips/Shape, Light, and Material Decomposition from Images using Monte Carlo Rendering and Denoising create mode 100644 data/2022/neurips/ShapeCrafter: A Recursive Text-Conditioned 3D Shape Generation Model create mode 100644 data/2022/neurips/Sharing Knowledge for Meta-learning with Feature Descriptions create mode 100644 data/2022/neurips/Sharp Analysis of Stochastic Optimization under Global Kurdyka-Lojasiewicz Inequality create mode 100644 data/2022/neurips/Sharper Convergence Guarantees for Asynchronous SGD for Distributed and Federated Learning create mode 100644 data/2022/neurips/Sharpness-Aware Training for Free create mode 100644 data/2022/neurips/Shield Decentralization for Safe Multi-Agent Reinforcement Learning create mode 100644 data/2022/neurips/ShuffleMixer: An Efficient ConvNet for Image Super-Resolution create mode 100644 data/2022/neurips/SignRFF: Sign Random Fourier Features create mode 100644 data/2022/neurips/Signal Processing for Implicit Neural Representations create mode 100644 data/2022/neurips/Signal Propagation in Transformers: Theoretical Perspectives and the Role of Rank Collapse create mode 100644 data/2022/neurips/Signal Recovery with Non-Expansive Generative Network Priors create mode 100644 data/2022/neurips/Simple Mechanisms for Welfare Maximization in Rich Advertising Auctions create mode 100644 data/2022/neurips/Simple Unsupervised Object-Centric Learning for Complex and Naturalistic Videos create mode 100644 data/2022/neurips/Simple and Optimal Greedy Online Contention Resolution Schemes create mode 100644 data/2022/neurips/Simplified Graph Convolution with Heterophily create mode 100644 data/2022/neurips/Simulation-guided Beam Search for Neural Combinatorial Optimization create mode 100644 data/2022/neurips/Simultaneous Missing Value Imputation and Structure Learning with Groups create mode 100644 data/2022/neurips/Single Loop Gaussian Homotopy Method for Non-convex Optimization create mode 100644 data/2022/neurips/Single Model Uncertainty Estimation via Stochastic Data Centering create mode 100644 data/2022/neurips/Single-Stage Visual Relationship Learning using Conditional Queries create mode 100644 data/2022/neurips/Single-pass Streaming Lower Bounds for Multi-armed Bandits Exploration with Instance-sensitive Sample Complexity create mode 100644 data/2022/neurips/Single-phase deep learning in cortico-cortical networks create mode 100644 data/2022/neurips/Singular Value Fine-tuning: Few-shot Segmentation requires Few-parameters Fine-tuning create mode 100644 data/2022/neurips/Size and depth of monotone neural networks: interpolation and approximation create mode 100644 data/2022/neurips/SizeShiftReg: a Regularization Method for Improving Size-Generalization in Graph Neural Networks create mode 100644 data/2022/neurips/Sketch-GNN: Scalable Graph Neural Networks with Sublinear Training Complexity create mode 100644 data/2022/neurips/SketchBoost: Fast Gradient Boosted Decision Tree for Multioutput Problems create mode 100644 data/2022/neurips/Sketching based Representations for Robust Image Classification with Provable Guarantees create mode 100644 data/2022/neurips/Skills Regularized Task Decomposition for Multi-task Offline Reinforcement Learning create mode 100644 data/2022/neurips/SkinCon: A skin disease dataset densely annotated by domain experts for fine-grained debugging and analysis create mode 100644 data/2022/neurips/Sleeper Agent: Scalable Hidden Trigger Backdoors for Neural Networks Trained from Scratch create mode 100644 data/2022/neurips/Smooth Fictitious Play in Stochastic Games with Perturbed Payoffs and Unknown Transitions create mode 100644 data/2022/neurips/Smoothed Embeddings for Certified Few-Shot Learning create mode 100644 data/2022/neurips/Smoothed Online Convex Optimization Based on Discounted-Normal-Predictor create mode 100644 data/2022/neurips/SnAKe: Bayesian Optimization with Pathwise Exploration create mode 100644 data/2022/neurips/So3krates: Equivariant attention for interactions on arbitrary length-scales in molecular systems create mode 100644 data/2022/neurips/SoLar: Sinkhorn Label Refinery for Imbalanced Partial-Label Learning create mode 100644 data/2022/neurips/Sobolev Acceleration and Statistical Optimality for Learning Elliptic Equations via Gradient Descent create mode 100644 data/2022/neurips/Social-Inverse: Inverse Decision-making of Social Contagion Management with Task Migrations create mode 100644 data/2022/neurips/Society of Agents: Regret Bounds of Concurrent Thompson Sampling create mode 100644 data/2022/neurips/SoftPatch: Unsupervised Anomaly Detection with Noisy Data create mode 100644 data/2022/neurips/Solving Quantitative Reasoning Problems with Language Models create mode 100644 data/2022/neurips/SoteriaFL: A Unified Framework for Private Federated Learning with Communication Compression create mode 100644 data/2022/neurips/Sound and Complete Causal Identification with Latent Variables Given Local Background Knowledge create mode 100644 data/2022/neurips/Sound and Complete Verification of Polynomial Networks create mode 100644 data/2022/neurips/SoundSpaces 2.0: A Simulation Platform for Visual-Acoustic Learning create mode 100644 data/2022/neurips/SparCL: Sparse Continual Learning on the Edge create mode 100644 data/2022/neurips/Sparse Fourier Backpropagation in Cryo-EM Reconstruction create mode 100644 data/2022/neurips/Sparse Gaussian Process Hyperparameters: Optimize or Integrate? create mode 100644 data/2022/neurips/Sparse Hypergraph Community Detection Thresholds in Stochastic Block Model create mode 100644 data/2022/neurips/Sparse Interaction Additive Networks via Feature Interaction Detection and Sparse Selection create mode 100644 data/2022/neurips/Sparse Probabilistic Circuits via Pruning and Growing create mode 100644 data/2022/neurips/Sparse Structure Search for Delta Tuning create mode 100644 data/2022/neurips/Sparse Winning Tickets are Data-Efficient Image Recognizers create mode 100644 data/2022/neurips/Sparse2Dense: Learning to Densify 3D Features for 3D Object Detection create mode 100644 data/2022/neurips/Sparsity in Continuous-Depth Neural Networks create mode 100644 data/2022/neurips/Spartan: Differentiable Sparsity via Regularized Transportation create mode 100644 data/2022/neurips/Spatial Mixture-of-Experts create mode 100644 data/2022/neurips/Spatial Pruned Sparse Convolution for Efficient 3D Object Detection create mode 100644 data/2022/neurips/Spectral Bias Outside the Training Set for Deep Networks in the Kernel Regime create mode 100644 data/2022/neurips/Spectrum Random Masking for Generalization in Image-based Reinforcement Learning create mode 100644 data/2022/neurips/Spending Thinking Time Wisely: Accelerating MCTS with Virtual Expansions create mode 100644 data/2022/neurips/Spherical Channels for Modeling Atomic Interactions create mode 100644 data/2022/neurips/Spherization Layer: Representation Using Only Angles create mode 100644 data/2022/neurips/Split-kl and PAC-Bayes-split-kl Inequalities for Ternary Random Variables create mode 100644 data/2022/neurips/Squeezeformer: An Efficient Transformer for Automatic Speech Recognition create mode 100644 data/2022/neurips/Stability Analysis and Generalization Bounds of Adversarial Training create mode 100644 data/2022/neurips/Stability and Generalization Analysis of Gradient Methods for Shallow Neural Networks create mode 100644 data/2022/neurips/Stability and Generalization for Markov Chain Stochastic Gradient Methods create mode 100644 data/2022/neurips/Stability and Generalization of Kernel Clustering: from Single Kernel to Multiple Kernel create mode 100644 data/2022/neurips/Staggered Rollout Designs Enable Causal Inference Under Interference Without Network Knowledge create mode 100644 data/2022/neurips/Staircase Attention for Recurrent Processing of Sequences create mode 100644 data/2022/neurips/Star Temporal Classification: Sequence Modeling with Partially Labeled Data create mode 100644 data/2022/neurips/Stars: Tera-Scale Graph Building for Clustering and Learning create mode 100644 data/2022/neurips/Statistical Learning and Inverse Problems: A Stochastic Gradient Approach create mode 100644 data/2022/neurips/Statistical, Robustness, and Computational Guarantees for Sliced Wasserstein Distances create mode 100644 data/2022/neurips/Statistically Meaningful Approximation: a Case Study on Approximating Turing Machines with Transformers create mode 100644 data/2022/neurips/Stimulative Training of Residual Networks: A Social Psychology Perspective of Loafing create mode 100644 data/2022/neurips/Stochastic Adaptive Activation Function create mode 100644 data/2022/neurips/Stochastic Halpern Iteration with Variance Reduction for Stochastic Monotone Inclusions create mode 100644 data/2022/neurips/Stochastic Multiple Target Sampling Gradient Descent create mode 100644 data/2022/neurips/Stochastic Online Learning with Feedback Graphs: Finite-Time and Asymptotic Optimality create mode 100644 data/2022/neurips/Stochastic Second-Order Methods Improve Best-Known Sample Complexity of SGD for Gradient-Dominated Functions create mode 100644 data/2022/neurips/Stochastic Window Transformer for Image Restoration create mode 100644 data/2022/neurips/Streaming Radiance Fields for 3D Video Synthesis create mode 100644 data/2022/neurips/StrokeRehab: A Benchmark Dataset for Sub-second Action Identification create mode 100644 data/2022/neurips/Structural Analysis of Branch-and-Cut and the Learnability of Gomory Mixed Integer Cuts create mode 100644 data/2022/neurips/Structural Kernel Search via Bayesian Optimization and Symbolical Optimal Transport create mode 100644 data/2022/neurips/Structural Knowledge Distillation for Object Detection create mode 100644 data/2022/neurips/Structural Pruning via Latency-Saliency Knapsack create mode 100644 data/2022/neurips/Structure-Aware Image Segmentation with Homotopy Warping create mode 100644 data/2022/neurips/Structure-Preserving 3D Garment Modeling with Neural Sewing Machines create mode 100644 data/2022/neurips/Structured Energy Network As a Loss create mode 100644 data/2022/neurips/Structured Recognition for Generative Models with Explaining Away create mode 100644 data/2022/neurips/Structuring Representations Using Group Invariants create mode 100644 data/2022/neurips/Structuring Uncertainty for Fine-Grained Sampling in Stochastic Segmentation Networks create mode 100644 data/2022/neurips/Sub-exponential time Sum-of-Squares lower bounds for Principal Components Analysis create mode 100644 data/2022/neurips/Subgame Solving in Adversarial Team Games create mode 100644 data/2022/neurips/Subgroup Robustness Grows On Trees: An Empirical Baseline Investigation create mode 100644 data/2022/neurips/Sublinear Algorithms for Hierarchical Clustering create mode 100644 data/2022/neurips/Submodular Maximization in Clean Linear Time create mode 100644 data/2022/neurips/Subquadratic Kronecker Regression with Applications to Tensor Decomposition create mode 100644 data/2022/neurips/Subsidiary Prototype Alignment for Universal Domain Adaptation create mode 100644 data/2022/neurips/Subspace Recovery from Heterogeneous Data with Non-isotropic Noise create mode 100644 data/2022/neurips/Subspace clustering in high-dimensions: Phase transitions & Statistical-to-Computational gap create mode 100644 data/2022/neurips/Supervised Training of Conditional Monge Maps create mode 100644 data/2022/neurips/Supervising the Multi-Fidelity Race of Hyperparameter Configurations create mode 100644 data/2022/neurips/Support Recovery in Sparse PCA with Incomplete Data create mode 100644 data/2022/neurips/Supported Policy Optimization for Offline Reinforcement Learning create mode 100644 data/2022/neurips/SurDis: A Surface Discontinuity Dataset for Wearable Technology to Assist Blind Navigation in Urban Environments create mode 100644 data/2022/neurips/Surprise Minimizing Multi-Agent Learning with Energy-based Models create mode 100644 data/2022/neurips/Sustainable Online Reinforcement Learning for Auto-bidding create mode 100644 data/2022/neurips/SwinTrack: A Simple and Strong Baseline for Transformer Tracking create mode 100644 data/2022/neurips/Sym-NCO: Leveraging Symmetricity for Neural Combinatorial Optimization create mode 100644 data/2022/neurips/Symbolic Distillation for Learned TCP Congestion Control create mode 100644 data/2022/neurips/Symmetry Teleportation for Accelerated Optimization create mode 100644 data/2022/neurips/Symmetry-induced Disentanglement on Graphs create mode 100644 data/2022/neurips/Symplectic Spectrum Gaussian Processes: Learning Hamiltonians from Noisy and Sparse Data create mode 100644 data/2022/neurips/Syndicated Bandits: A Framework for Auto Tuning Hyper-parameters in Contextual Bandit Algorithms create mode 100644 data/2022/neurips/Synergy-of-Experts: Collaborate to Improve Adversarial Robustness create mode 100644 data/2022/neurips/Synthetic Model Combination: An Instance-wise Approach to Unsupervised Ensemble Learning create mode 100644 data/2022/neurips/Systematic improvement of neural network quantum states using Lanczos create mode 100644 data/2022/neurips/TA-GATES: An Encoding Scheme for Neural Network Architectures create mode 100644 data/2022/neurips/TA-MoE: Topology-Aware Large Scale Mixture-of-Expert Training create mode 100644 data/2022/neurips/TANGO: Text-driven Photorealistic and Robust 3D Stylization via Lighting Decomposition create mode 100644 data/2022/neurips/TANKBind: Trigonometry-Aware Neural NetworKs for Drug-Protein Binding Structure Prediction create mode 100644 data/2022/neurips/TAP-Vid: A Benchmark for Tracking Any Point in a Video create mode 100644 data/2022/neurips/TCT: Convexifying Federated Learning using Bootstrapped Neural Tangent Kernels create mode 100644 data/2022/neurips/TGEA 2.0: A Large-Scale Diagnostically Annotated Dataset with Benchmark Tasks for Text Generation of Pretrained Language Models create mode 100644 data/2022/neurips/TOIST: Task Oriented Instance Segmentation Transformer with Noun-Pronoun Distillation create mode 100644 data/2022/neurips/TPU-KNN: K Nearest Neighbor Search at Peak FLOP s create mode 100644 data/2022/neurips/TREC: Transient Redundancy Elimination-based Convolution create mode 100644 data/2022/neurips/TTOpt: A Maximum Volume Quantized Tensor Train-based Optimization and its Application to Reinforcement Learning create mode 100644 data/2022/neurips/TUSK: Task-Agnostic Unsupervised Keypoints create mode 100644 data/2022/neurips/TVLT: Textless Vision-Language Transformer create mode 100644 data/2022/neurips/TaSIL: Taylor Series Imitation Learning create mode 100644 data/2022/neurips/TabNAS: Rejection Sampling for Neural Architecture Search on Tabular Datasets create mode 100644 data/2022/neurips/TaiSu: A 166M Large-scale High-Quality Dataset for Chinese Vision-Language Pre-training create mode 100644 "data/2022/neurips/Taming Fat-Tailed (\"Heavier-Tailed\" with Potentially Infinite Variance) Noise in Federated Learning" create mode 100644 data/2022/neurips/TarGF: Learning Target Gradient Field to Rearrange Objects without Explicit Goal Specification create mode 100644 data/2022/neurips/Target alignment in truncated kernel ridge regression create mode 100644 data/2022/neurips/Task Discovery: Finding the Tasks that Neural Networks Generalize on create mode 100644 data/2022/neurips/Task-Agnostic Graph Explanations create mode 100644 data/2022/neurips/Task-Free Continual Learning via Online Discrepancy Distance Learning create mode 100644 data/2022/neurips/Task-level Differentially Private Meta Learning create mode 100644 data/2022/neurips/Teach Less, Learn More: On the Undistillable Classes in Knowledge Distillation create mode 100644 data/2022/neurips/Teacher Forcing Recovers Reward Functions for Text Generation create mode 100644 data/2022/neurips/TempEL: Linking Dynamically Evolving and Newly Emerging Entities create mode 100644 data/2022/neurips/Template based Graph Neural Network with Optimal Transport Distances create mode 100644 data/2022/neurips/Tempo: Accelerating Transformer-Based Model Training through Memory Footprint Reduction create mode 100644 data/2022/neurips/Temporal Effective Batch Normalization in Spiking Neural Networks create mode 100644 data/2022/neurips/Temporal Latent Bottleneck: Synthesis of Fast and Slow Processing Mechanisms in Sequence Learning create mode 100644 data/2022/neurips/Temporally Disentangled Representation Learning create mode 100644 data/2022/neurips/Temporally-Consistent Survival Analysis create mode 100644 data/2022/neurips/Tenrec: A Large-scale Multipurpose Benchmark Dataset for Recommender Systems create mode 100644 data/2022/neurips/Tensor Program Optimization with Probabilistic Programs create mode 100644 data/2022/neurips/Tensor Wheel Decomposition and Its Tensor Completion Application create mode 100644 data/2022/neurips/Test Time Adaptation via Conjugate Pseudo-labels create mode 100644 data/2022/neurips/Test-Time Prompt Tuning for Zero-Shot Generalization in Vision-Language Models create mode 100644 data/2022/neurips/Test-Time Training with Masked Autoencoders create mode 100644 data/2022/neurips/Text Classification with Born's Rule create mode 100644 data/2022/neurips/Text-Adaptive Multiple Visual Prototype Matching for Video-Text Retrieval create mode 100644 data/2022/neurips/The BigScience ROOTS Corpus: A 1.6TB Composite Multilingual Dataset create mode 100644 data/2022/neurips/The Burer-Monteiro SDP method can fail even above the Barvinok-Pataki bound create mode 100644 data/2022/neurips/The Curse of Unrolling: Rate of Differentiating Through Optimization create mode 100644 data/2022/neurips/The Dollar Street Dataset: Images Representing the Geographic and Socioeconomic Diversity of the World create mode 100644 data/2022/neurips/The Effects of Regularization and Data Augmentation are Class Dependent create mode 100644 data/2022/neurips/The First Optimal Acceleration of High-Order Methods in Smooth Convex Optimization create mode 100644 data/2022/neurips/The First Optimal Algorithm for Smooth and Strongly-Convex-Strongly-Concave Minimax Optimization create mode 100644 data/2022/neurips/The Franz-Parisi Criterion and Computational Trade-offs in High Dimensional Statistics create mode 100644 data/2022/neurips/The Gyro-Structure of Some Matrix Manifolds create mode 100644 data/2022/neurips/The Hessian Screening Rule create mode 100644 data/2022/neurips/The Impact of Task Underspecification in Evaluating Deep Reinforcement Learning create mode 100644 data/2022/neurips/The Implicit Delta Method create mode 100644 data/2022/neurips/The Mechanism of Prediction Head in Non-contrastive Self-supervised Learning create mode 100644 data/2022/neurips/The Minority Matters: A Diversity-Promoting Collaborative Metric Learning Algorithm create mode 100644 data/2022/neurips/The Missing Invariance Principle found - the Reciprocal Twin of Invariant Risk Minimization create mode 100644 data/2022/neurips/The Nature of Temporal Difference Errors in Multi-step Distributional Reinforcement Learning create mode 100644 data/2022/neurips/The Neural Covariance SDE: Shaped Infinite Depth-and-Width Networks at Initialization create mode 100644 data/2022/neurips/The Neural Testbed: Evaluating Joint Predictions create mode 100644 data/2022/neurips/The Phenomenon of Policy Churn create mode 100644 data/2022/neurips/The Pitfalls of Regularization in Off-Policy TD Learning create mode 100644 data/2022/neurips/The Policy-gradient Placement and Generative Routing Neural Networks for Chip Design create mode 100644 data/2022/neurips/The Power and Limitation of Pretraining-Finetuning for Linear Regression under Covariate Shift create mode 100644 data/2022/neurips/The Privacy Onion Effect: Memorization is Relative create mode 100644 data/2022/neurips/The Query Complexity of Cake Cutting create mode 100644 data/2022/neurips/The Role of Baselines in Policy Gradient Optimization create mode 100644 data/2022/neurips/The Sample Complexity of One-Hidden-Layer Neural Networks create mode 100644 data/2022/neurips/The Stability-Efficiency Dilemma: Investigating Sequence Length Warmup for Training GPT Models create mode 100644 data/2022/neurips/The Surprising Effectiveness of PPO in Cooperative Multi-Agent Games create mode 100644 data/2022/neurips/The Unreasonable Effectiveness of Fully-Connected Layers for Low-Data Regimes create mode 100644 data/2022/neurips/The Unreliability of Explanations in Few-shot Prompting for Textual Reasoning create mode 100644 data/2022/neurips/The alignment property of SGD noise and how it helps select flat minima: A stability analysis create mode 100644 data/2022/neurips/The computational and learning benefits of Daleian neural networks create mode 100644 data/2022/neurips/The least-control principle for local learning at equilibrium create mode 100644 data/2022/neurips/The price of ignorance: how much does it cost to forget noise structure in low-rank matrix estimation? create mode 100644 data/2022/neurips/The price of unfairness in linear bandits with biased feedback create mode 100644 data/2022/neurips/The trade-offs of model size in large recommendation models : 100GB to 10MB Criteo-tb DLRM model create mode 100644 data/2022/neurips/Theoretical analysis of deep neural networks for temporally dependent observations create mode 100644 data/2022/neurips/Theoretically Better and Numerically Faster Distributed Optimization with Smoothness-Aware Quantization Techniques create mode 100644 data/2022/neurips/Theoretically Provable Spiking Neural Networks create mode 100644 data/2022/neurips/Theory and Approximate Solvers for Branched Optimal Transport with Multiple Sources create mode 100644 data/2022/neurips/Theseus: A Library for Differentiable Nonlinear Optimization create mode 100644 data/2022/neurips/Thinking Outside the Ball: Optimal Learning with Gradient Descent for Generalized Linear Stochastic Convex Optimization create mode 100644 data/2022/neurips/Thinned random measures for sparse graphs with overlapping communities create mode 100644 data/2022/neurips/This is the way: designing and compiling LEPISZCZE, a comprehensive NLP benchmark for Polish create mode 100644 data/2022/neurips/Thompson Sampling Efficiently Learns to Control Diffusion Processes create mode 100644 data/2022/neurips/Thor: Wielding Hammers to Integrate Language Models and Automated Theorem Provers create mode 100644 data/2022/neurips/Tiered Reinforcement Learning: Pessimism in the Face of Uncertainty and Constant Regret create mode 100644 data/2022/neurips/Tight Analysis of Extra-gradient and Optimistic Gradient Methods For Nonconvex Minimax Problems create mode 100644 data/2022/neurips/Tight Lower Bounds on Worst-Case Guarantees for Zero-Shot Learning with Attributes create mode 100644 data/2022/neurips/Tight Mutual Information Estimation With Contrastive Fenchel-Legendre Optimization create mode 100644 data/2022/neurips/Tikhonov Regularization is Optimal Transport Robust under Martingale Constraints create mode 100644 data/2022/neurips/Time-Conditioned Dances with Simplicial Complexes: Zigzag Filtration Curve based Supra-Hodge Convolution Networks for Time-series Forecasting create mode 100644 data/2022/neurips/To update or not to update? Neurons at equilibrium in deep models create mode 100644 data/2022/neurips/ToDD: Topological Compound Fingerprinting in Computer-Aided Drug Discovery create mode 100644 data/2022/neurips/TokenMixup: Efficient Attention-guided Token-level Data Augmentation for Transformers create mode 100644 data/2022/neurips/Top Two Algorithms Revisited create mode 100644 data/2022/neurips/Torsional Diffusion for Molecular Conformer Generation create mode 100644 data/2022/neurips/TotalSelfScan: Learning Full-body Avatars from Self-Portrait Videos of Faces, Hands, and Bodies create mode 100644 data/2022/neurips/Touch and Go: Learning from Human-Collected Vision and Touch create mode 100644 data/2022/neurips/Toward Equation of Motion for Deep Neural Networks: Continuous-time Gradient Descent and Discretization Error Analysis create mode 100644 data/2022/neurips/Toward Robust Spiking Neural Network Against Adversarial Perturbation create mode 100644 data/2022/neurips/Toward Understanding Privileged Features Distillation in Learning-to-Rank create mode 100644 data/2022/neurips/Toward a realistic model of speech processing in the brain with self-supervised learning create mode 100644 data/2022/neurips/Towards Better Evaluation for Dynamic Link Prediction create mode 100644 data/2022/neurips/Towards Consistency in Adversarial Classification create mode 100644 data/2022/neurips/Towards Disentangling Information Paths with Coded ResNeXt create mode 100644 data/2022/neurips/Towards Diverse and Faithful One-shot Adaption of Generative Adversarial Networks create mode 100644 data/2022/neurips/Towards Effective Multi-Modal Interchanges in Zero-Resource Sounding Object Localization create mode 100644 data/2022/neurips/Towards Efficient 3D Object Detection with Knowledge Distillation create mode 100644 data/2022/neurips/Towards Efficient Post-training Quantization of Pre-trained Language Models create mode 100644 data/2022/neurips/Towards Hard-pose Virtual Try-on via 3D-aware Global Correspondence Learning create mode 100644 data/2022/neurips/Towards Human-Level Bimanual Dexterous Manipulation with Reinforcement Learning create mode 100644 data/2022/neurips/Towards Improving Calibration in Object Detection Under Domain Shift create mode 100644 data/2022/neurips/Towards Improving Faithfulness in Abstractive Summarization create mode 100644 data/2022/neurips/Towards Learning Universal Hyperparameter Optimizers with Transformers create mode 100644 data/2022/neurips/Towards Lightweight Black-Box Attack Against Deep Neural Networks create mode 100644 data/2022/neurips/Towards Optimal Communication Complexity in Distributed Non-Convex Optimization create mode 100644 data/2022/neurips/Towards Out-of-Distribution Sequential Event Prediction: A Causal Treatment create mode 100644 data/2022/neurips/Towards Practical Control of Singular Values of Convolutional Layers create mode 100644 data/2022/neurips/Towards Practical Few-shot Query Sets: Transductive Minimum Description Length Inference create mode 100644 data/2022/neurips/Towards Reasonable Budget Allocation in Untargeted Graph Structure Attacks via Gradient Debias create mode 100644 data/2022/neurips/Towards Reliable Simulation-Based Inference with Balanced Neural Ratio Estimation create mode 100644 data/2022/neurips/Towards Robust Blind Face Restoration with Codebook Lookup Transformer create mode 100644 data/2022/neurips/Towards Safe Reinforcement Learning with a Safety Editor Policy create mode 100644 data/2022/neurips/Towards Theoretically Inspired Neural Initialization Optimization create mode 100644 data/2022/neurips/Towards Trustworthy Automatic Diagnosis Systems by Emulating Doctors' Reasoning with Deep Reinforcement Learning create mode 100644 data/2022/neurips/Towards Understanding Grokking: An Effective Theory of Representation Learning create mode 100644 data/2022/neurips/Towards Understanding the Condensation of Neural Networks at Initial Training create mode 100644 data/2022/neurips/Towards Understanding the Mixture-of-Experts Layer in Deep Learning create mode 100644 data/2022/neurips/Towards Versatile Embodied Navigation create mode 100644 data/2022/neurips/Towards Video Text Visual Question Answering: Benchmark and Baseline create mode 100644 data/2022/neurips/Towards a Standardised Performance Evaluation Protocol for Cooperative MARL create mode 100644 data/2022/neurips/Towards a Unified Framework for Uncertainty-aware Nonlinear Variable Selection with Theoretical Guarantees create mode 100644 data/2022/neurips/Tracking Functional Changes in Nonstationary Signals with Evolutionary Ensemble Bayesian Model for Robust Neural Decoding create mode 100644 data/2022/neurips/Tractable Function-Space Variational Inference in Bayesian Neural Networks create mode 100644 data/2022/neurips/Tractable Optimality in Episodic Latent MABs create mode 100644 data/2022/neurips/Trade-off between Payoff and Model Rewards in Shapley-Fair Collaborative Machine Learning create mode 100644 data/2022/neurips/Trading Off Resource Budgets For Improved Regret Bounds create mode 100644 data/2022/neurips/Trading off Image Quality for Robustness is not Necessary with Regularized Deterministic Autoencoders create mode 100644 data/2022/neurips/Trading off Utility, Informativeness, and Complexity in Emergent Communication create mode 100644 data/2022/neurips/Training Scale-Invariant Neural Networks on the Sphere Can Happen in Three Regimes create mode 100644 data/2022/neurips/Training Spiking Neural Networks with Event-driven Backpropagation create mode 100644 data/2022/neurips/Training Spiking Neural Networks with Local Tandem Learning create mode 100644 data/2022/neurips/Training Subset Selection for Weak Supervision create mode 100644 data/2022/neurips/Training Uncertainty-Aware Classifiers with Conformalized Deep Learning create mode 100644 data/2022/neurips/Training and Inference on Any-Order Autoregressive Models the Right Way create mode 100644 data/2022/neurips/Training language models to follow instructions with human feedback create mode 100644 data/2022/neurips/Training stochastic stabilized supralinear networks by dynamics-neutral growth create mode 100644 data/2022/neurips/Training with More Confidence: Mitigating Injected and Natural Backdoors During Training create mode 100644 data/2022/neurips/Trajectory Inference via Mean-field Langevin in Path Space create mode 100644 data/2022/neurips/Trajectory balance: Improved credit assignment in GFlowNets create mode 100644 data/2022/neurips/Trajectory of Mini-Batch Momentum: Batch Size Saturation and Convergence in High Dimensions create mode 100644 data/2022/neurips/Trajectory-guided Control Prediction for End-to-end Autonomous Driving: A Simple yet Strong Baseline create mode 100644 data/2022/neurips/TransBoost: Improving the Best ImageNet Performance using Deep Transduction create mode 100644 data/2022/neurips/TransTab: Learning Transferable Tabular Transformers Across Tables create mode 100644 data/2022/neurips/Transcormer: Transformer for Sentence Scoring with Sliding Language Modeling create mode 100644 data/2022/neurips/Transfer Learning on Heterogeneous Feature Spaces for Treatment Effects Estimation create mode 100644 data/2022/neurips/Transferring Fairness under Distribution Shifts via Fair Consistency Regularization create mode 100644 data/2022/neurips/Transferring Pre-trained Multimodal Representations with Cross-modal Similarity Matching create mode 100644 data/2022/neurips/Transform Once: Efficient Operator Learning in Frequency Domain create mode 100644 data/2022/neurips/Transformer Memory as a Differentiable Search Index create mode 100644 data/2022/neurips/Transformer-based Working Memory for Multiagent Reinforcement Learning with Action Parsing create mode 100644 data/2022/neurips/Transformers from an Optimization Perspective create mode 100644 data/2022/neurips/Transformers meet Stochastic Block Models: Attention with Data-Adaptive Sparsity and Cost create mode 100644 data/2022/neurips/Transition to Linearity of General Neural Networks with Directed Acyclic Graph Architecture create mode 100644 data/2022/neurips/Translation-equivariant Representation in Recurrent Networks with a Continuous Manifold of Attractors create mode 100644 data/2022/neurips/Trap and Replace: Defending Backdoor Attacks by Trapping Them into an Easy-to-Replace Subnetwork create mode 100644 data/2022/neurips/Tree Mover's Distance: Bridging Graph Metrics and Stability of Graph Neural Networks create mode 100644 data/2022/neurips/Tree ensemble kernels for Bayesian optimization with known constraints over mixed-feature spaces create mode 100644 data/2022/neurips/TreeMoCo: Contrastive Neuron Morphology Representation Learning create mode 100644 data/2022/neurips/Triangulation candidates for Bayesian optimization create mode 100644 data/2022/neurips/Trimmed Maximum Likelihood Estimation for Robust Generalized Linear Model create mode 100644 data/2022/neurips/Truly Deterministic Policy Optimization create mode 100644 data/2022/neurips/Truncated Matrix Power Iteration for Differentiable DAG Learning create mode 100644 data/2022/neurips/Truncated proposals for scalable and hassle-free simulation-based inference create mode 100644 data/2022/neurips/Trust Region Policy Optimization with Optimal Transport Discrepancies: Duality and Algorithm for Continuous Actions create mode 100644 data/2022/neurips/Trustworthy Monte Carlo create mode 100644 data/2022/neurips/Tsetlin Machine for Solving Contextual Bandit Problems create mode 100644 data/2022/neurips/Turbocharging Solution Concepts: Solving NEs, CEs and CCEs with Neural Equilibrium Solvers create mode 100644 data/2022/neurips/Turning the Tables: Biased, Imbalanced, Dynamic Tabular Datasets for ML Evaluation create mode 100644 data/2022/neurips/TweetNERD - End to End Entity Linking Benchmark for Tweets create mode 100644 data/2022/neurips/TwiBot-22: Towards Graph-Based Twitter Bot Detection create mode 100644 data/2022/neurips/Two-Stream Network for Sign Language Recognition and Translation create mode 100644 data/2022/neurips/Two-layer neural network on infinite dimensional data: global optimization guarantee in the mean-field regime create mode 100644 data/2022/neurips/UDC: Unified DNAS for Compressible TinyML Models for Neural Processing Units create mode 100644 data/2022/neurips/ULNeF: Untangled Layered Neural Fields for Mix-and-Match Virtual Try-On create mode 100644 data/2022/neurips/UMIX: Improving Importance Weighting for Subpopulation Shift via Uncertainty-Aware Mixup create mode 100644 data/2022/neurips/UQGAN: A Unified Model for Uncertainty Quantification of Deep Classifiers trained via Conditional GANs create mode 100644 data/2022/neurips/USB: A Unified Semi-supervised Learning Benchmark for Classification create mode 100644 data/2022/neurips/UViM: A Unified Modeling Approach for Vision with Learned Guiding Codes create mode 100644 data/2022/neurips/Uncalibrated Models Can Improve Human-AI Collaboration create mode 100644 data/2022/neurips/Uncertainty Estimation Using Riemannian Model Dynamics for Offline Reinforcement Learning create mode 100644 data/2022/neurips/Uncertainty Estimation for Multi-view Data: The Power of Seeing the Whole Picture create mode 100644 data/2022/neurips/Uncertainty-Aware Hierarchical Refinement for Incremental Implicitly-Refined Classification create mode 100644 data/2022/neurips/Uncertainty-Aware Reinforcement Learning for Risk-Sensitive Player Evaluation in Sports Game create mode 100644 data/2022/neurips/Uncoupled Learning Dynamics with O(log T) Swap Regret in Multiplayer Games create mode 100644 data/2022/neurips/Uncovering the Structural Fairness in Graph Contrastive Learning create mode 100644 data/2022/neurips/Understanding Aesthetics with Language: A Photo Critique Dataset for Aesthetic Assessment create mode 100644 data/2022/neurips/Understanding Benign Overfitting in Gradient-Based Meta Learning create mode 100644 data/2022/neurips/Understanding Cross-Domain Few-Shot Learning Based on Domain Similarity and Few-Shot Difficulty create mode 100644 data/2022/neurips/Understanding Deep Contrastive Learning via Coordinate-wise Optimization create mode 100644 data/2022/neurips/Understanding Hyperdimensional Computing for Parallel Single-Pass Learning create mode 100644 data/2022/neurips/Understanding Non-linearity in Graph Neural Networks from the Bayesian-Inference Perspective create mode 100644 data/2022/neurips/Understanding Programmatic Weak Supervision via Source-aware Influence Function create mode 100644 data/2022/neurips/Understanding Robust Learning through the Lens of Representation Similarities create mode 100644 data/2022/neurips/Understanding Square Loss in Training Overparametrized Neural Network Classifiers create mode 100644 data/2022/neurips/Understanding and Extending Subgraph GNNs by Rethinking Their Symmetries create mode 100644 data/2022/neurips/Understanding and Improving Robustness of Vision Transformers through Patch-based Negative Augmentation create mode 100644 data/2022/neurips/Understanding the Eluder Dimension create mode 100644 data/2022/neurips/Understanding the Evolution of Linear Regions in Deep Reinforcement Learning create mode 100644 data/2022/neurips/Understanding the Failure of Batch Normalization for Transformers in NLP create mode 100644 data/2022/neurips/Understanding the Generalization Benefit of Normalization Layers: Sharpness Reduction create mode 100644 data/2022/neurips/UnfoldML: Cost-Aware and Uncertainty-Based Dynamic 2D Prediction for Multi-Stage Classification create mode 100644 data/2022/neurips/Uni-Perceiver-MoE: Learning Sparse Generalist Models with Conditional MoEs create mode 100644 data/2022/neurips/UniCLIP: Unified Framework for Contrastive Language-Image Pre-training create mode 100644 data/2022/neurips/UniGAN: Reducing Mode Collapse in GANs using a Uniform Generator create mode 100644 data/2022/neurips/Uni[MASK]: Unified Inference in Sequential Decision Problems create mode 100644 data/2022/neurips/Unified Optimal Transport Framework for Universal Domain Adaptation create mode 100644 data/2022/neurips/Unifying Voxel-based Representation with Transformer for 3D Object Detection create mode 100644 data/2022/neurips/Unifying and Boosting Gradient-Based Training-Free Neural Architecture Search create mode 100644 data/2022/neurips/Universal Rates for Interactive Learning create mode 100644 data/2022/neurips/Universality of Group Convolutional Neural Networks Based on Ridgelet Analysis on Groups create mode 100644 data/2022/neurips/Universally Expressive Communication in Multi-Agent Reinforcement Learning create mode 100644 data/2022/neurips/Unknown-Aware Domain Adversarial Learning for Open-Set Domain Adaptation create mode 100644 data/2022/neurips/Unlabelled Sample Compression Schemes for Intersection-Closed Classes and Extremal Classes create mode 100644 data/2022/neurips/Unpacking Reward Shaping: Understanding the Benefits of Reward Engineering on Sample Complexity create mode 100644 data/2022/neurips/Unravelling the Performance of Physics-informed Graph Neural Networks for Dynamical Systems create mode 100644 data/2022/neurips/Unsupervised Adaptation from Repeated Traversals for Autonomous Driving create mode 100644 data/2022/neurips/Unsupervised Causal Generative Understanding of Images create mode 100644 data/2022/neurips/Unsupervised Cross-Task Generalization via Retrieval Augmentation create mode 100644 data/2022/neurips/Unsupervised Domain Adaptation for Semantic Segmentation using Depth Distribution create mode 100644 data/2022/neurips/Unsupervised Image-to-Image Translation with Density Changing Regularization create mode 100644 data/2022/neurips/Unsupervised Learning From Incomplete Measurements for Inverse Problems create mode 100644 data/2022/neurips/Unsupervised Learning for Combinatorial Optimization with Principled Objective Relaxation create mode 100644 data/2022/neurips/Unsupervised Learning of Equivariant Structure from Sequences create mode 100644 data/2022/neurips/Unsupervised Learning of Group Invariant and Equivariant Representations create mode 100644 data/2022/neurips/Unsupervised Learning of Shape Programs with Repeatable Implicit Parts create mode 100644 data/2022/neurips/Unsupervised Learning under Latent Label Shift create mode 100644 data/2022/neurips/Unsupervised Multi-Object Segmentation by Predicting Probable Motion Patterns create mode 100644 data/2022/neurips/Unsupervised Multi-View Object Segmentation Using Radiance Field Propagation create mode 100644 data/2022/neurips/Unsupervised Object Detection Pretraining with Joint Object Priors Generation and Detector Learning create mode 100644 data/2022/neurips/Unsupervised Object Representation Learning using Translation and Rotation Group Equivariant VAE create mode 100644 data/2022/neurips/Unsupervised Point Cloud Completion and Segmentation by Generative Adversarial Autoencoding Network create mode 100644 data/2022/neurips/Unsupervised Reinforcement Learning with Contrastive Intrinsic Control create mode 100644 data/2022/neurips/Unsupervised Representation Learning from Pre-trained Diffusion Probabilistic Models create mode 100644 data/2022/neurips/Unsupervised Skill Discovery via Recurrent Skill Training create mode 100644 data/2022/neurips/Unsupervised Visual Representation Learning via Mutual Information Regularized Assignment create mode 100644 data/2022/neurips/Untargeted Backdoor Watermark: Towards Harmless and Stealthy Dataset Copyright Protection create mode 100644 data/2022/neurips/Uplifting Bandits create mode 100644 data/2022/neurips/Use-Case-Grounded Simulations for Explanation Evaluation create mode 100644 data/2022/neurips/Using Embeddings for Causal Estimation of Peer Influence in Social Networks create mode 100644 data/2022/neurips/Using Mixup as a Regularizer Can Surprisingly Improve Accuracy & Out-of-Distribution Robustness create mode 100644 data/2022/neurips/Using Partial Monotonicity in Submodular Maximization create mode 100644 data/2022/neurips/Using natural language and program abstractions to instill human inductive biases in machines create mode 100644 data/2022/neurips/VAEL: Bridging Variational Autoencoders and Probabilistic Logic Programming create mode 100644 data/2022/neurips/VCT: A Video Compression Transformer create mode 100644 data/2022/neurips/VER: Scaling On-Policy RL Leads to the Emergence of Navigation in Embodied Rearrangement create mode 100644 data/2022/neurips/VF-PS: How to Select Important Participants in Vertical Federated Learning, Efficiently and Securely? create mode 100644 data/2022/neurips/VICE: Variational Interpretable Concept Embeddings create mode 100644 data/2022/neurips/VICRegL: Self-Supervised Learning of Local Visual Features create mode 100644 data/2022/neurips/VITA: Video Instance Segmentation via Object Token Association create mode 100644 data/2022/neurips/VLMbench: A Compositional Benchmark for Vision-and-Language Manipulation create mode 100644 data/2022/neurips/VLMo: Unified Vision-Language Pre-Training with Mixture-of-Modality-Experts create mode 100644 data/2022/neurips/VRL3: A Data-Driven Framework for Visual Deep Reinforcement Learning create mode 100644 data/2022/neurips/VTC-LFC: Vision Transformer Compression with Low-Frequency Components create mode 100644 data/2022/neurips/VaiPhy: a Variational Inference Based Algorithm for Phylogeny create mode 100644 data/2022/neurips/Value Function Decomposition for Iterative Design of Reinforcement Learning Agents create mode 100644 data/2022/neurips/Variable-rate hierarchical CPC leads to acoustic unit discovery in speech create mode 100644 data/2022/neurips/Variance Reduced ProxSkip: Algorithm, Theory and Application to Federated Learning create mode 100644 data/2022/neurips/Variational Model Perturbation for Source-Free Domain Adaptation create mode 100644 data/2022/neurips/Variational inference via Wasserstein gradient flows create mode 100644 data/2022/neurips/VectorAdam for Rotation Equivariant Geometry Optimization create mode 100644 data/2022/neurips/VeriDark: A Large-Scale Benchmark for Authorship Verification on the Dark Web create mode 100644 data/2022/neurips/Verification and search algorithms for causal DAGs create mode 100644 data/2022/neurips/Versatile Multi-stage Graph Neural Network for Circuit Representation create mode 100644 data/2022/neurips/ViSioNS: Visual Search in Natural Scenes Benchmark create mode 100644 data/2022/neurips/ViTPose: Simple Vision Transformer Baselines for Human Pose Estimation create mode 100644 data/2022/neurips/Video Diffusion Models create mode 100644 data/2022/neurips/Video PreTraining (VPT): Learning to Act by Watching Unlabeled Online Videos create mode 100644 data/2022/neurips/Video compression dataset and benchmark of learning-based video-quality metrics create mode 100644 data/2022/neurips/Video-based Human-Object Interaction Detection from Tubelet Tokens create mode 100644 data/2022/neurips/VideoMAE: Masked Autoencoders are Data-Efficient Learners for Self-Supervised Video Pre-Training create mode 100644 data/2022/neurips/ViewFool: Evaluating the Robustness of Visual Recognition to Adversarial Viewpoints create mode 100644 data/2022/neurips/VisCo Grids: Surface Reconstruction with Viscosity and Coarea Grids create mode 100644 data/2022/neurips/VisFIS: Visual Feature Importance Supervision with Right-for-the-Right-Reason Objectives create mode 100644 data/2022/neurips/Vision GNN: An Image is Worth Graph of Nodes create mode 100644 data/2022/neurips/Vision Transformers provably learn spatial structure create mode 100644 data/2022/neurips/Visual Clues: Bridging Vision and Language Foundations for Image Paragraph Captioning create mode 100644 data/2022/neurips/Visual Concepts Tokenization create mode 100644 data/2022/neurips/Visual Prompting via Image Inpainting create mode 100644 data/2022/neurips/Visual correspondence-based explanations improve AI robustness and human-AI team accuracy create mode 100644 data/2022/neurips/VoiceBlock: Privacy through Real-Time Adversarial Attacks with Audio-to-Audio Models create mode 100644 data/2022/neurips/VoxGRAF: Fast 3D-Aware Image Synthesis with Sparse Voxel Grids create mode 100644 data/2022/neurips/WT-MVSNet: Window-based Transformers for Multi-view Stereo create mode 100644 data/2022/neurips/Washing The Unwashable : On The (Im)possibility of Fairwashing Detection create mode 100644 data/2022/neurips/Wasserstein $K$-means for clustering probability distributions create mode 100644 data/2022/neurips/Wasserstein Iterative Networks for Barycenter Estimation create mode 100644 data/2022/neurips/Wasserstein Logistic Regression with Mixed Features create mode 100644 data/2022/neurips/Watermarking for Out-of-distribution Detection create mode 100644 data/2022/neurips/WaveBound: Dynamic Error Bounds for Stable Time Series Forecasting create mode 100644 data/2022/neurips/Wavelet Feature Maps Compression for Image-to-Image CNNs create mode 100644 data/2022/neurips/Wavelet Score-Based Generative Modeling create mode 100644 data/2022/neurips/Weak-shot Semantic Segmentation via Dual Similarity Transfer create mode 100644 data/2022/neurips/Weakly Supervised Representation Learning with Sparse Perturbations create mode 100644 data/2022/neurips/Weakly supervised causal representation learning create mode 100644 data/2022/neurips/Weakly-Supervised Multi-Granularity Map Learning for Vision-and-Language Navigation create mode 100644 data/2022/neurips/WebShop: Towards Scalable Real-World Web Interaction with Grounded Language Agents create mode 100644 data/2022/neurips/Weighted Distillation with Unlabeled Examples create mode 100644 data/2022/neurips/Weighted Mutual Learning with Diversity-Driven Model Compression create mode 100644 data/2022/neurips/WeightedSHAP: analyzing and improving Shapley based feature attributions create mode 100644 data/2022/neurips/Weisfeiler and Leman Go Walking: Random Walk Kernels Revisited create mode 100644 data/2022/neurips/What Can Transformers Learn In-Context? A Case Study of Simple Function Classes create mode 100644 data/2022/neurips/What Can the Neural Tangent Kernel Tell Us About Adversarial Robustness? create mode 100644 data/2022/neurips/What I Cannot Predict, I Do Not Understand: A Human-Centered Evaluation Framework for Explainability Methods create mode 100644 data/2022/neurips/What Makes Graph Neural Networks Miscalibrated? create mode 100644 "data/2022/neurips/What Makes a \"Good\" Data Augmentation in Knowledge Distillation - A Statistical Perspective" create mode 100644 data/2022/neurips/What You See is What You Classify: Black Box Attributions create mode 100644 data/2022/neurips/What You See is What You Get: Principled Deep Learning via Distributional Generalization create mode 100644 data/2022/neurips/What are the best Systems? New Perspectives on NLP Benchmarking create mode 100644 data/2022/neurips/What is Where by Looking: Weakly-Supervised Open-World Phrase-Grounding without Text Inputs create mode 100644 data/2022/neurips/What is a Good Metric to Study Generalization of Minimax Learners? create mode 100644 data/2022/neurips/What's the Harm? Sharp Bounds on the Fraction Negatively Affected by Treatment create mode 100644 data/2022/neurips/When Adversarial Training Meets Vision Transformers: Recipes from Training to Architecture create mode 100644 data/2022/neurips/When Combinatorial Thompson Sampling meets Approximation Regret create mode 100644 data/2022/neurips/When Do Flat Minima Optimizers Work? create mode 100644 data/2022/neurips/When Does Differentially Private Learning Not Suffer in High Dimensions? create mode 100644 data/2022/neurips/When Does Group Invariant Learning Survive Spurious Correlations? create mode 100644 data/2022/neurips/When Privacy Meets Partial Information: A Refined Analysis of Differentially Private Bandits create mode 100644 data/2022/neurips/When are Local Queries Useful for Robust Learning? create mode 100644 data/2022/neurips/When are Offline Two-Player Zero-Sum Markov Games Solvable? create mode 100644 data/2022/neurips/When does dough become a bagel? Analyzing the remaining mistakes on ImageNet create mode 100644 data/2022/neurips/When does return-conditioned supervised learning work for offline reinforcement learning? create mode 100644 data/2022/neurips/When to Ask for Help: Proactive Interventions in Autonomous Reinforcement Learning create mode 100644 data/2022/neurips/When to Intervene: Learning Optimal Intervention Policies for Critical Events create mode 100644 data/2022/neurips/When to Make Exceptions: Exploring Language Models as Accounts of Human Moral Judgment create mode 100644 data/2022/neurips/When to Trust Your Simulator: Dynamics-Aware Hybrid Offline-and-Online Reinforcement Learning create mode 100644 data/2022/neurips/When to Update Your Model: Constrained Model-based Reinforcement Learning create mode 100644 data/2022/neurips/Where do Models go Wrong? Parameter-Space Saliency Maps for Explainability create mode 100644 data/2022/neurips/Where to Pay Attention in Sparse Training for Feature Selection? create mode 100644 data/2022/neurips/Where2comm: Communication-Efficient Collaborative Perception via Spatial Confidence Maps create mode 100644 data/2022/neurips/Which Explanation Should I Choose? A Function Approximation Perspective to Characterizing Post Hoc Explanations create mode 100644 data/2022/neurips/Whitening Convergence Rate of Coupling-based Normalizing Flows create mode 100644 data/2022/neurips/Why Do Artificially Generated Data Help Adversarial Robustness create mode 100644 data/2022/neurips/Why Robust Generalization in Deep Learning is Difficult: Perspective of Expressive Power create mode 100644 data/2022/neurips/Why So Pessimistic? Estimating Uncertainties for Offline RL through Ensembles, and Why Their Independence Matters create mode 100644 data/2022/neurips/Why do We Need Large Batchsizes in Contrastive Learning? A Gradient-Bias Perspective create mode 100644 data/2022/neurips/Why do tree-based models still outperform deep learning on typical tabular data? create mode 100644 data/2022/neurips/Why neural networks find simple solutions: The many regularizers of geometric complexity create mode 100644 data/2022/neurips/Wild-Time: A Benchmark of in-the-Wild Distribution Shift over Time create mode 100644 data/2022/neurips/Will Bilevel Optimizers Benefit from Loops create mode 100644 data/2022/neurips/WinoGAViL: Gamified Association Benchmark to Challenge Vision-and-Language Models create mode 100644 data/2022/neurips/Wukong: A 100 Million Large-scale Chinese Cross-modal Pre-training Benchmark create mode 100644 data/2022/neurips/XTC: Extreme Compression for Pre-trained Transformers Made Simple and Efficient create mode 100644 data/2022/neurips/You Can't Count on Luck: Why Decision Transformers and RvS Fail in Stochastic Environments create mode 100644 data/2022/neurips/You Never Stop Dancing: Non-freezing Dance Generation via Bank-constrained Manifold Projection create mode 100644 data/2022/neurips/You Only Live Once: Single-Life Reinforcement Learning create mode 100644 data/2022/neurips/Your Out-of-Distribution Detection Method is Not Robust! create mode 100644 data/2022/neurips/Your Transformer May Not be as Powerful as You Expect create mode 100644 data/2022/neurips/ZARTS: On Zero-order Optimization for Neural Architecture Search create mode 100644 data/2022/neurips/ZIN: When and How to Learn Invariance Without Environment Partition? create mode 100644 data/2022/neurips/ZSON: Zero-Shot Object-Goal Navigation using Multimodal Goal Embeddings create mode 100644 data/2022/neurips/Zero-Shot 3D Drug Design by Sketching and Generating create mode 100644 data/2022/neurips/Zero-Shot Video Question Answering via Frozen Bidirectional Language Models create mode 100644 data/2022/neurips/Zero-Sum Stochastic Stackelberg Games create mode 100644 data/2022/neurips/Zero-shot Transfer Learning within a Heterogeneous Graph via Knowledge Transfer Networks create mode 100644 data/2022/neurips/ZeroC: A Neuro-Symbolic Model for Zero-shot Concept Recognition and Acquisition at Inference Time create mode 100644 data/2022/neurips/ZeroQuant: Efficient and Affordable Post-Training Quantization for Large-Scale Transformers create mode 100644 data/2022/neurips/Zeroth-Order Hard-Thresholding: Gradient Error vs. Expansivity create mode 100644 data/2022/neurips/Zeroth-Order Negative Curvature Finding: Escaping Saddle Points without Gradients create mode 100644 data/2022/neurips/Zonotope Domains for Lagrangian Neural Network Verification create mode 100644 data/2022/neurips/ZooD: Exploiting Model Zoo for Out-of-Distribution Generalization create mode 100644 data/2022/neurips/coVariance Neural Networks create mode 100644 data/2022/neurips/mRI: Multi-modal 3D Human Pose Estimation Dataset using mmWave, RGB-D, and Inertial Sensors create mode 100644 data/2022/neurips/pFL-Bench: A Comprehensive Benchmark for Personalized Federated Learning create mode 100644 data/2022/neurips/projUNN: efficient method for training deep networks with unitary matrices create mode 100644 data/2022/neurips/pyKT: A Python Library to Benchmark Deep Learning based Knowledge Tracing Models create mode 100644 data/2022/neurips/u-HuBERT: Unified Mixed-Modal Speech Pretraining And Zero-Shot Transfer to Unlabeled Modality create mode 100644 data/2022/neurips/xView3-SAR: Detecting Dark Fishing Activity Using Synthetic Aperture Radar Imagery create mode 100644 "data/2022/neurips/\360\237\217\230\357\270\217 ProcTHOR: Large-Scale Embodied AI Using Procedural Generation" create mode 100644 "data/2023/neurips/\"Why Not Looking backward?\" A Robust Two-Step Method to Automatically Terminate Bayesian Optimization" create mode 100644 data/2023/neurips/(Amplified) Banded Matrix Factorization: A unified approach to private training create mode 100644 data/2023/neurips/2Direction: Theoretically Faster Distributed Training with Bidirectional Communication Compression create mode 100644 data/2023/neurips/3D molecule generation by denoising voxel grids create mode 100644 data/2023/neurips/3D-LLM: Injecting the 3D World into Large Language Models create mode 100644 data/2023/neurips/4D Panoptic Scene Graph Generation create mode 100644 data/2023/neurips/A Bayesian Approach To Analysing Training Data Attribution In Deep Learning create mode 100644 data/2023/neurips/A Bounded Ability Estimation for Computerized Adaptive Testing create mode 100644 data/2023/neurips/A Cross-Moment Approach for Causal Effect Estimation create mode 100644 data/2023/neurips/A Data-Free Approach to Mitigate Catastrophic Forgetting in Federated Class Incremental Learning for Vision Tasks create mode 100644 data/2023/neurips/A Dataset for Analyzing Streaming Media Performance over HTTP 3 Browsers create mode 100644 data/2023/neurips/A Diffusion-Model of Joint Interactive Navigation create mode 100644 data/2023/neurips/A Dynamical System View of Langevin-Based Non-Convex Sampling create mode 100644 data/2023/neurips/A Finite-Sample Analysis of Payoff-Based Independent Learning in Zero-Sum Stochastic Games create mode 100644 data/2023/neurips/A Framework for Fast and Stable Representations of Multiparameter Persistent Homology Decompositions create mode 100644 data/2023/neurips/A General Framework for Robust G-Invariance in G-Equivariant Networks create mode 100644 data/2023/neurips/A General Theory of Correct, Incorrect, and Extrinsic Equivariance create mode 100644 data/2023/neurips/A Holistic Approach to Unifying Automatic Concept Extraction and Concept Importance Estimation create mode 100644 data/2023/neurips/A Massive Scale Semantic Similarity Dataset of Historical English create mode 100644 data/2023/neurips/A Measure-Theoretic Axiomatisation of Causality create mode 100644 data/2023/neurips/A Regularized Conditional GAN for Posterior Sampling in Image Recovery Problems create mode 100644 data/2023/neurips/A Riemannian Exponential Augmented Lagrangian Method for Computing the Projection Robust Wasserstein Distance create mode 100644 data/2023/neurips/A Robust Exact Algorithm for the Euclidean Bipartite Matching Problem create mode 100644 data/2023/neurips/A Scalable Neural Network for DSIC Affine Maximizer Auction Design create mode 100644 data/2023/neurips/A Scale-Invariant Sorting Criterion to Find a Causal Order in Additive Noise Models create mode 100644 data/2023/neurips/A Single-Loop Accelerated Extra-Gradient Difference Algorithm with Improved Complexity Bounds for Constrained Minimax Optimization create mode 100644 data/2023/neurips/A Spectral Algorithm for List-Decodable Covariance Estimation in Relative Frobenius Norm create mode 100644 data/2023/neurips/A Theoretical Analysis of Optimistic Proximal Policy Optimization in Linear Markov Decision Processes create mode 100644 data/2023/neurips/A Unified Algorithm Framework for Unsupervised Discovery of Skills based on Determinantal Point Process create mode 100644 data/2023/neurips/A Unified Framework for Uniform Signal Recovery in Nonlinear Generative Compressed Sensing create mode 100644 data/2023/neurips/A Unified Model and Dimension for Interactive Estimation create mode 100644 data/2023/neurips/A Unified, Scalable Framework for Neural Population Decoding create mode 100644 data/2023/neurips/A Unifying Perspective on Multi-Calibration: Game Dynamics for Multi-Objective Learning create mode 100644 data/2023/neurips/A fast heuristic to optimize time-space tradeoff for large models create mode 100644 data/2023/neurips/A graphon-signal analysis of graph neural networks create mode 100644 data/2023/neurips/A new perspective on building efficient and expressive 3D equivariant graph neural networks create mode 100644 data/2023/neurips/A polar prediction model for learning to represent visual transformations create mode 100644 data/2023/neurips/A unified framework for information-theoretic generalization bounds create mode 100644 data/2023/neurips/ADGym: Design Choices for Deep Anomaly Detection create mode 100644 data/2023/neurips/AGD: an Auto-switchable Optimizer using Stepwise Gradient Difference for Preconditioning Matrix create mode 100644 data/2023/neurips/AMDP: An Adaptive Detection Procedure for False Discovery Rate Control in High-Dimensional Mediation Analysis create mode 100644 data/2023/neurips/ANPL: Towards Natural Programming with Interactive Decomposition create mode 100644 data/2023/neurips/ANTN: Bridging Autoregressive Neural Networks and Tensor Networks for Quantum Many-Body Simulation create mode 100644 data/2023/neurips/AQuA: A Benchmarking Tool for Label Quality Assessment create mode 100644 data/2023/neurips/AR-Diffusion: Auto-Regressive Diffusion Model for Text Generation create mode 100644 data/2023/neurips/ASL Citizen: A Community-Sourced Dataset for Advancing Isolated Sign Language Recognition create mode 100644 data/2023/neurips/ASPEN: Breaking Operator Barriers for Efficient Parallelization of Deep Neural Networks create mode 100644 data/2023/neurips/ATMAN: Understanding Transformer Predictions Through Memory Efficient Attention Manipulation create mode 100644 data/2023/neurips/AUDIT: Audio Editing by Following Instructions with Latent Diffusion Models create mode 100644 data/2023/neurips/AVIS: Autonomous Visual Information Seeking with Large Language Model Agent create mode 100644 data/2023/neurips/AVeriTeC: A Dataset for Real-world Claim Verification with Evidence from the Web create mode 100644 data/2023/neurips/AbDiffuser: full-atom generation of in-vitro functioning antibodies create mode 100644 data/2023/neurips/Accelerated On-Device Forward Neural Network Training with Module-Wise Descending Asynchronism create mode 100644 data/2023/neurips/Accelerated Zeroth-order Method for Non-Smooth Stochastic Convex Optimization Problem with Infinite Variance create mode 100644 data/2023/neurips/Accelerating Motion Planning via Optimal Transport create mode 100644 data/2023/neurips/Accessing Higher Dimensions for Unsupervised Word Translation create mode 100644 data/2023/neurips/Achieving Cross Modal Generalization with Multimodal Unified Representation create mode 100644 data/2023/neurips/Action Inference by Maximising Evidence: Zero-Shot Imitation from Observation with World Models create mode 100644 data/2023/neurips/Active Vision Reinforcement Learning under Limited Visual Observability create mode 100644 data/2023/neurips/Active representation learning for general task space with applications in robotics create mode 100644 data/2023/neurips/Activity Grammars for Temporal Action Segmentation create mode 100644 data/2023/neurips/AdANNS: A Framework for Adaptive Semantic Search create mode 100644 data/2023/neurips/Adapting Neural Link Predictors for Data-Efficient Complex Query Answering create mode 100644 data/2023/neurips/Adaptive Principal Component Regression with Applications to Panel Data create mode 100644 data/2023/neurips/Adaptive Privacy Composition for Accuracy-first Mechanisms create mode 100644 data/2023/neurips/Adaptive Test-Time Personalization for Federated Learning create mode 100644 data/2023/neurips/Adaptive recurrent vision performs zero-shot computation scaling to unseen difficulty levels create mode 100644 data/2023/neurips/Add and Thin: Diffusion for Temporal Point Processes create mode 100644 data/2023/neurips/Addressing Negative Transfer in Diffusion Models create mode 100644 data/2023/neurips/Adversarial Counterfactual Environment Model Learning create mode 100644 data/2023/neurips/Adversarial Examples Exist in Two-Layer ReLU Networks for Low Dimensional Linear Subspaces create mode 100644 data/2023/neurips/Adversarial Self-Training Improves Robustness and Generalization for Gradual Domain Adaptation create mode 100644 data/2023/neurips/Adversarial Training from Mean Field Perspective create mode 100644 data/2023/neurips/Adversarially Robust Distributed Count Tracking via Partial Differential Privacy create mode 100644 data/2023/neurips/Advice Querying under Budget Constraint for Online Algorithms create mode 100644 data/2023/neurips/Affinity-Aware Graph Networks create mode 100644 data/2023/neurips/AirDelhi: Fine-Grained Spatio-Temporal Particulate Matter Dataset From Delhi For ML based Modeling create mode 100644 data/2023/neurips/AlberDICE: Addressing Out-Of-Distribution Joint Actions in Offline Multi-Agent RL via Alternating Stationary Distribution Correction Estimation create mode 100644 data/2023/neurips/Align Your Prompts: Test-Time Prompting with Distribution Alignment for Zero-Shot Generalization create mode 100644 data/2023/neurips/Aligning Gradient and Hessian for Neural Signed Distance Function create mode 100644 data/2023/neurips/Aligning Language Models with Human Preferences via a Bayesian Approach create mode 100644 data/2023/neurips/Alignment with human representations supports robust few-shot learning create mode 100644 data/2023/neurips/All Points Matter: Entropy-Regularized Distribution Alignment for Weakly-supervised 3D Segmentation create mode 100644 data/2023/neurips/Alternating Gradient Descent and Mixture-of-Experts for Integrated Multimodal Perception create mode 100644 data/2023/neurips/Alternating Updates for Efficient Transformers create mode 100644 data/2023/neurips/Alternation makes the adversary weaker in two-player games create mode 100644 data/2023/neurips/American Stories: A Large-Scale Structured Text Dataset of Historical U.S. Newspapers create mode 100644 data/2023/neurips/Amortized Reparametrization: Efficient and Scalable Variational Inference for Latent SDEs create mode 100644 "data/2023/neurips/An Alternating Optimization Method for Bilevel Problems under the Polyak-\305\201ojasiewicz Condition" create mode 100644 data/2023/neurips/An Efficient Dataset Condensation Plugin and Its Application to Continual Learning create mode 100644 data/2023/neurips/An Exploration-by-Optimization Approach to Best of Both Worlds in Linear Bandits create mode 100644 data/2023/neurips/An Improved Relaxation for Oracle-Efficient Adversarial Contextual Bandits create mode 100644 data/2023/neurips/An Optimization-based Approach To Node Role Discovery in Networks: Approximating Equitable Partitions create mode 100644 data/2023/neurips/An information-theoretic quantification of the content of communication between brain regions create mode 100644 data/2023/neurips/Analyzing Generalization of Neural Networks through Loss Path Kernels create mode 100644 data/2023/neurips/Analyzing the Sample Complexity of Self-Supervised Image Reconstruction Methods create mode 100644 data/2023/neurips/Anchor Data Augmentation create mode 100644 data/2023/neurips/Anonymous and Copy-Robust Delegations for Liquid Democracy create mode 100644 data/2023/neurips/Anytime Model Selection in Linear Bandits create mode 100644 data/2023/neurips/Anytime-Competitive Reinforcement Learning with Policy Prior create mode 100644 data/2023/neurips/Approximate Allocation Matching for Structural Causal Bandits with Unobserved Confounders create mode 100644 data/2023/neurips/Approximate inference of marginals using the IBIA framework create mode 100644 data/2023/neurips/Are Diffusion Models Vision-And-Language Reasoners? create mode 100644 data/2023/neurips/Are Vision Transformers More Data Hungry Than Newborn Visual Systems? create mode 100644 data/2023/neurips/Assessor360: Multi-sequence Network for Blind Omnidirectional Image Quality Assessment create mode 100644 data/2023/neurips/Auditing for Human Expertise create mode 100644 data/2023/neurips/Augmenting Language Models with Long-Term Memory create mode 100644 data/2023/neurips/Auslan-Daily: Australian Sign Language Translation for Daily Communication and News create mode 100644 data/2023/neurips/AutoGO: Automated Computation Graph Optimization for Neural Network Evolution create mode 100644 data/2023/neurips/Autodecoding Latent 3D Diffusion Models create mode 100644 data/2023/neurips/BEDD: The MineRL BASALT Evaluation and Demonstrations Dataset for Training and Benchmarking Agents that Solve Fuzzy Tasks create mode 100644 data/2023/neurips/BIOT: Biosignal Transformer for Cross-data Learning in the Wild create mode 100644 data/2023/neurips/BQ-NCO: Bisimulation Quotienting for Efficient Neural Combinatorial Optimization create mode 100644 data/2023/neurips/Back-Modality: Leveraging Modal Transformation for Data Augmentation create mode 100644 data/2023/neurips/Balancing Risk and Reward: A Batched-Bandit Strategy for Automated Phased Release create mode 100644 data/2023/neurips/Banana: Banach Fixed-Point Network for Pointcloud Segmentation with Inter-Part Equivariance create mode 100644 data/2023/neurips/Bandit Task Assignment with Unknown Processing Time create mode 100644 data/2023/neurips/BanditPAM++: Faster k-medoids Clustering create mode 100644 data/2023/neurips/BasisFormer: Attention-based Time Series Forecasting with Learnable and Interpretable Basis create mode 100644 data/2023/neurips/BayesTune: Bayesian Sparse Deep Model Fine-tuning create mode 100644 data/2023/neurips/Bayesian Active Causal Discovery with Multi-Fidelity Experiments create mode 100644 data/2023/neurips/Bayesian Learning via Q-Exponential Process create mode 100644 data/2023/neurips/Bayesian Metric Learning for Uncertainty Quantification in Image Retrieval create mode 100644 data/2023/neurips/Bayesian Risk-Averse Q-Learning with Streaming Observations create mode 100644 data/2023/neurips/Bayesian nonparametric (non-)renewal processes for analyzing neural spike train variability create mode 100644 data/2023/neurips/BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset create mode 100644 data/2023/neurips/Benchmarking Foundation Models with Language-Model-as-an-Examiner create mode 100644 data/2023/neurips/Better Private Linear Regression Through Better Private Feature Selection create mode 100644 data/2023/neurips/Beyond Exponential Graph: Communication-Efficient Topologies for Decentralized Learning via Finite-time Convergence create mode 100644 data/2023/neurips/Beyond Myopia: Learning from Positive and Unlabeled Data through Holistic Predictive Trends create mode 100644 data/2023/neurips/Bias in Evaluation Processes: An Optimization-Based Model create mode 100644 data/2023/neurips/Bicriteria Approximation Algorithms for the Submodular Cover Problem create mode 100644 data/2023/neurips/Bicriteria Multidimensional Mechanism Design with Side Information create mode 100644 data/2023/neurips/Bifurcations and loss jumps in RNN training create mode 100644 data/2023/neurips/BioMassters: A Benchmark Dataset for Forest Biomass Estimation using Multi-modal Satellite Time-series create mode 100644 data/2023/neurips/Bitstream-Corrupted Video Recovery: A Novel Benchmark Dataset and Method create mode 100644 data/2023/neurips/Black-Box Differential Privacy for Interactive ML create mode 100644 data/2023/neurips/Block Coordinate Plug-and-Play Methods for Blind Inverse Problems create mode 100644 data/2023/neurips/Block-State Transformers create mode 100644 data/2023/neurips/Boosting Adversarial Transferability by Achieving Flat Local Maxima create mode 100644 data/2023/neurips/Boosting Spectral Clustering on Incomplete Data via Kernel Correction and Affinity Learning create mode 100644 data/2023/neurips/Bootstrapped Training of Score-Conditioned Generator for Offline Design of Biological Sequences create mode 100644 data/2023/neurips/Boundary Guided Learning-Free Semantic Control with Diffusion Models create mode 100644 data/2023/neurips/Bounding training data reconstruction in DP-SGD create mode 100644 data/2023/neurips/Brain Diffusion for Visual Exploration: Cortical Discovery using Large Scale Generative Models create mode 100644 data/2023/neurips/Breaking the Communication-Privacy-Accuracy Tradeoff with f-Differential Privacy create mode 100644 data/2023/neurips/Bridging the Domain Gap: Self-Supervised 3D Scene Understanding with Foundation Models create mode 100644 data/2023/neurips/Bypassing spike sorting: Density-based decoding using spike localization from dense multielectrode probes create mode 100644 data/2023/neurips/Bypassing the Simulator: Near-Optimal Adversarial Linear Contextual Bandits create mode 100644 data/2023/neurips/C-Disentanglement: Discovering Causally-Independent Generative Factors under an Inductive Bias of Confounder create mode 100644 data/2023/neurips/CAP: Correlation-Aware Pruning for Highly-Accurate Sparse Vision Models create mode 100644 data/2023/neurips/CARE: Modeling Interacting Dynamics Under Temporal Environmental Variation create mode 100644 data/2023/neurips/CAST: Cross-Attention in Space and Time for Video Action Recognition create mode 100644 data/2023/neurips/CEIL: Generalized Contextual Imitation Learning create mode 100644 data/2023/neurips/CHAMMI: A benchmark for channel-adaptive models in microscopy imaging create mode 100644 data/2023/neurips/COCO-Counterfactuals: Automatically Constructed Counterfactual Examples for Image-Text Pairs create mode 100644 data/2023/neurips/COOM: A Game Benchmark for Continual Reinforcement Learning create mode 100644 data/2023/neurips/CORL: Research-oriented Deep Offline Reinforcement Learning Library create mode 100644 data/2023/neurips/CQM: Curriculum Reinforcement Learning with a Quantized World Model create mode 100644 data/2023/neurips/CRoSS: Diffusion Model Makes Controllable, Robust and Secure Image Steganography create mode 100644 data/2023/neurips/CWCL: Cross-Modal Transfer with Continuously Weighted Contrastive Loss create mode 100644 data/2023/neurips/Cal-DETR: Calibrated Detection Transformer create mode 100644 data/2023/neurips/Calibrate and Boost Logical Expressiveness of GNN Over Multi-Relational and Temporal Graphs create mode 100644 data/2023/neurips/Calibration by Distribution Matching: Trainable Kernel Calibration Metrics create mode 100644 data/2023/neurips/CamoPatch: An Evolutionary Strategy for Generating Camoflauged Adversarial Patches create mode 100644 data/2023/neurips/Cascading Bandits: Optimizing Recommendation Frequency in Delayed Feedback Environments create mode 100644 data/2023/neurips/Causal discovery from observational and interventional data across multiple environments create mode 100644 data/2023/neurips/Causal-structure Driven Augmentations for Text OOD Generalization create mode 100644 data/2023/neurips/Characterization and Learning of Causal Graphs with Small Conditioning Sets create mode 100644 data/2023/neurips/Characterization of Overfitting in Robust Multiclass Classification create mode 100644 data/2023/neurips/Chasing Fairness Under Distribution Shift: A Model Weight Perturbation Approach create mode 100644 data/2023/neurips/ChatGPT-Powered Hierarchical Comparisons for Image Classification create mode 100644 data/2023/neurips/Cheap and Quick: Efficient Vision-Language Instruction Tuning for Large Language Models create mode 100644 data/2023/neurips/Cheaply Estimating Inference Efficiency Metrics for Autoregressive Transformer Models create mode 100644 data/2023/neurips/CityRefer: Geography-aware 3D Visual Grounding Dataset on City-scale Point Cloud Data create mode 100644 data/2023/neurips/Class-Conditional Conformal Prediction with Many Classes create mode 100644 data/2023/neurips/Classical Simulation of Quantum Circuits: Parallel Environments and Benchmark create mode 100644 data/2023/neurips/ClimateLearn: Benchmarking Machine Learning for Weather and Climate Modeling create mode 100644 data/2023/neurips/ClusterFomer: Clustering As A Universal Visual Learner create mode 100644 data/2023/neurips/Clustering the Sketch: Dynamic Compression for Embedding Tables create mode 100644 data/2023/neurips/CoDA: Collaborative Novel Box Discovery and Cross-modal Alignment for Open-vocabulary 3D Object Detection create mode 100644 data/2023/neurips/CoDet: Co-occurrence Guided Region-Word Alignment for Open-Vocabulary Object Detection create mode 100644 data/2023/neurips/CoPriv: Network Protocol Co-Optimization for Communication-Efficient Private Inference create mode 100644 data/2023/neurips/Cognitive Steering in Deep Neural Networks via Long-Range Modulatory Feedback Connections create mode 100644 data/2023/neurips/Collaborative Alignment of NLP Models create mode 100644 data/2023/neurips/Collaborative Learning via Prediction Consensus create mode 100644 data/2023/neurips/Collaborative Score Distillation for Consistent Visual Editing create mode 100644 data/2023/neurips/Collaboratively Learning Linear Models with Structured Missing Data create mode 100644 data/2023/neurips/Collapsed Inference for Bayesian Deep Learning create mode 100644 data/2023/neurips/Compact Neural Volumetric Video Representations with Dynamic Codebooks create mode 100644 data/2023/neurips/Comparing Causal Frameworks: Potential Outcomes, Structural Models, Graphs, and Abstractions create mode 100644 data/2023/neurips/Complementary Benefits of Contrastive Learning and Self-Training Under Distribution Shift create mode 100644 data/2023/neurips/Computational Complexity of Learning Neural Networks: Smoothness and Degeneracy create mode 100644 data/2023/neurips/Concept Algebra for (Score-Based) Text-Controlled Generative Models create mode 100644 data/2023/neurips/Concept Distillation: Leveraging Human-Centered Explanations for Model Improvement create mode 100644 data/2023/neurips/Conditional Adapters: Parameter-efficient Transfer Learning with Fast Inference create mode 100644 data/2023/neurips/Conditional Mutual Information for Disentangled Representations in Reinforcement Learning create mode 100644 data/2023/neurips/Conformal Prediction for Uncertainty-Aware Planning with Diffusion Dynamics Model create mode 100644 data/2023/neurips/Connected Superlevel Set in (Deep) Reinforcement Learning and its Application to Minimax Theorems create mode 100644 data/2023/neurips/Connecting Certified and Adversarial Training create mode 100644 data/2023/neurips/Conservative State Value Estimation for Offline Reinforcement Learning create mode 100644 data/2023/neurips/Constructing Non-isotropic Gaussian Diffusion Model Using Isotropic Gaussian Diffusion Model for Image Editing create mode 100644 data/2023/neurips/Context Shift Reduction for Offline Meta-Reinforcement Learning create mode 100644 data/2023/neurips/Context-guided Embedding Adaptation for Effective Topic Modeling in Low-Resource Regimes create mode 100644 data/2023/neurips/Context-lumpable stochastic bandits create mode 100644 data/2023/neurips/Contextual Bandits and Imitation Learning with Preference-Based Active Queries create mode 100644 data/2023/neurips/Contextual Stochastic Bilevel Optimization create mode 100644 data/2023/neurips/Continuous-Time Functional Diffusion Processes create mode 100644 data/2023/neurips/Contrastive Moments: Unsupervised Halfspace Learning in Polynomial Time create mode 100644 data/2023/neurips/Contrastive Sampling Chains in Diffusion Models create mode 100644 data/2023/neurips/Controlling Text-to-Image Diffusion by Orthogonal Finetuning create mode 100644 data/2023/neurips/Convex-Concave Zero-Sum Stochastic Stackelberg Games create mode 100644 data/2023/neurips/Convolutional Neural Operators for robust and accurate learning of PDEs create mode 100644 data/2023/neurips/Convolutional State Space Models for Long-Range Spatiotemporal Modeling create mode 100644 data/2023/neurips/Core-sets for Fair and Diverse Data Summarization create mode 100644 data/2023/neurips/Correlative Information Maximization: A Biologically Plausible Approach to Supervised Deep Neural Networks without Weight Symmetry create mode 100644 data/2023/neurips/CorresNeRF: Image Correspondence Priors for Neural Radiance Fields create mode 100644 data/2023/neurips/Counterfactual Conservative Q Learning for Offline Multi-agent Reinforcement Learning create mode 100644 data/2023/neurips/Counterfactual Evaluation of Peer-Review Assignment Policies create mode 100644 data/2023/neurips/Coupled Reconstruction of Cortical Surfaces by Diffeomorphic Mesh Deformation create mode 100644 data/2023/neurips/Covariance-adaptive best arm identification create mode 100644 data/2023/neurips/Creating Multi-Level Skill Hierarchies in Reinforcement Learning create mode 100644 data/2023/neurips/Creating a Public Repository for Joining Private Data create mode 100644 data/2023/neurips/Cross-Domain Policy Adaptation via Value-Guided Data Filtering create mode 100644 data/2023/neurips/Cross-links Matter for Link Prediction: Rethinking the Debiased GNN from a Data Perspective create mode 100644 data/2023/neurips/D4Explainer: In-distribution Explanations of Graph Neural Network via Discrete Denoising Diffusion create mode 100644 data/2023/neurips/DAC-DETR: Divide the Attention Layers and Conquer create mode 100644 data/2023/neurips/DAMEX: Dataset-aware Mixture-of-Experts for visual understanding of mixture-of-datasets create mode 100644 data/2023/neurips/DASpeech: Directed Acyclic Transformer for Fast and High-quality Speech-to-Speech Translation create mode 100644 data/2023/neurips/DESSERT: An Efficient Algorithm for Vector Set Search with Vector Set Queries create mode 100644 data/2023/neurips/DIFFER: Decomposing Individual Reward for Fair Experience Replay in Multi-Agent Reinforcement Learning create mode 100644 data/2023/neurips/DIFUSCO: Graph-based Diffusion Solvers for Combinatorial Optimization create mode 100644 data/2023/neurips/DISCS: A Benchmark for Discrete Sampling create mode 100644 data/2023/neurips/DaTaSeg: Taming a Universal Multi-Dataset Multi-Task Segmentation Model create mode 100644 data/2023/neurips/Data Quality in Imitation Learning create mode 100644 data/2023/neurips/Data Selection for Language Models via Importance Resampling create mode 100644 data/2023/neurips/Data-driven Optimal Filtering for Linear Systems with Unknown Noise Covariances create mode 100644 data/2023/neurips/Dataset Diffusion: Diffusion-based Synthetic Data Generation for Pixel-Level Semantic Segmentation create mode 100644 data/2023/neurips/Debiasing Conditional Stochastic Optimization create mode 100644 data/2023/neurips/Decentralized Randomly Distributed Multi-agent Multi-armed Bandit with Heterogeneous Rewards create mode 100644 data/2023/neurips/Decision Stacks: Flexible Reinforcement Learning via Modular Generative Models create mode 100644 data/2023/neurips/Decoding the Enigma: Benchmarking Humans and AIs on the Many Facets of Working Memory create mode 100644 data/2023/neurips/Decompose Novel into Known: Part Concept Learning For 3D Novel Class Discovery create mode 100644 data/2023/neurips/Decompose a Task into Generalizable Subtasks in Multi-Agent Reinforcement Learning create mode 100644 data/2023/neurips/Deep Contract Design via Discontinuous Networks create mode 100644 data/2023/neurips/Deep Fractional Fourier Transform create mode 100644 data/2023/neurips/Deep Gaussian Markov Random Fields for Graph-Structured Dynamical Systems create mode 100644 data/2023/neurips/Deep Insights into Noisy Pseudo Labeling on Graph Data create mode 100644 data/2023/neurips/Deep Neural Collapse Is Provably Optimal for the Deep Unconstrained Features Model create mode 100644 data/2023/neurips/DeepACO: Neural-enhanced Ant Systems for Combinatorial Optimization create mode 100644 data/2023/neurips/DeepPCR: Parallelizing Sequential Operations in Neural Networks create mode 100644 data/2023/neurips/DeepSimHO: Stable Pose Estimation for Hand-Object Interaction via Physics Simulation create mode 100644 data/2023/neurips/Delegated Classification create mode 100644 data/2023/neurips/Demystifying Structural Disparity in Graph Neural Networks: Can One Size Fit All? create mode 100644 data/2023/neurips/Dense and Aligned Captions (DAC) Promote Compositional Reasoning in VL Models create mode 100644 data/2023/neurips/Depth-discriminative Metric Learning for Monocular 3D Object Detection create mode 100644 data/2023/neurips/Derandomized novelty detection with FDR control via conformal e-values create mode 100644 data/2023/neurips/Described Object Detection: Liberating Object Detection with Flexible Expressions create mode 100644 data/2023/neurips/DiT-3D: Exploring Plain Diffusion Transformers for 3D Shape Generation create mode 100644 data/2023/neurips/DiViNeT: 3D Reconstruction from Disparate Views using Neural Template Regularization create mode 100644 data/2023/neurips/Diff-Foley: Synchronized Video-to-Audio Synthesis with Latent Diffusion Models create mode 100644 data/2023/neurips/Diff-Instruct: A Universal Approach for Transferring Knowledge From Pre-trained Diffusion Models create mode 100644 data/2023/neurips/DiffAttack: Evasion Attacks Against Diffusion-Based Adversarial Purification create mode 100644 data/2023/neurips/DiffComplete: Diffusion-based Generative 3D Shape Completion create mode 100644 data/2023/neurips/DiffInfinite: Large Mask-Image Synthesis via Parallel Random Patch Diffusion in Histopathology create mode 100644 data/2023/neurips/DiffTraj: Generating GPS Trajectory with Diffusion Probabilistic Model create mode 100644 data/2023/neurips/Differentiable Clustering with Perturbed Spanning Forests create mode 100644 data/2023/neurips/Differentiable sorting for censored time-to-event data create mode 100644 "data/2023/neurips/Differentially Private Statistical Inference through \316\262-Divergence One Posterior Sampling" create mode 100644 data/2023/neurips/DiffuseBot: Breeding Soft Robots With Physics-Augmented Generative Diffusion Models create mode 100644 data/2023/neurips/Diffusion Model is an Effective Planner and Data Synthesizer for Multi-Task Reinforcement Learning create mode 100644 data/2023/neurips/Diffusion Self-Guidance for Controllable Image Generation create mode 100644 data/2023/neurips/Direct Preference-based Policy Optimization without Reward Modeling create mode 100644 data/2023/neurips/Dis-inhibitory neuronal circuits can control the sign of synaptic plasticity create mode 100644 data/2023/neurips/DisDiff: Unsupervised Disentanglement of Diffusion Probabilistic Models create mode 100644 data/2023/neurips/Disambiguated Attention Embedding for Multi-Instance Partial-Label Learning create mode 100644 data/2023/neurips/Discovering General Reinforcement Learning Algorithms with Adversarial Environment Design create mode 100644 data/2023/neurips/Discovering Hierarchical Achievements in Reinforcement Learning via Contrastive Learning create mode 100644 data/2023/neurips/Discovering Intrinsic Spatial-Temporal Logic Rules to Explain Human Actions create mode 100644 data/2023/neurips/Disentangled Wasserstein Autoencoder for T-Cell Receptor Engineering create mode 100644 data/2023/neurips/Disentangling Cognitive Diagnosis with Limited Exercise Labels create mode 100644 data/2023/neurips/Disentangling Voice and Content with Self-Supervision for Speaker Recognition create mode 100644 data/2023/neurips/Distilling Out-of-Distribution Robustness from Vision-Language Foundation Models create mode 100644 data/2023/neurips/Distributed Inference and Fine-tuning of Large Language Models Over The Internet create mode 100644 data/2023/neurips/Distributed Personalized Empirical Risk Minimization create mode 100644 data/2023/neurips/Distributionally Robust Ensemble of Lottery Tickets Towards Calibrated Sparse Network Training create mode 100644 data/2023/neurips/Diverse Shape Completion via Style Modulated Generative Adversarial Networks create mode 100644 data/2023/neurips/Diversified Outlier Exposure for Out-of-Distribution Detection via Informative Extrapolation create mode 100644 data/2023/neurips/Diversify Your Vision Datasets with Automatic Diffusion-based Augmentation create mode 100644 data/2023/neurips/Divide, Evaluate, and Refine: Evaluating and Improving Text-to-Image Alignment with Iterative VQA Feedback create mode 100644 data/2023/neurips/DoReMi: Optimizing Data Mixtures Speeds Up Language Model Pretraining create mode 100644 data/2023/neurips/Does Invariant Graph Learning via Environment Augmentation Learn Invariance? create mode 100644 data/2023/neurips/Does a sparse ReLU network training problem always admit an optimum ? create mode 100644 data/2023/neurips/Don't blame Dataset Shift! Shortcut Learning due to Gradients and Cross Entropy create mode 100644 data/2023/neurips/Don't just prune by magnitude! Your mask topology is a secret weapon create mode 100644 data/2023/neurips/Double Pessimism is Provably Efficient for Distributionally Robust Offline Reinforcement Learning: Generic Algorithm and Robust Partial Coverage create mode 100644 data/2023/neurips/Double Randomized Underdamped Langevin with Dimension-Independent Convergence Guarantee create mode 100644 data/2023/neurips/Double and Single Descent in Causal Inference with an Application to High-Dimensional Synthetic Control create mode 100644 data/2023/neurips/Doubly Robust Augmented Transfer for Meta-Reinforcement Learning create mode 100644 data/2023/neurips/Drift doesn't Matter: Dynamic Decomposition with Diffusion Reconstruction for Unstable Multivariate Time Series Anomaly Detection create mode 100644 data/2023/neurips/Dual Self-Awareness Value Decomposition Framework without Individual Global Max for Cooperative MARL create mode 100644 data/2023/neurips/DynGFN: Towards Bayesian Inference of Gene Regulatory Networks with GFlowNets create mode 100644 data/2023/neurips/DynPoint: Dynamic Neural Point For View Synthesis create mode 100644 data/2023/neurips/Dynamic Context Pruning for Efficient and Interpretable Autoregressive Transformers create mode 100644 data/2023/neurips/Dynamic Personalized Federated Learning with Adaptive Differential Privacy create mode 100644 data/2023/neurips/Dynamic Regret of Adversarial Linear Mixture MDPs create mode 100644 data/2023/neurips/Dynamic Sparsity Is Channel-Level Sparsity Learner create mode 100644 "data/2023/neurips/D\303\244RF: Boosting Radiance Fields from Sparse Input Views with Monocular Depth Adaptation" create mode 100644 data/2023/neurips/E2PNet: Event to Point Cloud Registration with Spatio-Temporal Representation Learning create mode 100644 data/2023/neurips/ECG-QA: A Comprehensive Question Answering Dataset Combined With Electrocardiogram create mode 100644 data/2023/neurips/EDGI: Equivariant Diffusion for Planning with Embodied Agents create mode 100644 data/2023/neurips/EFWI: Multiparameter Benchmark Datasets for Elastic Full Waveform Inversion of Geophysical Properties create mode 100644 data/2023/neurips/EHRSHOT: An EHR Benchmark for Few-Shot Evaluation of Foundation Models create mode 100644 data/2023/neurips/ELDEN: Exploration via Local Dependencies create mode 100644 data/2023/neurips/Easy Learning from Label Proportions create mode 100644 data/2023/neurips/Effective Bayesian Heteroscedastic Regression with Deep Neural Networks create mode 100644 data/2023/neurips/Effective Robustness against Natural Distribution Shifts for Models with Different Training Data create mode 100644 data/2023/neurips/Effectively Learning Initiation Sets in Hierarchical Reinforcement Learning create mode 100644 data/2023/neurips/Efficient Adversarial Contrastive Learning via Robustness-Aware Coreset Selection create mode 100644 data/2023/neurips/Efficient Algorithms for Generalized Linear Bandits with Heavy-tailed Rewards create mode 100644 data/2023/neurips/Efficient Bayesian Learning Curve Extrapolation using Prior-Data Fitted Networks create mode 100644 data/2023/neurips/Efficient Diffusion Policies For Offline Reinforcement Learning create mode 100644 data/2023/neurips/Efficient Learning of Linear Graph Neural Networks via Node Subsampling create mode 100644 data/2023/neurips/Efficient Model-Free Exploration in Low-Rank MDPs create mode 100644 data/2023/neurips/Efficient Potential-based Exploration in Reinforcement Learning using Inverse Dynamic Bisimulation Metric create mode 100644 data/2023/neurips/Efficient Robust Bayesian Optimization for Arbitrary Uncertain inputs create mode 100644 data/2023/neurips/Efficient Sampling of Stochastic Differential Equations with Positive Semi-Definite Models create mode 100644 data/2023/neurips/Efficient Test-Time Adaptation for Super-Resolution with Second-Order Degradation and Reconstruction create mode 100644 data/2023/neurips/Efficient Uncertainty Quantification and Reduction for Over-Parameterized Neural Networks create mode 100644 data/2023/neurips/Efficiently incorporating quintuple interactions into geometric deep learning force fields create mode 100644 data/2023/neurips/EgoTracks: A Long-term Egocentric Visual Object Tracking Dataset create mode 100644 data/2023/neurips/Eliminating Catastrophic Overfitting Via Abnormal Adversarial Examples Regularization create mode 100644 data/2023/neurips/Emergence of Shape Bias in Convolutional Neural Networks through Activation Sparsity create mode 100644 data/2023/neurips/Emergent Communication for Rules Reasoning create mode 100644 data/2023/neurips/Empowering Convolutional Neural Nets with MetaSin Activation create mode 100644 data/2023/neurips/End-To-End Latent Variational Diffusion Models for Inverse Problems in High Energy Physics create mode 100644 data/2023/neurips/Energy-Based Cross Attention for Bayesian Context Update in Text-to-Image Diffusion Models create mode 100644 data/2023/neurips/Energy-Efficient Scheduling with Predictions create mode 100644 data/2023/neurips/Enhancing Knowledge Transfer for Task Incremental Learning with Data-free Subnetwork create mode 100644 data/2023/neurips/Enhancing Motion Deblurring in High-Speed Scenes with Spike Streams create mode 100644 data/2023/neurips/Enhancing Sharpness-Aware Optimization Through Variance Suppression create mode 100644 data/2023/neurips/Entropic Neural Optimal Transport via Diffusion Processes create mode 100644 data/2023/neurips/Entropy-based Training Methods for Scalable Neural Implicit Samplers create mode 100644 data/2023/neurips/Episodic Multi-Task Learning with Heterogeneous Neural Processes create mode 100644 data/2023/neurips/Equal Opportunity of Coverage in Fair Regression create mode 100644 data/2023/neurips/Ess-InfoGAIL: Semi-supervised Imitation Learning from Imbalanced Demonstrations create mode 100644 data/2023/neurips/Estimating Causal Effects Identifiable from a Combination of Observations and Experiments create mode 100644 data/2023/neurips/Estimating Koopman operators with sketching to provably learn large scale dynamical systems create mode 100644 data/2023/neurips/Estimating Riemannian Metric with Noise-Contaminated Intrinsic Distance create mode 100644 data/2023/neurips/Evaluating Cognitive Maps and Planning in Large Language Models with CogEval create mode 100644 data/2023/neurips/Evaluating Neuron Interpretation Methods of NLP Models create mode 100644 data/2023/neurips/Evaluating Open-QA Evaluation create mode 100644 data/2023/neurips/Evaluating Post-hoc Explanations for Graph Neural Networks via Robustness Analysis create mode 100644 data/2023/neurips/Evaluating Robustness and Uncertainty of Graph Models Under Structural Distributional Shifts create mode 100644 data/2023/neurips/Evaluating Self-Supervised Learning for Molecular Graph Embeddings create mode 100644 data/2023/neurips/Evaluating and Improving Tool-Augmented Computation-Intensive Math Reasoning create mode 100644 data/2023/neurips/Evaluating the Robustness of Interpretability Methods through Explanation Invariance and Equivariance create mode 100644 data/2023/neurips/Expanding Small-Scale Datasets with Guided Imagination create mode 100644 data/2023/neurips/Experimental Designs for Heteroskedastic Variance create mode 100644 data/2023/neurips/Exploiting Correlated Auxiliary Feedback in Parameterized Bandits create mode 100644 data/2023/neurips/Exploiting hidden structures in non-convex games for convergence to Nash equilibrium create mode 100644 data/2023/neurips/Exploring Loss Functions for Time-based Training Strategy in Spiking Neural Networks create mode 100644 data/2023/neurips/Exploring Question Decomposition for Zero-Shot VQA create mode 100644 data/2023/neurips/Exponentially Convergent Algorithms for Supervised Matrix Factorization create mode 100644 data/2023/neurips/Exposing Attention Glitches with Flip-Flop Language Modeling create mode 100644 data/2023/neurips/Exposing flaws of generative model evaluation metrics and their unfair treatment of diffusion models create mode 100644 data/2023/neurips/Expressive Sign Equivariant Networks for Spectral Geometric Learning create mode 100644 data/2023/neurips/Expressivity-Preserving GNN Simulation create mode 100644 data/2023/neurips/FAMO: Fast Adaptive Multitask Optimization create mode 100644 data/2023/neurips/FETV: A Benchmark for Fine-Grained Evaluation of Open-Domain Text-to-Video Generation create mode 100644 data/2023/neurips/FIND: A Function Description Benchmark for Evaluating Interpretability Methods create mode 100644 data/2023/neurips/FIRAL: An Active Learning Algorithm for Multinomial Logistic Regression create mode 100644 data/2023/neurips/FLuID: Mitigating Stragglers in Federated Learning using Invariant Dropout create mode 100644 data/2023/neurips/Facing Off World Model Backbones: RNNs, Transformers, and S4 create mode 100644 data/2023/neurips/Failure-Aware Gaussian Process Optimization with Regret Bounds create mode 100644 data/2023/neurips/Fair Graph Distillation create mode 100644 data/2023/neurips/Fair Streaming Principal Component Analysis: Statistical and Algorithmic Viewpoint create mode 100644 data/2023/neurips/Fairly Recommending with Social Attributes: A Flexible and Controllable Optimization Approach create mode 100644 data/2023/neurips/Fairness Continual Learning Approach to Semantic Scene Understanding in Open-World Environments create mode 100644 data/2023/neurips/Faith and Fate: Limits of Transformers on Compositionality create mode 100644 data/2023/neurips/False Discovery Proportion control for aggregated Knockoffs create mode 100644 data/2023/neurips/Fast Approximation of Similarity Graphs with Kernel Density Estimation create mode 100644 data/2023/neurips/Fast Model DeBias with Machine Unlearning create mode 100644 data/2023/neurips/Fast Projected Newton-like Method for Precision Matrix Estimation under Total Positivity create mode 100644 data/2023/neurips/Fast Scalable and Accurate Discovery of DAGs Using the Best Order Score Search and Grow Shrink Trees create mode 100644 data/2023/neurips/Faster Differentially Private Convex Optimization via Second-Order Methods create mode 100644 data/2023/neurips/Faster Discrete Convex Function Minimization with Predictions: The M-Convex Case create mode 100644 data/2023/neurips/Faster approximate subgraph counts with privacy create mode 100644 data/2023/neurips/Feature Learning for Interpretable, Performant Decision Trees create mode 100644 data/2023/neurips/Feature Likelihood Score: Evaluating the Generalization of Generative Models Using Samples create mode 100644 data/2023/neurips/Feature Selection in the Contrastive Analysis Setting create mode 100644 data/2023/neurips/Fed-GraB: Federated Long-tailed Learning with Self-Adjusting Gradient Balancer create mode 100644 data/2023/neurips/FedGCN: Convergence-Communication Tradeoffs in Federated Training of Graph Convolutional Networks create mode 100644 data/2023/neurips/FedNAR: Federated Optimization with Normalized Annealing Regularization create mode 100644 data/2023/neurips/Federated Linear Bandits with Finite Adversarial Actions create mode 100644 data/2023/neurips/Finding Local Minima Efficiently in Decentralized Optimization create mode 100644 data/2023/neurips/Finding Safe Zones of Markov Decision Processes Policies create mode 100644 data/2023/neurips/Fine-Grained Cross-View Geo-Localization Using a Correlation-Aware Homography Estimator create mode 100644 data/2023/neurips/Fine-Grained Visual Prompting create mode 100644 data/2023/neurips/Finite Population Regression Adjustment and Non-asymptotic Guarantees for Treatment Effect Estimation create mode 100644 data/2023/neurips/Flat Seeking Bayesian Neural Networks create mode 100644 data/2023/neurips/Flocks of Stochastic Parrots: Differentially Private Prompt Learning for Large Language Models create mode 100644 data/2023/neurips/Flow Factorized Representation Learning create mode 100644 data/2023/neurips/Flow-Based Feature Fusion for Vehicle-Infrastructure Cooperative 3D Object Detection create mode 100644 data/2023/neurips/Flow: Per-instance Personalized Federated Learning create mode 100644 data/2023/neurips/Focus Your Attention when Few-Shot Classification create mode 100644 data/2023/neurips/Focus on Query: Adversarial Mining Transformer for Few-Shot Segmentation create mode 100644 data/2023/neurips/Follow-ups Also Matter: Improving Contextual Bandits via Post-serving Contexts create mode 100644 data/2023/neurips/Formulating Discrete Probability Flow Through Optimal Transport create mode 100644 data/2023/neurips/FourierGNN: Rethinking Multivariate Time Series Forecasting from a Pure Graph Perspective create mode 100644 data/2023/neurips/FourierHandFlow: Neural 4D Hand Representation Using Fourier Query Flow create mode 100644 data/2023/neurips/Framework and Benchmarks for Combinatorial and Mixed-variable Bayesian Optimization create mode 100644 data/2023/neurips/Frequency Domain-Based Dataset Distillation create mode 100644 data/2023/neurips/Frequency-domain MLPs are More Effective Learners in Time Series Forecasting create mode 100644 data/2023/neurips/From Cloze to Comprehension: Retrofitting Pre-trained Masked Language Models to Pre-trained Machine Reader create mode 100644 data/2023/neurips/From Trainable Negative Depth to Edge Heterophily in Graphs create mode 100644 data/2023/neurips/Full-Atom Protein Pocket Design via Iterative Refinement create mode 100644 data/2023/neurips/Function Space Bayesian Pseudocoreset for Bayesian Neural Networks create mode 100644 data/2023/neurips/Functional Equivalence and Path Connectivity of Reducible Hyperbolic Tangent Networks create mode 100644 data/2023/neurips/GAIA: Delving into Gradient-based Attribution Abnormality for Out-of-distribution Detection create mode 100644 data/2023/neurips/GAUCHE: A Library for Gaussian Processes in Chemistry create mode 100644 data/2023/neurips/GLEMOS: Benchmark for Instantaneous Graph Learning Model Selection create mode 100644 data/2023/neurips/GLOBER: Coherent Non-autoregressive Video Generation via GLOBal Guided Video DecodER create mode 100644 data/2023/neurips/GMSF: Global Matching Scene Flow create mode 100644 data/2023/neurips/GPEX, A Framework For Interpreting Artificial Neural Networks create mode 100644 data/2023/neurips/GPT-ST: Generative Pre-Training of Spatio-Temporal Graph Neural Networks create mode 100644 data/2023/neurips/GPT4Tools: Teaching Large Language Model to Use Tools via Self-instruction create mode 100644 data/2023/neurips/Gacs-Korner Common Information Variational Autoencoder create mode 100644 data/2023/neurips/Gaussian Membership Inference Privacy create mode 100644 data/2023/neurips/Gaussian Partial Information Decomposition: Bias Correction and Application to High-dimensional Data create mode 100644 data/2023/neurips/Gaussian Process Probes (GPP) for Uncertainty-Aware Probing create mode 100644 data/2023/neurips/GenImage: A Million-Scale Benchmark for Detecting AI-Generated Image create mode 100644 data/2023/neurips/GenS: Generalizable Neural Surface Reconstruction from Multi-View Images create mode 100644 data/2023/neurips/Generalized Bayesian Inference for Scientific Simulators via Amortized Cost Estimation create mode 100644 data/2023/neurips/Generalized Belief Transport create mode 100644 data/2023/neurips/Generalized Logit Adjustment: Calibrating Fine-tuned Models by Removing Label Bias in Foundation Models create mode 100644 data/2023/neurips/Generalized Weighted Path Consistency for Mastering Atari Games create mode 100644 data/2023/neurips/Generalized equivalences between subsampling and ridge regularization create mode 100644 data/2023/neurips/Generalized test utilities for long-tail performance in extreme multi-label classification create mode 100644 data/2023/neurips/Generating Behaviorally Diverse Policies with Latent Diffusion Models create mode 100644 data/2023/neurips/Generator Identification for Linear SDEs with Additive and Multiplicative Noise create mode 100644 data/2023/neurips/GeoDE: a Geographically Diverse Evaluation Dataset for Object Recognition create mode 100644 data/2023/neurips/Geodesic Multi-Modal Mixup for Robust Fine-Tuning create mode 100644 data/2023/neurips/Geometric Analysis of Matrix Sensing over Graphs create mode 100644 data/2023/neurips/Global Convergence Analysis of Local SGD for Two-layer Neural Network without Overparameterization create mode 100644 "data/2023/neurips/Global Identifiability of \360\235\223\2011-based Dictionary Learning via Matrix Volume Optimization" create mode 100644 data/2023/neurips/Global Structure-Aware Diffusion Process for Low-light Image Enhancement create mode 100644 data/2023/neurips/Gradient-Based Feature Learning under Structured Data create mode 100644 data/2023/neurips/Grammar Prompting for Domain-Specific Language Generation with Large Language Models create mode 100644 data/2023/neurips/Granger Components Analysis: Unsupervised learning of latent temporal dependencies create mode 100644 data/2023/neurips/Graph Neural Networks for Road Safety Modeling: Datasets and Evaluations for Accident Analysis create mode 100644 data/2023/neurips/Grassmann Manifold Flows for Stable Shape Generation create mode 100644 data/2023/neurips/Grounded Decoding: Guiding Text Generation with Grounded Models for Embodied Agents create mode 100644 data/2023/neurips/Group Fairness in Peer Review create mode 100644 data/2023/neurips/Guiding The Last Layer in Federated Learning with Pre-Trained Models create mode 100644 data/2023/neurips/H-InDex: Visual Reinforcement Learning with Hand-Informed Representations for Dexterous Manipulation create mode 100644 data/2023/neurips/HA-ViD: A Human Assembly Video Dataset for Comprehensive Assembly Knowledge Understanding create mode 100644 data/2023/neurips/HOH: Markerless Multimodal Human-Object-Human Handover Dataset with Large Object Count create mode 100644 data/2023/neurips/Hardware Resilience Properties of Text-Guided Image Classifiers create mode 100644 data/2023/neurips/Harnessing the power of choices in decision tree learning create mode 100644 data/2023/neurips/HiNeRV: Video Compression with Hierarchical Encoding-based Neural Representation create mode 100644 data/2023/neurips/Hierarchical Decomposition of Prompt-Based Continual Learning: Rethinking Obscured Sub-optimality create mode 100644 data/2023/neurips/Hierarchical Multi-Agent Skill Discovery create mode 100644 data/2023/neurips/Hierarchical Semi-Implicit Variational Inference with Application to Diffusion Model Acceleration create mode 100644 data/2023/neurips/High Precision Causal Model Evaluation with Conditional Randomization create mode 100644 data/2023/neurips/High-dimensional Asymptotics of Denoising Autoencoders create mode 100644 data/2023/neurips/Holistic Evaluation of Text-to-Image Models create mode 100644 data/2023/neurips/Homotopy-based training of NeuralODEs for accurate dynamics discovery create mode 100644 data/2023/neurips/How Far Can Camels Go? Exploring the State of Instruction Tuning on Open Resources create mode 100644 data/2023/neurips/How Re-sampling Helps for Long-Tail Learning? create mode 100644 data/2023/neurips/How does GPT-2 compute greater-than?: Interpreting mathematical abilities in a pre-trained language model create mode 100644 data/2023/neurips/How to Fine-tune the Model: Unified Model Shift and Model Bias Policy Optimization create mode 100644 data/2023/neurips/How to Scale Your EMA create mode 100644 data/2023/neurips/How to Turn Your Knowledge Graph Embeddings into Generative Models create mode 100644 data/2023/neurips/HubRouter: Learning Global Routing via Hub Generation and Pin-hub Connection create mode 100644 data/2023/neurips/Human-Guided Complexity-Controlled Abstractions create mode 100644 data/2023/neurips/Human-in-the-Loop Optimization for Deep Stimulus Encoding in Visual Prostheses create mode 100644 data/2023/neurips/Hyper-Skin: A Hyperspectral Dataset for Reconstructing Facial Skin-Spectra from RGB Images create mode 100644 data/2023/neurips/Hyperbolic Space with Hierarchical Margin Boosts Fine-Grained Learning from Coarse Labels create mode 100644 data/2023/neurips/IBA: Towards Irreversible Backdoor Attacks in Federated Learning create mode 100644 data/2023/neurips/ID and OOD Performance Are Sometimes Inversely Correlated on Real-world Datasets create mode 100644 data/2023/neurips/IPMix: Label-Preserving Data Augmentation Method for Training Robust Classifiers create mode 100644 data/2023/neurips/ImageNet-Hard: The Hardest Images Remaining from a Study of the Power of Zoom and Spatial Biases in Image Classification create mode 100644 data/2023/neurips/ImageReward: Learning and Evaluating Human Preferences for Text-to-Image Generation create mode 100644 data/2023/neurips/Imagine That! Abstract-to-Intricate Text-to-Image Synthesis with Scene Graph Hallucination Diffusion create mode 100644 data/2023/neurips/Imitation Learning from Imperfection: Theoretical Justifications and Algorithms create mode 100644 data/2023/neurips/Imitation Learning from Vague Feedback create mode 100644 data/2023/neurips/Implicit Bias of Gradient Descent for Logistic Regression at the Edge of Stability create mode 100644 data/2023/neurips/Implicit Manifold Gaussian Process Regression create mode 100644 data/2023/neurips/Implicit Regularization in Over-Parameterized Support Vector Machine create mode 100644 data/2023/neurips/Implicit Variational Inference for High-Dimensional Posteriors create mode 100644 data/2023/neurips/Implicit variance regularization in non-contrastive SSL create mode 100644 data/2023/neurips/Improved Algorithms for Stochastic Linear Bandits Using Tail Bounds for Martingale Mixtures create mode 100644 data/2023/neurips/Improved Bayes Risk Can Yield Reduced Social Welfare Under Competition create mode 100644 data/2023/neurips/Improved Best-of-Both-Worlds Guarantees for Multi-Armed Bandits: FTRL with General Regularizers and Multiple Optimal Arms create mode 100644 data/2023/neurips/Improved Convergence in High Probability of Clipped Gradient Methods with Heavy Tailed Noise create mode 100644 data/2023/neurips/Improving Robustness with Adaptive Weight Decay create mode 100644 data/2023/neurips/Improving neural network representations using human similarity judgments create mode 100644 data/2023/neurips/In Defense of Softmax Parametrization for Calibrated and Consistent Learning to Defer create mode 100644 data/2023/neurips/In-Context Impersonation Reveals Large Language Models' Strengths and Biases create mode 100644 data/2023/neurips/In-Context Learning Unlocked for Diffusion Models create mode 100644 data/2023/neurips/Individual Arbitrariness and Group Fairness create mode 100644 data/2023/neurips/Inferring Hybrid Neural Fluid Fields from Videos create mode 100644 data/2023/neurips/Inferring the Future by Imagining the Past create mode 100644 data/2023/neurips/InfoCD: A Contrastive Chamfer Distance Loss for Point Cloud Completion create mode 100644 data/2023/neurips/InfoPrompt: Information-Theoretic Soft Prompt Tuning for Natural Language Understanding create mode 100644 data/2023/neurips/Information Geometry of the Retinal Representation Manifold create mode 100644 data/2023/neurips/Information-guided Planning: An Online Approach for Partially Observable Problems create mode 100644 data/2023/neurips/Initialization-Dependent Sample Complexity of Linear Predictors and Neural Networks create mode 100644 data/2023/neurips/Inner Product-based Neural Network Similarity create mode 100644 data/2023/neurips/InsActor: Instruction-driven Physics-based Characters create mode 100644 data/2023/neurips/Inserting Anybody in Diffusion Models via Celeb Basis create mode 100644 data/2023/neurips/Interactive Multi-fidelity Learning for Cost-effective Adaptation of Language Model with Sparse Human Supervision create mode 100644 data/2023/neurips/Interactive Visual Reasoning under Uncertainty create mode 100644 data/2023/neurips/Interpretability at Scale: Identifying Causal Mechanisms in Alpaca create mode 100644 data/2023/neurips/Interpretable Prototype-based Graph Information Bottleneck create mode 100644 data/2023/neurips/Intrinsic Dimension Estimation for Robust Detection of AI-Generated Texts create mode 100644 data/2023/neurips/Invariant Learning via Probability of Sufficient and Necessary Causes create mode 100644 data/2023/neurips/Inverse Dynamics Pretraining Learns Good Representations for Multitask Imitation create mode 100644 data/2023/neurips/Inverse Reinforcement Learning with the Average Reward Criterion create mode 100644 data/2023/neurips/Iterative Reachability Estimation for Safe Reinforcement Learning create mode 100644 data/2023/neurips/Jaccard Metric Losses: Optimizing the Jaccard Index with Soft Labels create mode 100644 data/2023/neurips/Jailbroken: How Does LLM Safety Training Fail? create mode 100644 data/2023/neurips/K-Nearest-Neighbor Local Sampling Based Conditional Independence Testing create mode 100644 data/2023/neurips/KD-Zero: Evolving Knowledge Distiller for Any Teacher-Student Pairs create mode 100644 data/2023/neurips/Kernel Quadrature with Randomly Pivoted Cholesky create mode 100644 data/2023/neurips/Kiki or Bouba? Sound Symbolism in Vision-and-Language Models create mode 100644 data/2023/neurips/Kissing to Find a Match: Efficient Low-Rank Permutation Representation create mode 100644 data/2023/neurips/Knowledge Diffusion for Distillation create mode 100644 data/2023/neurips/Knowledge Distillation Performs Partial Variance Reduction create mode 100644 data/2023/neurips/L-CAD: Language-based Colorization with Any-level Descriptions using Diffusion Priors create mode 100644 data/2023/neurips/L2-Uniform Stability of Randomized Learning Algorithms: Sharper Generalization Bounds and Confidence Boosting create mode 100644 data/2023/neurips/LEACE: Perfect linear concept erasure in closed form create mode 100644 data/2023/neurips/LIBERO: Benchmarking Knowledge Transfer for Lifelong Robot Learning create mode 100644 data/2023/neurips/LLMScore: Unveiling the Power of Large Language Models in Text-to-Image Synthesis Evaluation create mode 100644 data/2023/neurips/LVM-Med: Learning Large-Scale Self-Supervised Vision Models for Medical Imaging via Second-order Graph Matching create mode 100644 data/2023/neurips/Label Poisoning is All You Need create mode 100644 data/2023/neurips/Label-Only Model Inversion Attacks via Knowledge Transfer create mode 100644 data/2023/neurips/Label-Retrieval-Augmented Diffusion Models for Learning from Noisy Labels create mode 100644 data/2023/neurips/LagrangeBench: A Lagrangian Fluid Mechanics Benchmarking Suite create mode 100644 data/2023/neurips/Langevin Quasi-Monte Carlo create mode 100644 data/2023/neurips/Language Is Not All You Need: Aligning Perception with Language Models create mode 100644 data/2023/neurips/Language Models Don't Always Say What They Think: Unfaithful Explanations in Chain-of-Thought Prompting create mode 100644 data/2023/neurips/Language Models Meet World Models: Embodied Experiences Enhance Language Models create mode 100644 data/2023/neurips/Language Quantized AutoEncoders: Towards Unsupervised Text-Image Alignment create mode 100644 data/2023/neurips/Language-based Action Concept Spaces Improve Video Self-Supervised Learning create mode 100644 data/2023/neurips/Large Language Models Are Semi-Parametric Reinforcement Learning Agents create mode 100644 data/2023/neurips/Large Language Models are Visual Reasoning Coordinators create mode 100644 data/2023/neurips/Large Language Models can Implement Policy Iteration create mode 100644 data/2023/neurips/LargeST: A Benchmark Dataset for Large-Scale Traffic Forecasting create mode 100644 data/2023/neurips/Last-Iterate Convergent Policy Gradient Primal-Dual Methods for Constrained MDPs create mode 100644 data/2023/neurips/Latent SDEs on Homogeneous Spaces create mode 100644 data/2023/neurips/Latent exploration for Reinforcement Learning create mode 100644 data/2023/neurips/Layer-Neighbor Sampling - Defusing Neighborhood Explosion in GNNs create mode 100644 data/2023/neurips/Learn to Categorize or Categorize to Learn? Self-Coding for Generalized Category Discovery create mode 100644 data/2023/neurips/Learning Adaptive Tensorial Density Fields for Clean Cryo-ET Reconstruction create mode 100644 data/2023/neurips/Learning Causal Models under Independent Changes create mode 100644 data/2023/neurips/Learning Cuts via Enumeration Oracles create mode 100644 data/2023/neurips/Learning DAGs from Data with Few Root Causes create mode 100644 data/2023/neurips/Learning Dense Flow Field for Highly-accurate Cross-view Camera Localization create mode 100644 data/2023/neurips/Learning Descriptive Image Captioning via Semipermeable Maximum Likelihood Estimation create mode 100644 data/2023/neurips/Learning Functional Transduction create mode 100644 data/2023/neurips/Learning Human Action Recognition Representations Without Real Humans create mode 100644 data/2023/neurips/Learning Invariant Molecular Representation in Latent Discrete Space create mode 100644 data/2023/neurips/Learning Large Graph Property Prediction via Graph Segment Training create mode 100644 data/2023/neurips/Learning Large-Scale MTP2 Gaussian Graphical Models via Bridge-Block Decomposition create mode 100644 data/2023/neurips/Learning Large-scale Neural Fields via Context Pruned Meta-Learning create mode 100644 data/2023/neurips/Learning List-Level Domain-Invariant Representations for Ranking create mode 100644 data/2023/neurips/Learning Motion Refinement for Unsupervised Face Animation create mode 100644 data/2023/neurips/Learning Rate Free Bayesian Inference in Constrained Domains create mode 100644 data/2023/neurips/Learning Regularized Monotone Graphon Mean-Field Games create mode 100644 data/2023/neurips/Learning Repeatable Speech Embeddings Using An Intra-class Correlation Regularizer create mode 100644 data/2023/neurips/Learning Shared Safety Constraints from Multi-task Demonstrations create mode 100644 data/2023/neurips/Learning Time-Invariant Representations for Individual Neurons from Population Dynamics create mode 100644 data/2023/neurips/Learning Trajectories are Generalization Indicators create mode 100644 data/2023/neurips/Learning Universal Policies via Text-Guided Video Generation create mode 100644 data/2023/neurips/Learning Visual Prior via Generative Pre-Training create mode 100644 data/2023/neurips/Learning from Active Human Involvement through Proxy Value Propagation create mode 100644 data/2023/neurips/Learning from Rich Semantics and Coarse Locations for Long-tailed Object Detection create mode 100644 data/2023/neurips/Learning non-Markovian Decision-Making from State-only Sequences create mode 100644 data/2023/neurips/Learning to Augment Distributions for Out-of-distribution Detection create mode 100644 data/2023/neurips/Learning to Group Auxiliary Datasets for Molecule create mode 100644 data/2023/neurips/Learning to Parameterize Visual Attributes for Open-set Fine-grained Retrieval create mode 100644 data/2023/neurips/Learning to Receive Help: Intervention-Aware Concept Embedding Models create mode 100644 data/2023/neurips/Learning to Taste: A Multimodal Wine Dataset create mode 100644 data/2023/neurips/Learning-to-Rank Meets Language: Boosting Language-Driven Ordering Alignment for Ordinal Classification create mode 100644 data/2023/neurips/Leveraging Pre-trained Large Language Models to Construct and Utilize World Models for Model-based Task Planning create mode 100644 data/2023/neurips/Leveraging the two-timescale regime to demonstrate convergence of neural networks create mode 100644 data/2023/neurips/Lexinvariant Language Models create mode 100644 data/2023/neurips/LinkerNet: Fragment Poses and Linker Co-Design with 3D Equivariant Diffusion create mode 100644 data/2023/neurips/LithoBench: Benchmarking AI Computational Lithography for Semiconductor Manufacturing create mode 100644 data/2023/neurips/Lo-Hi: Practical ML Drug Discovery Benchmark create mode 100644 data/2023/neurips/LoCoOp: Few-Shot Out-of-Distribution Detection via Prompt Learning create mode 100644 data/2023/neurips/Lockdown: Backdoor Defense for Federated Learning with Isolated Subspace Training create mode 100644 data/2023/neurips/LogSpecT: Feasible Graph Learning Model from Stationary Signals with Recovery Guarantees create mode 100644 data/2023/neurips/Logarithmic Bayes Regret Bounds create mode 100644 data/2023/neurips/Long-Term Fairness with Unknown Dynamics create mode 100644 data/2023/neurips/Loss Dynamics of Temporal Difference Reinforcement Learning create mode 100644 data/2023/neurips/Lossy Image Compression with Conditional Diffusion Models create mode 100644 data/2023/neurips/Low-shot Object Learning with Mutual Exclusivity Bias create mode 100644 data/2023/neurips/Lower Bounds on Adaptive Sensing for Matrix Recovery create mode 100644 data/2023/neurips/LuminAIRe: Illumination-Aware Conditional Image Repainting for Lighting-Realistic Generation create mode 100644 data/2023/neurips/Lung250M-4B: A Combined 3D Dataset for CT- and Point Cloud-Based Intra-Patient Lung Registration create mode 100644 data/2023/neurips/M2Hub: Unlocking the Potential of Machine Learning for Materials Discovery create mode 100644 data/2023/neurips/M5HisDoc: A Large-scale Multi-style Chinese Historical Document Analysis Benchmark create mode 100644 data/2023/neurips/MADLAD-400: A Multilingual And Document-Level Large Audited Dataset create mode 100644 data/2023/neurips/MEGABYTE: Predicting Million-byte Sequences with Multiscale Transformers create mode 100644 data/2023/neurips/MG-ViT: A Multi-Granularity Method for Compact and Efficient Vision Transformers create mode 100644 data/2023/neurips/MMD-Fuse: Learning and Combining Kernels for Two-Sample Testing Without Data Splitting create mode 100644 data/2023/neurips/MVDoppler: Unleashing the Power of Multi-View Doppler for MicroMotion-based Gait Classification create mode 100644 data/2023/neurips/Machine learning detects terminal singularities create mode 100644 data/2023/neurips/Macro Placement by Wire-Mask-Guided Black-Box Optimization create mode 100644 data/2023/neurips/Many-body Approximation for Non-negative Tensors create mode 100644 data/2023/neurips/Marich: A Query-efficient Distributionally Equivalent Model Extraction Attack create mode 100644 data/2023/neurips/Masked Space-Time Hash Encoding for Efficient Dynamic Scene Reconstruction create mode 100644 data/2023/neurips/MathNAS: If Blocks Have a Role in Mathematical Architecture Design create mode 100644 data/2023/neurips/Max-Sliced Mutual Information create mode 100644 data/2023/neurips/Maximization of Average Precision for Deep Learning with Adversarial Ranking Robustness create mode 100644 data/2023/neurips/Maximum Independent Set: Self-Training through Dynamic Programming create mode 100644 data/2023/neurips/May the Force be with You: Unified Force-Centric Pre-Training for 3D Molecular Conformations create mode 100644 data/2023/neurips/MeGraph: Capturing Long-Range Interactions by Alternating Local and Hierarchical Aggregation on Multi-Scaled Graph Hierarchy create mode 100644 data/2023/neurips/MedSat: A Public Health Dataset for England Featuring Medical Prescriptions and Satellite Imagery create mode 100644 data/2023/neurips/Meta-AdaM: An Meta-Learned Adaptive Optimizer with Momentum for Few-Shot Learning create mode 100644 data/2023/neurips/Meta-Learning with Neural Bandit Scheduler create mode 100644 data/2023/neurips/Meta-in-context learning in large language models create mode 100644 data/2023/neurips/Meta-learning families of plasticity rules in recurrent spiking networks using simulation-based inference create mode 100644 data/2023/neurips/Metis: Understanding and Enhancing In-Network Regular Expressions create mode 100644 data/2023/neurips/Michelangelo: Conditional 3D Shape Generation based on Shape-Image-Text Aligned Latent Representation create mode 100644 data/2023/neurips/Mind the spikes: Benign overfitting of kernels and neural networks in fixed dimension create mode 100644 data/2023/neurips/Minigrid & Miniworld: Modular & Customizable Reinforcement Learning Environments for Goal-Oriented Tasks create mode 100644 data/2023/neurips/Minimax Forward and Backward Learning of Evolving Tasks with Performance Guarantees create mode 100644 data/2023/neurips/Minimum Description Length and Generalization Guarantees for Representation Learning create mode 100644 data/2023/neurips/Minimum-Risk Recalibration of Classifiers create mode 100644 data/2023/neurips/Mitigating Over-smoothing in Transformers via Regularized Nonlocal Functionals create mode 100644 data/2023/neurips/Mitigating Test-Time Bias for Fair Image Retrieval create mode 100644 data/2023/neurips/Mitigating the Effect of Incidental Correlations on Part-based Learning create mode 100644 data/2023/neurips/Mitigating the Popularity Bias of Graph Collaborative Filtering: A Dimensional Collapse Perspective create mode 100644 data/2023/neurips/MixFormerV2: Efficient Fully Transformer Tracking create mode 100644 data/2023/neurips/Mnemosyne: Learning to Train Transformers with Transformers create mode 100644 data/2023/neurips/MoCa: Measuring Human-Language Model Alignment on Causal and Moral Judgment Tasks create mode 100644 data/2023/neurips/Modality-Agnostic Self-Supervised Learning with Meta-Learned Masked Auto-Encoder create mode 100644 data/2023/neurips/Modality-Independent Teachers Meet Weakly-Supervised Audio-Visual Event Parser create mode 100644 data/2023/neurips/Model and Feature Diversity for Bayesian Neural Networks in Mutual Learning create mode 100644 data/2023/neurips/Model-Based Control with Sparse Neural Dynamics create mode 100644 data/2023/neurips/Model-Based Reparameterization Policy Gradient Methods: Theory and Practical Algorithms create mode 100644 data/2023/neurips/Model-free Posterior Sampling via Learning Rate Randomization create mode 100644 data/2023/neurips/Modelling Cellular Perturbations with the Sparse Additive Mechanism Shift Variational Autoencoder create mode 100644 data/2023/neurips/Module-wise Adaptive Distillation for Multimodality Foundation Models create mode 100644 data/2023/neurips/Moment Matching Denoising Gibbs Sampling create mode 100644 data/2023/neurips/MomentDiff: Generative Video Moment Retrieval from Random to Real create mode 100644 data/2023/neurips/Momentum Provably Improves Error Feedback! create mode 100644 data/2023/neurips/Monarch Mixer: A Simple Sub-Quadratic GEMM-Based Architecture create mode 100644 data/2023/neurips/Monte Carlo Tree Search with Boltzmann Exploration create mode 100644 data/2023/neurips/Motion-X: A Large-scale 3D Expressive Whole-body Human Motion Dataset create mode 100644 data/2023/neurips/Multi-Agent Learning with Heterogeneous Linear Contextual Bandits create mode 100644 data/2023/neurips/Multi-Agent Meta-Reinforcement Learning: Sharper Convergence Rates with Task Similarity create mode 100644 data/2023/neurips/Multi-Prompt Alignment for Multi-Source Unsupervised Domain Adaptation create mode 100644 data/2023/neurips/Multi-body SE(3) Equivariance for Unsupervised Rigid Segmentation and Motion Estimation create mode 100644 data/2023/neurips/Multi-modal Queried Object Detection in the Wild create mode 100644 data/2023/neurips/Multi-scale Diffusion Denoised Smoothing create mode 100644 data/2023/neurips/Multiclass Boosting: Simple and Intuitive Weak Learning Criteria create mode 100644 data/2023/neurips/Multinomial Logistic Regression: Asymptotic Normality on Null Covariates in High-Dimensions create mode 100644 data/2023/neurips/Multiply Robust Federated Estimation of Targeted Average Treatment Effects create mode 100644 data/2023/neurips/NAVI: Category-Agnostic Image Collections with High-Quality 3D Shape and Pose Annotations create mode 100644 data/2023/neurips/NCDL: A Framework for Deep Learning on non-Cartesian Lattices create mode 100644 data/2023/neurips/Natural Actor-Critic for Robust Reinforcement Learning with Function Approximation create mode 100644 data/2023/neurips/Near-Linear Time Algorithm for the Chamfer Distance create mode 100644 data/2023/neurips/Near-Optimal k-Clustering in the Sliding Window Model create mode 100644 data/2023/neurips/Nearest Neighbour with Bandit Feedback create mode 100644 data/2023/neurips/Nearly Optimal Bounds for Cyclic Forgetting create mode 100644 data/2023/neurips/Necessary and Sufficient Conditions for Optimal Decision Trees using Dynamic Programming create mode 100644 data/2023/neurips/Neural Circuits for Fast Poisson Compressed Sensing in the Olfactory Bulb create mode 100644 data/2023/neurips/Neural Data Transformer 2: Multi-context Pretraining for Neural Spiking Activity create mode 100644 data/2023/neurips/Neural Foundations of Mental Simulation: Future Prediction of Latent Representations on Dynamic Scenes create mode 100644 data/2023/neurips/Neural Functional Transformers create mode 100644 data/2023/neurips/Neural Ideal Large Eddy Simulation: Modeling Turbulence with Neural Stochastic Differential Equations create mode 100644 data/2023/neurips/Neural Image Compression: Generalization, Robustness, and Spectral Biases create mode 100644 data/2023/neurips/Neural Lyapunov Control for Discrete-Time Systems create mode 100644 data/2023/neurips/Neural Modulation for Flash Memory: An Unsupervised Learning Framework for Improved Reliability create mode 100644 data/2023/neurips/Neural Priming for Sample-Efficient Adaptation create mode 100644 data/2023/neurips/Neural Processes with Stability create mode 100644 data/2023/neurips/Neural Sampling in Hierarchical Exponential-family Energy-based Models create mode 100644 data/2023/neurips/NeuralGF: Unsupervised Point Normal Estimation by Learning Neural Gradient Function create mode 100644 data/2023/neurips/NeuroEvoBench: Benchmarking Evolutionary Optimizers for Deep Learning Applications create mode 100644 data/2023/neurips/New Bounds for Hyperparameter Tuning of Regression Problems Across Instances create mode 100644 data/2023/neurips/No Change, No Gain: Empowering Graph Neural Networks with Expected Model Change Maximization for Active Learning create mode 100644 data/2023/neurips/Nominality Score Conditioned Time Series Anomaly Detection by Point Sequential Reconstruction create mode 100644 data/2023/neurips/Non-Asymptotic Analysis of a UCB-based Top Two Algorithm create mode 100644 data/2023/neurips/Norm-based Generalization Bounds for Sparse Neural Networks create mode 100644 data/2023/neurips/Normalization Layers Are All That Sharpness-Aware Minimization Needs create mode 100644 data/2023/neurips/Normalizing flow neural networks by JKO scheme create mode 100644 data/2023/neurips/Not All Neuro-Symbolic Concepts Are Created Equal: Analysis and Mitigation of Reasoning Shortcuts create mode 100644 data/2023/neurips/OBELICS: An Open Web-Scale Filtered Dataset of Interleaved Image-Text Documents create mode 100644 data/2023/neurips/ODE-based Recurrent Model-free Reinforcement Learning for POMDPs create mode 100644 data/2023/neurips/OV-PARTS: Towards Open-Vocabulary Part Segmentation create mode 100644 data/2023/neurips/Object Reprojection Error (ORE): Camera pose benchmarks from lightweight tracking annotations create mode 100644 data/2023/neurips/Occ3D: A Large-Scale 3D Occupancy Prediction Benchmark for Autonomous Driving create mode 100644 data/2023/neurips/OceanBench: The Sea Surface Height Edition create mode 100644 data/2023/neurips/Offline Imitation Learning with Variational Counterfactual Reasoning create mode 100644 data/2023/neurips/Offline RL with Discrete Proxy Representations for Generalizability in POMDPs create mode 100644 data/2023/neurips/On Convergence of Polynomial Approximations to the Gaussian Mixture Entropy create mode 100644 data/2023/neurips/On Differentially Private Sampling from Gaussian and Product Distributions create mode 100644 data/2023/neurips/On Generalization Bounds for Projective Clustering create mode 100644 data/2023/neurips/On Masked Pre-training and the Marginal Likelihood create mode 100644 data/2023/neurips/On Robust Streaming for Learning with Experts: Algorithms and Lower Bounds create mode 100644 data/2023/neurips/On permutation symmetries in Bayesian neural network posteriors: a variational perspective create mode 100644 data/2023/neurips/On student-teacher deviations in distillation: does it pay to disobey? create mode 100644 data/2023/neurips/On the Adversarial Robustness of Out-of-distribution Generalization Models create mode 100644 data/2023/neurips/On the Complexity of Differentially Private Best-Arm Identification with Fixed Confidence create mode 100644 data/2023/neurips/On the Connection between Pre-training Data Diversity and Fine-tuning Robustness create mode 100644 data/2023/neurips/On the Convergence to a Global Solution of Shuffling-Type Gradient Algorithms create mode 100644 data/2023/neurips/On the Generalization Error of Stochastic Mirror Descent for Quadratically-Bounded Losses: an Improved Analysis create mode 100644 data/2023/neurips/On the Identifiability and Interpretability of Gaussian Process Models create mode 100644 data/2023/neurips/On the Identifiability of Sparse ICA without Assuming Non-Gaussianity create mode 100644 data/2023/neurips/On the Overlooked Structure of Stochastic Gradients create mode 100644 data/2023/neurips/On the Planning Abilities of Large Language Models - A Critical Investigation create mode 100644 data/2023/neurips/On the Robustness of Removal-Based Feature Attributions create mode 100644 data/2023/neurips/On the Role of Noise in the Sample Complexity of Learning Recurrent Neural Networks: Exponential Gaps for Long Sequences create mode 100644 data/2023/neurips/On the Role of Randomization in Adversarially Robust Classification create mode 100644 data/2023/neurips/On the Trade-off of Intra- Inter-class Diversity for Supervised Pre-training create mode 100644 data/2023/neurips/On the spectral bias of two-layer linear networks create mode 100644 data/2023/neurips/One Risk to Rule Them All: A Risk-Sensitive Perspective on Model-Based Offline Reinforcement Learning create mode 100644 data/2023/neurips/One-for-All: Bridge the Gap Between Heterogeneous Architectures in Knowledge Distillation create mode 100644 data/2023/neurips/One-step differentiation of iterative algorithms create mode 100644 data/2023/neurips/OneNet: Enhancing Time Series Forecasting Models under Concept Drift by Online Ensembling create mode 100644 data/2023/neurips/Online Constrained Meta-Learning: Provable Guarantees for Generalization create mode 100644 data/2023/neurips/Online Control for Meta-optimization create mode 100644 data/2023/neurips/Online Label Shift: Optimal Dynamic Regret meets Practical Algorithms create mode 100644 data/2023/neurips/Online Learning under Adversarial Nonlinear Constraints create mode 100644 data/2023/neurips/Online POMDP Planning with Anytime Deterministic Guarantees create mode 100644 data/2023/neurips/Online robust non-stationary estimation create mode 100644 data/2023/neurips/Open Compound Domain Adaptation with Object Style Compensation for Semantic Segmentation create mode 100644 data/2023/neurips/Open Visual Knowledge Extraction via Relation-Oriented Multimodality Model Prompting create mode 100644 data/2023/neurips/OpenMask3D: Open-Vocabulary 3D Instance Segmentation create mode 100644 data/2023/neurips/OpenSTL: A Comprehensive Benchmark of Spatio-Temporal Predictive Learning create mode 100644 data/2023/neurips/Operation-Level Early Stopping for Robustifying Differentiable NAS create mode 100644 data/2023/neurips/Operator Learning with Neural Fields: Tackling PDEs on General Geometries create mode 100644 data/2023/neurips/Optimal Algorithms for the Inhomogeneous Spiked Wigner Model create mode 100644 data/2023/neurips/Optimal Block-wise Asymmetric Graph Construction for Graph-based Semi-supervised Learning create mode 100644 data/2023/neurips/Optimal Convergence Rate for Exact Policy Mirror Descent in Discounted Markov Decision Processes create mode 100644 data/2023/neurips/Optimal Extragradient-Based Algorithms for Stochastic Variational Inequalities with Separable Structure create mode 100644 data/2023/neurips/Optimal Time Complexities of Parallel Stochastic Optimization Methods Under a Fixed Computation Model create mode 100644 data/2023/neurips/Optimal Transport for Treatment Effect Estimation create mode 100644 data/2023/neurips/Optimal testing using combined test statistics across independent studies create mode 100644 data/2023/neurips/Optimistic Natural Policy Gradient: a Simple Efficient Policy Optimization Framework for Online RL create mode 100644 data/2023/neurips/Optimizing Prompts for Text-to-Image Generation create mode 100644 data/2023/neurips/Order Matters in the Presence of Dataset Imbalance for Multilingual Learning create mode 100644 data/2023/neurips/Out-of-distribution Detection Learning with Unreliable Out-of-distribution Sources create mode 100644 data/2023/neurips/P-Flow: A Fast and Data-Efficient Zero-Shot TTS through Speech Prompting create mode 100644 data/2023/neurips/PAC Learning Linear Thresholds from Label Proportions create mode 100644 data/2023/neurips/PDE-Refiner: Achieving Accurate Long Rollouts with Neural PDE Solvers create mode 100644 data/2023/neurips/PLANNER: Generating Diversified Paragraph via Latent Language Diffusion Model create mode 100644 data/2023/neurips/POMDP Planning for Object Search in Partially Unknown Environment create mode 100644 data/2023/neurips/PPi: Pretraining Brain Signal Model for Patient-independent Seizure Detection create mode 100644 data/2023/neurips/PUCA: Patch-Unshuffle and Channel Attention for Enhanced Self-Supervised Image Denoising create mode 100644 data/2023/neurips/ParaFuzz: An Interpretability-Driven Technique for Detecting Poisoned Samples in NLP create mode 100644 data/2023/neurips/Parallel Submodular Function Minimization create mode 100644 data/2023/neurips/Parallel-mentoring for Offline Model-based Optimization create mode 100644 data/2023/neurips/Paraphrasing evades detectors of AI-generated text, but retrieval is an effective defense create mode 100644 data/2023/neurips/Partial Label Learning with Dissimilarity Propagation guided Candidate Label Shrinkage create mode 100644 data/2023/neurips/Participatory Personalization in Classification create mode 100644 data/2023/neurips/Particle-based Variational Inference with Generalized Wasserstein Gradient Flow create mode 100644 data/2023/neurips/Patch Diffusion: Faster and More Data-Efficient Training of Diffusion Models create mode 100644 data/2023/neurips/Patch n' Pack: NaViT, a Vision Transformer for any Aspect Ratio and Resolution create mode 100644 data/2023/neurips/Phase diagram of early training dynamics in deep neural networks: effect of the learning rate, depth, and width create mode 100644 data/2023/neurips/Physion++: Evaluating Physical Scene Understanding that Requires Online Inference of Different Physical Properties create mode 100644 data/2023/neurips/PlanE: Representation Learning over Planar Graphs create mode 100644 data/2023/neurips/PoET: A generative model of protein families as sequences-of-sequences create mode 100644 data/2023/neurips/Policy Space Diversity for Non-Transitive Games create mode 100644 data/2023/neurips/PolyDiffuse: Polygonal Shape Reconstruction via Guided Set Diffusion Models create mode 100644 data/2023/neurips/Polynomially Over-Parameterized Convolutional Neural Networks Contain Structured Strong Winning Lottery Tickets create mode 100644 data/2023/neurips/Post Hoc Explanations of Language Models Can Improve Language Models create mode 100644 data/2023/neurips/Post-processing Private Synthetic Data for Improving Utility on Selected Measures create mode 100644 data/2023/neurips/PrObeD: Proactive Object Detection Wrapper create mode 100644 data/2023/neurips/PreDiff: Precipitation Nowcasting with Latent Diffusion Models create mode 100644 data/2023/neurips/Preconditioning Matters: Fast Global Convergence of Non-convex Matrix Factorization via Scaled Gradient Descent create mode 100644 data/2023/neurips/Predicting a Protein's Stability under a Million Mutations create mode 100644 data/2023/neurips/Prediction and Control in Continual Reinforcement Learning create mode 100644 data/2023/neurips/Primal-Attention: Self-attention through Asymmetric Kernel SVD in Primal Representation create mode 100644 data/2023/neurips/Privacy Amplification via Compression: Achieving the Optimal Privacy-Accuracy-Communication Trade-off in Distributed Mean Estimation create mode 100644 data/2023/neurips/Private (Stochastic) Non-Convex Optimization Revisited: Second-Order Stationary Points and Excess Risks create mode 100644 data/2023/neurips/Private estimation algorithms for stochastic block models and mixture models create mode 100644 data/2023/neurips/ProBio: A Protocol-guided Multimodal Dataset for Molecular Biology Lab create mode 100644 data/2023/neurips/ProPILE: Probing Privacy Leakage in Large Language Models create mode 100644 data/2023/neurips/Probabilistic Invariant Learning with Randomized Linear Classifiers create mode 100644 data/2023/neurips/Projection-Free Methods for Stochastic Simple Bilevel Optimization with Convex Lower-level Problem create mode 100644 data/2023/neurips/Projection-Free Online Convex Optimization via Efficient Newton Iterations create mode 100644 data/2023/neurips/PromptIR: Prompting for All-in-One Image Restoration create mode 100644 data/2023/neurips/Propagating Knowledge Updates to LMs Through Distillation create mode 100644 data/2023/neurips/ProteinGym: Large-Scale Benchmarks for Protein Fitness Prediction and Design create mode 100644 data/2023/neurips/ProteinInvBench: Benchmarking Protein Inverse Folding on Diverse Tasks, Models, and Metrics create mode 100644 data/2023/neurips/Provable benefits of annealing for estimating normalizing constants: Importance Sampling, Noise-Contrastive Estimation, and beyond create mode 100644 data/2023/neurips/Provable benefits of score matching create mode 100644 data/2023/neurips/Provable convergence guarantees for black-box variational inference create mode 100644 data/2023/neurips/Provably (More) Sample-Efficient Offline RL with Options create mode 100644 data/2023/neurips/Provably Bounding Neural Network Preimages create mode 100644 data/2023/neurips/Proximity-Informed Calibration for Deep Neural Networks create mode 100644 data/2023/neurips/Q-DM: An Efficient Low-bit Quantized Diffusion Model create mode 100644 data/2023/neurips/QH9: A Quantum Hamiltonian Prediction Benchmark for QM9 Molecules create mode 100644 data/2023/neurips/Quantizable Transformers: Removing Outliers by Helping Attention Heads Do Nothing create mode 100644 data/2023/neurips/Query-based Temporal Fusion with Explicit Motion for 3D Object Detection create mode 100644 data/2023/neurips/RAPHAEL: Text-to-Image Generation via Large Mixture of Diffusion Paths create mode 100644 data/2023/neurips/RECESS Vaccine for Federated Learning: Proactive Defense Against Model Poisoning Attacks create mode 100644 data/2023/neurips/RL-based Stateful Neural Adaptive Sampling and Denoising for Real-Time Path Tracing create mode 100644 data/2023/neurips/RaLEs: a Benchmark for Radiology Language Evaluations create mode 100644 data/2023/neurips/RanPAC: Random Projections and Pre-trained Models for Continual Learning create mode 100644 data/2023/neurips/Random Cuts are Optimal for Explainable k-Medians create mode 100644 data/2023/neurips/RangePerception: Taming LiDAR Range View for Efficient and Accurate 3D Object Detection create mode 100644 data/2023/neurips/Rank-1 Matrix Completion with Gradient Descent and Small Random Initialization create mode 100644 data/2023/neurips/ReMaX: Relaxing for Better Training on Efficient Panoptic Segmentation create mode 100644 data/2023/neurips/Read and Reap the Rewards: Learning to Play Atari with the Help of Instruction Manuals create mode 100644 data/2023/neurips/Real-Time Motion Prediction via Heterogeneous Polyline Transformer with Relative Pose Encoding create mode 100644 data/2023/neurips/Recasting Continual Learning as Sequence Modeling create mode 100644 data/2023/neurips/Recovering Simultaneously Structured Data via Non-Convex Iteratively Reweighted Least Squares create mode 100644 data/2023/neurips/Recurrent Temporal Revision Graph Networks create mode 100644 data/2023/neurips/Recursion in Recursion: Two-Level Nested Recursion for Length Generalization with Scalability create mode 100644 data/2023/neurips/Red Teaming Deep Neural Networks with Feature Synthesis Tools create mode 100644 data/2023/neurips/Regret Matching+: (In)Stability and Fast Convergence in Games create mode 100644 data/2023/neurips/Regret-Optimal Model-Free Reinforcement Learning for Discounted MDPs with Short Burn-In Time create mode 100644 data/2023/neurips/Rehearsal Learning for Avoiding Undesired Future create mode 100644 data/2023/neurips/Reimagining Synthetic Tabular Data Generation through Data-Centric AI: A Comprehensive Benchmark create mode 100644 data/2023/neurips/Reinforcement Learning for Fine-tuning Text-to-Image Diffusion Models create mode 100644 data/2023/neurips/Reinforcement Learning with Fast and Forgetful Memory create mode 100644 data/2023/neurips/Reining Generalization in Offline Reinforcement Learning via Representation Distinction create mode 100644 data/2023/neurips/Relative Entropic Optimal Transport: a (Prior-aware) Matching Perspective to (Unbalanced) Classification create mode 100644 data/2023/neurips/Reliable Off-Policy Learning for Dosage Combinations create mode 100644 data/2023/neurips/Repetition In Repetition Out: Towards Understanding Neural Text Degeneration from the Data Perspective create mode 100644 data/2023/neurips/Replicability in Reinforcement Learning create mode 100644 data/2023/neurips/Representation Equivalent Neural Operators: a Framework for Alias-free Operator Learning create mode 100644 data/2023/neurips/Reproducibility in Multiple Instance Learning: A Case For Algorithmic Unit Tests create mode 100644 data/2023/neurips/ResShift: Efficient Diffusion Model for Image Super-resolution by Residual Shifting create mode 100644 data/2023/neurips/Resetting the Optimizer in Deep RL: An Empirical Study create mode 100644 data/2023/neurips/Residual Alignment: Uncovering the Mechanisms of Residual Networks create mode 100644 data/2023/neurips/Resilient Constrained Learning create mode 100644 data/2023/neurips/Response Length Perception and Sequence Scheduling: An LLM-Empowered LLM Inference Pipeline create mode 100644 data/2023/neurips/Responsible AI (RAI) Games and Ensembles create mode 100644 data/2023/neurips/Restart Sampling for Improving Generative Processes create mode 100644 data/2023/neurips/Restless Bandits with Average Reward: Breaking the Uniform Global Attractor Assumption create mode 100644 data/2023/neurips/Rethinking Bias Mitigation: Fairer Architectures Make for Fairer Face Recognition create mode 100644 data/2023/neurips/Rethinking Incentives in Recommender Systems: Are Monotone Rewards Always Beneficial? create mode 100644 data/2023/neurips/Revealing the unseen: Benchmarking video action recognition under occlusion create mode 100644 data/2023/neurips/Revisiting Adversarial Robustness Distillation from the Perspective of Robust Fairness create mode 100644 data/2023/neurips/Revisiting Adversarial Training for ImageNet: Architectures, Training and Generalization across Threat Models create mode 100644 data/2023/neurips/Revisiting Out-of-distribution Robustness in NLP: Benchmarks, Analysis, and LLMs Evaluations create mode 100644 data/2023/neurips/Revisiting the Evaluation of Image Synthesis with GANs create mode 100644 data/2023/neurips/Reward Imputation with Sketching for Contextual Batched Bandits create mode 100644 data/2023/neurips/Reward-Directed Conditional Diffusion: Provable Distribution Estimation and Reward Improvement create mode 100644 data/2023/neurips/Rewarded soups: towards Pareto-optimal alignment by interpolating weights fine-tuned on diverse rewards create mode 100644 data/2023/neurips/Rewrite Caption Semantics: Bridging Semantic Gaps for Language-Supervised Semantic Segmentation create mode 100644 data/2023/neurips/Riemannian Residual Neural Networks create mode 100644 data/2023/neurips/Riemannian SAM: Sharpness-Aware Minimization on Riemannian Manifolds create mode 100644 data/2023/neurips/Riemannian stochastic optimization methods avoid strict saddle points create mode 100644 data/2023/neurips/Risk-Averse Active Sensing for Timely Outcome Prediction under Cost Pressure create mode 100644 data/2023/neurips/Robust Bayesian Satisficing create mode 100644 data/2023/neurips/Robust Data Pruning under Label Noise via Maximizing Re-labeling Accuracy create mode 100644 data/2023/neurips/Robust Learning with Progressive Data Expansion Against Spurious Correlation create mode 100644 data/2023/neurips/Robust Matrix Sensing in the Semi-Random Model create mode 100644 data/2023/neurips/Robust Mean Estimation Without Moments for Symmetric Distributions create mode 100644 data/2023/neurips/Robust Model Reasoning and Fitting via Dual Sparsity Pursuit create mode 100644 data/2023/neurips/Robust and Actively Secure Serverless Collaborative Learning create mode 100644 data/2023/neurips/Robust covariance estimation with missing values and cell-wise contamination create mode 100644 data/2023/neurips/Robust low-rank training via approximate orthonormal constraints create mode 100644 data/2023/neurips/SA-Solver: Stochastic Adams Solver for Fast Sampling of Diffusion Models create mode 100644 data/2023/neurips/SALSA VERDE: a machine learning attack on LWE with sparse small secrets create mode 100644 data/2023/neurips/SAME: Uncovering GNN Black Box with Structure-aware Shapley-based Multipiece Explanations create mode 100644 data/2023/neurips/SANFlow: Semantic-Aware Normalizing Flow for Anomaly Detection create mode 100644 data/2023/neurips/SE(3) Diffusion Model-based Point Cloud Registration for Robust 6D Object Pose Estimation create mode 100644 data/2023/neurips/SE(3) Equivariant Augmented Coupling Flows create mode 100644 data/2023/neurips/SEEDS: Exponential SDE Solvers for Fast High-Quality Sampling from Diffusion Models create mode 100644 data/2023/neurips/SEVA: Leveraging sketches to evaluate alignment between human and machine visual abstraction create mode 100644 data/2023/neurips/SLM: A Smoothed First-Order Lagrangian Method for Structured Constrained Nonconvex Optimization create mode 100644 data/2023/neurips/SLaM: Student-Label Mixing for Distillation with Unlabeled Examples create mode 100644 data/2023/neurips/SOC: Semantic-Assisted Object Cluster for Referring Video Object Segmentation create mode 100644 data/2023/neurips/SPA: A Graph Spectral Alignment Perspective for Domain Adaptation create mode 100644 data/2023/neurips/SPACE: Single-round Participant Amalgamation for Contribution Evaluation in Federated Learning create mode 100644 data/2023/neurips/SPQR: Controlling Q-ensemble Independence with Spiked Random Model for Reinforcement Learning create mode 100644 data/2023/neurips/STARSS23: An Audio-Visual Dataset of Spatial Recordings of Real Scenes with Spatiotemporal Annotations of Sound Events create mode 100644 data/2023/neurips/STEVE-1: A Generative Model for Text-to-Behavior in Minecraft create mode 100644 data/2023/neurips/SUPA: A Lightweight Diagnostic Simulator for Machine Learning in Particle Physics create mode 100644 data/2023/neurips/SaVeNet: A Scalable Vector Network for Enhanced Molecular Representation Learning create mode 100644 data/2023/neurips/SafeDICE: Offline Safe Imitation Learning with Non-Preferred Demonstrations create mode 100644 data/2023/neurips/Safety Verification of Decision-Tree Policies in Continuous Time create mode 100644 data/2023/neurips/Sample Complexity Bounds for Score-Matching: Causal Discovery and Generative Modeling create mode 100644 data/2023/neurips/Sample Complexity of Goal-Conditioned Hierarchical Reinforcement Learning create mode 100644 data/2023/neurips/Sample-efficient Multi-objective Molecular Optimization with GFlowNets create mode 100644 data/2023/neurips/SatBird: a Dataset for Bird Species Distribution Modeling using Remote Sensing and Citizen Science Data create mode 100644 data/2023/neurips/SatLM: Satisfiability-Aided Language Models Using Declarative Prompting create mode 100644 data/2023/neurips/Scalable 3D Captioning with Pretrained Models create mode 100644 data/2023/neurips/Scalable Fair Influence Maximization create mode 100644 data/2023/neurips/ScaleLong: Towards More Stable Training of Diffusion Model via Scaling Network Long Skip Connection create mode 100644 data/2023/neurips/Scaling Open-Vocabulary Object Detection create mode 100644 data/2023/neurips/Scaling Riemannian Diffusion Models create mode 100644 data/2023/neurips/Scaling laws for language encoding models in fMRI create mode 100644 data/2023/neurips/Scan and Snap: Understanding Training Dynamics and Token Composition in 1-layer Transformer create mode 100644 data/2023/neurips/Scenario Diffusion: Controllable Driving Scenario Generation With Diffusion create mode 100644 data/2023/neurips/Searching for Optimal Per-Coordinate Step-sizes with Multidimensional Backtracking create mode 100644 data/2023/neurips/Secure Out-of-Distribution Task Generalization with Energy-Based Models create mode 100644 data/2023/neurips/Seeing is not Believing: Robust Reinforcement Learning against Spurious Correlation create mode 100644 data/2023/neurips/SegRefiner: Towards Model-Agnostic Segmentation Refinement with Discrete Diffusion Process create mode 100644 data/2023/neurips/Segment Anything in High Quality create mode 100644 data/2023/neurips/Selective Amnesia: A Continual Learning Approach to Forgetting in Deep Generative Models create mode 100644 data/2023/neurips/Selectively Sharing Experiences Improves Multi-Agent Reinforcement Learning create mode 100644 data/2023/neurips/Self-Adaptive Motion Tracking against On-body Displacement of Flexible Sensors create mode 100644 data/2023/neurips/Self-Chained Image-Language Model for Video Localization and Question Answering create mode 100644 data/2023/neurips/Self-Correcting Bayesian Optimization through Bayesian Active Learning create mode 100644 data/2023/neurips/Self-Predictive Universal AI create mode 100644 data/2023/neurips/Self-supervised video pretraining yields robust and more human-aligned visual representations create mode 100644 data/2023/neurips/Sequential Subset Matching for Dataset Distillation create mode 100644 data/2023/neurips/Setting the Trap: Capturing and Defeating Backdoors in Pretrained Language Models through Honeypots create mode 100644 data/2023/neurips/Shape Non-rigid Kinematics (SNK): A Zero-Shot Method for Non-Rigid Shape Matching via Unsupervised Functional Map Regularized Reconstruction create mode 100644 data/2023/neurips/Sharpness Minimization Algorithms Do Not Only Minimize Sharpness To Achieve Better Generalization create mode 100644 data/2023/neurips/Should Under-parameterized Student Networks Copy or Average Teacher Weights? create mode 100644 data/2023/neurips/SimMMDG: A Simple and Effective Framework for Multi-modal Domain Generalization create mode 100644 data/2023/neurips/Simple, Scalable and Effective Clustering via One-Dimensional Projections create mode 100644 data/2023/neurips/Simplifying and Empowering Transformers for Large-Graph Representations create mode 100644 data/2023/neurips/Single-Call Stochastic Extragradient Methods for Structured Non-monotone Variational Inequalities: Improved Analysis under Weaker Conditions create mode 100644 data/2023/neurips/SituatedGen: Incorporating Geographical and Temporal Contexts into Generative Commonsense Reasoning create mode 100644 data/2023/neurips/Sketchy: Memory-efficient Adaptive Regularization with Frequent Directions create mode 100644 data/2023/neurips/Slot-guided Volumetric Object Radiance Fields create mode 100644 data/2023/neurips/Small batch deep reinforcement learning create mode 100644 data/2023/neurips/Smooth, exact rotational symmetrization for deep learning on point clouds create mode 100644 data/2023/neurips/Smoothed Analysis of Sequential Probability Assignment create mode 100644 data/2023/neurips/SoTTA: Robust Test-Time Adaptation on Noisy Data Streams create mode 100644 data/2023/neurips/Social Motion Prediction with Cognitive Hierarchies create mode 100644 data/2023/neurips/Softmax Output Approximation for Activation Memory-Efficient Training of Attention-based Networks create mode 100644 data/2023/neurips/Solving a Class of Non-Convex Minimax Optimization in Federated Learning create mode 100644 data/2023/neurips/Sparse Deep Learning for Time Series Data: Theory and Applications create mode 100644 data/2023/neurips/Sparse Modular Activation for Efficient Sequence Modeling create mode 100644 data/2023/neurips/Sparse Parameterization for Epitomic Dataset Distillation create mode 100644 data/2023/neurips/Sparsity-Preserving Differentially Private Training of Large Embedding Models create mode 100644 data/2023/neurips/Spatial-frequency channels, shape bias, and adversarial robustness create mode 100644 data/2023/neurips/Spatially Resolved Gene Expression Prediction from Histology Images via Bi-modal Contrastive Learning create mode 100644 data/2023/neurips/Spectral Entry-wise Matrix Estimation for Low-Rank Reinforcement Learning create mode 100644 data/2023/neurips/Spike-driven Transformer create mode 100644 data/2023/neurips/Spontaneous symmetry breaking in generative diffusion models create mode 100644 data/2023/neurips/Squared Neural Families: A New Class of Tractable Density Models create mode 100644 data/2023/neurips/Squeeze, Recover and Relabel: Dataset Condensation at ImageNet Scale From A New Perspective create mode 100644 data/2023/neurips/Stabilizing the Optimization of Neural Signed Distance Functions and Finer Shape Representation create mode 100644 data/2023/neurips/Stable Bias: Evaluating Societal Representations in Diffusion Models create mode 100644 data/2023/neurips/Stable Vectorization of Multiparameter Persistent Homology using Signed Barcodes as Measures create mode 100644 data/2023/neurips/StableFDG: Style and Attention Based Learning for Federated Domain Generalization create mode 100644 data/2023/neurips/State Sequences Prediction via Fourier Transform for Representation Learning create mode 100644 data/2023/neurips/State-Action Similarity-Based Representations for Off-Policy Evaluation create mode 100644 data/2023/neurips/State-space models with layer-wise nonlinearity are universal approximators with exponential decaying memory create mode 100644 data/2023/neurips/State2Explanation: Concept-Based Explanations to Benefit Agent Learning and User Understanding create mode 100644 data/2023/neurips/Static and Sequential Malicious Attacks in the Context of Selective Forgetting create mode 100644 data/2023/neurips/Statistical Analysis of Quantum State Learning Process in Quantum Neural Networks create mode 100644 data/2023/neurips/Statistical Knowledge Assessment for Large Language Models create mode 100644 data/2023/neurips/Statistical and Computational Trade-off in Multi-Agent Multi-Armed Bandits create mode 100644 data/2023/neurips/Statistically Valid Variable Importance Assessment through Conditional Permutations create mode 100644 "data/2023/neurips/Stein \316\240-Importance Sampling" create mode 100644 data/2023/neurips/Stochastic Optimal Control for Collective Variable Free Sampling of Molecular Transition Paths create mode 100644 data/2023/neurips/StoryBench: A Multifaceted Benchmark for Continuous Story Visualization create mode 100644 data/2023/neurips/Strategic Apple Tasting create mode 100644 data/2023/neurips/Streaming Algorithms and Lower Bounds for Estimating Correlation Clustering Cost create mode 100644 data/2023/neurips/Streaming PCA for Markovian Data create mode 100644 data/2023/neurips/Strong and Precise Modulation of Human Percepts via Robustified ANNs create mode 100644 data/2023/neurips/Structure from Duplicates: Neural Inverse Graphics from a Pile of Objects create mode 100644 data/2023/neurips/Structure of universal formulas create mode 100644 data/2023/neurips/Structured Neural Networks for Density Estimation and Causal Inference create mode 100644 data/2023/neurips/Structured Neural-PI Control with End-to-End Stability and Output Tracking Guarantees create mode 100644 data/2023/neurips/Structured Prediction with Stronger Consistency Guarantees create mode 100644 data/2023/neurips/StyleDrop: Text-to-Image Synthesis of Any Style create mode 100644 data/2023/neurips/StyleGAN knows Normal, Depth, Albedo, and More create mode 100644 data/2023/neurips/Sub-optimality of the Naive Mean Field approximation for proportional high-dimensional Linear Regression create mode 100644 data/2023/neurips/Subclass-Dominant Label Noise: A Counterexample for the Success of Early Stopping create mode 100644 data/2023/neurips/Successor-Predecessor Intrinsic Exploration create mode 100644 data/2023/neurips/Suggesting Variable Order for Cylindrical Algebraic Decomposition via Reinforcement Learning create mode 100644 data/2023/neurips/Survival Permanental Processes for Survival Analysis with Time-Varying Covariates create mode 100644 data/2023/neurips/SustainGym: Reinforcement Learning Environments for Sustainable Energy Systems create mode 100644 data/2023/neurips/SwapPrompt: Test-Time Prompt Adaptation for Vision-Language Models create mode 100644 data/2023/neurips/Swarm Reinforcement Learning for Adaptive Mesh Refinement create mode 100644 data/2023/neurips/SwiftSage: A Generative Agent with Fast and Slow Thinking for Complex Interactive Tasks create mode 100644 data/2023/neurips/Symmetry-Informed Geometric Representation for Molecules, Proteins, and Crystalline Materials create mode 100644 data/2023/neurips/Synthetic-to-Real Pose Estimation with Geometric Reconstruction create mode 100644 data/2023/neurips/Systematic Visual Reasoning through Object-Centric Relational Abstraction create mode 100644 data/2023/neurips/T2I-CompBench: A Comprehensive Benchmark for Open-world Compositional Text-to-image Generation create mode 100644 data/2023/neurips/TACO: Temporal Latent Action-Driven Contrastive Loss for Visual Reinforcement Learning create mode 100644 data/2023/neurips/TFLEX: Temporal Feature-Logic Embedding Framework for Complex Reasoning over Temporal Knowledge Graph create mode 100644 data/2023/neurips/TOA: Task-oriented Active VQA create mode 100644 data/2023/neurips/TRIAGE: Characterizing and auditing training data for improved regression create mode 100644 data/2023/neurips/Tackling Heavy-Tailed Rewards in Reinforcement Learning with Function Approximation: Minimax Optimal and Instance-Dependent Regret Bounds create mode 100644 data/2023/neurips/Tailoring Self-Attention for Graph via Rooted Subtrees create mode 100644 data/2023/neurips/Taming Local Effects in Graph-based Spatiotemporal Forecasting create mode 100644 data/2023/neurips/Tanimoto Random Features for Scalable Molecular Machine Learning create mode 100644 data/2023/neurips/Task Arithmetic in the Tangent Space: Improved Editing of Pre-Trained Models create mode 100644 data/2023/neurips/Temperature Balancing, Layer-wise Weight Analysis, and Neural Network Training create mode 100644 data/2023/neurips/Temporal Continual Learning with Prior Compensation for Human Motion Prediction create mode 100644 data/2023/neurips/Test-Time Amendment with a Coarse Classifier for Fine-Grained Classification create mode 100644 data/2023/neurips/Test-time Training for Matching-based Video Object Segmentation create mode 100644 data/2023/neurips/Text Alignment Is An Efficient Unified Model for Massive NLP Tasks create mode 100644 data/2023/neurips/Textually Pretrained Speech Language Models create mode 100644 data/2023/neurips/The Behavior and Convergence of Local Bayesian Optimization create mode 100644 data/2023/neurips/The Best of Both Worlds in Network Population Games: Reaching Consensus and Convergence to Equilibrium create mode 100644 data/2023/neurips/The Crucial Role of Normalization in Sharpness-Aware Minimization create mode 100644 data/2023/neurips/The Curious Price of Distributional Robustness in Reinforcement Learning with a Generative Model create mode 100644 data/2023/neurips/The Distortion of Binomial Voting Defies Expectation create mode 100644 data/2023/neurips/The Equivalence of Dynamic and Strategic Stability under Regularized Learning in Games create mode 100644 data/2023/neurips/The Gain from Ordering in Online Learning create mode 100644 data/2023/neurips/The Grand Illusion: The Myth of Software Portability and Implications for ML Progress create mode 100644 data/2023/neurips/The Graph Pencil Method: Mapping Subgraph Densities to Stochastic Block Models create mode 100644 data/2023/neurips/The Learnability of In-Context Learning create mode 100644 data/2023/neurips/The RefinedWeb Dataset for Falcon LLM: Outperforming Curated Corpora with Web Data Only create mode 100644 data/2023/neurips/The Rise of AI Language Pathologists: Exploring Two-level Prompt Learning for Few-shot Weakly-supervised Whole Slide Image Classification create mode 100644 data/2023/neurips/The Transient Nature of Emergent In-Context Learning in Transformers create mode 100644 data/2023/neurips/The Tunnel Effect: Building Data Representations in Deep Neural Networks create mode 100644 data/2023/neurips/The expressive power of pooling in Graph Neural Networks create mode 100644 data/2023/neurips/The noise level in linear regression with dependent data create mode 100644 data/2023/neurips/The probability flow ODE is provably fast create mode 100644 data/2023/neurips/The s-value: evaluating stability with respect to distributional shifts create mode 100644 data/2023/neurips/Theoretical Analysis of the Inductive Biases in Deep Convolutional Networks create mode 100644 data/2023/neurips/Three-Way Trade-Off in Multi-Objective Learning: Optimization, Generalization and Conflict-Avoidance create mode 100644 data/2023/neurips/Thrust: Adaptively Propels Large Language Models with External Knowledge create mode 100644 data/2023/neurips/Tight Bounds for Volumetric Spanners and Applications create mode 100644 data/2023/neurips/Tight Risk Bounds for Gradient Descent on Separable Data create mode 100644 data/2023/neurips/Time Series as Images: Vision Transformer for Irregularly Sampled Time Series create mode 100644 data/2023/neurips/Toolformer: Language Models Can Teach Themselves to Use Tools create mode 100644 data/2023/neurips/Top-Ambiguity Samples Matter: Understanding Why Deep Ensemble Works in Selective Classification create mode 100644 data/2023/neurips/TopoSRL: Topology preserving self-supervised Simplicial Representation Learning create mode 100644 data/2023/neurips/Topological RANSAC for instance verification and retrieval without fine-tuning create mode 100644 data/2023/neurips/Towards Better Dynamic Graph Learning: New Architecture and Unified Library create mode 100644 data/2023/neurips/Towards Data-Agnostic Pruning At Initialization: What Makes a Good Sparse Mask? create mode 100644 data/2023/neurips/Towards Data-Algorithm Dependent Generalization: a Case Study on Overparameterized Linear Regression create mode 100644 data/2023/neurips/Towards Distribution-Agnostic Generalized Category Discovery create mode 100644 data/2023/neurips/Towards Foundation Models for Scientific Machine Learning: Characterizing Scaling and Transfer Behavior create mode 100644 data/2023/neurips/Towards Higher Ranks via Adversarial Weight Pruning create mode 100644 data/2023/neurips/Towards In-context Scene Understanding create mode 100644 data/2023/neurips/Towards Label-free Scene Understanding by Vision Foundation Models create mode 100644 data/2023/neurips/Towards Last-layer Retraining for Group Robustness with Fewer Annotations create mode 100644 data/2023/neurips/Towards Revealing the Mystery behind Chain of Thought: A Theoretical Perspective create mode 100644 data/2023/neurips/Towards Semi-Structured Automatic ICD Coding via Tree-based Contrastive Learning create mode 100644 data/2023/neurips/Towards Stable Backdoor Purification through Feature Shift Tuning create mode 100644 data/2023/neurips/Towards a Comprehensive Benchmark for High-Level Synthesis Targeted to FPGAs create mode 100644 data/2023/neurips/Towards a Unified Analysis of Kernel-based Methods Under Covariate Shift create mode 100644 data/2023/neurips/Towards a Unified Framework of Contrastive Learning for Disentangled Representations create mode 100644 data/2023/neurips/Towards a fuller understanding of neurons with Clustered Compositional Explanations create mode 100644 data/2023/neurips/TpuGraphs: A Performance Prediction Dataset on Large Tensor Computational Graphs create mode 100644 data/2023/neurips/Train Hard, Fight Easy: Robust Meta Reinforcement Learning create mode 100644 data/2023/neurips/Training Chain-of-Thought via Latent-Variable Inference create mode 100644 "data/2023/neurips/Training Fully Connected Neural Networks is \342\210\203R-Complete" create mode 100644 data/2023/neurips/Training on Foveated Images Improves Robustness to Adversarial Attacks create mode 100644 data/2023/neurips/Training-free Diffusion Model Adaptation for Variable-Sized Text-to-Image Synthesis create mode 100644 data/2023/neurips/Trajectory Alignment: Understanding the Edge of Stability Phenomenon via Bifurcation Theory create mode 100644 data/2023/neurips/Transformed Low-Rank Parameterization Can Help Robust Generalization for Tensor Neural Networks create mode 100644 data/2023/neurips/Transient Neural Radiance Fields for Lidar View Synthesis and 3D Reconstruction create mode 100644 data/2023/neurips/TriRE: A Multi-Mechanism Learning Paradigm for Continual Knowledge Retention and Promotion create mode 100644 data/2023/neurips/Trial matching: capturing variability with data-constrained spiking neural networks create mode 100644 data/2023/neurips/TrojLLM: A Black-box Trojan Prompt Attack on Large Language Models create mode 100644 data/2023/neurips/Turbulence in Focus: Benchmarking Scaling Behavior of 3D Volumetric Super-Resolution with BLASTNet 2.0 Data create mode 100644 data/2023/neurips/Two Sides of One Coin: the Limits of Untuned SGD and the Power of Adaptive Methods create mode 100644 data/2023/neurips/Two-Stage Learning to Defer with Multiple Experts create mode 100644 data/2023/neurips/Type-to-Track: Retrieve Any Object via Prompt-based Tracking create mode 100644 data/2023/neurips/UDC-SIT: A Real-World Dataset for Under-Display Cameras create mode 100644 data/2023/neurips/UE4-NeRF: Neural Radiance Field for Real-Time Rendering of Large-Scale Scene create mode 100644 data/2023/neurips/UP-NeRF: Unconstrained Pose Prior-Free Neural Radiance Field create mode 100644 data/2023/neurips/UltraRE: Enhancing RecEraser for Recommendation Unlearning via Error Decomposition create mode 100644 data/2023/neurips/Unbiased learning of deep generative models with structured discrete representations create mode 100644 data/2023/neurips/Uncertainty-Aware Alignment Network for Cross-Domain Video-Text Retrieval create mode 100644 data/2023/neurips/Uncertainty-Aware Instance Reweighting for Off-Policy Learning create mode 100644 data/2023/neurips/Unconstrained Dynamic Regret via Sparse Coding create mode 100644 data/2023/neurips/Uncovering Prototypical Knowledge for Weakly Open-Vocabulary Semantic Segmentation create mode 100644 data/2023/neurips/Uncovering the Hidden Dynamics of Video Self-supervised Learning under Distribution Shifts create mode 100644 data/2023/neurips/Understanding Diffusion Objectives as the ELBO with Simple Data Augmentation create mode 100644 data/2023/neurips/Understanding Few-Shot Learning: Measuring Task Relatedness and Adaptation Difficulty via Attributes create mode 100644 data/2023/neurips/Understanding How Consistency Works in Federated Learning via Stage-wise Relaxed Initialization create mode 100644 data/2023/neurips/Understanding and Improving Feature Learning for Out-of-Distribution Generalization create mode 100644 data/2023/neurips/Understanding the Limitations of Deep Models for Molecular property prediction: Insights and Solutions create mode 100644 data/2023/neurips/Understanding the detrimental class-level effects of data augmentation create mode 100644 data/2023/neurips/UniPC: A Unified Predictor-Corrector Framework for Fast Sampling of Diffusion Models create mode 100644 data/2023/neurips/Uniform Convergence with Square-Root Lipschitz Loss create mode 100644 data/2023/neurips/Universality and Limitations of Prompt Tuning create mode 100644 data/2023/neurips/Unleashing the Power of Randomization in Auditing Differentially Private ML create mode 100644 data/2023/neurips/Unsupervised Anomaly Detection with Rejection create mode 100644 data/2023/neurips/Unsupervised Graph Neural Architecture Search with Disentangled Self-Supervision create mode 100644 data/2023/neurips/Unsupervised Image Denoising with Score Function create mode 100644 data/2023/neurips/Unsupervised Polychromatic Neural Representation for CT Metal Artifact Reduction create mode 100644 data/2023/neurips/Using Imperfect Surrogates for Downstream Inference: Design-based Supervised Learning for Social Science Applications of Large Language Models create mode 100644 data/2023/neurips/Utilitarian Algorithm Configuration create mode 100644 data/2023/neurips/VAST: A Vision-Audio-Subtitle-Text Omni-Modality Foundation Model and Dataset create mode 100644 data/2023/neurips/VLATTACK: Multimodal Adversarial Attacks on Vision-Language Tasks via Pre-trained Models create mode 100644 data/2023/neurips/VaRT: Variational Regression Trees create mode 100644 data/2023/neurips/Variational Annealing on Graphs for Combinatorial Optimization create mode 100644 data/2023/neurips/Video Prediction Models as Rewards for Reinforcement Learning create mode 100644 data/2023/neurips/Video-Mined Task Graphs for Keystep Recognition in Instructional Videos create mode 100644 data/2023/neurips/VisAlign: Dataset for Measuring the Alignment between AI and Humans in Visual Perception create mode 100644 data/2023/neurips/VisoGender: A dataset for benchmarking gender bias in image-text pronoun resolution create mode 100644 data/2023/neurips/Visual Instruction Inversion: Image Editing via Image Prompting create mode 100644 data/2023/neurips/Volume Feature Rendering for Fast Neural Radiance Field Reconstruction create mode 100644 "data/2023/neurips/Wasserstein Quantum Monte Carlo: A Novel Approach for Solving the Quantum Many-Body Schr\303\266dinger Equation" create mode 100644 data/2023/neurips/Waypoint Transformer: Reinforcement Learning via Supervised Learning with Intermediate Targets create mode 100644 data/2023/neurips/What Can We Learn from Unlearnable Datasets? create mode 100644 data/2023/neurips/What Truly Matters in Trajectory Prediction for Autonomous Driving? create mode 100644 data/2023/neurips/What a MESS: Multi-Domain Evaluation of Zero-Shot Semantic Segmentation create mode 100644 data/2023/neurips/When Demonstrations meet Generative World Models: A Maximum Likelihood Framework for Offline Inverse Reinforcement Learning create mode 100644 data/2023/neurips/When Do Neural Nets Outperform Boosted Trees on Tabular Data? create mode 100644 data/2023/neurips/When Does Confidence-Based Cascade Deferral Suffice? create mode 100644 data/2023/neurips/When Does Optimizing a Proper Loss Yield Calibration? create mode 100644 data/2023/neurips/Where Did I Come From? Origin Attribution of AI-Generated Images create mode 100644 data/2023/neurips/Where are we in the search for an Artificial Visual Cortex for Embodied Intelligence? create mode 100644 data/2023/neurips/Where2Explore: Few-shot Affordance Learning for Unseen Novel Categories of Articulated Objects create mode 100644 data/2023/neurips/Why Does Sharpness-Aware Minimization Generalize Better Than SGD? create mode 100644 data/2023/neurips/Why think step by step? Reasoning emerges from the locality of experience create mode 100644 data/2023/neurips/WildfireSpreadTS: A dataset of multi-modal time series for wildfire spread prediction create mode 100644 data/2023/neurips/Winner-Take-All Column Row Sampling for Memory Efficient Adaptation of Language Model create mode 100644 data/2023/neurips/Worst-case Performance of Popular Approximate Nearest Neighbor Search Implementations: Guarantees and Limitations create mode 100644 data/2023/neurips/XAGen: 3D Expressive Human Avatars Generation create mode 100644 data/2023/neurips/Zero-One Laws of Graph Neural Networks create mode 100644 data/2023/neurips/ZipLM: Inference-Aware Structured Pruning of Language Models create mode 100644 data/2023/neurips/f-Policy Gradients: A General Framework for Goal-Conditioned RL using f-Divergences create mode 100644 data/2023/neurips/k-Median Clustering via Metric Embedding: Towards Better Initialization with Differential Privacy create mode 100644 data/2023/neurips/rPPG-Toolbox: Deep Remote PPG Toolbox create mode 100644 data/2023/neurips/xTrimoGene: An Efficient and Scalable Representation Learner for Single-Cell RNA-Seq Data diff --git a/data/2018/vldb/Declarative Recursive Computation on an RDBMS b/data/2018/vldb/Declarative Recursive Computation on an RDBMS new file mode 100644 index 0000000000..2e630faf78 --- /dev/null +++ b/data/2018/vldb/Declarative Recursive Computation on an RDBMS @@ -0,0 +1 @@ +We explore the close relationship between the tensor-based computations performed during modern machine learning, and relational database computations. We consider how to make a very small set of changes to a modern RDBMS to make it suitable for distributed learning computations. Changes include adding better support for recursion, and optimization and execution of very large compute plans. We also show that there are key advantages to using an RDBMS as a machine learning platform. In particular, DBMSbased learning allows for trivial scaling to large data sets and especially large models, where different computational units operate on different parts of a model that may be too large to fit into RAM. \ No newline at end of file diff --git a/data/2020/neurips/(De)Randomized Smoothing for Certifiable Defense against Patch Attacks b/data/2020/neurips/(De)Randomized Smoothing for Certifiable Defense against Patch Attacks new file mode 100644 index 0000000000..c9f427f1b6 --- /dev/null +++ b/data/2020/neurips/(De)Randomized Smoothing for Certifiable Defense against Patch Attacks @@ -0,0 +1 @@ +Patch adversarial attacks on images, in which the attacker can distort pixels within a region of bounded size, are an important threat model since they provide a quantitative model for physical adversarial attacks. In this paper, we introduce a certifiable defense against patch attacks that guarantees for a given image and patch attack size, no patch adversarial examples exist. Our method is related to the broad class of randomized smoothing robustness schemes which provide high-confidence probabilistic robustness certificates. By exploiting the fact that patch attacks are more constrained than general sparse attacks, we derive meaningfully large robustness certificates. Additionally, the algorithm we propose is de-randomized, providing deterministic certificates. To the best of our knowledge, there exists only one prior method for certifiable defense against patch attacks, which relies on interval bound propagation. While this sole existing method performs well on MNIST, it has several limitations: it requires computationally expensive training, does not scale to ImageNet, and performs poorly on CIFAR-10. In contrast, our proposed method effectively addresses all of these issues: our classifier can be trained quickly, achieves high clean and certified robust accuracy on CIFAR-10, and provides certificates at the ImageNet scale. For example, for a 5*5 patch attack on CIFAR-10, our method achieves up to around 57.8% certified accuracy (with a classifier around 83.9% clean accuracy), compared to at most 30.3% certified accuracy for the existing method (with a classifier with around 47.8% clean accuracy), effectively establishing a new state-of-the-art. Code is available at this https URL. \ No newline at end of file diff --git a/data/2020/neurips/3D Multi-bodies: Fitting Sets of Plausible 3D Human Models to Ambiguous Image Data b/data/2020/neurips/3D Multi-bodies: Fitting Sets of Plausible 3D Human Models to Ambiguous Image Data new file mode 100644 index 0000000000..c95e7ae53c --- /dev/null +++ b/data/2020/neurips/3D Multi-bodies: Fitting Sets of Plausible 3D Human Models to Ambiguous Image Data @@ -0,0 +1 @@ +We consider the problem of obtaining dense 3D reconstructions of humans from single and partially occluded views. In such cases, the visual evidence is usually insufficient to identify a 3D reconstruction uniquely, so we aim at recovering several plausible reconstructions compatible with the input data. We suggest that ambiguities can be modelled more effectively by parametrizing the possible body shapes and poses via a suitable 3D model, such as SMPL for humans. We propose to learn a multi-hypothesis neural network regressor using a best-of-M loss, where each of the M hypotheses is constrained to lie on a manifold of plausible human poses by means of a generative model. We show that our method outperforms alternative approaches in ambiguous pose recovery on standard benchmarks for 3D humans, and in heavily occluded versions of these benchmarks. \ No newline at end of file diff --git a/data/2020/neurips/3D Self-Supervised Methods for Medical Imaging b/data/2020/neurips/3D Self-Supervised Methods for Medical Imaging new file mode 100644 index 0000000000..c9ae299201 --- /dev/null +++ b/data/2020/neurips/3D Self-Supervised Methods for Medical Imaging @@ -0,0 +1 @@ +Self-supervised learning methods have witnessed a recent surge of interest after proving successful in multiple application fields. In this work, we leverage these techniques, and we propose 3D versions for five different self-supervised methods, in the form of proxy tasks. Our methods facilitate neural network feature learning from unlabeled 3D images, aiming to reduce the required cost for expert annotation. The developed algorithms are 3D Contrastive Predictive Coding, 3D Rotation prediction, 3D Jigsaw puzzles, Relative 3D patch location, and 3D Exemplar networks. Our experiments show that pretraining models with our 3D tasks yields more powerful semantic representations, and enables solving downstream tasks more accurately and efficiently, compared to training the models from scratch and to pretraining them on 2D slices. We demonstrate the effectiveness of our methods on three downstream tasks from the medical imaging domain: i) Brain Tumor Segmentation from 3D MRI, ii) Pancreas Tumor Segmentation from 3D CT, and iii) Diabetic Retinopathy Detection from 2D Fundus images. In each task, we assess the gains in data-efficiency, performance, and speed of convergence. Interestingly, we also find gains when transferring the learned representations, by our methods, from a large unlabeled 3D corpus to a small downstream-specific dataset. We achieve results competitive to state-of-the-art solutions at a fraction of the computational expense. We publish our implementations for the developed algorithms (both 3D and 2D versions) as an open-source library, in an effort to allow other researchers to apply and extend our methods on their datasets. \ No newline at end of file diff --git a/data/2020/neurips/3D Shape Reconstruction from Vision and Touch b/data/2020/neurips/3D Shape Reconstruction from Vision and Touch new file mode 100644 index 0000000000..48ad4b44ee --- /dev/null +++ b/data/2020/neurips/3D Shape Reconstruction from Vision and Touch @@ -0,0 +1 @@ +When a toddler is presented a new toy, their instinctual behaviour is to pick it up and inspect it with their hand and eyes in tandem, clearly searching over its surface to properly understand what they are playing with. Here, touch provides high fidelity localized information while vision provides complementary global context. However, in 3D shape reconstruction, the complementary fusion of visual and haptic modalities remains largely unexplored. In this paper, we study this problem and present an effective chart-based approach to fusing vision and touch, which leverages advances in graph convolutional networks. To do so, we introduce a dataset of simulated touch and vision signals from the interaction between a robotic hand and a large array of 3D objects. Our results show that (1) leveraging both vision and touch signals consistently improves single-modality baselines; (2) our approach outperforms alternative modality fusion methods and strongly benefits from the proposed chart-based structure; (3) the reconstruction quality increases with the number of grasps provided; and (4) the touch information not only enhances the reconstruction at the touch site but also extrapolates to its local neighborhood. \ No newline at end of file diff --git a/data/2020/neurips/A B Testing in Dense Large-Scale Networks: Design and Inference b/data/2020/neurips/A B Testing in Dense Large-Scale Networks: Design and Inference new file mode 100644 index 0000000000..b98a812441 --- /dev/null +++ b/data/2020/neurips/A B Testing in Dense Large-Scale Networks: Design and Inference @@ -0,0 +1 @@ +Design of experiments and estimation of treatment effects in large-scale networks, in the presence of strong interference, is a challenging and important problem. Most existing methods' performance deteriorates as the density of the network increases. In this paper, we present a novel strategy for accurately estimating the causal effects of a class of treatments in a dense large-scale network. First, we design an approximate randomized controlled experiment by solving an optimization problem to allocate treatments in the presence of competition among neighboring nodes. Then we apply an importance sampling adjustment to correct for any leftover bias (from the approximation) in estimating average treatment effects. We provide theoretical guarantees, verify robustness in a simulation study, and validate the scalability and usefulness of our procedure in a real-world experiment on a large social network. \ No newline at end of file diff --git a/data/2020/neurips/A Bandit Learning Algorithm and Applications to Auction Design b/data/2020/neurips/A Bandit Learning Algorithm and Applications to Auction Design new file mode 100644 index 0000000000..0621844e6a --- /dev/null +++ b/data/2020/neurips/A Bandit Learning Algorithm and Applications to Auction Design @@ -0,0 +1 @@ +We consider online bandit learning in which at every time step, an algorithm has to make a decision and then observe only its reward. The goal is to design efficient (polynomial-time) algorithms that achieve a total reward approximately close to that of the best fixed decision in hindsight. In this paper, we introduce a new notion of ( λ, µ ) -concave functions and present a bandit learning algorithm that achieves a performance guarantee which is characterized as a function of the concavity parameters λ and µ . The algorithm is based on the mirror descent algorithm in which the update directions follow the gradient of the multilinear extensions of the reward functions. The regret bound induced by our algorithm is (cid:101) O ( √ T ) which is nearly optimal. We apply our algorithm to auction design, specifically to welfare maximization, revenue maximization, and no-envy learning in auctions. In welfare maximization, we show that a version of fictitious play in smooth auctions guarantees a competitive regret bound which is determined by the smooth parameters. In revenue maximization, we consider the simultaneous second-price auctions with reserve prices in multi-parameter environments. We give a bandit algorithm which achieves the total revenue at least 1 / 2 times that of the best fixed reserve prices in hind-sight. In no-envy learning, we study the bandit item selection problem where the player valuation is submodular and provide an efficient 1 / 2 -approximation no-envy algorithm. \ No newline at end of file diff --git a/data/2020/neurips/A Bayesian Nonparametrics View into Deep Representations b/data/2020/neurips/A Bayesian Nonparametrics View into Deep Representations new file mode 100644 index 0000000000..388b753b29 --- /dev/null +++ b/data/2020/neurips/A Bayesian Nonparametrics View into Deep Representations @@ -0,0 +1 @@ +We investigate neural network representations from a probabilistic perspective. Specifically, we leverage Bayesian nonparametrics to construct models of neural activations in Convolutional Neural Networks (CNNs) and latent representations in Variational Autoencoders (VAEs). This allows us to formulate a tractable complexity measure for distributions of neural activations and to explore global structure of latent spaces learned by VAEs. We use this machinery to uncover how memorization and two common forms of regularization, i.e. dropout and input augmentation, influence representational complexity in CNNs. We demonstrate that networks that can exploit patterns in data learn vastly less complex representations than networks forced to memorize. We also show marked differences between effects of input augmentation and dropout, with the latter strongly depending on network width. Next, we investigate latent representations learned by standard β-VAEs and Maximum Mean Discrepancy (MMD) β-VAEs. We show that aggregated posterior in standard VAEs quickly collapses to the diagonal prior when regularization strength increases. MMD-VAEs, on the other hand, learn more complex posterior distributions, even with strong regularization. While this gives a richer sample space, MMD-VAEs do not exhibit independence of latent dimensions. Finally, we leverage our probabilistic models as an effective sampling strategy for latent codes, improving quality of samples in VAEs with rich posteriors. \ No newline at end of file diff --git a/data/2020/neurips/A Bayesian Perspective on Training Speed and Model Selection b/data/2020/neurips/A Bayesian Perspective on Training Speed and Model Selection new file mode 100644 index 0000000000..f5d734b333 --- /dev/null +++ b/data/2020/neurips/A Bayesian Perspective on Training Speed and Model Selection @@ -0,0 +1 @@ +We take a Bayesian perspective to illustrate a connection between training speed and the marginal likelihood in linear models. This provides two major insights: first, that a measure of a model's training speed can be used to estimate its marginal likelihood. Second, that this measure, under certain conditions, predicts the relative weighting of models in linear model combinations trained to minimize a regression loss. We verify our results in model selection tasks for linear models and for the infinite-width limit of deep neural networks. We further provide encouraging empirical evidence that the intuition developed in these settings also holds for deep neural networks trained with stochastic gradient descent. Our results suggest a promising new direction towards explaining why neural networks trained with stochastic gradient descent are biased towards functions that generalize well. \ No newline at end of file diff --git a/data/2020/neurips/A Benchmark for Systematic Generalization in Grounded Language Understanding b/data/2020/neurips/A Benchmark for Systematic Generalization in Grounded Language Understanding new file mode 100644 index 0000000000..bf124dbf81 --- /dev/null +++ b/data/2020/neurips/A Benchmark for Systematic Generalization in Grounded Language Understanding @@ -0,0 +1 @@ +Human language users easily interpret expressions that describe unfamiliar situations composed from familiar parts ("greet the pink brontosaurus by the ferris wheel"). Modern neural networks, by contrast, struggle to interpret compositions unseen in training. In this paper, we introduce a new benchmark, gSCAN, for evaluating compositional generalization in models of situated language understanding. We take inspiration from standard models of meaning composition in formal linguistics. Going beyond an earlier related benchmark that focused on syntactic aspects of generalization, gSCAN defines a language grounded in the states of a grid world. This allows us to build novel generalization tasks that probe the acquisition of linguistically motivated rules. For example, agents must understand how adjectives such as 'small' are interpreted relative to the current world state or how adverbs such as 'cautiously' combine with new verbs. We test a strong multi-modal baseline model and a state-of-the-art compositional method finding that, in most cases, they fail dramatically when generalization requires systematic compositional rules. \ No newline at end of file diff --git a/data/2020/neurips/A Biologically Plausible Neural Network for Slow Feature Analysis b/data/2020/neurips/A Biologically Plausible Neural Network for Slow Feature Analysis new file mode 100644 index 0000000000..ded41f4c1f --- /dev/null +++ b/data/2020/neurips/A Biologically Plausible Neural Network for Slow Feature Analysis @@ -0,0 +1 @@ +Learning latent features from time series data is an important problem in both machine learning and brain function. One approach, called Slow Feature Analysis (SFA), leverages the slowness of many salient features relative to the rapidly varying input signals. Furthermore, when trained on naturalistic stimuli, SFA reproduces interesting properties of cells in the primary visual cortex and hippocampus, suggesting that the brain uses temporal slowness as a computational principle for learning latent features. However, despite the potential relevance of SFA for modeling brain function, there is currently no SFA algorithm with a biologically plausible neural network implementation, by which we mean an algorithm operates in the online setting and can be mapped onto a neural network with local synaptic updates. In this work, starting from an SFA objective, we derive an SFA algorithm, called Bio-SFA, with a biologically plausible neural network implementation. We validate Bio-SFA on naturalistic stimuli. \ No newline at end of file diff --git a/data/2020/neurips/A Boolean Task Algebra for Reinforcement Learning b/data/2020/neurips/A Boolean Task Algebra for Reinforcement Learning new file mode 100644 index 0000000000..7858a160d8 --- /dev/null +++ b/data/2020/neurips/A Boolean Task Algebra for Reinforcement Learning @@ -0,0 +1 @@ +We propose a framework for defining a Boolean algebra over the space of tasks. This allows us to formulate new tasks in terms of the negation, disjunction and conjunction of a set of base tasks. We then show that by learning goal-oriented value functions and restricting the transition dynamics of the tasks, an agent can solve these new tasks with no further learning. We prove that by composing these value functions in specific ways, we immediately recover the optimal policies for all tasks expressible under the Boolean algebra. We verify our approach in two domains, including a high-dimensional video game environment requiring function approximation, where an agent first learns a set of base skills, and then composes them to solve a super-exponential number of new tasks. \ No newline at end of file diff --git a/data/2020/neurips/A Catalyst Framework for Minimax Optimization b/data/2020/neurips/A Catalyst Framework for Minimax Optimization new file mode 100644 index 0000000000..2ee73861db --- /dev/null +++ b/data/2020/neurips/A Catalyst Framework for Minimax Optimization @@ -0,0 +1 @@ +We introduce a generic two-loop scheme for smooth minimax optimization with strongly-convex-concave objectives. Our approach applies the accelerated proximal point framework (or Catalyst) to the associated dual problem and takes full advantage of existing gradient-based algorithms to solve a sequence of well-balanced strongly-convex-strongly-concave minimax problems. Despite its simplicity, this leads to a family of near-optimal algorithms with improved complexity over all existing methods designed for strongly-convex-concave minimax problems. Additionally, we obtain the first variance-reduced algorithms for this class of minimax problems with finite-sum structure and establish faster convergence rate than batch algorithms. Furthermore, when extended to the nonconvex-concave minimax optimization, our algorithm again achieves the state-of-the-art complexity for finding a stationary point. We carry out several numerical experiments showcasing the superiority of the Catalyst framework in practice. \ No newline at end of file diff --git a/data/2020/neurips/A Causal View on Robustness of Neural Networks b/data/2020/neurips/A Causal View on Robustness of Neural Networks new file mode 100644 index 0000000000..671d79273c --- /dev/null +++ b/data/2020/neurips/A Causal View on Robustness of Neural Networks @@ -0,0 +1 @@ +We present a causal view on the robustness of neural networks against input manipulations, which applies not only to traditional classification tasks but also to general measurement data. Based on this view, we design a deep causal manipulation augmented model (deep CAMA) which explicitly models possible manipulations on certain causes leading to changes in the observed effect. We further develop data augmentation and test-time fine-tuning methods to improve deep CAMA's robustness. When compared with discriminative deep neural networks, our proposed model shows superior robustness against unseen manipulations. As a by-product, our model achieves disentangled representation which separates the representation of manipulations from those of other latent causes. \ No newline at end of file diff --git a/data/2020/neurips/A Class of Algorithms for General Instrumental Variable Models b/data/2020/neurips/A Class of Algorithms for General Instrumental Variable Models new file mode 100644 index 0000000000..fdeb1c4838 --- /dev/null +++ b/data/2020/neurips/A Class of Algorithms for General Instrumental Variable Models @@ -0,0 +1 @@ +Causal treatment effect estimation is a key problem that arises in a variety of real-world settings, from personalized medicine to governmental policy making. There has been a flurry of recent work in machine learning on estimating causal effects when one has access to an instrument. However, to achieve identifiability, they in general require one-size-fits-all assumptions such as an additive error model for the outcome. An alternative is partial identification, which provides bounds on the causal effect. Little exists in terms of bounding methods that can deal with the most general case, where the treatment itself can be continuous. Moreover, bounding methods generally do not allow for a continuum of assumptions on the shape of the causal effect that can smoothly trade off stronger background knowledge for more informative bounds. In this work, we provide a method for causal effect bounding in continuous distributions, leveraging recent advances in gradient-based methods for the optimization of computationally intractable objective functions. We demonstrate on a set of synthetic and real-world data that our bounds capture the causal effect when additive methods fail, providing a useful range of answers compatible with observation as opposed to relying on unwarranted structural assumptions. \ No newline at end of file diff --git a/data/2020/neurips/A Closer Look at Accuracy vs. Robustness b/data/2020/neurips/A Closer Look at Accuracy vs. Robustness new file mode 100644 index 0000000000..bafd5aa2f6 --- /dev/null +++ b/data/2020/neurips/A Closer Look at Accuracy vs. Robustness @@ -0,0 +1 @@ +Current methods for training robust networks lead to a drop in test accuracy, which has led prior works to posit that a robustness-accuracy tradeoff may be inevitable in deep learning. We take a closer look at this phenomenon and first show that real image datasets are actually separated. With this property in mind, we then prove that robustness and accuracy should both be achievable for benchmark datasets through locally Lipschitz functions, and hence, there should be no inherent tradeoff between robustness and accuracy. Through extensive experiments with robustness methods, we argue that the gap between theory and practice arises from two limitations of current methods: either they fail to impose local Lipschitzness or they are insufficiently generalized. We explore combining dropout with robust training methods and obtain better generalization. We conclude that achieving robustness and accuracy in practice may require using methods that impose local Lipschitzness and augmenting them with deep learning generalization techniques. Code available at this https URL \ No newline at end of file diff --git a/data/2020/neurips/A Closer Look at the Training Strategy for Modern Meta-Learning b/data/2020/neurips/A Closer Look at the Training Strategy for Modern Meta-Learning new file mode 100644 index 0000000000..540f9899ff --- /dev/null +++ b/data/2020/neurips/A Closer Look at the Training Strategy for Modern Meta-Learning @@ -0,0 +1 @@ +The support/query (S/Q) episodic training strategy has been widely used in modern meta-learning algorithms and is believed to improve their generalization ability to test environments. This paper conducts a theoretical investigation of this training strategy on generalization. From a stability perspective, we analyze the generalization error bound of generic meta-learning algorithms trained with such strategy. We show that the S/Q episodic training strategy naturally leads to a counterintuitive generalization bound of O (1 / √ n ) , which only depends on the task number n but independent of the inner-task sample size m . Under the common assumption m << n for few-shot learning, the bound of O (1 / √ n ) implies strong generalization guarantees for modern meta-learning algorithms in the few-shot regime. To further explore the influence of training strategies on generalization, we propose a leave-one-out (LOO) training strategy for meta-learning and compare it with S/Q training. Experiments on standard few-shot regression and classification tasks with popular meta-learning algorithms validate our analysis. \ No newline at end of file diff --git a/data/2020/neurips/A Combinatorial Perspective on Transfer Learning b/data/2020/neurips/A Combinatorial Perspective on Transfer Learning new file mode 100644 index 0000000000..8b6c852569 --- /dev/null +++ b/data/2020/neurips/A Combinatorial Perspective on Transfer Learning @@ -0,0 +1 @@ +Human intelligence is characterized not only by the capacity to learn complex skills, but the ability to rapidly adapt and acquire new skills within an ever-changing environment. In this work we study how the learning of modular solutions can allow for effective generalization to both unseen and potentially differently distributed data. Our main postulate is that the combination of task segmentation, modular learning and memory-based ensembling can give rise to generalization on an exponentially growing number of unseen tasks. We provide a concrete instantiation of this idea using a combination of: (1) the Forget-Me-Not Process, for task segmentation and memory based ensembling; and (2) Gated Linear Networks, which in contrast to contemporary deep learning techniques use a modular and local learning mechanism. We demonstrate that this system exhibits a number of desirable continual learning properties: robustness to catastrophic forgetting, no negative transfer and increasing levels of positive transfer as more tasks are seen. We show competitive performance against both offline and online methods on standard continual learning benchmarks. \ No newline at end of file diff --git a/data/2020/neurips/A Computational Separation between Private Learning and Online Learning b/data/2020/neurips/A Computational Separation between Private Learning and Online Learning new file mode 100644 index 0000000000..338080e5d0 --- /dev/null +++ b/data/2020/neurips/A Computational Separation between Private Learning and Online Learning @@ -0,0 +1 @@ +A recent line of work has shown a qualitative equivalence between differentially private PAC learning and online learning: A concept class is privately learnable if and only if it is online learnable with a finite mistake bound. However, both directions of this equivalence incur significant losses in both sample and computational efficiency. Studying a special case of this connection, Gonen, Hazan, and Moran (NeurIPS 2019) showed that uniform or highly sample-efficient pure-private learners can be time-efficiently compiled into online learners. We show that, assuming the existence of one-way functions, such an efficient conversion is impossible even for general pure-private learners with polynomial sample complexity. This resolves a question of Neel, Roth, and Wu (FOCS 2019). \ No newline at end of file diff --git a/data/2020/neurips/A Continuous-Time Mirror Descent Approach to Sparse Phase Retrieval b/data/2020/neurips/A Continuous-Time Mirror Descent Approach to Sparse Phase Retrieval new file mode 100644 index 0000000000..d546ca9c5e --- /dev/null +++ b/data/2020/neurips/A Continuous-Time Mirror Descent Approach to Sparse Phase Retrieval @@ -0,0 +1 @@ +We analyze continuous-time mirror descent applied to sparse phase retrieval, which is the problem of recovering sparse signals from a set of magnitude-only measurements. We apply mirror descent to the unconstrained empirical risk minimization problem (batch setting), using the square loss and square measurements. We provide a convergence analysis of the algorithm in this non-convex setting and prove that, with the hypentropy mirror map, mirror descent recovers any $k$-sparse vector $\mathbf{x}^\star\in\mathbb{R}^n$ with minimum (in modulus) non-zero entry on the order of $\| \mathbf{x}^\star \|_2/\sqrt{k}$ from $k^2$ Gaussian measurements, modulo logarithmic terms. This yields a simple algorithm which, unlike most existing approaches to sparse phase retrieval, adapts to the sparsity level, without including thresholding steps or adding regularization terms. Our results also provide a principled theoretical understanding for Hadamard Wirtinger flow [58], as Euclidean gradient descent applied to the empirical risk problem with Hadamard parametrization can be recovered as a first-order approximation to mirror descent in discrete time. \ No newline at end of file diff --git a/data/2020/neurips/A Contour Stochastic Gradient Langevin Dynamics Algorithm for Simulations of Multi-modal Distributions b/data/2020/neurips/A Contour Stochastic Gradient Langevin Dynamics Algorithm for Simulations of Multi-modal Distributions new file mode 100644 index 0000000000..2de55c3228 --- /dev/null +++ b/data/2020/neurips/A Contour Stochastic Gradient Langevin Dynamics Algorithm for Simulations of Multi-modal Distributions @@ -0,0 +1 @@ +We propose an adaptively weighted stochastic gradient Langevin dynamics algorithm (SGLD), so-called contour stochastic gradient Langevin dynamics (CSGLD), for Bayesian learning in big data statistics. The proposed algorithm is essentially a scalable dynamic importance sampler, which automatically flattens the target distribution such that the simulation for a multi-modal distribution can be greatly facilitated. Theoretically, we prove a stability condition and establish the asymptotic convergence of the self-adapting parameter to a unique fixed-point, regardless of the non-convexity of the original energy function; we also present an error analysis for the weighted averaging estimators. Empirically, the CSGLD algorithm is tested on multiple benchmark datasets including CIFAR10 and CIFAR100. The numerical results indicate its superiority over the existing state-of-the-art algorithms in training deep neural networks. \ No newline at end of file diff --git a/data/2020/neurips/A Convolutional Auto-Encoder for Haplotype Assembly and Viral Quasispecies Reconstruction b/data/2020/neurips/A Convolutional Auto-Encoder for Haplotype Assembly and Viral Quasispecies Reconstruction new file mode 100644 index 0000000000..7a771dec01 --- /dev/null +++ b/data/2020/neurips/A Convolutional Auto-Encoder for Haplotype Assembly and Viral Quasispecies Reconstruction @@ -0,0 +1 @@ +Haplotype assembly and viral quasispecies reconstruction are challenging tasks concerned with analysis of genomic mixtures using sequencing data. High-throughput sequencing technologies generate enormous amounts of short fragments (reads) which essentially oversample components of a mixture; the representation redundancy enables reconstruction of the components (haplotypes, viral strains). The reconstruction problem, known to be NP-hard, boils down to grouping together reads originating from the same component in a mixture. Existing methods struggle to solve this problem with required level of accuracy and low runtimes; the problem is becoming increasingly more challenging as the number and length of the components increase. This paper proposes a read clustering method based on a convolutional auto-encoder designed to first project sequenced fragments to a low-dimensional space and then estimate the probability of the read origin using learned embedded features. The components are reconstructed by finding consensus sequences that agglomerate reads from the same origin. Mini-batch stochastic gradient descent and dimension reduction of reads allow the proposed method to efficiently deal with massive numbers of long reads. Experiments on simulated, semi-experimental and experimental data demonstrate the ability of the proposed method to accurately reconstruct haplotypes and viral quasispecies, often demonstrating superior performance compared to state-of-the-art methods. \ No newline at end of file diff --git a/data/2020/neurips/A Decentralized Parallel Algorithm for Training Generative Adversarial Nets b/data/2020/neurips/A Decentralized Parallel Algorithm for Training Generative Adversarial Nets new file mode 100644 index 0000000000..b8500a3e9c --- /dev/null +++ b/data/2020/neurips/A Decentralized Parallel Algorithm for Training Generative Adversarial Nets @@ -0,0 +1 @@ +Generative Adversarial Networks (GANs) are powerful class of generative models in the deep learning community. Current practice on large-scale GAN training~\citep{brock2018large} utilizes large models and distributed large-batch training strategies, and is implemented on deep learning frameworks (e.g., TensorFlow, PyTorch, etc.) designed in a centralized manner. In the centralized network topology, every worker needs to communicate with the central node. However, when the network bandwidth is low or network latency is high, the performance would be significantly degraded. Despite recent progress on decentralized algorithms for training deep neural networks, it remains unclear whether it is possible to train GANs in a decentralized manner. The main difficulty lies at handling the nonconvex-nonconcave min-max optimization and the decentralized communication simultaneously. In this paper, we address this difficulty by designing the \textbf{first gradient-based decentralized parallel algorithm} which allows workers to have multiple rounds of communications in one iteration and to update the discriminator and generator simultaneously, and this design makes it amenable for the convergence analysis of the proposed decentralized algorithm. Theoretically, our proposed decentralized algorithm is able to solve a class of non-convex non-concave min-max problems with provable non-asymptotic convergence to first-order stationary point. Experimental results on GANs demonstrate the effectiveness of the proposed algorithm. \ No newline at end of file diff --git a/data/2020/neurips/A Dictionary Approach to Domain-Invariant Learning in Deep Networks b/data/2020/neurips/A Dictionary Approach to Domain-Invariant Learning in Deep Networks new file mode 100644 index 0000000000..e81a5f7b68 --- /dev/null +++ b/data/2020/neurips/A Dictionary Approach to Domain-Invariant Learning in Deep Networks @@ -0,0 +1 @@ +In this paper, we consider domain-invariant deep learning by explicitly modeling domain shifts with only a small amount of domain-specific parameters in a Convolutional Neural Network (CNN). By exploiting the observation that a convolutional filter can be well approximated as a linear combination of a small set of dictionary atoms, we show for the first time, both empirically and theoretically, that domain shifts can be effectively handled by decomposing a convolutional layer into a domain-specific atom layer and a domain-shared coefficient layer, while both remain convolutional. An input channel will now first convolve spatially only with each respective domain-specific dictionary atom to "absorb" domain variations, and then output channels are linearly combined using common decomposition coefficients trained to promote shared semantics across domains. We use toy examples, rigorous analysis, and real-world examples with diverse datasets and architectures, to show the proposed plug-in framework's effectiveness in cross and joint domain performance and domain adaptation. With the proposed architecture, we need only a small set of dictionary atoms to model each additional domain, which brings a negligible amount of additional parameters, typically a few hundred. \ No newline at end of file diff --git a/data/2020/neurips/A Discrete Variational Recurrent Topic Model without the Reparametrization Trick b/data/2020/neurips/A Discrete Variational Recurrent Topic Model without the Reparametrization Trick new file mode 100644 index 0000000000..0e720512eb --- /dev/null +++ b/data/2020/neurips/A Discrete Variational Recurrent Topic Model without the Reparametrization Trick @@ -0,0 +1 @@ +We show how to learn a neural topic model with discrete random variables---one that explicitly models each word's assigned topic---using neural variational inference that does not rely on stochastic backpropagation to handle the discrete variables. The model we utilize combines the expressive power of neural methods for representing sequences of text with the topic model's ability to capture global, thematic coherence. Using neural variational inference, we show improved perplexity and document understanding across multiple corpora. We examine the effect of prior parameters both on the model and variational parameters and demonstrate how our approach can compete and surpass a popular topic model implementation on an automatic measure of topic quality. \ No newline at end of file diff --git a/data/2020/neurips/A Dynamical Central Limit Theorem for Shallow Neural Networks b/data/2020/neurips/A Dynamical Central Limit Theorem for Shallow Neural Networks new file mode 100644 index 0000000000..81566804a9 --- /dev/null +++ b/data/2020/neurips/A Dynamical Central Limit Theorem for Shallow Neural Networks @@ -0,0 +1 @@ +Recent theoretical work has characterized the dynamics of wide shallow neural networks trained via gradient descent in an asymptotic regime called the mean-field limit as the number of parameters tends towards infinity. At initialization, the randomly sampled parameters lead to a deviation from the mean-field limit that is dictated by the classical Central Limit Theorem (CLT). However, the dynamics of training introduces correlations among the parameters, raising the question of how the fluctuations evolve during training. Here, we analyze the mean-field dynamics as a Wasserstein gradient flow and prove that the deviations from the mean-field limit scaled by the width, in the width-asymptotic limit, remain bounded throughout training. In particular, they eventually vanish in the CLT scaling if the mean-field dynamics converges to a measure that interpolates the training data. This observation has implications for both the approximation rate and the generalization: the upper bound we obtain is given by a Monte-Carlo type resampling error, which does not depend explicitly on the dimension. This bound motivates a regularizaton term on the 2-norm of the underlying measure, which is also connected to generalization via the variation-norm function spaces. \ No newline at end of file diff --git a/data/2020/neurips/A Fair Classifier Using Kernel Density Estimation b/data/2020/neurips/A Fair Classifier Using Kernel Density Estimation new file mode 100644 index 0000000000..dfa4701786 --- /dev/null +++ b/data/2020/neurips/A Fair Classifier Using Kernel Density Estimation @@ -0,0 +1 @@ +As machine learning becomes prevalent in a widening array of sensitive applications such as job hiring and criminal justice, one critical aspect in the design of machine learning classifiers is to ensure fairness: Guaranteeing the irrelevancy of a prediction to sensitive attributes such as gender and race. This work develops a kernel density estimation (KDE) methodology to faithfully respect the fairness constraint while yielding a tractable optimization problem that comes with high accuracy-fairness tradeoff. One key feature of this approach is that the fairness measure quantified based on KDE can be expressed as a differentiable function w.r.t. model parameters, thereby enabling the use of prominent gradient descent to readily solve an interested optimization problem. This work focuses on classifi-cation tasks and two well-known measures of group fairness: demographic parity and equalized odds. We empirically show that our algorithm achieves greater or comparable performances against prior fair classifers in accuracy-fairness tradeoff as well as in training stability on both synthetic and benchmark real datasets. \ No newline at end of file diff --git a/data/2020/neurips/A Feasible Level Proximal Point Method for Nonconvex Sparse Constrained Optimization b/data/2020/neurips/A Feasible Level Proximal Point Method for Nonconvex Sparse Constrained Optimization new file mode 100644 index 0000000000..bd3095ab1f --- /dev/null +++ b/data/2020/neurips/A Feasible Level Proximal Point Method for Nonconvex Sparse Constrained Optimization @@ -0,0 +1 @@ +Nonconvex sparse models have received significant attention in high-dimensional machine learning. In this paper, we study a new model consisting of a general convex or nonconvex objectives and a variety of continuous nonconvex sparsity-inducing constraints. For this constrained model, we propose a novel proximal point algorithm that solves a sequence of convex subproblems with gradually relaxed constraint levels. Each subproblem, having a proximal point objective and a convex surrogate constraint, can be efficiently solved based on a fast routine for projection onto the surrogate constraint. We establish the asymptotic convergence of the proposed algorithm to the Karush-Kuhn-Tucker (KKT) solutions. We also establish new convergence complexities to achieve an approximate KKT solution when the objective can be smooth/nonsmooth, deterministic/stochastic and convex/nonconvex with complexity that is on a par with gradient descent for unconstrained optimization problems in respective cases. To the best of our knowledge, this is the first study of the first-order methods with complexity guarantee for nonconvex sparse-constrained problems. We perform numerical experiments to demonstrate the effectiveness of our new model and efficiency of the proposed algorithm for large scale problems. \ No newline at end of file diff --git a/data/2020/neurips/A Finite-Time Analysis of Two Time-Scale Actor-Critic Methods b/data/2020/neurips/A Finite-Time Analysis of Two Time-Scale Actor-Critic Methods new file mode 100644 index 0000000000..2b8bac416f --- /dev/null +++ b/data/2020/neurips/A Finite-Time Analysis of Two Time-Scale Actor-Critic Methods @@ -0,0 +1 @@ +Actor-critic (AC) methods have exhibited great empirical success compared with other reinforcement learning algorithms, where the actor uses the policy gradient to improve the learning policy and the critic uses temporal difference learning to estimate the policy gradient. Under the two time-scale learning rate schedule, the asymptotic convergence of AC has been well studied in the literature. However, the non-asymptotic convergence and finite sample complexity of actor-critic methods are largely open. In this work, we provide a non-asymptotic analysis for two time-scale actor-critic methods under non-i.i.d. setting. We prove that the actor-critic method is guaranteed to find a first-order stationary point (i.e., $\|\nabla J(\boldsymbol{\theta})\|_2^2 \le \epsilon$) of the non-concave performance function $J(\boldsymbol{\theta})$, with $\mathcal{\tilde{O}}(\epsilon^{-2.5})$ sample complexity. To the best of our knowledge, this is the first work providing finite-time analysis and sample complexity bound for two time-scale actor-critic methods. \ No newline at end of file diff --git a/data/2020/neurips/A Flexible Framework for Designing Trainable Priors with Adaptive Smoothing and Game Encoding b/data/2020/neurips/A Flexible Framework for Designing Trainable Priors with Adaptive Smoothing and Game Encoding new file mode 100644 index 0000000000..7fb8806eac --- /dev/null +++ b/data/2020/neurips/A Flexible Framework for Designing Trainable Priors with Adaptive Smoothing and Game Encoding @@ -0,0 +1 @@ +We introduce a general framework for designing and training neural network layers whose forward passes can be interpreted as solving non-smooth convex optimization problems, and whose architectures are derived from an optimization algorithm. We focus on convex games, solved by local agents represented by the nodes of a graph and interacting through regularization functions. This approach is appealing for solving imaging problems, as it allows the use of classical image priors within deep models that are trainable end to end. The priors used in this presentation include variants of total variation, Laplacian regularization, bilateral filtering, sparse coding on learned dictionaries, and non-local self similarities. Our models are fully interpretable as well as parameter and data efficient. Our experiments demonstrate their effectiveness on a large diversity of tasks ranging from image denoising and compressed sensing for fMRI to dense stereo matching. \ No newline at end of file diff --git a/data/2020/neurips/A Game Theoretic Analysis of Additive Adversarial Attacks and Defenses b/data/2020/neurips/A Game Theoretic Analysis of Additive Adversarial Attacks and Defenses new file mode 100644 index 0000000000..ea36436bf9 --- /dev/null +++ b/data/2020/neurips/A Game Theoretic Analysis of Additive Adversarial Attacks and Defenses @@ -0,0 +1 @@ +Research in adversarial learning follows a cat and mouse game between attackers and defenders where attacks are proposed, they are mitigated by new defenses, and subsequently new attacks are proposed that break earlier defenses, and so on. However, it has remained unclear as to whether there are conditions under which no better attacks or defenses can be proposed. In this paper, we propose a game-theoretic framework for studying attacks and defenses which exist in equilibrium. Under a locally linear decision boundary model for the underlying binary classifier, we prove that the Fast Gradient Method attack and the Randomized Smoothing defense form a Nash Equilibrium. We then show how this equilibrium defense can be approximated given finitely many samples from a data-generating distribution, and derive a generalization bound for the performance of our approximation. \ No newline at end of file diff --git a/data/2020/neurips/A Game-Theoretic Analysis of the Empirical Revenue Maximization Algorithm with Endogenous Sampling b/data/2020/neurips/A Game-Theoretic Analysis of the Empirical Revenue Maximization Algorithm with Endogenous Sampling new file mode 100644 index 0000000000..4e7905979e --- /dev/null +++ b/data/2020/neurips/A Game-Theoretic Analysis of the Empirical Revenue Maximization Algorithm with Endogenous Sampling @@ -0,0 +1 @@ +The Empirical Revenue Maximization (ERM) is one of the most important price learning algorithms in auction design: as the literature shows it can learn approximately optimal reserve prices for revenue-maximizing auctioneers in both repeated auctions and uniform-price auctions. However, in these applications the agents who provide inputs to ERM have incentives to manipulate the inputs to lower the outputted price. We generalize the definition of an incentive-awareness measure proposed by Lavi et al (2019), to quantify the reduction of ERM's outputted price due to a change of $m\ge 1$ out of $N$ input samples, and provide specific convergence rates of this measure to zero as $N$ goes to infinity for different types of input distributions. By adopting this measure, we construct an efficient, approximately incentive-compatible, and revenue-optimal learning algorithm using ERM in repeated auctions against non-myopic bidders, and show approximate group incentive-compatibility in uniform-price auctions. \ No newline at end of file diff --git a/data/2020/neurips/A General Large Neighborhood Search Framework for Solving Integer Linear Programs b/data/2020/neurips/A General Large Neighborhood Search Framework for Solving Integer Linear Programs new file mode 100644 index 0000000000..d871c75fe2 --- /dev/null +++ b/data/2020/neurips/A General Large Neighborhood Search Framework for Solving Integer Linear Programs @@ -0,0 +1 @@ +Mixed integer programming provides a unifying framework for solving a medley of hard combinatorial optimization problems of practical interest. A mixed integer program (MIP) is typically solved using linear programming (LP) based branch-and-bound algorithm. Primal heuristics are a key component of MIP solvers and enable finding good feasible solutions early in the tree search. The literature is rich with a variety of hybrid primal heuristics that combine both heuristic and exact methods. In this thesis, we propose a new supervised large neighborhood search heuristic for the general MIP, as well as, a new analytical MIP-based primal heuristic for the linear ordering problem. We present our work in two parts. Part I: Supervised Neighborhood Selection for Mixed Integer Programs Large neighborhood search (LNS) heuristics are employed as improvement procedures within the branch-and-bound algorithm. They formulate the neighborhoods as an auxiliary MIP with a restricted search space, which is then solved to search for an improving solution. The neighborhoods are typically defined by handcrafted rules, guided by human intuition and offline experimentation. Alternatively, a neighborhood should be defined such that it has a high likelihood of success. We apply a data-driven approach to predict an ideal neighborhood for the neighborhood search. We propose a supervised large neighborhood search heuristic for the general mixed integer programs and compare it with Relaxation Induced Neighborhood Search (RINS), a popular LNS heuristic. Our heuristic not only finds an improving solution more often but also improves the solver performance on key metrics. Part II: MIP-based Primal Heuristic for the Linear Ordering Problem Linear ordering problem (LOP) is a quintessential combinatorial optimization problem and has been well studied in the literature. We present a new MIP-based primal heuristic for the LOP. The heuristic decomposes the LOP instance into sub-problems albeit sub-optimal ones, solves each sub-problem optimally and concatenates the partial solutions to construct a solution to the original problem instance. We present empirical results that show that our heuristics achieves good performance on benchmark instances. \ No newline at end of file diff --git a/data/2020/neurips/A General Method for Robust Learning from Batches b/data/2020/neurips/A General Method for Robust Learning from Batches new file mode 100644 index 0000000000..696b25599f --- /dev/null +++ b/data/2020/neurips/A General Method for Robust Learning from Batches @@ -0,0 +1 @@ +In many applications, data is collected in batches, some of which are corrupt or even adversarial. Recent work derived optimal robust algorithms for estimating discrete distributions in this setting. We consider a general framework of robust learning from batches, and determine the limits of both classification and distribution estimation over arbitrary, including continuous, domains. Building on these results, we derive the first robust agnostic computationally-efficient learning algorithms for piecewise-interval classification, and for piecewise-polynomial, monotone, log-concave, and gaussian-mixture distribution estimation. \ No newline at end of file diff --git a/data/2020/neurips/A Generalized Neural Tangent Kernel Analysis for Two-layer Neural Networks b/data/2020/neurips/A Generalized Neural Tangent Kernel Analysis for Two-layer Neural Networks new file mode 100644 index 0000000000..f931c61699 --- /dev/null +++ b/data/2020/neurips/A Generalized Neural Tangent Kernel Analysis for Two-layer Neural Networks @@ -0,0 +1 @@ +A recent breakthrough in deep learning theory shows that the training of over-parameterized deep neural networks can be characterized by a kernel function called \textit{neural tangent kernel} (NTK). However, it is known that this type of results does not perfectly match the practice, as NTK-based analysis requires the network weights to stay very close to their initialization throughout training, and cannot handle regularizers or gradient noises. In this paper, we provide a generalized neural tangent kernel analysis and show that noisy gradient descent with weight decay can still exhibit a "kernel-like" behavior. This implies that the training loss converges linearly up to a certain accuracy. We also establish a novel generalization error bound for two-layer neural networks trained by noisy gradient descent with weight decay. \ No newline at end of file diff --git a/data/2020/neurips/A Group-Theoretic Framework for Data Augmentation b/data/2020/neurips/A Group-Theoretic Framework for Data Augmentation new file mode 100644 index 0000000000..1c7beca9de --- /dev/null +++ b/data/2020/neurips/A Group-Theoretic Framework for Data Augmentation @@ -0,0 +1,2 @@ +Data augmentation is a widely used trick when training deep neural networks: in addition to the original data, properly transformed data are also added to the training set. However, to the best of our knowledge, a clear mathematical framework to explain the performance benefits of data augmentation is not available. +In this paper, we develop such a theoretical framework. We show data augmentation is equivalent to an averaging operation over the orbits of a certain group that keeps the data distribution approximately invariant. We prove that it leads to variance reduction. We study empirical risk minimization, and the examples of exponential families, linear regression, and certain two-layer neural networks. We also discuss how data augmentation could be used in problems with symmetry where other approaches are prevalent, such as in cryo-electron microscopy (cryo-EM). \ No newline at end of file diff --git a/data/2020/neurips/A Limitation of the PAC-Bayes Framework b/data/2020/neurips/A Limitation of the PAC-Bayes Framework new file mode 100644 index 0000000000..7c1c1bbe19 --- /dev/null +++ b/data/2020/neurips/A Limitation of the PAC-Bayes Framework @@ -0,0 +1,2 @@ +PAC-Bayes is a useful framework for deriving generalization bounds which was introduced by McAllester ('98). This framework has the flexibility of deriving distribution- and algorithm-dependent bounds, which are often tighter than VC-related uniform convergence bounds. In this manuscript we present a limitation for the PAC-Bayes framework. We demonstrate an easy learning task that is not amenable to a PAC-Bayes analysis. +Specifically, we consider the task of linear classification in 1D; it is well-known that this task is learnable using just $O(\log(1/\delta)/\epsilon)$ examples. On the other hand, we show that this fact can not be proved using a PAC-Bayes analysis: for any algorithm that learns 1-dimensional linear classifiers there exists a (realizable) distribution for which the PAC-Bayes bound is arbitrarily large. \ No newline at end of file diff --git a/data/2020/neurips/A Local Temporal Difference Code for Distributional Reinforcement Learning b/data/2020/neurips/A Local Temporal Difference Code for Distributional Reinforcement Learning new file mode 100644 index 0000000000..d8e5712a5a --- /dev/null +++ b/data/2020/neurips/A Local Temporal Difference Code for Distributional Reinforcement Learning @@ -0,0 +1 @@ +Recent theoretical and experimental results suggest that the dopamine system implements distributional temporal difference backups, allowing learning of the entire distributions of the long-run values of states rather than just their expected values. However, the distributional codes explored so far rely on a complex imputation step which crucially relies on spatial non-locality: in order to compute reward prediction errors, units must know not only their own state but also the states of the other units. It is far from clear how these steps could be implemented in realistic neural circuits. Here, we introduce the Laplace code: a local temporal difference code for distributional reinforcement learning that is representationally powerful and computationally straightforward. The code decomposes value distributions and prediction errors across three separated dimensions: reward magnitude (related to distributional quantiles), temporal discounting (related to the Laplace transform of future rewards) and time horizon (related to eligibility traces). Besides lending itself to a local learning rule, the decomposition recovers the temporal evolution of the immediate reward distribution, indicating all possible rewards at all future times. This increases representational capacity and allows for temporally-flexible computations that immediately adjust to changing horizons or discount factors. \ No newline at end of file diff --git a/data/2020/neurips/A Loss Function for Generative Neural Networks Based on Watson's Perceptual Model b/data/2020/neurips/A Loss Function for Generative Neural Networks Based on Watson's Perceptual Model new file mode 100644 index 0000000000..22e7918233 --- /dev/null +++ b/data/2020/neurips/A Loss Function for Generative Neural Networks Based on Watson's Perceptual Model @@ -0,0 +1 @@ +To train Variational Autoencoders (VAEs) to generate realistic imagery requires a loss function that reflects human perception of image similarity. We propose such a loss function based on Watson's perceptual model, which computes a weighted distance in frequency space and accounts for luminance and contrast masking. We extend the model to color images, increase its robustness to translation by using the Fourier Transform, remove artifacts due to splitting the image into blocks, and make it differentiable. In experiments, VAEs trained with the new loss function generated realistic, high-quality image samples. Compared to using the Euclidean distance and the Structural Similarity Index, the images were less blurry; compared to deep neural network based losses, the new approach required less computational resources and generated images with less artifacts. \ No newline at end of file diff --git a/data/2020/neurips/A Matrix Chernoff Bound for Markov Chains and Its Application to Co-occurrence Matrices b/data/2020/neurips/A Matrix Chernoff Bound for Markov Chains and Its Application to Co-occurrence Matrices new file mode 100644 index 0000000000..5a4530ff53 --- /dev/null +++ b/data/2020/neurips/A Matrix Chernoff Bound for Markov Chains and Its Application to Co-occurrence Matrices @@ -0,0 +1 @@ +We prove a Chernoff-type bound for sums of matrix-valued random variables sampled via a regular (aperiodic and irreducible) finite Markov chain. Specially, consider a random walk on a regular Markov chain and a Hermitian matrix-valued function on its state space. Our result gives exponentially decreasing bounds on the tail distributions of the extreme eigenvalues of the sample mean matrix. Our proof is based on the matrix expander (regular undirected graph) Chernoff bound [Garg et al. STOC ’18] and scalar Chernoff-Hoeffding bounds for Markov chains [Chung et al. STACS ’12]. Our matrix Chernoff bound for Markov chains can be applied to analyze the behavior of co-occurrence statistics for sequential data, which have been common and important data signals in machine learning. We show that given a regular Markov chain with \(n\) states and mixing time \(\tau\), we need a trajectory of length \(O(\tau (\log{n} + \log{\tau})/\epsilon^2)\) to achieve an estimator of the co-occurrence matrix with error bound \(\epsilon\). We conduct several experiments and the experimental results are consistent with the exponentially fast convergence rate from theoretical analysis. Our result gives the first bound on the convergence rate of the co-occurrence matrix and the first sample complexity analysis in graph representation learning. \ No newline at end of file diff --git a/data/2020/neurips/A Maximum-Entropy Approach to Off-Policy Evaluation in Average-Reward MDPs b/data/2020/neurips/A Maximum-Entropy Approach to Off-Policy Evaluation in Average-Reward MDPs new file mode 100644 index 0000000000..8a85010e4a --- /dev/null +++ b/data/2020/neurips/A Maximum-Entropy Approach to Off-Policy Evaluation in Average-Reward MDPs @@ -0,0 +1 @@ +This work focuses on off-policy evaluation (OPE) with function approximation in infinite-horizon undiscounted Markov decision processes (MDPs). For MDPs that are ergodic and linear (i.e. where rewards and dynamics are linear in some known features), we provide the first finite-sample OPE error bound, extending existing results beyond the episodic and discounted cases. In a more general setting, when the feature dynamics are approximately linear and for arbitrary rewards, we propose a new approach for estimating stationary distributions with function approximation. We formulate this problem as finding the maximum-entropy distribution subject to matching feature expectations under empirical dynamics. We show that this results in an exponential-family distribution whose sufficient statistics are the features, paralleling maximum-entropy approaches in supervised learning. We demonstrate the effectiveness of the proposed OPE approaches in multiple environments. \ No newline at end of file diff --git a/data/2020/neurips/A Measure-Theoretic Approach to Kernel Conditional Mean Embeddings b/data/2020/neurips/A Measure-Theoretic Approach to Kernel Conditional Mean Embeddings new file mode 100644 index 0000000000..c69f33b1ed --- /dev/null +++ b/data/2020/neurips/A Measure-Theoretic Approach to Kernel Conditional Mean Embeddings @@ -0,0 +1 @@ +We present a new operator-free, measure-theoretic definition of the conditional mean embedding as a random variable taking values in a reproducing kernel Hilbert space. While the kernel mean embedding of marginal distributions has been defined rigorously, the existing operator-based approach of the conditional version lacks a rigorous definition, and depends on strong assumptions that hinder its analysis. Our definition does not impose any of the assumptions that the operator-based counterpart requires. We derive a natural regression interpretation to obtain empirical estimates, and provide a thorough analysis of its properties, including universal consistency. As natural by-products, we obtain the conditional analogues of the Maximum Mean Discrepancy and Hilbert-Schmidt Independence Criterion, and demonstrate their behaviour via simulations. \ No newline at end of file diff --git a/data/2020/neurips/A Non-Asymptotic Analysis for Stein Variational Gradient Descent b/data/2020/neurips/A Non-Asymptotic Analysis for Stein Variational Gradient Descent new file mode 100644 index 0000000000..1c3116c859 --- /dev/null +++ b/data/2020/neurips/A Non-Asymptotic Analysis for Stein Variational Gradient Descent @@ -0,0 +1 @@ +We study the Stein Variational Gradient Descent (SVGD) algorithm, which optimises a set of particles to approximate a target probability distribution $\pi\propto e^{-V}$ on $\mathbb{R}^d$. In the population limit, SVGD performs gradient descent in the space of probability distributions on the KL divergence with respect to $\pi$, where the gradient is smoothed through a kernel integral operator. In this paper, we provide a novel finite time analysis for the SVGD algorithm. We obtain a descent lemma establishing that the algorithm decreases the objective at each iteration, and provably converges, with less restrictive assumptions on the step size than required in earlier analyses. We further provide a guarantee on the convergence rate in Kullback-Leibler divergence, assuming $\pi$ satisfies a Stein log-Sobolev inequality as in Duncan et al. (2019), which takes into account the geometry induced by the smoothed KL gradient. \ No newline at end of file diff --git a/data/2020/neurips/A Novel Approach for Constrained Optimization in Graphical Models b/data/2020/neurips/A Novel Approach for Constrained Optimization in Graphical Models new file mode 100644 index 0000000000..2ba092297f --- /dev/null +++ b/data/2020/neurips/A Novel Approach for Constrained Optimization in Graphical Models @@ -0,0 +1 @@ +We consider the following constrained maximization problem in discrete probabilistic graphical models (PGMs). Given two (possibly identical) PGMs M 1 and M 2 defined over the same set of variables and a real number q , find an assignment of values to all variables such that the probability of the assignment is maximized w.r.t. M 1 and is smaller than q w.r.t. M 2 . We show that several explanation and robust estimation queries over graphical models are special cases of this problem. We propose a class of approximate algorithms for solving this problem. Our algorithms are based on a graph concept called k -separator and heuristic algorithms for multiple choice knapsack and subset-sum problems. Our experiments show that our algorithms are superior to the following approach: encode the problem as a mixed integer linear program (MILP) and solve the latter using a state-of-the-art MILP solver such as SCIP. \ No newline at end of file diff --git a/data/2020/neurips/A Novel Automated Curriculum Strategy to Solve Hard Sokoban Planning Instances b/data/2020/neurips/A Novel Automated Curriculum Strategy to Solve Hard Sokoban Planning Instances new file mode 100644 index 0000000000..34a713e2f1 --- /dev/null +++ b/data/2020/neurips/A Novel Automated Curriculum Strategy to Solve Hard Sokoban Planning Instances @@ -0,0 +1 @@ +In recent years, we have witnessed tremendous progress in deep reinforcement learning (RL) for tasks such as Go, Chess, video games, and robot control. Nevertheless, other combinatorial domains, such as AI planning, still pose considerable challenges for RL approaches. The key difficulty in those domains is that a positive reward signal becomes {\em exponentially rare} as the minimal solution length increases. So, an RL approach loses its training signal. There has been promising recent progress by using a curriculum-driven learning approach that is designed to solve a single hard instance. We present a novel {\em automated} curriculum approach that dynamically selects from a pool of unlabeled training instances of varying task complexity guided by our {\em difficulty quantum momentum} strategy. We show how the smoothness of the task hardness impacts the final learning results. In particular, as the size of the instance pool increases, the ``hardness gap'' decreases, which facilitates a smoother automated curriculum based learning process. Our automated curriculum approach dramatically improves upon the previous approaches. We show our results on Sokoban, which is a traditional PSPACE-complete planning problem and presents a great challenge even for specialized solvers. Our RL agent can solve hard instances that are far out of reach for any previous state-of-the-art Sokoban solver. In particular, our approach can uncover plans that require hundreds of steps, while the best previous search methods would take many years of computing time to solve such instances. In addition, we show that we can further boost the RL performance with an intricate coupling of our automated curriculum approach with a curiosity-driven search strategy and a graph neural net representation. \ No newline at end of file diff --git a/data/2020/neurips/A Randomized Algorithm to Reduce the Support of Discrete Measures b/data/2020/neurips/A Randomized Algorithm to Reduce the Support of Discrete Measures new file mode 100644 index 0000000000..637ed0517d --- /dev/null +++ b/data/2020/neurips/A Randomized Algorithm to Reduce the Support of Discrete Measures @@ -0,0 +1 @@ +Given a discrete probability measure supported on $N$ atoms and a set of $n$ real-valued functions, there exists a probability measure that is supported on a subset of $n+1$ of the original $N$ atoms and has the same mean when integrated against each of the $n$ functions. If $ N \gg n$ this results in a huge reduction of complexity. We give a simple geometric characterization of barycenters via negative cones and derive a randomized algorithm that computes this new measure by "greedy geometric sampling". We then study its properties, and benchmark it on synthetic and real-world data to show that it can be very beneficial in the $N\gg n$ regime. A Python implementation is available at \url{this https URL}. \ No newline at end of file diff --git a/data/2020/neurips/A Ranking-based, Balanced Loss Function Unifying Classification and Localisation in Object Detection b/data/2020/neurips/A Ranking-based, Balanced Loss Function Unifying Classification and Localisation in Object Detection new file mode 100644 index 0000000000..49d13aa812 --- /dev/null +++ b/data/2020/neurips/A Ranking-based, Balanced Loss Function Unifying Classification and Localisation in Object Detection @@ -0,0 +1 @@ +We propose \textit{average Localisation-Recall-Precision} (aLRP), a unified, bounded, balanced and ranking-based loss function for both classification and localisation tasks in object detection. aLRP extends the Localisation-Recall-Precision (LRP) performance metric (Oksuz et al., 2018) inspired from how Average Precision (AP) Loss extends precision to a ranking-based loss function for classification (Chen et al., 2020). aLRP has the following distinct advantages: (i) aLRP is the first ranking-based loss function for both classification and localisation tasks. (ii) Thanks to using ranking for both tasks, aLRP naturally enforces high-quality localisation for high-precision classification. (iii) aLRP provides provable balance between positives and negatives. (iv) Compared to on average $\sim$6 hyperparameters in the loss functions of state-of-the-art detectors, aLRP Loss has only one hyperparameter, which we did not tune in practice. On the COCO dataset, aLRP Loss improves its ranking-based predecessor, AP Loss, up to around $5$ AP points, achieves $48.9$ AP without test time augmentation and outperforms all one-stage detectors. Code available at: this https URL . \ No newline at end of file diff --git a/data/2020/neurips/A Robust Functional EM Algorithm for Incomplete Panel Count Data b/data/2020/neurips/A Robust Functional EM Algorithm for Incomplete Panel Count Data new file mode 100644 index 0000000000..6439e382bf --- /dev/null +++ b/data/2020/neurips/A Robust Functional EM Algorithm for Incomplete Panel Count Data @@ -0,0 +1 @@ +Panel count data describes aggregated counts of recurrent events observed at discrete time points. To understand dynamics of health behaviors and predict future negative events, the field of quantitative behavioral research has evolved to increasingly rely upon panel count data collected via multiple self reports, for example, about frequencies of smoking using in-the-moment surveys on mobile devices. However, missing reports are common and present a major barrier to downstream statistical learning. As a first step, under a missing completely at random assumption (MCAR), we propose a simple yet widely applicable functional EM algorithm to estimate the counting process mean function, which is of central interest to behavioral scientists. The proposed approach wraps several popular panel count inference methods, seamlessly deals with incomplete counts and is robust to misspecification of the Poisson process assumption. Theoretical analysis of the proposed algorithm provides finite-sample guarantees by expanding parametric EM theory [3, 34] to the general non-parametric setting. We illustrate the utility of the proposed algorithm through numerical experiments and an analysis of smoking cessation data. We also discuss useful extensions to address deviations from the MCAR assumption and covariate effects. \ No newline at end of file diff --git a/data/2020/neurips/A Scalable Approach for Privacy-Preserving Collaborative Machine Learning b/data/2020/neurips/A Scalable Approach for Privacy-Preserving Collaborative Machine Learning new file mode 100644 index 0000000000..fd26aae946 --- /dev/null +++ b/data/2020/neurips/A Scalable Approach for Privacy-Preserving Collaborative Machine Learning @@ -0,0 +1 @@ +We consider a collaborative learning scenario in which multiple data-owners wish to jointly train a logistic regression model, while keeping their individual datasets private from the other parties. We propose COPML, a fully-decentralized training framework that achieves scalability and privacy-protection simultaneously. The key idea of COPML is to securely encode the individual datasets to distribute the computation load effectively across many parties and to perform the training computations as well as the model updates in a distributed manner on the securely encoded data. We provide the privacy analysis of COPML and prove its convergence. Furthermore, we experimentally demonstrate that COPML can achieve significant speedup in training over the benchmark protocols. Our protocol provides strong statistical privacy guarantees against colluding parties (adversaries) with unbounded computational power, while achieving up to $16\times$ speedup in the training time against the benchmark protocols. \ No newline at end of file diff --git a/data/2020/neurips/A Scalable MIP-based Method for Learning Optimal Multivariate Decision Trees b/data/2020/neurips/A Scalable MIP-based Method for Learning Optimal Multivariate Decision Trees new file mode 100644 index 0000000000..94282f5ba8 --- /dev/null +++ b/data/2020/neurips/A Scalable MIP-based Method for Learning Optimal Multivariate Decision Trees @@ -0,0 +1 @@ +Several recent publications report advances in training optimal decision trees (ODT) using mixed-integer programs (MIP), due to algorithmic advances in integer programming and a growing interest in addressing the inherent suboptimality of heuristic approaches such as CART. In this paper, we propose a novel MIP formulation, based on a 1-norm support vector machine model, to train a multivariate ODT for classification problems. We provide cutting plane techniques that tighten the linear relaxation of the MIP formulation, in order to improve run times to reach optimality. Using 36 data-sets from the University of California Irvine Machine Learning Repository, we demonstrate that our formulation outperforms its counterparts in the literature by an average of about 10% in terms of mean out-of-sample testing accuracy across the data-sets. We provide a scalable framework to train multivariate ODT on large data-sets by introducing a novel linear programming (LP) based data selection method to choose a subset of the data for training. Our method is able to routinely handle large data-sets with more than 7,000 sample points and outperform heuristics methods and other MIP based techniques. We present results on data-sets containing up to 245,000 samples. Existing MIP-based methods do not scale well on training data-sets beyond 5,500 samples. \ No newline at end of file diff --git a/data/2020/neurips/A Self-Tuning Actor-Critic Algorithm b/data/2020/neurips/A Self-Tuning Actor-Critic Algorithm new file mode 100644 index 0000000000..35bda8eeec --- /dev/null +++ b/data/2020/neurips/A Self-Tuning Actor-Critic Algorithm @@ -0,0 +1 @@ +Reinforcement learning algorithms are highly sensitive to the choice of hyperparameters, typically requiring significant manual effort to identify hyperparameters that perform well on a new domain. In this paper, we take a step towards addressing this issue by using metagradients to automatically adapt hyperparameters online by meta-gradient descent (Xu et al., 2018). We apply our algorithm, Self-Tuning Actor-Critic (STAC), to self-tune all the differentiable hyperparameters of an actor-critic loss function, to discover auxiliary tasks, and to improve off-policy learning using a novel leaky V-trace operator. STAC is simple to use, sample efficient and does not require a significant increase in compute. Ablative studies show that the overall performance of STAC improved as we adapt more hyperparameters. When applied to the Arcade Learning Environment (Bellemare et al. 2012), STAC improved the median human normalized score in $200$M steps from $243\%$ to $364\%$. When applied to the DM Control suite (Tassa et al., 2018), STAC improved the mean score in $30$M steps from $217$ to $389$ when learning with features, from $108$ to $202$ when learning from pixels, and from $195$ to $295$ in the Real-World Reinforcement Learning Challenge (Dulac-Arnold et al., 2020). \ No newline at end of file diff --git a/data/2020/neurips/A Simple Language Model for Task-Oriented Dialogue b/data/2020/neurips/A Simple Language Model for Task-Oriented Dialogue new file mode 100644 index 0000000000..1fca3d1040 --- /dev/null +++ b/data/2020/neurips/A Simple Language Model for Task-Oriented Dialogue @@ -0,0 +1 @@ +Task-oriented dialogue is often decomposed into three tasks: understanding user input, deciding actions, and generating a response. While such decomposition might suggest a dedicated model for each sub-task, we find a simple, unified approach leads to state-of-the-art performance on the MultiWOZ dataset. SimpleTOD is a simple approach to task-oriented dialogue that uses a single causal language model trained on all sub-tasks recast as a single sequence prediction problem. This allows SimpleTOD to fully leverage transfer learning from pre-trained, open domain, causal language models such as GPT-2. SimpleTOD improves over the prior state-of-the-art by 0.49 points in joint goal accuracy for dialogue state tracking. More impressively, SimpleTOD also improves the main metrics used to evaluate action decisions and response generation in an end-to-end setting for task-oriented dialog systems: inform rate by 8.1 points, success rate by 9.7 points, and combined score by 7.2 points. \ No newline at end of file diff --git a/data/2020/neurips/A Simple and Efficient Smoothing Method for Faster Optimization and Local Exploration b/data/2020/neurips/A Simple and Efficient Smoothing Method for Faster Optimization and Local Exploration new file mode 100644 index 0000000000..469f90378f --- /dev/null +++ b/data/2020/neurips/A Simple and Efficient Smoothing Method for Faster Optimization and Local Exploration @@ -0,0 +1 @@ +This work proposes a novel smoothing method, called Bend, Mix and Release (BMR) , that extends two well-known smooth approximations of the convex optimization literature: randomized smoothing and the Moreau envelope. The BMR smoothing method allows to trade-off between the computational simplicity of randomized smoothing (RS) and the approximation efficiency of the Moreau envelope (ME). More specifically, we show that BMR achieves up to a √ d multiplicative improvement compared to the approximation error of RS, where d is the dimension of the search space, while being less computation intensive than the ME. For non-convex objectives, BMR also has the desirable property to widen local minima, allowing optimization methods to reach small cracks and crevices of extremely irregular and non-convex functions, while being well-suited to a distributed setting. This novel smoothing method is then used to improve first-order non-smooth optimization (both convex and non-convex) by allowing for a local exploration of the search space. More specifically, our analysis sheds light on the similarities be-tween evolution strategies and BMR, creating a link between exploration strategies of zeroth-order methods and the regularity of first-order optimization problems. Finally, we evidence the impact of BMR through synthetic experiments. \ No newline at end of file diff --git a/data/2020/neurips/A Single Recipe for Online Submodular Maximization with Adversarial or Stochastic Constraints b/data/2020/neurips/A Single Recipe for Online Submodular Maximization with Adversarial or Stochastic Constraints new file mode 100644 index 0000000000..93a03de5d8 --- /dev/null +++ b/data/2020/neurips/A Single Recipe for Online Submodular Maximization with Adversarial or Stochastic Constraints @@ -0,0 +1 @@ +In this paper, we consider an online optimization problem in which the reward functions are DR-submodular, and in addition to maximizing the total reward, the sequence of decisions must satisfy some convex constraints on average. Specifically, at each round t ∈ { 1 , . . . , T } , upon committing to an action x t , a DR-submodular utility function f t ( · ) and a convex constraint function g t ( · ) are revealed, and the goal is to maximize the overall utility while ensuring the average of the constraint functions 1 T (cid:80) Tt =1 g t ( x t ) is non-positive. Such cumulative constraints arise naturally in applications where the average resource consumption is required to remain below a prespecified threshold. We study this problem under an adversarial model and a stochastic model for the convex constraints, where the functions g t can vary arbitrarily or according to an i.i.d. process over time slots t ∈ { 1 , . . . , T } , respectively. We propose a single algorithm which achieves sub-linear (with respect to T ) regret as well as sub-linear constraint violation bounds in both settings, without prior knowledge of the regime. Prior works have studied this problem in the special case of linear constraint functions. Our results not only improve upon the existing bounds under linear cumulative constraints, but also give the first sub-linear bounds for general convex long-term constraints. \ No newline at end of file diff --git a/data/2020/neurips/A Single-Loop Smoothed Gradient Descent-Ascent Algorithm for Nonconvex-Concave Min-Max Problems b/data/2020/neurips/A Single-Loop Smoothed Gradient Descent-Ascent Algorithm for Nonconvex-Concave Min-Max Problems new file mode 100644 index 0000000000..b1d2728f7e --- /dev/null +++ b/data/2020/neurips/A Single-Loop Smoothed Gradient Descent-Ascent Algorithm for Nonconvex-Concave Min-Max Problems @@ -0,0 +1 @@ +Nonconvex-concave min-max problem arises in many machine learning applications including minimizing a pointwise maximum of a set of nonconvex functions and robust adversarial training of neural networks. A popular approach to solve this problem is the gradient descent-ascent (GDA) algorithm which unfortunately can exhibit oscillation in case of nonconvexity. In this paper, we introduce a "smoothing" scheme which can be combined with GDA to stabilize the oscillation and ensure convergence to a stationary solution. We prove that the stabilized GDA algorithm can achieve an $O(1/\epsilon^2)$ iteration complexity for minimizing the pointwise maximum of a finite collection of nonconvex functions. Moreover, the smoothed GDA algorithm achieves an $O(1/\epsilon^4)$ iteration complexity for general nonconvex-concave problems. Extensions of this stabilized GDA algorithm to multi-block cases are presented. To the best of our knowledge, this is the first algorithm to achieve $O(1/\epsilon^2)$ for a class of nonconvex-concave problem. We illustrate the practical efficiency of the stabilized GDA algorithm on robust training. \ No newline at end of file diff --git a/data/2020/neurips/A Spectral Energy Distance for Parallel Speech Synthesis b/data/2020/neurips/A Spectral Energy Distance for Parallel Speech Synthesis new file mode 100644 index 0000000000..6610386e5e --- /dev/null +++ b/data/2020/neurips/A Spectral Energy Distance for Parallel Speech Synthesis @@ -0,0 +1 @@ +Speech synthesis is an important practical generative modeling problem that has seen great progress over the last few years, with likelihood-based autoregressive neural models now outperforming traditional concatenative systems. A downside of such autoregressive models is that they require executing tens of thousands of sequential operations per second of generated audio, making them ill-suited for deployment on specialized deep learning hardware. Here, we propose a new learning method that allows us to train highly parallel models of speech, without requiring access to an analytical likelihood function. Our approach is based on a generalized energy distance between the distributions of the generated and real audio. This spectral energy distance is a proper scoring rule with respect to the distribution over magnitude-spectrograms of the generated waveform audio and offers statistical consistency guarantees. The distance can be calculated from minibatches without bias, and does not involve adversarial learning, yielding a stable and consistent method for training implicit generative models. Empirically, we achieve state-of-the-art generation quality among implicit generative models, as judged by the recently-proposed cFDSD metric. When combining our method with adversarial techniques, we also improve upon the recently-proposed GAN-TTS model in terms of Mean Opinion Score as judged by trained human evaluators. \ No newline at end of file diff --git a/data/2020/neurips/A Statistical Framework for Low-bitwidth Training of Deep Neural Networks b/data/2020/neurips/A Statistical Framework for Low-bitwidth Training of Deep Neural Networks new file mode 100644 index 0000000000..f876fae692 --- /dev/null +++ b/data/2020/neurips/A Statistical Framework for Low-bitwidth Training of Deep Neural Networks @@ -0,0 +1 @@ +Fully quantized training (FQT), which uses low-bitwidth hardware by quantizing the activations, weights, and gradients of a neural network model, is a promising approach to accelerate the training of deep neural networks. One major challenge with FQT is the lack of theoretical understanding, in particular of how gradient quantization impacts convergence properties. In this paper, we address this problem by presenting a statistical framework for analyzing FQT algorithms. We view the quantized gradient of FQT as a stochastic estimator of its full precision counterpart, a procedure known as quantization-aware training (QAT). We show that the FQT gradient is an unbiased estimator of the QAT gradient, and we discuss the impact of gradient quantization on its variance. Inspired by these theoretical results, we develop two novel gradient quantizers, and we show that these have smaller variance than the existing per-tensor quantizer. For training ResNet-50 on ImageNet, our 5-bit block Householder quantizer achieves only 0.5% validation accuracy loss relative to QAT, comparable to the existing INT8 baseline. \ No newline at end of file diff --git a/data/2020/neurips/A Statistical Mechanics Framework for Task-Agnostic Sample Design in Machine Learning b/data/2020/neurips/A Statistical Mechanics Framework for Task-Agnostic Sample Design in Machine Learning new file mode 100644 index 0000000000..d8550b9761 --- /dev/null +++ b/data/2020/neurips/A Statistical Mechanics Framework for Task-Agnostic Sample Design in Machine Learning @@ -0,0 +1 @@ +In this paper, we present a statistical mechanics framework to understand the effect of sampling properties of training data on the generalization gap of machine learning (ML) algorithms. We connect the generalization gap to the spatial properties of a sample design characterized by the pair correlation function (PCF). In particular, we express generalization gap in terms of the power spectra of the sample design and that of the function to be learned. Using this framework, we show that space-filling sample designs, such as blue noise and Poisson disk sampling, which optimize spectral properties, outperform random designs in terms of the generalization gap and characterize this gain in a closed-form. Our analysis also sheds light on design principles for constructing optimal task-agnostic sample designs that minimize the generalization gap. We corroborate our findings using regression experiments with neural networks on: a) synthetic functions, and b) a complex scientific simulator for inertial confinement fusion (ICF). \ No newline at end of file diff --git a/data/2020/neurips/A Stochastic Path Integral Differential EstimatoR Expectation Maximization Algorithm b/data/2020/neurips/A Stochastic Path Integral Differential EstimatoR Expectation Maximization Algorithm new file mode 100644 index 0000000000..93c0589edd --- /dev/null +++ b/data/2020/neurips/A Stochastic Path Integral Differential EstimatoR Expectation Maximization Algorithm @@ -0,0 +1 @@ +The Expectation Maximization (EM) algorithm is of key importance for inference in latent variable models including mixture of regressors and experts, missing observations. This paper introduces a novel EM algorithm, called SPIDER-EM, for inference from a training set of size n, n ≫ 1. At the core of our algorithm is an estimator of the full conditional expectation in the E-step, adapted from the stochastic path-integrated differential estimator (SPIDER) technique. We derive finite-time complexity bounds for smooth non-convex likelihood: we show that for convergence to an ǫ-approximate stationary point, the complexity scales as K Opt (n, ǫ) = O(ǫ −1) and K CE (n, ǫ) = n + √ nO(ǫ −1), where K Opt (n, ǫ) and K CE (n, ǫ) are respectively the number of M-steps and the number of per-sample conditional expectations evaluations. This improves over the state-of-the-art algorithms. Numerical results support our findings. \ No newline at end of file diff --git a/data/2020/neurips/A Study on Encodings for Neural Architecture Search b/data/2020/neurips/A Study on Encodings for Neural Architecture Search new file mode 100644 index 0000000000..cef2b48831 --- /dev/null +++ b/data/2020/neurips/A Study on Encodings for Neural Architecture Search @@ -0,0 +1,2 @@ +Neural architecture search (NAS) has been extensively studied in the past few years. A popular approach is to represent each neural architecture in the search space as a directed acyclic graph (DAG), and then search over all DAGs by encoding the adjacency matrix and list of operations as a set of hyperparameters. Recent work has demonstrated that even small changes to the way each architecture is encoded can have a significant effect on the performance of NAS algorithms. +In this work, we present the first formal study on the effect of architecture encodings for NAS, including a theoretical grounding and an empirical study. First we formally define architecture encodings and give a theoretical characterization on the scalability of the encodings we study Then we identify the main encoding-dependent subroutines which NAS algorithms employ, running experiments to show which encodings work best with each subroutine for many popular algorithms. The experiments act as an ablation study for prior work, disentangling the algorithmic and encoding-based contributions, as well as a guideline for future work. Our results demonstrate that NAS encodings are an important design decision which can have a significant impact on overall performance. Our code is available at this https URL. \ No newline at end of file diff --git a/data/2020/neurips/A Theoretical Framework for Target Propagation b/data/2020/neurips/A Theoretical Framework for Target Propagation new file mode 100644 index 0000000000..7cec602bcf --- /dev/null +++ b/data/2020/neurips/A Theoretical Framework for Target Propagation @@ -0,0 +1 @@ +The success of deep learning, a brain-inspired form of AI, has sparked interest in understanding how the brain could similarly learn across multiple layers of neurons. However, the majority of biologically-plausible learning algorithms have not yet reached the performance of backpropagation (BP), nor are they built on strong theoretical foundations. Here, we analyze target propagation (TP), a popular but not yet fully understood alternative to BP, from the standpoint of mathematical optimization. Our theory shows that TP is closely related to Gauss-Newton optimization and thus substantially differs from BP. Furthermore, our analysis reveals a fundamental limitation of difference target propagation (DTP), a well-known variant of TP, in the realistic scenario of non-invertible neural networks. We provide a first solution to this problem through a novel reconstruction loss that improves feedback weight training, while simultaneously introducing architectural flexibility by allowing for direct feedback connections from the output to each hidden layer. Our theory is corroborated by experimental results that show significant improvements in performance and in the alignment of forward weight updates with loss gradients, compared to DTP. \ No newline at end of file diff --git a/data/2020/neurips/A Tight Lower Bound and Efficient Reduction for Swap Regret b/data/2020/neurips/A Tight Lower Bound and Efficient Reduction for Swap Regret new file mode 100644 index 0000000000..1a34e0a514 --- /dev/null +++ b/data/2020/neurips/A Tight Lower Bound and Efficient Reduction for Swap Regret @@ -0,0 +1 @@ +Swap regret, a generic performance measure of online decision-making algorithms, plays an important role in the theory of repeated games, along with a close connection to correlated equilibria in strategic games. This paper shows an Ω( √ TN log N ) -lower bound for swap regret, where T and N denote the numbers of time steps and available actions, respectively. Our lower bound is tight up to a constant, and resolves an open problem mentioned, e.g., in the book by Nisan et al. [28]. Besides, we present a computationally efficient reduction method that converts no-external-regret algorithms to no-swap-regret algorithms. This method can be applied not only to the full-information setting but also to the bandit setting and provides a better regret bound than previous results. \ No newline at end of file diff --git a/data/2020/neurips/A Topological Filter for Learning with Label Noise b/data/2020/neurips/A Topological Filter for Learning with Label Noise new file mode 100644 index 0000000000..f1df4e013a --- /dev/null +++ b/data/2020/neurips/A Topological Filter for Learning with Label Noise @@ -0,0 +1 @@ +Noisy labels can impair the performance of deep neural networks. To tackle this problem, in this paper, we propose a new method for filtering label noise. Unlike most existing methods relying on the posterior probability of a noisy classifier, we focus on the much richer spatial behavior of data in the latent representational space. By leveraging the high-order topological information of data, we are able to collect most of the clean data and train a high-quality model. Theoretically we prove that this topological approach is guaranteed to collect the clean data with high probability. Empirical results show that our method outperforms the state-of-the-arts and is robust to a broad spectrum of noise types and levels. \ No newline at end of file diff --git a/data/2020/neurips/A Unified Switching System Perspective and Convergence Analysis of Q-Learning Algorithms b/data/2020/neurips/A Unified Switching System Perspective and Convergence Analysis of Q-Learning Algorithms new file mode 100644 index 0000000000..10a655eed3 --- /dev/null +++ b/data/2020/neurips/A Unified Switching System Perspective and Convergence Analysis of Q-Learning Algorithms @@ -0,0 +1 @@ +This paper develops a novel and unified framework to analyze the convergence of a large family of Q-learning algorithms from the switching system perspective. We show that the nonlinear ODE models associated with Q-learning and many of its variants can be naturally formulated as affine switching systems . Building on their asymptotic stability, we obtain a number of interesting results: (i) we provide a simple ODE analysis for the convergence of asynchronous Q-learning under relatively weak assumptions; (ii) we establish the first convergence analysis of the averaging Q-learning algorithm, and (iii) we derive a new sufficient condition for the convergence of Q-learning with linear function approximation \ No newline at end of file diff --git a/data/2020/neurips/A Unified View of Label Shift Estimation b/data/2020/neurips/A Unified View of Label Shift Estimation new file mode 100644 index 0000000000..0258b85a67 --- /dev/null +++ b/data/2020/neurips/A Unified View of Label Shift Estimation @@ -0,0 +1 @@ +Label shift describes the setting where although the label distribution might change between the source and target domains, the class-conditional probabilities (of data given a label) do not. There are two dominant approaches for estimating the label marginal. BBSE, a moment-matching approach based on confusion matrices, is provably consistent and provides interpretable error bounds. However, a maximum likelihood estimation approach, which we call MLLS, dominates empirically. In this paper, we present a unified view of the two methods and the first theoretical characterization of the likelihood-based estimator. Our contributions include (i) conditions for consistency of MLLS, which include calibration of the classifier and a confusion matrix invertibility condition that BBSE also requires; (ii) a unified view of the methods, casting the confusion matrix as roughly equivalent to MLLS for a particular choice of calibration method; and (iii) a decomposition of MLLS's finite-sample error into terms reflecting the impacts of miscalibration and estimation error. Our analysis attributes BBSE's statistical inefficiency to a loss of information due to coarse calibration. We support our findings with experiments on both synthetic data and the MNIST and CIFAR10 image recognition datasets. \ No newline at end of file diff --git a/data/2020/neurips/A Unifying View of Optimism in Episodic Reinforcement Learning b/data/2020/neurips/A Unifying View of Optimism in Episodic Reinforcement Learning new file mode 100644 index 0000000000..274a2471bf --- /dev/null +++ b/data/2020/neurips/A Unifying View of Optimism in Episodic Reinforcement Learning @@ -0,0 +1 @@ +The principle of optimism in the face of uncertainty underpins many theoretically successful reinforcement learning algorithms. In this paper we provide a general framework for designing, analyzing and implementing such algorithms in the episodic reinforcement learning problem. This framework is built upon Lagrangian duality, and demonstrates that every model-optimistic algorithm that constructs an optimistic MDP has an equivalent representation as a value-optimistic dynamic programming algorithm. Typically, it was thought that these two classes of algorithms were distinct, with model-optimistic algorithms benefiting from a cleaner probabilistic analysis while value-optimistic algorithms are easier to implement and thus more practical. With the framework developed in this paper, we show that it is possible to get the best of both worlds by providing a class of algorithms which have a computationally efficient dynamic-programming implementation and also a simple probabilistic analysis. Besides being able to capture many existing algorithms in the tabular setting, our framework can also address largescale problems under realizable function approximation, where it enables a simple model-based analysis of some recently proposed methods. \ No newline at end of file diff --git a/data/2020/neurips/A Universal Approximation Theorem of Deep Neural Networks for Expressing Probability Distributions b/data/2020/neurips/A Universal Approximation Theorem of Deep Neural Networks for Expressing Probability Distributions new file mode 100644 index 0000000000..a5178cae10 --- /dev/null +++ b/data/2020/neurips/A Universal Approximation Theorem of Deep Neural Networks for Expressing Probability Distributions @@ -0,0 +1 @@ +This paper studies the universal approximation property of deep neural networks for representing probability distributions. Given a target distribution $\pi$ and a source distribution $p_z$ both defined on $\mathbb{R}^d$, we prove under some assumptions that there exists a deep neural network $g:\mathbb{R}^d\rightarrow \mathbb{R}$ with ReLU activation such that the push-forward measure $(\nabla g)_\# p_z$ of $p_z$ under the map $\nabla g$ is arbitrarily close to the target measure $\pi$. The closeness are measured by three classes of integral probability metrics between probability distributions: $1$-Wasserstein distance, maximum mean distance (MMD) and kernelized Stein discrepancy (KSD). We prove upper bounds for the size (width and depth) of the deep neural network in terms of the dimension $d$ and the approximation error $\varepsilon$ with respect to the three discrepancies. In particular, the size of neural network can grow exponentially in $d$ when $1$-Wasserstein distance is used as the discrepancy, whereas for both MMD and KSD the size of neural network only depends on $d$ at most polynomially. Our proof relies on convergence estimates of empirical measures under aforementioned discrepancies and semi-discrete optimal transport. \ No newline at end of file diff --git a/data/2020/neurips/A Variational Approach for Learning from Positive and Unlabeled Data b/data/2020/neurips/A Variational Approach for Learning from Positive and Unlabeled Data new file mode 100644 index 0000000000..339c015110 --- /dev/null +++ b/data/2020/neurips/A Variational Approach for Learning from Positive and Unlabeled Data @@ -0,0 +1 @@ +Learning binary classifiers only from positive and unlabeled (PU) data is an important and challenging task in many real-world applications, including web text classification, disease gene identification and fraud detection, where negative samples are difficult to verify experimentally. Most recent PU learning methods are developed based on the conventional misclassification risk of the supervised learning type, and they require to solve the intractable risk estimation problem by approximating the negative data distribution or the class prior. In this paper, we introduce a variational principle for PU learning that allows us to quantitatively evaluate the modeling error of the Bayesian classifier directly from given data. This leads to a loss function which can be efficiently calculated without any intermediate step or model, and a variational learning method can then be employed to optimize the classifier under general conditions. In addition, the discriminative performance and numerical stability of the variational PU learning method can be further improved by incorporating a margin maximizing loss function. We illustrate the effectiveness of the proposed variational method on a number of benchmark examples. \ No newline at end of file diff --git a/data/2020/neurips/A causal view of compositional zero-shot recognition b/data/2020/neurips/A causal view of compositional zero-shot recognition new file mode 100644 index 0000000000..6015f5e6ec --- /dev/null +++ b/data/2020/neurips/A causal view of compositional zero-shot recognition @@ -0,0 +1,2 @@ +People easily recognize new visual categories that are new combinations of known components. This compositional generalization capacity is critical for learning in real-world domains like vision and language because the long tail of new combinations dominates the distribution. Unfortunately, learning systems struggle with compositional generalization because they often build on features that are correlated with class labels even if they are not "essential" for the class. This leads to consistent misclassification of samples from a new distribution, like new combinations of known components. +Here we describe an approach for compositional generalization that builds on causal ideas. First, we describe compositional zero-shot learning from a causal perspective, and propose to view zero-shot inference as finding "which intervention caused the image?". Second, we present a causal-inspired embedding model that learns disentangled representations of elementary components of visual objects from correlated (confounded) training data. We evaluate this approach on two datasets for predicting new combinations of attribute-object pairs: A well-controlled synthesized images dataset and a real world dataset which consists of fine-grained types of shoes. We show improvements compared to strong baselines. \ No newline at end of file diff --git a/data/2020/neurips/A convex optimization formulation for multivariate regression b/data/2020/neurips/A convex optimization formulation for multivariate regression new file mode 100644 index 0000000000..b1826e441a --- /dev/null +++ b/data/2020/neurips/A convex optimization formulation for multivariate regression @@ -0,0 +1 @@ +Multivariate regression (or multi-task learning) concerns the task of predicting the value of multiple responses from a set of covariates. In this article, we pro-pose a convex optimization formulation for high-dimensional multivariate linear regression under a general error covariance structure. The main difficulty with simultaneous estimation of the regression coefficients and the error covariance matrix lies in the fact that the negative log-likelihood function is not convex. To overcome this difficulty, a new parameterization is proposed, under which the negative log-likelihood function is proved to be convex. For faster computation, two other alternative loss functions are also considered, and proved to be convex under the proposed parameterization. This new parameterization is also useful for covariate-adjusted Gaussian graphical modeling in which the inverse of the error covariance matrix is of interest. A joint non-asymptotic analysis of the regression coefficients and the error covariance matrix is carried out under the new parameterization. In particular, we show that the proposed method recovers the oracle estimator under sharp scaling conditions, and rates of convergence in terms of vector (cid:96) ∞ norm are also established. Empirically, the proposed methods outperform existing high-dimensional multivariate linear regression methods that are based on either minimizing certain non-convex criteria or certain two-step procedures. \ No newline at end of file diff --git a/data/2020/neurips/A game-theoretic analysis of networked system control for common-pool resource management using multi-agent reinforcement learning b/data/2020/neurips/A game-theoretic analysis of networked system control for common-pool resource management using multi-agent reinforcement learning new file mode 100644 index 0000000000..8de9373cdc --- /dev/null +++ b/data/2020/neurips/A game-theoretic analysis of networked system control for common-pool resource management using multi-agent reinforcement learning @@ -0,0 +1 @@ +Multi-agent reinforcement learning has recently shown great promise as an approach to networked system control. Arguably, one of the most difficult and important tasks for which large scale networked system control is applicable is common-pool resource management. Crucial common-pool resources include arable land, fresh water, wetlands, wildlife, fish stock, forests and the atmosphere, of which proper management is related to some of society's greatest challenges such as food security, inequality and climate change. Here we take inspiration from a recent research program investigating the game-theoretic incentives of humans in social dilemma situations such as the well-known tragedy of the commons. However, instead of focusing on biologically evolved human-like agents, our concern is rather to better understand the learning and operating behaviour of engineered networked systems comprising general-purpose reinforcement learning agents, subject only to nonbiological constraints such as memory, computation and communication bandwidth. Harnessing tools from empirical game-theoretic analysis, we analyse the differences in resulting solution concepts that stem from employing different information structures in the design of networked multi-agent systems. These information structures pertain to the type of information shared between agents as well as the employed communication protocol and network topology. Our analysis contributes new insights into the consequences associated with certain design choices and provides an additional dimension of comparison between systems beyond efficiency, robustness, scalability and mean control performance. \ No newline at end of file diff --git a/data/2020/neurips/A kernel test for quasi-independence b/data/2020/neurips/A kernel test for quasi-independence new file mode 100644 index 0000000000..2f770e9a38 --- /dev/null +++ b/data/2020/neurips/A kernel test for quasi-independence @@ -0,0 +1 @@ +We consider settings in which the data of interest correspond to pairs of ordered times, e.g, the birth times of the first and second child, the times at which a new user creates an account and makes the first purchase on a website, and the entry and survival times of patients in a clinical trial. In these settings, the two times are not independent (the second occurs after the first), yet it is still of interest to determine whether there exists significant dependence {\em beyond} their ordering in time. We refer to this notion as "quasi-(in)dependence". For instance, in a clinical trial, to avoid biased selection, we might wish to verify that recruitment times are quasi-independent of survival times, where dependencies might arise due to seasonal effects. In this paper, we propose a nonparametric statistical test of quasi-independence. Our test considers a potentially infinite space of alternatives, making it suitable for complex data where the nature of the possible quasi-dependence is not known in advance. Standard parametric approaches are recovered as special cases, such as the classical conditional Kendall's tau, and log-rank tests. The tests apply in the right-censored setting: an essential feature in clinical trials, where patients can withdraw from the study. We provide an asymptotic analysis of our test-statistic, and demonstrate in experiments that our test obtains better power than existing approaches, while being more computationally efficient. \ No newline at end of file diff --git a/data/2020/neurips/A mathematical model for automatic differentiation in machine learning b/data/2020/neurips/A mathematical model for automatic differentiation in machine learning new file mode 100644 index 0000000000..1dab52015b --- /dev/null +++ b/data/2020/neurips/A mathematical model for automatic differentiation in machine learning @@ -0,0 +1 @@ +Automatic differentiation, as implemented today, does not have a simple mathematical model adapted to the needs of modern machine learning. In this work we articulate the relationships between differentiation of programs as implemented in practice and differentiation of nonsmooth functions. To this end we provide a simple class of functions, a nonsmooth calculus, and show how they apply to stochastic approximation methods. We also evidence the issue of artificial critical points created by algorithmic differentiation and show how usual methods avoid these points with probability one. \ No newline at end of file diff --git a/data/2020/neurips/A mathematical theory of cooperative communication b/data/2020/neurips/A mathematical theory of cooperative communication new file mode 100644 index 0000000000..6fc796ee9a --- /dev/null +++ b/data/2020/neurips/A mathematical theory of cooperative communication @@ -0,0 +1 @@ +Cooperative communication plays a central role in theories of human cognition, language, development, culture, and human-robot interaction. Prior models of cooperative communication are algorithmic in nature and do not shed light on why cooperation may yield effective belief transmission and what limitations may arise due to differences between beliefs of agents. Through a connection to the theory of optimal transport, we establishing a mathematical framework for cooperative communication. We derive prior models as special cases, statistical interpretations of belief transfer plans, and proofs of robustness and instability. Computational simulations support and elaborate our theoretical results, and demonstrate fit to human behavior. The results show that cooperative communication provably enables effective, robust belief transmission which is required to explain feats of human learning and improve human-machine interaction. \ No newline at end of file diff --git a/data/2020/neurips/A mean-field analysis of two-player zero-sum games b/data/2020/neurips/A mean-field analysis of two-player zero-sum games new file mode 100644 index 0000000000..23a275fbe0 --- /dev/null +++ b/data/2020/neurips/A mean-field analysis of two-player zero-sum games @@ -0,0 +1 @@ +Finding Nash equilibria in two-player zero-sum continuous games is a central problem in machine learning, e.g. for training both GANs and robust models. The existence of pure Nash equilibria requires strong conditions which are not typically met in practice. Mixed Nash equilibria exist in greater generality and may be found using mirror descent. Yet this approach does not scale to high dimensions. To address this limitation, we parametrize mixed strategies as mixtures of particles, whose positions and weights are updated using gradient descent-ascent. We study this dynamics as an interacting gradient flow over measure spaces endowed with the Wasserstein-Fisher-Rao metric. We establish global convergence to an approximate equilibrium for the related Langevin gradient-ascent dynamic. We prove a law of large numbers that relates particle dynamics to mean-field dynamics. Our method identifies mixed equilibria in high dimensions and is demonstrably effective for training mixtures of GANs. \ No newline at end of file diff --git a/data/2020/neurips/A meta-learning approach to (re)discover plasticity rules that carve a desired function into a neural network b/data/2020/neurips/A meta-learning approach to (re)discover plasticity rules that carve a desired function into a neural network new file mode 100644 index 0000000000..2a2df27a45 --- /dev/null +++ b/data/2020/neurips/A meta-learning approach to (re)discover plasticity rules that carve a desired function into a neural network @@ -0,0 +1 @@ +The search for biologically faithful synaptic plasticity rules has resulted in a large body of models. They are usually inspired by – and fitted to – experimental data, but they rarely produce neural dynamics that serve complex functions. These failures suggest that current plasticity models are still under-constrained by existing data. Here, we present an alternative approach that uses meta-learning to discover plausible synaptic plasticity rules. Instead of experimental data, the rules are constrained by the functions they implement and the structure they are meant to produce. Briefly, we parameterize synaptic plasticity rules by a Volterra expansion and then use supervised learning methods (gradient descent or evolutionary strategies) to minimize a problem-dependent loss function that quantifies how effectively a candidate plasticity rule transforms an initially random network into one with the desired function. We first validate our approach by re-discovering previously described plasticity rules, starting at the single-neuron level and “Oja’s rule”, a simple Hebbian plasticity rule that captures the direction of most variability of inputs to a neuron (i.e., the first principal component). We expand the problem to the network level and ask the framework to find Oja’s rule together with an anti-Hebbian rule such that an initially random two-layer firing-rate network will recover several principal components of the input space after learning. Next, we move to networks of integrate-and-fire neurons with plastic inhibitory afferents. We train for rules that achieve a target firing rate by countering tuned excitation. Our algorithm discovers a specific subset of the manifold of rules that can solve this task. Our work is a proof of principle of an automated and unbiased approach to unveil synaptic plasticity rules that obey biological constraints and can solve complex functions. \ No newline at end of file diff --git a/data/2020/neurips/A new convergent variant of Q-learning with linear function approximation b/data/2020/neurips/A new convergent variant of Q-learning with linear function approximation new file mode 100644 index 0000000000..6ca967fbe2 --- /dev/null +++ b/data/2020/neurips/A new convergent variant of Q-learning with linear function approximation @@ -0,0 +1 @@ +In this work, we identify a novel set of conditions that ensure convergence with probability 1 of Q -learning with linear function approximation, by proposing a two time-scale variation thereof. In the faster time scale, the algorithm features an update similar to that of DQN, where the impact of bootstrapping is attenuated by using a Q -value estimate akin to that of the target network in DQN. The slower time-scale, in turn, can be seen as a modified target network update. We establish the convergence of our algorithm, provide an error bound and discuss our results in light of existing convergence results on reinforcement learning with function approximation. Finally, we illustrate the convergent behavior of our method in domains where standard Q -learning has previously been shown to diverge. \ No newline at end of file diff --git a/data/2020/neurips/A new inference approach for training shallow and deep generalized linear models of noisy interacting neurons b/data/2020/neurips/A new inference approach for training shallow and deep generalized linear models of noisy interacting neurons new file mode 100644 index 0000000000..eee470d6e9 --- /dev/null +++ b/data/2020/neurips/A new inference approach for training shallow and deep generalized linear models of noisy interacting neurons @@ -0,0 +1 @@ +Generalized linear models are one of the most efficient paradigms for predicting the correlated stochastic activity of neuronal networks in response to external stimuli, with applications in many brain areas. However, when dealing with complex stimuli, the inferred coupling parameters often do not generalize across different stimulus statistics, leading to degraded performance and blowup instabilities. Here, we develop a two-step inference strategy that allows us to train robust generalized linear models of interacting neurons, by explicitly separating the effects of correlations in the stimulus from network interactions in each training step. Applying this approach to the responses of retinal ganglion cells to complex visual stimuli, we show that, compared to classical methods, the models trained in this way exhibit improved performance, are more stable, yield robust interaction networks, and generalize well across complex visual statistics. The method can be extended to deep convolutional neural networks, leading to models with high predictive accuracy for both the neuron firing rates and their correlations. \ No newline at end of file diff --git a/data/2020/neurips/A novel variational form of the Schatten-$p$ quasi-norm b/data/2020/neurips/A novel variational form of the Schatten-$p$ quasi-norm new file mode 100644 index 0000000000..1b3ed68cf6 --- /dev/null +++ b/data/2020/neurips/A novel variational form of the Schatten-$p$ quasi-norm @@ -0,0 +1 @@ +The Schatten-$p$ quasi-norm with $p\in(0,1)$ has recently gained considerable attention in various low-rank matrix estimation problems offering significant benefits over relevant convex heuristics such as the nuclear norm. However, due to the nonconvexity of the Schatten-$p$ quasi-norm, minimization suffers from two major drawbacks: 1) the lack of theoretical guarantees and 2) the high computational cost which is demanded for the minimization task even for trivial tasks such as finding stationary points. In an attempt to reduce the high computational cost induced by Schatten-$p$ quasi-norm minimization, variational forms, which are defined over smaller-size matrix factors whose product equals the original matrix, have been proposed. Here, we propose and analyze a novel variational form of Schatten-$p$ quasi-norm which, for the first time in the literature, is defined for any continuous value of $p\in(0,1]$ and decouples along the columns of the factorized matrices. The proposed form can be considered as the natural generalization of the well-known variational form of the nuclear norm to the nonconvex case i.e., for $p\in(0,1)$. The resulting formulation gives way to SVD-free algorithms thus offering lower computational complexity than the one that is induced by the original definition of the Schatten-$p$ quasi-norm. A local optimality analysis is provided which shows~that we can arrive at a local minimum of the original Schatten-$p$ quasi-norm problem by reaching a local minimum of the matrix factorization based surrogate problem. In addition, for the case of the squared Frobenius loss with linear operators obeying the restricted isometry property (RIP), a rank-one update scheme is proposed, which offers a way to escape poor local minima. Finally, the efficiency of our approach is empirically shown on a matrix completion problem. \ No newline at end of file diff --git a/data/2020/neurips/A polynomial-time algorithm for learning nonparametric causal graphs b/data/2020/neurips/A polynomial-time algorithm for learning nonparametric causal graphs new file mode 100644 index 0000000000..be288efca8 --- /dev/null +++ b/data/2020/neurips/A polynomial-time algorithm for learning nonparametric causal graphs @@ -0,0 +1 @@ +We establish finite-sample guarantees for a polynomial-time algorithm for learning a nonlinear, nonparametric directed acyclic graphical (DAG) model from data. The analysis is model-free and does not assume linearity, additivity, independent noise, or faithfulness. Instead, we impose a condition on the residual variances that is closely related to previous work on linear models with equal variances. Compared to an optimal algorithm with oracle knowledge of the variable ordering, the additional cost of the algorithm is linear in the dimension $d$ and the number of samples $n$. Finally, we compare the proposed algorithm to existing approaches in a simulation study. \ No newline at end of file diff --git a/data/2020/neurips/A shooting formulation of deep learning b/data/2020/neurips/A shooting formulation of deep learning new file mode 100644 index 0000000000..26e14752ab --- /dev/null +++ b/data/2020/neurips/A shooting formulation of deep learning @@ -0,0 +1 @@ +Continuous-depth neural networks can be viewed as deep limits of discrete neural networks whose dynamics resemble a discretization of an ordinary differential equation (ODE). Although important steps have been taken to realize the advantages of such continuous formulations, most current techniques are not truly continuous-depth as they assume identical layers. Indeed, existing works throw into relief the myriad difficulties presented by an infinite-dimensional parameter space in learning a continuous-depth neural ODE. To this end, we introduce a shooting formulation which shifts the perspective from parameterizing a network layer-by-layer to parameterizing over optimal networks described only by a set of initial conditions. For scalability, we propose a novel particle-ensemble parametrization which fully specifies the optimal weight trajectory of the continuous-depth neural network. Our experiments show that our particle-ensemble shooting formulation can achieve competitive performance, especially on long-range forecasting tasks. Finally, though the current work is inspired by continuous-depth neural networks, the particle-ensemble shooting formulation also applies to discrete-time networks and may lead to a new fertile area of research in deep learning parametrization. \ No newline at end of file diff --git a/data/2020/neurips/A simple normative network approximates local non-Hebbian learning in the cortex b/data/2020/neurips/A simple normative network approximates local non-Hebbian learning in the cortex new file mode 100644 index 0000000000..537dd2fbed --- /dev/null +++ b/data/2020/neurips/A simple normative network approximates local non-Hebbian learning in the cortex @@ -0,0 +1 @@ +To guide behavior, the brain extracts relevant features from high-dimensional data streamed by sensory organs. Neuroscience experiments demonstrate that the processing of sensory inputs by cortical neurons is modulated by instructive signals which provide context and task-relevant information. Here, adopting a normative approach, we model these instructive signals as supervisory inputs guiding the projection of the feedforward data. Mathematically, we start with a family of Reduced-Rank Regression (RRR) objective functions which include Reduced Rank (minimum) Mean Square Error (RRMSE) and Canonical Correlation Analysis (CCA), and derive novel offline and online optimization algorithms, which we call Bio-RRR. The online algorithms can be implemented by neural networks whose synaptic learning rules resemble calcium plateau potential dependent plasticity observed in the cortex. We detail how, in our model, the calcium plateau potential can be interpreted as a backpropagating error signal. We demonstrate that, despite relying exclusively on biologically plausible local learning rules, our algorithms perform competitively with existing implementations of RRMSE and CCA. \ No newline at end of file diff --git a/data/2020/neurips/AI Feynman 2.0: Pareto-optimal symbolic regression exploiting graph modularity b/data/2020/neurips/AI Feynman 2.0: Pareto-optimal symbolic regression exploiting graph modularity new file mode 100644 index 0000000000..5cb148b6f8 --- /dev/null +++ b/data/2020/neurips/AI Feynman 2.0: Pareto-optimal symbolic regression exploiting graph modularity @@ -0,0 +1 @@ +We present an improved method for symbolic regression that seeks to fit data to formulas that are Pareto-optimal, in the sense of having the best accuracy for a given complexity. It improves on the previous state-of-the-art by typically being orders of magnitude more robust toward noise and bad data, and also by discovering many formulas that stumped previous methods. We develop a method for discovering generalized symmetries (arbitrary modularity in the computational graph of a formula) from gradient properties of a neural network fit. We use normalizing flows to generalize our symbolic regression method to probability distributions from which we only have samples, and employ statistical hypothesis testing to accelerate robust brute-force search. \ No newline at end of file diff --git a/data/2020/neurips/AOT: Appearance Optimal Transport Based Identity Swapping for Forgery Detection b/data/2020/neurips/AOT: Appearance Optimal Transport Based Identity Swapping for Forgery Detection new file mode 100644 index 0000000000..41cec6164d --- /dev/null +++ b/data/2020/neurips/AOT: Appearance Optimal Transport Based Identity Swapping for Forgery Detection @@ -0,0 +1 @@ +Recent studies have shown that the performance of forgery detection can be improved with diverse and challenging Deepfakes datasets. However, due to the lack of Deepfakes datasets with large variance in appearance, which can be hardly produced by recent identity swapping methods, the detection algorithm may fail in this situation. In this work, we provide a new identity swapping algorithm with large differences in appearance for face forgery detection. The appearance gaps mainly arise from the large discrepancies in illuminations and skin colors that widely exist in real-world scenarios. However, due to the difficulties of modeling the complex appearance mapping, it is challenging to transfer fine-grained appearances adaptively while preserving identity traits. This paper formulates appearance mapping as an optimal transport problem and proposes an Appearance Optimal Transport model (AOT) to formulate it in both latent and pixel space. Specifically, a relighting generator is designed to simulate the optimal transport plan. It is solved via minimizing the Wasserstein distance of the learned features in the latent space, enabling better performance and less computation than conventional optimization. To further refine the solution of the optimal transport plan, we develop a segmentation game to minimize the Wasserstein distance in the pixel space. A discriminator is introduced to distinguish the fake parts from a mix of real and fake image patches. Extensive experiments reveal that the superiority of our method when compared with state-of-the-art methods and the ability of our generated data to improve the performance of face forgery detection. \ No newline at end of file diff --git a/data/2020/neurips/ARMA Nets: Expanding Receptive Field for Dense Prediction b/data/2020/neurips/ARMA Nets: Expanding Receptive Field for Dense Prediction new file mode 100644 index 0000000000..53ad7b8f1f --- /dev/null +++ b/data/2020/neurips/ARMA Nets: Expanding Receptive Field for Dense Prediction @@ -0,0 +1 @@ +Global information is essential for dense prediction problems, whose goal is to compute a discrete or continuous label for each pixel in the images. Traditional convolutional layers in neural networks, initially designed for image classification, are restrictive in these problems since the filter size limits their receptive fields. In this work, we propose to replace any traditional convolutional layer with an autoregressive moving-average (ARMA) layer, a novel module with an adjustable receptive field controlled by the learnable autoregressive coefficients. Compared with traditional convolutional layers, our ARMA layer enables explicit interconnections of the output neurons and learns its receptive field by adapting the autoregressive coefficients of the interconnections. ARMA layer is adjustable to different types of tasks: for tasks where global information is crucial, it is capable of learning relatively large autoregressive coefficients to allow for an output neuron's receptive field covering the entire input; for tasks where only local information is required, it can learn small or near zero autoregressive coefficients and automatically reduces to a traditional convolutional layer. We show both theoretically and empirically that the effective receptive field of networks with ARMA layers (named as ARMA networks) expands with larger autoregressive coefficients. We also provably solve the instability problem of learning and prediction in the ARMA layer through a re-parameterization mechanism. Additionally, we demonstrate that ARMA networks substantially improve their baselines on challenging dense prediction tasks including video prediction and semantic segmentation. \ No newline at end of file diff --git a/data/2020/neurips/AViD Dataset: Anonymized Videos from Diverse Countries b/data/2020/neurips/AViD Dataset: Anonymized Videos from Diverse Countries new file mode 100644 index 0000000000..6bf8094d65 --- /dev/null +++ b/data/2020/neurips/AViD Dataset: Anonymized Videos from Diverse Countries @@ -0,0 +1 @@ +We introduce a new public video dataset for action recognition: Anonymized Videos from Diverse countries (AViD). Unlike existing public video datasets, AViD is a collection of action videos from many different countries. The motivation is to create a public dataset that would benefit training and pretraining of action recognition models for everybody, rather than making it useful for limited countries. Further, all the face identities in the AViD videos are properly anonymized to protect their privacy. It also is a static dataset where each video is licensed with the creative commons license. We confirm that most of the existing video datasets are statistically biased to only capture action videos from a limited number of countries. We experimentally illustrate that models trained with such biased datasets do not transfer perfectly to action videos from the other countries, and show that AViD addresses such problem. We also confirm that the new AViD dataset could serve as a good dataset for pretraining the models, performing comparably or better than prior datasets. \ No newline at end of file diff --git a/data/2020/neurips/Accelerating Training of Transformer-Based Language Models with Progressive Layer Dropping b/data/2020/neurips/Accelerating Training of Transformer-Based Language Models with Progressive Layer Dropping new file mode 100644 index 0000000000..cb2a85a95b --- /dev/null +++ b/data/2020/neurips/Accelerating Training of Transformer-Based Language Models with Progressive Layer Dropping @@ -0,0 +1 @@ +Recently, Transformer-based language models have demonstrated remarkable performance across many NLP domains. However, the unsupervised pre-training step of these models suffers from unbearable overall computational expenses. Current methods for accelerating the pre-training either rely on massive parallelism with advanced hardware or are not applicable to language modeling. In this work, we propose a method based on progressive layer dropping that speeds the training of Transformer-based language models, not at the cost of excessive hardware resources but from model architecture change and training technique boosted efficiency. Extensive experiments on BERT show that the proposed method achieves a 24% time reduction on average per sample and allows the pre-training to be 2.5 times faster than the baseline to get a similar accuracy on downstream tasks. While being faster, our pre-trained models are equipped with strong knowledge transferability, achieving comparable and sometimes higher GLUE score than the baseline when pre-trained with the same number of samples. \ No newline at end of file diff --git a/data/2020/neurips/Acceleration with a Ball Optimization Oracle b/data/2020/neurips/Acceleration with a Ball Optimization Oracle new file mode 100644 index 0000000000..9b0fdcc422 --- /dev/null +++ b/data/2020/neurips/Acceleration with a Ball Optimization Oracle @@ -0,0 +1 @@ +Consider an oracle which takes a point $x$ and returns the minimizer of a convex function $f$ in an $\ell_2$ ball of radius $r$ around $x$. It is straightforward to show that roughly $r^{-1}\log\frac{1}{\epsilon}$ calls to the oracle suffice to find an $\epsilon$-approximate minimizer of $f$ in an $\ell_2$ unit ball. Perhaps surprisingly, this is not optimal: we design an accelerated algorithm which attains an $\epsilon$-approximate minimizer with roughly $r^{-2/3} \log \frac{1}{\epsilon}$ oracle queries, and give a matching lower bound. Further, we implement ball optimization oracles for functions with locally stable Hessians using a variant of Newton's method. The resulting algorithm applies to a number of problems of practical and theoretical import, improving upon previous results for logistic and $\ell_\infty$ regression and achieving guarantees comparable to the state-of-the-art for $\ell_p$ regression. \ No newline at end of file diff --git a/data/2020/neurips/Achieving Equalized Odds by Resampling Sensitive Attributes b/data/2020/neurips/Achieving Equalized Odds by Resampling Sensitive Attributes new file mode 100644 index 0000000000..697f788c95 --- /dev/null +++ b/data/2020/neurips/Achieving Equalized Odds by Resampling Sensitive Attributes @@ -0,0 +1 @@ +We present a flexible framework for learning predictive models that approximately satisfy the equalized odds notion of fairness. This is achieved by introducing a general discrepancy functional that rigorously quantifies violations of this criterion. This differentiable functional is used as a penalty driving the model parameters towards equalized odds. To rigorously evaluate fitted models, we develop a formal hypothesis test to detect whether a prediction rule violates this property, the first such test in the literature. Both the model fitting and hypothesis testing leverage a resampled version of the sensitive attribute obeying equalized odds, by construction. We demonstrate the applicability and validity of the proposed framework both in regression and multi-class classification problems, reporting improved performance over state-of-the-art methods. Lastly, we show how to incorporate techniques for equitable uncertainty quantification---unbiased for each group under study---to communicate the results of the data analysis in exact terms. \ No newline at end of file diff --git a/data/2020/neurips/Active Invariant Causal Prediction: Experiment Selection through Stability b/data/2020/neurips/Active Invariant Causal Prediction: Experiment Selection through Stability new file mode 100644 index 0000000000..c8ae87cc3e --- /dev/null +++ b/data/2020/neurips/Active Invariant Causal Prediction: Experiment Selection through Stability @@ -0,0 +1 @@ +A fundamental difficulty of causal learning is that causal models can generally not be fully identified based on observational data only. Interventional data, that is, data originating from different experimental environments, improves identifiability. However, the improvement depends critically on the target and nature of the interventions carried out in each experiment. Since in real applications experiments tend to be costly, there is a need to perform the right interventions such that as few as possible are required. In this work we propose a new active learning (i.e. experiment selection) framework (A-ICP) based on Invariant Causal Prediction (ICP) (Peters et al., 2016). For general structural causal models, we characterize the effect of interventions on so-called stable sets, a notion introduced by (Pfister et al., 2019). We leverage these results to propose several intervention selection policies for A-ICP which quickly reveal the direct causes of a response variable in the causal graph while maintaining the error control inherent in ICP. Empirically, we analyze the performance of the proposed policies in both population and finite-regime experiments. \ No newline at end of file diff --git a/data/2020/neurips/Active Structure Learning of Causal DAGs via Directed Clique Trees b/data/2020/neurips/Active Structure Learning of Causal DAGs via Directed Clique Trees new file mode 100644 index 0000000000..ac32a6a858 --- /dev/null +++ b/data/2020/neurips/Active Structure Learning of Causal DAGs via Directed Clique Trees @@ -0,0 +1 @@ +A growing body of work has begun to study intervention design for efficient structure learning of causal directed acyclic graphs (DAGs). A typical setting is a causally sufficient setting, i.e. a system with no latent confounders, selection bias, or feedback, when the essential graph of the observational equivalence class (EC) is given as an input and interventions are assumed to be noiseless. Most existing works focus on worst-case or average-case lower bounds for the number of interventions required to orient a DAG. These worst-case lower bounds only establish that the largest clique in the essential graph could make it difficult to learn the true DAG. In this work, we develop a universal lower bound for single-node interventions that establishes that the largest clique is always a fundamental impediment to structure learning. Specifically, we present a decomposition of a DAG into independently orientable components through directed clique trees and use it to prove that the number of single-node interventions necessary to orient any DAG in an EC is at least the sum of half the size of the largest cliques in each chain component of the essential graph. Moreover, we present a two-phase intervention design algorithm that, under certain conditions on the chordal skeleton, matches the optimal number of interventions up to a multiplicative logarithmic factor in the number of maximal cliques. We show via synthetic experiments that our algorithm can scale to much larger graphs than most of the related work and achieves better worst-case performance than other scalable approaches. A code base to recreate these results can be found at this https URL \ No newline at end of file diff --git a/data/2020/neurips/AdaBelief Optimizer: Adapting Stepsizes by the Belief in Observed Gradients b/data/2020/neurips/AdaBelief Optimizer: Adapting Stepsizes by the Belief in Observed Gradients new file mode 100644 index 0000000000..d55db36846 --- /dev/null +++ b/data/2020/neurips/AdaBelief Optimizer: Adapting Stepsizes by the Belief in Observed Gradients @@ -0,0 +1 @@ +Most popular optimizers for deep learning can be broadly categorized as adaptive methods (e.g. Adam) and accelerated schemes (e.g. stochastic gradient descent (SGD) with momentum). For many models such as convolutional neural networks (CNNs), adaptive methods typically converge faster but generalize worse compared to SGD; for complex settings such as generative adversarial networks (GANs), adaptive methods are typically the default because of their stability.We propose AdaBelief to simultaneously achieve three goals: fast convergence as in adaptive methods, good generalization as in SGD, and training stability. The intuition for AdaBelief is to adapt the stepsize according to the "belief" in the current gradient direction. Viewing the exponential moving average (EMA) of the noisy gradient as the prediction of the gradient at the next time step, if the observed gradient greatly deviates from the prediction, we distrust the current observation and take a small step; if the observed gradient is close to the prediction, we trust it and take a large step. We validate AdaBelief in extensive experiments, showing that it outperforms other methods with fast convergence and high accuracy on image classification and language modeling. Specifically, on ImageNet, AdaBelief achieves comparable accuracy to SGD. Furthermore, in the training of a GAN on Cifar10, AdaBelief demonstrates high stability and improves the quality of generated samples compared to a well-tuned Adam optimizer. Code is available at this https URL \ No newline at end of file diff --git a/data/2020/neurips/AdaShare: Learning What To Share For Efficient Deep Multi-Task Learning b/data/2020/neurips/AdaShare: Learning What To Share For Efficient Deep Multi-Task Learning new file mode 100644 index 0000000000..8efc4abda4 --- /dev/null +++ b/data/2020/neurips/AdaShare: Learning What To Share For Efficient Deep Multi-Task Learning @@ -0,0 +1 @@ +Multi-task learning is an open and challenging problem in computer vision. The typical way of conducting multi-task learning with deep neural networks is either through handcrafting schemes that share all initial layers and branch out at an adhoc point or through using separate task-specific networks with an additional feature sharing/fusion mechanism. Unlike existing methods, we propose an adaptive sharing approach, called AdaShare, that decides what to share across which tasks for achieving the best recognition accuracy, while taking resource efficiency into account. Specifically, our main idea is to learn the sharing pattern through a task-specific policy that selectively chooses which layers to execute for a given task in the multi-task network. We efficiently optimize the task-specific policy jointly with the network weights using standard back-propagation. Experiments on three challenging and diverse benchmark datasets with a variable number of tasks well demonstrate the efficacy of our approach over state-of-the-art methods. \ No newline at end of file diff --git a/data/2020/neurips/AdaTune: Adaptive Tensor Program Compilation Made Efficient b/data/2020/neurips/AdaTune: Adaptive Tensor Program Compilation Made Efficient new file mode 100644 index 0000000000..f1834d269a --- /dev/null +++ b/data/2020/neurips/AdaTune: Adaptive Tensor Program Compilation Made Efficient @@ -0,0 +1 @@ +Deep learning models are computationally intense, and implementations often have to be highly optimized by experts or hardware vendors to be usable in practice. The DL compiler, together with Learning-to-Compile has proven to be a powerful technique for optimizing tensor programs. However, a limitation of this approach is that it still suffers from unbearably long overall optimization time. In this paper, we present a new method, called AdaTune, that significantly reduces the optimization time of tensor programs for high-performance deep learning inference. In particular, we propose an adaptive evaluation method that statistically early terminates a costly hardware measurement without losing much accuracy. We further devise a surrogate model with uncertainty quantification that allows the optimization to adapt to hardware and model heterogeneity better. Finally, we introduce a contextual optimizer that provides adaptive control of the exploration and exploitation to improve the transformation space searching effectiveness. We evaluate and compare the levels of optimization obtained by AutoTVM, a state-of-the-art Learning-to-Compile technique on top of TVM, and AdaTune. The experiment results show that AdaTune obtains up to 115% higher GFLOPS than the baseline under the same optimization time budget. Furthermore, AdaTune provides 1.3–3.9\(\times\) speedup in optimization time over the baseline to reach the same optimization quality for a range of models across different hardware architectures. \ No newline at end of file diff --git a/data/2020/neurips/Adam with Bandit Sampling for Deep Learning b/data/2020/neurips/Adam with Bandit Sampling for Deep Learning new file mode 100644 index 0000000000..73eebc9fde --- /dev/null +++ b/data/2020/neurips/Adam with Bandit Sampling for Deep Learning @@ -0,0 +1 @@ +Adam is a widely used optimization method for training deep learning models. It computes individual adaptive learning rates for different parameters. In this paper, we propose a generalization of Adam, called Adambs, that allows us to also adapt to different training examples based on their importance in the model's convergence. To achieve this, we maintain a distribution over all examples, selecting a mini-batch in each iteration by sampling according to this distribution, which we update using a multi-armed bandit algorithm. This ensures that examples that are more beneficial to the model training are sampled with higher probabilities. We theoretically show that Adambs improves the convergence rate of Adam---$O(\sqrt{\frac{\log n}{T} })$ instead of $O(\sqrt{\frac{n}{T}})$ in some cases. Experiments on various models and datasets demonstrate Adambs's fast convergence in practice. \ No newline at end of file diff --git a/data/2020/neurips/Adaptation Properties Allow Identification of Optimized Neural Codes b/data/2020/neurips/Adaptation Properties Allow Identification of Optimized Neural Codes new file mode 100644 index 0000000000..7864aeedfb --- /dev/null +++ b/data/2020/neurips/Adaptation Properties Allow Identification of Optimized Neural Codes @@ -0,0 +1 @@ +The adaptation of neural codes to the statistics of their environment is well captured by efficient coding approaches. Here we solve an inverse problem: characterizing the objective and constraint functions that efficient codes appear to be optimal for, on the basis of how they adapt to different stimulus distributions. We formulate a general efficient coding problem, with flexible objective and constraint functions and minimal parametric assumptions. Solving special cases of this model, we provide solutions to broad classes of Fisher information-based efficient coding problems, generalizing a wide range of previous results. We show that different objective function types impose qualitatively different adaptation behaviors, while constraints enforce characteristic deviations from classic efficient coding signatures. Despite interaction between these effects, clear signatures emerge for both unconstrained optimization problems and information-maximizing objective functions. Asking for a fixed-point of the neural code adaptation, we find an objective-independent characterization of constraints on the neural code. We use this result to propose an experimental paradigm that can characterize both the objective and constraint functions that an observed code appears to be optimized for. \ No newline at end of file diff --git a/data/2020/neurips/Adapting Neural Architectures Between Domains b/data/2020/neurips/Adapting Neural Architectures Between Domains new file mode 100644 index 0000000000..de18ac494c --- /dev/null +++ b/data/2020/neurips/Adapting Neural Architectures Between Domains @@ -0,0 +1 @@ +Neural architecture search (NAS) has demonstrated impressive performance in automatically designing high-performance neural networks. The power of deep neural networks is to be unleashed for analyzing a large volume of data (e.g. ImageNet), but the architecture search is often executed on another smaller dataset (e.g. CIFAR-10) to finish it in a feasible time. However, it is hard to guarantee that the optimal architecture derived on the proxy task could maintain its advantages on another more challenging dataset. This paper aims to improve the generalization of neural architectures via domain adaptation. We analyze the generalization bounds of the derived architecture and suggest its close relations with the validation error and the data distribution distance on both domains. These theoretical analyses lead to AdaptNAS, a novel and principled approach to adapt neural architectures between domains in NAS. Our experimental evaluation shows that only a small part of ImageNet will be sufficient for AdaptNAS to extend its architecture success to the entire ImageNet and outperform state-of-the-art comparison algorithms. \ No newline at end of file diff --git a/data/2020/neurips/Adapting to Misspecification in Contextual Bandits b/data/2020/neurips/Adapting to Misspecification in Contextual Bandits new file mode 100644 index 0000000000..972132afef --- /dev/null +++ b/data/2020/neurips/Adapting to Misspecification in Contextual Bandits @@ -0,0 +1 @@ +A major research direction in contextual bandits is to develop algorithms that are computationally efficient, yet support flexible, general-purpose function approximation. Algorithms based on modeling rewards have shown strong empirical performance, but typically require a well-specified model, and can fail when this assumption does not hold. Can we design algorithms that are efficient and flexible, yet degrade gracefully in the face of model misspecification? We introduce a new family of oracle-efficient algorithms for $\varepsilon$-misspecified contextual bandits that adapt to unknown model misspecification -- both for finite and infinite action settings. Given access to an online oracle for square loss regression, our algorithm attains optimal regret and -- in particular -- optimal dependence on the misspecification level, with no prior knowledge. Specializing to linear contextual bandits with infinite actions in $d$ dimensions, we obtain the first algorithm that achieves the optimal $O(d\sqrt{T} + \varepsilon\sqrt{d}T)$ regret bound for unknown misspecification level $\varepsilon$. On a conceptual level, our results are enabled by a new optimization-based perspective on the regression oracle reduction framework of Foster and Rakhlin, which we anticipate will find broader use. \ No newline at end of file diff --git a/data/2020/neurips/Adaptive Discretization for Model-Based Reinforcement Learning b/data/2020/neurips/Adaptive Discretization for Model-Based Reinforcement Learning new file mode 100644 index 0000000000..f42c9ed758 --- /dev/null +++ b/data/2020/neurips/Adaptive Discretization for Model-Based Reinforcement Learning @@ -0,0 +1,2 @@ +We introduce the technique of adaptive discretization to design efficient model-based episodic reinforcement learning algorithms in large (potentially continuous) state-action spaces. Our algorithm is based on optimistic one-step value iteration extended to maintain an adaptive discretization of the space. From a theoretical perspective, we provide worst-case regret bounds for our algorithm, which are competitive compared to the state-of-the-art model-based algorithms; moreover, our bounds are obtained via a modular proof technique, which can potentially extend to incorporate additional structure on the problem. +From an implementation standpoint, our algorithm has much lower storage and computational requirements, due to maintaining a more efficient partition of the state and action spaces. We illustrate this via experiments on several canonical control problems, which shows that our algorithm empirically performs significantly better than fixed discretization in terms of both faster convergence and lower memory usage. Interestingly, we observe empirically that while fixed-discretization model-based algorithms vastly outperform their model-free counterparts, the two achieve comparable performance with adaptive discretization. \ No newline at end of file diff --git a/data/2020/neurips/Adaptive Experimental Design with Temporal Interference: A Maximum Likelihood Approach b/data/2020/neurips/Adaptive Experimental Design with Temporal Interference: A Maximum Likelihood Approach new file mode 100644 index 0000000000..d1256b4522 --- /dev/null +++ b/data/2020/neurips/Adaptive Experimental Design with Temporal Interference: A Maximum Likelihood Approach @@ -0,0 +1,2 @@ +Suppose an online platform wants to compare a treatment and control policy, e.g., two different matching algorithms in a ridesharing system, or two different inventory management algorithms in an online retail site. Standard randomized controlled trials are typically not feasible, since the goal is to estimate policy performance on the entire system. Instead, the typical current practice involves dynamically alternating between the two policies for fixed lengths of time, and comparing the average performance of each over the intervals in which they were run as an estimate of the treatment effect. However, this approach suffers from *temporal interference*: one algorithm alters the state of the system as seen by the second algorithm, biasing estimates of the treatment effect. Further, the simple non-adaptive nature of such designs implies they are not sample efficient. +We develop a benchmark theoretical model in which to study optimal experimental design for this setting. We view testing the two policies as the problem of estimating the steady state difference in reward between two unknown Markov chains (i.e., policies). We assume estimation of the steady state reward for each chain proceeds via nonparametric maximum likelihood, and search for consistent (i.e., asymptotically unbiased) experimental designs that are efficient (i.e., asymptotically minimum variance). Characterizing such designs is equivalent to a Markov decision problem with a minimum variance objective; such problems generally do not admit tractable solutions. Remarkably, in our setting, using a novel application of classical martingale analysis of Markov chains via Poisson's equation, we characterize efficient designs via a succinct convex optimization problem. We use this characterization to propose a consistent, efficient online experimental design that adaptively samples the two Markov chains. \ No newline at end of file diff --git a/data/2020/neurips/Adaptive Gradient Quantization for Data-Parallel SGD b/data/2020/neurips/Adaptive Gradient Quantization for Data-Parallel SGD new file mode 100644 index 0000000000..8b362e07f2 --- /dev/null +++ b/data/2020/neurips/Adaptive Gradient Quantization for Data-Parallel SGD @@ -0,0 +1 @@ +Many communication-efficient variants of SGD use gradient quantization schemes. These schemes are often heuristic and fixed over the course of training. We empirically observe that the statistics of gradients of deep models change during the training. Motivated by this observation, we introduce two adaptive quantization schemes, ALQ and AMQ. In both schemes, processors update their compression schemes in parallel by efficiently computing sufficient statistics of a parametric distribution. We improve the validation accuracy by almost 2% on CIFAR-10 and 1% on ImageNet in challenging low-cost communication setups. Our adaptive methods are also significantly more robust to the choice of hyperparameters. \ No newline at end of file diff --git a/data/2020/neurips/Adaptive Graph Convolutional Recurrent Network for Traffic Forecasting b/data/2020/neurips/Adaptive Graph Convolutional Recurrent Network for Traffic Forecasting new file mode 100644 index 0000000000..3f49227635 --- /dev/null +++ b/data/2020/neurips/Adaptive Graph Convolutional Recurrent Network for Traffic Forecasting @@ -0,0 +1 @@ +Modeling complex spatial and temporal correlations in the correlated time series data is indispensable for understanding the traffic dynamics and predicting the future status of an evolving traffic system. Recent works focus on designing complicated graph neural network architectures to capture shared patterns with the help of pre-defined graphs. In this paper, we argue that learning node-specific patterns is essential for traffic forecasting while the pre-defined graph is avoidable. To this end, we propose two adaptive modules for enhancing Graph Convolutional Network (GCN) with new capabilities: 1) a Node Adaptive Parameter Learning (NAPL) module to capture node-specific patterns; 2) a Data Adaptive Graph Generation (DAGG) module to infer the inter-dependencies among different traffic series automatically. We further propose an Adaptive Graph Convolutional Recurrent Network (AGCRN) to capture fine-grained spatial and temporal correlations in traffic series automatically based on the two modules and recurrent networks. Our experiments on two real-world traffic datasets show AGCRN outperforms state-of-the-art by a significant margin without pre-defined graphs about spatial connections. \ No newline at end of file diff --git a/data/2020/neurips/Adaptive Importance Sampling for Finite-Sum Optimization and Sampling with Decreasing Step-Sizes b/data/2020/neurips/Adaptive Importance Sampling for Finite-Sum Optimization and Sampling with Decreasing Step-Sizes new file mode 100644 index 0000000000..6b64c78cfa --- /dev/null +++ b/data/2020/neurips/Adaptive Importance Sampling for Finite-Sum Optimization and Sampling with Decreasing Step-Sizes @@ -0,0 +1 @@ +Reducing the variance of the gradient estimator is known to improve the convergence rate of stochastic gradient-based optimization and sampling algorithms. One way of achieving variance reduction is to design importance sampling strategies. Recently, the problem of designing such schemes was formulated as an online learning problem with bandit feedback, and algorithms with sub-linear static regret were designed. In this work, we build on this framework and propose Avare, a simple and efficient algorithm for adaptive importance sampling for finite-sum optimization and sampling with decreasing step-sizes. Under standard technical conditions, we show that Avare achieves $\mathcal{O}(T^{2/3})$ and $\mathcal{O}(T^{5/6})$ dynamic regret for SGD and SGLD respectively when run with $\mathcal{O}(1/t)$ step sizes. We achieve this dynamic regret bound by leveraging our knowledge of the dynamics defined by the algorithm, and combining ideas from online learning and variance-reduced stochastic optimization. We validate empirically the performance of our algorithm and identify settings in which it leads to significant improvements. \ No newline at end of file diff --git a/data/2020/neurips/Adaptive Learning of Rank-One Models for Efficient Pairwise Sequence Alignment b/data/2020/neurips/Adaptive Learning of Rank-One Models for Efficient Pairwise Sequence Alignment new file mode 100644 index 0000000000..1facd6a333 --- /dev/null +++ b/data/2020/neurips/Adaptive Learning of Rank-One Models for Efficient Pairwise Sequence Alignment @@ -0,0 +1 @@ +Pairwise alignment of DNA sequencing data is a ubiquitous task in bioinformatics and typically represents a heavy computational burden. State-of-the-art approaches to speed up this task use hashing to identify short segments (k-mers) that are shared by pairs of reads, which can then be used to estimate alignment scores. However, when the number of reads is large, accurately estimating alignment scores for all pairs is still very costly. Moreover, in practice, one is only interested in identifying pairs of reads with large alignment scores. In this work, we propose a new approach to pairwise alignment estimation based on two key new ingredients. The first ingredient is to cast the problem of pairwise alignment estimation under a general framework of rank-one crowdsourcing models, where the workers' responses correspond to k-mer hash collisions. These models can be accurately solved via a spectral decomposition of the response matrix. The second ingredient is to utilise a multi-armed bandit algorithm to adaptively refine this spectral estimator only for read pairs that are likely to have large alignments. The resulting algorithm iteratively performs a spectral decomposition of the response matrix for adaptively chosen subsets of the read pairs. \ No newline at end of file diff --git a/data/2020/neurips/Adaptive Online Estimation of Piecewise Polynomial Trends b/data/2020/neurips/Adaptive Online Estimation of Piecewise Polynomial Trends new file mode 100644 index 0000000000..0532bd06e0 --- /dev/null +++ b/data/2020/neurips/Adaptive Online Estimation of Piecewise Polynomial Trends @@ -0,0 +1 @@ +We consider the framework of non-stationary stochastic optimization [Besbes et al, 2015] with squared error losses and noisy gradient feedback where the dynamic regret of an online learner against a time varying comparator sequence is studied. Motivated from the theory of non-parametric regression, we introduce a new variational constraint that enforces the comparator sequence to belong to a discrete $k^{th}$ order Total Variation ball of radius $C_n$. This variational constraint models comparators that have piece-wise polynomial structure which has many relevant practical applications [Tibshirani, 2014]. By establishing connections to the theory of wavelet based non-parametric regression, we design a polynomial time algorithm that achieves the nearly optimal dynamic regret of $\tilde{O}(n^{\frac{1}{2k+3}}C_n^{\frac{2}{2k+3}})$. The proposed policy is adaptive to the unknown radius $C_n$. Further, we show that the same policy is minimax optimal for several other non-parametric families of interest. \ No newline at end of file diff --git a/data/2020/neurips/Adaptive Probing Policies for Shortest Path Routing b/data/2020/neurips/Adaptive Probing Policies for Shortest Path Routing new file mode 100644 index 0000000000..6d2b351751 --- /dev/null +++ b/data/2020/neurips/Adaptive Probing Policies for Shortest Path Routing @@ -0,0 +1 @@ +Inspired by traffic routing applications, we consider the problem of finding the shortest path from a source s to a destination t in a graph, when the lengths of the edges are unknown. Instead, we are given hints or predictions of the edge lengths from a collection of ML models, trained possibly on historical data and other contexts in the network. Additionally, we assume that the true length of any candidate path can be obtained by probing an up-to-date snapshot of the network. However, each probe introduces a latency, and thus the goal is to minimize the number of probes while finding a near-optimal path with high probability. We formalize this problem and show assumptions under which it admits to efficient approximation algorithms. We verify these assumptions and validate the performance of our algorithms on real data. \ No newline at end of file diff --git a/data/2020/neurips/Adaptive Reduced Rank Regression b/data/2020/neurips/Adaptive Reduced Rank Regression new file mode 100644 index 0000000000..eecaf70cdc --- /dev/null +++ b/data/2020/neurips/Adaptive Reduced Rank Regression @@ -0,0 +1 @@ +Low rank regression has proven to be useful in a wide range of forecasting problems. However, in settings with a low signal-to-noise ratio, it is known to suffer from severe overfitting. This paper studies the reduced rank regression problem and presents algorithms with provable generalization guarantees. We use adaptive hard rank-thresholding in two different parts of the data analysis pipeline. First, we consider a low rank projection of the data to eliminate the components that are most likely to be noisy. Second, we perform a standard multivariate linear regression estimator on the data obtained in the first step, and subsequently consider a low-rank projection of the obtained regression matrix. Both thresholding is performed in a data-driven manner and is required to prevent severe overfitting as our lower bounds show. Experimental results show that our approach either outperforms or is competitive with existing baselines. \ No newline at end of file diff --git a/data/2020/neurips/Adaptive Sampling for Stochastic Risk-Averse Learning b/data/2020/neurips/Adaptive Sampling for Stochastic Risk-Averse Learning new file mode 100644 index 0000000000..9a6bb3ff85 --- /dev/null +++ b/data/2020/neurips/Adaptive Sampling for Stochastic Risk-Averse Learning @@ -0,0 +1 @@ +We consider the problem of training machine learning models in a risk-averse manner. In particular, we propose an adaptive sampling algorithm for stochastically optimizing the Conditional Value-at-Risk (CVaR) of a loss distribution. We use a distributionally robust formulation of the CVaR to phrase the problem as a zero-sum game between two players. Our approach solves the game using an efficient no-regret algorithm for each player. Critically, we can apply these algorithms to large-scale settings because the implementation relies on sampling from Determinantal Point Processes. Finally, we empirically demonstrate its effectiveness on large-scale convex and non-convex learning tasks. \ No newline at end of file diff --git a/data/2020/neurips/Adaptive Shrinkage Estimation for Streaming Graphs b/data/2020/neurips/Adaptive Shrinkage Estimation for Streaming Graphs new file mode 100644 index 0000000000..3ba0376dde --- /dev/null +++ b/data/2020/neurips/Adaptive Shrinkage Estimation for Streaming Graphs @@ -0,0 +1 @@ +Networks are a natural representation of complex systems across the sciences, and higher-order dependencies are central to the understanding and modeling of these systems. However, in many practical applications such as online social networks, networks are massive, dynamic, and naturally streaming, where pairwise interactions among vertices become available one at a time in some arbitrary order. The massive size and streaming nature of these networks allow only partial observation, since it is infeasible to analyze the entire network. Under such scenarios, it is challenging to study the higher-order structural and connectivity patterns of streaming networks. In this work, we consider the fundamental problem of estimating the higher-order dependencies using adaptive sampling. We propose a novel adaptive, single-pass sampling framework and unbiased estimators for higher-order network analysis of large streaming networks. Our algorithms exploit adaptive techniques to identify edges that are highly informative for efficiently estimating the higher-order structure of streaming networks from small sample data. We also introduce a novel James-Stein shrinkage estimator to reduce the estimation error. Our approach is fully analytic, computationally efficient, and can be incrementally updated in a streaming setting. Numerical experiments on large networks show that our approach is superior to baseline methods. \ No newline at end of file diff --git a/data/2020/neurips/AdvFlow: Inconspicuous Black-box Adversarial Attacks using Normalizing Flows b/data/2020/neurips/AdvFlow: Inconspicuous Black-box Adversarial Attacks using Normalizing Flows new file mode 100644 index 0000000000..70b3953e38 --- /dev/null +++ b/data/2020/neurips/AdvFlow: Inconspicuous Black-box Adversarial Attacks using Normalizing Flows @@ -0,0 +1 @@ +Deep learning classifiers are susceptible to well-crafted, imperceptible variations of their inputs, known as adversarial attacks. In this regard, the study of powerful attack models sheds light on the sources of vulnerability in these classifiers, hopefully leading to more robust ones. In this paper, we introduce AdvFlow: a novel black-box adversarial attack method on image classifiers that exploits the power of normalizing flows to model the density of adversarial examples around a given target image. We see that the proposed method generates adversaries that closely follow the clean data distribution, a property which makes their detection less likely. Also, our experimental results show competitive performance of the proposed approach with some of the existing attack methods on defended classifiers, outperforming them in both the number of queries and attack success rate. The code is available at this https URL. \ No newline at end of file diff --git a/data/2020/neurips/Advances in Black-Box VI: Normalizing Flows, Importance Weighting, and Optimization b/data/2020/neurips/Advances in Black-Box VI: Normalizing Flows, Importance Weighting, and Optimization new file mode 100644 index 0000000000..a44be84191 --- /dev/null +++ b/data/2020/neurips/Advances in Black-Box VI: Normalizing Flows, Importance Weighting, and Optimization @@ -0,0 +1 @@ +Recent research has seen several advances relevant to black-box VI, but the current state of automatic posterior inference is unclear. One such advance is the use of normalizing flows to define flexible posterior densities for deep latent variable models. Another direction is the integration of Monte-Carlo methods to serve two purposes; first, to obtain tighter variational objectives for optimization, and second, to define enriched variational families through sampling. However, both flows and variational Monte-Carlo methods remain relatively unexplored for black-box VI. Moreover, on a pragmatic front, there are several optimization considerations like step-size scheme, parameter initialization, and choice of gradient estimators, for which there are no clear guidance in the existing literature. In this paper, we postulate that black-box VI is best addressed through a careful combination of numerous algorithmic components. We evaluate components relating to optimization, flows, and Monte-Carlo methods on a benchmark of 30 models from the Stan model library. The combination of these algorithmic components significantly advances the state-of-the-art "out of the box" variational inference. \ No newline at end of file diff --git a/data/2020/neurips/Adversarial Attacks on Deep Graph Matching b/data/2020/neurips/Adversarial Attacks on Deep Graph Matching new file mode 100644 index 0000000000..3d9adc8e30 --- /dev/null +++ b/data/2020/neurips/Adversarial Attacks on Deep Graph Matching @@ -0,0 +1 @@ +Despite achieving remarkable performance, deep graph learning models, such as node classification and network embedding, suffer from harassment caused by small adversarial perturbations. However, the vulnerability analysis of graph matching under adversarial attacks has not been fully investigated yet. This paper proposes an adversarial attack model with two novel attack techniques to perturb the graph structure and degrade the quality of deep graph matching: (1) a kernel density estimation approach is utilized to estimate and maximize node densities to derive imperceptible perturbations, by pushing attacked nodes to dense regions in two graphs, such that they are indistinguishable from many neighbors; and (2) a meta learning-based projected gradient descent method is developed to well choose attack starting points and to improve the search performance for producing effective perturbations. We evaluate the effectiveness of the attack model on real datasets and validate that the attacks can be transferable to other graph learning models. \ No newline at end of file diff --git a/data/2020/neurips/Adversarial Attacks on Linear Contextual Bandits b/data/2020/neurips/Adversarial Attacks on Linear Contextual Bandits new file mode 100644 index 0000000000..890e4b218b --- /dev/null +++ b/data/2020/neurips/Adversarial Attacks on Linear Contextual Bandits @@ -0,0 +1 @@ +Contextual bandit algorithms are applied in a wide range of domains, from advertising to recommender systems, from clinical trials to education. In many of these domains, malicious agents may have incentives to attack the bandit algorithm to induce it to perform a desired behavior. For instance, an unscrupulous ad publisher may try to increase their own revenue at the expense of the advertisers; a seller may want to increase the exposure of their products, or thwart a competitor's advertising campaign. In this paper, we study several attack scenarios and show that a malicious agent can force a linear contextual bandit algorithm to pull any desired arm $T - o(T)$ times over a horizon of $T$ steps, while applying adversarial modifications to either rewards or contexts that only grow logarithmically as $O(\log T)$. We also investigate the case when a malicious agent is interested in affecting the behavior of the bandit algorithm in a single context (e.g., a specific user). We first provide sufficient conditions for the feasibility of the attack and we then propose an efficient algorithm to perform the attack. We validate our theoretical results on experiments performed on both synthetic and real-world datasets. \ No newline at end of file diff --git a/data/2020/neurips/Adversarial Bandits with Corruptions: Regret Lower Bound and No-regret Algorithm b/data/2020/neurips/Adversarial Bandits with Corruptions: Regret Lower Bound and No-regret Algorithm new file mode 100644 index 0000000000..e69de29bb2 diff --git a/data/2020/neurips/Adversarial Blocking Bandits b/data/2020/neurips/Adversarial Blocking Bandits new file mode 100644 index 0000000000..0c66c8253e --- /dev/null +++ b/data/2020/neurips/Adversarial Blocking Bandits @@ -0,0 +1 @@ +We consider a general adversarial multi-armed blocking bandit setting where each played arm can be blocked (unavailable) for some time periods and the reward per arm is given at each time period adversarially without obeying any distribution. The setting models scenarios of allocating scarce limited supplies (e.g., arms) where the supplies replenish and can be reused only after certain time periods. We first show that, in the optimization setting, when the blocking durations and rewards are known in advance, finding an optimal policy (e.g., determining which arm per round) that maximises the cumulative reward is strongly NP-hard, eliminating the possibility of a fully polynomial-time approximation scheme (FPTAS) for the problem unless P = NP. To complement our result, we show that a greedy algorithm that plays the best available arm at each round provides an approximation guarantee that depends on the blocking durations and the path variance of the rewards. In the bandit setting, when the blocking durations and rewards are not known, we design two algorithms, RGA and RGA-META, for the case of bounded duration an path variation. In particular, when the variation budget B_T is known in advance, RGA can achieve O(\sqrt{T(2\tilde{D}+K)B_{T}}) dynamic approximate regret. On the other hand, when B_T is not known, we show that the dynamic approximate regret of RGA-META is at most O((K+\tilde{D})^{1/4}\tilde{B}^{1/2}T^{3/4}) where \tilde{B} is the maximal path variation budget within each batch of RGA-META (which is provably in order of o(\sqrt{T}). We also prove that if either the variation budget or the maximal blocking duration is unbounded, the approximate regret will be at least Theta(T). We also show that the regret upper bound of RGA is tight if the blocking durations are bounded above by an order of O(1). \ No newline at end of file diff --git a/data/2020/neurips/Adversarial Crowdsourcing Through Robust Rank-One Matrix Completion b/data/2020/neurips/Adversarial Crowdsourcing Through Robust Rank-One Matrix Completion new file mode 100644 index 0000000000..2cfca4659b --- /dev/null +++ b/data/2020/neurips/Adversarial Crowdsourcing Through Robust Rank-One Matrix Completion @@ -0,0 +1 @@ +We consider the problem of reconstructing a rank-one matrix from a revealed subset of its entries when some of the revealed entries are corrupted with perturbations that are unknown and can be arbitrarily large. It is not known which revealed entries are corrupted. We propose a new algorithm combining alternating minimization with extreme-value filtering and provide sufficient and necessary conditions to recover the original rank-one matrix. In particular, we show that our proposed algorithm is optimal when the set of revealed entries is given by an Erdős-Renyi random graph. These results are then applied to the problem of classification from crowdsourced data under the assumption that while the majority of the workers are governed by the standard single-coin David-Skene model (i.e., they output the correct answer with a certain probability), some of the workers can deviate arbitrarily from this model. In particular, the "adversarial" workers could even make decisions designed to make the algorithm output an incorrect answer. Extensive experimental results show our algorithm for this problem, based on rank-one matrix completion with perturbations, outperforms all other state-of-the-art methods in such an adversarial scenario. \ No newline at end of file diff --git a/data/2020/neurips/Adversarial Distributional Training for Robust Deep Learning b/data/2020/neurips/Adversarial Distributional Training for Robust Deep Learning new file mode 100644 index 0000000000..952d20ef3b --- /dev/null +++ b/data/2020/neurips/Adversarial Distributional Training for Robust Deep Learning @@ -0,0 +1 @@ +Adversarial training (AT) is among the most effective techniques to improve model robustness by augmenting training data with adversarial examples. However, the adversarially trained models do not perform well enough on test data or under other attack algorithms unseen during training, which remains to be improved. In this paper, we introduce a novel adversarial distributional training (ADT) framework for learning robust models. Specifically, we formulate ADT as a minimax optimization problem, where the inner maximization aims to learn an adversarial distribution to characterize the potential adversarial examples around a natural one, and the outer minimization aims to train robust classifiers by minimizing the expected loss over the worst-case adversarial distributions. We conduct a theoretical analysis on how to solve the minimax problem, leading to a general algorithm for ADT. We further propose three different approaches to parameterize the adversarial distributions. Empirical results on various benchmarks validate the effectiveness of ADT compared with the state-of-the-art AT methods. \ No newline at end of file diff --git a/data/2020/neurips/Adversarial Example Games b/data/2020/neurips/Adversarial Example Games new file mode 100644 index 0000000000..2a509a8c93 --- /dev/null +++ b/data/2020/neurips/Adversarial Example Games @@ -0,0 +1 @@ +The existence of adversarial examples capable of fooling trained neural network classifiers calls for a much better understanding of possible attacks to guide the development of safeguards against them. This includes attack methods in the challenging non-interactive blackbox setting, where adversarial attacks are generated without any access, including queries, to the target model. Prior attacks in this setting have relied mainly on algorithmic innovations derived from empirical observations (e.g., that momentum helps), lacking principled transferability guarantees. In this work, we provide a theoretical foundation for crafting transferable adversarial examples to entire hypothesis classes. We introduce Adversarial Example Games (AEG), a framework that models the crafting of adversarial examples as a min-max game between a generator of attacks and a classifier. AEG provides a new way to design adversarial examples by adversarially training a generator and a classifier from a given hypothesis class (e.g., architecture). We prove that this game has an equilibrium, and that the optimal generator is able to craft adversarial examples that can attack any classifier from the corresponding hypothesis class. We demonstrate the efficacy of AEG on the MNIST and CIFAR-10 datasets, outperforming prior state-of-the-art approaches with an average relative improvement of $29.9\%$ and $47.2\%$ against undefended and robust models (Table 2 & 3) respectively. \ No newline at end of file diff --git a/data/2020/neurips/Adversarial Learning for Robust Deep Clustering b/data/2020/neurips/Adversarial Learning for Robust Deep Clustering new file mode 100644 index 0000000000..4b3fc86a00 --- /dev/null +++ b/data/2020/neurips/Adversarial Learning for Robust Deep Clustering @@ -0,0 +1 @@ +Deep clustering integrates embedding and clustering together to obtain the optimal nonlinear embedding space, which is more effective in real-world scenarios compared with conventional clustering methods. However, the robustness of the clustering network is prone to being attenuated especially when it encounters an adversarial attack. A small perturbation in the embedding space will lead to diverse clustering results since the labels are absent. In this paper, we propose a robust deep clustering method based on adversarial learning. Specifically, we first attempt to define adversarial samples in the embedding space for the clustering network. Meanwhile, we devise an adversarial attack strategy to explore samples that easily fool the clustering layers but do not impact the performance of the deep embedding. We then provide a simple yet efficient defense algorithm to improve the robustness of the clustering network. Experimental results on two popular datasets show that the proposed adversarial learning method can significantly enhance the robustness and further improve the overall clustering performance. Particularly, the proposed method is generally applicable to multiple existing clustering frameworks to boost their robustness. The source code is available at https://github.com/xdxuyang/ALRDC . \ No newline at end of file diff --git a/data/2020/neurips/Adversarial Robustness of Supervised Sparse Coding b/data/2020/neurips/Adversarial Robustness of Supervised Sparse Coding new file mode 100644 index 0000000000..fe19e3f3de --- /dev/null +++ b/data/2020/neurips/Adversarial Robustness of Supervised Sparse Coding @@ -0,0 +1 @@ +Several recent results provide theoretical insights into the phenomena of adversarial examples. Existing results, however, are often limited due to a gap between the simplicity of the models studied and the complexity of those deployed in practice. In this work, we strike a better balance by considering a model that involves learning a representation while at the same time giving a precise generalization bound and a robustness certificate. We focus on the hypothesis class obtained by combining a sparsity-promoting encoder coupled with a linear classifier, and show an interesting interplay between the expressivity and stability of the (supervised) representation map and a notion of margin in the feature space. We bound the robust risk (to $\ell_2$-bounded perturbations) of hypotheses parameterized by dictionaries that achieve a mild encoder gap on training data. Furthermore, we provide a robustness certificate for end-to-end classification. We demonstrate the applicability of our analysis by computing certified accuracy on real data, and compare with other alternatives for certified robustness. \ No newline at end of file diff --git a/data/2020/neurips/Adversarial Self-Supervised Contrastive Learning b/data/2020/neurips/Adversarial Self-Supervised Contrastive Learning new file mode 100644 index 0000000000..fb502012d9 --- /dev/null +++ b/data/2020/neurips/Adversarial Self-Supervised Contrastive Learning @@ -0,0 +1 @@ +Existing adversarial learning approaches mostly use class labels to generate adversarial samples that lead to incorrect predictions, which are then used to augment the training of the model for improved robustness. While some recent works propose semi-supervised adversarial learning methods that utilize unlabeled data, they still require class labels. However, do we really need class labels at all, for adversarially robust training of deep neural networks? In this paper, we propose a novel adversarial attack for unlabeled data, which makes the model confuse the instance-level identities of the perturbed data samples. Further, we present a self-supervised contrastive learning framework to adversarially train a robust neural network without labeled data, which aims to maximize the similarity between a random augmentation of a data sample and its instance-wise adversarial perturbation. We validate our method, Robust Contrastive Learning (RoCL), on multiple benchmark datasets, on which it obtains comparable robust accuracy over state-of-the-art supervised adversarial learning methods, and significantly improved robustness against the black box and unseen types of attacks. Moreover, with further joint fine-tuning with supervised adversarial loss, RoCL obtains even higher robust accuracy over using self-supervised learning alone. Notably, RoCL also demonstrate impressive results in robust transfer learning. \ No newline at end of file diff --git a/data/2020/neurips/Adversarial Soft Advantage Fitting: Imitation Learning without Policy Optimization b/data/2020/neurips/Adversarial Soft Advantage Fitting: Imitation Learning without Policy Optimization new file mode 100644 index 0000000000..d0538e41bb --- /dev/null +++ b/data/2020/neurips/Adversarial Soft Advantage Fitting: Imitation Learning without Policy Optimization @@ -0,0 +1 @@ +Adversarial imitation learning alternates between learning a discriminator -- which tells apart expert's demonstrations from generated ones -- and a generator's policy to produce trajectories that can fool this discriminator. This alternated optimization is known to be delicate in practice since it compounds unstable adversarial training with brittle and sample-inefficient reinforcement learning. We propose to remove the burden of the policy optimization steps by leveraging a novel discriminator formulation. Specifically, our discriminator is explicitly conditioned on two policies: the one from the previous generator's iteration and a learnable policy. When optimized, this discriminator directly learns the optimal generator's policy. Consequently, our discriminator's update solves the generator's optimization problem for free: learning a policy that imitates the expert does not require an additional optimization loop. This formulation effectively cuts by half the implementation and computational burden of adversarial imitation learning algorithms by removing the reinforcement learning phase altogether. We show on a variety of tasks that our simpler approach is competitive to prevalent imitation learning methods. \ No newline at end of file diff --git a/data/2020/neurips/Adversarial Sparse Transformer for Time Series Forecasting b/data/2020/neurips/Adversarial Sparse Transformer for Time Series Forecasting new file mode 100644 index 0000000000..6c357a95ac --- /dev/null +++ b/data/2020/neurips/Adversarial Sparse Transformer for Time Series Forecasting @@ -0,0 +1 @@ +Many approaches have been proposed for time series forecasting, in light of its significance in wide applications including business demand prediction. However, the existing methods suffer from two key limitations. Firstly, most point prediction models only predict an exact value of each time step without flexibility, which can hardly capture the stochasticity of data. Even probabilistic prediction using the likelihood estimation suffers these problems in the same way. Besides, most of them use the auto-regressive generative mode, where ground-truth is provided during training and replaced by the network’s own one-step ahead output during inference, causing the error accumulation in inference. Thus they may fail to forecast time series for long time horizon due to the error accumulation. To solve these issues, in this paper, we propose a new time series forecasting model – Adversarial Sparse Transformer (AST), based on Generative Adversarial Networks (GANs). Specifically, AST adopts a Sparse Transformer as the generator to learn a sparse attention map for time series forecasting, and uses a discriminator to improve the prediction performance at a sequence level. Extensive experiments on several real-world datasets show the effectiveness and efficiency of our method. \ No newline at end of file diff --git a/data/2020/neurips/Adversarial Style Mining for One-Shot Unsupervised Domain Adaptation b/data/2020/neurips/Adversarial Style Mining for One-Shot Unsupervised Domain Adaptation new file mode 100644 index 0000000000..699a2a4567 --- /dev/null +++ b/data/2020/neurips/Adversarial Style Mining for One-Shot Unsupervised Domain Adaptation @@ -0,0 +1 @@ +We aim at the problem named One-Shot Unsupervised Domain Adaptation. Unlike traditional Unsupervised Domain Adaptation, it assumes that only one unlabeled target sample can be available when learning to adapt. This setting is realistic but more challenging, in which conventional adaptation approaches are prone to failure due to the scarce of unlabeled target data. To this end, we propose a novel Adversarial Style Mining approach, which combines the style transfer module and task-specific module into an adversarial manner. Specifically, the style transfer module iteratively searches for harder stylized images around the one-shot target sample according to the current learning state, leading the task model to explore the potential styles that are difficult to solve in the almost unseen target domain, thus boosting the adaptation performance in a data-scarce scenario. The adversarial learning framework makes the style transfer module and task-specific module benefit each other during the competition. Extensive experiments on both cross-domain classification and segmentation benchmarks verify that ASM achieves state-of-the-art adaptation performance under the challenging one-shot setting. \ No newline at end of file diff --git a/data/2020/neurips/Adversarial Training is a Form of Data-dependent Operator Norm Regularization b/data/2020/neurips/Adversarial Training is a Form of Data-dependent Operator Norm Regularization new file mode 100644 index 0000000000..14ae06ffa5 --- /dev/null +++ b/data/2020/neurips/Adversarial Training is a Form of Data-dependent Operator Norm Regularization @@ -0,0 +1 @@ +We establish a theoretical link between adversarial training and operator norm regularization for deep neural networks. Specifically, we prove that $\ell_p$-norm constrained projected gradient ascent based adversarial training with an $\ell_q$-norm loss on the logits of clean and perturbed inputs is equivalent to data-dependent (p, q) operator norm regularization. This fundamental connection confirms the long-standing argument that a network's sensitivity to adversarial examples is tied to its spectral properties and hints at novel ways to robustify and defend against adversarial attacks. We provide extensive empirical evidence on state-of-the-art network architectures to support our theoretical results. \ No newline at end of file diff --git a/data/2020/neurips/Adversarial Weight Perturbation Helps Robust Generalization b/data/2020/neurips/Adversarial Weight Perturbation Helps Robust Generalization new file mode 100644 index 0000000000..8b629d54d6 --- /dev/null +++ b/data/2020/neurips/Adversarial Weight Perturbation Helps Robust Generalization @@ -0,0 +1 @@ +The study on improving the robustness of deep neural networks against adversarial examples grows rapidly in recent years. Among them, adversarial training is the most promising one, which flattens the input loss landscape (loss change with respect to input) via training on adversarially perturbed examples. However, how the widely used weight loss landscape (loss change with respect to weight) performs in adversarial training is rarely explored. In this paper, we investigate the weight loss landscape from a new perspective, and identify a clear correlation between the flatness of weight loss landscape and robust generalization gap. Several well-recognized adversarial training improvements, such as early stopping, designing new objective functions, or leveraging unlabeled data, all implicitly flatten the weight loss landscape. Based on these observations, we propose a simple yet effective Adversarial Weight Perturbation (AWP) to explicitly regularize the flatness of weight loss landscape, forming a double-perturbation mechanism in the adversarial training framework that adversarially perturbs both inputs and weights. Extensive experiments demonstrate that AWP indeed brings flatter weight loss landscape and can be easily incorporated into various existing adversarial training methods to further boost their adversarial robustness. \ No newline at end of file diff --git a/data/2020/neurips/Adversarial robustness via robust low rank representations b/data/2020/neurips/Adversarial robustness via robust low rank representations new file mode 100644 index 0000000000..a882558e00 --- /dev/null +++ b/data/2020/neurips/Adversarial robustness via robust low rank representations @@ -0,0 +1,4 @@ +Adversarial robustness measures the susceptibility of a classifier to imperceptible perturbations made to the inputs at test time. In this work we highlight the benefits of natural low rank representations that often exist for real data such as images, for training neural networks with certified robustness guarantees. +Our first contribution is for certified robustness to perturbations measured in $\ell_2$ norm. We exploit low rank data representations to provide improved guarantees over state-of-the-art randomized smoothing-based approaches on standard benchmark datasets such as CIFAR-10 and CIFAR-100. +Our second contribution is for the more challenging setting of certified robustness to perturbations measured in $\ell_\infty$ norm. We demonstrate empirically that natural low rank representations have inherent robustness properties, that can be leveraged to provide significantly better guarantees for certified robustness to $\ell_\infty$ perturbations in those representations. Our certificate of $\ell_\infty$ robustness relies on a natural quantity involving the $\infty \to 2$ matrix operator norm associated with the representation, to translate robustness guarantees from $\ell_2$ to $\ell_\infty$ perturbations. +A key technical ingredient for our certification guarantees is a fast algorithm with provable guarantees based on the multiplicative weights update method to provide upper bounds on the above matrix norm. Our algorithmic guarantees improve upon the state of the art for this problem, and may be of independent interest. \ No newline at end of file diff --git a/data/2020/neurips/Adversarially Robust Few-Shot Learning: A Meta-Learning Approach b/data/2020/neurips/Adversarially Robust Few-Shot Learning: A Meta-Learning Approach new file mode 100644 index 0000000000..2f37b1aa3b --- /dev/null +++ b/data/2020/neurips/Adversarially Robust Few-Shot Learning: A Meta-Learning Approach @@ -0,0 +1 @@ +Previous work on adversarially robust neural networks for image classification requires large training sets and computationally expensive training procedures. On the other hand, few-shot learning methods are highly vulnerable to adversarial examples. The goal of our work is to produce networks which both perform well at few-shot classification tasks and are simultaneously robust to adversarial examples. We develop an algorithm for producing adversarially robust meta-learners, and we thoroughly investigate factors which contribute to adversarial vulnerability. Moreover, our method achieves far superior robust performance on few-shot image classification tasks, such as Mini-ImageNet and CIFAR-FS, than robust transfer learning. \ No newline at end of file diff --git a/data/2020/neurips/Adversarially Robust Streaming Algorithms via Differential Privacy b/data/2020/neurips/Adversarially Robust Streaming Algorithms via Differential Privacy new file mode 100644 index 0000000000..6c7913c20d --- /dev/null +++ b/data/2020/neurips/Adversarially Robust Streaming Algorithms via Differential Privacy @@ -0,0 +1 @@ +A streaming algorithm is said to be adversarially robust if its accuracy guarantees are maintained even when the data stream is chosen maliciously, by an adaptive adversary. We establish a connection between adversarial robustness of streaming algorithms and the notion of differential privacy. This connection allows us to design new adversarially robust streaming algorithms that outperform the current state-of-the-art constructions for many interesting regimes of parameters. \ No newline at end of file diff --git a/data/2020/neurips/Adversarially-learned Inference via an Ensemble of Discrete Undirected Graphical Models b/data/2020/neurips/Adversarially-learned Inference via an Ensemble of Discrete Undirected Graphical Models new file mode 100644 index 0000000000..552666c042 --- /dev/null +++ b/data/2020/neurips/Adversarially-learned Inference via an Ensemble of Discrete Undirected Graphical Models @@ -0,0 +1 @@ +Undirected graphical models are compact representations of joint probability distributions over random variables. To carry out an inference task of interest, graphical models of arbitrary topology can be trained using empirical risk minimization. However, when faced with new tasks, these models (EGMs) often need to be re-trained. Instead, we propose an inference-agnostic adversarial training framework for producing an ensemble of graphical models (AGMs). The ensemble is optimized to generate data, and inference is learned as a by-product of this endeavor. AGMs perform comparably with EGMs on inference tasks that the latter were specifically optimized for. Most importantly, AGMs show significantly better generalization capabilities across inference tasks. AGMs are also on par with GibbsNet, a state-of-the-art deep neural architecture, which like AGMs, allows conditioning on any subset of random variables. Finally, AGMs allow fast data sampling, competitive with Gibbs sampling from EGMs. \ No newline at end of file diff --git a/data/2020/neurips/Agnostic $Q$-learning with Function Approximation in Deterministic Systems: Near-Optimal Bounds on Approximation Error and Sample Complexity b/data/2020/neurips/Agnostic $Q$-learning with Function Approximation in Deterministic Systems: Near-Optimal Bounds on Approximation Error and Sample Complexity new file mode 100644 index 0000000000..0907704f8f --- /dev/null +++ b/data/2020/neurips/Agnostic $Q$-learning with Function Approximation in Deterministic Systems: Near-Optimal Bounds on Approximation Error and Sample Complexity @@ -0,0 +1 @@ +The current paper studies the problem of agnostic Q -learning with function approx-1 imation in deterministic systems where the optimal Q -function is approximable 2 by a function in the class F with approximation error δ ≥ 0 . We propose a novel 3 recursion-based algorithm and show that if δ = O (cid:0) ρ/ √ dim E (cid:1) , then one can find 4 the optimal policy using O (dim E ) trajectories, where ρ is the gap between the 5 optimal Q -value of the best actions and that of the second-best actions and dim E 6 is the Eluder dimension of F . Our result has two implications: 7 1. In conjunction with the lower bound in [Du et al., 2020], our upper bound 8 suggests that the condition δ = (cid:101) Θ (cid:0) ρ/ √ dim E (cid:1) is necessary and sufficient for 9 algorithms with polynomial sample complexity. 10 2. In conjunction with the obvious lower bound in the tabular case, our upper 11 bound suggests that the sample complexity (cid:101) Θ (dim E ) is tight in the agnostic 12 setting. 13 Therefore, we help address the open problem on agnostic Q -learning proposed 14 in [Wen and Van Roy, 2013]. We further extend our algorithm to the stochastic 15 reward setting and obtain similar results. 16 \ No newline at end of file diff --git a/data/2020/neurips/Agnostic Learning of a Single Neuron with Gradient Descent b/data/2020/neurips/Agnostic Learning of a Single Neuron with Gradient Descent new file mode 100644 index 0000000000..51cde16748 --- /dev/null +++ b/data/2020/neurips/Agnostic Learning of a Single Neuron with Gradient Descent @@ -0,0 +1 @@ +We consider the problem of learning the best-fitting single neuron as measured by the expected square loss $\mathbb{E}_{(x,y)\sim \mathcal{D}}[(\sigma(w^\top x)-y)^2]$ over some unknown joint distribution $\mathcal{D}$ by using gradient descent to minimize the empirical risk induced by a set of i.i.d. samples $S\sim \mathcal{D}^n$. The activation function $\sigma$ is an arbitrary Lipschitz and non-decreasing function, making the optimization problem nonconvex and nonsmooth in general, and covers typical neural network activation functions and inverse link functions in the generalized linear model setting. In the agnostic PAC learning setting, where no assumption on the relationship between the labels $y$ and the input $x$ is made, if the optimal population risk is $\mathsf{OPT}$, we show that gradient descent achieves population risk $O(\mathsf{OPT}^{1/2})+\epsilon$ in polynomial time and sample complexity. When labels take the form $y = \sigma(v^\top x) + \xi$ for zero-mean sub-Gaussian noise $\xi$, we show that gradient descent achieves population risk $\mathsf{OPT} + \epsilon$. Our sample complexity and runtime guarantees are (almost) dimension independent, and when $\sigma$ is strictly increasing and Lipschitz, require no distributional assumptions beyond boundedness. For ReLU, we show the same results under a nondegeneracy assumption for the marginal distribution of the input. To the best of our knowledge, this is the first result for agnostic learning of a single neuron using gradient descent. \ No newline at end of file diff --git a/data/2020/neurips/Agnostic Learning with Multiple Objectives b/data/2020/neurips/Agnostic Learning with Multiple Objectives new file mode 100644 index 0000000000..3bf21e16da --- /dev/null +++ b/data/2020/neurips/Agnostic Learning with Multiple Objectives @@ -0,0 +1 @@ +Most machine learning tasks are inherently multi-objective. This means that the learner has to come up with a model that performs well across a number of base objectives L 1 , . . . , L p , as opposed to a single one. Since optimizing with respect to multiple objectives at the same time is often computationally expensive, the base objectives are often combined in an ensemble (cid:80) pk =1 λ k L k , thereby reducing the problem to scalar optimization. The mixture weights λ k are set to uniform or some other fixed distribution, based on the learner’s preferences. We argue that learning with a fixed distribution on the mixture weights runs the risk of overfitting to some individual objectives and significantly harming others, despite performing well on an entire ensemble. Moreover, in reality, the true preferences of a learner across multiple objectives are often unknown or hard to express as a specific distribution. Instead, we propose a new framework of Agnostic Learning with Multiple Objectives (ALMO), where a model is optimized for any weights in the mixture of base objectives. We present data-dependent Rademacher complexity guarantees for learning in the ALMO framework, which are used to guide a scalable optimization algorithm and the corresponding regularization. We present convergence guarantees for this algorithm, assuming convexity of the loss functions and the underlying hypothesis space. We further implement the algorithm in a popular symbolic gradient computation framework and empirically demonstrate on a number of datasets the benefits of ALMO framework versus learning with a fixed mixture weights distribution. \ No newline at end of file diff --git a/data/2020/neurips/Agree to Disagree: Adaptive Ensemble Knowledge Distillation in Gradient Space b/data/2020/neurips/Agree to Disagree: Adaptive Ensemble Knowledge Distillation in Gradient Space new file mode 100644 index 0000000000..02c1a6024c --- /dev/null +++ b/data/2020/neurips/Agree to Disagree: Adaptive Ensemble Knowledge Distillation in Gradient Space @@ -0,0 +1 @@ +Distilling knowledge from an ensemble of teacher models is expected to have a more promising performance than that from a single one. Current methods mainly adopt a vanilla average rule, i.e. , to simply take the average of all teacher losses for training the student network. However, this approach treats teachers equally and ignores the diversity among them. When conflicts or competitions exist among teachers, which is common, the inner compromise might hurt the distillation performance. In this paper, we examine the diversity of teacher models in the gradient space and regard the ensemble knowledge distillation as a multi-objective optimization problem so that we can determine a better optimization direction for the training of student network. Besides, we also introduce a tolerance parameter to accommodate disagreement among teachers. In this way, our method can be seen as a dynamic weighting method for each teacher in the ensemble. Extensive experiments validate the effectiveness of our method for both logits-based and feature-based cases. \ No newline at end of file diff --git a/data/2020/neurips/Algorithmic recourse under imperfect causal knowledge: a probabilistic approach b/data/2020/neurips/Algorithmic recourse under imperfect causal knowledge: a probabilistic approach new file mode 100644 index 0000000000..e9e4891631 --- /dev/null +++ b/data/2020/neurips/Algorithmic recourse under imperfect causal knowledge: a probabilistic approach @@ -0,0 +1 @@ +Recent work has discussed the limitations of counterfactual explanations to recommend actions for algorithmic recourse, and argued for the need of taking causal relationships between features into consideration. Unfortunately, in practice, the true underlying structural causal model is generally unknown. In this work, we first show that it is impossible to guarantee recourse without access to the true structural equations. To address this limitation, we propose two probabilistic approaches to select optimal actions that achieve recourse with high probability given limited causal knowledge (e.g., only the causal graph). The first captures uncertainty over structural equations under additive Gaussian noise, and uses Bayesian model averaging to estimate the counterfactual distribution. The second removes any assumptions on the structural equations by instead computing the average effect of recourse actions on individuals similar to the person who seeks recourse, leading to a novel subpopulation-based interventional notion of recourse. We then derive a gradient-based procedure for selecting optimal recourse actions, and empirically show that the proposed approaches lead to more reliable recommendations under imperfect causal knowledge than non-probabilistic baselines. \ No newline at end of file diff --git a/data/2020/neurips/All Word Embeddings from One Embedding b/data/2020/neurips/All Word Embeddings from One Embedding new file mode 100644 index 0000000000..dcb6e7472b --- /dev/null +++ b/data/2020/neurips/All Word Embeddings from One Embedding @@ -0,0 +1 @@ +In neural network-based models for natural language processing (NLP), the largest part of the parameters often consists of word embeddings. Conventional models prepare a large embedding matrix whose size depends on the vocabulary size. Therefore, storing these models in memory and disk storage is costly. In this study, to reduce the total number of parameters, the embeddings for all words are represented by transforming a shared embedding. The proposed method, ALONE (all word embeddings from one), constructs the embedding of a word by modifying the shared embedding with a filter vector, which is word-specific but non-trainable. Then, we input the constructed embedding into a feed-forward neural network to increase its expressiveness. Naively, the filter vectors occupy the same memory size as the conventional embedding matrix, which depends on the vocabulary size. To solve this issue, we also introduce a memory-efficient filter construction approach. We indicate our ALONE can be used as word representation sufficiently through an experiment on the reconstruction of pre-trained word embeddings. In addition, we also conduct experiments on NLP application tasks: machine translation and summarization. We combined ALONE with the current state-of-the-art encoder-decoder model, the Transformer, and achieved comparable scores on WMT 2014 English-to-German translation and DUC 2004 very short summarization with less parameters. \ No newline at end of file diff --git a/data/2020/neurips/All-or-nothing statistical and computational phase transitions in sparse spiked matrix estimation b/data/2020/neurips/All-or-nothing statistical and computational phase transitions in sparse spiked matrix estimation new file mode 100644 index 0000000000..288775b897 --- /dev/null +++ b/data/2020/neurips/All-or-nothing statistical and computational phase transitions in sparse spiked matrix estimation @@ -0,0 +1 @@ +We determine statistical and computational limits for estimation of a rank-one matrix (the spike) corrupted by an additive gaussian noise matrix, in a sparse limit, where the underlying hidden vector (that constructs the rank-one matrix) has a number of non-zero components that scales sub-linearly with the total dimension of the vector, and the signal-to-noise ratio tends to infinity at an appropriate speed. We prove explicit low-dimensional variational formulas for the asymptotic mutual information between the spike and the observed noisy matrix and analyze the approximate message passing algorithm in the sparse regime. For Bernoulli and Bernoulli-Rademacher distributed vectors, and when the sparsity and signal strength satisfy an appropriate scaling relation, we find all-or-nothing phase transitions for the asymptotic minimum and algorithmic mean-square errors. These jump from their maximum possible value to zero, at well defined signal-to-noise thresholds whose asymptotic values we determine exactly. In the asymptotic regime the statistical-to-algorithmic gap diverges indicating that sparse recovery is hard for approximate message passing. \ No newline at end of file diff --git a/data/2020/neurips/Almost Optimal Model-Free Reinforcement Learningvia Reference-Advantage Decomposition b/data/2020/neurips/Almost Optimal Model-Free Reinforcement Learningvia Reference-Advantage Decomposition new file mode 100644 index 0000000000..6cdc744af4 --- /dev/null +++ b/data/2020/neurips/Almost Optimal Model-Free Reinforcement Learningvia Reference-Advantage Decomposition @@ -0,0 +1 @@ +We study the reinforcement learning problem in the setting of finite-horizon episodic Markov Decision Processes (MDPs) with $S$ states, $A$ actions, and episode length $H$. We propose a model-free algorithm UCB-Advantage and prove that it achieves $\tilde{O}(\sqrt{H^2SAT})$ regret where $T = KH$ and $K$ is the number of episodes to play. Our regret bound improves upon the results of [Jin et al., 2018] and matches the best known model-based algorithms as well as the information theoretic lower bound up to logarithmic factors. We also show that UCB-Advantage achieves low local switching cost and applies to concurrent reinforcement learning, improving upon the recent results of [Bai et al., 2019]. \ No newline at end of file diff --git a/data/2020/neurips/Almost Surely Stable Deep Dynamics b/data/2020/neurips/Almost Surely Stable Deep Dynamics new file mode 100644 index 0000000000..20a762aa04 --- /dev/null +++ b/data/2020/neurips/Almost Surely Stable Deep Dynamics @@ -0,0 +1 @@ +We introduce a method for learning provably stable deep neural network based dynamic models from observed data. Specifically, we consider discrete-time stochastic dynamic models, as they are of particular interest in practical applications such as estimation and control. However, these aspects exacerbate the challenge of guaranteeing stability. Our method works by embedding a Lyapunov neural network into the dynamic model, thereby inherently satisfying the stability criterion. To this end, we propose two approaches and apply them in both the deterministic and stochastic settings: one exploits convexity of the Lyapunov function, while the other enforces stability through an implicit output layer. We demonstrate the utility of each approach through numerical examples. \ No newline at end of file diff --git a/data/2020/neurips/An Analysis of SVD for Deep Rotation Estimation b/data/2020/neurips/An Analysis of SVD for Deep Rotation Estimation new file mode 100644 index 0000000000..dcac791e2b --- /dev/null +++ b/data/2020/neurips/An Analysis of SVD for Deep Rotation Estimation @@ -0,0 +1 @@ +Symmetric orthogonalization via SVD, and closely related procedures, are well-known techniques for projecting matrices onto $O(n)$ or $SO(n)$. These tools have long been used for applications in computer vision, for example optimal 3D alignment problems solved by orthogonal Procrustes, rotation averaging, or Essential matrix decomposition. Despite its utility in different settings, SVD orthogonalization as a procedure for producing rotation matrices is typically overlooked in deep learning models, where the preferences tend toward classic representations like unit quaternions, Euler angles, and axis-angle, or more recently-introduced methods. Despite the importance of 3D rotations in computer vision and robotics, a single universally effective representation is still missing. Here, we explore the viability of SVD orthogonalization for 3D rotations in neural networks. We present a theoretical analysis that shows SVD is the natural choice for projecting onto the rotation group. Our extensive quantitative analysis shows simply replacing existing representations with the SVD orthogonalization procedure obtains state of the art performance in many deep learning applications covering both supervised and unsupervised training. \ No newline at end of file diff --git a/data/2020/neurips/An Asymptotically Optimal Primal-Dual Incremental Algorithm for Contextual Linear Bandits b/data/2020/neurips/An Asymptotically Optimal Primal-Dual Incremental Algorithm for Contextual Linear Bandits new file mode 100644 index 0000000000..c919fac233 --- /dev/null +++ b/data/2020/neurips/An Asymptotically Optimal Primal-Dual Incremental Algorithm for Contextual Linear Bandits @@ -0,0 +1 @@ +In the contextual linear bandit setting, algorithms built on the optimism principle fail to exploit the structure of the problem and have been shown to be asymptotically suboptimal. In this paper, we follow recent approaches of deriving asymptotically optimal algorithms from problem-dependent regret lower bounds and we introduce a novel algorithm improving over the state-of-the-art along multiple dimensions. We build on a reformulation of the lower bound, where context distribution and exploration policy are decoupled, and we obtain an algorithm robust to unbalanced context distributions. Then, using an incremental primal-dual approach to solve the Lagrangian relaxation of the lower bound, we obtain a scalable and computationally efficient algorithm. Finally, we remove forced exploration and build on confidence intervals of the optimization problem to encourage a minimum level of exploration that is better adapted to the problem structure. We demonstrate the asymptotic optimality of our algorithm, while providing both problem-dependent and worst-case finite-time regret guarantees. Our bounds scale with the logarithm of the number of arms, thus avoiding the linear dependence common in all related prior works. Notably, we establish minimax optimality for any learning horizon in the special case of non-contextual linear bandits. Finally, we verify that our algorithm obtains better empirical performance than state-of-the-art baselines. \ No newline at end of file diff --git a/data/2020/neurips/An Efficient Adversarial Attack for Tree Ensembles b/data/2020/neurips/An Efficient Adversarial Attack for Tree Ensembles new file mode 100644 index 0000000000..6783debacc --- /dev/null +++ b/data/2020/neurips/An Efficient Adversarial Attack for Tree Ensembles @@ -0,0 +1 @@ +We study the problem of efficient adversarial attacks on tree based ensembles such as gradient boosting decision trees (GBDTs) and random forests (RFs). Since these models are non-continuous step functions and gradient does not exist, most existing efficient adversarial attacks are not applicable. Although decision-based black-box attacks can be applied, they cannot utilize the special structure of trees. In our work, we transform the attack problem into a discrete search problem specially designed for tree ensembles, where the goal is to find a valid "leaf tuple" that leads to mis-classification while having the shortest distance to the original input. With this formulation, we show that a simple yet effective greedy algorithm can be applied to iteratively optimize the adversarial example by moving the leaf tuple to its neighborhood within hamming distance 1. Experimental results on several large GBDT and RF models with up to hundreds of trees demonstrate that our method can be thousands of times faster than the previous mixed-integer linear programming (MILP) based approach, while also providing smaller (better) adversarial examples than decision-based black-box attacks on general $\ell_p$ ($p=1, 2, \infty$) norm perturbations. Our code is available at this https URL. \ No newline at end of file diff --git a/data/2020/neurips/An Efficient Asynchronous Method for Integrating Evolutionary and Gradient-based Policy Search b/data/2020/neurips/An Efficient Asynchronous Method for Integrating Evolutionary and Gradient-based Policy Search new file mode 100644 index 0000000000..67dc81fdb6 --- /dev/null +++ b/data/2020/neurips/An Efficient Asynchronous Method for Integrating Evolutionary and Gradient-based Policy Search @@ -0,0 +1 @@ +Deep reinforcement learning (DRL) algorithms and evolution strategies (ES) have been applied to various tasks, showing excellent performances. These have the opposite properties, with DRL having good sample efficiency and poor stability, while ES being vice versa. Recently, there have been attempts to combine these algorithms, but these methods fully rely on synchronous update scheme, making it not ideal to maximize the benefits of the parallelism in ES. To solve this challenge, asynchronous update scheme was introduced, which is capable of good time-efficiency and diverse policy exploration. In this paper, we introduce an Asynchronous Evolution Strategy-Reinforcement Learning (AES-RL) that maximizes the parallel efficiency of ES and integrates it with policy gradient methods. Specifically, we propose 1) a novel framework to merge ES and DRL asynchronously and 2) various asynchronous update methods that can take all advantages of asynchronism, ES, and DRL, which are exploration and time efficiency, stability, and sample efficiency, respectively. The proposed framework and update methods are evaluated in continuous control benchmark work, showing superior performance as well as time efficiency compared to the previous methods. \ No newline at end of file diff --git a/data/2020/neurips/An Efficient Framework for Clustered Federated Learning b/data/2020/neurips/An Efficient Framework for Clustered Federated Learning new file mode 100644 index 0000000000..408c0abc38 --- /dev/null +++ b/data/2020/neurips/An Efficient Framework for Clustered Federated Learning @@ -0,0 +1 @@ +We address the problem of federated learning (FL) where users are distributed and partitioned into clusters. This setup captures settings where different groups of users have their own objectives (learning tasks) but by aggregating their data with others in the same cluster (same learning task), they can leverage the strength in numbers in order to perform more efficient federated learning. For this new framework of clustered federated learning, we propose the Iterative Federated Clustering Algorithm (IFCA), which alternately estimates the cluster identities of the users and optimizes model parameters for the user clusters via gradient descent. We analyze the convergence rate of this algorithm first in a linear model with squared loss and then for generic strongly convex and smooth loss functions. We show that in both settings, with good initialization, IFCA is guaranteed to converge, and discuss the optimality of the statistical error rate. In particular, for the linear model with two clusters, we can guarantee that our algorithm converges as long as the initialization is slightly better than random. When the clustering structure is ambiguous, we propose to train the models by combining IFCA with the weight sharing technique in multi-task learning. In the experiments, we show that our algorithm can succeed even if we relax the requirements on initialization with random initialization and multiple restarts. We also present experimental results showing that our algorithm is efficient in non-convex problems such as neural networks. We demonstrate the benefits of IFCA over the baselines on several clustered FL benchmarks. \ No newline at end of file diff --git a/data/2020/neurips/An Empirical Process Approach to the Union Bound: Practical Algorithms for Combinatorial and Linear Bandits b/data/2020/neurips/An Empirical Process Approach to the Union Bound: Practical Algorithms for Combinatorial and Linear Bandits new file mode 100644 index 0000000000..1a5c9168a7 --- /dev/null +++ b/data/2020/neurips/An Empirical Process Approach to the Union Bound: Practical Algorithms for Combinatorial and Linear Bandits @@ -0,0 +1 @@ +This paper proposes near-optimal algorithms for the pure-exploration linear bandit problem in the fixed confidence and fixed budget settings. Leveraging ideas from the theory of suprema of empirical processes, we provide an algorithm whose sample complexity scales with the geometry of the instance and avoids an explicit union bound over the number of arms. Unlike previous approaches which sample based on minimizing a worst-case variance (e.g. G-optimal design), we define an experimental design objective based on the Gaussian-width of the underlying arm set. We provide a novel lower bound in terms of this objective that highlights its fundamental role in the sample complexity. The sample complexity of our fixed confidence algorithm matches this lower bound, and in addition is computationally efficient for combinatorial classes, e.g. shortest-path, matchings and matroids, where the arm sets can be exponentially large in the dimension. Finally, we propose the first algorithm for linear bandits in the the fixed budget setting. Its guarantee matches our lower bound up to logarithmic factors. \ No newline at end of file diff --git a/data/2020/neurips/An Equivalence between Loss Functions and Non-Uniform Sampling in Experience Replay b/data/2020/neurips/An Equivalence between Loss Functions and Non-Uniform Sampling in Experience Replay new file mode 100644 index 0000000000..fcdf04ec7d --- /dev/null +++ b/data/2020/neurips/An Equivalence between Loss Functions and Non-Uniform Sampling in Experience Replay @@ -0,0 +1 @@ +Prioritized Experience Replay (PER) is a deep reinforcement learning technique in which agents learn from transitions sampled with non-uniform probability proportionate to their temporal-difference error. We show that any loss function evaluated with non-uniformly sampled data can be transformed into another uniformly sampled loss function with the same expected gradient. Surprisingly, we find in some environments PER can be replaced entirely by this new loss function without impact to empirical performance. Furthermore, this relationship suggests a new branch of improvements to PER by correcting its uniformly sampled loss function equivalent. We demonstrate the effectiveness of our proposed modifications to PER and the equivalent loss function in several MuJoCo and Atari environments. \ No newline at end of file diff --git a/data/2020/neurips/An Imitation from Observation Approach to Transfer Learning with Dynamics Mismatch b/data/2020/neurips/An Imitation from Observation Approach to Transfer Learning with Dynamics Mismatch new file mode 100644 index 0000000000..cf84af4c9e --- /dev/null +++ b/data/2020/neurips/An Imitation from Observation Approach to Transfer Learning with Dynamics Mismatch @@ -0,0 +1 @@ +We examine the problem of transferring a policy learned in a source environment to a target environment with different dynamics, particularly in the case where it is critical to reduce the amount of interaction with the target environment during learning. This problem is particularly important in sim-to-real transfer because simulators inevitably model real-world dynamics imperfectly. In this paper, we show that one existing solution to this transfer problem - grounded action transformation - is closely related to the problem of imitation from observation (IfO): learning behaviors that mimic the observations of behavior demonstrations. After establishing this relationship, we hypothesize that recent state-of-the-art approaches from the IfO literature can be effectively repurposed for grounded transfer learning.To validate our hypothesis we derive a new algorithm - generative adversarial reinforced action transformation (GARAT) - based on adversarial imitation from observation techniques. We run experiments in several domains with mismatched dynamics, and find that agents trained with GARAT achieve higher returns in the target environment compared to existing black-box transfer methods \ No newline at end of file diff --git a/data/2020/neurips/An Improved Analysis of (Variance-Reduced) Policy Gradient and Natural Policy Gradient Methods b/data/2020/neurips/An Improved Analysis of (Variance-Reduced) Policy Gradient and Natural Policy Gradient Methods new file mode 100644 index 0000000000..296f07f814 --- /dev/null +++ b/data/2020/neurips/An Improved Analysis of (Variance-Reduced) Policy Gradient and Natural Policy Gradient Methods @@ -0,0 +1 @@ +In this paper, we revisit and improve the convergence of policy gradient (PG), natural PG (NPG) methods, and their variance-reduced variants, under general smooth policy parametrizations. More specifically, with the Fisher information matrix of the policy being positive definite: i) we show that a state-of-the-art variance-reduced PG method, which has only been shown to converge to stationary points, converges to the globally optimal value up to some inherent function approximation error due to policy parametrization; ii) we show that NPG enjoys a lower sample complexity; iii) we propose SRVR-NPG, which incorporates variance-reduction into the NPG update. Our improvements follow from an observation that the convergence of (variance-reduced) PG and NPG methods can improve each other: the stationary convergence analysis of PG can be applied to NPG as well, and the global convergence analysis of NPG can help to establish the global convergence of (variance-reduced) PG methods. Our analysis carefully integrates the advantages of these two lines of works. Thanks to this improvement, we have also made variance-reduction for NPG possible, with both global convergence and an efficient finite-sample complexity. \ No newline at end of file diff --git a/data/2020/neurips/An Improved Analysis of Stochastic Gradient Descent with Momentum b/data/2020/neurips/An Improved Analysis of Stochastic Gradient Descent with Momentum new file mode 100644 index 0000000000..a4a4320d2a --- /dev/null +++ b/data/2020/neurips/An Improved Analysis of Stochastic Gradient Descent with Momentum @@ -0,0 +1 @@ +SGD with momentum (SGDM) has been widely applied in many machine learning tasks, and it is often applied with dynamic stepsizes and momentum weights tuned in a stagewise manner. Despite of its empirical advantage over SGD, the role of momentum is still unclear in general since previous analyses on SGDM either provide worse convergence bounds than those of SGD, or assume Lipschitz or quadratic objectives, which fail to hold in practice. Furthermore, the role of dynamic parameters have not been addressed. In this work, we show that SGDM converges as fast as SGD for smooth objectives under both strongly convex and nonconvex settings. We also prove that multistage strategy is beneficial for SGDM compared to using fixed parameters. Finally, we verify these theoretical claims by numerical experiments. \ No newline at end of file diff --git a/data/2020/neurips/An Optimal Elimination Algorithm for Learning a Best Arm b/data/2020/neurips/An Optimal Elimination Algorithm for Learning a Best Arm new file mode 100644 index 0000000000..aa55097abe --- /dev/null +++ b/data/2020/neurips/An Optimal Elimination Algorithm for Learning a Best Arm @@ -0,0 +1 @@ +We consider the classic problem of $(\epsilon,\delta)$-PAC learning a best arm where the goal is to identify with confidence $1-\delta$ an arm whose mean is an $\epsilon$-approximation to that of the highest mean arm in a multi-armed bandit setting. This problem is one of the most fundamental problems in statistics and learning theory, yet somewhat surprisingly its worst-case sample complexity is not well understood. In this paper, we propose a new approach for $(\epsilon,\delta)$-PAC learning a best arm. This approach leads to an algorithm whose sample complexity converges to \emph{exactly} the optimal sample complexity of $(\epsilon,\delta)$-learning the mean of $n$ arms separately and we complement this result with a conditional matching lower bound. More specifically: \ No newline at end of file diff --git a/data/2020/neurips/An Unbiased Risk Estimator for Learning with Augmented Classes b/data/2020/neurips/An Unbiased Risk Estimator for Learning with Augmented Classes new file mode 100644 index 0000000000..41213d9155 --- /dev/null +++ b/data/2020/neurips/An Unbiased Risk Estimator for Learning with Augmented Classes @@ -0,0 +1 @@ +In this paper, we study the problem of learning with augmented classes (LAC), where new classes that do not appear in the training dataset might emerge in the testing phase. The mixture of known classes and new classes in the testing distribution makes the LAC problem quite challenging. Our discovery is that by exploiting cheap and vast unlabeled data, the testing distribution can be estimated in the training stage, which paves us a way to develop algorithms with nice statistical properties. Specifically, we propose an unbiased risk estimator over the testing distribution for the LAC problem, and further develop an efficient algorithm to perform the empirical risk minimization. Both asymptotic and non-asymptotic analyses are provided as theoretical guarantees. The efficacy of the proposed algorithm is also confirmed by experiments. \ No newline at end of file diff --git a/data/2020/neurips/An Unsupervised Information-Theoretic Perceptual Quality Metric b/data/2020/neurips/An Unsupervised Information-Theoretic Perceptual Quality Metric new file mode 100644 index 0000000000..1a5dee1b1e --- /dev/null +++ b/data/2020/neurips/An Unsupervised Information-Theoretic Perceptual Quality Metric @@ -0,0 +1 @@ +Tractable models of human perception have proved to be challenging to build. Hand-designed models such as MS-SSIM remain popular predictors of human image quality judgements due to their simplicity and speed. Recent modern deep learning approaches can perform better, but they rely on supervised data which can be costly to gather: large sets of class labels such as ImageNet, image quality ratings, or both. We combine recent advances in information-theoretic objective functions with a computational architecture informed by the physiology of the human visual system and unsupervised training on pairs of video frames, yielding our Perceptual Information Metric (PIM). We show that PIM is competitive with supervised metrics on the recent and challenging BAPPS image quality assessment dataset. We also perform qualitative experiments using the ImageNet-C dataset, and establish that our approach is robust with respect to architectural details. \ No newline at end of file diff --git a/data/2020/neurips/An analytic theory of shallow networks dynamics for hinge loss classification b/data/2020/neurips/An analytic theory of shallow networks dynamics for hinge loss classification new file mode 100644 index 0000000000..28999d6155 --- /dev/null +++ b/data/2020/neurips/An analytic theory of shallow networks dynamics for hinge loss classification @@ -0,0 +1 @@ +Neural networks have been shown to perform incredibly well in classification tasks over structured high-dimensional datasets. However, the learning dynamics of such networks is still poorly understood. In this paper we study in detail the training dynamics of a simple type of neural network: a single hidden layer trained to perform a classification task. We show that in a suitable mean-field limit this case maps to a single-node learning problem with a time-dependent dataset determined self-consistently from the average nodes population. We specialize our theory to the prototypical case of a linearly separable data and a linear hinge loss, for which the dynamics can be explicitly solved in the infinite dataset limit. This allows us to address in a simple setting several phenomena appearing in modern networks such as slowing down of training dynamics, crossover between rich and lazy learning, and overfitting. Finally, we assess the limitations of mean-field theory by studying the case of large but finite number of nodes and of training samples. \ No newline at end of file diff --git a/data/2020/neurips/An efficient nonconvex reformulation of stagewise convex optimization problems b/data/2020/neurips/An efficient nonconvex reformulation of stagewise convex optimization problems new file mode 100644 index 0000000000..b675c1b39c --- /dev/null +++ b/data/2020/neurips/An efficient nonconvex reformulation of stagewise convex optimization problems @@ -0,0 +1 @@ +Convex optimization problems with staged structure appear in several contexts, including optimal control, verification of deep neural networks, and isotonic regression. Off-the-shelf solvers can solve these problems but may scale poorly. We develop a nonconvex reformulation designed to exploit this staged structure. Our reformulation has only simple bound constraints, enabling solution via projected gradient methods and their accelerated variants. The method automatically generates a sequence of primal and dual feasible solutions to the original convex problem, making optimality certification easy. We establish theoretical properties of the nonconvex formulation, showing that it is (almost) free of spurious local minima and has the same global optimum as the convex problem. We modify PGD to avoid spurious local minimizers so it always converges to the global minimizer. For neural network verification, our approach obtains small duality gaps in only a few gradient steps. Consequently, it can quickly solve large-scale verification problems faster than both off-the-shelf and specialized solvers. \ No newline at end of file diff --git a/data/2020/neurips/An implicit function learning approach for parametric modal regression b/data/2020/neurips/An implicit function learning approach for parametric modal regression new file mode 100644 index 0000000000..e98be9eaf0 --- /dev/null +++ b/data/2020/neurips/An implicit function learning approach for parametric modal regression @@ -0,0 +1 @@ +For multi-valued functions---such as when the conditional distribution on targets given the inputs is multi-modal---standard regression approaches are not always desirable because they provide the conditional mean. Modal regression algorithms address this issue by instead finding the conditional mode(s). Most, however, are nonparametric approaches and so can be difficult to scale. Further, parametric approximators, like neural networks, facilitate learning complex relationships between inputs and targets. In this work, we propose a parametric modal regression algorithm. We use the implicit function theorem to develop an objective, for learning a joint function over inputs and targets. We empirically demonstrate on several synthetic problems that our method (i) can learn multi-valued functions and produce the conditional modes, (ii) scales well to high-dimensional inputs, and (iii) can even be more effective for certain uni-modal problems, particularly for high-frequency functions. We demonstrate that our method is competitive in a real-world modal regression problem and two regular regression datasets. \ No newline at end of file diff --git a/data/2020/neurips/An operator view of policy gradient methods b/data/2020/neurips/An operator view of policy gradient methods new file mode 100644 index 0000000000..15614d0742 --- /dev/null +++ b/data/2020/neurips/An operator view of policy gradient methods @@ -0,0 +1 @@ +We cast policy gradient methods as the repeated application of two operators: a policy improvement operator $\mathcal{I}$, which maps any policy $\pi$ to a better one $\mathcal{I}\pi$, and a projection operator $\mathcal{P}$, which finds the best approximation of $\mathcal{I}\pi$ in the set of realizable policies. We use this framework to introduce operator-based versions of traditional policy gradient methods such as REINFORCE and PPO, which leads to a better understanding of their original counterparts. We also use the understanding we develop of the role of $\mathcal{I}$ and $\mathcal{P}$ to propose a new global lower bound of the expected return. This new perspective allows us to further bridge the gap between policy-based and value-based methods, showing how REINFORCE and the Bellman optimality operator, for example, can be seen as two sides of the same coin. \ No newline at end of file diff --git a/data/2020/neurips/Analysis and Design of Thompson Sampling for Stochastic Partial Monitoring b/data/2020/neurips/Analysis and Design of Thompson Sampling for Stochastic Partial Monitoring new file mode 100644 index 0000000000..751bcb2f8f --- /dev/null +++ b/data/2020/neurips/Analysis and Design of Thompson Sampling for Stochastic Partial Monitoring @@ -0,0 +1 @@ +We investigate finite stochastic partial monitoring, which is a general model for sequential learning with limited feedback. While Thompson sampling is one of the most promising algorithms on a variety of online decision-making problems, its properties for stochastic partial monitoring have not been theoretically investigated, and the existing algorithm relies on a heuristic approximation of the posterior distribution. To mitigate these problems, we present a novel Thompson-sampling-based algorithm, which enables us to exactly sample the target parameter from the posterior distribution. Besides, we prove that the new algorithm achieves the logarithmic problem-dependent expected pseudo-regret $\mathrm{O}(\log T)$ for a linearized variant of the problem with local observability. This result is the first regret bound of Thompson sampling for partial monitoring, which also becomes the first logarithmic regret bound of Thompson sampling for linear bandits. \ No newline at end of file diff --git a/data/2020/neurips/Analytic Characterization of the Hessian in Shallow ReLU Models: A Tale of Symmetry b/data/2020/neurips/Analytic Characterization of the Hessian in Shallow ReLU Models: A Tale of Symmetry new file mode 100644 index 0000000000..7433154446 --- /dev/null +++ b/data/2020/neurips/Analytic Characterization of the Hessian in Shallow ReLU Models: A Tale of Symmetry @@ -0,0 +1 @@ +We consider the optimization problem associated with fitting two-layers ReLU networks with respect to the squared loss, where labels are generated by a target network. We leverage the rich symmetry structure to analytically characterize the Hessian at various families of spurious minima in the natural regime where the number of inputs $d$ and the number of hidden neurons $k$ is finite. In particular, we prove that for $d\ge k$ standard Gaussian inputs: (a) of the $dk$ eigenvalues of the Hessian, $dk - O(d)$ concentrate near zero, (b) $\Omega(d)$ of the eigenvalues grow linearly with $k$. Although this phenomenon of extremely skewed spectrum has been observed many times before, to our knowledge, this is the first time it has been established {rigorously}. Our analytic approach uses techniques, new to the field, from symmetry breaking and representation theory, and carries important implications for our ability to argue about statistical generalization through local curvature. \ No newline at end of file diff --git a/data/2020/neurips/Analytical Probability Distributions and Exact Expectation-Maximization for Deep Generative Networks b/data/2020/neurips/Analytical Probability Distributions and Exact Expectation-Maximization for Deep Generative Networks new file mode 100644 index 0000000000..7403533fc9 --- /dev/null +++ b/data/2020/neurips/Analytical Probability Distributions and Exact Expectation-Maximization for Deep Generative Networks @@ -0,0 +1 @@ +Deep Generative Networks (DGNs) with probabilistic modeling of their output and latent space are currently trained via Variational Autoencoders (VAEs). In the absence of a known analytical form for the posterior and likelihood expectation, VAEs resort to approximations, including (Amortized) Variational Inference (AVI) and Monte-Carlo sampling. We exploit the Continuous Piecewise Affine property of modern DGNs to derive their posterior and marginal distributions as well as the latter’s first two moments. These findings enable us to derive an analytical Expectation-Maximization (EM) algorithm for gradient-free DGN learning. We demonstrate empirically that EM training of DGNs produces greater likelihood than VAE training. Our new framework will guide the design of new VAE AVI that better approximates the true posterior and open new avenues to apply standard statistical tools for model comparison, anomaly detection, and missing data imputation. \ No newline at end of file diff --git a/data/2020/neurips/Applications of Common Entropy for Causal Inference b/data/2020/neurips/Applications of Common Entropy for Causal Inference new file mode 100644 index 0000000000..9b5b4e1534 --- /dev/null +++ b/data/2020/neurips/Applications of Common Entropy for Causal Inference @@ -0,0 +1 @@ +We study the problem of discovering the simplest latent variable that can make two observed discrete variables conditionally independent. The minimum entropy required for such a latent is known as common entropy in information theory. We extend this notion to Rényi common entropy by minimizing the Rényi entropy of the latent variable. To efficiently compute common entropy, we propose an iterative algorithm that can be used to discover the trade-off between the entropy of the latent variable and the conditional mutual information of the observed variables. We show two applications of common entropy in causal inference: First, under the assumption that there are no low-entropy mediators, it can be used to distinguish causation from spurious correlation among almost all joint distributions on simple causal graphs with two observed variables. Second, common entropy can be used to improve constraint-based methods such as PC or FCI algorithms in the small-sample regime, where these methods are known to struggle. We propose a modification to these constraint-based methods to assess if a separating set found by these algorithms are valid using common entropy. We finally evaluate our algorithms on synthetic and real data to establish their performance. \ No newline at end of file diff --git a/data/2020/neurips/Approximate Cross-Validation for Structured Models b/data/2020/neurips/Approximate Cross-Validation for Structured Models new file mode 100644 index 0000000000..a37245a865 --- /dev/null +++ b/data/2020/neurips/Approximate Cross-Validation for Structured Models @@ -0,0 +1 @@ +Many modern data analyses benefit from explicitly modeling dependence structure in data -- such as measurements across time or space, ordered words in a sentence, or genes in a genome. Cross-validation is the gold standard to evaluate these analyses but can be prohibitively slow due to the need to re-run already-expensive learning algorithms many times. Previous work has shown approximate cross-validation (ACV) methods provide a fast and provably accurate alternative in the setting of empirical risk minimization. But this existing ACV work is restricted to simpler models by the assumptions that (i) data are independent and (ii) an exact initial model fit is available. In structured data analyses, (i) is always untrue, and (ii) is often untrue. In the present work, we address (i) by extending ACV to models with dependence structure. To address (ii), we verify -- both theoretically and empirically -- that ACV quality deteriorates smoothly with noise in the initial fit. We demonstrate the accuracy and computational benefits of our proposed methods on a diverse set of real-world applications. \ No newline at end of file diff --git a/data/2020/neurips/Approximate Cross-Validation with Low-Rank Data in High Dimensions b/data/2020/neurips/Approximate Cross-Validation with Low-Rank Data in High Dimensions new file mode 100644 index 0000000000..018a6ea8d8 --- /dev/null +++ b/data/2020/neurips/Approximate Cross-Validation with Low-Rank Data in High Dimensions @@ -0,0 +1 @@ +Many recent advances in machine learning are driven by a challenging trifecta: large data size $N$; high dimensions; and expensive algorithms. In this setting, cross-validation (CV) serves as an important tool for model assessment. Recent advances in approximate cross validation (ACV) provide accurate approximations to CV with only a single model fit, avoiding traditional CV's requirement for repeated runs of expensive algorithms. Unfortunately, these ACV methods can lose both speed and accuracy in high dimensions -- unless sparsity structure is present in the data. Fortunately, there is an alternative type of simplifying structure that is present in most data: approximate low rank (ALR). Guided by this observation, we develop a new algorithm for ACV that is fast and accurate in the presence of ALR data. Our first key insight is that the Hessian matrix -- whose inverse forms the computational bottleneck of existing ACV methods -- is ALR. We show that, despite our use of the \emph{inverse} Hessian, a low-rank approximation using the largest (rather than the smallest) matrix eigenvalues enables fast, reliable ACV. Our second key insight is that, in the presence of ALR data, error in existing ACV methods roughly grows with the (approximate, low) rank rather than with the (full, high) dimension. These insights allow us to prove theoretical guarantees on the quality of our proposed algorithm -- along with fast-to-compute upper bounds on its error. We demonstrate the speed and accuracy of our method, as well as the usefulness of our bounds, on a range of real and simulated data sets. \ No newline at end of file diff --git a/data/2020/neurips/Approximate Heavily-Constrained Learning with Lagrange Multiplier Models b/data/2020/neurips/Approximate Heavily-Constrained Learning with Lagrange Multiplier Models new file mode 100644 index 0000000000..79611fab0b --- /dev/null +++ b/data/2020/neurips/Approximate Heavily-Constrained Learning with Lagrange Multiplier Models @@ -0,0 +1 @@ +In machine learning applications such as ranking fairness or fairness over intersec-tional groups, one often encounters optimization problems with extremely large numbers of constraints. In particular, with ranking fairness tasks, there may even be a variable number of constraints, e.g. one for each query in the training set. In these cases, the standard approach of optimizing a Lagrangian while maintaining one Lagrange multiplier per constraint may no longer be practical. Our proposal is to associate a feature vector with each constraint, and to learn a “multiplier model” that maps each such vector to the corresponding Lagrange multiplier. We prove optimality, approximate feasibility and generalization guarantees under assumptions on the flexibility of the multiplier model, and empirically demonstrate that our method is effective on real-world case studies. \ No newline at end of file diff --git a/data/2020/neurips/Approximation Based Variance Reduction for Reparameterization Gradients b/data/2020/neurips/Approximation Based Variance Reduction for Reparameterization Gradients new file mode 100644 index 0000000000..2d39a90635 --- /dev/null +++ b/data/2020/neurips/Approximation Based Variance Reduction for Reparameterization Gradients @@ -0,0 +1 @@ +Flexible variational distributions improve variational inference but are harder to optimize. In this work we present a control variate that is applicable for any reparameterizable distribution with known mean and covariance matrix, e.g. Gaussians with any covariance structure. The control variate is based on a quadratic approximation of the model, and its parameters are set using a double-descent scheme by minimizing the gradient estimator's variance. We empirically show that this control variate leads to large improvements in gradient variance and optimization convergence for inference with non-factorized variational distributions. \ No newline at end of file diff --git a/data/2020/neurips/Assessing SATNet's Ability to Solve the Symbol Grounding Problem b/data/2020/neurips/Assessing SATNet's Ability to Solve the Symbol Grounding Problem new file mode 100644 index 0000000000..6c89330b3b --- /dev/null +++ b/data/2020/neurips/Assessing SATNet's Ability to Solve the Symbol Grounding Problem @@ -0,0 +1 @@ +SATNet is an award-winning MAXSAT solver that can be used to infer logical rules and integrated as a differentiable layer in a deep neural network. It had been shown to solve Sudoku puzzles visually from examples of puzzle digit images, and was heralded as an impressive achievement towards the longstanding AI goal of combining pattern recognition with logical reasoning. In this paper, we clarify SATNet's capabilities by showing that in the absence of intermediate labels that identify individual Sudoku digit images with their logical representations, SATNet completely fails at visual Sudoku (0% test accuracy). More generally, the failure can be pinpointed to its inability to learn to assign symbols to perceptual phenomena, also known as the symbol grounding problem, which has long been thought to be a prerequisite for intelligent agents to perform real-world logical reasoning. We propose an MNIST based test as an easy instance of the symbol grounding problem that can serve as a sanity check for differentiable symbolic solvers in general. Naive applications of SATNet on this test lead to performance worse than that of models without logical reasoning capabilities. We report on the causes of SATNet's failure and how to prevent them. \ No newline at end of file diff --git a/data/2020/neurips/Assisted Learning: A Framework for Multi-Organization Learning b/data/2020/neurips/Assisted Learning: A Framework for Multi-Organization Learning new file mode 100644 index 0000000000..6ffb067876 --- /dev/null +++ b/data/2020/neurips/Assisted Learning: A Framework for Multi-Organization Learning @@ -0,0 +1 @@ +In an increasing number of AI scenarios, collaborations among different organizations or agents (e.g., human and robots, mobile units) are often essential to accomplish an organization-specific mission. However, to avoid leaking useful and possibly proprietary information, organizations typically enforce stringent security constraints on sharing modeling algorithms and data, which significantly limits collaborations. In this work, we introduce the Assisted Learning framework for organizations to assist each other in supervised learning tasks without revealing any organization’s algorithm, data, or even task. An organization seeks assistance by broadcasting task-specific but nonsensitive statistics and incorporating others’ feedback in one or more iterations to eventually improve its predictive performance. Theoretical and experimental studies, including real-world medical benchmarks, show that Assisted Learning can often achieve near-oracle learning performance as if data and training processes were centralized. \ No newline at end of file diff --git a/data/2020/neurips/Asymmetric Shapley values: incorporating causal knowledge into model-agnostic explainability b/data/2020/neurips/Asymmetric Shapley values: incorporating causal knowledge into model-agnostic explainability new file mode 100644 index 0000000000..914f49fc5c --- /dev/null +++ b/data/2020/neurips/Asymmetric Shapley values: incorporating causal knowledge into model-agnostic explainability @@ -0,0 +1 @@ +Explaining AI systems is fundamental both to the development of high performing models and to the trust placed in them by their users. The Shapley framework for explainability has strength in its general applicability combined with its precise, rigorous foundation: it provides a common, model-agnostic language for AI explainability and uniquely satisfies a set of intuitive mathematical axioms. However, Shapley values are too restrictive in one significant regard: they ignore all causal structure in the data. We introduce a less restrictive framework, Asymmetric Shapley values (ASVs), which are rigorously founded on a set of axioms, applicable to any AI system, and flexible enough to incorporate any causal structure known to be respected by the data. We demonstrate that ASVs can (i) improve model explanations by incorporating causal information, (ii) provide an unambiguous test for unfair discrimination in model predictions, (iii) enable sequentially incremental explanations in time-series models, and (iv) support feature-selection studies without the need for model retraining. \ No newline at end of file diff --git a/data/2020/neurips/Asymptotic Guarantees for Generative Modeling Based on the Smooth Wasserstein Distance b/data/2020/neurips/Asymptotic Guarantees for Generative Modeling Based on the Smooth Wasserstein Distance new file mode 100644 index 0000000000..0d90c2e10e --- /dev/null +++ b/data/2020/neurips/Asymptotic Guarantees for Generative Modeling Based on the Smooth Wasserstein Distance @@ -0,0 +1 @@ +Minimum distance estimation (MDE) gained recent attention as a formulation of (implicit) generative modeling. It considers minimizing, over model parameters, a statistical distance between the empirical data distribution and the model. This formulation lends itself well to theoretical analysis, but typical results are hindered by the curse of dimensionality. To overcome this and devise a scalable finite-sample statistical MDE theory, we adopt the framework of smooth 1-Wasserstein distance (SWD) $\mathsf{W}_1^{(\sigma)}$. The SWD was recently shown to preserve the metric and topological structure of classic Wasserstein distances, while enjoying dimension-free empirical convergence rates. In this work, we conduct a thorough statistical study of the minimum smooth Wasserstein estimators (MSWEs), first proving the estimator's measurability and asymptotic consistency. We then characterize the limit distribution of the optimal model parameters and their associated minimal SWD. These results imply an $O(n^{-1/2})$ generalization bound for generative modeling based on MSWE, which holds in arbitrary dimension. Our main technical tool is a novel high-dimensional limit distribution result for empirical $\mathsf{W}_1^{(\sigma)}$. The characterization of a nondegenerate limit stands in sharp contrast with the classic empirical 1-Wasserstein distance, for which a similar result is known only in the one-dimensional case. The validity of our theory is supported by empirical results, posing the SWD as a potent tool for learning and inference in high dimensions. \ No newline at end of file diff --git a/data/2020/neurips/Asymptotic normality and confidence intervals for derivatives of 2-layers neural network in the random features model b/data/2020/neurips/Asymptotic normality and confidence intervals for derivatives of 2-layers neural network in the random features model new file mode 100644 index 0000000000..2400d6b56f --- /dev/null +++ b/data/2020/neurips/Asymptotic normality and confidence intervals for derivatives of 2-layers neural network in the random features model @@ -0,0 +1 @@ +This paper studies two-layers Neural Networks (NN), where the first layer contains random weights, and the second layer is trained using Ridge regularization. This model has been the focus of numerous recent works, showing that despite its simplicity, it captures some of the empirically observed behaviors of NN in the over-parametrized regime, such as the double-descent curve where the generalization error decreases as the number of weights increases to + ∞ . This paper establishes asymptotic distribution results for this 2-layers NN model in the regime where the ratios pn and dn have finite limits, where n is the sample size, p the ambient dimension and d is the width of the first layer. We show that a weighted average of the derivatives of the trained NN at the observed data is asymptotically normal, in a setting with Lipschitz activation functions in a linear regression response with Gaussian features under possibly non-linear perturbations. We then leverage this asymptotic normality result to construct confidence intervals (CIs) for single components of the unknown regression vector. The novelty of our results are threefold: (1) Despite the nonlinearity induced by the activation function, we characterize the asymptotic distribution of a weighted average of the gradients of the network after training; (2) It provides the first frequentist uncertainty quantification guarantees, in the form of valid ( 1 - α )-CIs, based on NN estimates; (3) It shows that the double-descent phenomenon occurs in terms of the length of the CIs, with the length increasing and then decreasing as d n (cid:37) + ∞ for certain fixed values of p n . We also provide a toolbox to predict the length of CIs numerically, which lets \ No newline at end of file diff --git a/data/2020/neurips/Asymptotically Optimal Exact Minibatch Metropolis-Hastings b/data/2020/neurips/Asymptotically Optimal Exact Minibatch Metropolis-Hastings new file mode 100644 index 0000000000..abe50e78c6 --- /dev/null +++ b/data/2020/neurips/Asymptotically Optimal Exact Minibatch Metropolis-Hastings @@ -0,0 +1 @@ +Metropolis-Hastings (MH) is a commonly-used MCMC algorithm, but it can be intractable on large datasets due to requiring computations over the whole dataset. In this paper, we study minibatch MH methods, which instead use subsamples to enable scaling. We observe that most existing minibatch MH methods are inexact (i.e. they may change the target distribution), and show that this inexactness can cause arbitrarily large errors in inference. We propose a new exact minibatch MH method, TunaMH, which exposes a tunable trade-off between its batch size and its theoretically guaranteed convergence rate. We prove a lower bound on the batch size that any minibatch MH method must use to retain exactness while guaranteeing fast convergence-the first such bound for minibatch MH-and show TunaMH is asymptotically optimal in terms of the batch size. Empirically, we show TunaMH outperforms other exact minibatch MH methods on robust linear regression, truncated Gaussian mixtures, and logistic regression. \ No newline at end of file diff --git a/data/2020/neurips/Attack of the Tails: Yes, You Really Can Backdoor Federated Learning b/data/2020/neurips/Attack of the Tails: Yes, You Really Can Backdoor Federated Learning new file mode 100644 index 0000000000..a67672c37c --- /dev/null +++ b/data/2020/neurips/Attack of the Tails: Yes, You Really Can Backdoor Federated Learning @@ -0,0 +1 @@ +Due to its decentralized nature, Federated Learning (FL) lends itself to adversarial attacks in the form of backdoors during training. The goal of a backdoor is to corrupt the performance of the trained model on specific sub-tasks (e.g., by classifying green cars as frogs). A range of FL backdoor attacks have been introduced in the literature, but also methods to defend against them, and it is currently an open question whether FL systems can be tailored to be robust against backdoors. In this work, we provide evidence to the contrary. We first establish that, in the general case, robustness to backdoors implies model robustness to adversarial examples, a major open problem in itself. Furthermore, detecting the presence of a backdoor in a FL model is unlikely assuming first order oracles or polynomial time. We couple our theoretical results with a new family of backdoor attacks, which we refer to as edge-case backdoors. An edge-case backdoor forces a model to misclassify on seemingly easy inputs that are however unlikely to be part of the training, or test data, i.e., they live on the tail of the input distribution. We explain how these edge-case backdoors can lead to unsavory failures and may have serious repercussions on fairness, and exhibit that with careful tuning at the side of the adversary, one can insert them across a range of machine learning tasks (e.g., image classification, OCR, text prediction, sentiment analysis). \ No newline at end of file diff --git a/data/2020/neurips/AttendLight: Universal Attention-Based Reinforcement Learning Model for Traffic Signal Control b/data/2020/neurips/AttendLight: Universal Attention-Based Reinforcement Learning Model for Traffic Signal Control new file mode 100644 index 0000000000..7a74667bba --- /dev/null +++ b/data/2020/neurips/AttendLight: Universal Attention-Based Reinforcement Learning Model for Traffic Signal Control @@ -0,0 +1 @@ +We propose AttendLight, an end-to-end Reinforcement Learning (RL) algorithm for the problem of traffic signal control. Previous approaches for this problem have the shortcoming that they require training for each new intersection with a different structure or traffic flow distribution. AttendLight solves this issue by training a single, universal model for intersections with any number of roads, lanes, phases (possible signals), and traffic flow. To this end, we propose a deep RL model which incorporates two attention models. The first attention model is introduced to handle different numbers of roads-lanes; and the second attention model is intended for enabling decision-making with any number of phases in an intersection. As a result, our proposed model works for any intersection configuration, as long as a similar configuration is represented in the training set. Experiments were conducted with both synthetic and real-world standard benchmark data-sets. The results we show cover intersections with three or four approaching roads; one-directional/bi-directional roads with one, two, and three lanes; different number of phases; and different traffic flows. We consider two regimes: (i) single-environment training, single-deployment, and (ii) multi-environment training, multi-deployment. AttendLight outperforms both classical and other RL-based approaches on all cases in both regimes. \ No newline at end of file diff --git a/data/2020/neurips/Attention-Gated Brain Propagation: How the brain can implement reward-based error backpropagation b/data/2020/neurips/Attention-Gated Brain Propagation: How the brain can implement reward-based error backpropagation new file mode 100644 index 0000000000..8988cabb49 --- /dev/null +++ b/data/2020/neurips/Attention-Gated Brain Propagation: How the brain can implement reward-based error backpropagation @@ -0,0 +1 @@ +Much recent work has focused on biologically plausible variants of supervised learning algorithms. However, there is no teacher in the motor cortex that instructs the motor neurons and learning in the brain depends on reward and punishment. We demonstrate a biologically plausible reinforcement learning scheme for deep networks with an arbitrary number of layers. The network chooses an action by selecting a unit in the output layer and uses feedback connections to assign credit to the units in successively lower layers that are responsible for this action. After the choice, the network receives reinforcement and there is no teacher correcting the errors. We show how the new learning scheme – Attention-Gated Brain Propagation (BrainProp) – is mathematically equivalent to error backpropagation, for one output unit at a time. We demonstrate successful learning of deep fully connected, convolutional and locally connected networks on classical and hard image-classification benchmarks; MNIST, CIFAR10, CIFAR100 and Tiny ImageNet. BrainProp achieves an accuracy that is equivalent to that of standard error-backpropagation, and better than state-of-the-art biologically inspired learning schemes. Additionally, the trial-and-error nature of learning is associated with limited additional training time so that BrainProp is a factor of 1-3.5 times slower. Our results thereby provide new insights into how deep learning may be implemented in the brain. \ No newline at end of file diff --git a/data/2020/neurips/Attribute Prototype Network for Zero-Shot Learning b/data/2020/neurips/Attribute Prototype Network for Zero-Shot Learning new file mode 100644 index 0000000000..2d8950f443 --- /dev/null +++ b/data/2020/neurips/Attribute Prototype Network for Zero-Shot Learning @@ -0,0 +1 @@ +From the beginning of zero-shot learning research, visual attributes have been shown to play an important role. In order to better transfer attribute-based knowledge from known to unknown classes, we argue that an image representation with integrated attribute localization ability would be beneficial for zero-shot learning. To this end, we propose a novel zero-shot representation learning framework that jointly learns discriminative global and local features using only class-level attributes. While a visual-semantic embedding layer learns global features, local features are learned through an attribute prototype network that simultaneously regresses and decorrelates attributes from intermediate features. We show that our locality augmented image representations achieve a new state-of-the-art on three zero-shot learning benchmarks. As an additional benefit, our model points to the visual evidence of the attributes in an image, e.g. for the CUB dataset, confirming the improved attribute localization ability of our image representation. The code will be publicaly available at this https URL. \ No newline at end of file diff --git a/data/2020/neurips/Attribution Preservation in Network Compression for Reliable Network Interpretation b/data/2020/neurips/Attribution Preservation in Network Compression for Reliable Network Interpretation new file mode 100644 index 0000000000..33a2817f21 --- /dev/null +++ b/data/2020/neurips/Attribution Preservation in Network Compression for Reliable Network Interpretation @@ -0,0 +1 @@ +Neural networks embedded in safety-sensitive applications such as self-driving cars and wearable health monitors rely on two important techniques: input attribution for hindsight analysis and network compression to reduce its size for edge-computing. In this paper, we show that these seemingly unrelated techniques conflict with each other as network compression deforms the produced attributions, which could lead to dire consequences for mission-critical applications. This phenomenon arises due to the fact that conventional network compression methods only preserve the predictions of the network while ignoring the quality of the attributions. To combat the attribution inconsistency problem, we present a framework that can preserve the attributions while compressing a network. By employing the Weighted Collapsed Attribution Matching regularizer, we match the attribution maps of the network being compressed to its pre-compression former self. We demonstrate the effectiveness of our algorithm both quantitatively and qualitatively on diverse compression methods. \ No newline at end of file diff --git a/data/2020/neurips/Audeo: Audio Generation for a Silent Performance Video b/data/2020/neurips/Audeo: Audio Generation for a Silent Performance Video new file mode 100644 index 0000000000..ae179ca362 --- /dev/null +++ b/data/2020/neurips/Audeo: Audio Generation for a Silent Performance Video @@ -0,0 +1 @@ +We present a novel system that gets as an input video frames of a musician playing the piano and generates the music for that video. Generation of music from visual cues is a challenging problem and it is not clear whether it is an attainable goal at all. Our main aim in this work is to explore the plausibility of such a transformation and to identify cues and components able to carry the association of sounds with visual events. To achieve the transformation we built a full pipeline named `\textit{Audeo}' containing three components. We first translate the video frames of the keyboard and the musician hand movements into raw mechanical musical symbolic representation Piano-Roll (Roll) for each video frame which represents the keys pressed at each time step. We then adapt the Roll to be amenable for audio synthesis by including temporal correlations. This step turns out to be critical for meaningful audio generation. As a last step, we implement Midi synthesizers to generate realistic music. \textit{Audeo} converts video to audio smoothly and clearly with only a few setup constraints. We evaluate \textit{Audeo} on `in the wild' piano performance videos and obtain that their generated music is of reasonable audio quality and can be successfully recognized with high precision by popular music identification software. \ No newline at end of file diff --git a/data/2020/neurips/Auditing Differentially Private Machine Learning: How Private is Private SGD? b/data/2020/neurips/Auditing Differentially Private Machine Learning: How Private is Private SGD? new file mode 100644 index 0000000000..0e4e576be6 --- /dev/null +++ b/data/2020/neurips/Auditing Differentially Private Machine Learning: How Private is Private SGD? @@ -0,0 +1 @@ +We investigate whether Differentially Private SGD offers better privacy in practice than what is guaranteed by its state-of-the-art analysis. We do so via novel data poisoning attacks, which we show correspond to realistic privacy attacks. While previous work (Ma et al., arXiv 2019) proposed this connection between differential privacy and data poisoning as a defense against data poisoning, our use as a tool for understanding the privacy of a specific mechanism is new. More generally, our work takes a quantitative, empirical approach to understanding the privacy afforded by specific implementations of differentially private algorithms that we believe has the potential to complement and influence analytical work on differential privacy. \ No newline at end of file diff --git a/data/2020/neurips/Auto Learning Attention b/data/2020/neurips/Auto Learning Attention new file mode 100644 index 0000000000..abcb7692a0 --- /dev/null +++ b/data/2020/neurips/Auto Learning Attention @@ -0,0 +1 @@ +Attention modules have been demonstrated effective in strengthening the representation ability of a neural network via reweighting spatial or channel features or stacking both operations sequentially. However, designing the structures of different attention operations requires a bulk of computation and extensive expertise. In this paper, we devise an Auto Learning Attention (AutoLA) method, which is the first attempt on automatic attention design. Specifically, we define a novel attention module named high order group attention (HOGA) as a directed acyclic graph (DAG) where each group represents a node, and each edge represents an operation of heterogeneous attentions. A typical HOGA architecture can be searched automatically via the differential AutoLA method within 1 GPU day using the ResNet-20 backbone on CIFAR10. Further, the searched attention module can generalize to various backbones as a plug-and-play component and outperforms popular manually designed channel and spatial attentions for many vision tasks, including image classification on CIFAR100 and ImageNet, object detection and human keypoint detection on COCO dataset. Code is available at https://github.com/btma48/AutoLA. \ No newline at end of file diff --git a/data/2020/neurips/Auto-Panoptic: Cooperative Multi-Component Architecture Search for Panoptic Segmentation b/data/2020/neurips/Auto-Panoptic: Cooperative Multi-Component Architecture Search for Panoptic Segmentation new file mode 100644 index 0000000000..b92bfe1f5f --- /dev/null +++ b/data/2020/neurips/Auto-Panoptic: Cooperative Multi-Component Architecture Search for Panoptic Segmentation @@ -0,0 +1 @@ +Panoptic segmentation is posed as a new popular test-bed for the state-of-the-art holistic scene understanding methods with the requirement of simultaneously segmenting both foreground things and background stuff. The state-of-the-art panoptic segmentation network exhibits high structural complexity in different network components, i.e. backbone, proposal-based foreground branch, segmentation-based background branch, and feature fusion module across branches, which heavily relies on expert knowledge and tedious trials. In this work, we propose an efficient, cooperative and highly automated framework to simultaneously search for all main components including backbone, segmentation branches, and feature fusion module in a unified panoptic segmentation pipeline based on the prevailing one-shot Network Architecture Search (NAS) paradigm. Notably, we extend the common single-task NAS into the multi-component scenario by taking the advantage of the newly proposed intra-modular search space and problem-oriented inter-modular search space, which helps us to obtain an optimal network architecture that not only performs well in both instance segmentation and semantic segmentation tasks but also be aware of the reciprocal relations between foreground things and background stuff classes. To relieve the vast computation burden incurred by applying NAS to complicated network architectures, we present a novel path-priority greedy search policy to find a robust, transferrable architecture with significantly reduced searching overhead. Our searched architecture, namely Auto-Panoptic, achieves the new state-of-the-art on the challenging COCO and ADE20K benchmarks. Moreover, extensive experiments are conducted to demonstrate the effectiveness of path-priority policy and transferability of Auto-Panoptic across different datasets. Codes and models are available at: this https URL. \ No newline at end of file diff --git a/data/2020/neurips/AutoBSS: An Efficient Algorithm for Block Stacking Style Search b/data/2020/neurips/AutoBSS: An Efficient Algorithm for Block Stacking Style Search new file mode 100644 index 0000000000..4cf02543c5 --- /dev/null +++ b/data/2020/neurips/AutoBSS: An Efficient Algorithm for Block Stacking Style Search @@ -0,0 +1 @@ +Neural network architecture design mostly focuses on the new convolutional operator or special topological structure of network block, little attention is drawn to the configuration of stacking each block, called Block Stacking Style (BSS). Recent studies show that BSS may also have an unneglectable impact on networks, thus we design an efficient algorithm to search it automatically. The proposed method, AutoBSS, is a novel AutoML algorithm based on Bayesian optimization by iteratively refining and clustering Block Stacking Style Code (BSSC), which can find optimal BSS in a few trials without biased evaluation. On ImageNet classification task, ResNet50/MobileNetV2/EfficientNet-B0 with our searched BSS achieve 79.29%/74.5%/77.79%, which outperform the original baselines by a large margin. More importantly, experimental results on model compression, object detection and instance segmentation show the strong generalizability of the proposed AutoBSS, and further verify the unneglectable impact of BSS on neural networks. \ No newline at end of file diff --git a/data/2020/neurips/AutoPrivacy: Automated Layer-wise Parameter Selection for Secure Neural Network Inference b/data/2020/neurips/AutoPrivacy: Automated Layer-wise Parameter Selection for Secure Neural Network Inference new file mode 100644 index 0000000000..5f7a9958c3 --- /dev/null +++ b/data/2020/neurips/AutoPrivacy: Automated Layer-wise Parameter Selection for Secure Neural Network Inference @@ -0,0 +1,2 @@ +Hybrid Privacy-Preserving Neural Network (HPPNN) implementing linear layers by Homomorphic Encryption (HE) and nonlinear layers by Garbled Circuit (GC) is one of the most promising secure solutions to emerging Machine Learning as a Service (MLaaS). Unfortunately, a HPPNN suffers from long inference latency, e.g., $\sim100$ seconds per image, which makes MLaaS unsatisfactory. Because HE-based linear layers of a HPPNN cost $93\%$ inference latency, it is critical to select a set of HE parameters to minimize computational overhead of linear layers. Prior HPPNNs over-pessimistically select huge HE parameters to maintain large noise budgets, since they use the same set of HE parameters for an entire network and ignore the error tolerance capability of a network. +In this paper, for fast and accurate secure neural network inference, we propose an automated layer-wise parameter selector, AutoPrivacy, that leverages deep reinforcement learning to automatically determine a set of HE parameters for each linear layer in a HPPNN. The learning-based HE parameter selection policy outperforms conventional rule-based HE parameter selection policy. Compared to prior HPPNNs, AutoPrivacy-optimized HPPNNs reduce inference latency by $53\%\sim70\%$ with negligible loss of accuracy. \ No newline at end of file diff --git a/data/2020/neurips/AutoSync: Learning to Synchronize for Data-Parallel Distributed Deep Learning b/data/2020/neurips/AutoSync: Learning to Synchronize for Data-Parallel Distributed Deep Learning new file mode 100644 index 0000000000..4fc84ec271 --- /dev/null +++ b/data/2020/neurips/AutoSync: Learning to Synchronize for Data-Parallel Distributed Deep Learning @@ -0,0 +1 @@ +The rationale behind Eq. 1 is as follows: (1) Since many runtime systems (e.g. TensorFlow [1] or PyTorch [3]) introduce scheduling or parallelization between communication and computation, in practice, there are significant overlaps between the two components; (2) in data-parallel training, it is commonly observed that one component usually dominates the other [4]. These make using the maximum of them as the estimation of the total time reasonable. \ No newline at end of file diff --git a/data/2020/neurips/Autoencoders that don't overfit towards the Identity b/data/2020/neurips/Autoencoders that don't overfit towards the Identity new file mode 100644 index 0000000000..fc902473ca --- /dev/null +++ b/data/2020/neurips/Autoencoders that don't overfit towards the Identity @@ -0,0 +1 @@ +Autoencoders (AE) aim to reproduce the output from the input. They may hence tend to overfit towards learning the identity-function between the input and output, i.e., they may predict each feature in the output from itself in the input. This is not useful, however, when AEs are used for prediction tasks in the presence of noise in the data. It may seem intuitively evident that this kind of overfitting is prevented by training a denoising AE [36], as the dropped-out features have to be predicted from the other features. In this paper, we consider linear autoencoders, as they facilitate analytic solutions, and first show that denoising / dropout actually prevents the overfitting towards the identity-function only to the degree that it is penalized by the induced L2-norm regularization. In the main theorem of this paper, we show that the emphasized denoising AE [37] is indeed capable of completely eliminating the overfitting towards the identity-function. Our derivations reveal several new insights, including the closed-form solution of the full-rank model, as well as a new (near-)orthogonality constraint in the low-rank model. While this constraint is conceptually very different from the regularizers recently proposed in [11, 42, 14], their resulting effects on the learned embeddings are empirically similar. Our experiments on three well-known data-sets corroborate the various theoretical insights derived in this paper. \ No newline at end of file diff --git a/data/2020/neurips/Autofocused oracles for model-based design b/data/2020/neurips/Autofocused oracles for model-based design new file mode 100644 index 0000000000..37033823a4 --- /dev/null +++ b/data/2020/neurips/Autofocused oracles for model-based design @@ -0,0 +1 @@ +Data-driven design is making headway into a number of application areas, including protein, small-molecule, and materials engineering. The design goal is to construct an object with desired properties, such as a protein that binds to a target more tightly than previously observed. To that end, costly experimental measurements are being replaced with calls to a high-capacity regression model trained on labeled data, which can be leveraged in an in silico search for promising design candidates. However, the design goal necessitates moving into regions of the input space beyond where such models were trained. Therefore, one can ask: should the regression model be altered as the design algorithm explores the input space, in the absence of new data acquisition? Herein, we answer this question in the affirmative. In particular, we (i) formalize the data-driven design problem as a non-zero-sum game, (ii) leverage this formalism to develop a strategy for retraining the regression model as the design algorithm proceeds---what we refer to as autofocusing the model, and (iii) demonstrate the promise of autofocusing empirically. \ No newline at end of file diff --git a/data/2020/neurips/Automatic Curriculum Learning through Value Disagreement b/data/2020/neurips/Automatic Curriculum Learning through Value Disagreement new file mode 100644 index 0000000000..36b8dfc0d9 --- /dev/null +++ b/data/2020/neurips/Automatic Curriculum Learning through Value Disagreement @@ -0,0 +1 @@ +Continually solving new, unsolved tasks is the key to learning diverse behaviors. Through reinforcement learning (RL), we have made massive strides towards solving tasks that have a single goal. However, in the multi-task domain, where an agent needs to reach multiple goals, the choice of training goals can largely affect sample efficiency. When biological agents learn, there is often an organized and meaningful order to which learning happens. Inspired by this, we propose setting up an automatic curriculum for goals that the agent needs to solve. Our key insight is that if we can sample goals at the frontier of the set of goals that an agent is able to reach, it will provide a significantly stronger learning signal compared to randomly sampled goals. To operationalize this idea, we introduce a goal proposal module that prioritizes goals that maximize the epistemic uncertainty of the Q-function of the policy. This simple technique samples goals that are neither too hard nor too easy for the agent to solve, hence enabling continual improvement. We evaluate our method across 13 multi-goal robotic tasks and 5 navigation tasks, and demonstrate performance gains over current state-of-the-art methods. \ No newline at end of file diff --git a/data/2020/neurips/Automatic Perturbation Analysis for Scalable Certified Robustness and Beyond b/data/2020/neurips/Automatic Perturbation Analysis for Scalable Certified Robustness and Beyond new file mode 100644 index 0000000000..98d4f72395 --- /dev/null +++ b/data/2020/neurips/Automatic Perturbation Analysis for Scalable Certified Robustness and Beyond @@ -0,0 +1 @@ +Linear relaxation based perturbation analysis (LiRPA) for neural networks, which computes provable linear bounds of output neurons given a certain amount of input perturbation, has become a core component in robustness verification and certified defense. The majority of LiRPA-based methods focus on simple feed-forward networks and need particular manual derivations and implementations when extended to other architectures. In this paper, we develop an automatic framework to enable perturbation analysis on any neural network structures, by generalizing existing LiRPA algorithms such as CROWN to operate on general computational graphs. The flexibility, differentiability and ease of use of our framework allow us to obtain state-of-the-art results on LiRPA based certified defense on fairly complicated networks like DenseNet, ResNeXt and Transformer that are not supported by prior works. Our framework also enables loss fusion, a technique that significantly reduces the computational complexity of LiRPA for certified defense. For the first time, we demonstrate LiRPA based certified defense on Tiny ImageNet and Downscaled ImageNet where previous approaches cannot scale to due to the relatively large number of classes. Our work also yields an open-source library for the community to apply LiRPA to areas beyond certified defense without much LiRPA expertise, e.g., we create a neural network with a probably flat optimization landscape by applying LiRPA to network parameters. Our opensource library is available at https://github.com/KaidiXu/auto_LiRPA. \ No newline at end of file diff --git a/data/2020/neurips/Automatically Learning Compact Quality-aware Surrogates for Optimization Problems b/data/2020/neurips/Automatically Learning Compact Quality-aware Surrogates for Optimization Problems new file mode 100644 index 0000000000..0330d86bb4 --- /dev/null +++ b/data/2020/neurips/Automatically Learning Compact Quality-aware Surrogates for Optimization Problems @@ -0,0 +1 @@ +Solving optimization problems with unknown parameters often requires learning a predictive model to predict the values of the unknown parameters and then solving the problem using these values. Recent work has shown that including the optimization problem as a layer in the model training pipeline results in predictions of the unobserved parameters that lead to higher decision quality. Unfortunately, this process comes at a large computational cost because the optimization problem must be solved and differentiated through in each training iteration; furthermore, it may also sometimes fail to improve solution quality due to non-smoothness issues that arise when training through a complex optimization layer. To address these shortcomings, we learn a low-dimensional surrogate model of a large optimization problem by representing the feasible space in terms of meta-variables, each of which is a linear combination of the original variables. By training a low-dimensional surrogate model end-to-end, and jointly with the predictive model, we achieve: i) a large reduction in training and inference time; and ii) improved performance by focusing attention on the more important variables in the optimization and learning in a smoother space. Empirically, we demonstrate these improvements on a non-convex adversary modeling task, a submodular recommendation task and a convex portfolio optimization task. \ No newline at end of file diff --git a/data/2020/neurips/Autoregressive Score Matching b/data/2020/neurips/Autoregressive Score Matching new file mode 100644 index 0000000000..8fe77f7e67 --- /dev/null +++ b/data/2020/neurips/Autoregressive Score Matching @@ -0,0 +1 @@ +Autoregressive models use chain rule to define a joint probability distribution as a product of conditionals. These conditionals need to be normalized, imposing constraints on the functional families that can be used. To increase flexibility, we propose autoregressive conditional score models (AR-CSM) where we parameterize the joint distribution in terms of the derivatives of univariate log-conditionals (scores), which need not be normalized. To train AR-CSM, we introduce a new divergence between distributions named Composite Score Matching (CSM). For AR-CSM models, this divergence between data and model distributions can be computed and optimized efficiently, requiring no expensive sampling or adversarial training. Compared to previous score matching algorithms, our method is more scalable to high dimensional data and more stable to optimize. We show with extensive experimental results that it can be applied to density estimation on synthetic data, image generation, image denoising, and training latent variable models with implicit encoders. \ No newline at end of file diff --git a/data/2020/neurips/Auxiliary Task Reweighting for Minimum-data Learning b/data/2020/neurips/Auxiliary Task Reweighting for Minimum-data Learning new file mode 100644 index 0000000000..1cd3fbbeb9 --- /dev/null +++ b/data/2020/neurips/Auxiliary Task Reweighting for Minimum-data Learning @@ -0,0 +1 @@ +Supervised learning requires a large amount of training data, limiting its application where labeled data is scarce. To compensate for data scarcity, one possible method is to utilize auxiliary tasks to provide additional supervision for the main task. Assigning and optimizing the importance weights for different auxiliary tasks remains an crucial and largely understudied research question. In this work, we propose a method to automatically reweight auxiliary tasks in order to reduce the data requirement on the main task. Specifically, we formulate the weighted likelihood function of auxiliary tasks as a surrogate prior for the main task. By adjusting the auxiliary task weights to minimize the divergence between the surrogate prior and the true prior of the main task, we obtain a more accurate prior estimation, achieving the goal of minimizing the required amount of training data for the main task and avoiding a costly grid search. In multiple experimental settings (e.g. semi-supervised learning, multi-label classification), we demonstrate that our algorithm can effectively utilize limited labeled data of the main task with the benefit of auxiliary tasks compared with previous task reweighting methods. We also show that under extreme cases with only a few extra examples (e.g. few-shot domain adaptation), our algorithm results in significant improvement over the baseline. \ No newline at end of file diff --git a/data/2020/neurips/AvE: Assistance via Empowerment b/data/2020/neurips/AvE: Assistance via Empowerment new file mode 100644 index 0000000000..45d7349bae --- /dev/null +++ b/data/2020/neurips/AvE: Assistance via Empowerment @@ -0,0 +1 @@ +One difficulty in using artificial agents for human-assistive applications lies in the challenge of accurately assisting with a person's goal(s). Existing methods tend to rely on inferring the human's goal, which is challenging when there are many potential goals or when the set of candidate goals is difficult to identify. We propose a new paradigm for assistance by instead increasing the human's ability to control their environment, and formalize this approach by augmenting reinforcement learning with human empowerment. This task-agnostic objective preserves the person's autonomy and ability to achieve any eventual state. We test our approach against assistance based on goal inference, highlighting scenarios where our method overcomes failure modes stemming from goal ambiguity or misspecification. As existing methods for estimating empowerment in continuous domains are computationally hard, precluding its use in real time learned assistance, we also propose an efficient empowerment-inspired proxy metric. Using this, we are able to successfully demonstrate our method in a shared autonomy user study for a challenging simulated teleoperation task with human-in-the-loop training. \ No newline at end of file diff --git a/data/2020/neurips/Avoiding Side Effects By Considering Future Tasks b/data/2020/neurips/Avoiding Side Effects By Considering Future Tasks new file mode 100644 index 0000000000..4946141472 --- /dev/null +++ b/data/2020/neurips/Avoiding Side Effects By Considering Future Tasks @@ -0,0 +1 @@ +Designing reward functions is difficult: the designer has to specify what to do (what it means to complete the task) as well as what not to do (side effects that should be avoided while completing the task). To alleviate the burden on the reward designer, we propose an algorithm to automatically generate an auxiliary reward function that penalizes side effects. This auxiliary objective rewards the ability to complete possible future tasks, which decreases if the agent causes side effects during the current task. The future task reward can also give the agent an incentive to interfere with events in the environment that make future tasks less achievable, such as irreversible actions by other agents. To avoid this interference incentive, we introduce a baseline policy that represents a default course of action (such as doing nothing), and use it to filter out future tasks that are not achievable by default. We formally define interference incentives and show that the future task approach with a baseline policy avoids these incentives in the deterministic case. Using gridworld environments that test for side effects and interference, we show that our method avoids interference and is more effective for avoiding side effects than the common approach of penalizing irreversible actions. \ No newline at end of file diff --git a/data/2020/neurips/Avoiding Side Effects in Complex Environments b/data/2020/neurips/Avoiding Side Effects in Complex Environments new file mode 100644 index 0000000000..1784bffefa --- /dev/null +++ b/data/2020/neurips/Avoiding Side Effects in Complex Environments @@ -0,0 +1 @@ +Reward function specification can be difficult, even in simple environments. Realistic environments contain millions of states. Rewarding the agent for making a widget may be easy, but penalizing the multitude of possible negative side effects is hard. In toy environments, Attainable Utility Preservation (AUP) avoids side effects by penalizing shifts in the ability to achieve randomly generated goals. We scale this approach to large, randomly generated environments based on Conway's Game of Life. By preserving optimal value for a single randomly generated reward function, AUP incurs modest overhead, completes the specified task, and avoids side effects. \ No newline at end of file diff --git a/data/2020/neurips/Axioms for Learning from Pairwise Comparisons b/data/2020/neurips/Axioms for Learning from Pairwise Comparisons new file mode 100644 index 0000000000..449b5aa73d --- /dev/null +++ b/data/2020/neurips/Axioms for Learning from Pairwise Comparisons @@ -0,0 +1 @@ +To be well-behaved, systems that process preference data must satisfy certain conditions identified by economic decision theory and by social choice theory. In ML, preferences and rankings are commonly learned by fitting a probabilistic model to noisy preference data. The behavior of this learning process from the view of economic theory has previously been studied for the case where the data consists of rankings. In practice, it is more common to have only pairwise comparison data, and the formal properties of the associated learning problem are more challenging to analyze. We show that a large class of random utility models (including the Thurstone–Mosteller Model), when estimated using the MLE, satisfy a Pareto efficiency condition. These models also satisfy a strong monotonicity property, which implies that the learning process is responsive to input data. On the other hand, we show that these models fail certain other consistency conditions from social choice theory, and in particular do not always follow the majority opinion. Our results inform existing and future applications of random utility models for societal decision making. \ No newline at end of file diff --git a/data/2020/neurips/BAIL: Best-Action Imitation Learning for Batch Deep Reinforcement Learning b/data/2020/neurips/BAIL: Best-Action Imitation Learning for Batch Deep Reinforcement Learning new file mode 100644 index 0000000000..72758ad5ff --- /dev/null +++ b/data/2020/neurips/BAIL: Best-Action Imitation Learning for Batch Deep Reinforcement Learning @@ -0,0 +1 @@ +There has recently been a surge in research in batch Deep Reinforcement Learning (DRL), which aims for learning a high-performing policy from a given dataset without additional interactions with the environment. We propose a new algorithm, Best-Action Imitation Learning (BAIL), which strives for both simplicity and performance. BAIL learns a V function, uses the V function to select actions it believes to be high-performing, and then uses those actions to train a policy network using imitation learning. For the MuJoCo benchmark, we provide a comprehensive experimental study of BAIL, comparing its performance to four other batch Q-learning and imitation-learning schemes for a large variety of batch datasets. Our experiments show that BAIL's performance is much higher than the other schemes, and is also computationally much faster than the batch Q-learning schemes. \ No newline at end of file diff --git a/data/2020/neurips/BERT Loses Patience: Fast and Robust Inference with Early Exit b/data/2020/neurips/BERT Loses Patience: Fast and Robust Inference with Early Exit new file mode 100644 index 0000000000..55e4ad560d --- /dev/null +++ b/data/2020/neurips/BERT Loses Patience: Fast and Robust Inference with Early Exit @@ -0,0 +1 @@ +In this paper, we propose Patience-based Early Exit, a straightforward yet effective inference method that can be used as a plug-and-play technique to simultaneously improve the efficiency and robustness of a pretrained language model (PLM). To achieve this, our approach couples an internal-classifier with each layer of a PLM and dynamically stops inference when the intermediate predictions of the internal classifiers remain unchanged for a pre-defined number of steps. Our approach improves inference efficiency as it allows the model to make a prediction with fewer layers. Meanwhile, experimental results with an ALBERT model show that our method can improve the accuracy and robustness of the model by preventing it from overthinking and exploiting multiple classifiers for prediction, yielding a better accuracy-speed trade-off compared to existing early exit methods. \ No newline at end of file diff --git a/data/2020/neurips/BOSS: Bayesian Optimization over String Spaces b/data/2020/neurips/BOSS: Bayesian Optimization over String Spaces new file mode 100644 index 0000000000..6f14831994 --- /dev/null +++ b/data/2020/neurips/BOSS: Bayesian Optimization over String Spaces @@ -0,0 +1 @@ +This article develops a Bayesian optimization (BO) method which acts directly over raw strings, proposing the first uses of string kernels and genetic algorithms within BO loops. Recent applications of BO over strings have been hindered by the need to map inputs into a smooth and unconstrained latent space. Learning this projection is computationally and data-intensive. Our approach instead builds a powerful Gaussian process surrogate model based on string kernels, naturally supporting variable length inputs, and performs efficient acquisition function maximization for spaces with syntactical constraints. Experiments demonstrate considerably improved optimization over existing approaches across a broad range of constraints, including the popular setting where syntax is governed by a context-free grammar. \ No newline at end of file diff --git a/data/2020/neurips/BRP-NAS: Prediction-based NAS using GCNs b/data/2020/neurips/BRP-NAS: Prediction-based NAS using GCNs new file mode 100644 index 0000000000..b8049b09ef --- /dev/null +++ b/data/2020/neurips/BRP-NAS: Prediction-based NAS using GCNs @@ -0,0 +1 @@ +Neural architecture search (NAS) enables researchers to automatically explore broad design spaces in order to improve efficiency of neural networks. This efficiency is especially important in the case of on-device deployment, where improvements in accuracy should be balanced out with computational demands of a model. In practice, performance metrics of model are computationally expensive to obtain. Previous work uses a proxy (e.g., number of operations) or a layer-wise measurement of neural network layers to estimate end-to-end hardware performance but the imprecise prediction diminishes the quality of NAS. To address this problem, we propose BRP-NAS, an efficient hardware-aware NAS enabled by an accurate performance predictor-based on graph convolutional network (GCN). What is more, we investigate prediction quality on different metrics and show that sample efficiency of the predictor-based NAS can be improved by considering binary relations of models and an iterative data selection strategy. We show that our proposed method outperforms all prior methods on NAS-Bench-101, NAS-Bench-201 and DARTS. Finally, to raise awareness of the fact that accurate latency estimation is not a trivial task, we release LatBench -- a latency dataset of NAS-Bench-201 models running on a broad range of devices. \ No newline at end of file diff --git a/data/2020/neurips/Backpropagating Linearly Improves Transferability of Adversarial Examples b/data/2020/neurips/Backpropagating Linearly Improves Transferability of Adversarial Examples new file mode 100644 index 0000000000..e8c22b3c0d --- /dev/null +++ b/data/2020/neurips/Backpropagating Linearly Improves Transferability of Adversarial Examples @@ -0,0 +1 @@ +The vulnerability of deep neural networks (DNNs) to adversarial examples has drawn great attention from the community. In this paper, we study the transferability of such examples, which lays the foundation of many black-box attacks on DNNs. We revisit a not so new but definitely noteworthy hypothesis of Goodfellow et al.'s and disclose that the transferability can be enhanced by improving the linearity of DNNs in an appropriate manner. We introduce linear backpropagation (LinBP), a method that performs backpropagation in a more linear fashion using off-the-shelf attacks that exploit gradients. More specifically, it calculates forward as normal but backpropagates loss as if some nonlinear activations are not encountered in the forward pass. Experimental results demonstrate that this simple yet effective method obviously outperforms current state-of-the-arts in crafting transferable adversarial examples on CIFAR-10 and ImageNet, leading to more effective attacks on a variety of DNNs. \ No newline at end of file diff --git a/data/2020/neurips/Bad Global Minima Exist and SGD Can Reach Them b/data/2020/neurips/Bad Global Minima Exist and SGD Can Reach Them new file mode 100644 index 0000000000..16c06cdd60 --- /dev/null +++ b/data/2020/neurips/Bad Global Minima Exist and SGD Can Reach Them @@ -0,0 +1 @@ +Several recent works have aimed to explain why severely overparameterized models, generalize well when trained by Stochastic Gradient Descent (SGD). The emergent consensus explanation has two parts: the first is that there are "no bad local minima", while the second is that SGD performs implicit regularization by having a bias towards low complexity models. We revisit both of these ideas in the context of image classification with common deep neural network architectures. Our first finding is that there exist bad global minima, i.e., models that fit the training set perfectly, yet have poor generalization. Our second finding is that given only unlabeled training data, we can easily construct initializations that will cause SGD to quickly converge to such bad global minima. For example, on CIFAR, CINIC10, and (Restricted) ImageNet, this can be achieved by starting SGD at a model derived by fitting random labels on the training data: while subsequent SGD training (with the correct labels) will reach zero training error, the resulting model will exhibit a test accuracy degradation of up to 40% compared to training from a random initialization. Finally, we show that regularization seems to provide SGD with an escape route: once heuristics such as data augmentation are used, starting from a complex model (adversarial initialization) has no effect on the test accuracy. \ No newline at end of file diff --git a/data/2020/neurips/Balanced Meta-Softmax for Long-Tailed Visual Recognition b/data/2020/neurips/Balanced Meta-Softmax for Long-Tailed Visual Recognition new file mode 100644 index 0000000000..b8207000b0 --- /dev/null +++ b/data/2020/neurips/Balanced Meta-Softmax for Long-Tailed Visual Recognition @@ -0,0 +1 @@ +Deep classifiers have achieved great success in visual recognition. However, real-world data is long-tailed by nature, leading to the mismatch between training and testing distributions. In this paper, we show that Softmax function, though used in most classification tasks, gives a biased gradient estimation under the long-tailed setup. This paper presents Balanced Softmax, an elegant unbiased extension of Softmax, to accommodate the label distribution shift between training and testing. Theoretically, we derive the generalization bound for multiclass Softmax regression and show our loss minimizes the bound. In addition, we introduce Balanced Meta-Softmax, applying a complementary Meta Sampler to estimate the optimal class sample rate and further improve long-tailed learning. In our experiments, we demonstrate that Balanced Meta-Softmax outperforms state-of-the-art long-tailed classification solutions on both visual recognition and instance segmentation tasks. \ No newline at end of file diff --git a/data/2020/neurips/Bandit Linear Control b/data/2020/neurips/Bandit Linear Control new file mode 100644 index 0000000000..a3e44ee833 --- /dev/null +++ b/data/2020/neurips/Bandit Linear Control @@ -0,0 +1 @@ +We consider the problem of controlling a known linear dynamical system under stochastic noise, adversarially chosen costs, and bandit feedback. Unlike the full feedback setting where the entire cost function is revealed after each decision, here only the cost incurred by the learner is observed. We present a new and efficient algorithm that, for strongly convex and smooth costs, obtains regret that grows with the square root of the time horizon $T$. We also give extensions of this result to general convex, possibly non-smooth costs, and to non-stochastic system noise. A key component of our algorithm is a new technique for addressing bandit optimization of loss functions with memory. \ No newline at end of file diff --git a/data/2020/neurips/Bandit Samplers for Training Graph Neural Networks b/data/2020/neurips/Bandit Samplers for Training Graph Neural Networks new file mode 100644 index 0000000000..9351f3e39d --- /dev/null +++ b/data/2020/neurips/Bandit Samplers for Training Graph Neural Networks @@ -0,0 +1 @@ +Several sampling algorithms with variance reduction have been proposed for accelerating the training of Graph Convolution Networks (GCNs). However, due to the intractable computation of optimal sampling distribution, these sampling algorithms are suboptimal for GCNs and are not applicable to more general graph neural networks (GNNs) where the message aggregator contains learned weights rather than fixed weights, such as Graph Attention Networks (GAT). The fundamental reason is that the embeddings of the neighbors or learned weights involved in the optimal sampling distribution are changing during the training and not known a priori, but only partially observed when sampled, thus making the derivation of an optimal variance reduced samplers non-trivial. In this paper, we formulate the optimization of the sampling variance as an adversary bandit problem, where the rewards are related to the node embeddings and learned weights, and can vary constantly. Thus a good sampler needs to acquire variance information about more neighbors (exploration) while at the same time optimizing the immediate sampling variance (exploit). We theoretically show that our algorithm asymptotically approaches the optimal variance within a factor of 3. We show the efficiency and effectiveness of our approach on multiple datasets. \ No newline at end of file diff --git a/data/2020/neurips/BanditPAM: Almost Linear Time k-Medoids Clustering via Multi-Armed Bandits b/data/2020/neurips/BanditPAM: Almost Linear Time k-Medoids Clustering via Multi-Armed Bandits new file mode 100644 index 0000000000..2e0573e3b9 --- /dev/null +++ b/data/2020/neurips/BanditPAM: Almost Linear Time k-Medoids Clustering via Multi-Armed Bandits @@ -0,0 +1 @@ +Clustering is a ubiquitous task in data science. Compared to the commonly used $k$-means clustering algorithm, $k$-medoids clustering algorithms require the cluster centers to be actual data points and support arbitrary distance metrics, allowing for greater interpretability and the clustering of structured objects. Current state-of-the-art $k$-medoids clustering algorithms, such as Partitioning Around Medoids (PAM), are iterative and are quadratic in the dataset size $n$ for each iteration, being prohibitively expensive for large datasets. We propose Bandit-PAM, a randomized algorithm inspired by techniques from multi-armed bandits, that significantly improves the computational efficiency of PAM. We theoretically prove that Bandit-PAM reduces the complexity of each PAM iteration from $O(n^2)$ to $O(n \log n)$ and returns the same results with high probability, under assumptions on the data that often hold in practice. We empirically validate our results on several large-scale real-world datasets, including a coding exercise submissions dataset from this http URL, the 10x Genomics 68k PBMC single-cell RNA sequencing dataset, and the MNIST handwritten digits dataset. We observe that Bandit-PAM returns the same results as PAM while performing up to 200x fewer distance computations. The improvements demonstrated by Bandit-PAM enable $k$-medoids clustering on a wide range of applications, including identifying cell types in large-scale single-cell data and providing scalable feedback for students learning computer science online. We also release Python and C++ implementations of our algorithm. \ No newline at end of file diff --git a/data/2020/neurips/Barking up the right tree: an approach to search over molecule synthesis DAGs b/data/2020/neurips/Barking up the right tree: an approach to search over molecule synthesis DAGs new file mode 100644 index 0000000000..a78880772c --- /dev/null +++ b/data/2020/neurips/Barking up the right tree: an approach to search over molecule synthesis DAGs @@ -0,0 +1 @@ +When designing new molecules with particular properties, it is not only important what to make but crucially how to make it. These instructions form a synthesis directed acyclic graph (DAG), describing how a large vocabulary of simple building blocks can be recursively combined through chemical reactions to create more complicated molecules of interest. In contrast, many current deep generative models for molecules ignore synthesizability. We therefore propose a deep generative model that better represents the real world process, by directly outputting molecule synthesis DAGs. We argue that this provides sensible inductive biases, ensuring that our model searches over the same chemical space that chemists would also have access to, as well as interpretability. We show that our approach is able to model chemical space well, producing a wide range of diverse molecules, and allows for unconstrained optimization of an inherently constrained problem: maximize certain chemical properties such that discovered molecules are synthesizable. \ No newline at end of file diff --git a/data/2020/neurips/Batch normalization provably avoids ranks collapse for randomly initialised deep networks b/data/2020/neurips/Batch normalization provably avoids ranks collapse for randomly initialised deep networks new file mode 100644 index 0000000000..e5a68ca9c5 --- /dev/null +++ b/data/2020/neurips/Batch normalization provably avoids ranks collapse for randomly initialised deep networks @@ -0,0 +1 @@ +Randomly initialized neural networks are known to become harder to train with increasing depth, unless architectural enhancements like residual connections and batch normalization are used. We here investigate this phenomenon by revisiting the connection between random initialization in deep networks and spectral instabilities in products of random matrices. Given the rich literature on random matrices, it is not surprising to find that the rank of the intermediate representations in unnormalized networks collapses quickly with depth. In this work we highlight the fact that batch normalization is an effective strategy to avoid rank collapse for both linear and ReLU networks. Leveraging tools from Markov chain theory, we derive a meaningful lower rank bound in deep linear networks. Empirically, we also demonstrate that this rank robustness generalizes to ReLU nets. Finally, we conduct an extensive set of experiments on real-world data sets, which confirm that rank stability is indeed a crucial condition for training modern-day deep neural architectures. \ No newline at end of file diff --git a/data/2020/neurips/Batched Coarse Ranking in Multi-Armed Bandits b/data/2020/neurips/Batched Coarse Ranking in Multi-Armed Bandits new file mode 100644 index 0000000000..528c66c483 --- /dev/null +++ b/data/2020/neurips/Batched Coarse Ranking in Multi-Armed Bandits @@ -0,0 +1 @@ +We study the problem of coarse ranking in the multi-armed bandits (MAB) setting, where we have a set of arms each of which is associated with an unknown distribution. The task is to partition the arms into clusters of predefined sizes, such that the mean of any arm in the i -th cluster is larger than that of any arm in the j -th cluster for any j > i . Coarse ranking generalizes a number of basic problems in MAB (e.g., best arm identification) and has many real-world applications. We initiate the study of the problem in the batched model where we can only have a small number of policy changes. We study both the fixed budget and fixed confidence variants in MAB, and propose algorithms and prove impossibility results which together give almost tight tradeoffs between the total number of arms pulls and the number of policy changes. We have tested our algorithms in both real and synthetic data; our experimental results have demonstrated the efficiency of the proposed methods. \ No newline at end of file diff --git a/data/2020/neurips/Baxter Permutation Process b/data/2020/neurips/Baxter Permutation Process new file mode 100644 index 0000000000..8d8f475483 --- /dev/null +++ b/data/2020/neurips/Baxter Permutation Process @@ -0,0 +1 @@ +In this paper, a Bayesian nonparametric (BNP) model for Baxter permutations (BPs), termed BP process (BPP) is proposed and applied to relational data analysis. The BPs are a well-studied class of permutations, and it has been demonstrated that there is one-to-one correspondence between BPs and several interesting objects including floorplan partitioning (FP), which constitutes a subset of rectangular partitioning (RP). Accordingly, the BPP can be used as an FP model. We combine the BPP with a multi-dimensional extension of the stick-breaking process called the block-breaking process to fill the gap between FP and RP, and obtain a stochastic process on arbitrary RPs. Compared with conventional BNP models for arbitrary RPs, the proposed model is simpler and has a high affinity with Bayesian inference. \ No newline at end of file diff --git a/data/2020/neurips/BayReL: Bayesian Relational Learning for Multi-omics Data Integration b/data/2020/neurips/BayReL: Bayesian Relational Learning for Multi-omics Data Integration new file mode 100644 index 0000000000..ec3a69c446 --- /dev/null +++ b/data/2020/neurips/BayReL: Bayesian Relational Learning for Multi-omics Data Integration @@ -0,0 +1 @@ +High-throughput molecular profiling technologies have produced high-dimensional multi-omics data, enabling systematic understanding of living systems at the genome scale. Studying molecular interactions across different data types helps reveal signal transduction mechanisms across different classes of molecules. In this paper, we develop a novel Bayesian representation learning method that infers the relational interactions across multi-omics data types. Our method, Bayesian Relational Learning (BayReL) for multi-omics data integration, takes advantage of a priori known relationships among the same class of molecules, modeled as a graph at each corresponding view, to learn view-specific latent variables as well as a multi-partite graph that encodes the interactions across views. Our experiments on several real-world datasets demonstrate enhanced performance of BayReL in inferring meaningful interactions compared to existing baselines. \ No newline at end of file diff --git a/data/2020/neurips/Bayes Consistency vs. H-Consistency: The Interplay between Surrogate Loss Functions and the Scoring Function Class b/data/2020/neurips/Bayes Consistency vs. H-Consistency: The Interplay between Surrogate Loss Functions and the Scoring Function Class new file mode 100644 index 0000000000..dbbeb4ec33 --- /dev/null +++ b/data/2020/neurips/Bayes Consistency vs. H-Consistency: The Interplay between Surrogate Loss Functions and the Scoring Function Class @@ -0,0 +1 @@ +A fundamental question in multiclass classification concerns understanding the consistency properties of surrogate risk minimization algorithms, which minimize a (often convex) surrogate to the multiclass 0-1 loss. In particular, the framework of calibrated surrogates has played an important role in analyzing Bayes consistency of such algorithms, i.e. in studying convergence to a Bayes optimal classifier (Zhang, 2004; Tewari and Bartlett, 2007). However, follow-up work has suggested this framework can be of limited value when studying H -consistency; in particular, concerns have been raised that even when the data comes from an underlying linear model, minimizing certain convex calibrated surrogates over linear scoring functions fails to recover the true model (Long and Servedio, 2013). In this paper, we investigate this apparent conundrum. We find that while some calibrated surrogates can indeed fail to provide H -consistency when minimized over a naturallooking but naïvely chosen scoring function class F , the situation can potentially be remedied by minimizing them over a more carefully chosen class of scoring functions F . In particular, for the popular one-vs-all hinge and logistic surrogates, both of which are calibrated (and therefore provide Bayes consistency) under realizable models, but were previously shown to pose problems for realizable H -consistency, we derive a form of scoring function class F that enables Hconsistency. When H is the class of linear models, the class F consists of certain piecewise linear scoring functions that are characterized by the same number of parameters as in the linear case, and minimization over which can be performed using an adaptation of the min-pooling idea from neural network training. Our experiments confirm that the one-vs-all surrogates, when trained over this class of nonlinear scoring functions F , yield better linear multiclass classifiers than when trained over standard linear scoring functions. \ No newline at end of file diff --git a/data/2020/neurips/Bayesian Attention Modules b/data/2020/neurips/Bayesian Attention Modules new file mode 100644 index 0000000000..d38d3467a9 --- /dev/null +++ b/data/2020/neurips/Bayesian Attention Modules @@ -0,0 +1 @@ +Attention modules, as simple and effective tools, have not only enabled deep neural networks to achieve state-of-the-art results in many domains, but also enhanced their interpretability. Most current models use deterministic attention modules due to their simplicity and ease of optimization. Stochastic counterparts, on the other hand, are less popular despite their potential benefits. The main reason is that stochastic attention often introduces optimization issues or requires significant model changes. In this paper, we propose a scalable stochastic version of attention that is easy to implement and optimize. We construct simplex-constrained attention distributions by normalizing reparameterizable distributions, making the training process differentiable. We learn their parameters in a Bayesian framework where a data-dependent prior is introduced for regularization. We apply the proposed stochastic attention modules to various attention-based models, with applications to graph node classification, visual question answering, image captioning, machine translation, and language understanding. Our experiments show the proposed method brings consistent improvements over the corresponding baselines. \ No newline at end of file diff --git a/data/2020/neurips/Bayesian Bits: Unifying Quantization and Pruning b/data/2020/neurips/Bayesian Bits: Unifying Quantization and Pruning new file mode 100644 index 0000000000..0ba8c6140b --- /dev/null +++ b/data/2020/neurips/Bayesian Bits: Unifying Quantization and Pruning @@ -0,0 +1 @@ +We introduce Bayesian Bits, a practical method for joint mixed precision quantization and pruning through gradient based optimization. Bayesian Bits employs a novel decomposition of the quantization operation, which sequentially considers doubling the bit width. At each new bit width, the residual error between the full precision value and the previously rounded value is quantized. We then decide whether or not to add this quantized residual error for a higher effective bit width and lower quantization noise. By starting with a power-of-two bit width, this decomposition will always produce hardware-friendly configurations, and through an additional 0-bit option, serves as a unified view of pruning and quantization. Bayesian Bits then introduces learnable stochastic gates, which collectively control the bit width of the given tensor. As a result, we can obtain low bit solutions by performing approximate inference over the gates, with prior distributions that encourage most of them to be switched off. We experimentally validate our proposed method on several benchmark datasets and show that we can learn pruned, mixed precision networks that provide a better trade-off between accuracy and efficiency than their static bit width equivalents. \ No newline at end of file diff --git a/data/2020/neurips/Bayesian Causal Structural Learning with Zero-Inflated Poisson Bayesian Networks b/data/2020/neurips/Bayesian Causal Structural Learning with Zero-Inflated Poisson Bayesian Networks new file mode 100644 index 0000000000..5f25b2d03a --- /dev/null +++ b/data/2020/neurips/Bayesian Causal Structural Learning with Zero-Inflated Poisson Bayesian Networks @@ -0,0 +1 @@ +Without loss of generality, we assume that the nodes are labeled such that there is no directed edge in E from later node to earlier node. Such labeling is also known as perfect/topological ordering of DAG G. Define a set of edges that are connected to node j and have opposite directions in E and E′, re(j) = {k ∈ V : ejk = ekj = 1} for j = 1, . . . , p. For k ∈ re(j), E includes an edge k → j, while E′ has the reverse edge j → k. If re(j) = ∅ for all j, there exists no pair of nodes (j, k) such that ejk = ekj = 1. This means E = E ′, because Markov equivalent DAGs have the same skeleton. We will show by mathematical induction that re(j) = ∅ for all j, which contradicts the assumption that E 6= E′. For node p that is the last element of the perfect ordering of G, we have pa(p) = pa′(p) ∪ re(p) due to the same skeleton of G and G′. Taking the difference of the equality (1) at (x1, x2, . . . , xp + 1) and (x1, x2, . . . , xp) yields, ∑ \ No newline at end of file diff --git a/data/2020/neurips/Bayesian Deep Ensembles via the Neural Tangent Kernel b/data/2020/neurips/Bayesian Deep Ensembles via the Neural Tangent Kernel new file mode 100644 index 0000000000..18e260eb2b --- /dev/null +++ b/data/2020/neurips/Bayesian Deep Ensembles via the Neural Tangent Kernel @@ -0,0 +1 @@ +We explore the link between deep ensembles and Gaussian processes (GPs) through the lens of the Neural Tangent Kernel (NTK): a recent development in understanding the training dynamics of wide neural networks (NNs). Previous work has shown that even in the infinite width limit, when NNs become GPs, there is no GP posterior interpretation to a deep ensemble trained with squared error loss. We introduce a simple modification to standard deep ensembles training, through addition of a computationally-tractable, randomised and untrainable function to each ensemble member, that enables a posterior interpretation in the infinite width limit. When ensembled together, our trained NNs give an approximation to a posterior predictive distribution, and we prove that our Bayesian deep ensembles make more conservative predictions than standard deep ensembles in the infinite width limit. Finally, using finite width NNs we demonstrate that our Bayesian deep ensembles faithfully emulate the analytic posterior predictive when available, and can outperform standard deep ensembles in various out-of-distribution settings, for both regression and classification tasks. \ No newline at end of file diff --git a/data/2020/neurips/Bayesian Deep Learning and a Probabilistic Perspective of Generalization b/data/2020/neurips/Bayesian Deep Learning and a Probabilistic Perspective of Generalization new file mode 100644 index 0000000000..9aa70b2259 --- /dev/null +++ b/data/2020/neurips/Bayesian Deep Learning and a Probabilistic Perspective of Generalization @@ -0,0 +1 @@ +The key distinguishing property of a Bayesian approach is marginalization, rather than using a single setting of weights. Bayesian marginalization can particularly improve the accuracy and calibration of modern deep neural networks, which are typically underspecified by the data, and can represent many compelling but different solutions. We show that deep ensembles provide an effective mechanism for approximate Bayesian marginalization, and propose a related approach that further improves the predictive distribution by marginalizing within basins of attraction, without significant overhead. We also investigate the prior over functions implied by a vague distribution over neural network weights, explaining the generalization properties of such models from a probabilistic perspective. From this perspective, we explain results that have been presented as mysterious and distinct to neural network generalization, such as the ability to fit images with random labels, and show that these results can be reproduced with Gaussian processes. We also show that Bayesian model averaging alleviates double descent, resulting in monotonic performance improvements with increased flexibility. Finally, we provide a Bayesian perspective on tempering for calibrating predictive distributions. \ No newline at end of file diff --git a/data/2020/neurips/Bayesian Meta-Learning for the Few-Shot Setting via Deep Kernels b/data/2020/neurips/Bayesian Meta-Learning for the Few-Shot Setting via Deep Kernels new file mode 100644 index 0000000000..e604001fad --- /dev/null +++ b/data/2020/neurips/Bayesian Meta-Learning for the Few-Shot Setting via Deep Kernels @@ -0,0 +1 @@ +Recently, different machine learning methods have been introduced to tackle the challenging few-shot learning scenario that is, learning from a small labeled dataset related to a specific task. Common approaches have taken the form of meta-learning: learning to learn on the new problem given the old. Following the recognition that meta-learning is implementing learning in a multi-level model, we present a Bayesian treatment for the meta-learning inner loop through the use of deep kernels. As a result we can learn a kernel that transfers to new tasks; we call this Deep Kernel Transfer (DKT). This approach has many advantages: is straightforward to implement as a single optimizer, provides uncertainty quantification, and does not require estimation of task-specific parameters. We empirically demonstrate that DKT outperforms several state-of-the-art algorithms in few-shot classification, and is the state of the art for cross-domain adaptation and regression. We conclude that complex meta-learning routines can be replaced by a simpler Bayesian model without loss of accuracy. \ No newline at end of file diff --git a/data/2020/neurips/Bayesian Multi-type Mean Field Multi-agent Imitation Learning b/data/2020/neurips/Bayesian Multi-type Mean Field Multi-agent Imitation Learning new file mode 100644 index 0000000000..50ec68911a --- /dev/null +++ b/data/2020/neurips/Bayesian Multi-type Mean Field Multi-agent Imitation Learning @@ -0,0 +1 @@ +Multi-agent Imitation learning (MAIL) refers to the problem that agents learn to perform a task interactively in a multi-agent system through observing and mimicking expert demonstrations, without any knowledge of a reward function from the environment. MAIL has received a lot of attention due to promising results achieved on synthesized tasks, with the potential to be applied to complex real-world multi-agent tasks. Key challenges for MAIL include sample efficiency and scalability. In this paper, we proposed Bayesian multi-type mean field multi-agent imitation learning (BM3IL). Our method improves sample efficiency through establishing a Bayesian formulation for MAIL, and enhances scalability through introducing a new multi-type mean field approximation. We demonstrate the performance of our algorithm through benchmarking with three state-of-the-art multi-agent imitation learning algorithms on several tasks, including solving a multi-agent traffic optimization problem in a real-world transportation network. Experimental results indicate that our algorithm significantly outperforms all other algorithms in all scenarios. \ No newline at end of file diff --git a/data/2020/neurips/Bayesian Optimization for Iterative Learning b/data/2020/neurips/Bayesian Optimization for Iterative Learning new file mode 100644 index 0000000000..72d904d4d6 --- /dev/null +++ b/data/2020/neurips/Bayesian Optimization for Iterative Learning @@ -0,0 +1 @@ +The performance of deep (reinforcement) learning systems crucially depends on the choice of hyperparameters. Their tuning is notoriously expensive, typically requiring an iterative training process to run for numerous steps to convergence. Traditional tuning algorithms only consider the final performance of hyperparameters acquired after many expensive iterations and ignore intermediate information from earlier training steps. In this paper, we present a Bayesian optimization(BO) approach which exploits the iterative structure of learning algorithms for efficient hyperparameter tuning. We propose to learn an evaluation function compressing learning progress at any stage of the training process into a single numeric score according to both training success and stability. Our BO framework is then tradeoff the benefit of assessing a hyperparameter setting over additional training steps against their computation cost. We further increase model efficiency by selectively including scores from different training steps for any evaluated hyperparameter set. We demonstrate the efficiency of our algorithm by tuning hyperparameters for the training of deep reinforcement learning agents and convolutional neural networks. Our algorithm outperforms all existing baselines in identifying optimal hyperparameters in minimal time. \ No newline at end of file diff --git a/data/2020/neurips/Bayesian Optimization of Risk Measures b/data/2020/neurips/Bayesian Optimization of Risk Measures new file mode 100644 index 0000000000..049076f655 --- /dev/null +++ b/data/2020/neurips/Bayesian Optimization of Risk Measures @@ -0,0 +1 @@ +In this study, we propose a novel multi-objective Bayesian optimization (MOBO) method to efficiently identify the Pareto front (PF) defined by risk measures for black-box functions under the presence of input uncertainty (IU). Existing BO methods for Pareto optimization in the presence of IU are risk-specific or without theoretical guarantees, whereas our proposed method addresses general risk measures and has theoretical guarantees. The basic idea of the proposed method is to assume a Gaussian process (GP) model for the black-box function and to construct high-probability bounding boxes for the risk measures using the GP model. Furthermore, in order to reduce the uncertainty of non-dominated bounding boxes, we propose a method of selecting the next evaluation point using a maximin distance defined by the maximum value of a quasi distance based on bounding boxes. As theoretical analysis, we prove that the algorithm can return an arbitrary-accurate solution in a finite number of iterations with high probability, for various risk measures such as Bayes risk, worst-case risk, and value-at-risk. We also give a theoretical analysis that takes into account approximation errors because there exist non-negligible approximation errors (e.g., finite approximation of PFs and sampling-based approximation of bounding boxes) in practice. We confirm that the proposed method outperforms compared with existing methods not only in the setting with IU but also in the setting of ordinary MOBO through numerical experiments. \ No newline at end of file diff --git a/data/2020/neurips/Bayesian Probabilistic Numerical Integration with Tree-Based Models b/data/2020/neurips/Bayesian Probabilistic Numerical Integration with Tree-Based Models new file mode 100644 index 0000000000..67f59c7045 --- /dev/null +++ b/data/2020/neurips/Bayesian Probabilistic Numerical Integration with Tree-Based Models @@ -0,0 +1 @@ +Bayesian quadrature (BQ) is a method for solving numerical integration problems in a Bayesian manner, which allows user to quantify their uncertainty about the solution. The standard approach to BQ is based on Gaussian process (GP) approximation of the integrand. As a result, BQ approach is inherently limited to cases where GP approximations can be done in an efficient manner, thus often prohibiting high-dimensional or non-smooth target functions. This paper proposes to tackle this issue with a new Bayesian numerical integration algorithm based on Bayesian Additive Regression Trees (BART) priors, which we call BART-Int. BART priors are easy to tune and well-suited for discontinuous functions. We demonstrate that they also lend themselves naturally to a sequential design setting and that explicit convergence rates can be obtained in a variety of settings. The advantages and disadvantages of this new methodology are highlighted on a set of benchmark tests including the Genz functions, and on a Bayesian survey design problem. \ No newline at end of file diff --git a/data/2020/neurips/Bayesian Pseudocoresets b/data/2020/neurips/Bayesian Pseudocoresets new file mode 100644 index 0000000000..60b4abe7cc --- /dev/null +++ b/data/2020/neurips/Bayesian Pseudocoresets @@ -0,0 +1 @@ +A Bayesian pseudocoreset is a small synthetic dataset for which the posterior over parameters approximates that of the original dataset. While promising, the scalability of Bayesian pseudocoresets is not yet validated in realistic problems such as image classification with deep neural networks. On the other hand, dataset distillation methods similarly construct a small dataset such that the optimization using the synthetic dataset converges to a solution with performance competitive with optimization using full data. Although dataset distillation has been empirically verified in large-scale settings, the framework is restricted to point estimates, and their adaptation to Bayesian inference has not been explored. This paper casts two representative dataset distillation algorithms as approximations to methods for constructing pseudocoresets by minimizing specific divergence measures: reverse KL divergence and Wasserstein distance. Furthermore, we provide a unifying view of such divergence measures in Bayesian pseudocoreset construction. Finally, we propose a novel Bayesian pseudocoreset algorithm based on minimizing forward KL divergence. Our empirical results demonstrate that the pseudocoresets constructed from these methods reflect the true posterior even in high-dimensional Bayesian inference problems. \ No newline at end of file diff --git a/data/2020/neurips/Bayesian Robust Optimization for Imitation Learning b/data/2020/neurips/Bayesian Robust Optimization for Imitation Learning new file mode 100644 index 0000000000..ecb2daa314 --- /dev/null +++ b/data/2020/neurips/Bayesian Robust Optimization for Imitation Learning @@ -0,0 +1 @@ +One of the main challenges in imitation learning is determining what action an agent should take when outside the state distribution of the demonstrations. Inverse reinforcement learning (IRL) can enable generalization to new states by learning a parameterized reward function, but these approaches still face uncertainty over the true reward function and corresponding optimal policy. Existing safe imitation learning approaches based on IRL deal with this uncertainty using a maxmin framework that optimizes a policy under the assumption of an adversarial reward function, whereas risk-neutral IRL approaches either optimize a policy for the mean or MAP reward function. While completely ignoring risk can lead to overly aggressive and unsafe policies, optimizing in a fully adversarial sense is also problematic as it can lead to overly conservative policies that perform poorly in practice. To provide a bridge between these two extremes, we propose Bayesian Robust Optimization for Imitation Learning (BROIL). BROIL leverages Bayesian reward function inference and a user specific risk tolerance to efficiently optimize a robust policy that balances expected return and conditional value at risk. Our empirical results show that BROIL provides a natural way to interpolate between return-maximizing and risk-minimizing behaviors and outperforms existing risk-sensitive and risk-neutral inverse reinforcement learning algorithms. \ No newline at end of file diff --git a/data/2020/neurips/Bayesian filtering unifies adaptive and non-adaptive neural network optimization methods b/data/2020/neurips/Bayesian filtering unifies adaptive and non-adaptive neural network optimization methods new file mode 100644 index 0000000000..d4d709d367 --- /dev/null +++ b/data/2020/neurips/Bayesian filtering unifies adaptive and non-adaptive neural network optimization methods @@ -0,0 +1 @@ +We formulate the problem of neural network optimization as Bayesian filtering, where the observations are the backpropagated gradients. While neural network optimization has previously been studied using natural gradient methods which are closely related to Bayesian inference, they were unable to recover standard optimizers such as Adam and RMSprop with a root-mean-square gradient normalizer, instead getting a mean-square normalizer. To recover the root-mean-square normalizer, we find it necessary to account for the temporal dynamics of all the other parameters as they are geing optimized. The resulting optimizer, AdaBayes, adaptively transitions between SGD-like and Adam-like behaviour, automatically recovers AdamW, a state of the art variant of Adam with decoupled weight decay, and has generalisation performance competitive with SGD. \ No newline at end of file diff --git a/data/2020/neurips/Belief Propagation Neural Networks b/data/2020/neurips/Belief Propagation Neural Networks new file mode 100644 index 0000000000..35f14f9fe0 --- /dev/null +++ b/data/2020/neurips/Belief Propagation Neural Networks @@ -0,0 +1 @@ +Learned neural solvers have successfully been used to solve combinatorial optimization and decision problems. More general counting variants of these problems, however, are still largely solved with hand-crafted solvers. To bridge this gap, we introduce belief propagation neural networks (BPNNs), a class of parameterized operators that operate on factor graphs and generalize Belief Propagation (BP). In its strictest form, a BPNN layer (BPNN-D) is a learned iterative operator that provably maintains many of the desirable properties of BP for any choice of the parameters. Empirically, we show that by training BPNN-D learns to perform the task better than the original BP: it converges 1.7x faster on Ising models while providing tighter bounds. On challenging model counting problems, BPNNs compute estimates 100's of times faster than state-of-the-art handcrafted methods, while returning an estimate of comparable quality. \ No newline at end of file diff --git a/data/2020/neurips/Belief-Dependent Macro-Action Discovery in POMDPs using the Value of Information b/data/2020/neurips/Belief-Dependent Macro-Action Discovery in POMDPs using the Value of Information new file mode 100644 index 0000000000..fd8ac0f21b --- /dev/null +++ b/data/2020/neurips/Belief-Dependent Macro-Action Discovery in POMDPs using the Value of Information @@ -0,0 +1 @@ +This work introduces macro-action discovery using value-of-information (VoI) for robust and efficient planning in partially observable Markov decision processes (POMDPs). POMDPs are a powerful framework for planning under uncertainty. Previous approaches have used high-level macro-actions within POMDP policies to reduce planning complexity. However, macro-action design is often heuristic and rarely comes with performance guarantees. Here, we present a method for extracting belief-dependent, variable-length macro-actions directly from a low-level POMDP model. We construct macro-actions by chaining sequences of open-loop actions together when the task-specific value of information (VoI) — the change in expected task performance caused by observations in the current planning iteration — is low. Importantly, we provide performance guarantees on the resulting VoI macro-action policies in the form of bounded regret relative to the optimal policy. In simulated tracking experiments, we achieve higher reward than both closed-loop and hand-coded macro-action baselines, selectively using VoI macro-actions to reduce planning complexity while maintaining near-optimal task performance. \ No newline at end of file diff --git a/data/2020/neurips/Benchmarking Deep Inverse Models over time, and the Neural-Adjoint method b/data/2020/neurips/Benchmarking Deep Inverse Models over time, and the Neural-Adjoint method new file mode 100644 index 0000000000..618827afa2 --- /dev/null +++ b/data/2020/neurips/Benchmarking Deep Inverse Models over time, and the Neural-Adjoint method @@ -0,0 +1 @@ +We consider the task of solving generic inverse problems, where one wishes to determine the hidden parameters of a natural system that will give rise to a particular set of measurements. Recently many new approaches based upon deep learning have arisen generating impressive results. We conceptualize these models as different schemes for efficiently, but randomly, exploring the space of possible inverse solutions. As a result, the accuracy of each approach should be evaluated as a function of time rather than a single estimated solution, as is often done now. Using this metric, we compare several state-of-the-art inverse modeling approaches on four benchmark tasks: two existing tasks, one simple task for visualization and one new task from metamaterial design. Finally, inspired by our conception of the inverse problem, we explore a solution that uses a deep learning model to approximate the forward model, and then uses backpropagation to search for good inverse solutions. This approach, termed the neural-adjoint, achieves the best performance in many scenarios. \ No newline at end of file diff --git a/data/2020/neurips/Benchmarking Deep Learning Interpretability in Time Series Predictions b/data/2020/neurips/Benchmarking Deep Learning Interpretability in Time Series Predictions new file mode 100644 index 0000000000..43b78137f8 --- /dev/null +++ b/data/2020/neurips/Benchmarking Deep Learning Interpretability in Time Series Predictions @@ -0,0 +1 @@ +Saliency methods are used extensively to highlight the importance of input features in model predictions. These methods are mostly used in vision and language tasks, and their applications to time series data is relatively unexplored. In this paper, we set out to extensively compare the performance of various saliency-based interpretability methods across diverse neural architectures, including Recurrent Neural Network, Temporal Convolutional Networks, and Transformers in a new benchmark of synthetic time series data. We propose and report multiple metrics to empirically evaluate the performance of saliency methods for detecting feature importance over time using both precision (i.e., whether identified features contain meaningful signals) and recall (i.e., the number of features with signal identified as important). Through several experiments, we show that (i) in general, network architectures and saliency methods fail to reliably and accurately identify feature importance over time in time series data, (ii) this failure is mainly due to the conflation of time and feature domains, and (iii) the quality of saliency maps can be improved substantially by using our proposed two-step temporal saliency rescaling (TSR) approach that first calculates the importance of each time step before calculating the importance of each feature at a time step. \ No newline at end of file diff --git a/data/2020/neurips/Beta Embeddings for Multi-Hop Logical Reasoning in Knowledge Graphs b/data/2020/neurips/Beta Embeddings for Multi-Hop Logical Reasoning in Knowledge Graphs new file mode 100644 index 0000000000..e75b07d4c1 --- /dev/null +++ b/data/2020/neurips/Beta Embeddings for Multi-Hop Logical Reasoning in Knowledge Graphs @@ -0,0 +1 @@ +One of the fundamental problems in Artificial Intelligence is to perform complex multi-hop logical reasoning over the facts captured by a knowledge graph (KG). This problem is challenging, because KGs can be massive and incomplete. Recent approaches embed KG entities in a low dimensional space and then use these embeddings to find the answer entities. However, it has been an outstanding challenge of how to handle arbitrary first-order logic (FOL) queries as present methods are limited to only a subset of FOL operators. In particular, the negation operator is not supported. An additional limitation of present methods is also that they cannot naturally model uncertainty. Here, we present BetaE, a probabilistic embedding framework for answering arbitrary FOL queries over KGs. BetaE is the first method that can handle a complete set of first-order logical operations: conjunction ($\wedge$), disjunction ($\vee$), and negation ($\neg$). A key insight of BetaE is to use probabilistic distributions with bounded support, specifically the Beta distribution, and embed queries/entities as distributions, which as a consequence allows us to also faithfully model uncertainty. Logical operations are performed in the embedding space by neural operators over the probabilistic embeddings. We demonstrate the performance of BetaE on answering arbitrary FOL queries on three large, incomplete KGs. While being more general, BetaE also increases relative performance by up to 25.4% over the current state-of-the-art KG reasoning methods that can only handle conjunctive queries without negation. \ No newline at end of file diff --git a/data/2020/neurips/Beta R-CNN: Looking into Pedestrian Detection from Another Perspective b/data/2020/neurips/Beta R-CNN: Looking into Pedestrian Detection from Another Perspective new file mode 100644 index 0000000000..b1ecb13be1 --- /dev/null +++ b/data/2020/neurips/Beta R-CNN: Looking into Pedestrian Detection from Another Perspective @@ -0,0 +1 @@ +Recently significant progress has been made in pedestrian detection, but it remains challenging to achieve high performance in occluded and crowded scenes. It could be attributed mostly to the widely used representation of pedestrians, i.e ., 2D axis-aligned bounding box, which just describes the approximate location and size of the object. Bounding box models the object as a uniform distribution within the boundary, making pedestrians indistinguishable in occluded and crowded scenes due to much noise. To eliminate the problem, we propose a novel representation based on 2D beta distribution, named Beta Representation. It pictures a pedestrian by explicitly constructing the relationship between full-body and visible boxes, and emphasizes the center of visual mass by assigning different probability values to pixels. As a result, Beta Representation is much better for distinguishing highly-overlapped instances in crowded scenes with a new NMS strategy named BetaNMS. What’s more, to fully exploit Beta Representation, a novel pipeline Beta R-CNN equipped with BetaHead and BetaMask is proposed, leading to high detection performance in occluded and crowded scenes. \ No newline at end of file diff --git a/data/2020/neurips/Better Full-Matrix Regret via Parameter-Free Online Learning b/data/2020/neurips/Better Full-Matrix Regret via Parameter-Free Online Learning new file mode 100644 index 0000000000..c77964840c --- /dev/null +++ b/data/2020/neurips/Better Full-Matrix Regret via Parameter-Free Online Learning @@ -0,0 +1 @@ +We provide online convex optimization algorithms that guarantee improved full-matrix regret bounds. These algorithms extend prior work in several ways. First, we seamlessly allow for the incorporation of constraints without requiring unknown oracle-tuning for any learning rate parameters. Second, we improve the regret analysis of the full-matrix AdaGrad algorithm by suggesting a better learning rate value and showing how to tune the learning rate to this value on-the-fly. Third, all our bounds are obtained via a general framework for constructing regret bounds that depend on an arbitrary sequence of norms. \ No newline at end of file diff --git a/data/2020/neurips/Better Set Representations For Relational Reasoning b/data/2020/neurips/Better Set Representations For Relational Reasoning new file mode 100644 index 0000000000..d74fb90eee --- /dev/null +++ b/data/2020/neurips/Better Set Representations For Relational Reasoning @@ -0,0 +1 @@ +Incorporating relational reasoning into neural networks has greatly expanded their capabilities and scope. One defining trait of relational reasoning is that it operates on a set of entities, as opposed to standard vector representations. Existing end-to-end approaches typically extract entities from inputs by directly interpreting the latent feature representations as a set. We show that these approaches do not respect set permutational invariance and thus have fundamental representational limitations. To resolve this limitation, we propose a simple and general network module called a Set Refiner Network (SRN). We first use synthetic image experiments to demonstrate how our approach effectively decomposes objects without explicit supervision. Then, we insert our module into existing relational reasoning models and show that respecting set invariance leads to substantial gains in prediction performance and robustness on several relational reasoning tasks. \ No newline at end of file diff --git a/data/2020/neurips/Beyond Homophily in Graph Neural Networks: Current Limitations and Effective Designs b/data/2020/neurips/Beyond Homophily in Graph Neural Networks: Current Limitations and Effective Designs new file mode 100644 index 0000000000..65ebe79edb --- /dev/null +++ b/data/2020/neurips/Beyond Homophily in Graph Neural Networks: Current Limitations and Effective Designs @@ -0,0 +1 @@ +We investigate the representation power of graph neural networks in the semi-supervised node classification task under heterophily or low homophily, i.e., in networks where connected nodes may have different class labels and dissimilar features. Many popular GNNs fail to generalize to this setting, and are even outperformed by models that ignore the graph structure (e.g., multilayer perceptrons). Motivated by this limitation, we identify a set of key designs -- ego- and neighbor-embedding separation, higher-order neighborhoods, and combination of intermediate representations -- that boost learning from the graph structure under heterophily. We combine them into a graph neural network, H2GCN, which we use as the base method to empirically evaluate the effectiveness of the identified designs. Going beyond the traditional benchmarks with strong homophily, our empirical analysis shows that the identified designs increase the accuracy of GNNs by up to 40% and 27% over models without them on synthetic and real networks with heterophily, respectively, and yield competitive performance under homophily. \ No newline at end of file diff --git a/data/2020/neurips/Beyond Individualized Recourse: Interpretable and Interactive Summaries of Actionable Recourses b/data/2020/neurips/Beyond Individualized Recourse: Interpretable and Interactive Summaries of Actionable Recourses new file mode 100644 index 0000000000..9918d06a74 --- /dev/null +++ b/data/2020/neurips/Beyond Individualized Recourse: Interpretable and Interactive Summaries of Actionable Recourses @@ -0,0 +1 @@ +As predictive models are increasingly being deployed in high-stakes decision-making, there has been a lot of interest in developing algorithms which can provide recourses to affected individuals. While developing such tools is important, it is even more critical to analyse and interpret a predictive model, and vet it thoroughly to ensure that the recourses it offers are meaningful and non-discriminatory before it is deployed in the real world. To this end, we propose a novel model agnostic framework called Actionable Recourse Summaries (AReS) to construct global counterfactual explanations which provide an interpretable and accurate summary of recourses for the entire population. We formulate a novel objective which simultaneously optimizes for correctness of the recourses and interpretability of the explanations, while minimizing overall recourse costs across the entire population. More specifically, our objective enables us to learn, with optimality guarantees on recourse correctness, a small number of compact rule sets each of which capture recourses for well defined subpopulations within the data. We also demonstrate theoretically that several of the prior approaches proposed to generate recourses for individuals are special cases of our framework. Experimental evaluation with real world datasets and user studies demonstrate that our framework can provide decision makers with a comprehensive overview of recourses corresponding to any black box model, and consequently help detect undesirable model biases and discrimination. \ No newline at end of file diff --git a/data/2020/neurips/Beyond Lazy Training for Over-parameterized Tensor Decomposition b/data/2020/neurips/Beyond Lazy Training for Over-parameterized Tensor Decomposition new file mode 100644 index 0000000000..24f91827cf --- /dev/null +++ b/data/2020/neurips/Beyond Lazy Training for Over-parameterized Tensor Decomposition @@ -0,0 +1 @@ +Over-parametrization is an important technique in training neural networks. In both theory and practice, training a larger network allows the optimization algorithm to avoid bad local optimal solutions. In this paper we study a closely related tensor decomposition problem: given an $l$-th order tensor in $(R^d)^{\otimes l}$ of rank $r$ (where $r\ll d$), can variants of gradient descent find a rank $m$ decomposition where $m > r$? We show that in a lazy training regime (similar to the NTK regime for neural networks) one needs at least $m = \Omega(d^{l-1})$, while a variant of gradient descent can find an approximate tensor when $m = O^*(r^{2.5l}\log d)$. Our results show that gradient descent on over-parametrized objective could go beyond the lazy training regime and utilize certain low-rank structure in the data. \ No newline at end of file diff --git a/data/2020/neurips/Beyond Perturbations: Learning Guarantees with Arbitrary Adversarial Test Examples b/data/2020/neurips/Beyond Perturbations: Learning Guarantees with Arbitrary Adversarial Test Examples new file mode 100644 index 0000000000..678f310e56 --- /dev/null +++ b/data/2020/neurips/Beyond Perturbations: Learning Guarantees with Arbitrary Adversarial Test Examples @@ -0,0 +1 @@ +We present a transductive learning algorithm that takes as input training examples from a distribution $P$ and arbitrary (unlabeled) test examples, possibly chosen by an adversary. This is unlike prior work that assumes that test examples are small perturbations of $P$. Our algorithm outputs a selective classifier, which abstains from predicting on some examples. By considering selective transductive learning, we give the first nontrivial guarantees for learning classes of bounded VC dimension with arbitrary train and test distributions---no prior guarantees were known even for simple classes of functions such as intervals on the line. In particular, for any function in a class $C$ of bounded VC dimension, we guarantee a low test error rate and a low rejection rate with respect to $P$. Our algorithm is efficient given an Empirical Risk Minimizer (ERM) for $C$. Our guarantees hold even for test examples chosen by an unbounded white-box adversary. We also give guarantees for generalization, agnostic, and unsupervised settings. \ No newline at end of file diff --git a/data/2020/neurips/Beyond accuracy: quantifying trial-by-trial behaviour of CNNs and humans by measuring error consistency b/data/2020/neurips/Beyond accuracy: quantifying trial-by-trial behaviour of CNNs and humans by measuring error consistency new file mode 100644 index 0000000000..0abee77e3c --- /dev/null +++ b/data/2020/neurips/Beyond accuracy: quantifying trial-by-trial behaviour of CNNs and humans by measuring error consistency @@ -0,0 +1 @@ +A central problem in cognitive science and behavioural neuroscience as well as in machine learning and artificial intelligence research is to ascertain whether two or more decision makers (e.g. brains or algorithms) use the same strategy. Accuracy alone cannot distinguish between strategies: two systems may achieve similar accuracy with very different strategies. The need to differentiate beyond accuracy is particularly pressing if two systems are at or near ceiling performance, like Convolutional Neural Networks (CNNs) and humans on visual object recognition. Here we introduce trial-by-trial error consistency, a quantitative analysis for measuring whether two decision making systems systematically make errors on the same inputs. Making consistent errors on a trial-by-trial basis is a necessary condition if we want to ascertain similar processing strategies between decision makers. Our analysis is applicable to compare algorithms with algorithms, humans with humans, and algorithms with humans. When applying error consistency to visual object recognition we obtain three main findings: (1.) Irrespective of architecture, CNNs are remarkably consistent with one another (2.) The consistency between CNNs and human observers, however, is little above what can be expected by chance alone--indicating that humans and CNNs are likely implementing very different strategies (3.) CORnet-S, a recurrent model termed the "current best model of the primate ventral visual stream", fails to capture essential characteristics of human behavioural data and behaves essentially like a ResNet-50 in our analysis--that is, just like a standard feedforward network. Taken together, error consistency analysis suggests that the strategies used by human and machine vision are still very different--but we envision our general-purpose error consistency analysis to serve as a fruitful tool for quantifying future progress. \ No newline at end of file diff --git a/data/2020/neurips/Beyond the Mean-Field: Structured Deep Gaussian Processes Improve the Predictive Uncertainties b/data/2020/neurips/Beyond the Mean-Field: Structured Deep Gaussian Processes Improve the Predictive Uncertainties new file mode 100644 index 0000000000..b1ddc3cd05 --- /dev/null +++ b/data/2020/neurips/Beyond the Mean-Field: Structured Deep Gaussian Processes Improve the Predictive Uncertainties @@ -0,0 +1 @@ +Deep Gaussian Processes learn probabilistic data representations for supervised learning by cascading multiple Gaussian Processes. While this model family promises flexible predictive distributions, exact inference is not tractable. Approximate inference techniques trade off the ability to closely resemble the posterior distribution against speed of convergence and computational efficiency. We propose a novel Gaussian variational family that allows for retaining covariances between latent processes while achieving fast convergence by marginalising out all global latent variables. After providing a proof of how this marginalisation can be done for general covariances, we restrict them to the ones we empirically found to be most important in order to also achieve computational efficiency. We provide an efficient implementation of our new approach and apply it to several regression benchmark datasets. We find that it yields more accurate predictive distributions, in particular for test data points that are distant from the training set. \ No newline at end of file diff --git a/data/2020/neurips/Bi-level Score Matching for Learning Energy-based Latent Variable Models b/data/2020/neurips/Bi-level Score Matching for Learning Energy-based Latent Variable Models new file mode 100644 index 0000000000..41937e723d --- /dev/null +++ b/data/2020/neurips/Bi-level Score Matching for Learning Energy-based Latent Variable Models @@ -0,0 +1 @@ +Score matching (SM) provides a compelling approach to learn energy-based models (EBMs) by avoiding the calculation of partition function. However, it remains largely open to learn energy-based latent variable models (EBLVMs), except some special cases. This paper presents a bi-level score matching (BiSM) method to learn EBLVMs with general structures by reformulating SM as a bi-level optimization problem. The higher level introduces a variational posterior of the latent variables and optimizes a modified SM objective, and the lower level optimizes the variational posterior to fit the true posterior. To solve BiSM efficiently, we develop a stochastic optimization algorithm with gradient unrolling. Theoretically, we analyze the consistency of BiSM and the convergence of the stochastic algorithm. Empirically, we show the promise of BiSM in Gaussian restricted Boltzmann machines and highly nonstructural EBLVMs parameterized by deep convolutional neural networks. BiSM is comparable to the widely adopted contrastive divergence and SM methods when they are applicable; and can learn complex EBLVMs with intractable posteriors to generate natural images. \ No newline at end of file diff --git a/data/2020/neurips/Bias no more: high-probability data-dependent regret bounds for adversarial bandits and MDPs b/data/2020/neurips/Bias no more: high-probability data-dependent regret bounds for adversarial bandits and MDPs new file mode 100644 index 0000000000..6300a65b83 --- /dev/null +++ b/data/2020/neurips/Bias no more: high-probability data-dependent regret bounds for adversarial bandits and MDPs @@ -0,0 +1,2 @@ +We develop a new approach to obtaining high probability regret bounds for online learning with bandit feedback against an adaptive adversary. While existing approaches all require carefully constructing optimistic and biased loss estimators, our approach uses standard unbiased estimators and relies on a simple increasing learning rate schedule, together with the help of logarithmically homogeneous self-concordant barriers and a strengthened Freedman's inequality. +Besides its simplicity, our approach enjoys several advantages. First, the obtained high-probability regret bounds are data-dependent and could be much smaller than the worst-case bounds, which resolves an open problem asked by Neu (2015). Second, resolving another open problem of Bartlett et al. (2008) and Abernethy and Rakhlin (2009), our approach leads to the first general and efficient algorithm with a high-probability regret bound for adversarial linear bandits, while previous methods are either inefficient or only applicable to specific action sets. Finally, our approach can also be applied to learning adversarial Markov Decision Processes and provides the first algorithm with a high-probability small-loss bound for this problem. \ No newline at end of file diff --git a/data/2020/neurips/Bidirectional Convolutional Poisson Gamma Dynamical Systems b/data/2020/neurips/Bidirectional Convolutional Poisson Gamma Dynamical Systems new file mode 100644 index 0000000000..41622b4720 --- /dev/null +++ b/data/2020/neurips/Bidirectional Convolutional Poisson Gamma Dynamical Systems @@ -0,0 +1 @@ +, \ No newline at end of file diff --git a/data/2020/neurips/Big Bird: Transformers for Longer Sequences b/data/2020/neurips/Big Bird: Transformers for Longer Sequences new file mode 100644 index 0000000000..61943313d2 --- /dev/null +++ b/data/2020/neurips/Big Bird: Transformers for Longer Sequences @@ -0,0 +1 @@ +Transformers-based models, such as BERT, have been one of the most successful deep learning models for NLP. Unfortunately, one of their core limitations is the quadratic dependency (mainly in terms of memory) on the sequence length due to their full attention mechanism. To remedy this, we propose, BigBird, a sparse attention mechanism that reduces this quadratic dependency to linear. We show that BigBird is a universal approximator of sequence functions and is Turing complete, thereby preserving these properties of the quadratic, full attention model. Along the way, our theoretical analysis reveals some of the benefits of having $O(1)$ global tokens (such as CLS), that attend to the entire sequence as part of the sparse attention mechanism. The proposed sparse attention can handle sequences of length up to 8x of what was previously possible using similar hardware. As a consequence of the capability to handle longer context, BigBird drastically improves performance on various NLP tasks such as question answering and summarization. We also propose novel applications to genomics data. \ No newline at end of file diff --git a/data/2020/neurips/Big Self-Supervised Models are Strong Semi-Supervised Learners b/data/2020/neurips/Big Self-Supervised Models are Strong Semi-Supervised Learners new file mode 100644 index 0000000000..3d8a39831c --- /dev/null +++ b/data/2020/neurips/Big Self-Supervised Models are Strong Semi-Supervised Learners @@ -0,0 +1 @@ +One paradigm for learning from few labeled examples while making best use of a large amount of unlabeled data is unsupervised pretraining followed by supervised fine-tuning. Although this paradigm uses unlabeled data in a task-agnostic way, in contrast to most previous approaches to semi-supervised learning for computer vision, we show that it is surprisingly effective for semi-supervised learning on ImageNet. A key ingredient of our approach is the use of a big (deep and wide) network during pretraining and fine-tuning. We find that, the fewer the labels, the more this approach (task-agnostic use of unlabeled data) benefits from a bigger network. After fine-tuning, the big network can be further improved and distilled into a much smaller one with little loss in classification accuracy by using the unlabeled examples for a second time, but in a task-specific way. The proposed semi-supervised learning algorithm can be summarized in three steps: unsupervised pretraining of a big ResNet model using SimCLRv2 (a modification of SimCLR), supervised fine-tuning on a few labeled examples, and distillation with unlabeled examples for refining and transferring the task-specific knowledge. This procedure achieves 73.9\% ImageNet top-1 accuracy with just 1\% of the labels ($\le$13 labeled images per class) using ResNet-50, a $10\times$ improvement in label efficiency over the previous state-of-the-art. With 10\% of labels, ResNet-50 trained with our method achieves 77.5\% top-1 accuracy, outperforming standard supervised training with all of the labels. \ No newline at end of file diff --git a/data/2020/neurips/Biological credit assignment through dynamic inversion of feedforward networks b/data/2020/neurips/Biological credit assignment through dynamic inversion of feedforward networks new file mode 100644 index 0000000000..f64b04c9c3 --- /dev/null +++ b/data/2020/neurips/Biological credit assignment through dynamic inversion of feedforward networks @@ -0,0 +1 @@ +Learning depends on changes in synaptic connections deep inside the brain. In multilayer networks, these changes are triggered by error signals fed back from the output, generally through a stepwise inversion of the feedforward processing steps. The gold standard for this process -- backpropagation -- works well in artificial neural networks, but is biologically implausible. Several recent proposals have emerged to address this problem, but many of these biologically-plausible schemes are based on learning an independent set of feedback connections. This complicates the assignment of errors to each synapse by making it dependent upon a second learning problem, and by fitting inversions rather than guaranteeing them. Here, we show that feedforward network transformations can be effectively inverted through dynamics. We derive this dynamic inversion from the perspective of feedback control, where the forward transformation is reused and dynamically interacts with fixed or random feedback to propagate error signals during the backward pass. Importantly, this scheme does not rely upon a second learning problem for feedback because accurate inversion is guaranteed through the network dynamics. We map these dynamics onto generic feedforward networks, and show that the resulting algorithm performs well on several supervised and unsupervised datasets. We also link this dynamic inversion to Gauss-Newton optimization, suggesting a biologically-plausible approximation to second-order learning. Overall, our work introduces an alternative perspective on credit assignment in the brain, and proposes a special role for temporal dynamics and feedback control during learning. \ No newline at end of file diff --git a/data/2020/neurips/Biologically Inspired Mechanisms for Adversarial Robustness b/data/2020/neurips/Biologically Inspired Mechanisms for Adversarial Robustness new file mode 100644 index 0000000000..ee2e04f399 --- /dev/null +++ b/data/2020/neurips/Biologically Inspired Mechanisms for Adversarial Robustness @@ -0,0 +1 @@ +A convolutional neural network strongly robust to adversarial perturbations at reasonable computational and performance cost has not yet been demonstrated. The primate visual ventral stream seems to be robust to small perturbations in visual stimuli but the underlying mechanisms that give rise to this robust perception are not understood. In this work, we investigate the role of two biologically plausible mechanisms in adversarial robustness. We demonstrate that the non-uniform sampling performed by the primate retina and the presence of multiple receptive fields with a range of receptive field sizes at each eccentricity improve the robustness of neural networks to small adversarial perturbations. We verify that these two mechanisms do not suffer from gradient obfuscation and study their contribution to adversarial robustness through ablation studies. \ No newline at end of file diff --git a/data/2020/neurips/Black-Box Certification with Randomized Smoothing: A Functional Optimization Based Framework b/data/2020/neurips/Black-Box Certification with Randomized Smoothing: A Functional Optimization Based Framework new file mode 100644 index 0000000000..ba0cf71fdb --- /dev/null +++ b/data/2020/neurips/Black-Box Certification with Randomized Smoothing: A Functional Optimization Based Framework @@ -0,0 +1 @@ +Randomized classifiers have been shown to provide a promising approach for achieving certified robustness against adversarial attacks in deep learning. However, most existing methods only leverage Gaussian smoothing noise and only work for $\ell_2$ perturbation. We propose a general framework of adversarial certification with non-Gaussian noise and for more general types of attacks, from a unified functional optimization perspective. Our new framework allows us to identify a key trade-off between accuracy and robustness via designing smoothing distributions, helping to design new families of non-Gaussian smoothing distributions that work more efficiently for different $\ell_p$ settings, including $\ell_1$, $\ell_2$ and $\ell_\infty$ attacks. Our proposed methods achieve better certification results than previous works and provide a new perspective on randomized smoothing certification. \ No newline at end of file diff --git a/data/2020/neurips/Black-Box Optimization with Local Generative Surrogates b/data/2020/neurips/Black-Box Optimization with Local Generative Surrogates new file mode 100644 index 0000000000..b99a6e17df --- /dev/null +++ b/data/2020/neurips/Black-Box Optimization with Local Generative Surrogates @@ -0,0 +1 @@ +We propose a novel method for gradient-based optimization of black-box simulators using differentiable local surrogate models. In fields such as physics and engineering, many processes are modeled with non-differentiable simulators with intractable likelihoods. Optimization of these forward models is particularly challenging, especially when the simulator is stochastic. To address such cases, we introduce the use of deep generative models to iteratively approximate the simulator in local neighborhoods of the parameter space. We demonstrate that these local surrogates can be used to approximate the gradient of the simulator, and thus enable gradient-based optimization of simulator parameters. In cases where the dependence of the simulator on the parameter space is constrained to a low dimensional submanifold, we observe that our method attains minima faster than baseline methods, including Bayesian optimization, numerical optimization, and approaches using score function gradient estimators. \ No newline at end of file diff --git a/data/2020/neurips/Black-Box Ripper: Copying black-box models using generative evolutionary algorithms b/data/2020/neurips/Black-Box Ripper: Copying black-box models using generative evolutionary algorithms new file mode 100644 index 0000000000..40fa1dbafd --- /dev/null +++ b/data/2020/neurips/Black-Box Ripper: Copying black-box models using generative evolutionary algorithms @@ -0,0 +1 @@ +We study the task of replicating the functionality of black-box neural models, for which we only know the output class probabilities provided for a set of input images. We assume back-propagation through the black-box model is not possible and its training images are not available, e.g. the model could be exposed only through an API. In this context, we present a teacher-student framework that can distill the black-box (teacher) model into a student model with minimal accuracy loss. To generate useful data samples for training the student, our framework (i) learns to generate images on a proxy data set (with images and classes different from those used to train the black-box) and (ii) applies an evolutionary strategy to make sure that each generated data sample exhibits a high response for a specific class when given as input to the black box. Our framework is compared with several baseline and state-of-the-art methods on three benchmark data sets. The empirical evidence indicates that our model is superior to the considered baselines. Although our method does not back-propagate through the black-box network, it generally surpasses state-of-the-art methods that regard the teacher as a glass-box model. Our code is available at: this https URL. \ No newline at end of file diff --git a/data/2020/neurips/Blind Video Temporal Consistency via Deep Video Prior b/data/2020/neurips/Blind Video Temporal Consistency via Deep Video Prior new file mode 100644 index 0000000000..09e987213c --- /dev/null +++ b/data/2020/neurips/Blind Video Temporal Consistency via Deep Video Prior @@ -0,0 +1 @@ +Applying image processing algorithms independently to each video frame often leads to temporal inconsistency in the resulting video. To address this issue, we present a novel and general approach for blind video temporal consistency. Our method is only trained on a pair of original and processed videos directly instead of a large dataset. Unlike most previous methods that enforce temporal consistency with optical flow, we show that temporal consistency can be achieved by training a convolutional network on a video with the Deep Video Prior. Moreover, a carefully designed iteratively reweighted training strategy is proposed to address the challenging multimodal inconsistency problem. We demonstrate the effectiveness of our approach on 7 computer vision tasks on videos. Extensive quantitative and perceptual experiments show that our approach obtains superior performance than state-of-the-art methods on blind video temporal consistency. Our source codes are publicly available at this http URL. \ No newline at end of file diff --git a/data/2020/neurips/BlockGAN: Learning 3D Object-aware Scene Representations from Unlabelled Images b/data/2020/neurips/BlockGAN: Learning 3D Object-aware Scene Representations from Unlabelled Images new file mode 100644 index 0000000000..6c7533e37e --- /dev/null +++ b/data/2020/neurips/BlockGAN: Learning 3D Object-aware Scene Representations from Unlabelled Images @@ -0,0 +1 @@ +We present BlockGAN, an image generative model that learns object-aware 3D scene representations directly from unlabelled 2D images. Current work on scene representation learning either ignores scene background or treats the whole scene as one object. Meanwhile, work that considers scene compositionality treats scene objects only as image patches or 2D layers with alpha maps. Inspired by the computer graphics pipeline, we design BlockGAN to learn to first generate 3D features of background and foreground objects, then combine them into 3D features for the wholes cene, and finally render them into realistic images. This allows BlockGAN to reason over occlusion and interaction between objects' appearance, such as shadow and lighting, and provides control over each object's 3D pose and identity, while maintaining image realism. BlockGAN is trained end-to-end, using only unlabelled single images, without the need for 3D geometry, pose labels, object masks, or multiple views of the same scene. Our experiments show that using explicit 3D features to represent objects allows BlockGAN to learn disentangled representations both in terms of objects (foreground and background) and their properties (pose and identity). \ No newline at end of file diff --git a/data/2020/neurips/BoTorch: A Framework for Efficient Monte-Carlo Bayesian Optimization b/data/2020/neurips/BoTorch: A Framework for Efficient Monte-Carlo Bayesian Optimization new file mode 100644 index 0000000000..f84c41dc01 --- /dev/null +++ b/data/2020/neurips/BoTorch: A Framework for Efficient Monte-Carlo Bayesian Optimization @@ -0,0 +1 @@ +One of the earliest commonly-used packages is Spearmint [94], which implements a variety of modeling techniques such as MCMC hyperparameter sampling and input warping [95]. Spearmint also supports parallel optimization via fantasies, and constrained optimization with the expected improvement and predictive entropy search acquisition functions [31, 38]. Spearmint was among the first libraries to make BO easily accessible to the end user. \ No newline at end of file diff --git a/data/2020/neurips/Boosting Adversarial Training with Hypersphere Embedding b/data/2020/neurips/Boosting Adversarial Training with Hypersphere Embedding new file mode 100644 index 0000000000..54a6bafaa0 --- /dev/null +++ b/data/2020/neurips/Boosting Adversarial Training with Hypersphere Embedding @@ -0,0 +1 @@ +Adversarial training (AT) is one of the most effective defenses against adversarial attacks for deep learning models. In this work, we advocate incorporating the hypersphere embedding (HE) mechanism into the AT procedure by regularizing the features onto compact manifolds, which constitutes a lightweight yet effective module to blend in the strength of representation learning. Our extensive analyses reveal that AT and HE are well coupled to benefit the robustness of the adversarially trained models from several aspects. We validate the effectiveness and adaptability of HE by embedding it into the popular AT frameworks including PGD-AT, ALP, and TRADES, as well as the FreeAT and FastAT strategies. In the experiments, we evaluate our methods under a wide range of adversarial attacks on the CIFAR-10 and ImageNet datasets, which verifies that integrating HE can consistently enhance the model robustness for each AT framework with little extra computation. \ No newline at end of file diff --git a/data/2020/neurips/Boosting First-Order Methods by Shifting Objective: New Schemes with Faster Worst-Case Rates b/data/2020/neurips/Boosting First-Order Methods by Shifting Objective: New Schemes with Faster Worst-Case Rates new file mode 100644 index 0000000000..9b40ea66c3 --- /dev/null +++ b/data/2020/neurips/Boosting First-Order Methods by Shifting Objective: New Schemes with Faster Worst-Case Rates @@ -0,0 +1 @@ +We propose a new methodology to design first-order methods for unconstrained strongly convex problems, i.e., to design for a shifted objective function. Several technical lemmas are provided as the building blocks for designing new methods. By shifting objective, the analysis is tightened, which leaves space for faster rates, and also simplified. Following this methodology, we derived several new accelerated schemes for problems that equipped with various first-order oracles, and all of the derived methods have faster worst case convergence rates than their existing counterparts. Experiments on machine learning tasks are conducted to evaluate the new methods. \ No newline at end of file diff --git a/data/2020/neurips/Bootstrap Your Own Latent - A New Approach to Self-Supervised Learning b/data/2020/neurips/Bootstrap Your Own Latent - A New Approach to Self-Supervised Learning new file mode 100644 index 0000000000..b24f54f0e1 --- /dev/null +++ b/data/2020/neurips/Bootstrap Your Own Latent - A New Approach to Self-Supervised Learning @@ -0,0 +1 @@ +We introduce Bootstrap Your Own Latent (BYOL), a new approach to self-supervised image representation learning. BYOL relies on two neural networks, referred to as online and target networks, that interact and learn from each other. From an augmented view of an image, we train the online network to predict the target network representation of the same image under a different augmented view. At the same time, we update the target network with a slow-moving average of the online network. While state-of-the art methods rely on negative pairs, BYOL achieves a new state of the art without them. BYOL reaches $74.3\%$ top-1 classification accuracy on ImageNet using a linear evaluation with a ResNet-50 architecture and $79.6\%$ with a larger ResNet. We show that BYOL performs on par or better than the current state of the art on both transfer and semi-supervised benchmarks. Our implementation and pretrained models are given on GitHub. \ No newline at end of file diff --git a/data/2020/neurips/Bootstrapping neural processes b/data/2020/neurips/Bootstrapping neural processes new file mode 100644 index 0000000000..8652c57686 --- /dev/null +++ b/data/2020/neurips/Bootstrapping neural processes @@ -0,0 +1 @@ +Unlike in the traditional statistical modeling for which a user typically hand-specify a prior, Neural Processes (NPs) implicitly define a broad class of stochastic processes with neural networks. Given a data stream, NP learns a stochastic process that best describes the data. While this "data-driven" way of learning stochastic processes has proven to handle various types of data, NPs still rely on an assumption that uncertainty in stochastic processes is modeled by a single latent variable, which potentially limits the flexibility. To this end, we propose the Boostrapping Neural Process (BNP), a novel extension of the NP family using the bootstrap. The bootstrap is a classical data-driven technique for estimating uncertainty, which allows BNP to learn the stochasticity in NPs without assuming a particular form. We demonstrate the efficacy of BNP on various types of data and its robustness in the presence of model-data mismatch. \ No newline at end of file diff --git a/data/2020/neurips/Boundary thickness and robustness in learning models b/data/2020/neurips/Boundary thickness and robustness in learning models new file mode 100644 index 0000000000..94d1a5da1a --- /dev/null +++ b/data/2020/neurips/Boundary thickness and robustness in learning models @@ -0,0 +1 @@ +Robustness of machine learning models to various adversarial and non-adversarial corruptions continues to be of interest. In this paper, we introduce the notion of the boundary thickness of a classifier, and we describe its connection with and usefulness for model robustness. Thick decision boundaries lead to improved performance, while thin decision boundaries lead to overfitting (e.g., measured by the robust generalization gap between training and testing) and lower robustness. We show that a thicker boundary helps improve robustness against adversarial examples (e.g., improving the robust test accuracy of adversarial training) as well as so-called out-of-distribution (OOD) transforms, and we show that many commonly-used regularization and data augmentation procedures can increase boundary thickness. On the theoretical side, we establish that maximizing boundary thickness during training is akin to the so-called mixup training. Using these observations, we show that noise-augmentation on mixup training further increases boundary thickness, thereby combating vulnerability to various forms of adversarial attacks and OOD transforms. We can also show that the performance improvement in several lines of recent work happens in conjunction with a thicker boundary. \ No newline at end of file diff --git a/data/2020/neurips/BoxE: A Box Embedding Model for Knowledge Base Completion b/data/2020/neurips/BoxE: A Box Embedding Model for Knowledge Base Completion new file mode 100644 index 0000000000..96f2a4a8b5 --- /dev/null +++ b/data/2020/neurips/BoxE: A Box Embedding Model for Knowledge Base Completion @@ -0,0 +1 @@ +Knowledge base completion (KBC) aims to automatically infer missing facts by exploiting information already present in a knowledge base (KB). A promising approach for KBC is to embed knowledge into latent spaces and make predictions from learned embeddings. However, existing embedding models are subject to at least one of the following limitations: (1) theoretical inexpressivity, (2) lack of support for prominent inference patterns (e.g., hierarchies), (3) lack of support for KBC over higher-arity relations, and (4) lack of support for incorporating logical rules. Here, we propose a spatio-translational embedding model, called BoxE, that simultaneously addresses all these limitations. BoxE embeds entities as points, and relations as a set of hyper-rectangles (or boxes), which spatially characterize basic logical properties. This seemingly simple abstraction yields a fully expressive model offering a natural encoding for many desired logical properties. BoxE can both capture and inject rules from rich classes of rule languages, going well beyond individual inference patterns. By design, BoxE naturally applies to higher-arity KBs. We conduct a detailed experimental analysis, and show that BoxE achieves state-of-the-art performance, both on benchmark knowledge graphs and on more general KBs, and we empirically show the power of integrating logical rules. \ No newline at end of file diff --git a/data/2020/neurips/Breaking Reversibility Accelerates Langevin Dynamics for Non-Convex Optimization b/data/2020/neurips/Breaking Reversibility Accelerates Langevin Dynamics for Non-Convex Optimization new file mode 100644 index 0000000000..0364ad3ef0 --- /dev/null +++ b/data/2020/neurips/Breaking Reversibility Accelerates Langevin Dynamics for Non-Convex Optimization @@ -0,0 +1 @@ +Langevin dynamics (LD) has been proven to be a powerful technique for optimizing a non-convex objective as an efficient algorithm to find local minima while eventually visiting a global minimum on longer time-scales. LD is based on the first-order Langevin diffusion which is reversible in time. We study two variants that are based on non-reversible Langevin diffusions: the underdamped Langevin dynamics (ULD) and the Langevin dynamics with a non-symmetric drift (NLD). Adopting the techniques of Tzen et al. (2018) for LD to non-reversible diffusions, we show that for a given local minimum that is within an arbitrary distance from the initialization, with high probability, either the ULD trajectory ends up somewhere outside a small neighborhood of this local minimum within a recurrence time which depends on the smallest eigenvalue of the Hessian at the local minimum or they enter this neighborhood by the recurrence time and stay there for a potentially exponentially long escape time. The ULD algorithm improves upon the recurrence time obtained for LD in Tzen et al. (2018) with respect to the dependency on the smallest eigenvalue of the Hessian at the local minimum. Similar results and improvements are obtained for the NLD algorithm. We also show that non-reversible variants can exit the basin of attraction of a local minimum faster in discrete time when the objective has two local minima separated by a saddle point and quantify the amount of improvement. Our analysis suggests that non-reversible Langevin algorithms are \ No newline at end of file diff --git a/data/2020/neurips/Breaking the Communication-Privacy-Accuracy Trilemma b/data/2020/neurips/Breaking the Communication-Privacy-Accuracy Trilemma new file mode 100644 index 0000000000..c76d6c5ce8 --- /dev/null +++ b/data/2020/neurips/Breaking the Communication-Privacy-Accuracy Trilemma @@ -0,0 +1 @@ +Two major challenges in distributed learning and estimation are 1) preserving the privacy of the local samples; and 2) communicating them efficiently to a central server, while achieving high accuracy for the end-to-end task. While there has been significant interest in addressing each of these challenges separately in the recent literature, treatments that simultaneously address both challenges are still largely missing. In this paper, we develop novel encoding and decoding mechanisms that simultaneously achieve optimal privacy and communication efficiency in various canonical settings. In particular, we consider the problems of mean estimation and frequency estimation under $\varepsilon $ -local differential privacy and $b$ -bit communication constraints. For mean estimation, we propose the SQKR mechanism, a scheme based on Kashin’s representation and random sampling, with order-optimal estimation error under both constraints. We further apply SQKR to distributed SGD and obtain a communication efficient and (locally) differentially private distributed SGD protocol. For frequency estimation, we present the RHR mechanism, a scheme that leverages the recursive structure of Walsh-Hadamard matrices and achieves order-optimal estimation error for all privacy levels and communication budgets. As a by-product, we also construct a distribution estimation mechanism that is rate-optimal for all privacy regimes and communication constraints, extending recent work that is limited to $b=1$ and $\varepsilon =O(1)$ . Our results demonstrate that intelligent encoding under joint privacy and communication constraints can yield a performance that matches the optimal accuracy achievable under either constraint alone. In other words, the optimal performance is determined by the more stringent of the two constraints, and the less stringent constraint can be satisfied for free. \ No newline at end of file diff --git a/data/2020/neurips/Breaking the Sample Size Barrier in Model-Based Reinforcement Learning with a Generative Model b/data/2020/neurips/Breaking the Sample Size Barrier in Model-Based Reinforcement Learning with a Generative Model new file mode 100644 index 0000000000..18734cd984 --- /dev/null +++ b/data/2020/neurips/Breaking the Sample Size Barrier in Model-Based Reinforcement Learning with a Generative Model @@ -0,0 +1 @@ +This paper studies a central issue in modern reinforcement learning, the sample efficiency, and makes progress toward solving an idealistic scenario that assumes access to a generative model or a simulator. Despite a large number of prior works tackling this problem, a complete picture of the trade-offs between sample complexity and statistical accuracy has yet to be determined. In particular, all prior results suffer from a severe sample size barrier in the sense that their claimed statistical guarantees hold only when the sample size exceeds some enormous threshold. The current paper overcomes this barrier and fully settles this problem; more specifically, we establish the minimax optimality of the model-based approach for any given target accuracy level. To the best of our knowledge, this work delivers the first minimax-optimal guarantees that accommodate the entire range of sample sizes (beyond which finding a meaningful policy is information theoretically infeasible). \ No newline at end of file diff --git a/data/2020/neurips/Bridging Imagination and Reality for Model-Based Deep Reinforcement Learning b/data/2020/neurips/Bridging Imagination and Reality for Model-Based Deep Reinforcement Learning new file mode 100644 index 0000000000..a20b0341e4 --- /dev/null +++ b/data/2020/neurips/Bridging Imagination and Reality for Model-Based Deep Reinforcement Learning @@ -0,0 +1 @@ +Sample efficiency has been one of the major challenges for deep reinforcement learning. Recently, model-based reinforcement learning has been proposed to address this challenge by performing planning on imaginary trajectories with a learned world model. However, world model learning may suffer from overfitting to training trajectories, and thus model-based value estimation and policy search will be pone to be sucked in an inferior local policy. In this paper, we propose a novel model-based reinforcement learning algorithm, called BrIdging Reality and Dream (BIRD). It maximizes the mutual information between imaginary and real trajectories so that the policy improvement learned from imaginary trajectories can be easily generalized to real trajectories. We demonstrate that our approach improves sample efficiency of model-based planning, and achieves state-of-the-art performance on challenging visual control benchmarks. \ No newline at end of file diff --git a/data/2020/neurips/Bridging the Gap between Sample-based and One-shot Neural Architecture Search with BONAS b/data/2020/neurips/Bridging the Gap between Sample-based and One-shot Neural Architecture Search with BONAS new file mode 100644 index 0000000000..209758648c --- /dev/null +++ b/data/2020/neurips/Bridging the Gap between Sample-based and One-shot Neural Architecture Search with BONAS @@ -0,0 +1 @@ +Neural Architecture Search (NAS) has shown great potentials in finding better neural network designs. Sample-based NAS is the most reliable approach which aims at exploring the search space and evaluating the most promising architectures. However, it is computationally very costly. As a remedy, the one-shot approach has emerged as a popular technique for accelerating NAS using weight-sharing. However, due to the weight-sharing of vastly different networks, the one-shot approach is less reliable than the sample-based approach. In this work, we propose BONAS (Bayesian Optimized Neural Architecture Search), a sample-based NAS framework which is accelerated using weight-sharing to evaluate multiple related architectures simultaneously. Specifically, we apply Graph Convolutional Network predictor as a surrogate model for Bayesian Optimization to select multiple related candidate models in each iteration. We then apply weight-sharing to train multiple candidate models simultaneously. This approach not only accelerates the traditional sample-based approach significantly, but also keeps its reliability. This is because weight-sharing among related architectures are more reliable than those in the one-shot approach. Extensive experiments are conducted to verify the effectiveness of our method over many competing algorithms. \ No newline at end of file diff --git a/data/2020/neurips/Building powerful and equivariant graph neural networks with structural message-passing b/data/2020/neurips/Building powerful and equivariant graph neural networks with structural message-passing new file mode 100644 index 0000000000..ca0f31bcdf --- /dev/null +++ b/data/2020/neurips/Building powerful and equivariant graph neural networks with structural message-passing @@ -0,0 +1 @@ +Message-passing has proved to be an effective way to design graph neural networks, as it is able to leverage both permutation equivariance and an inductive bias towards learning local structures to achieve good generalization. However, current message-passing architectures have a limited representation power and fail to learn basic topological properties of graphs. We address this problem and propose a new message-passing framework that is powerful while preserving permutation equivariance. Specifically, we propagate unique node identifiers in the form of a one-hot encoding in order to learn a local context matrix around each node. This enables to learn rich local information about both features and topology, which can be pooled to obtain node representations. Experimentally, we find our model to be superior at predicting various graph topological properties, opening the way to novel powerful architectures that are both equivariant and computationally efficient. \ No newline at end of file diff --git a/data/2020/neurips/Byzantine Resilient Distributed Multi-Task Learning b/data/2020/neurips/Byzantine Resilient Distributed Multi-Task Learning new file mode 100644 index 0000000000..6e9377bdfc --- /dev/null +++ b/data/2020/neurips/Byzantine Resilient Distributed Multi-Task Learning @@ -0,0 +1 @@ +Distributed multi-task learning provides significant advantages in multi-agent networks with heterogeneous data sources where agents aim to learn distinct but correlated models simultaneously. However, distributed algorithms for learning relatedness among tasks are not resilient in the presence of Byzantine agents. In this paper, we present an approach for Byzantine resilient distributed multi-task learning. We propose an efficient online weight assignment rule by measuring the accumulated loss using an agent's data and its neighbors' models. A small accumulated loss indicates a large similarity between the two tasks. In order to ensure the Byzantine resilience of the aggregation at a normal agent, we introduce a step for filtering out larger losses. We analyze the approach for convex models and show that normal agents converge resiliently towards their true targets. Further, an agent's learning performance using the proposed weight assignment rule is guaranteed to be at least as good as in the non-cooperative case as measured by the expected regret. Finally, we demonstrate the approach using three case studies, including regression and classification problems, and show that our method exhibits good empirical performance for non-convex models, such as convolutional neural networks. \ No newline at end of file diff --git a/data/2020/neurips/CASTLE: Regularization via Auxiliary Causal Graph Discovery b/data/2020/neurips/CASTLE: Regularization via Auxiliary Causal Graph Discovery new file mode 100644 index 0000000000..a0d8db47b6 --- /dev/null +++ b/data/2020/neurips/CASTLE: Regularization via Auxiliary Causal Graph Discovery @@ -0,0 +1 @@ +Regularization improves generalization of supervised models to out-of-sample data. Prior works have shown that prediction in the causal direction (effect from cause) results in lower testing error than the anti-causal direction. However, existing regularization methods are agnostic of causality. We introduce Causal Structure Learning (CASTLE) regularization and propose to regularize a neural network by jointly learning the causal relationships between variables. CASTLE learns the causal directed acyclical graph (DAG) as an adjacency matrix embedded in the neural network's input layers, thereby facilitating the discovery of optimal predictors. Furthermore, CASTLE efficiently reconstructs only the features in the causal DAG that have a causal neighbor, whereas reconstruction-based regularizers suboptimally reconstruct all input features. We provide a theoretical generalization bound for our approach and conduct experiments on a plethora of synthetic and real publicly available datasets demonstrating that CASTLE consistently leads to better out-of-sample predictions as compared to other popular benchmark regularizers. \ No newline at end of file diff --git a/data/2020/neurips/CHIP: A Hawkes Process Model for Continuous-time Networks with Scalable and Consistent Estimation b/data/2020/neurips/CHIP: A Hawkes Process Model for Continuous-time Networks with Scalable and Consistent Estimation new file mode 100644 index 0000000000..4db50f1a6e --- /dev/null +++ b/data/2020/neurips/CHIP: A Hawkes Process Model for Continuous-time Networks with Scalable and Consistent Estimation @@ -0,0 +1 @@ +In many application settings involving networks, such as messages between users of an on-line social network or transactions between traders in financial markets, the observed data consist of timestamped relational events, which form a continuous-time network. We propose the Community Hawkes Independent Pairs (CHIP) generative model for such networks. We show that applying spectral clustering to an aggregated adjacency matrix constructed from the CHIP model provides consistent community detection for a growing number of nodes and time duration. We also develop consistent and computationally efficient estimators for the model parameters. We demonstrate that our proposed CHIP model and estimation procedure scales to large networks with tens of thousands of nodes and provides superior fits than existing continuous-time network models on several real networks. \ No newline at end of file diff --git a/data/2020/neurips/CLEARER: Multi-Scale Neural Architecture Search for Image Restoration b/data/2020/neurips/CLEARER: Multi-Scale Neural Architecture Search for Image Restoration new file mode 100644 index 0000000000..e328e4f621 --- /dev/null +++ b/data/2020/neurips/CLEARER: Multi-Scale Neural Architecture Search for Image Restoration @@ -0,0 +1 @@ +Multi-scale neural networks have shown effectiveness in image restoration tasks, which are usually designed and integrated in a handcrafted manner. Different from the existing labor-intensive handcrafted architecture design paradigms, we present a novel method, termed as multi-sCaLe nEural ARchitecture sEarch for image Restoration (CLEARER), which is a specifically designed neural architecture search (NAS) for image restoration. Our contributions are twofold. On one hand, we design a multi-scale search space that consists of three task-flexible modules. Namely, 1) Parallel module that connects multi-resolution neural blocks in parallel, while preserving the channels and spatial-resolution in each neural block, 2) Transition module remains the existing multi-resolution features while extending them to a lower resolution, 3) Fusion module integrates multi-resolution features by passing the features of the parallel neural blocks to the current neural blocks. On the other hand, we present novel losses which could 1) balance the tradeoff between the model complexity and performance, which is highly expected to image restoration; and 2) relax the discrete architecture parameters into a continuous distribution which approximates to either 0 or 1. As a result, a differentiable strategy could be employed to search when to fuse or extract multi-resolution features, while the discretization issue faced by the gradient-based NAS could be alleviated. The proposed CLEARER could search a promising architecture in two GPU hours. Extensive experiments show the promising performance of our method comparing with nine image denoising methods and eight image deraining approaches in quantitative and qualitative evaluations. The codes are available at https://github.com/limit-scu . \ No newline at end of file diff --git a/data/2020/neurips/COBE: Contextualized Object Embeddings from Narrated Instructional Video b/data/2020/neurips/COBE: Contextualized Object Embeddings from Narrated Instructional Video new file mode 100644 index 0000000000..d8c2ccaf40 --- /dev/null +++ b/data/2020/neurips/COBE: Contextualized Object Embeddings from Narrated Instructional Video @@ -0,0 +1 @@ +Many objects in the real world undergo dramatic variations in visual appearance. For example, a tomato may be red or green, sliced or chopped, fresh or fried, liquid or solid. Training a single detector to accurately recognize tomatoes in all these different states is challenging. On the other hand, contextual cues (e.g., the presence of a knife, a cutting board, a strainer or a pan) are often strongly indicative of how the object appears in the scene. Recognizing such contextual cues is useful not only to improve the accuracy of object detection or to determine the state of the object, but also to understand its functional properties and to infer ongoing or upcoming human-object interactions. A fully-supervised approach to recognizing object states and their contexts in the real-world is unfortunately marred by the long-tailed, open-ended distribution of the data, which would effectively require massive amounts of annotations to capture the appearance of objects in all their different forms. Instead of relying on manually-labeled data for this task, we propose a new framework for learning Contextualized OBject Embeddings (COBE) from automatically-transcribed narrations of instructional videos. We leverage the semantic and compositional structure of language by training a visual detector to predict a contextualized word embedding of the object and its associated narration. This enables the learning of an object representation where concepts relate according to a semantic language metric. Our experiments show that our detector learns to predict a rich variety of contextual object information, and that it is highly effective in the settings of few-shot and zero-shot learning. \ No newline at end of file diff --git a/data/2020/neurips/COOT: Cooperative Hierarchical Transformer for Video-Text Representation Learning b/data/2020/neurips/COOT: Cooperative Hierarchical Transformer for Video-Text Representation Learning new file mode 100644 index 0000000000..10c9dc5cd7 --- /dev/null +++ b/data/2020/neurips/COOT: Cooperative Hierarchical Transformer for Video-Text Representation Learning @@ -0,0 +1 @@ +Many real-world video-text tasks involve different levels of granularity, such as frames and words, clip and sentences or videos and paragraphs, each with distinct semantics. In this paper, we propose a Cooperative hierarchical Transformer (COOT) to leverage this hierarchy information and model the interactions between different levels of granularity and different modalities. The method consists of three major components: an attention-aware feature aggregation layer, which leverages the local temporal context (intra-level, e.g., within a clip), a contextual transformer to learn the interactions between low-level and high-level semantics (inter-level, e.g. clip-video, sentence-paragraph), and a cross-modal cycle-consistency loss to connect video and text. The resulting method compares favorably to the state of the art on several benchmarks while having few parameters. All code is available open-source at this https URL \ No newline at end of file diff --git a/data/2020/neurips/COPT: Coordinated Optimal Transport on Graphs b/data/2020/neurips/COPT: Coordinated Optimal Transport on Graphs new file mode 100644 index 0000000000..910481c6f4 --- /dev/null +++ b/data/2020/neurips/COPT: Coordinated Optimal Transport on Graphs @@ -0,0 +1 @@ +We introduce COPT, a novel distance metric between graphs defined via an optimization routine, computing a coordinated pair of optimal transport maps simultaneously. This gives an unsupervised way to learn general-purpose graph representation, applicable to both graph sketching and graph comparison. COPT involves simultaneously optimizing dual transport plans, one between the vertices of two graphs, and another between graph signal probability distributions. We show theoretically that our method preserves important global structural information on graphs, in particular spectral information, and analyze connections to existing studies. Empirically, COPT outperforms state of the art methods in graph classification on both synthetic and real datasets. \ No newline at end of file diff --git a/data/2020/neurips/COT-GAN: Generating Sequential Data via Causal Optimal Transport b/data/2020/neurips/COT-GAN: Generating Sequential Data via Causal Optimal Transport new file mode 100644 index 0000000000..54f74fe5cc --- /dev/null +++ b/data/2020/neurips/COT-GAN: Generating Sequential Data via Causal Optimal Transport @@ -0,0 +1 @@ +We introduce COT-GAN, an adversarial algorithm to train implicit generative models optimized for producing sequential data. The loss function of this algorithm is formulated using ideas from Causal Optimal Transport (COT), which combines classic optimal transport methods with an additional temporal causality constraint. Remarkably, we find that this causality condition provides a natural framework to parameterize the cost function that is learned by the discriminator as a robust (worst-case) distance, and an ideal mechanism for learning time dependent data distributions. Following Genevay et al.\ (2018), we also include an entropic penalization term which allows for the use of the Sinkhorn algorithm when computing the optimal transport cost. Our experiments show effectiveness and stability of COT-GAN when generating both low- and high-dimensional time series data. The success of the algorithm also relies on a new, improved version of the Sinkhorn divergence which demonstrates less bias in learning. \ No newline at end of file diff --git a/data/2020/neurips/CSER: Communication-efficient SGD with Error Reset b/data/2020/neurips/CSER: Communication-efficient SGD with Error Reset new file mode 100644 index 0000000000..12fce13ef2 --- /dev/null +++ b/data/2020/neurips/CSER: Communication-efficient SGD with Error Reset @@ -0,0 +1 @@ +The scalability of Distributed Stochastic Gradient Descent (SGD) is today limited by communication bottlenecks. We propose a novel SGD variant: Communication-efficient SGD with Error Reset, or CSER. The key idea in CSER is first a new technique called "error reset" that adapts arbitrary compressors for SGD, producing bifurcated local models with periodic reset of resulting local residual errors. Second we introduce partial synchronization for both the gradients and the models, leveraging advantages from them. We prove the convergence of CSER for smooth non-convex problems. Empirical results show that when combined with highly aggressive compressors, the CSER algorithms: i) cause no loss of accuracy, and ii) accelerate the training by nearly $10\times$ for CIFAR-100, and by $4.5\times$ for ImageNet. \ No newline at end of file diff --git a/data/2020/neurips/CSI: Novelty Detection via Contrastive Learning on Distributionally Shifted Instances b/data/2020/neurips/CSI: Novelty Detection via Contrastive Learning on Distributionally Shifted Instances new file mode 100644 index 0000000000..a324622e8d --- /dev/null +++ b/data/2020/neurips/CSI: Novelty Detection via Contrastive Learning on Distributionally Shifted Instances @@ -0,0 +1 @@ +Novelty detection, i.e., identifying whether a given sample is drawn from outside the training distribution, is essential for reliable machine learning. To this end, there have been many attempts at learning a representation well-suited for novelty detection and designing a score based on such representation. In this paper, we propose a simple, yet effective method named contrasting shifted instances (CSI), inspired by the recent success on contrastive learning of visual representations. Specifically, in addition to contrasting a given sample with other instances as in conventional contrastive learning methods, our training scheme contrasts the sample with distributionally-shifted augmentations of itself. Based on this, we propose a new detection score that is specific to the proposed training scheme. Our experiments demonstrate the superiority of our method under various novelty detection scenarios, including unlabeled one-class, unlabeled multi-class and labeled multi-class settings, with various image benchmark datasets. \ No newline at end of file diff --git a/data/2020/neurips/CaSPR: Learning Canonical Spatiotemporal Point Cloud Representations b/data/2020/neurips/CaSPR: Learning Canonical Spatiotemporal Point Cloud Representations new file mode 100644 index 0000000000..9b29750e33 --- /dev/null +++ b/data/2020/neurips/CaSPR: Learning Canonical Spatiotemporal Point Cloud Representations @@ -0,0 +1 @@ +We propose CaSPR, a method to learn object-centric canonical spatiotemporal point cloud representations of dynamically moving or evolving objects. Our goal is to enable information aggregation over time and the interrogation of object state at any spatiotemporal neighborhood in the past, observed or not. Different from previous work, CaSPR learns representations that support spacetime continuity, are robust to variable and irregularly spacetime-sampled point clouds, and generalize to unseen object instances. Our approach divides the problem into two subtasks. First, we explicitly encode time by mapping an input point cloud sequence to a spatiotemporally-canonicalized object space. We then leverage this canonicalization to learn a spatiotemporal latent representation using neural ordinary differential equations and a generative model of dynamically evolving shapes using continuous normalizing flows. We demonstrate the effectiveness of our method on several applications including shape reconstruction, camera pose estimation, continuous spatiotemporal sequence reconstruction, and correspondence estimation from irregularly or intermittently sampled observations. \ No newline at end of file diff --git a/data/2020/neurips/Calibrated Reliable Regression using Maximum Mean Discrepancy b/data/2020/neurips/Calibrated Reliable Regression using Maximum Mean Discrepancy new file mode 100644 index 0000000000..4a5879eb08 --- /dev/null +++ b/data/2020/neurips/Calibrated Reliable Regression using Maximum Mean Discrepancy @@ -0,0 +1 @@ +Accurate quantification of uncertainty is crucial for real-world applications of machine learning. However, modern deep neural networks still produce unreliable predictive uncertainty, often yielding over-confident predictions. In this paper, we are concerned with getting well-calibrated predictions in regression tasks. We propose the calibrated regression method using the maximum mean discrepancy by minimizing the kernel embedding measure. Theoretically, the calibration error of our method asymptotically converges to zero when the sample size is large enough. Experiments on non-trivial real datasets show that our method can produce well-calibrated and sharp prediction intervals, which outperforms the related state-of-the-art methods. \ No newline at end of file diff --git a/data/2020/neurips/Calibrating CNNs for Lifelong Learning b/data/2020/neurips/Calibrating CNNs for Lifelong Learning new file mode 100644 index 0000000000..7d85f07d57 --- /dev/null +++ b/data/2020/neurips/Calibrating CNNs for Lifelong Learning @@ -0,0 +1 @@ +We present an approach for lifelong/continual learning of convolutional neural networks (CNN) that does not suffer from the problem of catastrophic forgetting when moving from one task to the other. We show that the activation maps generated by the CNN trained on the old task can be calibrated using very few calibration parameters, to become relevant to the new task. Based on this, we calibrate the activation maps produced by each network layer using spatial and channel-wise calibration modules and train only these calibration parameters for each new task in order to perform lifelong learning. Our calibration modules intro-duce significantly less computation and parameters as compared to the approaches that dynamically expand the network. Our approach is immune to catastrophic forgetting since we store the task-adaptive calibration parameters, which contain all the task-specific knowledge and is exclusive to each task. Further, our approach does not require storing data samples from the old tasks, which is done by many replay based methods. We perform extensive experiments on multiple benchmark datasets (SVHN, CIFAR, ImageNet, and MS-Celeb), all of which show substan-tial improvements over state-of-the-art methods (e.g., a 29% absolute increase in accuracy on CIFAR-100 with 10 classes at a time). On large-scale datasets, our approach yields 23.8% and 9.7% absolute increase in accuracy on ImageNet-100 and MS-Celeb-10K datasets, respectively, by employing very few (0.51% and 0.35% of model parameters) task-adaptive calibration parameters. \ No newline at end of file diff --git a/data/2020/neurips/Calibrating Deep Neural Networks using Focal Loss b/data/2020/neurips/Calibrating Deep Neural Networks using Focal Loss new file mode 100644 index 0000000000..9cfce1730d --- /dev/null +++ b/data/2020/neurips/Calibrating Deep Neural Networks using Focal Loss @@ -0,0 +1 @@ +Miscalibration - a mismatch between a model's confidence and its correctness - of Deep Neural Networks (DNNs) makes their predictions hard to rely on. Ideally, we want networks to be accurate, calibrated and confident. We show that, as opposed to the standard cross-entropy loss, focal loss [Lin et. al., 2017] allows us to learn models that are already very well calibrated. When combined with temperature scaling, whilst preserving accuracy, it yields state-of-the-art calibrated models. We provide a thorough analysis of the factors causing miscalibration, and use the insights we glean from this to justify the empirically excellent performance of focal loss. To facilitate the use of focal loss in practice, we also provide a principled approach to automatically select the hyperparameter involved in the loss function. We perform extensive experiments on a variety of computer vision and NLP datasets, and with a wide variety of network architectures, and show that our approach achieves state-of-the-art calibration without compromising on accuracy in almost all cases. Code is available at this https URL \ No newline at end of file diff --git a/data/2020/neurips/Calibration of Shared Equilibria in General Sum Partially Observable Markov Games b/data/2020/neurips/Calibration of Shared Equilibria in General Sum Partially Observable Markov Games new file mode 100644 index 0000000000..e59a32d46d --- /dev/null +++ b/data/2020/neurips/Calibration of Shared Equilibria in General Sum Partially Observable Markov Games @@ -0,0 +1 @@ +Training multi-agent systems (MAS) to achieve realistic equilibria gives us a useful tool to understand and model real-world systems. We consider a general sum partially observable Markov game where agents of different types share a single policy network, conditioned on agent-specific information. This paper aims at i) formally understanding equilibria reached by such agents, and ii) matching emergent phenomena of such equilibria to real-world targets. Parameter sharing with decentralized execution has been introduced as an efficient way to train multiple agents using a single policy network. However, the nature of resulting equilibria reached by such agents is not yet understood: we introduce the novel concept of \textit{Shared equilibrium} as a symmetric pure Nash equilibrium of a certain Functional Form Game (FFG) and prove convergence to the latter for a certain class of games using self-play. In addition, it is important that such equilibria satisfy certain constraints so that MAS are \textit{calibrated} to real world data for practical use: we solve this problem by introducing a novel dual-Reinforcement Learning based approach that fits emergent behaviors of agents in a Shared equilibrium to externally-specified targets, and apply our methods to a $n$-player market example. We do so by calibrating parameters governing distributions of agent types rather than individual agents, which allows both behavior differentiation among agents and coherent scaling of the shared policy network to multiple agents. \ No newline at end of file diff --git a/data/2020/neurips/Can Graph Neural Networks Count Substructures? b/data/2020/neurips/Can Graph Neural Networks Count Substructures? new file mode 100644 index 0000000000..c72b765b49 --- /dev/null +++ b/data/2020/neurips/Can Graph Neural Networks Count Substructures? @@ -0,0 +1 @@ +The ability to detect and count certain substructures in graphs is important for solving many tasks on graph-structured data, especially in the contexts of computational chemistry and biology as well as social network analysis. Inspired by this, we propose to study the expressive power of graph neural networks (GNNs) via their ability to count attributed graph substructures, extending recent works that examine their power in graph isomorphism testing and function approximation. We distinguish between two types of substructure counting: induced-subgraph-count and subgraph-count, and establish both positive and negative answers for popular GNN architectures. Specifically, we prove that Message Passing Neural Networks (MPNNs), 2-Weisfeiler-Lehman (2-WL) and 2-Invariant Graph Networks (2-IGNs) cannot perform induced-subgraph-count of substructures consisting of 3 or more nodes, while they can perform subgraph-count of star-shaped substructures. As an intermediary step, we prove that 2-WL and 2-IGNs are equivalent in distinguishing non-isomorphic graphs, partly answering an open problem raised in Maron et al. (2019). We also prove positive results for k-WL and k-IGNs as well as negative results for k-WL with a finite number of iterations. We then conduct experiments that support the theoretical results for MPNNs and 2-IGNs. Moreover, motivated by substructure counting, we propose a local relational pooling approach with inspirations from Murphy et al. (2019) and demonstrate that it is not only effective for substructure counting but also able to achieve competitive performance on real-world tasks. \ No newline at end of file diff --git a/data/2020/neurips/Can I Trust My Fairness Metric? Assessing Fairness with Unlabeled Data and Bayesian Inference b/data/2020/neurips/Can I Trust My Fairness Metric? Assessing Fairness with Unlabeled Data and Bayesian Inference new file mode 100644 index 0000000000..d3817b16c9 --- /dev/null +++ b/data/2020/neurips/Can I Trust My Fairness Metric? Assessing Fairness with Unlabeled Data and Bayesian Inference @@ -0,0 +1 @@ +We investigate the problem of reliably assessing group fairness when labeled examples are few but unlabeled examples are plentiful. We propose a general Bayesian framework that can augment labeled data with unlabeled data to produce more accurate and lower-variance estimates compared to methods based on labeled data alone. Our approach estimates calibrated scores for unlabeled examples in each group using a hierarchical latent variable model conditioned on labeled examples. This in turn allows for inference of posterior distributions with associated notions of uncertainty for a variety of group fairness metrics. We demonstrate that our approach leads to significant and consistent reductions in estimation error across multiple well-known fairness datasets, sensitive attributes, and predictive models. The results show the benefits of using both unlabeled data and Bayesian inference in terms of assessing whether a prediction model is fair or not. \ No newline at end of file diff --git a/data/2020/neurips/Can Implicit Bias Explain Generalization? Stochastic Convex Optimization as a Case Study b/data/2020/neurips/Can Implicit Bias Explain Generalization? Stochastic Convex Optimization as a Case Study new file mode 100644 index 0000000000..0a29d3f2f0 --- /dev/null +++ b/data/2020/neurips/Can Implicit Bias Explain Generalization? Stochastic Convex Optimization as a Case Study @@ -0,0 +1 @@ +The notion of implicit bias, or implicit regularization, has been suggested as a means to explain the surprising generalization ability of modern-days overparameterized learning algorithms. This notion refers to the tendency of the optimization algorithm towards a certain structured solution that often generalizes well. Recently, several papers have studied implicit regularization and were able to identify this phenomenon in various scenarios. We revisit this paradigm in arguably the simplest non-trivial setup, and study the implicit bias of Stochastic Gradient Descent (SGD) in the context of Stochastic Convex Optimization. As a first step, we provide a simple construction that rules out the existence of a \emph{distribution-independent} implicit regularizer that governs the generalization ability of SGD. We then demonstrate a learning problem that rules out a very general class of \emph{distribution-dependent} implicit regularizers from explaining generalization, which includes strongly convex regularizers as well as non-degenerate norm-based regularizations. Certain aspects of our constructions point out to significant difficulties in providing a comprehensive explanation of an algorithm's generalization performance by solely arguing about its implicit regularization properties. \ No newline at end of file diff --git a/data/2020/neurips/Can Q-Learning with Graph Networks Learn a Generalizable Branching Heuristic for a SAT Solver? b/data/2020/neurips/Can Q-Learning with Graph Networks Learn a Generalizable Branching Heuristic for a SAT Solver? new file mode 100644 index 0000000000..b692a26b08 --- /dev/null +++ b/data/2020/neurips/Can Q-Learning with Graph Networks Learn a Generalizable Branching Heuristic for a SAT Solver? @@ -0,0 +1 @@ +We present Graph-Q -SAT, a branching heuristic for a Boolean SAT solver trained with value-based reinforcement learning (RL) using Graph Neural Networks for function approximation. Solvers using Graph-Q -SAT are complete SAT solvers that either provide a satisfying assignment or proof of unsatisfiability, which is required for many SAT applications. The branching heuristics commonly used in SAT solvers make poor decisions during their warm-up period, whereas Graph-Q -SAT is trained to examine the structure of the particular problem instance to make better decisions early in the search. Training Graph-Q -SAT is data efficient and does not require elaborate dataset preparation or feature engineering. We train Graph-Q -SAT using RL interfacing with MiniSat solver and show that Graph-Q -SAT can reduce the number of iterations required to solve SAT problems by 2-3X. Furthermore, it generalizes to unsatisfiable SAT instances, as well as to problems with 5X more variables than it was trained on. We show that for larger problems, reductions in the number of iterations lead to wall clock time reductions, the ultimate goal when designing heuristics. We also show positive zero-shot transfer behavior when testing Graph-Q -SAT on a task family different from that used for training. While more work is needed to apply Graph-Q -SAT to reduce wall clock time in modern SAT solving settings, it is a compelling proof-of-concept showing that RL equipped with Graph Neural Networks can learn a generalizable branching heuristic for SAT search. \ No newline at end of file diff --git a/data/2020/neurips/Can Temporal-Difference and Q-Learning Learn Representation? A Mean-Field Theory b/data/2020/neurips/Can Temporal-Difference and Q-Learning Learn Representation? A Mean-Field Theory new file mode 100644 index 0000000000..6db5f7e5b4 --- /dev/null +++ b/data/2020/neurips/Can Temporal-Difference and Q-Learning Learn Representation? A Mean-Field Theory @@ -0,0 +1,3 @@ +Temporal-difference and Q-learning play a key role in deep reinforcement learning, where they are empowered by expressive nonlinear function approximators such as neural networks. At the core of their empirical successes is the learned feature representation, which embeds rich observations, e.g., images and texts, into the latent space that encodes semantic structures. Meanwhile, the evolution of such a feature representation is crucial to the convergence of temporal-difference and Q-learning. +In particular, temporal-difference learning converges when the function approximator is linear in a feature representation, which is fixed throughout learning, and possibly diverges otherwise. We aim to answer the following questions: When the function approximator is a neural network, how does the associated feature representation evolve? If it converges, does it converge to the optimal one? +We prove that, utilizing an overparameterized two-layer neural network, temporal-difference and Q-learning globally minimize the mean-squared projected Bellman error at a sublinear rate. Moreover, the associated feature representation converges to the optimal one, generalizing the previous analysis of Cai et al. (2019) in the neural tangent kernel regime, where the associated feature representation stabilizes at the initial one. The key to our analysis is a mean-field perspective, which connects the evolution of a finite-dimensional parameter to its limiting counterpart over an infinite-dimensional Wasserstein space. Our analysis generalizes to soft Q-learning, which is further connected to policy gradient. \ No newline at end of file diff --git a/data/2020/neurips/Can the Brain Do Backpropagation? - Exact Implementation of Backpropagation in Predictive Coding Networks b/data/2020/neurips/Can the Brain Do Backpropagation? - Exact Implementation of Backpropagation in Predictive Coding Networks new file mode 100644 index 0000000000..e34155ec3c --- /dev/null +++ b/data/2020/neurips/Can the Brain Do Backpropagation? - Exact Implementation of Backpropagation in Predictive Coding Networks @@ -0,0 +1 @@ +Backpropagation (BP) has been the most successful algorithm used to train artificial neural networks. However, there are several gaps between BP and learning in biologically plausible neuronal networks of the brain (learning in the brain, or simply BL, for short), in particular, (1) it has been unclear to date, if BP can be implemented exactly via BL, (2) there is a lack of local plasticity in BP, i.e., weight updates require information that is not locally available, while BL utilizes only locally available information, and (3) there is a lack of autonomy in BP, i.e., some external control over the neural network is required (e.g., switching between prediction and learning stages requires changes to dynamics and synaptic plasticity rules), while BL works fully autonomously. Bridging such gaps, i.e., understanding how BP can be approximated by BL, has been of major interest in both neuroscience and machine learning. Despite tremendous efforts, however, no previous model has bridged the gaps at a degree of demonstrating an equivalence to BP, instead, only approximations to BP have been shown. Here, we present for the first time a framework within BL that bridges the above crucial gaps. We propose a BL model that (1) produces exactly the same updates of the neural weights as BP, while (2) employing local plasticity, i.e., all neurons perform only local computations, done simultaneously. We then modify it to an alternative BL model that (3) also works fully autonomously. Overall, our work provides important evidence for the debate on the long-disputed question whether the brain can perform BP. \ No newline at end of file diff --git a/data/2020/neurips/Canonical 3D Deformer Maps: Unifying parametric and non-parametric methods for dense weakly-supervised category reconstruction b/data/2020/neurips/Canonical 3D Deformer Maps: Unifying parametric and non-parametric methods for dense weakly-supervised category reconstruction new file mode 100644 index 0000000000..1ef454e44e --- /dev/null +++ b/data/2020/neurips/Canonical 3D Deformer Maps: Unifying parametric and non-parametric methods for dense weakly-supervised category reconstruction @@ -0,0 +1 @@ +We propose the Canonical 3D Deformer Map, a new representation of the 3D shape of common object categories that can be learned from a collection of 2D images of independent objects. Our method builds in a novel way on concepts from parametric deformation models, non-parametric 3D reconstruction, and canonical embeddings, combining their individual advantages. In particular, it learns to associate each image pixel with a deformation model of the corresponding 3D object point which is canonical, i.e. intrinsic to the identity of the point and shared across objects of the category. The result is a method that, given only sparse 2D supervision at training time, can, at test time, reconstruct the 3D shape and texture of objects from single views, while establishing meaningful dense correspondences between object instances. It also achieves state-of-the-art results in dense 3D reconstruction on public in-the-wild datasets of faces, cars, and birds. \ No newline at end of file diff --git a/data/2020/neurips/Cascaded Text Generation with Markov Transformers b/data/2020/neurips/Cascaded Text Generation with Markov Transformers new file mode 100644 index 0000000000..a18cdb6d7a --- /dev/null +++ b/data/2020/neurips/Cascaded Text Generation with Markov Transformers @@ -0,0 +1 @@ +The two dominant approaches to neural text generation are fully autoregressive models, using serial beam search decoding, and non-autoregressive models, using parallel decoding with no output dependencies. This work proposes an autoregressive model with sub-linear parallel time generation. Noting that conditional random fields with bounded context can be decoded in parallel, we propose an efficient cascaded decoding approach for generating high-quality output. To parameterize this cascade, we introduce a Markov transformer, a variant of the popular fully autoregressive model that allows us to simultaneously decode with specific autoregressive context cutoffs. This approach requires only a small modification from standard autoregressive training, while showing competitive accuracy/speed tradeoff compared to existing methods on five machine translation datasets. \ No newline at end of file diff --git a/data/2020/neurips/Causal Discovery from Soft Interventions with Unknown Targets: Characterization and Learning b/data/2020/neurips/Causal Discovery from Soft Interventions with Unknown Targets: Characterization and Learning new file mode 100644 index 0000000000..6cb6dc4ab9 --- /dev/null +++ b/data/2020/neurips/Causal Discovery from Soft Interventions with Unknown Targets: Characterization and Learning @@ -0,0 +1 @@ +One fundamental problem in the empirical sciences is of reconstructing the causal structure that underlies a phenomenon of interest through observation and experimentation. While there exists a plethora of methods capable of learning the equivalence class of causal structures that are compatible with observations, it is less well-understood how to systematically combine observations and experiments to reconstruct the underlying structure. In this paper, we investigate the task of structural learning in non-Markovian systems (i.e., when latent variables a ff ect more than one observable) from a combination of observational and soft experimental data when the interventional targets are unknown. Using causal invariances found across the collection of observational and interventional distributions (not only conditional independences), we define a property called Ψ -Markov that connects these distributions to a pair consisting of (1) a causal graph D and (2) a set of interventional targets I . Building on this property, our main contributions are two-fold: First, we provide a graphical characterization that allows one to test whether two causal graphs with possibly di ff erent sets of interventional targets belong to the same Ψ -Markov equivalence class. Second, we develop an algorithm capable of harnessing the collection of data to learn the corresponding equivalence class. We then prove that this algorithm is sound and complete, in the sense that it is the most informative in the sample limit, i.e., it discovers as many tails and arrowheads as can be oriented within a Ψ -Markov equivalence class. \ No newline at end of file diff --git a/data/2020/neurips/Causal Discovery in Physical Systems from Videos b/data/2020/neurips/Causal Discovery in Physical Systems from Videos new file mode 100644 index 0000000000..c3dfb03d84 --- /dev/null +++ b/data/2020/neurips/Causal Discovery in Physical Systems from Videos @@ -0,0 +1 @@ +Causal discovery is at the core of human cognition. It enables us to reason about the environment and make counterfactual predictions about unseen scenarios, that can vastly differ from our previous experiences. We consider the task of causal discovery from videos in an end-to-end fashion without supervision on the ground-truth graph structure. In particular, our goal is to discover the structural dependencies among environmental and object variables: inferring the type and strength of interactions that have a causal effect on the behavior of the dynamical system. Our model consists of (a) a perception module that extracts a semantically meaningful and temporally consistent keypoint representation from images, (b) an inference module for determining the graph distribution induced by the detected keypoints, and (c) a dynamics module that can predict the future by conditioning on the inferred graph. We assume access to different configurations and environmental conditions, i.e., data from unknown interventions on the underlying system; thus, we can hope to discover the correct underlying causal graph without explicit interventions. We evaluate our method in a planar multi-body interaction environment and scenarios involving fabrics of different shapes like shirts and pants. Experiments demonstrate that our model can correctly identify the interactions from a short sequence of images and make long-term future predictions. The causal structure assumed by the model also allows it to make counterfactual predictions and extrapolate to systems of unseen interaction graphs or graphs of various sizes. \ No newline at end of file diff --git a/data/2020/neurips/Causal Estimation with Functional Confounders b/data/2020/neurips/Causal Estimation with Functional Confounders new file mode 100644 index 0000000000..ab2b2af25d --- /dev/null +++ b/data/2020/neurips/Causal Estimation with Functional Confounders @@ -0,0 +1 @@ +Causal inference relies on two fundamental assumptions: ignorability and positivity. We study causal inference when the true confounder value can be expressed as a function of the observed data; we call this setting estimation with functional confounders (EFC). In this setting ignorability is satisfied, however positivity is violated, and causal inference is impossible in general. We consider two scenarios where causal effects are estimable. First, we discuss interventions on a part of the treatment called functional interventions and a sufficient condition for effect estimation of these interventions called functional positivity. Second, we develop conditions for nonparametric effect estimation based on the gradient fields of the functional confounder and the true outcome function. To estimate effects under these conditions, we develop Level-set Orthogonal Descent Estimation (LODE). Further, we prove error bounds on LODE's effect estimates, evaluate our methods on simulated and real data, and empirically demonstrate the value of EFC. \ No newline at end of file diff --git a/data/2020/neurips/Causal Intervention for Weakly-Supervised Semantic Segmentation b/data/2020/neurips/Causal Intervention for Weakly-Supervised Semantic Segmentation new file mode 100644 index 0000000000..3a1c19fef0 --- /dev/null +++ b/data/2020/neurips/Causal Intervention for Weakly-Supervised Semantic Segmentation @@ -0,0 +1 @@ +We present a causal inference framework to improve Weakly-Supervised Semantic Segmentation (WSSS). Specifically, we aim to generate better pixel-level pseudo-masks by using only image-level labels -- the most crucial step in WSSS. We attribute the cause of the ambiguous boundaries of pseudo-masks to the confounding context, e.g., the correct image-level classification of "horse" and "person" may be not only due to the recognition of each instance, but also their co-occurrence context, making the model inspection (e.g., CAM) hard to distinguish between the boundaries. Inspired by this, we propose a structural causal model to analyze the causalities among images, contexts, and class labels. Based on it, we develop a new method: Context Adjustment (CONTA), to remove the confounding bias in image-level classification and thus provide better pseudo-masks as ground-truth for the subsequent segmentation model. On PASCAL VOC 2012 and MS-COCO, we show that CONTA boosts various popular WSSS methods to new state-of-the-arts. \ No newline at end of file diff --git a/data/2020/neurips/Causal Shapley Values: Exploiting Causal Knowledge to Explain Individual Predictions of Complex Models b/data/2020/neurips/Causal Shapley Values: Exploiting Causal Knowledge to Explain Individual Predictions of Complex Models new file mode 100644 index 0000000000..a743f8bb8e --- /dev/null +++ b/data/2020/neurips/Causal Shapley Values: Exploiting Causal Knowledge to Explain Individual Predictions of Complex Models @@ -0,0 +1,2 @@ +Shapley values underlie one of the most popular model-agnostic methods within explainable artificial intelligence. These values are designed to attribute the difference between a model's prediction and an average baseline to the different features used as input to the model. Being based on solid game-theoretic principles, Shapley values uniquely satisfy several desirable properties, which is why they are increasingly used to explain the predictions of possibly complex and highly non-linear machine learning models. Shapley values are well calibrated to a user's intuition when features are independent, but may lead to undesirable, counterintuitive explanations when the independence assumption is violated. +In this paper, we propose a novel framework for computing Shapley values that generalizes recent work that aims to circumvent the independence assumption. By employing Pearl's do-calculus, we show how these 'causal' Shapley values can be derived for general causal graphs without sacrificing any of their desirable properties. Moreover, causal Shapley values enable us to separate the contribution of direct and indirect effects. We provide a practical implementation for computing causal Shapley values based on causal chain graphs when only partial information is available and illustrate their utility on a real-world example. \ No newline at end of file diff --git a/data/2020/neurips/Causal analysis of Covid-19 Spread in Germany b/data/2020/neurips/Causal analysis of Covid-19 Spread in Germany new file mode 100644 index 0000000000..22983d3da9 --- /dev/null +++ b/data/2020/neurips/Causal analysis of Covid-19 Spread in Germany @@ -0,0 +1 @@ +In this work, we study the causal relations among German regions in terms of the spread of Covid-19 since the beginning of the pandemic, taking into account the restriction policies that were applied by the different federal states. We propose and prove a new theorem for a causal feature selection method for time series data, robust to latent confounders, which we subsequently apply on Covid-19 case numbers. We present findings about the spread of the virus in Germany and the causal impact of restriction measures, discussing the role of various policies in containing the spread. Since our results are based on rather limited target time series (only the numbers of reported cases), care should be exercised in interpreting them. However, it is encouraging that already such limited data seems to contain causal signals. This suggests that as more data becomes available, our causal approach may contribute towards meaningful causal analysis of political interventions on the development of Covid-19, and thus also towards the development of rational and data-driven methodologies for choosing interventions. \ No newline at end of file diff --git a/data/2020/neurips/Certifiably Adversarially Robust Detection of Out-of-Distribution Data b/data/2020/neurips/Certifiably Adversarially Robust Detection of Out-of-Distribution Data new file mode 100644 index 0000000000..9d7f423488 --- /dev/null +++ b/data/2020/neurips/Certifiably Adversarially Robust Detection of Out-of-Distribution Data @@ -0,0 +1 @@ +Deep neural networks are known to be overconfident when applied to out-of-distribution (OOD) inputs which clearly do not belong to any class. This is a problem in safety-critical applications since a reliable assessment of the uncertainty of a classifier is a key property, allowing the system to trigger human intervention or to transfer into a safe state. In this paper, we aim for certifiable worst case guarantees for OOD detection by enforcing not only low confidence at the OOD point but also in an $l_\infty$-ball around it. For this purpose, we use interval bound propagation (IBP) to upper bound the maximal confidence in the $l_\infty$-ball and minimize this upper bound during training time. We show that non-trivial bounds on the confidence for OOD data generalizing beyond the OOD dataset seen at training time are possible. Moreover, in contrast to certified adversarial robustness which typically comes with significant loss in prediction performance, certified guarantees for worst case OOD detection are possible without much loss in accuracy. \ No newline at end of file diff --git a/data/2020/neurips/Certified Defense to Image Transformations via Randomized Smoothing b/data/2020/neurips/Certified Defense to Image Transformations via Randomized Smoothing new file mode 100644 index 0000000000..8cc6c982c2 --- /dev/null +++ b/data/2020/neurips/Certified Defense to Image Transformations via Randomized Smoothing @@ -0,0 +1 @@ +We extend randomized smoothing to cover parameterized transformations (e.g., rotations, translations) and certify robustness in the parameter space (e.g., rotation angle). This is particularly challenging as interpolation and rounding effects mean that image transformations do not compose, in turn preventing direct certification of the perturbed image (unlike certification with $\ell^p$ norms). We address this challenge by introducing three different defenses, each with a different guarantee (heuristic, distributional and individual) stemming from the method used to bound the interpolation error. Importantly, in the individual case, we show how to efficiently compute the inverse of an image transformation, enabling us to provide individual guarantees in the online setting. We provide an implementation of all methods at https://github.com/eth-sri/transformation-smoothing. \ No newline at end of file diff --git a/data/2020/neurips/Certified Monotonic Neural Networks b/data/2020/neurips/Certified Monotonic Neural Networks new file mode 100644 index 0000000000..5183a89d7f --- /dev/null +++ b/data/2020/neurips/Certified Monotonic Neural Networks @@ -0,0 +1 @@ +Learning monotonic models with respect to a subset of the inputs is a desirable feature to effectively address the fairness, interpretability, and generalization issues in practice. Existing methods for learning monotonic neural networks either require specifically designed model structures to ensure monotonicity, which can be too restrictive/complicated, or enforce monotonicity by adjusting the learning process, which cannot provably guarantee the learned model is monotonic on selected features. In this work, we propose to certify the monotonicity of the general piece-wise linear neural networks by solving a mixed integer linear programming problem.This provides a new general approach for learning monotonic neural networks with arbitrary model structures. Our method allows us to train neural networks with heuristic monotonicity regularizations, and we can gradually increase the regularization magnitude until the learned network is certified monotonic. Compared to prior works, our approach does not require human-designed constraints on the weight space and also yields more accurate approximation. Empirical studies on various datasets demonstrate the efficiency of our approach over the state-of-the-art methods, such as Deep Lattice Networks. \ No newline at end of file diff --git a/data/2020/neurips/Certified Robustness of Graph Convolution Networks for Graph Classification under Topological Attacks b/data/2020/neurips/Certified Robustness of Graph Convolution Networks for Graph Classification under Topological Attacks new file mode 100644 index 0000000000..d7a8ba6541 --- /dev/null +++ b/data/2020/neurips/Certified Robustness of Graph Convolution Networks for Graph Classification under Topological Attacks @@ -0,0 +1 @@ +Graph convolution networks (GCNs) have become effective models for graph classification. Similar to many deep networks, GCNs are vulnerable to adversarial attacks on graph topology and node attributes. Recently, a number of effective attack and defense algorithms have been developed, but certificates of robustness against topological perturbations are currently available only for PageRank and label/feature propagation, while none has been designed for GCNs. We propose the first algorithm for certifying the robustness of GCNs to topological attacks in the application of graph classification . Our method is based on Lagrange dual-ization and convex envelope, which result in tight approximation bounds that are efficiently computable by dynamic programming. When used in conjunction with robust training, it allows an increased number of graphs to be certified as robust. \ No newline at end of file diff --git a/data/2020/neurips/Certifying Confidence via Randomized Smoothing b/data/2020/neurips/Certifying Confidence via Randomized Smoothing new file mode 100644 index 0000000000..fb6b032958 --- /dev/null +++ b/data/2020/neurips/Certifying Confidence via Randomized Smoothing @@ -0,0 +1 @@ +Randomized smoothing has been shown to provide good certified-robustness guarantees for high-dimensional classification problems. It uses the probabilities of predicting the top two most-likely classes around an input point under a smoothing distribution to generate a certified radius for a classifier's prediction. However, most smoothing methods do not give us any information about the \emph{confidence} with which the underlying classifier (e.g., deep neural network) makes a prediction. In this work, we propose a method to generate certified radii for the prediction confidence of the smoothed classifier. We consider two notions for quantifying confidence: average prediction score of a class and the margin by which the average prediction score of one class exceeds that of another. We modify the Neyman-Pearson lemma (a key theorem in randomized smoothing) to design a procedure for computing the certified radius where the confidence is guaranteed to stay above a certain threshold. Our experimental results on CIFAR-10 and ImageNet datasets show that using information about the distribution of the confidence scores allows us to achieve a significantly better certified radius than ignoring it. Thus, we demonstrate that extra information about the base classifier at the input point can help improve certified guarantees for the smoothed classifier. \ No newline at end of file diff --git a/data/2020/neurips/Certifying Strategyproof Auction Networks b/data/2020/neurips/Certifying Strategyproof Auction Networks new file mode 100644 index 0000000000..a2e2e26b55 --- /dev/null +++ b/data/2020/neurips/Certifying Strategyproof Auction Networks @@ -0,0 +1 @@ +Optimal auctions maximize a seller's expected revenue subject to individual rationality and strategyproofness for the buyers. Myerson's seminal work in 1981 settled the case of auctioning a single item; however, subsequent decades of work have yielded little progress moving beyond a single item, leaving the design of revenue-maximizing auctions as a central open problem in the field of mechanism design. A recent thread of work in "differentiable economics" has used tools from modern deep learning to instead learn good mechanisms. We focus on the RegretNet architecture, which can represent auctions with arbitrary numbers of items and participants; it is trained to be empirically strategyproof, but the property is never exactly verified leaving potential loopholes for market participants to exploit. We propose ways to explicitly verify strategyproofness under a particular valuation profile using techniques from the neural network verification literature. Doing so requires making several modifications to the RegretNet architecture in order to represent it exactly in an integer program. We train our network and produce certificates in several settings, including settings for which the optimal strategyproof mechanism is not known. \ No newline at end of file diff --git a/data/2020/neurips/Chaos, Extremism and Optimism: Volume Analysis of Learning in Games b/data/2020/neurips/Chaos, Extremism and Optimism: Volume Analysis of Learning in Games new file mode 100644 index 0000000000..7660d590ed --- /dev/null +++ b/data/2020/neurips/Chaos, Extremism and Optimism: Volume Analysis of Learning in Games @@ -0,0 +1,3 @@ +We present volume analyses of Multiplicative Weights Updates (MWU) and Optimistic Multiplicative Weights Updates (OMWU) in zero-sum as well as coordination games. Such analyses provide new insights into these game dynamical systems, which seem hard to achieve via the classical techniques within Computer Science and Machine Learning. +The first step is to examine these dynamics not in their original space (simplex of actions) but in a dual space (aggregate payoff space of actions). The second step is to explore how the volume of a set of initial conditions evolves over time when it is pushed forward according to the algorithm. This is reminiscent of approaches in Evolutionary Game Theory where replicator dynamics, the continuous-time analogue of MWU, is known to always preserve volume in all games. Interestingly, when we examine discrete-time dynamics, both the choice of the game and the choice of the algorithm play a critical role. So whereas MWU expands volume in zero-sum games and is thus Lyapunov chaotic, we show that OMWU contracts volume, providing an alternative understanding for its known convergent behavior. However, we also prove a no-free-lunch type of theorem, in the sense that when examining coordination games the roles are reversed: OMWU expands volume exponentially fast, whereas MWU contracts. +Using these tools, we prove two novel, rather negative properties of MWU in zero-sum games: (1) Extremism: even in games with unique fully mixed Nash equilibrium, the system recurrently gets stuck near pure-strategy profiles, despite them being clearly unstable from game theoretic perspective. (2) Unavoidability: given any set of good points (with your own interpretation of "good"), the system cannot avoid bad points indefinitely. \ No newline at end of file diff --git a/data/2020/neurips/Characterizing Optimal Mixed Policies: Where to Intervene and What to Observe b/data/2020/neurips/Characterizing Optimal Mixed Policies: Where to Intervene and What to Observe new file mode 100644 index 0000000000..3973b81f3d --- /dev/null +++ b/data/2020/neurips/Characterizing Optimal Mixed Policies: Where to Intervene and What to Observe @@ -0,0 +1 @@ +Intelligent agents are continuously faced with the challenge of optimizing a policy based on what they can observe (see) and which actions they can take (do) in the environment where they are deployed. Most policy can be parametrized in terms of these two dimensions, i.e., as a function of what can be seen and done given a certain situation, which we call a mixed policy . In this paper, we investigate several properties of the class of mixed policies and provide an efficient and effective characterization, including optimality and non-redundancy. Specifically, we introduce a graphical criterion to identify unnecessary contexts for a set of actions, leading to a natural characterization of non-redundancy of mixed policies. We then derive sufficient conditions under which one strategy can dominate the other with respect to their maximum achievable expected rewards (optimality). This characterization leads to a fundamental understanding of the space of mixed policies and a possible refinement of the agent’s strategy so that it converges to the optimum faster and more robustly. One surprising result of the causal characterization is that the agent following a more standard approach—intervening on all intervenable variables and observing all available contexts—may be hurting itself, and will never achieve an optimal performance. \ No newline at end of file diff --git a/data/2020/neurips/Characterizing emergent representations in a space of candidate learning rules for deep networks b/data/2020/neurips/Characterizing emergent representations in a space of candidate learning rules for deep networks new file mode 100644 index 0000000000..94b3ab7f85 --- /dev/null +++ b/data/2020/neurips/Characterizing emergent representations in a space of candidate learning rules for deep networks @@ -0,0 +1 @@ +We apply singular value decomposition (SVD) to the dataset’s input-output correlation matrix to extract the component of the input-output mapping for different hierarchical levels. Suppose the desired (target) output matrix is given by Y as shown in the main paper Fig. 1b, and input matrix is X where examples are placed in columns. In X, each object’s perceptual representation x (a column vector, where μ = 1...P indexing objects) is encoded by a one-hot input vector (Kronecker delta δμi). Thus, the input-output correlation matrix is Σ = YX . We use SVD on Σ, i.e., Σ = USV , which results in three key elements fully characterizing the input-output mapping to be learned (visualized in Supp. Fig. 1). For the case of hierarchically structured data from a binary tree, the SVD structure conforms to hierarchical distinctions in the dataset [4]. The first element is U, a feature-synthesizer matrix in which each column (a particular semantic dimension or ‘mode’) contains positive (negative) values for semantic features that objects categorized along this mode do (do not) possess. The second element is S, a singular value matrix that has nonzero values only on the diagonal, and these values are arranged in a descending order. And the last element is V , whose rows are object-analyzer vectors, whereby a binary code is assigned to classify objects according to each semantic mode (e.g., the 2nd row of V indicates the first 4 objects are plants, while the last 4 objects are animals, see Fig. 1b in the main paper). \ No newline at end of file diff --git a/data/2020/neurips/Choice Bandits b/data/2020/neurips/Choice Bandits new file mode 100644 index 0000000000..cf99d08a52 --- /dev/null +++ b/data/2020/neurips/Choice Bandits @@ -0,0 +1 @@ +In this work, we study sequential choice bandits with feedback. We propose bandit algorithms for a platform that personalizes users' experience to maximize its rewards. For each action directed to a given user, the platform is given a positive reward, which is a non-decreasing function of the action, if this action is below the user's threshold. Users are equipped with a patience budget, and actions that are above the threshold decrease the user's patience. When all patience is lost, the user abandons the platform. The platform attempts to learn the thresholds of the users in order to maximize its rewards, based on two different feedback models describing the information pattern available to the platform at each action. We define a notion of regret by determining the best action to be taken when the platform knows that the user's threshold is in a given interval. We then propose bandit algorithms for the two feedback models and show that upper and lower bounds on the regret are of the order of $\tilde{O}(N^{2/3})$ and $\tilde\Omega(N^{2/3})$, respectively, where $N$ is the total number of users. Finally, we show that the waiting time of any user before receiving a personalized experience is uniform in $N$. \ No newline at end of file diff --git a/data/2020/neurips/CircleGAN: Generative Adversarial Learning across Spherical Circles b/data/2020/neurips/CircleGAN: Generative Adversarial Learning across Spherical Circles new file mode 100644 index 0000000000..6dc3de578e --- /dev/null +++ b/data/2020/neurips/CircleGAN: Generative Adversarial Learning across Spherical Circles @@ -0,0 +1 @@ +We present a novel discriminator for GANs that improves realness and diversity of generated samples by learning a structured hypersphere embedding space using spherical circles. The proposed discriminator learns to populate realistic samples around the longest spherical circle, i.e., a great circle, while pushing unrealistic samples toward the poles perpendicular to the great circle. Since longer circles occupy larger area on the hypersphere, they encourage more diversity in representation learning, and vice versa. Discriminating samples based on their corresponding spherical circles can thus naturally induce diversity to generated samples. We also extend the proposed method for conditional settings with class labels by creating a hypersphere for each category and performing class-wise discrimination and update. In experiments, we validate the effectiveness for both unconditional and conditional generation on standard benchmarks, achieving the state of the art. \ No newline at end of file diff --git a/data/2020/neurips/Classification Under Misspecification: Halfspaces, Generalized Linear Models, and Evolvability b/data/2020/neurips/Classification Under Misspecification: Halfspaces, Generalized Linear Models, and Evolvability new file mode 100644 index 0000000000..f01bd6fcbf --- /dev/null +++ b/data/2020/neurips/Classification Under Misspecification: Halfspaces, Generalized Linear Models, and Evolvability @@ -0,0 +1 @@ +In this paper, we revisit the problem of distribution-independently learning half-spaces under Massart noise with rate ⌘ . Recent work [DGT19] resolved a long-standing problem in this model of efficiently learning to error ⌘ + ✏ for any ✏ > 0 , by giving an improper learner that partitions space into poly( d, 1 / ✏ ) regions. Here we give a much simpler algorithm and settle a number of outstanding open questions: (1) We give the first proper learner for Massart halfspaces that achieves ⌘ + ✏ . (2) Based on (1), we develop a blackbox knowledge distillation procedure to convert an arbitrarily complex classifier to an equally good proper classifier. (3) By leveraging a simple but overlooked connection to evolvability , we show any SQ algorithm requires super-polynomially many queries to achieve OPT + ✏ . We then zoom out to study generalized linear models and give an efficient algorithm for learning under a challenging new corruption model generalizing Massart noise. Lastly, we empirically evaluate our algorithm for Massart halfspaces and find it exhibits some intriguing fairness properties. \ No newline at end of file diff --git a/data/2020/neurips/Classification with Valid and Adaptive Coverage b/data/2020/neurips/Classification with Valid and Adaptive Coverage new file mode 100644 index 0000000000..6b829a4fb9 --- /dev/null +++ b/data/2020/neurips/Classification with Valid and Adaptive Coverage @@ -0,0 +1 @@ +Conformal inference, cross-validation+, and the jackknife+ are hold-out methods that can be combined with virtually any machine learning algorithm to construct prediction sets with guaranteed marginal coverage. In this paper, we develop specialized versions of these techniques for categorical and unordered response labels that, in addition to providing marginal coverage, are also fully adaptive to complex data distributions, in the sense that they perform favorably in terms of approximate conditional coverage compared to alternative methods. The heart of our contribution is a novel conformity score, which we explicitly demonstrate to be powerful and intuitive for classification problems, but whose underlying principle is potentially far more general. Experiments on synthetic and real data demonstrate the practical value of our theoretical guarantees, as well as the statistical advantages of the proposed methods over the existing alternatives. \ No newline at end of file diff --git a/data/2020/neurips/Closing the Dequantization Gap: PixelCNN as a Single-Layer Flow b/data/2020/neurips/Closing the Dequantization Gap: PixelCNN as a Single-Layer Flow new file mode 100644 index 0000000000..2b51809d1e --- /dev/null +++ b/data/2020/neurips/Closing the Dequantization Gap: PixelCNN as a Single-Layer Flow @@ -0,0 +1 @@ +Flow models have recently made great progress at modeling ordinal discrete data such as images and audio. Due to the continuous nature of flow models, dequantization is typically applied when using them for such discrete data, resulting in lower bound estimates of the likelihood. In this paper, we introduce subset flows, a class of flows that can tractably transform finite volumes and thus allow exact computation of likelihoods for discrete data. Based on subset flows, we identify ordinal discrete autoregressive models, including WaveNets, PixelCNNs and Transformers, as single-layer flows. We use the flow formulation to compare models trained and evaluated with either the exact likelihood or its dequantization lower bound. Finally, we study multilayer flows composed of PixelCNNs and non-autoregressive coupling layers and demonstrate state-of-the-art results on CIFAR-10 for flow models trained with dequantization. \ No newline at end of file diff --git a/data/2020/neurips/Co-Tuning for Transfer Learning b/data/2020/neurips/Co-Tuning for Transfer Learning new file mode 100644 index 0000000000..6181d90e2d --- /dev/null +++ b/data/2020/neurips/Co-Tuning for Transfer Learning @@ -0,0 +1 @@ +Fine-tuning pre-trained deep neural networks (DNNs) to a target dataset, also known as transfer learning, is widely used in computer vision and NLP. Because task-specific layers mainly contain categorical information and categories vary with datasets, practitioners only partially transfer pre-trained models by discarding task-specific layers and fine-tuning bottom layers. However, it is a reckless loss to simply discard task-specific parameters which take up as many as 20% of the total parameters in pre-trained models. To fully transfer pre-trained models, we propose a two-step framework named Co-Tuning : (i) learn the relationship between source categories and target categories from the pre-trained model with calibrated predictions; (ii) target labels (one-hot labels), as well as source labels (probabilistic labels) translated by the category relationship, collaboratively supervise the fine-tuning process. A simple instantiation of the framework shows strong empirical results in four visual classification tasks and one NLP classification task, bringing up to 20% relative improvement. While state-of-the-art fine-tuning techniques mainly focus on how to impose regularization when data are not abundant, Co-Tuning works not only in medium-scale datasets (100 samples per class) but also in large-scale datasets (1000 samples per class) where regularization-based methods bring no gains over the vanilla fine-tuning. Co-Tuning relies on a typically valid assumption that the pre-trained dataset is diverse enough, implying its broad application areas. \ No newline at end of file diff --git a/data/2020/neurips/Co-exposure Maximization in Online Social Networks b/data/2020/neurips/Co-exposure Maximization in Online Social Networks new file mode 100644 index 0000000000..449d3dad8c --- /dev/null +++ b/data/2020/neurips/Co-exposure Maximization in Online Social Networks @@ -0,0 +1 @@ +Social media has created new ways for citizens to stay informed on societal matters and participate in political discourse. However, with its algorithmically-curated and virally-propagating content, social media has contributed further to the polarization of opinions by reinforcing users’ existing viewpoints. An emerging line of research seeks to understand how content-recommendation algorithms can be re-designed to mitigate societal polarization amplified by social-media interactions. In this paper, we study the problem of allocating seed users to opposing campaigns: by drawing on the equal-time rule of political campaigning on traditional media, our goal is to allocate seed users to campaigners with the aim to maximize the expected number of users who are co-exposed to both campaigns. We show that the problem of maximizing co-exposure is NP -hard and its objective function is neither submodular nor supermodular. However, by exploiting a connection to a submodular function that acts as a lower bound to the objective, we are able to devise a greedy algorithm with provable approximation guarantee. We further provide a scalable instantiation of our approximation algorithm by introducing a novel extension to the notion of random reverse-reachable sets for efficiently estimating the expected co-exposure. We experimentally demonstrate the quality of our proposal on real-world social networks. \ No newline at end of file diff --git a/data/2020/neurips/CoADNet: Collaborative Aggregation-and-Distribution Networks for Co-Salient Object Detection b/data/2020/neurips/CoADNet: Collaborative Aggregation-and-Distribution Networks for Co-Salient Object Detection new file mode 100644 index 0000000000..ff68dc5d08 --- /dev/null +++ b/data/2020/neurips/CoADNet: Collaborative Aggregation-and-Distribution Networks for Co-Salient Object Detection @@ -0,0 +1 @@ +Co-Salient Object Detection (CoSOD) aims at discovering salient objects that repeatedly appear in a given query group containing two or more relevant images. One challenging issue is how to effectively capture co-saliency cues by modeling and exploiting inter-image relationships. In this paper, we present an end-to-end collaborative aggregation-and-distribution network (CoADNet) to capture both salient and repetitive visual patterns from multiple images. First, we integrate saliency priors into the backbone features to suppress the redundant background information through an online intra-saliency guidance structure. After that, we design a two-stage aggregate-and-distribute architecture to explore group-wise semantic interactions and produce the co-saliency features. In the first stage, we propose a group-attentional semantic aggregation module that models inter-image relationships to generate the group-wise semantic representations. In the second stage, we propose a gated group distribution module that adaptively distributes the learned group semantics to different individuals in a dynamic gating mechanism. Finally, we develop a group consistency preserving decoder tailored for the CoSOD task, which maintains group constraints during feature decoding to predict more consistent full-resolution co-saliency maps. The proposed CoADNet is evaluated on four prevailing CoSOD benchmark datasets, which demonstrates the remarkable performance improvement over ten state-of-the-art competitors. \ No newline at end of file diff --git a/data/2020/neurips/CoMIR: Contrastive Multimodal Image Representation for Registration b/data/2020/neurips/CoMIR: Contrastive Multimodal Image Representation for Registration new file mode 100644 index 0000000000..10eafbf468 --- /dev/null +++ b/data/2020/neurips/CoMIR: Contrastive Multimodal Image Representation for Registration @@ -0,0 +1 @@ +We propose contrastive coding to learn shared, dense image representations, referred to as CoMIRs (Contrastive Multimodal Image Representations). CoMIRs enable the registration of multimodal images where existing registration methods often fail due to a lack of sufficiently similar image structures. CoMIRs reduce the multimodal registration problem to a monomodal one, in which general intensity-based, as well as feature-based, registration algorithms can be applied. The method involves training one neural network per modality on aligned images, using a contrastive loss based on noise-contrastive estimation (InfoNCE). Unlike other contrastive coding methods, used for, e.g., classification, our approach generates image-like representations that contain the information shared between modalities. We introduce a novel, hyperparameter-free modification to InfoNCE, to enforce rotational equivariance of the learnt representations, a property essential to the registration task. We assess the extent of achieved rotational equivariance and the stability of the representations with respect to weight initialization, training set, and hyperparameter settings, on a remote sensing dataset of RGB and near-infrared images. We evaluate the learnt representations through registration of a biomedical dataset of bright-field and second-harmonic generation microscopy images; two modalities with very little apparent correlation. The proposed approach based on CoMIRs significantly outperforms registration of representations created by GAN-based image-to-image translation, as well as a state-of-the-art, application-specific method which takes additional knowledge about the data into account. Code is available at: this https URL. \ No newline at end of file diff --git a/data/2020/neurips/CoSE: Compositional Stroke Embeddings b/data/2020/neurips/CoSE: Compositional Stroke Embeddings new file mode 100644 index 0000000000..7ee3855cb3 --- /dev/null +++ b/data/2020/neurips/CoSE: Compositional Stroke Embeddings @@ -0,0 +1 @@ +We present a generative model for stroke-based drawing tasks which is able to model complex free-form structures. While previous approaches rely on sequence-based models for drawings of basic objects or handwritten text, we propose a model that treats drawings as a collection of strokes that can be composed into complex structures such as diagrams (e.g., flow-charts). At the core of the approach lies a novel auto-encoder that projects variable-length strokes into a latent space of fixed dimension. This representation space allows a relational model, operating in latent space, to better capture the relationship between strokes and to predict subsequent strokes. We demonstrate qualitatively and quantitatively that our proposed approach is able to model the appearance of individual strokes, as well as the compositional structure of larger diagram drawings. Our approach is suitable for interactive use cases such as auto-completing diagrams. \ No newline at end of file diff --git a/data/2020/neurips/CodeCMR: Cross-Modal Retrieval For Function-Level Binary Source Code Matching b/data/2020/neurips/CodeCMR: Cross-Modal Retrieval For Function-Level Binary Source Code Matching new file mode 100644 index 0000000000..9937f1e932 --- /dev/null +++ b/data/2020/neurips/CodeCMR: Cross-Modal Retrieval For Function-Level Binary Source Code Matching @@ -0,0 +1 @@ +Previous work indicates that CFG is a sensitive feature, because it changes greatly in different optimizations. However, the results have declined somewhat in our research, but within acceptable limits. To train a cross-platform model, we extract the token sequences based on IDA Pro microcode IR (Intermediate Representation). The columns are combinations of different compilers (gcc/clang), different platforms (x86/x64/arm/arm64), and different optimizations (O0/O1/O2/O3). The average recall@1 score of the 32 combinations is 88.9%. The lowest recall@1 score among the 32 combinations is 85.1% on gcc-arm64-O3, which is acceptable. \ No newline at end of file diff --git a/data/2020/neurips/Coded Sequential Matrix Multiplication For Straggler Mitigation b/data/2020/neurips/Coded Sequential Matrix Multiplication For Straggler Mitigation new file mode 100644 index 0000000000..63d46d580c --- /dev/null +++ b/data/2020/neurips/Coded Sequential Matrix Multiplication For Straggler Mitigation @@ -0,0 +1 @@ +In this work, we consider a sequence of $J$ matrix multiplication jobs which needs to be distributed by a master across multiple worker nodes. For $i\in \{1,2,\ldots,J\}$ , job- $i$ begins in round- $i$ and has to be completed by round- $(i+T)$ . In order to provide resiliency against slow workers (stragglers), previous works focus on coding across workers, which is the special case of $T=0$ . We propose here two schemes with $T > 0$ , which allow for coding across workers as well as the dimension of time. Our first scheme is a modification of the polynomial coding scheme introduced by Yu et al. and places no assumptions on the straggler model. Exploitation of the temporal dimension helps the scheme handle a larger set of straggler patterns than the polynomial coding scheme, for a given computational load per worker per round. The second scheme assumes a particular straggler model to further improve performance (in terms of encoding/decoding complexity). We develop theoretical results establishing (i) optimality of our proposed schemes for certain classes of straggler patterns and (ii) improved performance for the case of i.i.d. stragglers. These are further validated by experiments, where we implement our schemes to train neural networks. \ No newline at end of file diff --git a/data/2020/neurips/CogLTX: Applying BERT to Long Texts b/data/2020/neurips/CogLTX: Applying BERT to Long Texts new file mode 100644 index 0000000000..35bea0162e --- /dev/null +++ b/data/2020/neurips/CogLTX: Applying BERT to Long Texts @@ -0,0 +1 @@ +BERT is incapable of processing long texts due to its quadratically increasing memory and time consumption. The most natural ways to address this problem, such as slicing the text by a sliding window or simplifying transformers, suffer from insufficient long-range attentions or need customized CUDA kernels. The maximum length limit in BERT reminds us the limited capacity (5 ∼ 9 chunks) of the working memory of humans –— then how do human beings Cog nize L ong T e X ts? Founded on the cognitive theory stemming from Baddeley [2], the proposed CogLTX 1 framework identifies key sentences by training a judge model, concatenates them for reasoning, and enables multi-step reasoning via rehearsal and decay . Since relevance annotations are usually unavailable, we propose to use interventions to create supervision. As a general algorithm, CogLTX outperforms or gets comparable results to SOTA models on various downstream tasks with memory overheads independent of the length of text. \ No newline at end of file diff --git a/data/2020/neurips/CogMol: Target-Specific and Selective Drug Design for COVID-19 Using Deep Generative Models b/data/2020/neurips/CogMol: Target-Specific and Selective Drug Design for COVID-19 Using Deep Generative Models new file mode 100644 index 0000000000..d7dd735282 --- /dev/null +++ b/data/2020/neurips/CogMol: Target-Specific and Selective Drug Design for COVID-19 Using Deep Generative Models @@ -0,0 +1 @@ +The novel nature of SARS-CoV-2 calls for the development of efficient de novo drug design approaches. In this study, we propose an end-to-end framework, named CogMol (Controlled Generation of Molecules), for designing new drug-like small molecules targeting novel viral proteins with high affinity and off-target selectivity. CogMol combines adaptive pre-training of a molecular SMILES Variational Autoencoder (VAE) and an efficient multi-attribute controlled sampling scheme that uses guidance from attribute predictors trained on latent features. To generate novel and optimal drug-like molecules for unseen viral targets, CogMol leverages a protein-molecule binding affinity predictor that is trained using SMILES VAE embeddings and protein sequence embeddings learned unsupervised from a large corpus. CogMol framework is applied to three SARS-CoV-2 target proteins: main protease, receptor-binding domain of the spike protein, and non-structural protein 9 replicase. The generated candidates are novel at both molecular and chemical scaffold levels when compared to the training data. CogMol also includes insilico screening for assessing toxicity of parent molecules and their metabolites with a multi-task toxicity classifier, synthetic feasibility with a chemical retrosynthesis predictor, and target structure binding with docking simulations. Docking reveals favorable binding of generated molecules to the target protein structure, where 87-95 % of high affinity molecules showed docking free energy < -6 kcal/mol. When compared to approved drugs, the majority of designed compounds show low parent molecule and metabolite toxicity and high synthetic feasibility. In summary, CogMol handles multi-constraint design of synthesizable, low-toxic, drug-like molecules with high target specificity and selectivity, and does not need target-dependent fine-tuning of the framework or target structure information. \ No newline at end of file diff --git a/data/2020/neurips/Coherent Hierarchical Multi-Label Classification Networks b/data/2020/neurips/Coherent Hierarchical Multi-Label Classification Networks new file mode 100644 index 0000000000..c514c3d6a8 --- /dev/null +++ b/data/2020/neurips/Coherent Hierarchical Multi-Label Classification Networks @@ -0,0 +1 @@ +Hierarchical multi-label classification (HMC) is a challenging classification task extending standard multi-label classification problems by imposing a hierarchy constraint on the classes. In this paper, we propose C-HMCNN(h), a novel approach for HMC problems, which, given a network h for the underlying multi-label classification problem, exploits the hierarchy information in order to produce predictions coherent with the constraint and improve performance. We conduct an extensive experimental analysis showing the superior performance of C-HMCNN(h) when compared to state-of-the-art models. \ No newline at end of file diff --git a/data/2020/neurips/CoinDICE: Off-Policy Confidence Interval Estimation b/data/2020/neurips/CoinDICE: Off-Policy Confidence Interval Estimation new file mode 100644 index 0000000000..0000faa16c --- /dev/null +++ b/data/2020/neurips/CoinDICE: Off-Policy Confidence Interval Estimation @@ -0,0 +1 @@ +We study high-confidence behavior-agnostic off-policy evaluation in reinforcement learning, where the goal is to estimate a confidence interval on a target policy's value, given only access to a static experience dataset collected by unknown behavior policies. Starting from a function space embedding of the linear program formulation of the $Q$-function, we obtain an optimization problem with generalized estimating equation constraints. By applying the generalized empirical likelihood method to the resulting Lagrangian, we propose CoinDICE, a novel and efficient algorithm for computing confidence intervals. Theoretically, we prove the obtained confidence intervals are valid, in both asymptotic and finite-sample regimes. Empirically, we show in a variety of benchmarks that the confidence interval estimates are tighter and more accurate than existing methods. \ No newline at end of file diff --git a/data/2020/neurips/CoinPress: Practical Private Mean and Covariance Estimation b/data/2020/neurips/CoinPress: Practical Private Mean and Covariance Estimation new file mode 100644 index 0000000000..5e16262a76 --- /dev/null +++ b/data/2020/neurips/CoinPress: Practical Private Mean and Covariance Estimation @@ -0,0 +1 @@ +We present simple differentially private estimators for the mean and covariance of multivariate sub-Gaussian data that are accurate at small sample sizes. We demonstrate the effectiveness of our algorithms both theoretically and empirically using synthetic and real-world datasets---showing that their asymptotic error rates match the state-of-the-art theoretical bounds, and that they concretely outperform all previous methods. Specifically, previous estimators either have weak empirical accuracy at small sample sizes, perform poorly for multivariate data, or require the user to provide strong a priori estimates for the parameters. \ No newline at end of file diff --git a/data/2020/neurips/ColdGANs: Taming Language GANs with Cautious Sampling Strategies b/data/2020/neurips/ColdGANs: Taming Language GANs with Cautious Sampling Strategies new file mode 100644 index 0000000000..4730acc33d --- /dev/null +++ b/data/2020/neurips/ColdGANs: Taming Language GANs with Cautious Sampling Strategies @@ -0,0 +1 @@ +Training regimes based on Maximum Likelihood Estimation (MLE) suffer from known limitations, often leading to poorly generated text sequences. At the root of these limitations is the mismatch between training and inference, i.e. the so-called exposure bias, exacerbated by considering only the reference texts as correct, while in practice several alternative formulations could be as good. Generative Adversarial Networks (GANs) can mitigate those limitations but the discrete nature of text has hindered their application to language generation: the approaches proposed so far, based on Reinforcement Learning, have been shown to underperform MLE. Departing from previous works, we analyze the exploration step in GANs applied to text generation, and show how classical sampling results in unstable training. We propose to consider alternative exploration strategies in a GAN framework that we name ColdGANs, where we force the sampling to be close to the distribution modes to get smoother learning dynamics. For the first time, to the best of our knowledge, the proposed language GANs compare favorably to MLE, and obtain improvements over the state-of-the-art on three generative tasks, namely unconditional text generation, question generation, and abstractive summarization. \ No newline at end of file diff --git a/data/2020/neurips/Collapsing Bandits and Their Application to Public Health Intervention b/data/2020/neurips/Collapsing Bandits and Their Application to Public Health Intervention new file mode 100644 index 0000000000..f5ca428aae --- /dev/null +++ b/data/2020/neurips/Collapsing Bandits and Their Application to Public Health Intervention @@ -0,0 +1 @@ +We propose and study Collpasing Bandits, a new restless multi-armed bandit (RMAB) setting in which each arm follows a binary-state Markovian process with a special structure: when an arm is played, the state is fully observed, thus "collapsing" any uncertainty, but when an arm is passive, no observation is made, thus allowing uncertainty to evolve. The goal is to keep as many arms in the "good" state as possible by planning a limited budget of actions per round. Such Collapsing Bandits are natural models for many healthcare domains in which workers must simultaneously monitor patients and deliver interventions in a way that maximizes the health of their patient cohort. Our main contributions are as follows: (i) Building on the Whittle index technique for RMABs, we derive conditions under which the Collapsing Bandits problem is indexable. Our derivation hinges on novel conditions that characterize when the optimal policies may take the form of either "forward" or "reverse" threshold policies. (ii) We exploit the optimality of threshold policies to build fast algorithms for computing the Whittle index, including a closed-form. (iii) We evaluate our algorithm on several data distributions including data from a real-world healthcare task in which a worker must monitor and deliver interventions to maximize their patients' adherence to tuberculosis medication. Our algorithm achieves a 3-order-of-magnitude speedup compared to state-of-the-art RMAB techniques while achieving similar performance. \ No newline at end of file diff --git a/data/2020/neurips/Collegial Ensembles b/data/2020/neurips/Collegial Ensembles new file mode 100644 index 0000000000..142a094d21 --- /dev/null +++ b/data/2020/neurips/Collegial Ensembles @@ -0,0 +1 @@ +Modern neural network performance typically improves as model size increases. A recent line of research on the Neural Tangent Kernel (NTK) of over-parameterized networks indicates that the improvement with size increase is a product of a better conditioned loss landscape. In this work, we investigate a form of over-parameterization achieved through ensembling, where we define collegial ensembles (CE) as the aggregation of multiple independent models with identical architectures, trained as a single model. We show that the optimization dynamics of CE simplify dramatically when the number of models in the ensemble is large, resembling the dynamics of wide models, yet scale much more favorably. We use recent theoretical results on the finite width corrections of the NTK to perform efficient architecture search in a space of finite width CE that aims to either minimize capacity, or maximize trainability under a set of constraints. The resulting ensembles can be efficiently implemented in practical architectures using group convolutions and block diagonal layers. Finally, we show how our framework can be used to analytically derive optimal group convolution modules originally found using expensive grid searches, without having to train a single model. \ No newline at end of file diff --git a/data/2020/neurips/Color Visual Illusions: A Statistics-based Computational Model b/data/2020/neurips/Color Visual Illusions: A Statistics-based Computational Model new file mode 100644 index 0000000000..e899995a86 --- /dev/null +++ b/data/2020/neurips/Color Visual Illusions: A Statistics-based Computational Model @@ -0,0 +1 @@ +Visual illusions may be explained by the likelihood of patches in real-world images, as argued by input-driven paradigms in Neuro-Science. However, neither the data nor the tools existed in the past to extensively support these explanations. The era of big data opens a new opportunity to study input-driven approaches. We introduce a tool that computes the likelihood of patches, given a large dataset to learn from. Given this tool, we present a model that supports the approach and explains lightness and color visual illusions in a unified manner. Furthermore, our model generates visual illusions in natural images, by applying the same tool, reversely. \ No newline at end of file diff --git a/data/2020/neurips/Combining Deep Reinforcement Learning and Search for Imperfect-Information Games b/data/2020/neurips/Combining Deep Reinforcement Learning and Search for Imperfect-Information Games new file mode 100644 index 0000000000..e4e69783cb --- /dev/null +++ b/data/2020/neurips/Combining Deep Reinforcement Learning and Search for Imperfect-Information Games @@ -0,0 +1 @@ +The combination of deep reinforcement learning and search at both training and test time is a powerful paradigm that has led to a number of a successes in single-agent settings and perfect-information games, best exemplified by the success of AlphaZero. However, algorithms of this form have been unable to cope with imperfect-information games. This paper presents ReBeL, a general framework for self-play reinforcement learning and search for imperfect-information games. In the simpler setting of perfect-information games, ReBeL reduces to an algorithm similar to AlphaZero. Results show ReBeL leads to low exploitability in benchmark imperfect-information games and achieves superhuman performance in heads-up no-limit Texas hold'em poker, while using far less domain knowledge than any prior poker AI. We also prove that ReBeL converges to a Nash equilibrium in two-player zero-sum games in tabular settings. \ No newline at end of file diff --git a/data/2020/neurips/Community detection in sparse time-evolving graphs with a dynamical Bethe-Hessian b/data/2020/neurips/Community detection in sparse time-evolving graphs with a dynamical Bethe-Hessian new file mode 100644 index 0000000000..5fedc70f49 --- /dev/null +++ b/data/2020/neurips/Community detection in sparse time-evolving graphs with a dynamical Bethe-Hessian @@ -0,0 +1 @@ +This article considers the problem of community detection in sparse dynamical graphs in which the community structure evolves over time. A fast spectral algorithm based on an extension of the Bethe-Hessian matrix is proposed, which benefits from the positive correlation in the class labels and in their temporal evolution and is designed to be applicable to any dynamical graph with a community structure. Under the dynamical degree-corrected stochastic block model, in the case of two classes of equal size, we demonstrate and support with extensive simulations that our proposed algorithm is capable of making non-trivial community reconstruction as soon as theoretically possible, thereby reaching the optimal detectability thresholdand provably outperforming competing spectral methods. \ No newline at end of file diff --git "a/data/2020/neurips/Community detection using fast low-cardinality semidefinite programming\342\200\251" "b/data/2020/neurips/Community detection using fast low-cardinality semidefinite programming\342\200\251" new file mode 100644 index 0000000000..b37be71ba6 --- /dev/null +++ "b/data/2020/neurips/Community detection using fast low-cardinality semidefinite programming\342\200\251" @@ -0,0 +1 @@ +Modularity maximization has been a fundamental tool for understanding the community structure of a network, but the underlying optimization problem is nonconvex and NP-hard to solve. State-of-the-art algorithms like the Louvain or Leiden methods focus on different heuristics to help escape local optima, but they still depend on a greedy step that moves node assignment locally and is prone to getting trapped. In this paper, we propose a new class of low-cardinality algorithm that generalizes the local update to maximize a semidefinite relaxation derived from max-k-cut. This proposed algorithm is scalable, empirically achieves the global semidefinite optimality for small cases, and outperforms the state-of-the-art algorithms in real-world datasets with little additional time cost. From the algorithmic perspective, it also opens a new avenue for scaling-up semidefinite programming when the solutions are sparse instead of low-rank. \ No newline at end of file diff --git a/data/2020/neurips/Compact task representations as a normative model for higher-order brain activity b/data/2020/neurips/Compact task representations as a normative model for higher-order brain activity new file mode 100644 index 0000000000..66d99575c2 --- /dev/null +++ b/data/2020/neurips/Compact task representations as a normative model for higher-order brain activity @@ -0,0 +1 @@ +Higher-order brain areas such as the frontal cortices are considered essential for the flexible solution of tasks. However, the precise computational role of these areas is still debated. Indeed, even for the simplest of tasks, we cannot really explain how the measured brain activity, which evolves over time in complicated ways, relates to the task structure. Here, we follow a normative approach, based on integrating the principle of efficient coding with the framework of Markov decision processes (MDP). More specifically, we focus on MDPs whose state is based on action-observation histories, and we show how to compress the state space such that unnecessary redundancy is eliminated, while task-relevant information is preserved. We show that the efficiency of a state space representation depends on the (long-term) behavioural goal of the agent, and we distinguish between model-based and habitual agents. We apply our approach to simple tasks that require short-term memory, and we show that the efficient state space representations reproduce the key dynamical features of recorded neural activity in frontal areas (such as ramping, sequentiality, persistence). If we additionally assume that neural systems are subject to cost-accuracy tradeoffs, we find a surprising match to neural data on a population level. \ No newline at end of file diff --git a/data/2020/neurips/Comparator-Adaptive Convex Bandits b/data/2020/neurips/Comparator-Adaptive Convex Bandits new file mode 100644 index 0000000000..eff32f1e64 --- /dev/null +++ b/data/2020/neurips/Comparator-Adaptive Convex Bandits @@ -0,0 +1 @@ +We study bandit convex optimization methods that adapt to the norm of the comparator, a topic that has only been studied before for its full-information counterpart. Specifically, we develop convex bandit algorithms with regret bounds that are small whenever the norm of the comparator is small. We first use techniques from the full-information setting to develop comparator-adaptive algorithms for linear bandits. Then, we extend the ideas to convex bandits with Lipschitz or smooth loss functions, using a new single-point gradient estimator and carefully designed surrogate losses. \ No newline at end of file diff --git a/data/2020/neurips/Complex Dynamics in Simple Neural Networks: Understanding Gradient Flow in Phase Retrieval b/data/2020/neurips/Complex Dynamics in Simple Neural Networks: Understanding Gradient Flow in Phase Retrieval new file mode 100644 index 0000000000..e328964666 --- /dev/null +++ b/data/2020/neurips/Complex Dynamics in Simple Neural Networks: Understanding Gradient Flow in Phase Retrieval @@ -0,0 +1 @@ +Despite the widespread use of gradient-based algorithms for optimizing high-dimensional non-convex functions, understanding their ability of finding good minima instead of being trapped in spurious ones remains to a large extent an open problem. Here we focus on gradient flow dynamics for phase retrieval from random measurements. When the ratio of the number of measurements over the input dimension is small the dynamics remains trapped in spurious minima with large basins of attraction. We find analytically that above a critical ratio those critical points become unstable developing a negative direction toward the signal. By numerical experiments we show that in this regime the gradient flow algorithm is not trapped; it drifts away from the spurious critical points along the unstable direction and succeeds in finding the global minimum. Using tools from statistical physics we characterize this phenomenon, which is related to a BBP-type transition in the Hessian of the spurious minima. \ No newline at end of file diff --git a/data/2020/neurips/Compositional Explanations of Neurons b/data/2020/neurips/Compositional Explanations of Neurons new file mode 100644 index 0000000000..d0eac2d450 --- /dev/null +++ b/data/2020/neurips/Compositional Explanations of Neurons @@ -0,0 +1 @@ +We describe a procedure for explaining neurons in deep representations by identifying compositional logical concepts that closely approximate neuron behavior. Compared to prior work that uses atomic labels as explanations, analyzing neurons compositionally allows us to more precisely and expressively characterize their behavior. We use this procedure to answer several questions on interpretability in models for vision and natural language processing. First, we examine the kinds of abstractions learned by neurons. In image classification, we find that many neurons learn highly abstract but semantically coherent visual concepts, while other polysemantic neurons detect multiple unrelated features; in natural language inference (NLI), neurons learn shallow lexical heuristics from dataset biases. Second, we see whether compositional explanations give us insight into model performance: vision neurons that detect human-interpretable concepts are positively correlated with task performance, while NLI neurons that fire for shallow heuristics are negatively correlated with task performance. Finally, we show how compositional explanations provide an accessible way for end users to produce simple "copy-paste" adversarial examples that change model behavior in predictable ways. \ No newline at end of file diff --git a/data/2020/neurips/Compositional Generalization by Learning Analytical Expressions b/data/2020/neurips/Compositional Generalization by Learning Analytical Expressions new file mode 100644 index 0000000000..0b7473a877 --- /dev/null +++ b/data/2020/neurips/Compositional Generalization by Learning Analytical Expressions @@ -0,0 +1 @@ +Compositional generalization is a basic but essential intellective capability of human beings, which allows us to recombine known parts readily. However, existing neural network based models have been proven to be extremely deficient in such a capability. Inspired by work in cognition which argues compositionality can be captured by variable slots with symbolic functions, we present a refreshing view that connects a memory-augmented neural model with analytical expressions, to achieve compositional generalization. Our model consists of two cooperative neural modules Composer and Solver, fitting well with the cognitive argument while still being trained in an end-to-end manner via a hierarchical reinforcement learning algorithm. Experiments on a well-known benchmark SCAN demonstrate that our model seizes a great ability of compositional generalization, solving all challenges addressed by previous works with 100% accuracies. \ No newline at end of file diff --git a/data/2020/neurips/Compositional Generalization via Neural-Symbolic Stack Machines b/data/2020/neurips/Compositional Generalization via Neural-Symbolic Stack Machines new file mode 100644 index 0000000000..d43944df28 --- /dev/null +++ b/data/2020/neurips/Compositional Generalization via Neural-Symbolic Stack Machines @@ -0,0 +1 @@ +Despite achieving tremendous success, existing deep learning models have exposed limitations in compositional generalization, the capability to learn compositional rules and apply them to unseen cases in a systematic manner. To tackle this issue, we propose the Neural-Symbolic Stack Machine (NeSS). It contains a neural network to generate traces, which are then executed by a symbolic stack machine enhanced with sequence manipulation operations. NeSS combines the expressive power of neural sequence models with the recursion supported by the symbolic stack machine. Without training supervision on execution traces, NeSS achieves 100% generalization performance in four domains: the SCAN benchmark of language-driven navigation tasks, the task of few-shot learning of compositional instructions, the compositional machine translation benchmark, and context-free grammar parsing tasks. \ No newline at end of file diff --git a/data/2020/neurips/Compositional Visual Generation with Energy Based Models b/data/2020/neurips/Compositional Visual Generation with Energy Based Models new file mode 100644 index 0000000000..63cbe75c66 --- /dev/null +++ b/data/2020/neurips/Compositional Visual Generation with Energy Based Models @@ -0,0 +1 @@ +A vital aspect of human intelligence is the ability to compose increasingly complex concepts out of simpler ideas, enabling both rapid learning and adaptation of knowledge. In this paper we show that energy-based models can exhibit this ability by directly combining probability distributions. Samples from the combined distribution correspond to compositions of concepts. For example, given one distribution for smiling face images, and another for male faces, we can combine them to generate smiling male faces. This allows us to generate natural images that simultaneously satisfy conjunctions, disjunctions, and negations of concepts. We evaluate compositional generation abilities of our model on the CelebA dataset of natural faces and synthetic 3D scene images. We showcase the breadth of unique capabilities of our model, such as the ability to continually learn and incorporate new concepts, or infer compositions of concept properties underlying an image. \ No newline at end of file diff --git a/data/2020/neurips/Compositional Zero-Shot Learning via Fine-Grained Dense Feature Composition b/data/2020/neurips/Compositional Zero-Shot Learning via Fine-Grained Dense Feature Composition new file mode 100644 index 0000000000..be8af1b865 --- /dev/null +++ b/data/2020/neurips/Compositional Zero-Shot Learning via Fine-Grained Dense Feature Composition @@ -0,0 +1 @@ +We develop a novel generative model for zero-shot learning to recognize finegrained unseen classes without training samples. Our observation is that generating holistic features of unseen classes fails to capture every attribute needed to distinguish small differences among classes. We propose a feature composition framework that learns to extract attribute-based features from training samples and combines them to construct fine-grained features for unseen classes. Feature composition allows us to not only selectively compose features of unseen classes from only relevant training samples, but also obtain diversity among composed features via changing samples used for composition. In addition, instead of building a global feature of an unseen class, we use all attribute-based features to form a dense representation consisting of fine-grained attribute details. To recognize unseen classes, we propose a novel training scheme that uses a discriminative model to construct features that are subsequently used to train itself. Therefore, we directly train the discriminative model on composed features without learning separate generative models. We conduct experiments on four popular datasets of DeepFashion, AWA2, CUB, and SUN, showing that our method significantly improves the state of the art. \ No newline at end of file diff --git a/data/2020/neurips/Comprehensive Attention Self-Distillation for Weakly-Supervised Object Detection b/data/2020/neurips/Comprehensive Attention Self-Distillation for Weakly-Supervised Object Detection new file mode 100644 index 0000000000..1613d6ca93 --- /dev/null +++ b/data/2020/neurips/Comprehensive Attention Self-Distillation for Weakly-Supervised Object Detection @@ -0,0 +1 @@ +Weakly Supervised Object Detection (WSOD) has emerged as an effective tool to train object detectors using only the image-level category labels. However, without object-level labels, WSOD detectors are prone to detect bounding boxes on salient objects, clustered objects and discriminative object parts. Moreover, the image-level category labels do not enforce consistent object detection across different transformations of the same images. To address the above issues, we propose a Comprehensive Attention Self-Distillation (CASD) training approach for WSOD. To balance feature learning among all object instances, CASD computes the comprehensive attention aggregated from multiple transformations and feature layers of the same images. To enforce consistent spatial supervision on objects, CASD conducts self-distillation on the WSOD networks, such that the comprehensive attention is approximated simultaneously by multiple transformations and feature layers of the same images. CASD produces new state-of-the-art WSOD results on standard benchmarks such as PASCAL VOC 2007/2012 and MS-COCO. \ No newline at end of file diff --git a/data/2020/neurips/Compressing Images by Encoding Their Latent Representations with Relative Entropy Coding b/data/2020/neurips/Compressing Images by Encoding Their Latent Representations with Relative Entropy Coding new file mode 100644 index 0000000000..6f15839369 --- /dev/null +++ b/data/2020/neurips/Compressing Images by Encoding Their Latent Representations with Relative Entropy Coding @@ -0,0 +1 @@ +Variational Autoencoders (VAEs) have seen widespread use in learned image compression. They are used to learn expressive latent representations on which downstream compression methods can operate with high efficiency. Recently proposed 'bits-back' methods can indirectly encode the latent representation of images with codelength close to the relative entropy between the latent posterior and the prior. However, due to the underlying algorithm, these methods can only be used for lossless compression, and they only achieve their nominal efficiency when compressing multiple images simultaneously; they are inefficient for compressing single images. As an alternative, we propose a novel method, Relative Entropy Coding (REC), that can directly encode the latent representation with codelength close to the relative entropy for single images, supported by our empirical results obtained on the Cifar10, ImageNet32 and Kodak datasets. Moreover, unlike previous bits-back methods, REC is immediately applicable to lossy compression, where it is competitive with the state-of-the-art on the Kodak dataset. \ No newline at end of file diff --git a/data/2020/neurips/Computing Valid p-value for Optimal Changepoint by Selective Inference using Dynamic Programming b/data/2020/neurips/Computing Valid p-value for Optimal Changepoint by Selective Inference using Dynamic Programming new file mode 100644 index 0000000000..e1e14b0488 --- /dev/null +++ b/data/2020/neurips/Computing Valid p-value for Optimal Changepoint by Selective Inference using Dynamic Programming @@ -0,0 +1 @@ +There is a vast body of literature related to methods for detecting changepoints (CP). However, less attention has been paid to assessing the statistical reliability of the detected CPs. In this paper, we introduce a novel method to perform statistical inference on the significance of the CPs, estimated by a Dynamic Programming (DP)-based optimal CP detection algorithm. Based on the selective inference (SI) framework, we propose an exact (non-asymptotic) approach to compute valid p-values for testing the significance of the CPs. Although it is well-known that SI has low statistical power because of over-conditioning, we address this disadvantage by introducing parametric programming techniques. Then, we propose an efficient method to conduct SI with the minimum amount of conditioning, leading to high statistical power. We conduct experiments on both synthetic and real-world datasets, through which we offer evidence that our proposed method is more powerful than existing methods, has decent performance in terms of computational efficiency, and provides good results in many practical applications. \ No newline at end of file diff --git a/data/2020/neurips/Conditioning and Processing: Techniques to Improve Information-Theoretic Generalization Bounds b/data/2020/neurips/Conditioning and Processing: Techniques to Improve Information-Theoretic Generalization Bounds new file mode 100644 index 0000000000..cc14e4c19f --- /dev/null +++ b/data/2020/neurips/Conditioning and Processing: Techniques to Improve Information-Theoretic Generalization Bounds @@ -0,0 +1 @@ +Obtaining generalization bounds for learning algorithms is one of the main subjects studied in theoretical machine learning. In recent years, information-theoretic bounds on generalization have gained the attention of researchers. This approach provides an insight into learning algorithms by considering the mutual information between the model and the training set. In this paper, a probabilistic graphical representation of this approach is adopted and two general techniques to improve the bounds are introduced, namely conditioning and processing. In conditioning, a random variable in the graph is considered as given, while in processing a random variable is substituted with one of its children. These techniques can be used to improve the bounds by either sharpening them or increasing their applicability. It is demonstrated that the proposed framework provides a simple and unified way to explain a variety of recent tightening results. New improved bounds derived by utilizing these techniques are also proposed. \ No newline at end of file diff --git a/data/2020/neurips/Confidence sequences for sampling without replacement b/data/2020/neurips/Confidence sequences for sampling without replacement new file mode 100644 index 0000000000..4b6afefaa3 --- /dev/null +++ b/data/2020/neurips/Confidence sequences for sampling without replacement @@ -0,0 +1 @@ +Many practical tasks involve sampling sequentially without replacement (WoR) from a finite population of size $N$, in an attempt to estimate some parameter $\theta^\star$. Accurately quantifying uncertainty throughout this process is a nontrivial task, but is necessary because it often determines when we stop collecting samples and confidently report a result. We present a suite of tools for designing confidence sequences (CS) for $\theta^\star$. A CS is a sequence of confidence sets $(C_n)_{n=1}^N$, that shrink in size, and all contain $\theta^\star$ simultaneously with high probability. We first exploit a relationship between Bayesian posteriors and martingales to construct a (frequentist) CS for the parameters of a hypergeometric distribution. We then present Hoeffding- and empirical-Bernstein-type time-uniform CSs and fixed-time confidence intervals for sampling WoR which improve on previous bounds in the literature. \ No newline at end of file diff --git a/data/2020/neurips/Conformal Symplectic and Relativistic Optimization b/data/2020/neurips/Conformal Symplectic and Relativistic Optimization new file mode 100644 index 0000000000..d81128b440 --- /dev/null +++ b/data/2020/neurips/Conformal Symplectic and Relativistic Optimization @@ -0,0 +1 @@ +Arguably, the two most popular accelerated or momentum-based optimization methods in machine learning are Nesterov’s accelerated gradient and Polyaks’s heavy ball, both corresponding to different discretizations of a particular second order differential equation with friction. Such connections with continuous-time dynamical systems have been instrumental in demystifying acceleration phenomena in optimization. Here we study structure-preserving discretizations for a certain class of dissipative (conformal) Hamiltonian systems, allowing us to analyse the symplectic structure of both Nesterov and heavy ball, besides providing several new insights into these methods. Moreover, we propose a new algorithm based on a dissipative relativistic system that normalizes the momentum and may result in more stable/faster optimization. Importantly, such a method generalizes both Nesterov and heavy ball, each being recovered as distinct limiting cases, and has potential advantages at no additional cost. \ No newline at end of file diff --git a/data/2020/neurips/Confounding-Robust Policy Evaluation in Infinite-Horizon Reinforcement Learning b/data/2020/neurips/Confounding-Robust Policy Evaluation in Infinite-Horizon Reinforcement Learning new file mode 100644 index 0000000000..296af08253 --- /dev/null +++ b/data/2020/neurips/Confounding-Robust Policy Evaluation in Infinite-Horizon Reinforcement Learning @@ -0,0 +1 @@ +Off-policy evaluation of sequential decision policies from observational data is necessary in applications of batch reinforcement learning such as education and healthcare. In such settings, however, unobserved variables confound observed actions, rendering exact evaluation of new policies impossible, i.e., unidentifiable. We develop a robust approach that estimates sharp bounds on the (unidentifiable) value of a given policy in an infinite-horizon problem given data from another policy with unobserved confounding, subject to a sensitivity model. We consider stationary or baseline unobserved confounding and compute bounds by optimizing over the set of all stationary state-occupancy ratios that agree with a new partially identified estimating equation and the sensitivity model. We prove convergence to the sharp bounds as we collect more confounded data. Although checking set membership is a linear program, the support function is given by a difficult nonconvex optimization problem. We develop approximations based on nonconvex projected gradient descent and demonstrate the resulting bounds empirically. \ No newline at end of file diff --git a/data/2020/neurips/Conic Descent and its Application to Memory-efficient Optimization over Positive Semidefinite Matrices b/data/2020/neurips/Conic Descent and its Application to Memory-efficient Optimization over Positive Semidefinite Matrices new file mode 100644 index 0000000000..702ff7c833 --- /dev/null +++ b/data/2020/neurips/Conic Descent and its Application to Memory-efficient Optimization over Positive Semidefinite Matrices @@ -0,0 +1 @@ +We present an extension of the conditional gradient method to problems whose feasible sets are convex cones. We provide a convergence analysis for the method and for variants with nonconvex objectives, and we extend the analysis to practical cases with effective line search strategies. For the specific case of the positive semidefinite cone, we present a memory-efficient version based on randomized matrix sketches and advocate a heuristic greedy step that greatly improves its practical performance. Numerical results on phase retrieval and matrix completion problems indicate that our method can offer substantial advantages over traditional conditional gradient and Burer-Monteiro approaches \ No newline at end of file diff --git a/data/2020/neurips/Consequences of Misaligned AI b/data/2020/neurips/Consequences of Misaligned AI new file mode 100644 index 0000000000..0e7d7fc3e6 --- /dev/null +++ b/data/2020/neurips/Consequences of Misaligned AI @@ -0,0 +1 @@ +AI systems often rely on two key components: a specified goal or reward function and an optimization algorithm to compute the optimal behavior for that goal. This approach is intended to provide value for a principal: the user on whose behalf the agent acts. The objectives given to these agents often refer to a partial specification of the principal's goals. We consider the cost of this incompleteness by analyzing a model of a principal and an agent in a resource constrained world where the $L$ attributes of the state correspond to different sources of utility for the principal. We assume that the reward function given to the agent only has support on $J