The past five years have seen a huge increase in the capabilities of deep neural networks. Maintaining this rate of progress however, faces some steep challenges, and awaits fundamental insights. As our models become more complex, and venture into areas such as unsupervised learning or reinforcement learning, designing improvements becomes more laborious, and success can be brittle and hard to transfer to new settings.
This workshop seeks to highlight recent works that use theory as well as systematic experiments to isolate the fundamental questions that need to be addressed in deep learning. These have helped flesh out core questions on topics such as generalization, adversarial robustness, large batch training, generative adversarial nets, and optimization, and point towards elements of the theory of deep learning that is expected to emerge in the future.
The workshop aims to enhance this confluence of theory and practice, highlighting influential work with these methods, future open directions, and core fundamental problems. There will be an emphasis on discussion, via panels and round tables, to identify future research directions that are promising and tractable.
- Peter Bartlett (UC Berkeley)
- Yoshua Bengio (University of Montreal)
- Ian Goodfellow (Google Brain)
- Sham Kakade (University of Washington)
- Percy Liang (Stanford University)
- Doina Precup (McGill University)
Time | Event |
---|---|
8:35 - 8:45 | Opening Remarks |
8:45 - 9:15 | Yoshua Bengio: Generalization, Memorization and SGD |
9:15 - 9:45 | Ian Goodfellow: Bridging Theory and Practice of GANs |
9:45 - 10:00 | Spotlights 1 |
10:00 - 10:30 | Peter Bartlett: Generalization in Deep Networks |
10:30 - 11:00 | Coffee |
11:00 - 11:30 | Doina Precup: Experimental design in Deep Reinforcement Learning |
11:30 - 11:45 | Spotlights 2 |
11:45 - 1:30 | Lunch + first poster session |
1:30 - 2:00 | Percy Liang: Fighting Black Boxes, Adversaries, and Bugs in Deep Learning |
2:00 - 3:00 | Contributed talks |
3:00 - 4:00 | Coffee + second poster session |
4:00 - 4:30 | Sham Kakade: Towards Bridging Theory and Practice in DeepRL |
4:30 - 5:30 | Panel: Peter Bartlett, Yoshua Bengio, Sham Kakade, Percy Liang, Ruslan Salakhutdinov |
- Generalization in Deep Networks: The Role of Distance from Initialization
Vaishnavh Nagarajan and Zico Kolter - Entropy-SG(L)D optimizes the prior of a (valid) PAC-Bayes bound
Gintare Karolina Dziugaite and Daniel Roy - Large Batch Training of Deep Neural Networks with Layer-wise Adaptive Rate Scaling
Boris Ginsburg, Yang You, and Igor Gitman
- Measuring the Robustness of Neural Networks via Minimal Adversarial Examples
Sumanth Dathathri, Stephan Zheng, Sicun Gao, and Richard M. Murray - A Classification-Based Perspective on GAN Distributions
Shibani Santurkar, Ludwig Schmidt, and Aleksander Madry - Learning One-hidden-layer Neural Networks with Landscape Design
Rong Ge, Jason Lee, and Tengyu Ma
- Don't Decay the Learning Rate, Increase the Batch Size
Samuel L. Smith, Pieter-Jan Kindermans, and Quoc V. Le - Meta-Learning and Universality: Deep Representations and Gradient Descent Can Approximate Any Learning Algorithm
Chelsea Finn and Sergey Levine - Hyperparameter Optimization: A Spectral Approach
Elad Hazan, Adam Klivans, and Yang Yuan - Learning Implicit Generative Models with Method of Learned Moments
Suman Ravuri, Shakir Mohamed, Mihaela Rosca, and Oriol Vinyals
The posters are listed in order of submission.
- Rethinking Feature Discrimination and Polymerization for Large-scale Recognition
Yu Liu, Hongyang Li, and Xiaogang Wang - High dimensional dynamics of generalization error in neural networks
Madhu Advani and Andrew Saxe - On the importance of single directions for generalization
Ari S. Morcos, David G.T. Barrett, Neil C. Rabinowitz, and Matthew Botvinick - When is a Convolutional Filter Easy to Learn?
Simon Du, Jason Lee, and Yuandong Tian - A Pitfall of Unsupervised Pre-Training
Michele Alberti, Mathias Seuret, Rolf Ingold, and Marcus Liwicki - Competitive Learning Enriches Learning Representation and Accelerates the Fine-tuning of CNNs
Takashi Shinozaki - Reducing Duplicate Filters in Deep Neural Networks
Aruni Roychowdhury, Prakhar Sharma, and Erik Learned-Miller - Generalization Bounds of SGLD for Non-convex Learning: Two Theoretical Viewpoints
Wenlong Mou, Liwei Wang, Xiyu Zhai, and Kai Zheng - Towards Provable Learning of Polynomial Neural Networks Using Low-Rank Matrix Estimation
Mohammadreza Soltani and Chinmay Hegde - Learning Depth-Three Neural Networks in Polynomial Time
Surbhi Goel and Adam Klivans - Analyzing GANs with Generative Scattering Networks
Tomas Angles and Stephane Mallat - On Characterizing the Capacity of Neural Networks Using Algebraic Topology
William Guss and Ruslan Salakhutdinov - Deep Learning is Robust to Massive Label Noise
David Rolnick, Andreas Veit, Serge Belongie, and Nir Shavit - Network Approximation using Tensor Sketching
Shiva Kasiviswanathan, Nina Narodytska, and Hongxia Jin - Evaluation of Random Neural Layers in Deep Neural Networks
Corentin Hardy, Erwan Le Merrer, Gerardo Rubino, and Bruno Sericola - Training GANs with Optimism
Constantinos Daskalakis, Andrew Ilyas, Vasilis Syrgkanis, and Haoyang Zeng - GAN Connoisseur: Can GANs Learn Simple 1D Parametric Distributions?
Manzil Zaheer, Chun-Liang Li, Barnabas Poczos, and Ruslan Salakhutdinov - LSH Softmax: Sub-Linear Learning and Inference of the Softmax Layer in Deep Architectures
Daniel Levy, Danlu Chen, and Stefano Ermon - Global optimality conditions for deep neural networks
Chulhee Yun, Suvrit Sra, and Ali Jadbabaie - Visualizing the Loss Landscape of Neural Nets
Hao Li, Zheng Xu, Gavin Taylor, and Tom Goldstein - Semi-Supervised Learning with IPM-based GANs: an Empirical Study
Tom Sercu and Youssef Mroueh - Theoretical limitations of Encoder-Decoder GAN architectures
Sanjeev Arora, Andrej Risteski, and Yi Zhang - An Online Learning Approach to Generative Adversarial Networks
Paulina Grnarova, Kfir Levy, Aurelien Lucchi, Thomas Hofmann, and Andreas Krause - mixup: Beyond Empirical Risk Minimization
Hongyi Zhang, Moustapha Cisse, Yann Dauphin, and David Lopez-Paz - An Empirical Study of the Generalization Behavior of Generative Adversarial Networks
Hongyu Ren, Shengjia Zhao, Jiaming Song, Lijie Fan, and Stefano Ermon - Towards a testable notion of generalization for generative adversarial networks
Robert Cornish, Hongseok Yang, and Frank Wood - Sparse Coding and Autoencoders
Akshay Rangamani, Anirbit Mukherjee, Ashish Arora, Amitabh Basu, Tejaswini Ganapathi, Peter Chin, and Trac Tran - Investigating the working of text classifiers
Devendra Singh Sachan, Manzil Zaheer, and Russ Salakhutdinov - Intriguing Properties of Adversarial Examples
Ekin Dogus Cubuk, Barret Zoph, Samuel S. Schoenholz, and Quoc V. Le - Sparse-Gen: Combining Sparse Recovery and Generative Modeling for Compressed Sensing
Manik Dhar, Aditya Grover, and Stefano Ermon - Experiments & non-convex theory for sign-based methods in stochastic optimisation
Jeremy Bernstein, Kamyar Azizzadenesheli, Yu-Xiang Wang, and Anima Anandkumar - Exactly solvable nonlinear recurrent networks for finite state computation
Christopher Stock and Surya Ganguli - Training ultra-deep CNNs with critical initialization
Lechao Xiao, Yasaman Bahri, Sam Schoenholz, and Jeffrey Pennington
We invite researchers to submit anonymous extended abstracts of up to 4 pages (excluding references). No specific formatting is required. Authors may use the NIPS style file, or any other style as long as they have standard font size (11pt) and margins (1in).
Submit on https://easychair.org/conferences/?conf=dltp2017.
- Submission Deadline: (EXTENDED) Wednesday November 1st
- Notification: Saturday November 25th
- Workshop: Saturday December 9th
- Sanjeev Arora (Princeton University)
- Maithra Raghu (Cornell University and Google Brain)
- Ruslan Salakhutdinov (Carnegie Mellon University)
- Ludwig Schmidt (MIT)
- Oriol Vinyals (DeepMind)
Please email [email protected] with any questions.