In this lecture, we will discuss key concepts in meta-learning.
Guest lecture by Hugo Larochelle
Lecture Slides: PDF, additional slides
Reference:
See slides for references to relevant papers.
In this lecture, we will discuss self-supervised learning. We will discuss how to create representation beyond the supervised pre-training paradigm, and we are going to see how effective pretext tasks can be designed and how to train with contrastive objectives.
Guest lecture by Olivier J. Hénaff
Lecture Slides: PDF Google slides
Reference:
See slides for references to relevant papers.
In this lecture, we will discuss the most recent advances in generative modeling brought upon by denoising diffusion based models.
Guest lecture by Vikram Voleti
Lecture Slides: PDF Google slides
Reference:
See slides for references to relevant papers.
In this lecture, we will discuss Generative Adversarial Networks (GANs). GANs are a recent and very popular generative model paradigm. We will discuss the GAN formalism, some theory and practical considerations.
Lecture Slides: PDF Keynote
(slides built on top of Aaron Courville's course slides)
Reference:
Sections 20.10.4 of the Deep Learning textbook.
Generative Adversarial Networks by Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, Yoshua Bengio (NIPS 2014).
f-GAN: Training Generative Neural Samplers using Variational Divergence Minimization by Sebastian Nowozin, Botond Cseke and Ryota Tomioka (NIPS 2016).
NIPS 2016 Tutorial: Generative Adversarial Networks by Ian Goodfellow, arXiv:1701.00160v1, 2016
Adversarially Learned Inference by Vincent Dumoulin , Ishmael Belghazi , Ben Poole, Olivier Mastropietro, Alex Lamb, Martin Arjovsky and Aaron Courville (ICLR 2017).
Many others references in the slides.
Introduction to Normalizing Flows. We will see how such methods can be used as generative models by inverting the transformation of the data distribution into a prior distribution.
Lecture Slides: PDF Keynote
(slides built on top of Aaron Courville's course slides)
Reference:
Chapter 20.10.2 of the Deep Learning textbook.
Chapter 1-2 (for the core idea) and Chapter 6 (for applications) Normalizing Flows for Probabilistic Modeling and Inference by George Papamakarios and friends.
In this lecture, we will discuss a family of latent variable models known as the Variational Autoencoders (VAE). We’ll see how a deep latent gaussian model can be seen as an autoencoder via amortized variational inference, and how such an autoencoder can be used as a generative model. At the end, we’ll take a look at variants of VAE and different ways to improve inference.
Lecture Slides: PDF Keynote
(slides built on top of Aaron Courville's course slides)
Reference:
Chapter 20.10.3 of the Deep Learning textbook.
Chapter 2 of An Introduction to Variational Autoencoders by Kingma and Welling
Inference Suboptimality in Variational Autoencoders by Chris Cremer (ICML 2018)
Importance Weighted Autoencoders by Yuri Burda (ICLR 2016)
Variational Inference, lecture note by David Blei. Section 1-6.
Blog post Variational Autoencoder Explained by Goker Erdogan
Blog post Families of Generative Models by Andre Cianflone
Introduction to autoencoders and different types of autoregressive generative models.
Lecture Slides:
Autoencoders: PDF Keynote
Autoregressive Generative Models : PDF Keynote
(slides built on top of Aaron Courville's course slides)
Reference:
Chapter 13-14 of the Deep Learning textbook.
Sections 20.10.5-20.10.10 of the Deep Learning textbook.
The Neural Autoregressive Distribution Estimator by Hugo Larochelle and Iain Murray (AISTAT2011)
MADE: Masked Autoencoder for Distribution Estimation by Mathieu Germain, Karol Gregor, Iain Murray, Hugo Larochelle (ICML2015).
Pixel Recurrent Neural Networks by Aaron van den Oord, Nal Kalchbrenner, Koray Kavukcuoglu (ICML2016)
Conditional Image Generation with PixelCNN Decoders by Aaron van den Oord, Nal Kalchbrenner, Oriol Vinyals, Lasse Espeholt, Alex Graves, Koray Kavukcuoglu (NIPS2016)
Blog post explaining the blind spot problem in PixelCNNs and how it is fixed
Detailed discussion of regularization methods and their interpretation.
Lecture Slides:
Part I: Parameter Regularization: PDF Keynote
Part II: Dropout: PDF Keynote
(slides built on top of Aaron Courville's course slides)
Reference:
Chapter 7 of the Deep Learning textbook.
Understanding deep learning requires rethinking generalization (ICLR 2017) by Chiyuan Zhang, Samy Bengio, Moritz Hardt, Benjamin Recht, Oriol Vinyals
Discussion on popular and practical first-order optimization methods and discussion on normalization methods. We will not discuss second-order optimization methods but slides are provided below for your own reading (optional).
Lecture Slides: PDF Keynote (slides built on top of Aaron Courville's course slides)
(Optional) Slides for second-order optimization methods
Reference:
Chapter 8 of the Deep Learning textbook.
Why Momentum Really Works. Gabriel Goh, Distill 2017.
Introduction to Recurrent Neural Networks and related models.
Lecture Slides: PDF Keynote (slides built on top of Aaron Courville's course slides)
Reference:
Chapter 10 of the Deep Learning textbook (sections. 10.1-10.11)
Blog post on Understanding LSTM Networks by Chris Olah
This lecture will be an introductory tutorial to PyTorch. A Colab notebook will be provided for the tutorial. You are encouraged to bring your laptop to the class and follow along.
Material: Colab Notebook
This tutorial will cover
the torch.Tensor class, and important attributes and operations
automatic differentiation in pytorch
torch.nn and torch.optim modules
training MLPs and ConvNets on MNIST
Introduction to Convolutional Neural Networks.
Backprop in CNNs (Slides are from Hiroshi Kuwajima’s Memo on Backpropagation in Convolutional Neural Networks.) -- We won't go over these in class, but they are required reading and will be the basis of one of the questions in Assignment 1.
Lecture Slides: PDF Keynote (slides built on top of Aaron Courville's course slides)
Reference:
Chapter 9 of the Deep Learning textbook, Sections 9.10 and 9.11 are optional.
Andrej Karpathy’s excellent tutorial on CNNs.
Paper on convolution arithmetic by Vincent Dumoulin and Francesco Visin.
WaveNet Blog presenting dilated convolutions animation and samples.
Blog on Deconvolution and Checkerboard Artifacts by Augustus Odena, Vincent Dumoulin and Chris Olah.
Detailed discussion on how to train a feedforward neural network.
Lecture Slides: PDF Keynote (slides built on top of Aaron Courville's course slides)
Reference:
Chapter 6 of the Deep Learning textbook (by Ian Goodfellow, Yoshua Bengio and Aaron Courville).
Beginning of detailed introduction to neural networks.
Lecture Slides: PDF Keynote (slides built on top of Aaron Courville's course slides)
Reference:
Chapter 6 of the Deep Learning textbook (by Ian Goodfellow, Yoshua Bengio and Aaron Courville).
Provide an overview of latest developments in Deep Learning, what Deep Learning is and highlight some foundational material which serve as prerequisites for this course: linear algebra, calculus, and the basics of machine learning.
Lecture Slides: PDF Keynote (slides built on top of Aaron Courville's course slides)
Reference:
Chapters 1-5 of the Deep Learning textbook (by Ian Goodfellow, Yoshua Bengio and Aaron Courville).