Slides:
In the practice of learning from data, we make many assumptions - some fundamental to the theory of ML, some practical, and some implicit. This lecture attempts to identify some of these assumptions, and ways we can deal with breaking them. It covers the IID assumptions, systematic generalization (touching related ideas in causality), distributional shift, online/continual/open-set learning, and mentions some results in statistical learning theory and empirical investigations of deep network learning behaviour as well as in FATES (Fairness, Accountability, Transparency, Ethics, and Safety).
Slides:
References:
A lot of the recent progress on many AI tasks was enable in part by the availability of large quantities of labeled data. Yet, humans are able to learn concepts from as little as a handful of examples. Meta-learning is a very promising framework for addressing the problem of generalizing from small amounts of data, known as few-shot learning. In meta-learning, our model is itself a learning algorithm: it takes as input a training set and outputs a classifier. For few-shot learning, it is (meta-)trained directly to produce classifiers with good generalization performance for problems with very little labeled data. In this talk, I'll present an overview of the recent research that has made exciting progress on this topic (including my own) and, if time permits, will discuss the challenges as well as research opportunities that remain.
Slides:Meta-Learning slides
In this lecture, we will discuss Generative Adversarial Networks (GANs). GANs are a recent and very popular generative model paradigm. We will discuss the GAN formalism, some theory and practical considerations.
Slides:
Reference: (* = you are responsible for this material)
In this lecture, we will finish the inference suboptimality part of the VAE lecture, have a crash course on Normalizing Flows, and see how they can be used (1) to reduce the approximation gap of VAEs by using a more flexible family of variational distributions, and (2) as a generative model by inverting the transformation of the data distribution into a prior distribution.
Slides:
Reference: (* = you are responsible for this material)
In this lecture, Chin-Wei will talk about a family of latent variable models known as the Variational Autoencoders (VAE). We’ll see how a deep latent gaussian model can be seen as an autoencoder via amortized variational inference, and how such an autoencoder can be used as a generative model. At the end, we’ll take a look at variants of VAE and different ways to improve inference.
Slides:
Reference: (* = you are responsible for this material)
In this lecture we will take a closer look at a form of neural network known as an Autoencoder. We will also begin our look at generative models with Autoregressive Models.
Slides:
Reference: (* = you are responsible for this material)
In this talk Arian Hosseini will look at self attention and the transformer model. We will see how they work, dig deep into them, see analysis and performances, and their applications (in language, vision and speech).
Slides:
Reference:
In this lecture prepared by Dzmitry (Dima) Bahdanau, I will discuss attention in neural networks.
Slides:
Reference: (* = you are responsible for this material)
Devansh Arpit will introduce a number of normalization techniques that have become very popular in training deep neural networks.
Slides and Notes:
Reference: (* = you are responsible for this material)
In this lecture, we will discussion both popular and practical first-order optimization methods and - if time permits - discuss some approximate second-order methods and their interpretation.
Slides:
Reference: (* = you are responsible for this material)
In these lectures, we will have a rather detailed discussion of regularization methods and their interpretation.
Slides:
Reference: (* = you are responsible for this material)
In this lecture we introduce Recurrent Neural Networks and related models.
Lecture 08 RNNs (slides derived from Hugo Larochelle)
Reference: (* = you are responsible for this material)
Today we conclude our discussion of convolutional neural networks.
Lecture 05 CNNs II (Slides are from Hiroshi Kuwajima’s Memo on Backpropagation in Convolutional Neural Networks.)
Reference: (* = you are responsible for this material)
In these lectures we will have a PyTorch Tutorial and a question answering session.
The tutorial will cover
In this lecture we finish up our discussion of training neural networks and we introduce Convolutional Neural Networks.
Lecture 04 CNNs I (some slides are modified from Hugo Larochelle’s course notes)
Reference: (* = you are responsible for all of this material)
In this lecture we continue with our introduction to neural networks. Specifically we will discuss how to train neural networks: i.e. the Backpropagation Algorithm
Lecture 03 training NNs (slides modified from Hugo Larochelle’s course notes)
Reference: (you are responsible for all of this material)
In this lecture we finish our overview of Machine Learning and begin our detailed introduction to Neural Networks.
Lecture 02 artificial neurons (slides from Hugo Larochelle’s course notes)
Reference: (you are responsible for all of this material)
The first class is January 7th, 2019. We discuss the plan for the course and the pedagogical method chosen. We also briefly review some foundational material, covering linear algebra, calculus, and the basics of machine learning.
Lecture 01 slides (slides built on Hugo Larochelle’s slides)
Reference: