Machine Learning for the Working Mathematician

The Machine Learning for the Working Mathematician seminar is an introduction to ways in which machine learning (and in particular deep learning) has been used to solve problems in mathematics. The seminar is an initiative of the Sydney Mathematical Research Institute (SMRI).

We aim for a toolbox of simple examples, where one can get a grasp on what machine learning can and cannot do. We want to emphasise techniques in machine learning as tools that can be used in mathematics research, rather than a source of problems in themselves. The first six weeks or so will be introductory, and the second six weeks will feature talks from experts on applications.

Two nice examples of recent work that give the 'flavour' of the seminar are:

The seminar is organised by Joel Gibson, Georg Gottwald, and Geordie Williamson.

Seminar Schedule

The seminar takes place during the first half of 2022, at the University of Sydney. The first few weeks will run on Thursdays at 3pm-5pm in Carslaw 273 (see a map), starting from the first week of the semester. There is no need to sign up, everyone is welcome! Come along in person or log on to zoom: the password is the first 8 letters of the word BackPropagation (note the capitalisation). Each lecture will be recorded and made available online.

The workshop sessions run at Friday 3pm-4pm in Carslaw 273 (same room as the seminar), where attendees can get some hands-on experience with applying machine learning to mathematical problems. The workshop sessions will involve Google Colab notebooks. The workshop sessions will be in-person only, but the notebooks will be made available online.

NEW! We have a discussion board, so we can keep chatting after the lectures and workshops. Sign up - the 8-letter invite code is the same as the Zoom password.


In this talk, we will develop a conceptual approach towards inverse problems in imaging sciences by combining the model-based method of sparse regularization by shearlets with the data-driven method of deep learning. Our solvers pay particular attention to the singularity structures of the data. Focussing then on the inverse problem of (limited-angle) computed tomography, we will show that our algorithms significantly outperform previous methodologies, including methods entirely based on deep learning. Finally, we will also touch upon the issue of how to interpret the results of such algorithms, and present a novel, state-of-the-art explainability method based on information theory.

References

The main reference for the first half of the course will be Deep Learning by Ian Goodfellow, Yoshua Bengio, and Aaron Courville, freely available online. More specific references given in lectures will also be listed here, in the seminar schedule.

Additional colabs

Here we provide links to a few additional colabs which might be interesting.

Classifying descents in Sn : Here we train a simple neural network to classify the descent sets of a permutation.

Playing with parity : Here we train a simple neurl net to learn the parity function, and look at how well it generalizes.

Header image by Mary and Andrew, Wikimedia commons