Understanding human intelligence and how to replicate it in machines is arguably one of the greatest problems in science. Learning, its principles and computational implementations, is at the very core of intelligence. During the last two decades, for the first time, artificial intelligence systems have been developed that begin to solve complex tasks, until recently the exclusive domain of biological organisms, such as computer vision, speech recognition or natural language understanding: cameras recognize faces, smart phones understand voice commands, smart speakers/assistants answer questions and cars can see and avoid obstacles. The machine learning algorithms that are at the roots of these success stories are trained with examples rather than programmed to solve a task. However, a comprehensive theory of learning is still incomplete, as shown by the several puzzles of deep learning. An eventual theory of learning that explains why and how deep networks work and what their limitations are, may thus enable the development of even much more powerful learning approaches and even inform our understanding of human intelligence.
In this spirit, the course covers foundations and recent advances in statistical machine learning theory, with the dual goal a) of providing students with the theoretical knowledge and the intuitions needed to use effective machine learning solutions and b) to prepare more advanced students to contribute to progress in the field. This year the emphasis is again on b).
The course is organized about the core idea of supervised learning as an inverse problem, with stability as the key property required for good generalization performance of an algorithm.
The content is roughly divided into three parts. The first part --which is now shorter than in previous years -- is about classical regularization (regularized least squares, kernel machines, SVM, logistic regression, square and exponential loss) large margin/minimum norm, stochastic gradient methods, overparametrization, implicit regularization and also approximation/estimation errors. The second part is about deep networks: approximation theory -- which functions can be represented more efficiently by deep networks than shallow networks -- optimization theory -- why can stochastic gradient descent easily find global minima -- and estimation error -- how generalization in deep networks can be explained in terms of the complexity control implicit in SGD. The third part is about a few topics of current research, starting with the connections between learning theory and the brain, which was the original inspiration for modern networks and may provide ideas for future developments and breakthroughs in the theory and the algorithms of leaning. Throughout the course, and especially in the final classes, we will have occasional talks by leading researchers on selected advanced research topics. This class is a step in a NSF funded project in which a team of deep learning researchers from GeorgiaTech, Columbia. Princeton and MIT is developing courses that leverages the recent advances in our understanding of deep learning.
Apart for the first part on regularization, which is an essential part of any introduction to the field of machine learning, this year course is designed for students with a good background in ML.
We will make extensive use of basic notions of calculus, linear algebra and probability. The essentials are covered in class and in the math camp material. We will introduce a few concepts in functional/convex analysis and optimization. Note that this is an advanced graduate course and some exposure on introductory Machine Learning concepts or courses is expected: for course 6 students prerequisites are 6.041 and 18.06 and (6.036 or 6.401 or 6.867). Students are also expected to have basic familiarity with MATLAB/Octave.
Requirements for grading are attending lectures/participation (10%), three problem sets (30%) and a final project (60%). Use of LLMs -- such as Chat-GPT -- is allowed but we ask students to report if they do use it, to write the prompts they used and to provide feedback on its usefulness.
Classes will be conducted in-person this year (Fall 2023), unless MIT policy changes.
Grading policies, pset and project tentative dates: (slides)
Problem Set 1, out: Tue. Sept. 19, due: Tue. Oct. 03 (before class)
Problem Set 2, out: Thu. Oct. 03, due: Thu. Oct. 17 (before class)
Problem Set 3, out: Thu. Oct. 17, due: Thu. Oct. 31 (before class)
Submission instructions: Follow the instructions included with the problem set. Use the latex template for the report. Submit your report online through canvas by the due date/time.
Final project reports (5 pages for individuals, 8 pages for teams, NeurIPS style) are due on Tue. Dec. 12.
List of Wikipedia entries, created or edited as part of projects during previous course offerings.
This document has more information about navigating student resources at MIT