9.520/6.7910: Statistical Learning Theory
and Applications

Fall 2024

Course Schedule

Reading List

Notes covering the classes will be provided in the form of independent chapters of a book currently in draft format. Additional information will be given through the slides associated with classes (where applicable). The books/papers listed below are useful general reference reading, especially from the theoretical viewpoint. A list of additional suggested readings will also be provided separately for each class.

Book (draft)

  • L. Rosasco and T. Poggio, Machine Learning: a Regularization Approach, MIT-9.520 Lectures Notes, Manuscript, Dec. 2017 (provided).

Primary References

Papers of Interest

  • T. Poggio, H. Mhaskar, L. Rosasco, B. Miranda, and Q. Liao. Why and When can Deep-but not Shallow-Networks Avoid the Curse of Dimensionality: A Review. International Journal of Automation and Computing, 1-17, 2017.
  • T. Poggio, M. Fraser  Compositional sparsity of learnable functions, Bulleting of the American Mathematical Society, 2024.
  • M. Xu, A. Rangamani, Q. Liao, T. Galanti, T. Poggio  .Dynamics in Deep Classifiers Trained with the Square Loss: Normalization, Low Rank, Neural Collapse, and Generalization Bounds, Research, vol 6, 2023.
  • Y. LeCun, Y. Bengio and G. Hinton, Deep Learning, ​Nature, 436-444, 2015.
  • D. Silver et al, Mastering the game of Go with deep neural networks and tree search. Nature, 484-489, 2016.
  • J. Jumper et al, Highly accurate protein structure prediction with AlphaFold. Nature, 583-589. 2021. 
  • A. Vaswani et al, Attention is all you need. Advances in neural information processing systems, 30. 2017.

Resources and links