Resources
Textbook Information
This course will follow a few reference books and some papers
listed below:
Books (DRM-Free Electronic Copies available through MST Library)
Books and Lecture Notes (Accessible Electronic Copies for Free)
- Statistical Learning Theory, by Bruce Hajek and Maxim Raginsky, Lecture Notes for ECE 543, University of Illinois, Urbana-Champaign, IL, USA, 2021.
- Deep Learning Theory, by Matus J. Telgarsky, Lecture Notes for CS 540, Version: 2021-10-27 v0.0-e7150f2d (alpha), University of Illinois, Urbana-Champaign, IL, USA, 2021.
- Dive into Deep Learning, by Aston Zhang, Zachary C. Lipton, Mu Li, and Alexander J. Smola.
- Deep Learning, by Ian Goodfellow, Yoshua Bengio, Aaron Courville.
- Deep Learning: Foundations and Concepts, by Christopher M. Bishop and Hugh Bishop.
References for Programming Assignments
Relevant Articles (for Self Reading and Case Studies)
T1: Fundamentals of Learning
- E.D. Sontag. ``VC Dimension of Neural Networks," in Neural Networks and Machine Learning, pp. 69-95, 1998.
- M. Minsky and S. Papert, "Perceptrons: An Introduction to Computational Geometry," 2nd edition with corrections (first edition 1969), The MIT Press, Cambridge MA, ISBN 0-262-63022-2, 1972.
- K. Hornik, M. Stinchcombe, H. White, "Multilayer Feedforward Networks are Universal Approximators," Neural Networks, vol. 2, Pergamon Press, pp. 359–366, 1989.
- G. Cybenko, "Approximation by Superpositions of a Sigmoidal Function," Mathematics of Control, Signals, and Systems, vol. 2, no. 4, pp. 303–314, 1989.
- K. Hornik, "Approximation Capabilities of Multilayer Feedforward Networks". Neural Networks, vol. 4, no. 2, pp. 251–257, 1991.
- Y. LeCun, L. Bottou, Y. Bengio and P. Haffner, "Gradient-Based Learning Applied to Document Recognition," in Proceedings of the IEEE, vol. 86, no. 11, pp. 2278-2324, November 1998
T2: Convolutional Neural Networks
T3: Transformers
T4: Graph Neural Networks
T5: Deep Reinforcement Learning
T6: Generative Models
|