Course Content
Basics of pattern recognition, Neural networks. Introduction to deep learning, convolutional networks, Applications in audio and image processing. All coding done in Python.Contact Details
Instructor | Sriram Ganapathy |
Office | Electrical Engineering C 334 (2nd Floor) |
deeplearning doT cce2018 aT gmail doT com |
Textbooks
- “Pattern Recognition and Machine Learning”, C.M. Bishop, 2nd Edition, Springer, 2007.
- “Neural Networks”, C.M. Bishop, 2nd Edition, Springer, 1995.
- “Deep Learning”, I. Goodfellow, Y, Bengio, A. Courville, MIT Press, 2016. html
Pre-requisites
- Random Process/Probablity and Statistics - Recommended Book "Probability and Statistics" by Stark and Woods. Video Lectures
- Linear Algebra/Matrix Theory - Recommended book "Linear Algebra", by G. Strang - Author's webpage for Videos and Text
- Basic Programming with Python - numpy based array and matrix operations.
- Calculus
Slides
11-01-2018        |
Introduction to Deep Learning Course. Examples. Roadmap of the course. | slides | |
18-01-2018 | Basics of Machine Learning - Decision and Inference Problems, Joint probability and posterior probabilities. Likelihood and priors. Loss matrix. Rule of maximum posterior probability. Loss function for regression. |
slides | |
01-02-2018 | Matrix Derivatives. Maximum Likelihood estimation and Gaussian Example. Linear Models for Classification. Perpendicular distance of a point from a surface. Logistic regression. Ref - PRML, Bishop, Chap. 3. Appendix C |
slides | |
08-02-2018 |
Logistic regression two class motivation. Posterior probability, sigmoid function, properties. Maximum likelihood for two class logistic regression. Cross entropy error for two class.
Ref - PRML, Bishop, Chap. 4, Sec. 4.2 |
slides | |
15-02-2018 | Logistic regression for K classes, softmax function. Non-convex optimization (local and global minima), Gradient Descent - motivation and algorithm. Ref - PRML, Bishop, Sec. 4.2 and NN, Bishop, Sec. 7.5 |
Code for Logistic Regression | |
22-02-2018 | First Assignment posted - Due on March 04, 2018. |
HW1 | |
22-02-2018 |
Training and Validation data sets. Logistic Regression Code Discussion. Perceptron and 1 Hidden Layer Neural Networks. Non-linear separability with hidden layer network.
Ref - NN, Bishop. Sec. 3.5 |
slides | |
01-03-2018 |
Multi-layer perceptrons, type of hidden layer and output layer activations - sigmoid, tanh, relu, softmax functions. Error functions in MLPs. Backpropagation learning in MLP. MLP for logistic regression in Keras.
Ref - NN, Bishop. Sec. 4.8 |
slides Code for MLP |
08-03-2018 | Midterm Exam in Class. |
15-03-2018 |
Backpropagation in multi-layer deep neural networks. Universal approximation properties of single hidden layer networks. Need for depth. The trade-off between depth and width of networks. Representation learning in DNNs. Hierachical data abstractions. Example in Images
Ref - "Deep Learning", I. Goodfellow. Chap. 6 https://arxiv.org/abs/1311.2901 |
slides | |
22-03-2018 |
Convolutional neural networks. Kernels and convolutional operations. Maxpooling and subsampling. Backpropagation in CNN layers. CNN example for MNIST.
Ref - "Deep Learning", I. Goodfellow. Chap. 9 |
slides Code for CNN |
|
02-04-2018 |
Second assignment due on April 12.
|
HW2.pdf | |
05-04-2018 |
Recurrent neural networks, back propagation in recurrent neural networks. Different recurrent architectures - teacher forcing networks, encoder/decoder networks, bidirectional networks.
|
slides |
|
12-04-2018 |
Vanishing gradient problem in RNNs. Long short term memory networks. Unsupervised representation learning - Restricted Boltzmann machines, Autoencoders. Discussion of mid-term exam
|
slides |
26-04-2018 | Final Exam in Class at 6pm. |