Course Content
Basics of pattern recognition, Neural networks. Introduction to deep learning, convolutional networks, Applications in audio, image and text. All coding done in Python.Contact Details
| Instructor | Sriram Ganapathy | 
| Office | Electrical Engineering C 334 (2nd Floor) | 
| deeplearning doT cce2019 aT gmail doT com | |
| Teaching Assistants | Shreyas R, Purvi A (Electrical Engineering C 328 (2nd Floor)) | 
Textbooks
- “Pattern Recognition and Machine Learning”, C.M. Bishop, 2nd Edition, Springer, 2007.
- “Neural Networks”, C.M. Bishop, 2nd Edition, Springer, 1995.
- “Deep Learning”, I. Goodfellow, Y, Bengio, A. Courville, MIT Press, 2016. html
Pre-requisites
- Random Process/Probablity and Statistics - Recommended Book "Probability and Statistics" by Stark and Woods. Video Lectures
- Linear Algebra/Matrix Theory - Recommended book "Linear Algebra", by G. Strang - Author's webpage for Videos and Text
- Basic Programming with Python - numpy based array and matrix operations.
- Calculus
Announcements
- Midterm will be on March 7, 2019 in class hours.
- No class on the election day (April 18, 2019).
- Final exam will be on May 2, 2019, 6-7:45 pm. Venue: Lecture Hall 10, Lecture Hall Complex, Center for Continuing Education (CCE), near Prakruti Cafe, Indian Institute of Science, Bangalore-560012.
Slides
| 10-01-2019     | Introduction to Deep Learning Course. Examples. Roadmap of the course. | slides | |
| 17-01-2019    | Basics of Machine Learning - Decision and Inference Problems, Joint probability and posterior probabilities. Likelihood and priors. Loss matrix. Rule of maximum posterior probability. Loss function for regression. | slides | |
| 31-01-2019    | Matrix Derivatives. Maximum Likelihood estimation and Gaussian Example. Linear Models for Classification. Perpendicular distance of a point from a surface. Logistic regression. Ref - PRML, Bishop, Chap. 3. Appendix C | slides | |
| 07-02-2019    | Logistic regression two class motivation. Posterior probability, sigmoid function, properties. Maximum likelihood for two class logistic regression. Cross entropy error for $K$ classes. Logistic regression for K classes, softmax function. Non-convex optimization (local and global minima), Gradient Descent - motivation and algorithm. Ref - PRML, Bishop, Chap. 4, Sec. 4.2 | slides | |
| 14-02-2019    | Summary of previous lectures. Gradient descent for $K$ class logistic regression. Implementing logistic regression for MNIST dataset. Perceptron Model and motivation. Introduction to single hidden layer neural networks. Training and Validation data sets. Logistic Regression Code Discussion. Ref - PRML, Bishop, Sec. 4.2 and NN, Bishop, Sec. 3.5, 7.5 | Code for Logistic Regression | slides | 
| 15-02-2019 | First Assignment posted - Due on Feb 24, 2019. | Assignment-1 (with submission instructions) | |
| 21-02-2019 | Perceptron and 1 Hidden Layer Neural Networks. Non-linear separability with hidden layer network. Type of hidden layer and output layer activations - sigmoid, tanh, relu, softmax functions. Error functions in MLPs. Ref - NN, Bishop. Sec. 3.5, Sec. 4.8 | slides | |
| 28-02-2019 | Backpropagation in multi-layer deep neural networks. Universal approximation properties of single hidden layer networks. Need for depth. The trade-off between depth and width of networks. Representation learning in DNNs. Hierachical data abstractions. Example in Images Ref - "Deep Learning", I. Goodfellow. Chap. 6 https://arxiv.org/abs/1311.2901 | slides Code for Single Layer NN | |
| 07-03-2019 | Midterm Exam posted - Due by 19:45 hrs (90 minutes). | DL2019_Midterm1 | 14-03-2018 | Convolutional neural networks. Kernels and convolutional operations. Maxpooling and subsampling. Backpropagation in CNN layers. CNN example for MNIST. Ref - "Deep Learning", I. Goodfellow. Chap. 9 | slides | 
| 18-03-2019 | Second assignment due on March 30 (Can attempt CNN questions after next class). | HW2.pdf | |
| 21-03-2019 | Solutions to Midterm 1 Posted. | DL2019_Midterm1_Solutions | 21-03-2018 | Back Propagation in CNNs. Choice of Kernels and convolutional operations. CNN example for MNIST. Midterm Exam Discussion. Code for CNNRef - "Deep Learning", I. Goodfellow. Chap. 9 | 
| 28-03-2019 | Recurrent neural networks, back propagation in recurrent neural networks. Different recurrent architectures - teacher forcing networks, encoder/decoder networks, bidirectional networks. Ref - "Deep Learning", I. Goodfellow. Chap. 10 | slides | |
| 04-04-2019 | Issue of Vanishing and Exploding Gradients, Long Short Term Memory Networks (LSTM), Attention Mechanism in Neural Networks Ref - "Deep Learning", I. Goodfellow. Chap. 10 | slides | |
| 11-04-2019 | Advancted Topics, Network in Network, Convolutional LSTM networks, Unsupervised learning, autoencoders, adversarial learning Ref - "Deep Learning", I. Goodfellow. Chap. 16 | slides | |
| 25-04-2019 | Applications of Deep learning for natural language processing, image processing and speech processing. | slides | |
| 26-04-2019 | Third assignment posted, due on May 8. | HW3.pdf |