This is the primary course website for Mathematical Foundations of Deep Neural Networks (심층신경망의 수학적 기초), M1407.001200, Fall 2022. We will also use eTL as a secondary website.

- This year (2022), lectures will in a hybrid format. You may attend in person or virtually.
- We will have make-up lectures on 09/02 11:00-12:15 and 09/07 11:00-12:15. (Note, the hour is slightly different from the initial announcement.) You can attend through Zoom or in person, or you may watch the recorded video. (Live attendance is not required.) The lecture will take place in the 43-1-306 recording studio. Because the room has a capacity of 20 students, those who wish to attend in person must sign up through the link posted on eTL.
- No class on 09/29, 10/18, and 10/20.

- Python Tutorial 1, Jupyter notebook code.
- Python Tutorial 2, Jupyter notebook code.
- Python Tutorial 3, Jupyter notebook code.
- Chapter 0: Prologue
- Chapter 1: Optimization and Stochastic Gradient Descent, Jupyter notebook code.
- Chapter 2: Shallow Neural Networks, Jupyter notebook code.
- Chapter 3: Convolutional Neural Networks, Jupyter notebook code.

This class will have weekly homework assignments. Submit completed assignments through eTL.

- Homework 1, Due 09/07, 5pm, wideMinima.py, conv1D.py.
- Homework 2, Due 09/14, 5pm, twolayerSGD.py.
- Homework 3, Due 09/21, 5pm, threelayerSGD.py.
- Homework 4, Due 09/28, 5pm, lenet_original.py.
- Homework 5, Due 10/05, 5pm, mlp_backprop.py, label_memorization.py.
- Homework 6, Due 10/12, 5pm, ddescent.py.

- [Week 1] Optimization and stochastic gradient descent
- [Week 2] Shallow neural networks and logistic regression.
- [Week 3] Multi-layer perceptron. Softmax regression.
- [Week 4] Convolutional layers, pooling layers, GPU computing, LeNet
- [Week 5] Data augmentation, regularization techniques: dropout, weight decay, early stopping
- [Week 6] Weight initialization, VGGNet, backprop
- [Week 7] Optimizers (ADAM, RMSProp), NiN network, GoogLeNet
- [Week 8] Batch normalization, ResNet, DenseNet
- [Week 9] ResNext, SENet, DNCNN, super-resolution, inverse problem
- [Week 10-11] Flow models
- [Week 12-13] Variational auto-encoders
- [Week 14-15] Generative adversarial networks

Course material will be posted on this website. eTL will be used for announcements, homework submission, and receiving homework and exam scores.

Ernest K. Ryu, 27-205,

Joo Young Choi

Haneol Kijm

Sehyun Kwon

Jongmin Lee

Dongyeob Kim

Jae yeon Kim

Wonseok Lee

Yeeseok Oh

Kunwoo Na

Jaeseung Park

Hybrid lecture, in-person at 43-1-201 and Zoom, Tuesdays and Thursdays 5:00–6:15pm. We thank the Faculty of Liberal Education (기초교육원) for helping us execute the hybrid lecture. Live (in-person or online) attendance is required.

The midterm and final exams will be hand-written (no computers), in-person, and 4 hours long.

- Midterm exam: Saturday, 10/15, 8:00am–12:00pm.
- Final exam: Saturday, 12/17, 1:00–5:00pm.
- Sample Exam: 2021 Midterm.
- Sample Exam: 2021 Final.

Homework 30%, midterm exam 30%, final exam 40%.

Good knowledge of the following subjects is required.

- Vector calculus: profeciency with gradients and chain rules.
- Linear algebra: profeciency with matrix-vector products.
- Basic probability: random variables, expectation, mean, variance, discrete and continuous random variables, Gaussian random variables.

No prior backround in machine learning or deep learning is required.

Although programming is an essential part of deep learning, I do not want to make substantial Python programming background a prerequisite for this course. Therefore, I will record and upload a Python tutorial for students to study before the start of the semester. I expect this tutorial to be about 5 hours long, and it will cover the necessary Python programming concepts such as variables, loops, functions, classes, objects, inheritance, exceptions, Python lists, Python tuples, Python decorators, Python iterators, the NumPy library, and the Matplotlib library. Of course, students already familiar with these concepts need not spend time with the tutorial.

- Python Tutorial 1, Jupyter notebook code.
- Python Tutorial 2, Jupyter notebook code.
- Python Tutorial 3, Jupyter notebook code.

Programming in deep learning tends to be relatively simple if one utilizes the frameworks such as PyTorch. These frameworks do utilize advanced programming concepts such as inheritance, decorators, and iterators, but if you are roughly familiar with these concepts, I do not expect the Python programming to be a major challenge in this course.

The course does not have a designated reference or a textbook. However, you may find the following references useful: Fei Fei Li's CS 231n at Stanford, Pieter Abbeel's Deep Unsupervised Learning at Berkeley, *Dive into Deep Learning* by Zhang, Lipton, Li, and Smola, and *Deep Learning* by Goodfellow, Bengio, and Courville.

Throughout this class, you will spend many hours computing gradients. While the chain rule is taught in any vector calculus curriculum, chain rule formulae in vector and matrix forms are not. Fundamentally, you should first carry out the gradient calculations elementwise and then collect the result into a vectorized form. Nevertheless, you may find the following references useful: Matrix calculus notes 1, matrix calculus notes 2, and matrix calculus notes 3 (Chapters 1–3).

- Python Tutorial 1, Jupyter notebook code.
- Python Tutorial 2, Jupyter notebook code.
- Python Tutorial 3, Jupyter notebook code.
- Chapter 1: Optimization and Stochastic Gradient Descent, Jupyter notebook code.
- Chapter 2: Shallow Neural Networks, Jupyter notebook code.
- Chapter 3: Convolutional Neural Networks, Jupyter notebook code.
- Chapter 4: CNNs for Non-Classification Supervised Learning, Jupyter notebook code.
- Chapter 5: Unsupervised Learning, Jupyter notebook code.
- Chapter A: Basics of Monte Carlo, Jupyter notebook code.
- Chapter 6: Additional Things.

- Homework 1, wideMinima.py, conv1D.py.
- Homework 2, twolayerSGD.py.
- Homework 3, threelayerSGD.py.
- Homework 4, lenet_original.py.
- Homework 5, mlp_backprop.py, label_memorization.py.
- Homework 6, bn_remove.py, smallNetSaved.
- Homework 7, anomaly_detection.py.
- Homework 8, normalizingFlow1d.py.
- Homework 9, flow_inpainting.py, nice.pt, minimax_logistic.py.