January 14, 2025

CS 412: Intro to Machine Learning (Spring 2025)

📡 •Instructor: Aadirupa Saha
🔬 •Teaching Assistant: Akhil S Nair

📅 Course Schedule

DateTopicReadingMaterialsNotes
Jan 21Introduction (Course Logistics)
Basics of Supervised Learning
PML 1.2
ESL 2.1, 2.2
📊 CS412-Intro.pdfHW1.pdf (Self-Assessment)
[Submit on Gradescope]
Due: Jan 28
Jan 23Understanding loss functionsPML 4.3Scribe1: Edomwonyi, Uwadia
Jan 28Hypothesis (Func) Class
ERM
Linear Regression
PML 11.1, 11.2,11.3
ESL 3.2
Scribe2: Shelke, HarshHW1 due today!
Jan 30Overfitting & Regularizers
Lin-Reg: Linear Regression (contd)
PML 4.3, 5.4Scribe3: Hulu, Charis/ Lokesh
Feb 4Logistic Regression
Multiclass Logistic Regression (MLR)
PML 10.2.1, 10.2.2
ESL 4.4.1
Scribe4: Dahagam, SujayHW2 out [link]
(UIC credentials rqd.)
Finalize scribe assignment
Feb 6Multiclass logistic (contd)
MLE (Bernoulli)
MAP (Bernoulli)
PML 4.2, 4.5
Slides CMU 10701
Tom Mitchell's note
Scribe5: Bhat, AmithBonus-Quiz [qz-feb6.pdf]
(submit to gradescope)
Feedback form
Announcements
Feb 8
[Extra Class-1]
MLE for Regression
MAP = Regularization
PML 4.2.5 - 4.2.7, 4.5
CMU 10-315 notes
Nice blog
Scribe6: Sappidi, YugeshOptional Reading:
Conjugate Priors
Nice blog
Duke STA114
Feb 11Regularized logistic regression thru' MAP
Convex functions
PML 10.1-10.3
CvxO 1.1-1.3
Notes-IFT 6085
Scribe7: Ferdowsi, FarhadHW2 due!
Feb 13Properties of Cvx funcs
Gradient descent (GD)
Convergence rates
CvxO 1.1-1.3, 3.1
Notes by A. Ahmadi
Notes by C Wang
Scribe8: Ferdowsi, FarhadExtra Reading
1. Matrix Norms
Feb 18Convergence of GD
Newton's method
CvxO 3.1, 3.2, 3.4, 5.3.2
Nice notes on Cvx Optimization
Notes-CS 6820
Scribe9: Datta Sai VVNExtra Reading:
1. Matrix Derivative
2. Taylor Series
3. Newton's Method
Feb 20GD convergence analysis
SGD + Convergence guarantees
Batched SGD
Variants of GD
CvxO 3.1-3.2
Notes on SGD
Notes on Heavy Ball & Nesterov's Accelerated GD
Scribe10: Aniket WagdeHW3 out [link]
Announcements
Extra Reading:
1. Comprehensive study of SGD & Batched-SGD
2. Nice demo
Feb 25Max-Margin Formulation
SVM objective
KKT conditions
PML 17.3.1-17.3.2
Caltech CS156 Slides-1
KKT conditions (CMU-10-725)
Notes by Wang & Pavlu
Scribe11: Harsh KothariExtra Reading:
1. Slides CMU 10-601
Feb 27KKT conditions (contd)
Dual Optimization
SVM with strong duality
Strong Duality and Slater's Conditions
Caltech CS156 Slides-2
Scribe12: Boggavarapu, LokeshExtra Reading:
1. Primer on QP
2. Strong Duality
Mar 1
[Extra Class-2]
Kernel SVM
Kernel Properties
Examples
PML 17.3.4
CMU-10701 slides
Scribe13: Madishetty, AravindExtra Reading:
1. SVM-vs-Perceptron
2. Another amazing demo w/ Kernels!
Mar 4Soft-Margin SVM
SVM Regression
Perceptron
PML 17.3.3
COMP-652 slides
CMU-15-859
Scribe14: Lukau, PleaseHW3 due!
Mar 6Perceptron Mistake bounds
Perceptron w/o perfect linear separator
Kernel Perceptron
PML 10.2.5, 13.2
Princeton CoS495
CMU 10-607
Scribe15: Rithish Reddy ChichiliHW3 solutions: [link]
Mar 13Mid Term
Mar 15
[Extra Class-3]
Winnow's Algorithm
Mistake Bounds
Winnow: CS 4540
Mistake Bound: CS260
Winnow Example
Scribe16: Nemi Chakrawarthy Bhupathiraju / Simran MishraExtra Reading:
1. Proving lower bounds CMSC 35900 (Sec 2)
2. Advanced topics on Winnow & Perceptron: PLG Chap 12
Mar 18Boosting
Adaboost
Mistake Bounds
Boosting: CSCI699
Slides, Rob Schapire
Scribe17: Gabriel ZhangExtra Reading:
1. Original paper by Freund & Schapire'99
2. Advanced topics: Tutorial (Sec 9)
3. General winnow: Alg1, Alg2
Mar 20Midterm solutions
PCA
HW4 out: [link]
Project list out: [link]
Mar 25Spring Break!
Mar 27Spring Break!
April 1Orthonormal Basis (OB)
PCA (Min-Error Formulation)
OB: Gilbert Strang [Chap 3.4]
PCA: PRML 12.1-12.2
Scribe19: Jiachen Tao / AksunExtra Reading:
Basics of LA
Basics of Vector Space
April 3PCA (Max-Var Formulation)
Eigen Values-Vectors (EV)
Clustering
MA262 EV basics
PCA: PML 20.1
Clustering: J. Cadler Slides
E Liberty lecture
Scribe20: Mishra, SimranExtra Reading:
Nice video tutorial on EV!
• K-Medoids: CSC 380 slides
• KPCA: CSCI 5512 slides
April 10K-Means
Spectral Clustering (SC)
K-Means: CS217
PML 21
Scribe21: Wang, HaoxuanExtra Reading:
More on Graph Laplacian
April 12Spectral clustering (SC)
Neural Net Intro
SC: ML10.701, CSE902Scribe22: Wang, HaoxuanExtra Reading:
Advanced SC
April 15Feed-Forward NN (FFNN)
Back-Propagation (BP)
CS217: FFNN, BP
COS 324: FFNN & BP
Scribe23: Pipim, Charles / Rithish Reddy ChichiliHW4 due!
HW5 out: [link]
April 17Backpropagation (contd)
RNN
CNN (Intro)
CS217-RNN
RNN Slides
EECS-498: RNN Slides
Scribe24: Salvi, YuktaExtra Reading:
Detailed Applied YT course on DL
April 22CNN
Dropout regularization
Vanishing & Exploding Gradients (VE-Grads)
RNN-LSTM
CS217-CNN
COS 324: CNN
EECS-498: CNN Slides
VE-Grads in RNN
LSTM slides
Scribe25: Agnihotri, AkshunExtra Reading:
Regularization in NNs
April 24Gaussian Processes (GP)GP: CS229 notes
ESE-680 notes
Scribe26: Bhat, AmithOMD.pdf
OCO.pdf
April 26
[Optional Extra class-1]
Transformers for LMsCMU 11-785 slides
Attention Paper (Google)
Lecture topic not included in the final exam!
CS4780 Slides
April 29Online Learning Basics-1
- Halving (HA)
- Wtd-Majority (WMA)
HA.pdf
WMA.pdf
Scribe28: Kothari, Harsh
May 1Online Learning Basics-2
- EXP-Wt
- OMD (OCO)
Exp-Wt.pdfScribe29: Datta, Sai VVN[Topic not included in the final exam]

Extra Reading:
OMD.pdf
OCO.pdf
May 3HW5 due!
May 6Final ExamTime: 1-2:30 pm CT
Room: LC F6
May 9Project PresentationsFinal project report due!
Time: 3-5:30 pm CT

📚 Course Description

This is a comprehensive, math-intensive machine learning course that bridges theoretical foundations with practical applications. The course covers core supervised learning algorithms including linear/logistic regression, support vector machines, neural networks, and ensemble methods like boosting. Students will dive deep into optimization techniques (gradient descent, Newton's method), convex analysis, kernel methods, dimensionality reduction (PCA), clustering, and modern topics like deep learning architectures (CNNs, RNNs, LSTMs) and online learning algorithms.

The course emphasizes rigorous mathematical understanding alongside hands-on implementation. Students will engage through programming assignments in Python/MATLAB, scribe detailed lecture notes using LaTeX, and work on a significant course project. Topics progress from fundamental concepts like loss functions and regularization to advanced methods including spectral clustering, Gaussian processes, and transformer architectures. The curriculum balances theoretical analysis (convergence proofs, mistake bounds, duality theory) with practical machine learning skills.

Upon completion, students will have a solid theoretical foundation in machine learning mathematics, practical experience implementing algorithms from scratch, and the ability to analyze and apply ML methods to real-world problems. Success requires a strong background in probability, statistics, and linear algebra, along with programming proficiency. This course prepares students for advanced ML research, industry roles requiring deep technical understanding, or pursuing research-level graduate studies in ML and AI.

⚠️ Prerequisites

Expect this to be a fairly math intensive course. Please familiarize yourself with the basics of Probability-Statistics (PS) and Linear-Algebra (LA). Recommended introductory lectures to check if you are comfortable with the basics:

Familiarity with LaTeX for scientific writing for scribing the lecture notes (only for MS students). You can learn the basics from A Simple Quickstart Guide. Many other online tutorials are available for beginners — feel free to explore and use whichever best suits your needs.

Programming assignments (in Python or Matlab). Be prepared to code: [Python ML Tutorials], [Google Colab] (many online tutorials available for beginners).

A strong grasp of the foundational material outlined above is expected of all students taking the course for credit. Insufficient preparation may adversely affect your ability to engage with the course content and perform successfully in assessments, which may impact your final grades.

📖 Resources

Class lecture will be based on, but not limited to, the following books:

  • [ESL] The Elements of Statistical Learning by Hastie, Tibshirani, and Friedman [Book website]
  • [MLTM] Machine Learning by Tom Mitchell [Online copy]
  • [PML] Probabilistic Machine Learning: An Introduction, by Kevin Murphy [Book website]
  • [PLG] Prediction, Learning and Games by Nicolo Cesa-Bianchi and Gabor Lugosi, Cambridge University Press, 2006 [Local Copy from E1 245]
  • [CvxO] Convex Optimization: Algorithms and Complexity by Sebastien Bubeck [Book Website]
  • [PyAG] (Optional): Hands-On Machine Learning with Scikit-Learn & Tensorflow by Aurelien Geron [Online, Github]
  • [PRML] (Optional): Pattern Recognition and Machine Learning by Christopher Bishop [Free copy]
  • [CIML] (Optional): A Course in Machine Learning by Hal Daume III [Online copy], [Errata]
  • [UML] (Optional): Understanding Machine Learning: From Theory to Algorithms by Shai Ben-David, Shai Shalev-Shwartz [Online copy]
  • [OCO] (Optional) Introduction to Online Convex Optimization by Elad Hazan [Book website]

🎯 Course Logistics

  • 📍 Location: Lecture Complex F6
  • ⏰ Schedule: Tuesday & Thursday, 2:00 - 3:15 PM
  • 🏛️ Office Hours: Thurdays 5:00–6:00 PM or by appointment
  • 📧 Contact: Email or Piazza