January 14, 2025
CS 412: Intro to Machine Learning (Spring 2025)
📅 Course Schedule
Date | Topic | Reading | Materials | Notes |
---|---|---|---|---|
Jan 21 | Introduction (Course Logistics) Basics of Supervised Learning | PML 1.2 ESL 2.1, 2.2 | 📊 CS412-Intro.pdf | HW1.pdf (Self-Assessment) [Submit on Gradescope] Due: Jan 28 |
Jan 23 | Understanding loss functions | PML 4.3 | Scribe1: Edomwonyi, Uwadia | |
Jan 28 | Hypothesis (Func) Class ERM Linear Regression | PML 11.1, 11.2,11.3 ESL 3.2 | Scribe2: Shelke, Harsh | HW1 due today! |
Jan 30 | Overfitting & Regularizers Lin-Reg: Linear Regression (contd) | PML 4.3, 5.4 | Scribe3: Hulu, Charis/ Lokesh | |
Feb 4 | Logistic Regression Multiclass Logistic Regression (MLR) | PML 10.2.1, 10.2.2 ESL 4.4.1 | Scribe4: Dahagam, Sujay | HW2 out [link] (UIC credentials rqd.) Finalize scribe assignment |
Feb 6 | Multiclass logistic (contd) MLE (Bernoulli) MAP (Bernoulli) | PML 4.2, 4.5 Slides CMU 10701 Tom Mitchell's note | Scribe5: Bhat, Amith | Bonus-Quiz [qz-feb6.pdf] (submit to gradescope) Feedback form Announcements |
Feb 8 [Extra Class-1] | MLE for Regression MAP = Regularization | PML 4.2.5 - 4.2.7, 4.5 CMU 10-315 notes Nice blog | Scribe6: Sappidi, Yugesh | Optional Reading: Conjugate Priors • Nice blog • Duke STA114 |
Feb 11 | Regularized logistic regression thru' MAP Convex functions | PML 10.1-10.3 CvxO 1.1-1.3 Notes-IFT 6085 | Scribe7: Ferdowsi, Farhad | HW2 due! |
Feb 13 | Properties of Cvx funcs Gradient descent (GD) Convergence rates | CvxO 1.1-1.3, 3.1 Notes by A. Ahmadi Notes by C Wang | Scribe8: Ferdowsi, Farhad | Extra Reading 1. Matrix Norms |
Feb 18 | Convergence of GD Newton's method | CvxO 3.1, 3.2, 3.4, 5.3.2 Nice notes on Cvx Optimization Notes-CS 6820 | Scribe9: Datta Sai VVN | Extra Reading: 1. Matrix Derivative 2. Taylor Series 3. Newton's Method |
Feb 20 | GD convergence analysis SGD + Convergence guarantees Batched SGD Variants of GD | CvxO 3.1-3.2 Notes on SGD Notes on Heavy Ball & Nesterov's Accelerated GD | Scribe10: Aniket Wagde | HW3 out [link] Announcements Extra Reading: 1. Comprehensive study of SGD & Batched-SGD 2. Nice demo |
Feb 25 | Max-Margin Formulation SVM objective KKT conditions | PML 17.3.1-17.3.2 Caltech CS156 Slides-1 KKT conditions (CMU-10-725) Notes by Wang & Pavlu | Scribe11: Harsh Kothari | Extra Reading: 1. Slides CMU 10-601 |
Feb 27 | KKT conditions (contd) Dual Optimization SVM with strong duality | Strong Duality and Slater's Conditions Caltech CS156 Slides-2 | Scribe12: Boggavarapu, Lokesh | Extra Reading: 1. Primer on QP 2. Strong Duality |
Mar 1 [Extra Class-2] | Kernel SVM Kernel Properties Examples | PML 17.3.4 CMU-10701 slides | Scribe13: Madishetty, Aravind | Extra Reading: 1. SVM-vs-Perceptron 2. Another amazing demo w/ Kernels! |
Mar 4 | Soft-Margin SVM SVM Regression Perceptron | PML 17.3.3 COMP-652 slides CMU-15-859 | Scribe14: Lukau, Please | HW3 due! |
Mar 6 | Perceptron Mistake bounds Perceptron w/o perfect linear separator Kernel Perceptron | PML 10.2.5, 13.2 Princeton CoS495 CMU 10-607 | Scribe15: Rithish Reddy Chichili | HW3 solutions: [link] |
Mar 13 | Mid Term | |||
Mar 15 [Extra Class-3] | Winnow's Algorithm Mistake Bounds | Winnow: CS 4540 Mistake Bound: CS260 Winnow Example | Scribe16: Nemi Chakrawarthy Bhupathiraju / Simran Mishra | Extra Reading: 1. Proving lower bounds CMSC 35900 (Sec 2) 2. Advanced topics on Winnow & Perceptron: PLG Chap 12 |
Mar 18 | Boosting Adaboost Mistake Bounds | Boosting: CSCI699 Slides, Rob Schapire | Scribe17: Gabriel Zhang | Extra Reading: 1. Original paper by Freund & Schapire'99 2. Advanced topics: Tutorial (Sec 9) 3. General winnow: Alg1, Alg2 |
Mar 20 | Midterm solutions PCA | HW4 out: [link] Project list out: [link] | ||
Mar 25 | Spring Break! | |||
Mar 27 | Spring Break! | |||
April 1 | Orthonormal Basis (OB) PCA (Min-Error Formulation) | OB: Gilbert Strang [Chap 3.4] PCA: PRML 12.1-12.2 | Scribe19: Jiachen Tao / Aksun | Extra Reading: • Basics of LA • Basics of Vector Space |
April 3 | PCA (Max-Var Formulation) Eigen Values-Vectors (EV) Clustering | MA262 EV basics PCA: PML 20.1 Clustering: J. Cadler Slides E Liberty lecture | Scribe20: Mishra, Simran | Extra Reading: • Nice video tutorial on EV! • K-Medoids: CSC 380 slides • KPCA: CSCI 5512 slides |
April 10 | K-Means Spectral Clustering (SC) | K-Means: CS217 PML 21 | Scribe21: Wang, Haoxuan | Extra Reading: • More on Graph Laplacian |
April 12 | Spectral clustering (SC) Neural Net Intro | SC: ML10.701, CSE902 | Scribe22: Wang, Haoxuan | Extra Reading: • Advanced SC |
April 15 | Feed-Forward NN (FFNN) Back-Propagation (BP) | CS217: FFNN, BP COS 324: FFNN & BP | Scribe23: Pipim, Charles / Rithish Reddy Chichili | HW4 due! HW5 out: [link] |
April 17 | Backpropagation (contd) RNN CNN (Intro) | CS217-RNN RNN Slides EECS-498: RNN Slides | Scribe24: Salvi, Yukta | Extra Reading: • Detailed Applied YT course on DL |
April 22 | CNN Dropout regularization Vanishing & Exploding Gradients (VE-Grads) RNN-LSTM | CS217-CNN COS 324: CNN EECS-498: CNN Slides VE-Grads in RNN LSTM slides | Scribe25: Agnihotri, Akshun | Extra Reading: • Regularization in NNs |
April 24 | Gaussian Processes (GP) | GP: CS229 notes ESE-680 notes | Scribe26: Bhat, Amith | OMD.pdf OCO.pdf |
April 26 [Optional Extra class-1] | Transformers for LMs | CMU 11-785 slides Attention Paper (Google) | Lecture topic not included in the final exam! CS4780 Slides | |
April 29 | Online Learning Basics-1 - Halving (HA) - Wtd-Majority (WMA) | HA.pdf WMA.pdf | Scribe28: Kothari, Harsh | |
May 1 | Online Learning Basics-2 - EXP-Wt - OMD (OCO) | Exp-Wt.pdf | Scribe29: Datta, Sai VVN | [Topic not included in the final exam] Extra Reading: OMD.pdf OCO.pdf |
May 3 | HW5 due! | |||
May 6 | Final Exam | Time: 1-2:30 pm CT Room: LC F6 | ||
May 9 | Project Presentations | Final project report due! Time: 3-5:30 pm CT |
📚 Course Description
This is a comprehensive, math-intensive machine learning course that bridges theoretical foundations with practical applications. The course covers core supervised learning algorithms including linear/logistic regression, support vector machines, neural networks, and ensemble methods like boosting. Students will dive deep into optimization techniques (gradient descent, Newton's method), convex analysis, kernel methods, dimensionality reduction (PCA), clustering, and modern topics like deep learning architectures (CNNs, RNNs, LSTMs) and online learning algorithms.
The course emphasizes rigorous mathematical understanding alongside hands-on implementation. Students will engage through programming assignments in Python/MATLAB, scribe detailed lecture notes using LaTeX, and work on a significant course project. Topics progress from fundamental concepts like loss functions and regularization to advanced methods including spectral clustering, Gaussian processes, and transformer architectures. The curriculum balances theoretical analysis (convergence proofs, mistake bounds, duality theory) with practical machine learning skills.
Upon completion, students will have a solid theoretical foundation in machine learning mathematics, practical experience implementing algorithms from scratch, and the ability to analyze and apply ML methods to real-world problems. Success requires a strong background in probability, statistics, and linear algebra, along with programming proficiency. This course prepares students for advanced ML research, industry roles requiring deep technical understanding, or pursuing research-level graduate studies in ML and AI.
⚠️ Prerequisites
Expect this to be a fairly math intensive course. Please familiarize yourself with the basics of Probability-Statistics (PS) and Linear-Algebra (LA). Recommended introductory lectures to check if you are comfortable with the basics:
- PS review:
- LA review:
Familiarity with LaTeX for scientific writing for scribing the lecture notes (only for MS students). You can learn the basics from A Simple Quickstart Guide. Many other online tutorials are available for beginners — feel free to explore and use whichever best suits your needs.
Programming assignments (in Python or Matlab). Be prepared to code: [Python ML Tutorials], [Google Colab] (many online tutorials available for beginners).
A strong grasp of the foundational material outlined above is expected of all students taking the course for credit. Insufficient preparation may adversely affect your ability to engage with the course content and perform successfully in assessments, which may impact your final grades.
📖 Resources
Class lecture will be based on, but not limited to, the following books:
- [ESL] The Elements of Statistical Learning by Hastie, Tibshirani, and Friedman [Book website]
- [MLTM] Machine Learning by Tom Mitchell [Online copy]
- [PML] Probabilistic Machine Learning: An Introduction, by Kevin Murphy [Book website]
- [PLG] Prediction, Learning and Games by Nicolo Cesa-Bianchi and Gabor Lugosi, Cambridge University Press, 2006 [Local Copy from E1 245]
- [CvxO] Convex Optimization: Algorithms and Complexity by Sebastien Bubeck [Book Website]
- [PyAG] (Optional): Hands-On Machine Learning with Scikit-Learn & Tensorflow by Aurelien Geron [Online, Github]
- [PRML] (Optional): Pattern Recognition and Machine Learning by Christopher Bishop [Free copy]
- [CIML] (Optional): A Course in Machine Learning by Hal Daume III [Online copy], [Errata]
- [UML] (Optional): Understanding Machine Learning: From Theory to Algorithms by Shai Ben-David, Shai Shalev-Shwartz [Online copy]
- [OCO] (Optional) Introduction to Online Convex Optimization by Elad Hazan [Book website]
🎯 Course Logistics
- 📍 Location: Lecture Complex F6
- ⏰ Schedule: Tuesday & Thursday, 2:00 - 3:15 PM
- 🏛️ Office Hours: Thurdays 5:00–6:00 PM or by appointment
- 📧 Contact: Email or Piazza