Course Overview
This course provides an introduction to machine learning with a special focus on engineering applications. The course starts with a mathematical background required for machine learning and covers approaches for supervised learning (linear models, kernel methods, decision trees, neural networks) and unsupervised learning (clustering, dimensionality reduction), as well as theoretical foundations of machine learning (learning theory, optimization). Evaluation will consist of mathematical problem sets and programming projects targeting real-world engineering applications.
Prerequisites
This course is intended for graduate students and qualified undergraduate students with a strong mathematical and programming background. Undergraduate level training or coursework in algorithms, linear algebra, calculus, probability, and statistics is suggested. A background in programming will also be necessary for the problem sets; students are expected to be familiar with python or learn it during the course. At CMU, this course is most similar to MLD's 10-601 or 10-701, though this course is meant specifically for students in engineering.
Textbooks
There will be no required textbooks, though we suggest the following to help you to study (all available online):- (KM): Machine Learning: A Probabilistic Perspective, Kevin Murphy. Online access is free through CMU’s library. Note that to access the library, you may need to be on CMU’s network or VPN.
- (ESL): Elements of Statistical Learning Trevor Hastie, Robert Tibshirani and Jerome Friedman.
- (TM): Machine Learning, Tom Mitchell.
- (CIML): A Course in Machine Learning, Hal Daumé III.
- (MJ): An Introdution to Probabilistic Graphical Models, Michael I. Jordan.
Piazza
We will use Piazza for class discussions. Please go to the course Piazza site to join the course forum (note: you must use a cmu.edu email account to join the forum). We strongly encourage students to post on this forum rather than emailing the course staff directly (this will be more efficient for both students and staff). Students should use Piazza to:
- Ask clarifying questions about the course material.
- Share useful resources with classmates (so long as they do not contain homework solutions).
- Look for students to form study groups.
- Answer questions posted by other students to solidify your own understanding of the material.
Grading Policy
Grades will be based on the following components:
- Problem Sets (50%): There will be 7 problem sets. Each problem set will have equal weight. We will drop the problem set with your lowest score.
- Late submissions will not be accepted.
- There is one exception to this rule: You are given 5 "late days" (self-granted 24-hr extensions) which you can use to give yourself extra time without penalty. At most two late days can be used per assignment. This will be monitored automatically via Gradescope.
- Solutions will be graded on both correctness and clarity. If you cannot solve a problem completely, you will get more partial credit by identifying the gaps in your argument than by attempting to cover them up.
- Midterm (20%), Final (30%): These exams will cover material from the lectures and the problem sets.
- Bonus: On Piazza, the top student "endorsed answer" answerers can earn bonus points.
Gradescope: We will use Gradescope to collect PDF submissions of each problem set. Upon uploading your PDF, Gradescope will ask you to identify which page(s) contains your solution for each problem - this is a great way to double check that you haven't left anything out. The course staff will manually grade your submission, and you'll receive feedback explaining your final marks.
Regrade Requests: If you believe an error was made during grading, you'll be able to submit a regrade request on Gradescope. For each homework, regrade requests will be open for only 1 week after the grades have been published. This is to encourage you to check the feedback you've received early!Academic Integrity Policy
Group studying and collaborating on problem sets are encouraged, as working together is a great way to understand new material. Students are free to discuss the homework problems with anyone under the following conditions:- Students must write their own solutions and understand the solutions that they wrote down.
- Students must list the names of their collaborators (i.e., anyone with whom the assignment was discussed).
- Students may not use old solution sets from other classes under any circumstances, unless the instructor grants special permission.
Using LaTeX
Students are strongly encouraged to use LaTeX for problem sets. LaTeX makes it simple to typeset mathematical equations and is extremely useful for graduate students to know. Most of the academic papers you read were written with LaTeX, and probably most of the textbooks too. Here is an excellent LaTeX tutorial and here are instructions for installing LaTeX on your machine.
Acknowledgments
This course is based in part on material developed by Fei Sha, Ameet Talwalkar, Matt Gormley, and Emily Fox. We also thank Anit Sahu and Joao Saude for their help with course development.
Schedule (Subject to Change)
Date | Topics | Reading | HW |
---|---|---|---|
8/26 | Introduction and Preliminaries [Slides] |
KM, Ch. 1 |
|
8/28 | Probability, MLE/MAP, and Linear Algebra [Slides] | TM, Estimating Probabilities KM, Ch. 2 (for a refresh in probability) Math4ML (review/refresher) Vectors, Matrices, and Least Squares Matrix Cookbook | |
8/30 | Recitation: Review of Preliminaries [Slides] | HW 1 Release | |
9/2 | Labor Day - No Class | ||
9/4 | Linear Regression, Part I [Slides] | KM, Ch. 7.1-7.3 Deep Learning Book, Ch. 5* | |
9/6 | Recitation [Slides] [Code] | ||
9/9 | Linear Regression, Part II [Slides] | KM, Ch. 7.4-7.6 Intro to regression | HW 1 Due HW 2 Release |
9/11 | Overfitting, Bias/variance Trade-off, Evaluation [Slides] [Annotated Slides] | Deep Learning, Ch. 5.2-5.4 KM, Ch. 6.4 |
|
9/13 | Recitation [Slides] | ||
9/16 | Naive Bayes [Slides] [Annotated Slides] | CIML, Ch. 9 KM, Ch. 3.5 |
|
9/18 | Logistic Regression [Slides] | KM, Ch. 8.1-8.4, 8.6 Discriminative vs. Generative |
|
9/20 | Recitation [Slides] [Solution] | ||
9/23 | Multi-class Classification [Slides] | KM, Ch. 8.5 | HW 2 Due HW 3 Release |
9/25 | SVM, Part I [Slides] | ESL, Ch. 12 KM Ch. 14.5 |
|
9/27 | Recitation [Slides] | ||
9/30 | SVM, Part II [Slides] [Duality Supplement] | Idiot's Guide to SVM Duality Supplement |
|
10/2 | Nearest Neighbors [Slides] [Annotated Slides] | CIML, Ch. 3.1-3.2 | |
10/4 | Recitation [Slides] | ||
10/7 | Decision Trees [Slides] | CIML, Ch. 1.3 KM, Ch. 16.2 ESL, Ch. 9.2 |
HW 3 Due HW 4 Release |
10/9 | Boosting, random forests [Slides] [Annotated Slides] | ESL, Ch. 10.1, 10.4-10.6 | |
10/11 | Recitation [Slides] [Solution] | ||
10/14-10/18 | Fall Break - No Classes | ||
10/21 | Neural Networks, Part I [Slides] | Learning Deep Architectures for AI ImageNet |
HW 4 Due HW 5 Release |
10/23 | Midterm | ||
10/25 | Recitation [Slides] | ||
10/28 | Neural Networks, Part II [Slides] | Neural Networks and Deep Learning, Ch.3 Regularization for Deep Learning |
|
10/30 | Neural Networks, Part III [Slides] | Neural Networks and Deep Learning, Ch.3 Regularization for Deep Learning |
|
11/1 | Recitation [Slides] | ||
11/4 | Clustering, Part I [Slides] | CIML, Ch. 15.1 | HW 5 due HW 6 Release |
11/6 | Clustering, Part II [Slides] [Annotated Slides] | ESL, Ch. 14.3.1-14.3.9 | |
11/8 | Recitation [Slides] | ||
11/11 | Dimensionality Reduction [Slides] [Annotated Slides] | PCA Independent Component Analysis |
|
11/13 | EM [Slides] [Annotated Slides] | ||
11/15 | Recitation | ||
11/18 | Online Learning (Bandits) [Slides] [Annotated Slides] | HW 6 due HW 7 Release |
|
11/20 | Reinforcement Learning, Part I [Slides] [Annotated Slides] | ||
11/22 | Recitation | ||
11/25 | Reinforcement Learning, Part II | ||
11/27-11/29 | Thanksgiving - No Classes | ||
12/2 | Guest Lecture | ||
12/4 | Final Review | ||
12/6 | Recitation | ||
12/9-12/13 | Final Exams Week |