0% found this document useful (0 votes)
142 views6 pages

Machine Learning: COMS 4771 Fall 2018

This document provides information about the Machine Learning course offered in Fall 2018 including: - Time and location of lectures on Mondays and Wednesdays from 2:40-3:55pm in room 301 Pupin - Instructor and TAs contact information and office hours - Announcements about questions on the course and signing up for Piazza and Gradescope - A tentative schedule of lecture topics - Assessment breakdown of homeworks, exams, and textbook - Resources including recommended books, software, and background knowledge - Homework submission and collaboration policies - Academic conduct rules and violations penalties

Uploaded by

Jiahong He
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
142 views6 pages

Machine Learning: COMS 4771 Fall 2018

This document provides information about the Machine Learning course offered in Fall 2018 including: - Time and location of lectures on Mondays and Wednesdays from 2:40-3:55pm in room 301 Pupin - Instructor and TAs contact information and office hours - Announcements about questions on the course and signing up for Piazza and Gradescope - A tentative schedule of lecture topics - Assessment breakdown of homeworks, exams, and textbook - Resources including recommended books, software, and background knowledge - Homework submission and collaboration policies - Academic conduct rules and violations penalties

Uploaded by

Jiahong He
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 6

Machine Learning

COMS 4771 Fall 2018


Time & Mon and Wed 2:40p-3:55p at 301 Pupin
location
location:: Laboratories
Instructor
Instructor:: Nakul Verma ( verma [at] cs.columbia.edu )
Office hours
hours:: Mon 4-5p
TA Office Vincent Liu (zl2474) -- Mon 5-7p
hours
hours:: Nav Ravindranath (ngr2114) -- Tue 2-4p
Jacob Austin (ja3067) -- Wed 9-11a
Amit Bhat (ab4520) -- Wed 4-6p
Manksh Gupta (mg3835) -- Thu 10-12p
Yibo Jiang (yj2460) -- Fri 10-12p

Location: Mudd 1st floor TA room.


QA forum
forum:: Piazza
Homework & Gradescope
grading
grading::

Announcements
09/01: Have questions about this course? Read the FAQ here.
09/01: The website is up. Please signup at Piazza and Gradescope.

Tentative Schedule

Lecture Introduction, Maximum Likelihood Estimation, Classification via


1 Probabilistic Modeling, Bayes Classifier, Naive Bayes, Evaluating
Classifiers
[lec 1]
notes/reading: MLE basics | Multivariate Gaussian | Naive Bayes |
Optimality of Bayes classifier | CML 1.1-1.2, 1.4, 2.1 | DL Part I

Lecture Generative vs. Discriminative classifiers, Nearest Neighbor classifier,


2 Coping with drawbacks of k-NN, Decision Trees, Model Complexity and
Overfitting
[lec 2]
notes/reading: NN and k-d trees | decision trees | CML 1.3, 2.2-2.5, 3.1-3.2

Lecture Decision boundaries for classification, Linear decision boundaries (Linear


3 classification), The Perceptron algorithm, Coping with non-linear
boundaries, Kernel feature transform, Kernel trick
[lec 3]
notes/reading: CML 4.1-4.8, 11.1-11.2

Lecture Support Vector Machines, Large margin formulation, Constrained


4 Optimization, Lagrange Duality, Convexity, Duality Theorems, Dual SVMs
[lec 4]
notes/reading: SVM basics | Lagrange Duality | CVX book | CML 7.7, 11.5-
6

Lecture Regression, Parametric vs. non-parametric regression, Ordinary least


5 squares regression, Logistic regression, Lasso and ridge regression, Optimal
regressor, Kernel regression, consistency of kernel regression
[lec 5]
notes/reading: Logistic regression | Kernel regression

Exam #1

Lecture Statistical theory of learning, PAC-learnability, Occam's razor theorem, VC


6 dimension, VC theorem, Concentration of measure
[lec 6]
notes/reading: PAC and VC tutorial | Intro. Learning Theory | CML 12.3,
12.5-6

Lecture Unsupervised Learning, Clustering, k-means, Hierarchical clustering,


7 Gaussian mixture modeling, Expectation Maximization Algorithm
[lec 7]
notes/reading: GMMs and EM | CML 15.1

Lecture Dimensionality Reduction, Principal Components Analysis (PCA), non-


linear dimension reduction (manifold learning)
8 [lec 8] notes/reading: PCA tutorial | CML 15.2 | Non-linear Dimension
Reduction

Lecture Graphical Models, Bayesian Networks, Markov Random Fields, Inference


9 and learning on graphical models, Markov Chains, Hidden Markov Models
(HMMs)
[lec 9] notes/reading: Graphical Models tutorial | HMM tutorial

Exam #2

Course Info

Topic
Topic:: Machine learning
Assessment
Assessment:Homeworks
: (40%)
Exam #1 (30%)
Exam #2 (30%)
Textbook
Textbook:: There is no textbook for the course. The relevant reading material will
be posted with the lectures. You may find the books in Resources
section helpful.

Resources
Books on The Elements of Statistical Learning by Hastie, Tibshirani and
ML
ML:: Friedman (link)
Pattern Recognition and Machine Learning by Bishop (link)
A Course in Machine Learning by Daume (link)
Deep Learning by Goodfellow, Bengio and Courville (link)
Software
Software:: MATLAB: download info, learning the basics.
Python: download info, tutorial
LaTeX: download info, learning the basics, typesetting common math
formulas.
Background
Background:
Probability:
: Events, random variables, expectations, joint, conditional
and marginal distributions, and independence. (refresher 1, refresher 2)

Statistics: Bayes' Rule, Priors, Posteriors, Maximum Likelihood


Principle (MLE), Basic distributions such as Bernoulli, Binomial,
Multinomial, Poisson, Gaussian. (refresher, reference sheet)

Linear Algebra: Vector spaces, subspaces, matrix inversion, matrix


multiplication, linear independence, rank, determinants,
orthonormality, basis, solving systems of linear equations. (refresher 1,
refresher 2, refresher 3, refresher 4)

Multivariate Calculus: Take derivatives and integrals of common


functions, gradient, Jacobian, Hessian, compute maxima and minima
of common functions. (basic calculus identities, multivariable
differentiation, extrema refresher, refresher 1, refresher 2)

Mathematical maturity: Ability to communicate technical ideas


clearly. (refresher 1, refresher 2)

Programming: Ability to program in a high-level language, and


familiarity with basic algorithm design and coding principles.

Misc.
Misc.:: The Matrix Cookbook
Convex Optimization

Policies
Homework Homeworks will contain a mix of programming and written
submissionassignments.
&
formatting
formatting::The written segment of the homeworks must be typesetted as a PDF
document, with all mathematical formulas properly formatted. It is
strongly recommended that you prepare your write-up as a LaTeX
document (you can use the following templates: homework.tex,
homework.cls and homework.pdf); see the Resources section for more
details on LaTeX. An MS Word writeup is also fine; you must make
sure that all math formulas are typesetted properly.

The programming segment can be done in any high-level


programming language. You are expected to have good working
knowledge of the programming language you choose, and will receive
very little 'debugging' help from the staff. You may use supporting
math/scientific libraries for auxiliary tasks (such as matrix
multiplication, matrix inversion, etc.), however you must write code
for the core algorithms yourself.

Please include your name and UNI on the first page of the written
assignment and at the top level comment of your programming
assignment.

The written segment of the homework (including plots and


comparative experimental studies) must be submitted via Gradescope,
and (if the homeworks specifies) the a tarball of the programming files
should be handed to the TA by the specified due dates. No late
homeworks will be accepted.

Collaboration
Collaboration:
Discussion
: of the homework problems is encouraged, but you must
write the solution individually or in small groups of 2-3 students (as
specified in the Homeworks). Detailed discussion of the solution must
only be discussed within the group. General discussion on problem
clarification and possible approaches can be discussed with others over
Piazza. Each group must write up their own solutions independently.
Late No late homeworks will be accepted.
submissions
submissions::
Academic Students are expected to adhere to the Academic Honesty policy of the
rules of Computer Science Department, this policy can be found in full here.
conduct
conduct::
Violations
Violations::Violation of any portion of these policies will result in a penalty to be
assessed at the instructor's discretion. This may include receiving a
zero grade for the assignment in question and a failing grade for the
whole course, even for the first infraction.

You might also like