EECS 545 - EECS @ Michigan - University of Michigan [PDF]

EECS 545: Machine Learning. Instructor: Honglak Lee, Clayton Scott. Coverage The goal of machine learning is to develop

5 downloads 23 Views 88KB Size

Recommend Stories


EECS 498
If your life's work can be accomplished in your lifetime, you're not thinking big enough. Wes Jacks

MICHIGAN PUbLISHING UNIVERSITY OF MICHIGAN PRESS
Never wish them pain. That's not who you are. If they caused you pain, they must have pain inside. Wish

Untitled - EECS Berkeley
What we think, what we become. Buddha

University of Michigan University Library
Ask yourself: When was the last time you did something that you were afraid of? Next

University of Michigan University Library
Open your mouth only if what you are going to say is more beautiful than the silience. BUDDHA

EECS 221: Multicore Programming
How wonderful it is that nobody need wait a single moment before starting to improve the world. Anne

University of Michigan University Library
You often feel tired, not because you've done too much, but because you've done too little of what sparks

University of Michigan University Library
Ego says, "Once everything falls into place, I'll feel peace." Spirit says "Find your peace, and then

EECS 516 Home Page
Don’t grieve. Anything you lose comes round in another form. Rumi

Western Michigan University
Ask yourself: How am I spending too much time on things that aren't my priorities? Next

Idea Transcript


Search

ACADEMICS+ADMISSIONS RESEARCH PEOPLE INDUSTRY NEWS+AWARDS EVENTS

ABOUT

RESOURCES GIVING

Academics +Admissions

Back to Course List

Why EECS at Michigan?

EECS 545: Machine Learning

Course Information

Instructor: Honglak Lee, Clayton Scott

Class Schedules & Books

Coverage The goal of machine learning is to develop computer algorithms that can learn from data or past experience to predict well on the new unseen data. In the past few decades, machine learning has become a powerful tool in artificial intelligence and data mining, and it has made major impacts in many real-world applications.

Undergraduate Programs Graduate Programs Admissions Student Visits Financial Aid

Research Areas + Quick Links +

This course will give a graduate-level introduction of machine learning and provide foundations of machine learning, mathematical derivation and implementation of the algorithms, and their applications. Topics include supervised learning, unsupervised learning, learning theory, graphical models, and reinforcement learning. This course will also cover recent research topics such as sparsity and feature selection, Bayesian techniques, and deep learning. In addition to mathematical foundations, this course will also put an emphasis on practical applications of machine learning to artificial intelligence and data mining, such as computer vision, data mining, speech recognition, text processing, bioinformatics, and robot perception and control. The course will require an open-ended research project. Textbook(s) Bishop, Christopher M. Pattern Recognition and Machine Learning. New York, NY: Springer, 2006. Syllabus Topics to be covered (tentative) Introduction (1 lecture) Overview Probability review Loss function Maximum likelihood MAP Regression (2 lectures) Linear regression Gradient descent and stochastic gradient Newton method Probabilistic interpretation of linear regression: Maximum likelihood Classification (2 lectures) k-nearst neighbors (kNN) Naive Bayes Linear discriminant analysis/ Gaussian discriminant analysis Logistic regression Generalized linear models, softmax regression Kernel methods (4 lectures) Kernel density estimation, kernel regression Support vector machines Convex optimization Gaussian processes Regularization (2 lectures) L2 regularization L1 regularization, sparsity and feature selection Bias-Variance tradeoff Overfitting Cross validation, model selection Advice for developing machine learning algorithms Neural networks (1 lecture) Perceptron MLP and back-propagation Learning theory (2 lectures) Sample complexity VC dimension PAC learning Error bounds Graphical models (4 lectures) Bayesian networks Representation Exact inference Sampling based inference Learning in Bayesian networks Maximum likelihood estimation Expectation maximization Hidden Markov Models (HMM) Structure learning Bayesian inference and learning Markov networks Inference and learning Unsupervised learning (4 lectures) Clustering: K-means Gaussian mixtures Expectation Maximization (revisited) PCA Dimensionality reduction: ISOMAP, LLE ICA Sparse coding Boltzmann machines and autoencoders, Deep belief networks Reinforcement learning (3 lectures) MDP Value iteration and policy iteration Dynamic programming Value function approximation TD learning

Contact Us College of Engineering

University of Michigan

© 2018 The Regents of the University of Michigan

Smile Life

When life gives you a hundred reasons to cry, show life that you have a thousand reasons to smile

Get in touch

© Copyright 2015 - 2024 PDFFOX.COM - All rights reserved.