Information Processing and Learning
10-704, Spring 2012
Teaching Assistant: Min Xu
Class Assistant: Michelle Martin
|
|
|
|
Lecture:
|
Tuesday and Thursday, 10:30 - 11:50 am, 4303 GHC, (Notes)
|
Recitation:
|
Wednesday 3-4pm, 8102 GHC
|
Office hrs:
|
Instructor: Mondays 3-4 pm, 8207 GHC
TA: Friday 2-3 pm, 8013 GHC Atrium
|
|
Course Description:
|
What's the connection between how many bits we can send over a channel and how accurately we can classify documents or fit a curve to data? Is there any connection between decision trees, prefix codes and wavelet transforms? What about error-correcting codes, graphical models and compressed sensing?
This course will explore such questions that link the fields of signal
processing and machine learning, both of which aim to extract
information from signals or data. The goal of this inter-disciplinary
course is to highlight the concepts common to these fields that together
enable efficient information processing and learning.
The topics will range from basics of information theory, entropy and
fundamental limits of data compression, channel capacity & least
informative priors, rate-distortion
theory, Kolmogorov complexity & online learning,
hypothesis testing - information theoretic
limits and lower bounds in machine learning, sequential testing, function
approximation using fourier and wavelet transforms,
as well as advanced topics including connections between error-correcting
codes, inference in graphical models and compressed sensing, as time permits.
|
Prerequisites:
|
Fundamentals of Probability, Statistics, Linear Algebra and Real analysis
|
Textbook:
|
|
Grading:
|
- Homeworks (40%)
- Project (35%)
- Two Short Quiz (15%)
- Scribing (10%)
|
Tentative Syllabus Outline:
|
This outline is subject to significant revision during the lectures.
For actual lecture topics and notes, please see here.
---------------------------------------------------------------------
Jan 16 - May 4 (15 weeks + 1 week spring break)
---------------------------------------------------------------------
Information Theoretic Foundations
---------------------------------------------------------------------
|
week 1 -
|
Introduction
Basics of info theory - entropy, relative entropy and mutual information
|
week 2 -
|
Data processing inequality & Sufficient statistics
Fano's Inequality
|
week 3 -
|
Max entropy distributions & Exponential families
Asymptotic equipartition property
|
week 4 -
|
Source coding/fundamental limits of data compression
Prefix codes & Kraft Inequality
|
week 5 -
|
Lossy source coding & rate distortion theory
|
week 6 -
|
Channel capacity & least informative priors
Joint source channel coding
|
week 7 -
|
Universal source coding & Online learning
Kolmogorov complexity, Occam's razor & minimum description length
|
---------------------------------------------------------------------
Decision Theory
---------------------------------------------------------------------
|
week 8 -
|
hypothesis testing - Likelihood Ratio Tests, GLRT, Neyman Pearson framework
|
week 9 -
|
information theoretic limits of hypothesis testing
& lower bounds in machine learning problems
|
week 10 -
|
Multiple hypothesis testing
FWER (Family Wise Error Rate), FDR (False Discovery Rate)
|
week 11 -
|
sequential/active testing
|
---------------------------------------------------------------------
Estimation theory
---------------------------------------------------------------------
|
week 12 -
|
function spaces & approximation theory
linear and nonlinear estimators
|
week 13 -
|
wavelets & decision trees
|
---------------------------------------------------------------------
Advanced topics
---------------------------------------------------------------------
|
week 14 -
|
connections between error-correcting codes, message passing,
inference in graphical models & compressed sensing
|
week 15 -
|
project presentations
|
---------------------------------------------------------------------
|
|
|