Information Processing and Learning

10-704, Spring 2015

Akshay Krishnamurthy, Aarti Singh

Teaching Assistant: Kirthevasan Kandasamy
Class Assistant: Sandra Winkler



Lecture:

Tuesday and Thursday, 1:30 - 2:50 pm, Wean 4623

Recitation:

Monday, 6-7 pm, Doherty 2105

Office hrs:

Aarti: Wednesday 12-1pm GHC 8207
Akshay: Friday 3-4pm GHC 7507
Kirthevasan: Monday 4-5pm GHC8015 or 6-7pm DH2105


Course Description:

What's the connection between how many bits we can send over a channel and how accurately we can classify documents or fit a curve to data? Is there any link between decision trees, prefix codes and wavelet transforms? What about max-entropy and maximum likelihood, or universal coding and online learning?

This inter-disciplinary course will explore these and other questions that link the fields of information theory, signal processing, and machine learning, all of which aim to understand the information contained in data. The goal is to highlight the common concepts and establish concrete links between these fields that enable efficient information processing and learning.

We will do a short but introductory review of basic information theory, including entropy and fundamental limits of data compression, data processing and Fano's inequalities, channel capacity, and rate-distortion theory. Then we will dig into the connections to learning including: estimation of information theoretic quantities (such as entropy, mutual information, and divergence) and their applications in learning, information theoretic lower bounds for machine learning problems, duality of max entropy and maximum likelihood, connections between clustering and rate-distortion theory, universal coding and online learning, active learning and feedback channels, and more.

We expect that this course will cater to both students that have taken a basic information theory course and those that have not.

Emails:
Prerequisites: Fundamentals of Probability, Statistics, Linear Algebra and Real analysis

Textbooks:
Grading:
  • Homeworks (40%)
  • Project (25%)
  • Online Q&A (10%)
  • Two Short Quiz (20%)
  • Scribing (5%)
Homeworks: All homeworks, quizzes and solutions are posted here.

Course Project: Information about the course project is available here.

Schedule of Lectures: Lecture schedule, scribed notes and HWs are available here.

Tentative Schedule:
Week 1: Basics of information theory
Week 2: Some applications in machine learning
Week 3: Estimation of information theoretic quantities
Week 4: Maximum entropy distributions and exponential families, I-geometry
Week 5: Source coding and compression
Week 6: Model selection and connections to source coding
Week 7: Universal source coding and online learning
Week 8: Sufficient statistics, Information Bottleneck principle
Week 9: Channel coding and Redundancy-Capacity Theorem
Week 10: Fano's Inequality and minimax theory
Week 11: Minimax lower bounds for parametric and nonparametric problems
Week 12: Strong data processing inequalities and minimax lower bounds
Week 13: Classification and hypothesis testing
Week 14: Graphical models, LDPC codes, and active learning