Workshop Organizers: Rahul Sukthankar, Compaq CRL and Robotics Institute, Carnegie Mellon Larry Wasserman, Department of Statistics, Carnegie Mellon Rich Caruana, Center for Automated Learning and Discovery, Carnegie Mellon |
|
Workshops |
Cross-validation and bootstrap are popular methods for estimating generalization error based on resampling a limited pool of data, and have become widely-used for model selection. The aim of this workshop is to bring together researchers from both matchine learning and statistics in an informal setting to discuss current issues in resampling-based techniques. These include:
Morning Session | |
---|---|
7:30-7:45 | Introduction (Rahul Sukthankar) |
7:45-8:45 | Invited speaker: Brad Efron, The Cost of Model Selection |
8:45-9:15 | Olivier Chapelle, Vladimir Vapnik, Yoshua Bengio: Model Selection for Small Sample Regression |
9:15-9:45 | John Langford: PAC Bounds for Holdout Procedures |
9:45-10:00 | Break |
10:00-10:30 | Carl Rasmussen, Pedro Højen-Sørensen: Empirical Model Comparison: Bayesian Analysis of Disjoint Cross Validation for Continuous Losses |
10:30-11:00 | Masashi Sugiyama, Hidemitsu Ogawa: Subspace Information Criterion - Unbiased Generalization Error Estimator for Linear Regression |
Evening Session | |
---|---|
16:30-17:00 | Bart Bakker, Tom Heskes: Model Clustering and Resampling |
17:00-17:30 | Matthew Mullin, Rahul Sukthankar: Complete Cross Validation for Nearest Neighbor Classifiers |
17:30-17:45 | Break |
17:45-18:15 | Ashutosh Garg, Ira Cohen, Thomas Huang: Sampling Based EM Algorithm |
18:15-18:45 | Koji Tsuda, Gunnar Rätsch, Sebastian Mika, Klaus-Robert Müller: Learning to Predict the Leave-one-out Error |
18:45-19:00 | Wrap-up discussion |
Brad Efron | talk slides |
Olivier Chapelle, Vladimir Vapnik, Yoshua Bengio | preprint |
John Langford | COLT-99
paper Webpage with additional information (including NIPS presentation) |
Masashi Sugiyama, Hidemitsu Ogawa | Neural Computation article |
Matthew Mullin, Rahul Sukthankar | ICML paper |
|
|
|
|