did-you-know? rent-now

Amazon no longer offers textbook rentals. We do!

did-you-know? rent-now

Amazon no longer offers textbook rentals. We do!

We're the #1 textbook rental company. Let us show you why.

9780262017091

Machine Learning in Non-Stationary Environments

by ;
  • ISBN13:

    9780262017091

  • ISBN10:

    0262017091

  • Format: Hardcover
  • Copyright: 2012-03-30
  • Publisher: Mit Pr
  • Purchase Benefits
  • Free Shipping Icon Free Shipping On Orders Over $35!
    Your order must be $35 or more to qualify for free economy shipping. Bulk sales, PO's, Marketplace items, eBooks and apparel do not qualify for this offer.
  • eCampus.com Logo Get Rewarded for Ordering Your Textbooks! Enroll Now
List Price: $19.75

Summary

As the power of computing has grown over the past few decades, the field of machine learning has advanced rapidly in both theory and practice. Machine learning methods are usually based on the assumption that the data generation mechanism does not change over time. Yet real-world applications of machine learning, including image recognition, natural language processing, speech recognition, robot control, and bioinformatics, often violate this common assumption. Dealing with non-stationarity is one of modern machine learning's greatest challenges. This book focuses on a specific non-stationary environment known as covariate shift, in which the distributions of inputs (queries) change but the conditional distribution of outputs (answers) is unchanged, and presents machine learning theory, algorithms, and applications to overcome this variety of non-stationarity. After reviewing the state-of-the-art research in the field, the authors discuss topics that include learning under covariate shift, model selection, importance estimation, and active learning. They describe such real world applications of covariate shift adaption as brain-computer interface, speaker identification, and age prediction from facial images. With this book, they aim to encourage future research in machine learning, statistics, and engineering that strives to create truly autonomous learning machines able to learn under non-stationarity.

Author Biography

Masashi Sugiyama is Associate Professor in the Department of Computer Science at Tokyo Institute of Technology. Motoaki Kawanabe is a Postdoctoral Researcher in Intelligent Data Analysis at the Fraunhofer FIRST Institute, Berlin. In October 2011, he moved to Advanced Telecommunications Research Institute International (ATR) in Kyoto, Japan.

Table of Contents

Forewordp. xi
Prefacep. xiii
Introduction
Introduction and Problem Formulationp. 3
Machine Learning under Covariate Shiftp. 3
Quick Tour of Covariate Shift Adaptationp. 5
Problem Formulationp. 7
Function Learning from Examplesp. 7
Loss Functionsp. 8
Generalization Errorp. 9
Covariate Shiftp. 9
Models for Function Learningp. 10
Specification of Modelsp. 13
Structure of This Bookp. 14
Part II: Learning under Covariate Shiftp. 14
Part III: Learning Causing Covariate Shiftp. 17
Learning Under Covariate Shift
Function Approximationp. 21
Importance-Weighting Techniques for Covariate Shift Adaptationp. 22
Importance-Weighted ERMp. 22
Adaptive IWERMp. 23
Regularized IWERMp. 23
Examples of Importance-Weighted Regression Methodsp. 25
Squared Loss: Least-Squares Regressionp. 26
Absolute Loss: Least-Absolute Regressionp. 30
Huber Loss: Huber Regressionp. 31
Deadzone-Linear Loss: Support Vector Regressionp. 33
Examples of Importance-Weighted Classification Methodsp. 35
Squared Loss: Fisher Discriminant Analysisp. 36
Logistic Loss: Logistic Regression Classifierp. 38
Hinge Loss: Support Vector Machinep. 39
Exponential Loss: Boostingp. 40
Numerical Examplesp. 40
Regressionp. 40
Classificationp. 41
Summary and Discussionp. 45
Model Selectionp. 47
Importance-Weighted Akaike Information Criterionp. 47
Importance-Weighted Subspace Information Criterionp. 50
Input Dependence vs. Input Independence in Generalization Error Analysisp. 51
Approximately Correct Modelsp. 53
Input-Dependent Analysis of Generalization Errorp. 54
Importance-Weighted Cross-Validationp. 64
Numerical Examplesp. 66
Regressionp. 66
Classificationp. 69
Summary and Discussionp. 70
Importance Estimationp. 73
Kernel Density Estimationp. 73
Kernel Mean Matchingp. 75
Logistic Regressionp. 76
Kullback-Leibler Importance Estimation Procedurep. 78
Algorithmp. 78
Model Selection by Cross-Validationp. 81
Basis Function Designp. 82
Least-Squares Importance Fittingp. 83
Algorithmp. 83
Basis Function Design and Model Selectionp. 84
Regularization Path Trackingp. 85
Unconstrained Least-Squares Importance Fittingp. 87
Algorithmp. 87
Analytic Computation of Leave-One-Out Cross-Validationp. 88
Numerical Examplesp. 88
Settingp. 90
Importance Estimation by KLIEPp. 90
Covariate Shift Adaptation by IWLS and IWCVp. 92
Experimental Comparisonp. 94
Summaryp. 101
Direct Density-Ratio Estimation with Dimensionality Reductionp. 103
Density Difference in Hetero-Distributional Subspacep. 103
Characterization of Hetero-Distributional Subspacep. 104
Identifying Hetero-Distributional Subspacep. 106
Basic Ideap. 106
Fisher Discriminant Analysisp. 108
Local Fisher Discriminant Analysisp. 109
Using LFDA for Finding Hetero-Distributional Subspacep. 112
Density-Ratio Estimation in the Hetero-Distributional Subspacep. 113
Numerical Examplesp. 113
Illustrative Examplep. 113
Performance Comparison Using Artificial Data Setsp. 117
Summaryp. 121
Relation to Sample Selection Biasp. 125
Heckman's Sample Selection Modelp. 125
Distributional Change and Sample Selection Biasp. 129
The Two-Step Algorithmp. 131
Relation to Covariate Shift Approachp. 134
Applications of Covariate Shift Adaptationp. 137
Brain-Computer Interfacep. 137
Backgroundp. 137
Experimental Setupp. 138
Experimental Resultsp. 140
Speaker Identificationp. 142
Backgroundp. 142
Formulationp. 142
Experimental Resultsp. 144
Natural Language Processingp. 149
Formulationp. 149
Experimental Resultsp. 151
Perceived Age Prediction from Face Imagesp. 152
Backgroundp. 152
Formulationp. 153
Incorporating Characteristics of Human Age Perceptionp. 153
Experimental Resultsp. 155
Human Activity Recognition from Accelerometric Datap. 157
Backgroundp. 157
Importance-Weighted Least-Squares Probabilistic Classifierp. 157
Experimental Results.p. 160
Sample Reuse in Reinforcement Learningp. 165
Markov Decision Problemsp. 165
Policy Iterationp. 166
Value Function Approximationp. 167
Sample Reuse by Covariate Shift Adaptationp. 168
On-Policy vs. Off-Policyp. 169
Importance Weighting in Value Function Approximationp. 170
Automatic Selection of the Flattening Parameterp. 174
Sample Reuse Policy Iterationp. 175
Robot Control Experimentsp. 176
Learning Causing Covariate Shift
Active Learningp. 183
Preliminariesp. 183
Setupp. 183
Decomposition of Generalization Errorp. 185
Basic Strategy of Active Learningp. 188
Population-Based Active Learning Methodsp. 188
Classical Method of Active Learning for Correct Modelsp. 189
Limitations of Classical Approach and Countermeasuresp. 190
Input-Independent Variance-Only Methodp. 191
Input-Dependent Variance-Only Methodp. 193
Input-Independent Bias-and-Variance Approachp. 195
Numerical Examples of Population-Based Active Learning Methodsp. 198
Setupp. 198
Accuracy of Generalization Error Estimationp. 200
Obtained Generalization Errorp. 202
Pool-Based Active Learning Methodsp. 204
Classical Active Learning Method for Correct Models and Its Limitationsp. 204
Input-Independent Variance-Only Methodp. 205
Input-Dependent Variance-Only Methodp. 206
Input-Independent Bias-and-Variance Approachp. 207
Numerical Examples of Pool-Based Active Learning Methodsp. 209
Summary and Discussionp. 212
Active Learning with Model Selectionp. 215
Direct Approach and the Active Learning/Model Selection Dilemmap. 215
Sequential Approachp. 216
Batch Approachp. 218
Ensemble Active Learningp. 219
Numerical Examplesp. 220
Settingp. 220
Analysis of Batch Approachp. 221
Analysis of Sequential Approachp. 222
Comparison of Obtained Generalization Errorp. 222
Summary and Discussionp. 223
Applications of Active Learningp. 225
Design of Efficient Exploration Strategies in Reinforcement Learningp. 225
Efficient Exploration with Active Learningp. 225
Reinforcement Learning Revisitedp. 226
Decomposition of Generalization Errorp. 228
Estimating Generalization Error for Active Learningp. 229
Designing Sampling Policiesp. 230
Active Learning in Policy Iterationp. 231
Robot Control Experimentsp. 232
Wafer Alignment in Semiconductor Exposure Apparatusp. 234
Conclusions
Conclusions and Future Prospectsp. 241
Conclusionsp. 241
Future Prospectsp. 242
Appendix: List of Symbols and Abbreviationsp. 243
Bibliographyp. 247
Indexp. 259
Table of Contents provided by Ingram. All Rights Reserved.

Supplemental Materials

What is included with this book?

The New copy of this book will include any supplemental materials advertised. Please check the title of the book to determine if it should include any access cards, study guides, lab manuals, CDs, etc.

The Used, Rental and eBook copies of this book are not guaranteed to include any supplemental materials. Typically, only the book itself is included. This is true even if the title states it includes any access cards, study guides, lab manuals, CDs, etc.

Rewards Program