Show simple item record

dc.contributor.advisorZiebart, Brian
dc.creatorLiu, Anqi
dc.date.accessioned2018-11-27T16:29:03Z
dc.date.available2018-11-27T16:29:03Z
dc.date.created2018-08
dc.date.issued2018-08-02
dc.date.submittedAugust 2018
dc.identifier.urihttp://hdl.handle.net/10027/22988
dc.description.abstractIn real world machine learning applications, it is often not very realistic to assume that the training data distribution aligns with the testing data distribution. A relaxation is to assume the distribution shift only occurs on the input variable (covariates), while the conditional output distribution given the input variable (covariates) remains the same. This is called the covariate shift setting. Besides various examples of covariate shift in supervised learning tasks, one of the typical covariate shift scenarios is the sampling bias problem in pool-based active learning, in which the learner selects the labeled set, thus introducing a different input distribution from the unlabeled pool in each step of learning and prediction. In this thesis, we propose a general framework for robust prediction under covariate shift. Rather than focusing on minimizing a reweighted empirical loss on training data, we manage to more directly optimize the expected test loss with a minimax approach. The resulting predictor provides more randomized predictions on test data when it lacks training data distribution support and therefore avoids possible loss induced by over optimistic extrapolation of other predictors. This framework accounts for facilitating different loss function minimization and incorporating different feature functions and feature generalization assumptions. We discuss how the framework reduces to specific forms and the corresponding approaches to estimate the parameters. Moreover, we investigate active learning using robust prediction when the active learning step is constructed as a special case of robust covariate shift problem. We conduct experiments on synthetic biased benchmark datasets and natural covariate shift datasets to show performance of the robust prediction on real data. Additionally, we evaluate pool-based active learning using robust prediction on benchmark real data sets. We demonstrate a number of benefits over existing methods.
dc.format.mimetypeapplication/pdf
dc.subjectCovariate Shift
dc.subjectActive Learning
dc.subjectRobust Learning
dc.subjectClassification
dc.subjectKernel Methods
dc.titleRobust Prediction Methods for Covariate Shift and Active Learning
dc.typeThesis
thesis.degree.departmentComputer Science
thesis.degree.grantorUniversity of Illinois at Chicago
thesis.degree.levelDoctoral
thesis.degree.namePhD, Doctor of Philosophy
dc.contributor.committeeMemberLiu, Bing
dc.contributor.committeeMemberYu, Philip
dc.contributor.committeeMemberReyzin, Lev
dc.contributor.committeeMemberDudik, Miro
dc.type.materialtext
dc.contributor.chairZiebart, Brian


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record