Skip navigation
Full metadata record
DC FieldValueLanguage
dc.contributor.authorSaar-Tsechansky, Maytal-
dc.contributor.authorProvost, Foster-
dc.date.accessioned2005-11-29T20:53:58Z-
dc.date.available2005-11-29T20:53:58Z-
dc.date.issued2000-
dc.identifier.urihttp://hdl.handle.net/2451/14167-
dc.description.abstractFor many supervised learning tasks, the cost of acquiring training data is dominated by the cost of class labeling. In this work, we explore active learning for class probability estimation (CPE). Active learning acquires data incrementally, using the model learned so far to help identify especially useful additional data for labeling. We present a new method for active learning, BootstrapLV, which chooses new data based on the variance in probability estimates from bootstrap samples. We then show empirically that the method reduces the number of data items that must be labeled, across a wide variety of data sets. We also compare Bootstrap-LV with Uncertainty Sampling, an existing active-learning method for maximizing classification accuracy, and show not only that BootstrapLV dominates for CPE but also that it is quite competitive even for accuracy maximization.en
dc.format.extent2365028 bytes-
dc.format.mimetypeapplication/pdf-
dc.languageEnglishEN
dc.language.isoen_US-
dc.publisherStern School of Business, New York Universityen
dc.relation.ispartofseriesIS-00-05-
dc.titleVariance-based Active Learningen
dc.typeWorking Paperen
dc.description.seriesInformation Systems Working Papers SeriesEN
Appears in Collections:IOMS: Information Systems Working Papers

Files in This Item:
File Description SizeFormat 
IS-00-05.pdf2.31 MBAdobe PDFView/Open


Items in FDA are protected by copyright, with all rights reserved, unless otherwise indicated.