520 likes | 534 Views
This paper discusses the limitations of standard active learning algorithms and proposes a re-active learning approach that combines uncertainty sampling with relabeling. It introduces the concept of impact sampling and demonstrates its effectiveness in improving label uncertainty and data quality. The paper also presents a case study using two different datasets to validate the proposed approach.
E N D
Re-active Learning: Active Learning with Re-labeling Christopher H. Lin University of Washington Mausam IIT Delhi Daniel S. Weld University of Washington
Majority Vote Parrot Parakeet Parrot Parrot
Relabel? Parakeet Parrot VS New label? Parakeet
MORE NOISY DATA LESS BETTER DATA
MORE NOISY DATA LESS BETTER DATA [Sheng et al. 2008, Lin et al. 2014]
Re-active Learning Contributions Standard Active Learning Algorithms Fail Uncertainty Sampling [Lewis and Catlett 1994] Expected Error Reduction [Roy and McCallum 2001] Re-active Learning Algorithms Extensions of Uncertainty Sampling Impact Sampling
h* True Hypothesis
h* h Current Hypothesis
h* h Uncertainty Sampling [Lewis and Catlett (1994)]
h* h Suppose labeled many times already!
h* h Uncertainty Sampling labels these two examples Infinitely many times!
Fundamental Problem: Does not use all sources of information h* h Uncertainty Sampling labels these two examples Infinitely many times!
Re-active Learning Contributions Standard Active Learning Algorithms Fail Uncertainty Sampling [Lewis and Catlett 1994] Expected Error Reduction [Roy and McCallum 2001] Re-active Learning Algorithms Extensions of Uncertainty Sampling Impact Sampling
Expected Error Reduction (EER) [Roy and McCallum (2001)] Also suffers from infinite looping!
Re-active Learning Contributions Standard Active Learning Algorithms Fail Uncertainty Sampling [Lewis and Catlett 1994] Expected Error Reduction [Roy and McCallum 2001] Re-active Learning Algorithms Extensions of Uncertainty Sampling Impact Sampling
How to fix? Consider the aggregate label uncertainty! ML h* h High # annotations = LOW UNCERTAINTY
How to fix? Consider the aggregate label uncertainty! ML Low # annotations = HIGH UNCERTAINTY h* h High # annotations = LOW UNCERTAINTY
Alpha-weighted uncertainty sampling (1-α) . Classifier uncertainty + α . Aggregate Label uncertainty
Fixed-Relabeling Uncertainty Sampling • Pick new unlabeled example using classifier uncertainty • Get a fixed number of labels for that example
Re-active Learning Contributions Standard Active Learning Algorithms Fail Uncertainty Sampling [Lewis and Catlett 1994] Expected Error Reduction [Roy and McCallum 2001] Re-active Learning Algorithms Extensions of Uncertainty Sampling Impact Sampling
h Current Hypothesis
Labeled Labeled h
Labeled Labeled h What is the impact of labeling this example?
Labeled Labeled h Impact of labeling this example a diamond
Labeled Labeled h Ψ (x) Impact of labeling this example a diamond
Labeled Labeled h Impact of labeling this example a circle
Labeled Labeled h Ψ (x) Impact of labeling this example a circle
Total Expected Impact of h Ψ (x)
Total Expected Impact of h Ψ (x) h Ψ (x)
Total Expected Impact of h Ψ (x) h Ψ (x) Ψ (x) = P(x = ) Ψ(x) + P(x = ) Ψ (x)
Use classifier’s belief as prior. Bayesian update using annotations. Ψ (x) = P(x = ) Ψ(x) + P(x = ) Ψ (x)
Assuming annotation accuracy > 0.5: As # annotations (x) goes to infinity, Ψ(x) goes 0.
Theorem In many noiseless settings, when relabeling is unnecessary, impact sampling = uncertainty sampling
Theorem In many noiseless settings, when relabeling is unnecessary, impact sampling = uncertainty sampling When relabeling is necessary: impact sampling = uncertainty sampling
Consider an example with the following labels: Aggregated Label via majority vote
Before: After adding an additional label: NO CHANGE
Pseudolookahead Let r be the minimum number of labels to flip the aggregate label.
Pseudolookahead Let r be the minimum number of labels to flip the aggregate label.
Pseudolookahead Let r be the minimum number of labels to flip the aggregate label. r = 3
Pseudolookahead Ψ(x) = Ψ (x) / r Redefine r
Pseudolookahead Ψ(x) = Ψ (x) / r Redefine r Careful Optimism!
Budget = 1000 Label Accuracy = 75% 10,30,50,70,90 Features
EER impact Alpha-uncertainty Fixed-uncertainty uncertainty passive Gaussian (num features = 90)
impact uncertainty passive Arrhythmia (num features = 279)