1 / 35

Active Learning and the Importance of Feedback in Sampling

Active Learning and the Importance of Feedback in Sampling. Rui Castro Rebecca Willett and Robert Nowak. Motivation – “twenty questions”. Goal: Accurately “learn” a concept, as fast as possible, by strategically focusing in regions of interest. Active Sampling in Regression.

hamish
Télécharger la présentation

Active Learning and the Importance of Feedback in Sampling

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Active Learningand the Importance of Feedback in Sampling Rui Castro Rebecca Willett and Robert Nowak

  2. Motivation – “twenty questions” Goal: Accurately “learn” a concept, as fast as possible, by strategically focusing in regions of interest

  3. Active Sampling in Regression Learning by asking carefully chosen questions, constructed using information gleaned from previous observations

  4. Passive Sampling Sample locations are chosen a priori, before any observations are made

  5. Active Sampling Sample locations are chosen as a function of previous observations

  6. Problem Formulation

  7. Passive vs. Active Passive Sampling: Active Sampling:

  8. Estimation and Sampling Strategies Goal: The estimator : The sampling strategy :

  9. Classical Smoothness Spaces Functions with homogeneous complexity over the entire domain • Hölder smooth function class

  10. Smooth Functions – minimax lower bound Theorem (Castro, RW, Nowak ’05) The performance one can achieve with active learning is the same achievable with passive learning!!!

  11. Inhomogeneous Functions Homogenous functions spread-out complexity Inhomogeneous functions localized complexity The relevant features of inhomogeneous functions are very localized in space, making active sampling promising

  12. best possible rate Piecewise Constant Functions – d≥2

  13. Passive Learning in the PC Class Estimation using Recursive Dyadic Partitions (RDP) Prune the partition, adapting to the data Recursively divide the domain into hypercubes Decorate each partition set with a constant Distribute sample points uniformly over [0,1]d

  14. empirical risk measures fit of the data Complexity penalty RDP-based Algorithm Choose an RDP that fits the data well, but it is not overly complicated This estimator can be computed efficiently using a tree-pruning algorithm.

  15. approximation error complexity penalty balancing the two terms Error Bounds Oracle bounding techniques, akin to the work of Barron’91, can be used to upper bound the performance of our estimator

  16. Active Sampling in the PC class Active Sampling Key: learn the location of the boundary Use Recursive Dyadic Partitions to find the boundary

  17. Active Sampling in the PC Class Stage 1: “Oversample” at coarse resolution • n/2 samples uniformly distributed • Limit the resolution: many more samples than cells • biased, but very low variance result • (high approximation error, but low estimation error) • “boundary zone” is • reliably detected

  18. Active Sampling in the PC Class Stage 2: Critically sample in boundary zone • n/2 samples uniformly distributed • within boundary zone • construct fine partition around boundary • prune partition according to • standard multiscale methods • high resolution • estimate of boundary

  19. Main Theorem Main Theorem (Castro ’05): * * Cusp-free boundaries cannot behave like the graph of |x|1/2 at the origin, but milder “kinks” like |x| at 0 are allowable.

  20. Sketch of the Proof - Approach

  21. Controlling the Bias Cells intersecting the boundary may be pruned if ‘aligned’ with cell edge • Solution: • Repeat Stage 1 d-times, using d slightly offset partitions • Small cells remaining in any of the d+1 partitions are passed on to Stage 2 Potential Problem Area Not a problem after shift

  22. Multi-Stage Approach Iterating the approach yields a L-step method Compare with minimax lower bound:

  23. Learning PC Functions - Summary Passive Sampling: Active Sampling: This rates are nearly achieved using RDP-based estimators, that are easily implemented and have low computational complexity.

  24. Spatial Adaptivity and Active Learning Spatially adaptive estimators based on “sparse” model selection (e.g., wavelet thresholding) may provide automatic mechanisms for guiding active learning processes Instead of choosing “where-to-sample” one can also choose “where-to-compute” to actively reduce computation. Can active learning provably work in even more realistic situations and under little or no prior assumptions ?

  25. Piecewise Constant Functions – d =1 Consider first the simplest non-homogenous function class step function This is a parametric class

  26. Passive Sampling Distribute sample points uniformly over [0,1] and use a maximum likelihood estimator

  27. Active Sampling

  28. Learning Rates – d =1 Passive Sampling: Active Sampling: (Burnashev & Zigangirov ’74)

  29. Sketch of the Proof - Stage 1 Error due to approximation of the boundary regions estimation error Intuition tells us that this should be the error we experience away from the boundary

  30. Sketch of the Proof - Stage 1 Key: Limit the resolution of the RDPs 1/k 1/k This is the performance away from the boundary

  31. Sketch of the Proof - Stage 1 Are we finding more than the boundary? Lemma: At least we are not detecting too many areas outside the boundary.

  32. Sketch of the Proof - Stage 2 n/2 more samples distributed uniformly over the boundary Total error contribution from boundary zone:

  33. Sketch of the Proof – Overall Error Error away from the boundary Error in the boundary region Balancing the two errors yields

More Related