1 / 64

Active Learning by Labeling Features

Active Learning by Labeling Features. By. Gregory Druck Dept. of Computer Science University of Massachusetts Amherst, MA 01003 gdruck@cs.umass.edu. By. Burr Settles Dept. of Biostatistics & Medical Informatics Dept. of Computer Sciences University of Wisconsin Madison, WI 53706

machiko-rin
Télécharger la présentation

Active Learning by Labeling Features

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Active Learning by Labeling Features

  2. By • Gregory Druck • Dept. of Computer Science • University of Massachusetts • Amherst, MA 01003 • gdruck@cs.umass.edu

  3. By • Burr Settles • Dept. of Biostatistics & • Medical Informatics • Dept. of Computer Sciences • University of Wisconsin • Madison, WI 53706 • bsettles@cs.wisc.edu

  4. And • Andrew McCallum • Dept. of Computer Science • University of Massachusetts • Amherst, MA 01003 • mccallum@cs.umass.edu

  5. Abstract • Methods that learn from priorinformation about inputfeatures such as generalized expectation (GE) can train accurate models with very littleeffort. • This paperproposes an active learning approach, in which the machine solicits “labels” on features ratherthan instances.

  6. Abstract • In both simulated and realuser experiments on two sequence labeling tasks, • Our active learning with featuresoutperforms: • passive learningwithfeatures. ( active VS passive) • active learning with instances. (featuresVS instances)

  7. Abstract • Preliminary experiments suggest that: novel interfaces whichintelligently solicit labels on multiplefeatures , facilitate more efficientannotation.

  8. 1. Introduction • The application of machine learning to new problems is slowed by the need for labeled trainingdata.

  9. 1. Introduction • When output variables are structured, annotation can be particularly difficult and time-consuming. • For example, when training a conditional random field (Lafferty et al., 2001) to extract fields such as rent, contact, features, utilitiesfrom apartment classifieds, labeling 22 instances(2,540 tokens) provides only 66.1% accuracy.

  10. 1. Introduction • Recent work has used unlabeled data and limited prior information about input features to bootstrap accurate structured output models.

  11. 1. Introduction • For example, both Haghighi and Klein (2006) and Mann and McCallum (2008) have demonstrated results better than 66.1% on the apartments task described before by using only a list of 33 highly discriminative features and the labels they indicate • However, these methods have only been applied in scenarios in which the user supplies such prior knowledgebefore learning begins.

  12. 1. Introduction • In traditional active learning (Settles, 2009) machine queries the user for only the labels of instances • Here we solicit “labels” on input features • A labeled input feature denotes that a particular input feature, for example: the word "call", is highlyindicative of a particular label, such as "contact".

  13. Table 1. an excerptof a feature active learning session • Table 1: Two iterations of feature active learning. • Each table shows the features labeled, and the resulting change in accuracy • Note that the word includedwas labeled as both utilities and features, and* denotes a regularexpression feature.

  14. 1. Introduction • In this paper, we advocate using generalizedexpectation (GE) criteria (Mann and McCallum,2008) for learning with labeled features. • We provide an alternate treatment of the GE objective function used by Mann and McCallum (2008) and a novel speedup to the gradient computation. • Wethen provide a pool-basedfeature active learning algorithm that includes an option to skip queries, for cases in which a feature has no clear label. • We propose and evaluate featurequeryselection algorithms that aim to reduce model uncertainty, and compare to several baselines.

  15. 2. Conditional Random Fields • In this sectionwe describe the underlyingprobabilisticmodel for all methods in this paper. • We model the probability of the label sequence conditionedon the inputsequence usingfirst-order linear-chain conditional randomfields(CRFs) (Lafferty et al., 2001). • This probability where Zx is the partition function, featurefunctions fj consider the entire inputsequence and at mosttwoconsecutiveoutputvariables.

  16. 2. Conditional Random Fields • Provided a training data distribution we estimate CRF parameters by maximizing the conditional log likelihood of the training data. • We use numericaloptimization to maximize , which requires the gradient of with respect to the parameters.

  17. 2. Conditional Random Fields • the partial derivative with respect to parameter j is equal to the difference between the empirical expectation of Fj and the model expectation of Fj, where • We also include a zero-mean variance Gaussian prior on parameters in all experiments. 10 is a default value that works well in many settings.

  18. 2.1 Learning with missing labels • The training set may contain partially labeled sequences. Let z denote missing labels. • We estimateparameters with this data by maximizing themarginal log-likelihood of the observed labels. • We refer to this training method as maximum marginal likelihood (MML); it has also been explored by Quattoni et al. (2007).

  19. 2.1 Learning with missing labels • The gradient of can also be written as the difference of two expectations. • The first isan expectation over the empirical distribution ofxand y, and the model distribution of z. • The secondis a double expectation over the empirical distribution of x and the model distribution of y and z.

  20. 2.1 Learning with missing labels • We train models using with expectedgradient (Salakhutdinov et al., 2003). • To additionally leverage unlabeled data, we compare with entropy regularization (ER). • ERadds a term to the objective function that encouragesconfidentpredictions on unlabeled data. • Training of linear-chain CRFs with ER is described by Jiao et al. (2006).

  21. 3. Generalized Expectation Criteria • In this section, we givea briefoverview of generalized expectation criteria (GE) (Mann and McCallum, 2008; Druck et al., 2008) • and explainhowwe can use GE to learnCRF(conditional random fields) parameters with estimates of feature expectations and unlabeled data.

  22. 3. Generalized Expectation Criteria • GE criteria are terms in a parameter estimation objective function that express preferences on the value of a model expectation of some function. • Given a score function S, an empirical distribution , a model distribution , and a constraintfunction , the value of a GE criterion is

  23. 3. Generalized Expectation Criteria • GE provides a flexible framework for parameter estimation because each of these elements can take an arbitrary form. • The most important differencebetween GE and other parameter estimation methods is that it does not require a one-to-onecorrespondence between constraint functions Gk andmodel feature functions. • We leverage this flexibility to estimateparameters of feature-rich CRFswith a very small set of expectationconstraints.

  24. 3. Generalized Expectation Criteria • Constraint functions Gk can be normalized sothat the sum of the expectations of a set of functions is 1. • In this case, S may measure the divergence between the expectation of the constraintfunction and a target expectation . where

  25. 3. Generalized Expectation Criteria • It can be shown that the partialderivative of with respect to parameter j is proportional to the predictedcovariance between the model feature function Fj and the constraint function Gk.

  26. 3. Generalized Expectation Criteria • The partialderivative shows that GElearnsparametervalues for model feature functions basedon their predictedcovariance with the constraintfunctions. • GE can thus be interpreted as a bootstrapping method that uses the limitedtraining signal to learn about parametersfor related modelfeature functions.

  27. 3.1 Learning with feature-label distributions • Mann and McCallum (2008) apply GE to a linear-chain, first-order CRF. • In this section we providean alternate treatment that arrives at the same objective function from the general form describedin the previous section.

  28. 3.1 Learning with feature-label distributions • Often, feature functions in a first-order linear-chain CRF f are binary, and are the conjunctionof an observational test q(x,i) and a label pair test this notation for an indicator function that returns 1 if the condition in braces is satisfied, returns 0 otherwise.

  29. 3.1 Learning with feature-label distributions • The constraint functions Gk we use here decompose and operate similarly, except that they onlyinclude a test for a single label. • Single label constraints are easier for users to estimate and makeGE training more efficient. • Label transition structure can be learnedautomatically from single label constraints through the covariance-based parameter update of Equation 2.

  30. 3.1 Learning with feature-label distributions • For convenience,we can write Gyk to denote the constraint function that combines observation test k with a testfor label y. • We also add a normalization constant • Under this construction the expectation of Gyk isthe predicted conditional probability that the labelat some arbitrary position iis y when the observational test at i succeeds,

  31. 3.1 Learning with feature-label distributions • If we have a set of constraint functions { Gky : }, and we use the scorefunction in Equation 1, then the GE objective function specifies the minimization of the KL divergence between the model and target distributions over labels conditioned on the success of the observational test. • Computing the first term of the covariance in Equation 2 requires a marginal distribution over three labels, two of which will be consecutive, but the other of which could appear anywhere in the sequence.

  32. 3.1 Learning with feature-label distributions • We can compute this marginal using the algorithm of Mann and McCallum (2008). • As previously described, this algorithm is for a sequence of length n. • However, we make the following novel observation: we do notneed to compute the extra lattices for feature label pairs with , since this makes Equation 2 equal to zero.

  33. 3.1 Learning with feature-label distributions • In Mann and McCallum (2008), probabilities were smoothed so that . • If we assume that only a small number of labels m havenon-zero probability, then the time complexity ofthe gradient computation is . • In thispaper typically , while is 11 or 13. In experiments in this paper, using this optimization does not significantlyaffect final accuracy.

  34. 3.1 Learning with feature-label distributions • We use numericaloptimization to estimatemodel parameters. • In general GE objective functions are not convex. • Consequently, we initialize 0th-order CRF parameters using a sliding window logistic regression model trained with GE. • We also include a Gaussian prior on parameterswith in the objective function.

  35. 3.2 Learning with labeled features • The training procedure described above requiresa set of observational tests or input features withtarget distributions over labels. • Estimating a distribution could be a difficult task for an annotator. • Consequently, we abstract away from specifyinga distribution by allowing the user to assign labelsto features (c.f. Haghighi and Klein (2006) , Drucket al. (2008)).

  36. 3.2 Learning with labeled features • For example, we say that the wordfeature callhas label contact. • A label for a featuresimply indicates that the feature is a good indicatorof the label. • Note that features can have multiplelabels, for examplethe feature includedhas two labels in Table 1. • We convert an input featurewith a set of labels L into a distribution by assigning probability for each and probability 0 for each. • By assigning 0 probability tolabels ,we can use the speed-up described in the previous section.

  37. 4. Active Learning by Labeling Features • Feature active learning, presented in Algorithm 1,is a pool-based active learning algorithm (Lewisand Gale, 1994) (with a pool of features ratherthan instances). • The novel components of thealgorithm are an option to skipa query and thenotion that skipping and labeling have differentcosts. • The option to skip is important when using feature queries if a user don't knowhow to label some features.

  38. 4. Active Learning by Labeling Features • We advocate using GE forthe train procedure. • while the iteration costc < the maximum cost cmax, the feature query q that maximizes the query selection metric is selected. • The acceptfunction determines if the labeler will label q. • If q is labeled, itis added to the set of labeled features C, and thelabel cost clabel is added to c. • Otherwise, the skipcost cskip is added to c. • This process continues forN iterations.

  39. 4. Active Learning by Labeling Features

  40. 4.1 Feature query selection methods • In this section we propose feature query selection methods. • Queries with a higher scores are considered better candidates. • Note again that by features we mean observational tests qk(x,i). • Note: these are not feature selection methods since we are determining the featuresfor which supervisory feedback will be most helpful to the model, rather than determining whichfeatures will be part of the model.

  41. 4.1 Feature query selection methods • We propose to select queries that provide thelargest reduction in model uncertainty. • Expected Information Gain (EIG) of a query is theexpectation of the reduction in model uncertaintyoverall possible responses.

  42. 4.1 Feature query selection methods • we propose a tractable strategy for reducing model uncertainty, motivated by traditionaluncertaintysampling (Lewis and Gale, 1994). • Weassume that when a user responds to a query, thereduction in uncertainty will be equal to the Total Uncertainty (TU), the sum of the marginalentropies at the positions where the feature occurs.

  43. 4.1 Feature query selection methods • Total uncertainty, however, is highlybiased towards selecting frequentfeatures. • A mean uncertainty variant, normalized by the feature’scount, would tend to choose very infrequent features. • Consequently we propose a tradeoffbetween the two extremes, called weighted uncertainty (WU), that scales the mean uncertainty bythe log count of the feature in the corpus.

  44. 4.1 Feature query selection methods • Finally, we also suggest an uncertainty-based metric called diverse uncertainty (DU) that encourages diversity among queries by multiplying TUby the mean dissimilarity between the feature andpreviously labeled features. • For sequence labelingtasks, we can measure the relatedness of featuresusing distributional similarity. sim(qk; qj) returns the cosine similarity between context vectors of words occurring in a window of +-3.

  45. 4.1 Feature query selection methods • We contrast the notion of uncertainty describedabove with another type of uncertainty: the entropy of the predicted label distribution for the feature, or expectation uncertainty (EU).As abovewe also multiply by the log feature count. • EU is flawed because it will have a large value fornon-discriminative features.

  46. 5. Simulated User Experiments • In this section we experiment with an automated oracle labeler. • When presented an instance query, the oracle simply provides the true labels. • Whenpresented a feature query, the oracle first decideswhether to skip the query. • We have found thatusers are morelikely to label features that are relevant for only a few labels.

  47. 5. Simulated User Experiments • We estimate the effort of different labeling actions with preliminary experiments in which weobserve users labeling data for tenminutes • Userstook an average of 4 seconds to label a feature, 2seconds to skip a feature, and 0.7 seconds to label a token. • cmax = 60, cskip = 2 and clabel = 4 • Forinstance active learning, we use Algorithm 1 but without the skip option, and set clabel= 0.7

  48. 5. Simulated User Experiments • Weuse N = 10 iterations, so the entire experiment simulates 10 minutes of annotation time. • For efficiency, we consider the 500 most frequent unlabeled features in each iteration. • To start, ten randomly selected seed labeled features are provided.

  49. 5. Simulated User Experiments • We use random (rand) selection,uncertaintysampling (US) (using sequence entropy, normalized by sequence length) and information density (ID) (Settles and Craven, 2008) to select instance queries. • We use Entropy Regularization(ER) (Jiao et al., 2006) to leverage unlabeled instances. • We weight the ER term by choosing thebest weight in multiplied by for each data set and query selection method. • Seedinstances are provided suchthat the simulated labeling time is equivalent to labeling 10 features.

  50. 5. Simulated User Experiments • We evaluate on two sequence labeling tasks. • The apartments task involves segmenting 300 apartment classified ads into 11 fields including features, rent, neighborhood, and contact. • The corareferences task is to extract 13 BibTeX fields such as authorand booktitlefrom 500 research paper references. • All results are averaged over ten random 80:20 splits of the data.

More Related