1 / 30

Natural Language Processing

Natural Language Processing. Lecture Notes 10 Chapter 19 Computational Lexical Semantics Part 1: Supervised Word-Sense Disambiguation. Computational Lexical Semantics. Word-Sense Disambiguation (WSD) Computing word similarity Semantic role labeling

jcurran
Télécharger la présentation

Natural Language Processing

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Natural Language Processing Lecture Notes 10 Chapter 19 Computational Lexical Semantics Part 1: Supervised Word-Sense Disambiguation

  2. Computational Lexical Semantics • Word-Sense Disambiguation (WSD) • Computing word similarity • Semantic role labeling • AKA (also known as): case role or thematic role assignment

  3. WSD • As we saw, thematic roles and selectional restrictions depend on word senses • WSD could help machine translation, question answering, information extraction, information retrieval, and text classification

  4. WSD • Prototypical: • Input: a word instance in context and a fixed inventory of possible senses for that word • Output: one of the senses • The sense inventory depends on the task. E.g., For MT from English to Spanish, the sense inventory for an English word might be the set of possible Spanish translations

  5. Two Current Tasks • SENSEVAL – community competition, organized by ACL SIGLex • Lexical sample task: a pre-selected set of target words and fixed sense inventories for them • Supervised machine learning • One classifier per word (“line”, “interest”, “plant”) • All-words: system must sense tag all the content words in the corpus (which appear in the lexicon) • Data sparseness problems, so supervised learning not as feasible • So many polysemous words, that one classifier-per-word becomes less feasible

  6. Supervised ML Approaches • A training corpus of words manually tagged in context with senses is used to train a classifier that can tag words in new text

  7. Representations • Most supervised ML approaches represent the training data as • Vectors of feature/value pairs • So our first task is to extract training data from a corpus with respect to a particular instance of a target word • This typically consists of features of a window of text surrounding the target

  8. Representations • This is where ML and NLP intersect • If you stick to trivial surface features that are easy to extract from a text, then most of the work is in the ML system • If you decide to use features that require more analysis (say parse trees) then the ML part may be doing less work (relatively)

  9. Surface Representations • Collocational and co-occurrence information • Collocational • Encode features about the words that appear in specific positions to the right and left of the target word • Often limited to the words themselves as well as their part of speech • Warning: “collocation” may also mean ‘a word statistically correlated with the classification’, even in the statistical NLP community • Co-occurrence (bag-of-words) • Features characterizing the words that occur anywhere in the window regardless of position

  10. Examples • Example text (WSJ) • An electric guitar and bass player stand off to one side not really part of the scene, just as a sort of nod to gringo expectations perhaps • Assume a window of +/- 2 from the target

  11. Examples • Example text • An electric guitar andbassplayer stand off to one side not really part of the scene, just as a sort of nod to gringo expectations perhaps • Assume a window of +/- 2 from the target

  12. Collocational • Position-specific information about the words in the window • guitar and bassplayer stand • [guitar, NN, and, CJC, player, NN, stand, VVB] • In other words, a vector consisting of • […, position-i word, position-i part-of-speech…]

  13. Co-occurrence (bag-of-words) • Information about the words that occur within the window (an unordered set, ignoring positions) • Choose a subset of the words in the training data, W (e.g., not including stop words, and words which frequently appear with the target words) • Vector of binary features: feature w-i is 1 if w-i is in the window, 0 otherwise • Capture topic/domain information

  14. Bag-Of-Words Example • Assume we’ve settled on a possible vocabulary of 12 words that includes guitar and player but not and and stand • guitar and bassplayer stand • [0,0,0,1,0,0,0,0,0,1,0,0]

  15. Classifiers • Once we cast the WSD problem as a classification problem, then all sorts of techniques are possible • Naïve Bayes (the right thing to try first) • Decision lists • Decision trees • Boosting • Support vector machines • Nearest neighbor methods • …

  16. Classifiers • The choice of technique, in part, depends on the set of features that have been used • Some techniques work better/worse with features with numerical values • Some techniques work better/worse with features that have large numbers of possible values • For example, the feature the word to the left has a large number of possible values

  17. Naïve Bayes • Argmax P(sense|feature vector) • Rewriting with Bayes and assuming independence of the features

  18. Naïve Bayes • P(s) … just the prior of that sense. • As with part of speech tagging, not all senses will occur with equal frequency • P(vj|s) … conditional probability of some particular feature value given a particular sense • You can get both of these from a tagged corpus with the features encoded • Smoothing is needed for words in test data but not in training data

  19. Naïve Bayes Test • On a corpus of examples of uses of the word line, naïve Bayes achieved about 73% correct • Good? • Standalone: Prec,Recall,Acc,F • Baseline: most frequent sense • Baseline: the lesk algorithm (later) • Improve an NLP task: does the WSD-aware task system perform significantly better?

  20. Most-Frequent Sense Baseline • Depends on the corpus • McCarthy et al. (best paper at ACL-2005): • Unsupervised method for finding the most frequent sense in a corpus

  21. Naïve Bayes • One problem with NB (and other algorithms such as SVM) is that their conclusions are opaque • Difficult to examine the results and understand why the WSD system decided what it did • Decision lists and decision trees are more transparent. We’ll look at decision lists…

  22. Decision Lists • Like case, switch statements in PLs

  23. Learning DLs: one approach • Restrict the lists to rules that test a single feature (1-dl rules) • Evaluate each possible test and rank them based on how well they work. • Log-likelihood ratio: how discriminative a feature is between two senses • Abs(log(p(S1/F)p(S2|F)) • Use the top N tests, in order, as the decision list

  24. Yarowsky (1994) • On a binary(homonymy) distinction used the following metric to rank the tests • This gives about 95% on this test… • Is this better than the 73% on line we noted earlier?

  25. Bootstrapping • What if you don’t have enough data to train a system… • Bootstrap • Pick a word that you as an analyst think will co-occur with your target word in particular sense • Grep through your corpus for your target word and the hypothesized word • Assume that the target tag is the right one

  26. Bootstrapping • For bass • Assume play occurs with the music sense and fish occurs with the fish sense

  27. Bass Results

  28. Bootstrapping • Perhaps better • Use the little training data you have to train an inadequate system • Use that system to tag new data. • Use that larger set of training data to train a new system

  29. Problems • Given these general ML approaches, how many classifiers do I need to perform WSD robustly • One for each ambiguous word in the language • How do you decide what set of tags/labels/senses to use for a given word? • Depends on the application

  30. WordNet Bass • Tagging with this set of senses is probably impossibly hard and probably overkill for any realistic application • bass - (the lowest part of the musical range) • bass, bass part - (the lowest part in polyphonic music) • bass, basso - (an adult male singer with the lowest voice) • sea bass, bass - (flesh of lean-fleshed saltwater fish of the family Serranidae) • freshwater bass, bass - (any of various North American lean-fleshed freshwater fishes especially of the genus Micropterus) • bass, bass voice, basso - (the lowest adult male singing voice) • bass - (the member with the lowest range of a family of musical instruments) • bass -(nontechnical name for any of numerous edible marine and freshwater spiny-finned fishes)

More Related