1 / 45

Pushpak Bhattacharyya CSE Dept., IIT Bombay 20 th Jan , 2011

CS460/626 : Natural Language Processing/Speech, NLP and the Web (Lecture 8 – Closure on WSD; IWSD). Pushpak Bhattacharyya CSE Dept., IIT Bombay 20 th Jan , 2011. WordNet Sub-Graph. Hyponymy. Dwelling,abode. Hypernymy. Meronymy. kitchen. Hyponymy. bckyard. bedroom. M e r o n

pancho
Télécharger la présentation

Pushpak Bhattacharyya CSE Dept., IIT Bombay 20 th Jan , 2011

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. CS460/626 : Natural Language Processing/Speech, NLP and the Web(Lecture 8– Closure on WSD; IWSD) Pushpak BhattacharyyaCSE Dept., IIT Bombay 20th Jan, 2011

  2. WordNet Sub-Graph Hyponymy Dwelling,abode Hypernymy Meronymy kitchen Hyponymy bckyard bedroom M e r o n y m y house,home veranda A place that serves as the living quarters of one or mor efamilies Hyponymy study guestroom hermitage cottage Gloss

  3. WSD USING CONCEPTUAL DENSITY (Agirre and Rigau, 1996) • Select a sense based on the relatedness of that word-sense to the context. • Relatedness is measured in terms of conceptual distance • (i.e. how close the concept represented by the word and the concept represented by its context words are) • This approach uses a structured hierarchical semantic net (WordNet) for finding the conceptual distance. • Smaller the conceptual distance higher will be the conceptual density. • (i.e. if all words in the context are strong indicators of a particular concept then that concept will have a higher density.) 3

  4. CONCEPTUAL DENSITY FORMULA Wish list • The conceptual distance between two words should be proportional to the length of the path between the two words in the hierarchical tree (WordNet). • The conceptual distance between two words should be proportional to the depth of the concepts in the hierarchy. entity Sub-Tree d (depth) location finance h (height) of the concept “location” bank-2 bank-1 money where, c= concept nhyp= mean number of hyponyms h= height of the sub-hierarchy m= no. of senses of the word and senses of context words contained in the sub-ierarchy CD= Conceptual Density and 0.2 is the smoothing factor 4

  5. CONCEPTUAL DENSITY (cntd) • The dots in the figure represent the senses of the word to be disambiguated or the senses of the words in context. • The CD formula will yield highest density for the sub-hierarchy containing more senses. • The sense of W contained in the sub-hierarchy with the highest CD will be chosen. 5

  6. CONCEPTUAL DENSITY (EXAMPLE) administrative_unit body CD = 0.062 The jury(2) praised the administration(3) and operation (8) of Atlanta Police Department(1) division CD = 0.256 committee department government department local department jury operation police department jury administration Step 1: Make a lattice of the nouns in the context, their senses and hypernyms. Step 2: Compute the conceptual density of resultant concepts (sub-hierarchies). 6 Step 3: The concept with the highest CD is selected. Step 4: Select the senses below the selected concept as the correct sense for the respective words.

  7. WSD USING RANDOM WALK ALGORITHM (Page Rank) (sinha and Mihalcea, 2007) 0.46 0.97 0.42 S3 S3 S3 a b a Bell ring church Sunday c 0.49 e 0.35 0.63 S2 S2 S2 f k g h i 0.58 0.92 0.56 l 0.67 S1 S1 S1 S1 j Step 1: Add a vertex for each possible sense of each word in the text. Step 2: Add weighted edges using definition based semantic similarity (Lesk’s method). Step 3: Apply graph based ranking algorithm to find score of each vertex (i.e. for each word sense). 7 Step 4: Select the vertex (sense) which has the highest score.

  8. A look at Page Rank (from Wikipedia) Developed at Stanford University by Larry Page (hence the name Page-Rank) and Sergey Brin as part of a research project about a new kind of search engine The first paper about the project, describing PageRank and the initial prototype of the Google search engine, was published in 1998 Shortly after, Page and Brin founded Google Inc., the company behind the Google search engine While just one of many factors that determine the ranking of Google search results, PageRank continues to provide the basis for all of Google's web search tools

  9. A look at Page Rank (cntd) PageRank is a probability distribution used to represent the likelihood that a person randomly clicking on links will arrive at any particular page. Assume a small universe of four web pages: A, B, C and D. The initial approximation of PageRank would be evenly divided between these four documents. Hence, each document would begin with an estimated PageRank of 0.25. If pages B, C, and D each only link to A, they would each confer 0.25 PageRank to A. All PageRankPR( ) in this simplistic system would thus gather to A because all links would be pointing to A. PR(A)=PR(B)+PR(C)+PR(D) This is 0.75.

  10. A look at Page Rank (cntd) Suppose that page B has a link to page C as well as to page A, while page D has links to all three pages The value of the link-votes is divided among all the outbound links on a page. Thus, page B gives a vote worth 0.125 to page A and a vote worth 0.125 to page C. Only one third of D's PageRank is counted for A's PageRank (approximately 0.083). PR(A)=PR(B)/2+PR(C)/1+PR(D)/3 In general, PR(U)= ΣPR(V)/L(V), where B(u) is the set of pages u is linked to, and VεB(U) L(V) is the number of links from V

  11. A look at Page Rank (damping factor) The PageRank theory holds that even an imaginary surfer who is randomly clicking on links will eventually stop clicking. The probability, at any step, that the person will continue is a damping factor d. PR(U)= (1-d)/N + d.ΣPR(V)/L(V), VεB(U) N=size of document collection

  12. For WSD: Page Rank • Given a graph G = (V,E) • In(Vi) = predecessors of Vi • Out(Vi) = successors of Vi • In a weighted graph, the walker randomly selects an outgoing edge with higher probability of selecting edges with higher weight. 12

  13. Other Link Based Algorithms • HITS algorithm invented by Jon Kleinberg (used by Teoma and now Ask.com) • IBM CLEVER project • TrustRank algorithm.

  14. KB Approaches – Comparisons

  15. KB Approaches –Conclusions • Drawbacks of WSD using Selectional Restrictions • Needs exhaustive Knowledge Base. • Drawbacks of Overlap based approaches • Dictionary definitions are generally very small. • Dictionary entries rarely take into account the distributional constraints of different word senses (e.g. selectional preferences, kinds of prepositions, etc.  cigarette and ash never co-occur in a dictionary). • Suffer from the problem of sparse match. • Proper nouns are not present in a MRD. Hence these approaches fail to capture the strong clues provided by proper nouns.

  16. SUPERVISED APPROACHES

  17. NAÏVE BAYES • The Algorithm find the winner sense using sˆ= argmaxs ε senses Pr(s|Vw) • ‘Vw’ is a feature vector consisting of: • POS of w • Semantic & Syntactic features of w • Collocation vector (set of words around it)  typically consists of next word(+1), next-to-next word(+2), -2, -1 & their POS's • Co-occurrence vector (number of times w occurs in bag of words around it) • Applying Bayes rule and naive independence assumption sˆ= argmax s ε senses Pr(s).Πi=1nPr(Vwi|s) 17

  18. BAYES RULE AND INDEPENDENCE ASSUMPTION sˆ= argmaxs ε senses Pr(s|Vw) where Vwis the feature vector. • Apply Bayes rule: Pr(s|Vw)=Pr(s).Pr(Vw|s)/Pr(Vw) • Pr(Vw|s) can be approximated by independence assumption: Pr(Vw|s) = Pr(Vw1|s).Pr(Vw2|s,Vw1)...Pr(Vwn|s,Vw1,..,Vwn-1) =Πi=1nPr(Vwi|s) Thus, sˆ= argmaxsÎsenses Pr(s).Πi=1nPr(Vwi|s) sˆ= argmaxs ε senses Pr(s|Vw)

  19. ESTIMATING PARAMETERS • Parameters in the probabilistic WSD are: • Pr(s) • Pr(Vwi|s) • Senses are marked with respect to sense repository (WORDNET) Pr(s) = count(s,w) / count(w) Pr(Vwi|s) = Pr(Vwi,s)/Pr(s) = c(Vwi,s,w)/c(s,w)

  20. Supervised Approaches – Comparisons

  21. Supervised Approaches –Observations • General Comments • Use corpus evidence instead of relying of dictionary defined senses. • Can capture important clues provided by proper nouns because proper nouns do appear in a corpus. • Naïve Bayes • Suffers from data sparseness. • Since the scores are a product of probabilities, some weak features might pull down the overall score for a sense. • A large number of parameters need to be trained. • Decision Lists • A word-specific classifier. A separate classifier needs to be trained for each word. • Uses the single most predictive feature which eliminates the drawback of Naïve Bayes.

  22. Parameter Projection and Iterative WSD Language and Domain Adaptation

  23. Pioneering work at IITB on Multilingual WSD MiteshKhapra, SaurabhSohoney, AnupKulkarni and Pushpak Bhattacharyya, Value for Money: Balancing Annotation Effort, Lexicon Building and Accuracy for Multilingual WSD, Computational Linguistics Conference (COLING 2010), Beijing, China, August 2010. MiteshKhapra, AnupKulkarni, SaurabhSohoney and Pushpak Bhattacharyya, All Words Domain Adapted WSD: Finding a Middle Ground between Supervision and Unsupervision, Conference of Association of Computational Linguistics (ACL 2010), Uppsala, Sweden, July 2010. MiteshKhapra, Sapan Shah, PiyushKedia and Pushpak Bhattacharyya, Domain-Specific Word Sense Disambiguation Combining Corpus Based and Wordnet Based Parameters, 5th International Conference on Global Wordnet (GWC2010), Mumbai, Jan, 2010. MiteshKhapra, Sapan Shah, PiyushKedia and Pushpak Bhattacharyya, Projecting Parameters for Multilingual Word Sense Disambiguation, Empirical Methods in Natural Language Prfocessing (EMNLP09), Singapore, August, 2009.

  24. Motivation • Parallel corpora, wordnets and sense annotated corpora are scarce resources. • Challenges: Lack of resources, multiplicity of Indian languages. • Can we do annotation work in one language and find ways of reusing it for other languages? • Can a more resource fortunate language help a less resource fortunate language? CFILT - IITB

  25. Introduction • Aim: Perform WSD in a multilingual setting involving Hindi, Marathi, Bengali and Tamil • The wordnet and sense marked corpora of Hindi are used for all these languages • Methodology rests on a novel multilingual dictionary framework • Parameters are projected from Hindi to other languages • The domains of interest are Tourism and Health CFILT - IITB

  26. Related Work (1/2) • Knowledge Based Approaches • Lesk’s Algorithm, Walker’s algorithm, Conceptual density, PageRank • Fundamentally overlap based algorithms • Suffer from data sparsity, dictionary definitions being generally small • Broad-coverage algorithms, but, suffer from poor accuracies • Supervised Approaches • WSD using SVM, k-NN,Decision Lists • Typically word-specific classifiers with high accuracies • Need large training corpora - unsuitable for resource scarce languages CFILT - IITB

  27. Related Work (2/2) • Semi-Supervised/Unsupervised Approaches • Hyperlex,Decision Lists • Do not need large annotated corpora but are word-specific classifiers. • Not suited for broad-coverage • Hybrid approaches (Motivation for our work) • Structural Semantic Interconnections • Combine more than one knowledge sources (wordnet as well as a small amount of tagged corpora) • Suitable for broad-coverage CFILT - IITB No single existing solution to WSD completely meets our requirements of multilinguality, high domain accuracy and good performance in the face of not-so-large annotated corpora.

  28. Parameters for WSD (1/4) • Motivating example • The river flows through this region to meet the sea. • S1: (n) sea (a division of an ocean or a large body of salt water partially enclosed by land) • S2: (n) ocean, sea (anything apparently limitless in quantity or volume) • S3: (n) sea (turbulent water with swells of considerable size) "heavy seas“ CFILT - IITB What are the parameters that influence the choice of the correct sense for the word sea?

  29. Parameters for WSD (2/4) • Domain specific distributions • In the Tourism domain the “water-body” sense is more prevalent than the other senses • Domain-specific sense distribution information should be harnessed • Dominance of senses in a domain • {place, country, city, area}, {flora, fauna}, {mode of transport}, {fine arts} are dominant senses in the Tourism domain • A sense which belongs to the sub-tree of a dominant sense should be given a higher score than the other senses CFILT - IITB A synset node in the wordnet hypernymy hierarchy is called Dominantif the synsets in the sub-tree below the synset are frequently occurring in the domain corpora.

  30. Parameters for WSD (3/4) • Corpus Co-occurrence statistics • Co-occurring monosemous and/or already disambiguated words in the context help in disambiguation. • Example: The frequency of co-occurrence of river (monosemous) with “water-body” sense of sea is high • Semantic distance • Shortest path length between two synsets in the wordnet graph • An edge on this shortest path can be any semantic relation (hypernymy, hyponymy, meronymy, holonymy, etc.) • Conceptual distance between noun synsets CFILT - IITB

  31. Parameters for WSD (4/4) Summarizing parameters, • Wordnet-dependent parameters • belongingness-to-dominant-concept • conceptual-distance • semantic-distance • Corpus-dependent parameters • sense distributions • corpus co-occurrence CFILT - IITB

  32. Building a case for Parameter Projection • Wordnet-dependent parameters depend on the graph-based structure of wordnet • Corpus-dependent parameters depend on various statistics learnt from a sense marked corpora • Both the tasks, • Constructing a wordnet from scratch • Collecting sense marked corpora for multiple languages are tedious and expensive CFILT - IITB Can the effort required in constructing semantic graphs for multiple wordnets and collecting sense marked corpora in multiple languages be avoided?

  33. Synset based Multilingual Dictionary (1/2)RajatMohanty, Pushpak Bhattacharyya, PrabhakarPande, ShraddhaKalele, MiteshKhapra and Aditya Sharma. 2008. Synset Based Multilingual Dictionary: Insights, Applications and Challenges. Global Wordnet Conference, Szeged, Hungary, January 22-25. • Unlike traditional dictionary, synsets are linked, and after that the words inside the synsets are linked • Hindi is used as the central language – the synsets of all languages link to the corresponding Hindi synset. CFILT - IITB Advantage: The synsets in a particular column automatically inherit the various semantic relations of the Hindi wordnet – the wordnet based parameters thus get projected

  34. लड़का /HW1 ladakaa, बालक/HW2 baalak, बच्चा/HW3 bachcha, छोरा /HW4 choraa मुलगा/MW1mulagaa, पोरगा/MW2 poragaa, पोर /MW3 pora male-child /HW1, boy /HW2 English Synset Marathi Synset Hindi Synset Synset based Multilingual Dictionary (2/2) • Cross-linkages are set up manually from the words of a synset to the words of a linked synset of the central language • Such cross-linkages actually solve the problem of lexical choice in translating from text of one language to another. CFILT - IITB

  35. Sense Marked corpora Snapshot of a Marathi sense tagged paragraph

  36. saagar (sea) {water body} Sense_2650 samudra (sea) {water body} Sense_8231 saagar (sea) {abundance} saagar (sea) {abundance} Parameter Projection using MultiDict -P(Sense|Word) parameter (1/2) • P({water-body}|saagar) is given by • Using the cross-liked Hindi words we get P({water-body}|saagar) is • In general, CFILT - IITB

  37. Parameter Projection using MultiDict -P(Sense|Word) parameter (2/2) • For HindiMarathi • Average KL Divergence=0.29 • Spearman’s Correlation Coefficient=0.77 • For HindiBengali • Average KL Divergence=0.05 • Spearman’s Correlation Coefficient=0.82 CFILT - IITB There is a high degree of similarity between the distributions learnt using projection and those learnt from the self corpus.

  38. Comparison of projected and true sense distribution statistics for some Marathi words

  39. Parameter Projection using MultiDict -Co-occurrence parameter • Within a domain, the statistics of co-occurrence of senses remain the same across languages. • Co-occurrence of the synsets {cloud} and {sky} is almost same in the Marathi and Hindi corpus. CFILT - IITB

  40. Comparison of projected and true sense co-occurrences statistics for some Marathi words

  41. Algorithms for WSD – Iterative WSD Motivated by the Energy expression in Hopfield network CFILT - IITB

  42. Algorithms for WSD – Modified PageRank Modification Instead of using the overlap in dictionary definitions as edge weights, the wordnet and corpus based parameters are used to calculate edge weights CFILT - IITB

  43. Experimental Setup • Datasets • Tourism corpora in 4 languages (viz., Hindi, Marathi, Bengali and Tamil) • Healthcorpora in 2 languages (Hindi and Marathi) • A 4-fold cross validation was done for all the languages in both the domains CFILT - IITB

  44. Results CFILT - IITB

  45. Observations • IWSD performs better than PageRank • There is a drop in performance when we use parameter projection instead of using self corpora • Despite the drop in accuracy the performance is still better than the wordnet baseline • The performance is consistent in both the domains • One could trade accuracy with the cost of creating sense annotated corpora CFILT - IITB

More Related