1 / 26

A Generative Model for Parsing Natural Language to Meaning Representations

A Generative Model for Parsing Natural Language to Meaning Representations. Luke S. Zettlemoyer Massachusetts Institute of Technology. Wei Lu, Hwee Tou Ng, Wee Sun Lee National University of Singapore. Classic Goal of NLP: Understanding Natural Language.

Télécharger la présentation

A Generative Model for Parsing Natural Language to Meaning Representations

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. A Generative Model for Parsing Natural Language to Meaning Representations Luke S. Zettlemoyer Massachusetts Institute of Technology Wei Lu, Hwee Tou Ng, Wee Sun Lee National University of Singapore

  2. Classic Goal of NLP: Understanding Natural Language • Mapping Natural Language (NL) to Meaning Representations (MR) How many states do not have rivers ? … … … … … … … Meaning Representation Natural Language Sentence

  3. Meaning Representation (MR) QUERY:answer(NUM) NUM:count(STATE) STATE:exclude(STATE STATE) STATE:state(all) STATE:loc_1(RIVER) RIVER:river(all) How many states do not have rivers ?

  4. MR production • Meaning representation production (MR production) • Example: NUM:count(STATE) • Semantic category: NUM • Function symbol: count • Child semantic category: STATE • At most 2 child semantic categories

  5. Task Description • Training data: NL-MR pairs • Input: A new NL sentence • Output: An MR

  6. Challenge • Mapping of individual NL words to their associated MR productions is not given in the NL-MR pairs

  7. Mapping Words to MR Productions QUERY:answer(NUM) NUM:count(STATE) STATE:exclude(STATE STATE) STATE:state(all) STATE:loc_1(RIVER) RIVER:river(all) how many states do not have rivers ? 7

  8. Talk Outline • Generative model • Goal: flexible model that can parse a wide range of input sentences • Efficient algorithms for EM training and decoding • In practice: correct output is often in top-k list, but is not always the best scoring option • Reranking • Global features • Evaluation • Generative model combined with reranking technique achieves state-of-the-art performance

  9. Hybrid Tree NL-MR Pair QUERY:answer(NUM) NUM:count(STATE) STATE:exclude(STATE STATE) Hybrid sequences STATE:state(all) STATE:loc_1(RIVER) RIVER:river(all) How many states do not have rivers ?

  10. Model Parameters QUERY:answer(NUM) w: the NL sentence m: the MR T: the hybrid tree NUM:count(STATE) ? How many STATE:exclude(STATE STATE) STATE:state(all) do not STATE:loc_1(RIVER) states have RIVER:river(all) rivers P(w,m,T) =P(QUERY:answer(NUM)|-,arg=1) *P(NUM ?|QUERY:answer(NUM)) *P(NUM:count(STATE)|QUERY:answer(NUM),arg=1) *P(How many STATE|NUM:count(STATE)) *P(STATE:exclude(STATE STATE)|NUM:count(STATE),arg=1) *P(STATE1do notSTATE2|STATE:exclude(STATE STATE)) *P(STATE:state(all)|STATE:exclude(STATE STATE),arg=1) *P(states|STATE:state(all)) *P(STATE:loc_1(RIVER)|STATE:exclude(STATE STATE),arg=2) *P(have RIVER|STATE:loc_1(RIVER)) *P(RIVER:river(all)|STATE:loc_1(RIVER),arg=1) *P(rivers|RIVER:river(all)) MR Model Parameters ρ(m’|m,arg=k)

  11. Model Parameters QUERY:answer(NUM) w: the NL sentence m: the MR T: the hybrid tree NUM:count(STATE) ? How many STATE:exclude(STATE STATE) STATE:state(all) do not STATE:loc_1(RIVER) states have RIVER:river(all) rivers P(How many STATE|NUM:count(STATE)) = P(mwY|NUM:count(STATE)) * P(How|NUM:count(STATE),BEGIN) * P(many|NUM:count(STATE),How) * P(STATE|NUM:count(STATE),many) * P(END|NUM:count(STATE),STATE) Pattern Parameters Φ(r|m)

  12. Hybrid Patterns • M is an MR production, w is a word sequence • Y and Z are respectively the first and second child MR production • Note: [] denotes optional

  13. Model Parameters QUERY:answer(NUM) w: the NL sentence m: the MR T: the hybrid tree NUM:count(STATE) ? How many STATE:exclude(STATE STATE) STATE:state(all) do not STATE:loc_1(RIVER) states have RIVER:river(all) rivers P(How many STATE|NUM:count(STATE)) = P(mwY|NUM:count(STATE)) * P(How|NUM:count(STATE),BEGIN) * P(many|NUM:count(STATE),How) * P(STATE|NUM:count(STATE),many) * P(END|NUM:count(STATE),STATE) Emission Parameters θ(t|m,Λ)

  14. Assumptions : Model I, II, III NUM:count(STATE) BEGIN How many STATE END Model I Model II Model III Θ(ti|M,Λ) = P(ti|M) Θ(ti|M,Λ) = P(ti|M,ti-1) Θ(ti|M,Λ) = [P(ti|M,ti-1) + P(ti|M)] * 0.5 Unigram Model Bigram Model Mixgram Model

  15. Model Parameters • MR model parameters Σmi ρ(mi|mj,arg=k) = 1 They model the meaning representation • Emission parameters ΣtΘ(t|mj,Λ) = 1 They model the emission of words and semantic categories of MR productions. Λ is the context. • Pattern parameters ΣrΦ(r|mj) = 1 They model the selection of hybrid patterns

  16. Parameter Estimation • MR model parameters are easy to estimate • Learning the emission parameters and pattern parameters is challenging • Inside-outside algorithm with EM • Naïve implementation: O(n6m) • n: number of words in an NL sentence • m: number of MR productions in an MR • Improved efficient algorithm • Two-layer dynamic programming • Improved time complexity: O(n3m)

  17. Decoding • Given an NL sentence w, find the optimal MR M*: M* = argmaxm P(m|w) = argmaxmΣT P(m,T |w) = argmaxmΣT P(w,m,T ) • We find the most likely hybrid tree M* = argmaxmmaxT P(w,m,T ) • Similar DP techniques employed • Implemented Exact top-k decoding algorithm

  18. Reranking • Weakness of the generative model • Lacks the ability to model long range dependencies • Reranking with the averaged perceptron • Output space • Hybrid trees from exact top-k (k=50) decoding algorithm for each training/testing instance’s NL sentence • Single correct reference • Output of Viterbi algorithm for each training instance • Feature functions • Features 1-5 are indicator functions, while feature 6 is real-valued. • Threshold b that prunes unreliable predictions even when they score the highest, to optimize F-measure

  19. Reranking Features: Examples QUERY:answer(NUM) log(P(w,m,T)) NUM:count(STATE) ? How many STATE:exclude(STATE STATE) STATE:state(all) do not STATE:loc_1(RIVER) states have RIVER:river(all) rivers Feature 1: Hybrid Rule: A MR production and its child hybrid sequence Feature 2: Expanded Hybrid Rule: A MR production and its child hybrid sequence expanded Feature 3: Long-range Unigram: A MR production and a NL word appearing below in tree Feature 4: Grandchild Unigram: A MR production and its grandchild NL word Feature 5: Two Level Unigram: A MR production, its parent production, and its child NL word Feature 6: Model Log-Probability: Logarithm of base model’s joint probability

  20. Related Work • SILT (2005) by Kate, Wong, and Mooney • A system that learns deterministic rules to transform either sentences or their syntactic parse trees to meaning structures • WASP (2006) by Wong and Mooney • A system motivated by statistical machine translation techniques • KRISP (2006) by Kate and Mooney • A discriminative approach where meaning representation structures are constructed from the natural language strings hierarchically

  21. Evaluation Metrics • Precision # correct output structures # output structures • Recall # correct output structures # input sentences • F measure 2 1/Precision + 1/Recall

  22. Evaluations • Comparison over three models • I/II/III: Unigram/Bigram/Mixgram model; +R: w/ reranking • Reranking is shown to be effective • Overall, model III with reranking performs the best

  23. Evaluations • Comparison with other models • On Geoquery: • Able to handle more than 25% of the inputs that could not be handled by previous systems • Error reduction rate of 22%

  24. Evaluations • Comparison on other languages • Achieves performance comparable to previous system

  25. Contributions • Introduced a hybrid tree representation framework for this task • Proposed a new generative model that can be applied to the task of transforming NL sentences to MRs • Developed a new dynamic programming algorithm for efficient training and decoding • The approach, augmented with reranking, achieves state-of-the-art performance on benchmark corpora, with a notable improvement in recall

  26. Questions?

More Related