1 / 168

Textual Entailment

Textual Entailment . Ido Dagan Bar Ilan University Israel. Fabio Massimo Zanzotto University of Rome Italy. Dan Roth, University of Illinois, Urbana-Champaign USA. ACL -2007 . Outline. Motivation and Task Definition A Skeletal review of Textual Entailment Systems

kelda
Télécharger la présentation

Textual Entailment

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Textual Entailment Ido Dagan Bar Ilan University Israel Fabio Massimo Zanzotto University of Rome Italy Dan Roth, University of Illinois, Urbana-Champaign USA ACL -2007

  2. Outline • Motivation and Task Definition • A Skeletal review of Textual Entailment Systems • Knowledge Acquisition Methods • Applications of Textual Entailment • A Textual Entailment view of Applied Semantics

  3. I. Motivation and Task Definition

  4. Motivation • Text applications require semantic inference • A common framework for applied semantics is needed, but still missing • Textual entailment may provide such framework

  5. Desiderata for Modeling Framework • A framework for a target level of language processing should provide: • Generic (feasible) module for applications • Unified (agreeable) paradigm for investigating language phenomena • Most semantics research is scattered • WSD, NER, SRL, lexical semantics relations… (e.g. vs. syntax) • Dominating approach - interpretation

  6. Variability Ambiguity Natural Language and Meaning Meaning Language

  7. Variability of Semantic Expression Model variabilityas relations between text expressions: • Equivalence: text1  text2 (paraphrasing) • Entailment:text1  text2 the general case The Dow Jones Industrial Average closed up 255 Dow ends up Dow gains 255 points Stock market hits a record high Dow climbs 255

  8. Typical Application Inference: Entailment QuestionExpected answer formWhoboughtOverture? >> XboughtOverture • Similar for IE: X acquire Y • Similar for “semantic” IR: t: Overture was bought for … • Summarization (multi-document) – identify redundant info • MT evaluation (and recent ideas for MT) • Educational applications Overture’s acquisitionby Yahoo Yahoo bought Overture entails hypothesized answer text

  9. KRAQ'05 Workshop - KNOWLEDGE and REASONING for ANSWERING QUESTIONS (IJCAI-05) CFP: • Reasoning aspects:    * information fusion,    * search criteria expansion models     * summarization and intensional answers,    * reasoning under uncertainty or with incomplete knowledge, • Knowledge representation and integration:    * levels of knowledge involved (e.g. ontologies, domain knowledge),    * knowledge extraction models and techniques to optimize response accuracy… but similar needs for other applications – can entailment provide a common empirical framework?

  10. Classical Entailment Definition • Chierchia & McConnell-Ginet (2001):A text t entails a hypothesis h if h is true in every circumstance (possible world) in which t is true • Strict entailment - doesn't account for some uncertainty allowed in applications

  11. “Almost certain” Entailments t:The technological triumph known as GPS … was incubated in the mind of Ivan Getting. h: Ivan Getting invented the GPS.

  12. Applied Textual Entailment • A directional relation between two text fragments: Text (t) and Hypothesis (h): • Operational (applied) definition: • Human gold standard - as in NLP applications • Assuming common background knowledge – which is indeed expected from applications

  13. Probabilistic Interpretation Definition: • t probabilistically entailshif: • P(h istrue | t) > P(h istrue) • tincreases the likelihood of h being true • ≡ Positive PMI – t provides information on h’s truth • P(h istrue | t ):entailment confidence • The relevant entailment score for applications • In practice: “most likely” entailment expected

  14. The Role of Knowledge • For textual entailment to hold we require: • text AND knowledgeh but • knowledge should not entail h alone • Systems are not supposed to validate h’s truth regardless of t (e.g. by searching h on the web)

  15. PASCAL Recognizing Textual Entailment (RTE) ChallengesEU FP-6 Funded PASCAL Network of Excellence 2004-7 Bar-Ilan University ITC-irst and CELCT, Trento MITRE Microsoft Research

  16. Generic Dataset by Application Use • 7 application settings in RTE-1, 4 in RTE-2/3 • QA • IE • “Semantic” IR • Comparable documents / multi-doc summarization • MT evaluation • Reading comprehension • Paraphrase acquisition • Most data created from actual applications output • RTE-2/3: 800 examples in development and test sets • 50-50% YES/NO split

  17. RTE Examples

  18. Participation and Impact • Very successful challenges, world wide: • RTE-1 – 17 groups • RTE-2 – 23 groups • ~150 downloads • RTE-3 – 25 groups • Joint workshop at ACL-07 • High interest in the research community • Papers, conference sessions and areas, PhD’s, influence on funded projects • Textual Entailment special issue at JNLE • ACL-07 tutorial

  19. Methods and Approaches (RTE-2) • Measure similarity match between t and h (coverage of h by t): • Lexical overlap (unigram, N-gram, subsequence) • Lexical substitution (WordNet, statistical) • Syntactic matching/transformations • Lexical-syntactic variations (“paraphrases”) • Semantic role labeling and matching • Global similarity parameters (e.g. negation, modality) • Cross-pair similarity • Detect mismatch (for non-entailment) • Interpretation to logic representation + logic inference

  20. Dominant approach: Supervised Learning • Features model similarity and mismatch • Classifier determines relative weights of information sources • Train on development set and auxiliary t-h corpora Similarity Features:Lexical, n-gram,syntactic semantic, global Classifier YES t,h NO Feature vector

  21. RTE-2 Results Average: 60% Median: 59%

  22. Analysis • For the first time: methods that carry some deeper analysis seemed (?) to outperform shallow lexical methods  Cf. Kevin Knight’s invited talk at EACL-06, titled: Isn’t linguistic Structure Important, Asked the Engineer • Still, most systems, which do utilize deep analysis, did not score significantly better than the lexical baseline

  23. Why? • System reports point at: • Lack of knowledge (syntactic transformation rules, paraphrases, lexical relations, etc.) • Lack of training data • It seems that systems that coped better with these issues performed best: • Hickl et al. - acquisition of large entailment corpora for training • Tatu et al. – large knowledge bases (linguistic and world knowledge)

  24. Some suggested research directions • Knowledge acquisition • Unsupervised acquisition of linguistic and world knowledge from general corpora and web • Acquiring larger entailment corpora • Manual resources and knowledge engineering • Inference • Principled framework for inference and fusion of information levels • Are we happy with bags of features?

  25. Complementary Evaluation Modes • “Seek” mode: • Input: h and corpus • Output: all entailing t ’s in corpus • Captures information seeking needs, but requires post-run annotation (TREC-style) • Entailment subtasks evaluations • Lexical, lexical-syntactic, logical, alignment… • Contribution to various applications • QA – Harabagiu & Hickl, ACL-06; RE – Romano et al., EACL-06

  26. II. A Skeletal review of Textual Entailment Systems

  27. Textual Entailment Entails Subsumed by Eyeing the huge market potential, currently led by Google, Yahoo took over search company Overture Services Inc. last year  Yahoo acquired Overture Overture is a search company Google is a search company Google owns Overture ………. Phrasal verb paraphrasing Entity matching Alignment Semantic Role Labeling How? Integration

  28. A general Strategy for Textual Entailment Given a sentence T Given a sentence H e Re-represent T Re-represent T Lexical Syntactic Semantic Lexical Syntactic Semantic Knowledge Base semantic; structural & pragmatic Transformations/rules Representation Decision Find the set of Transformations/Features of the new representation (or: use these to create a cost function) that allow embedding of H in T. Re-represent T Re-represent T Re-represent T Re-represent T Re-represent T Re-represent T Re-represent T

  29. Details of The Entailment Strategy • Knowledge Sources • Syntactic mapping rules • Lexical resources • Semantic Phenomena specific modules • RTE specific knowledge sources • Additional Corpora/Web resources • Control Strategy & Decision Making • Single pass/iterative processing • Strict vs. Parameter based • Justification • What can be said about the decision? • Preprocessing • Multiple levels of lexical pre-processing • Syntactic Parsing • Shallow semantic parsing • Annotating semantic phenomena • Representation • Bag of words, n-grams through tree/graphs based representation • Logical representations

  30. The Case of Shallow Lexical Approaches • Knowledge Sources • Shallow Lexical resources – typically Wordnet • Control Strategy & Decision Making • Single pass • Compute Similarity; use threshold tuned on a development set (could be per task) • Justification • It works • Preprocessing • Identify Stop Words • Representation • Bag of words

  31. Shallow Lexical Approaches (Example) • Lexical/word-based semantic overlap: score based on matching each word in H with some word in T • Word similarity measure: may use WordNet • May take account of subsequences, word order • ‘Learn’ threshold on maximum word-based match score Clearly, this may not appeal to what we think as understanding, and it is easy to generate cases for which this does not work well. However, it works (surprisingly) well with respect to current evaluation metrics Text:The Cassini spacecraft arrived at Titan in July, 2006. Text:NASA’s Cassini-Huygens spacecraft traveled to Saturn in 2006. Text:The Cassini spacecraft has taken images that show rivers on Saturn’s moon Titan. Hyp: The Cassini spacecraft has reached Titan.

  32. An Algorithm: LocalLexcialMatching • For each word in Hypothesis, Text • if word matches stopword – remove word • if no words left in Hypothesis or Text return 0 • numberMatched = 0; • for each word W_H in Hypothesis • for each word W_T in Text • HYP_LEMMAS = Lemmatize(W_H); • TEXT_LEMMAS = Lemmatize(W_T); • Use Wordnet’s • if any term in HYP_LEMMAS matches any term in TEXT_LEMMAS • using LexicalCompare() • numberMatched++; • Return: numberMatched/|HYP_Lemmas|

  33. An Algorithm: LocalLexicalMatching (Cont.) LLM Performance: RTE2: Dev: 63.00 Test: 60.50 RTE 3: Dev: 67.50 Test: 65.63 • LexicalCompare() • if(LEMMA_H == LEMMA_T) • return TRUE; • if(HypernymDistanceFromTo(textWord, hypothesisWord) <= 3) • return TRUE; • if(MeronymyDistanceFromTo(textWord, hypothesisWord) <= 3) • returnTRUE; • if(MemberOfDistanceFromTo(textWord, hypothesisWord) <= 3) • return TRUE: • if(SynonymOf(textWord, hypothesisWord) • return TRUE; • Notes: • LexicalCompare is Asymmetric & makes use of single relation type • Additional differences could be attributed to stop word list (e.g, including aux verbs) • Straightforward improvements such as bi-grams do not help. • More sophisticated lexical knowledge (entities; time) should help.

  34. Details of The Entailment Strategy (Again) • Knowledge Sources • Syntactic mapping rules • Lexical resources • Semantic Phenomena specific modules • RTE specific knowledge sources • Additional Corpora/Web resources • Control Strategy & Decision Making • Single pass/iterative processing • Strict vs. Parameter based • Justification • What can be said about the decision? • Preprocessing • Multiple levels of lexical pre-processing • Syntactic Parsing • Shallow semantic parsing • Annotating semantic phenomena • Representation • Bag of words, n-grams through tree/graphs based representation • Logical representations

  35. Preprocessing • Syntactic Processing: • Syntactic Parsing (Collins; Charniak; CCG) • Dependency Parsing (+types) • Lexical Processing • Tokenization; lemmatization • For each word in Hypothesis, Text • Phrasal verbs • Idiom processing • Named Entities + Normalization • Date/Time arguments + Normalization • Semantic Processing • Semantic Role Labeling • Nominalization • Modality/Polarity/Factive • Co-reference Only a few systems } often used only during decision making } often used only during decision making

  36. Details of The Entailment Strategy (Again) • Knowledge Sources • Syntactic mapping rules • Lexical resources • Semantic Phenomena specific modules • RTE specific knowledge sources • Additional Corpora/Web resources • Control Strategy & Decision Making • Single pass/iterative processing • Strict vs. Parameter based • Justification • What can be said about the decision? • Preprocessing • Multiple levels of lexical pre-processing • Syntactic Parsing • Shallow semantic parsing • Annotating semantic phenomena • Representation • Bag of words, n-grams through tree/graphs based representation • Logical representations

  37. Basic Representations MeaningRepresentation Inference Logical Forms Semantic Representation Representation Syntactic Parse Local Lexical Raw Text Textual Entailment • Most approaches augment the basic structure defined by the processing level with additional annotation and make use of a tree/graph/frame-based system.

  38. Basic Representations (Syntax) Syntactic Parse Local Lexical Hyp: The Cassini spacecraft has reached Titan.

  39. ARG_0 PRED ARG_1 PRED ARG_2 ARG_1 PRED ARG_1 PRED AM_TMP ARG_1 ARG_0 ARG_2 Basic Representations (Shallow Semantics: Pred-Arg ) T: The government purchase of the Roanoke building, a former prison, took place in 1902. H: The Roanoke building, which was a former prison, was bought by the government in 1902. take The govt. purchase… prison place in 1902 purchase The Roanoke building buy The Roanoke … prison In 1902 The government be a former prison The Roanoke building

  40. Basic Representations (Logical Representation) [Bos & Markert] The semantic representation language is a first-order fragment a language used in Discourse Representation Theory (DRS), conveying argument structure with a neo-Davidsonian analysis and Including the recursive DRS structure to cover negation, disjunction, and implication.

  41. Representing Knowledge Sources • Rather straight forward in the Logical Framework: • Tree/Graph base representation may also use rule based transformations to encode different kinds of knowledge, sometimes represented as generic or knowledge based tree transformations.

  42. Representing Knowledge Sources (cont.) • In general, there is a mix of procedural and rule based encoding of knowledge sources • Done by hanging more information on parse tree or predicate argument representation • Or different frame-based annotation systems for encoding information, that are processed procedurally.

  43. Details of The Entailment Strategy (Again) • Knowledge Sources • Syntactic mapping rules • Lexical resources • Semantic Phenomena specific modules • RTE specific knowledge sources • Additional Corpora/Web resources • Control Strategy & Decision Making • Single pass/iterative processing • Strict vs. Parameter based • Justification • What can be said about the decision? • Preprocessing • Multiple levels of lexical pre-processing • Syntactic Parsing • Shallow semantic parsing • Annotating semantic phenomena • Representation • Bag of words, n-grams through tree/graphs based representation • Logical representations

  44. Knowledge Sources • The knowledge sources available to the system are the most significant component of supporting TE. • Different systems draw differently the line between preprocessing capabilities and knowledge resources. • The way resources are handled is also different across different approaches.

  45. Enriching Preprocessing • In addition to syntactic parsing several approaches enrich the representation with various linguistics resources • Pos tagging • Stemming • Predicate argument representation: verb predicates and nominalization • Entity Annotation: Stand alone NERs with a variable number of classes • Acronym handling and Entity Normalization: mapping mentions of the same entity mentioned in different ways to a single ID. • Co-reference resolution • Dates, times and numeric values; identification and normalization. • Identification of semantic relations: complex nominals, genitives, adjectival phrases, and adjectival clauses. • Event identification and frame construction.

  46. Lexical Resources • Recognizing that a word or a phrase in S entails a word or a phrase in H is essential in determining Textual Entailment. • Wordnet is the most commonly used resoruce • In most cases, a Wordnet based similarity measure between words is used. This is typically a symmetric relation. • Lexical chains over Wordnet are used; in some cases, care is taken to disallow some chains of specific relations. • Extended Wordnet is being used to make use of Entities • Derivation relation which links verbs with their corresponding nominalized nouns.

  47. Lexical Resources (Cont.) • Lexical Paraphrasing Rules • A number of efforts to acquire relational paraphrase rules are under way, and several systems are making use of resources such as DIRT and TEASE. • Some systems seems to have acquired paraphrase rules that are in the RTE corpus • person killed --> claimed one life • hand reins over to --> give starting job to • same-sex marriage --> gay nuptials • cast ballots in the election -> vote • dominant firm --> monopoly power • death toll --> kill • try to kill --> attack • lost their lives --> were killed • left people dead --> people were killed

  48. Semantic Phenomena • A large number of semantic phenomena have been identified as significant to Textual Entailment. • A large number of them are being handled (in a restricted way) by some of the systems. Very little quantification per-phenomena has been done, if at all. • Semantic implications of interpreting syntactic structures • Conjunctions • Jake and Jill ran up the hill Jake ran up the hill • Jake and Jill met on the hill *Jake met on the hill • Clausal modifiers • But celebrations were muted as many Iranians observed a Shi'ite mourning month. • Many Iranians observed a Shi'ite mourning month. • Semantic Role Labeling handles this phenomena automatically

  49. Semantic Phenomena (Cont.) • Relative clauses • The assailants fired six bullets at the car, which carried Vladimir Skobtsov. • The car carried Vladimir Skobtsov. • Semantic Role Labeling handles this phenomena automatically • Appositives • Frank Robinson, a one-time manager of the Indians, has the distinction for the NL. • Frank Robinson is a one-time manager of the Indians. • Passive • We have been approached by the investment banker. • The investment banker approached us. • Semantic Role Labeling handles this phenomena automatically • Genitive modifier • Malaysia's crude palm oil output is estimated to have risen.. • The crude palm oil output of Malasia is estimated to have risen .

  50. Logical Structure • Factivity : Uncovering the context in which a verb phrase is embedded • The terrorists tried to enter the building. • The terrorists entered the building. • Polaritynegative markers or a negation-denoting verb (e.g. deny, refuse, fail) • The terrorists failed to enter the building. • The terrorists entered the building. • Modality/Negation Dealing with modal auxiliary verbs (can, must, should), that modify verbs’ meanings and with the identification of the scope of negation. • Superlatives/Comperatives/Monotonicity: inflecting adjectives or adverbs. • Quantifiers, determiners and articles

More Related