380 likes | 387 Vues
Expanding WordNet to create a coherent representation of text by encoding world knowledge and constructing logical forms, enabling deep understanding beyond parsing and word senses.
E N D
Augmenting WordNet for Deep Understanding of Text Peter Clark, Phil Harrison, Bill Murray, John Thompson (Boeing) Christiane Fellbaum (Princeton Univ) Jerry Hobbs (ISI/USC)
“The soldier died” “The soldier was shot” “There was a fight” … “A soldier was killed in a gun battle” “Deep Understanding” • Not (just) parsing + word senses • Construction of a coherent representation of the scene the text describes • Challenge: much of that representation is not in the text
“Deep Understanding” Because A battle involves a fight. Soldiers use guns. Guns shoot. Guns can kill. If you are killed, you are dead. …. How do we get this knowledge into the machine? How do we exploit it? “The soldier died” “The soldier was shot” “There was a fight” … “A soldier was killed in a gun battle”
“Deep Understanding” Because A battle involves a fight. Soldiers use guns. Guns shoot. Guns can kill. If you are killed, you are dead. …. Several partially useful resources exist. WordNet is already used a lot…can we extend it? “The soldier died” “The soldier was shot” “There was a fight” … “A soldier was killed in a gun battle”
The Initial Vision • Our vision: Rapidly expand WordNet to be more of a knowledge-base Question-answering software to demonstrate its use
The Evolution of WordNet lexical resource • v1.0 (1986) • synsets (concepts) + hypernym (isa) links • v1.7 (2001) • add in additional relationships • has-part • causes • member-of • entails-doing (“subevent”) • v2.0 (2003) • introduce the instance/class distinction • Paris isa Capital-City is-type-of City • add in some derivational links • explode related-to explosion • … • v10.0 (200?) • ????? knowledge base?
Augmenting WordNet • World Knowledge • Sense-disambiguate the glosses (by hand) • Convert the glosses to logic • Similar to LCC’s Extended WordNet attempt • Axiomatize “core theories” • WordNet links • Morphosemantic links • Purpose links • Experiments
LFToolkit: Generate logical form fragments Lexical output rules produce logical form fragments strong drive for success strong(x1) & drive(x2) & for(x3,x4) & success(x5) Converting the Glosses to Logic “ambition#n2: A strong drive for success” Convert gloss to form “word is gloss” Parse (Charniak)
Converting the Glosses to Logic “ambition#n2: A strong drive for success” Convert gloss to form “word is gloss” Parse (Charniak) LFToolkit: Generate logical form fragments Identify equalities, add senses
Converting the Glosses to Logic Composition rules identify variables x2=x3 x4=x5 x1=x2 Lexical output rules produce logical form fragments A strong drive for success strong(x1) & drive(x2) & for(x3,x4) & success(x5) Identify equalities, add senses
ambition#n2(x1) → a(x1) & strong#a1(x1) & drive#n2(x1) & for(x1,x2) & success#a3(x2) Converting the Glosses to Logic “ambition#n2: A strong drive for success” Convert gloss to form “word is gloss” Parse (Charniak) LFToolkit: Generate logical form fragments Identify equalities, add senses
“hammer#n2: tool used to deliver an impulsive force by striking” hammer#n2(x1) → tool#n1(x1) & use#v1(e1,x2,x1) & to(e1,e2) & deliver#v2(e2,x3) & driving#a1(x3) & force#n1(x3) & by(e3,e4) & strike#v3(e4,x4). → Hammers hit things?? Converting the Glosses to Logic • Sometimes works well! • But often not. Primary problems: • Errors in the language processing • Only capture definitional knowledge • “flowery” language, many gaps, metonymy, ambiguity; If logic closely follows syntax → “logico-babble”
Augmenting WordNet • World Knowledge • Sense-disambiguate the glosses (by hand) • Convert the glosses to logic • Axiomatize “core theories” • WordNet links • Morphosemantic links • Purpose links • Experiments
Core Theories • Many domain-specific facts are instantiations of more general, “core” knowledge • By encoding this core knowledge, get leverage • eg 517 “vehicle” noun (senses), 185 “cover” verb (senses) • Approach: • Analysis and grouping of words in Core WordNet • Identification and encoding of underlying theories
Core Theories Composite Entities: perfect, empty, relative, secondary, similar, odd, ... Scales: step, degree, level, intensify, high, major, considerable, ... Events: constraint, secure, generate, fix, power, development, ... Space: grade, inside, lot, top, list, direction, turn, enlarge, long, ... Time: year, day, summer, recent, old, early, present, then, often, ... Cognition: imagination, horror, rely, remind, matter, estimate, idea, ... Communication: journal, poetry, announcement, gesture, charter, ... Persons and their Activities: leisure, childhood, glance, cousin, jump, ... Microsocial: virtue, separate, friendly, married, company, name, ... Material World: smoke, shell, stick, carbon, blue, burn, dry, tough, ... Geo: storm, moon, pole, world, peak, site, village, sea, island, ... Artifacts: bell, button, van, shelf, machine, film, floor, glass, chair, ... Food: cheese, potato, milk, break, cake, meat, beer, bake, spoil, ... Macrosocial: architecture, airport, headquarters, prosecution, ... Economic: import, money, policy, poverty, profit, venture, owe, ...
Augmenting WordNet • World Knowledge • Sense-disambiguate the glosses (by hand) • Convert the glosses to logic • Axiomatize “core theories” • WordNet links • Morphosemantic links • Purpose links • Experiments
Can solve with WN’s derivation links: (“attack”) attack_v3 aggression_n4 (←“violence”) derivation link “aggress”/“aggression” Morphosemantic Links • Often need to cross part-of-speech T: A council worker cleans up after Tuesday's violence in Budapest. H: There were attacks in Budapest on Tuesday.
(“pay”) pay_v1 payment_n1 (→ “transaction”) “pay”/“payment” (“pay”) pay_v1 payer_n1 (→ “person”) “pay”/“payer” Problem: The type of relation matters for derivatives! (Event? Agent?..) A pays B → The payment (event-noun) by A A is the payer (agent-noun) of B etc. Morphosemantic Links • But can go wrong! T: Paying was slow H1: The transaction was slow H2: *Theperson was slow [NOT entailed]
Morphosemantic Links • Task: Classify the 22,000 links in WordNet: • Semi-automatic process • Exploit taxonomy and morphology • 15 semantic types used • agent, undergoer, instrument, result, material, destination, location, result, by-means-of, event, uses, state, property, body-part, vehicle. Verb SynsetNoun SynsetRelationship hammer_v1 hammer_n1 instrument execute_v1 execution_n1 event (equal) sign_v2 signatory_n1 agent
isa(soldier01,soldier_n1), isa(…… object(kill01,soldier01) during(kill01,battle01) instrument(battle01,gun01) “soldier”(soldier01), “kill”(….. object(kill01,soldier01), “in”(kill01,battle01), modifier(battle01,gun01). “A soldier was killed in a gun battle” Initial Logic Final Logic Task: Recognizing Entailment • Experiment with WordNet, logical glosses, DIRT • Text interpretation to logic using Boeing’s NLP system • Entailment: T → H if: • T is subsumed by H (“cat eats mouse” → “animal was eaten”) • An elaboration of T using inference rules is subsumed by H • (“cat eats mouse” → “cat swallows mouse”) • No statistical similarity metrics
Successful Examples with the Glosses • Good example 14.H4 T: Britain puts curbs on immigrant labor from Bulgaria and Romania. H: Britain restricted workers from Bulgaria.
Successful Examples with the Glosses • Good example 14.H4 T: Britain puts curbs on immigrant labor from Bulgaria and Romania. H: Britain restricted workers from Bulgaria. WN: limit_v1:"restrict“: place limits on. T: Britain puts curbs on immigrant labor from Bulgaria and Romania. H: Britain placed limits on workers from Bulgaria. → ENTAILED (correct)
Successful Examples with the Glosses • Another (somewhat) good example 56.H3 T: The administration managed to track down the perpetrators. H: The perpetrators were being chased by the administration.
Successful Examples with the Glosses • Another (somewhat) good example 56.H3 T: The administration managed to track down the perpetrators. H: The perpetrators were being chased by the administration. WN: hunt_v1 “hunt” “track down”: pursue for food or sport T: The administration managed to pursue the perpetrators [for food or sport!]. H: The perpetrators were being chased by the administration. → ENTAILED (correct)
Unsuccessful examples with the glosses • More common: Being “tantalizingly close” 16.H3 T: Satomi Mitarai bled to death. H: His blood flowed out of his body.
Unsuccessful examples with the glosses • More common: Being “tantalizingly close” 16.H3 T: Satomi Mitarai bled to death. H: His blood flowed out of his body. WordNet: bleed_v1: "shed blood", "bleed", "hemorrhage": lose blood from one's body So close! Need to also know: “lose liquid from container” → “liquid flows out of container” usually
Unsuccessful examples with the glosses • More common: Being “tantalizingly close” 20.H2 T: The National Philharmonic orchestra draws large crowds. H: Large crowds were drawn to listen to the orchestra.
Unsuccessful examples with the glosses • More common: Being “tantalizingly close” 20.H2 T: The National Philharmonic orchestra draws large crowds. H: Large crowds were drawn to listen to the orchestra. WordNet: WN: orchestra = collection of musicians WN: musician: plays musical instrument WN: music = sound produced by musical instruments WN: listen = hear = perceive sound So close!
But need an agent (X verb Y -> X is agent-noun of Y) Got: result-noun (“invention” is result of “invent”) So no entailment (correct!) Success with Morphosemantic Links • Good example 66.H100 T: The Zoopraxiscope was invented by Mulbridge. H*: Mulbridge was the invention of the Zoopraxiscope. [NOT entailed] WordNet too permissive!: (“invent”) invent_v1 invention_n1 (“invention”) derivation link “invent”/ “invention”
Successful Examples with DIRT 54.H1 • Good example T: The president visited Iraq in September. H: The president traveled to Iraq. DIRT: IF Y is visited by X THEN X flocks to Y WordNet: "flock" is a type of "travel" Entailed [correct]
Unsuccessful Examples with DIRT • Bad rule 55.H100 T: The US troops stayed in Iraq although the war was over. H*: The US troops left Iraq when the war was over. [NOT entailed] DIRT: IF Y stays in X THEN Y leaves X Entailed [incorrect]
“Straight- Forward” Overall Results • Note: Eschewing statistics! • BPI test suite (61%):
Useful Overall Results • Note: Eschewing statistics! • BPI test suite (61%):
Overall Results • Note: Eschewing statistics! • BPI test suite (61%): Occasionally useful
Overall Results • Note: Eschewing statistics! • BPI test suite (61%): Often useful but unreliable • RTE3: 55%
Summary • “Understanding” • Constructing a coherent model of the scene being described • Much is implicit in text → Need lots of world knowledge • Augmenting WordNet • Made some steps forward: • More connectivity • Logicalized glosses • But still need a lot more knowledge!