580 likes | 592 Vues
A best-fit approach simplifies grammar by exploiting language understanding. Analyzing argument omission in Mandarin Chinese construction grammar framework. Model for productive argument omission.
E N D
A Best-Fit Approach for Productive Analysis of Omitted Arguments Eva Mok & John Bryant University of California, Berkeley International Computer Science Institute
Simplify grammar by exploiting the language understanding process Omission of arguments in Mandarin Chinese Construction grammar framework Model of language understanding Our best-fit approach
Productive Argument Omission (in Mandarin) 1 • Mother (I) give you this (a toy). 2 • You give auntie[the peach]. 3 • Oh (go on)! You give[auntie] [that]. 4 • [I]give[you] [some peach]. CHILDES Beijing Corpus (Tardiff, 1993; Tardiff, 1996)
Arguments are omitted with different probabilities All arguments omitted: 30.6% No arguments omitted: 6.1%
Construction grammar approach Kay & Fillmore 1999; Goldberg 1995 Grammaticality: form and function Basic unit of analysis: construction, i.e. a pairing of form and meaning constraints Not purely lexically compositional Implies early use of semantics in processing Embodied Construction Grammar (ECG) (Bergen & Chang, 2005)
If the analysis process is smart, then... The grammar needs only state one construction Omission of constituents is flexibly allowed The analysis process figures out what was omitted
Best-fit analysis process takes burden off the grammar representation Analyzer: Discourse & Situational Context Constructions Utterance incremental, competition-based, psycholinguistically plausible Semantic Specification: image schemas, frames, action schemas Simulation
Competition-based analyzer finds the best analysis An analysis is made up of: A constructional tree A set of resolutions A semantic specification The best fit has the highest combined score
Combined score that determines best-fit Syntactic Fit: Constituency relations Combine with preferences on non-local elements Conditioned on syntactic context Antecedent Fit: Ability to find referents in the context Conditioned on syntactic information, feature agreement Semantic Fit: Semantic bindings for frame roles Frame roles’ fillers are scored
Analyzing ni3 gei3 yi2 (You give auntie) Syntactic Fit: P(Theme omitted | ditransitive cxn) = 0.65 P(Recipient omitted | ditransitive cxn) = 0.42 Two of the competing analyses: (1-0.78)*(1-0.42)*0.65 = 0.08 (1-0.78)*(1-0.65)*0.42 = 0.03
Using frame and lexical information to restrict type of reference
Can the omitted argument be recovered from context? Antecedent Fit: • Discourse & Situational Context • child mother • peach auntie • table ?
How good of a theme is a peach? How about an aunt? • Semantic Fit:
The argument omission patterns shown earlier can be covered with just ONE construction Each cxn is annotated with probabilities of omission Language-specific default probability can be set P(omitted|cxn): 0.78 0.42 0.65
Research goal Language Data • A computationally-precise modeling framework for learning early constructions Learning New Construction Linguistic Knowledge
Frequent argument omission in pro-drop languages • Mandarin example: ni3 gei3 yi2 (“you give auntie”) • Even in English, there are often no spoken antecedents to pronouns in conversations Learner must integrate cues from intentions, gestures, prior discourse, etc
A short dialogue • bie2 mo3 wai4+tou2 a: #1_3 ! (別抹外頭啊) • NEG-IMP apply forehead • Don’t apply [lotion to your] forehead • mo3 wai4+tou2 ke3 jiu4 bu4 hao3+kan4 le a: . (抹外頭可就不好看了啊) • apply forehead LINKER LINKER NEG good looking CRS SFP • [If you] apply [lotion to your] forehead then [you will] not be pretty … • ze ya a: # bie2 gei3 ma1+ma wang3 lian3 shang4 mo:3 e: ! (嘖呀啊 # 別給媽媽往臉上抹呃) • INTERJ # NEG-IMP BEN mother CV-DIR face on apply • INTERJ # Don’t apply [the lotion] on [your mom’s] face (for mom) • [- low pitch motherese] ma1+ma bu4 mo:3 you:2 . (媽媽不抹油) • mother NEG apply lotion • Mom doesn’t apply (use) lotion
Goals, refined • Demonstrate learning given • embodied meaning representation • structured representation of context • Based on • Usage-based learning • Domain-general statistical learning mechanism • Generalization / linguistic category formation
Towards a precise computational model • Modeling early grammar learning • Context model & Simulation • Data annotation • Finding the best analysis for learning • Hypothesizing and reorganizing constructions • Pilot results
Embodied Construction Grammar construction yi2-N subcaseof Morpheme form constraints self.f.orth <-- "yi2" meaning : @Aunt evokes RD as rd constraints self.m <--> rd.referent self.m <--> rd.ontological_category
“you” specifies discourse role construction ni3-N subcase of Morpheme form constraints self.f.orth <-- "ni3" meaning : @Human evokes RD as rd constraints self.m <--> rd.referent self.m <--> rd.ontological_category rd.discourse_participant_role <-- @Addressee rd.set_size <-- @Singleton
The meaning of “give” is a schema with roles schema Transfer subcaseof Action roles giver : @Entity recipient : @Entity theme : @Entity constraints giver <--> protagonist construction gei3-V2 subcaseof Morpheme form constraints self.f.orth <-- "gei3" meaning : Give schema Give subcaseof Transfer constraints inherent_aspect <-- @Inherent_Achievement giver <-- @Animate recipient <-- @Animate theme <-- @Manipulable_Inanimate_Object
Finally, you-give-aunt links up the roles construction ni3-gei3-yi2 subcaseof Finite_Clause constructional constituents n : ni3-N g : gei3-V2 y : yi2-N form constraints n.f meets g.f g.f meets y.f meaning : Give constraints self.m <--> g.m self.m.giver <--> n.m self.m.recipient <--> y.m
Discourse & Situational Context The learning loop: Hypothesize & Reorganize World Knowledge Utterance Linguistic Knowledge reorganize Analysis reinforcement Context Fitting hypothesize PartialSemSpec
ni3 Addressee gei3 Give yi2 Aunt If the learner has a ditransitive cxn Form Meaning Context XIXI MOT addressee speaker giver meets Discourse Segment meets recipient INV theme omitted Peach
ni3 Addressee gei3 Give yi2 Aunt Context fitting recovers more relations Form Meaning Context XIXI MOT addressee speaker giver meets giver Discourse Segment Give meets recipient recipient attentional-focus INV theme theme omitted Peach
MOT addressee speaker ni3 Addressee meets giver giver Discourse Segment gei3 Give meets recipient recipient attentional-focus theme yi2 Aunt Peach But the learner does not yet have phrasal cxns Form Meaning Context XIXI Give INV
ni3 Addressee meets giver gei3 Give meets recipient yi2 Aunt Context bootstraps learning Form Meaning construction ni3-gei3-yi2 subcase of Finite_Clause constructional constituents n : ni3 g : gei3 y : yi2 form constraints n.f meets g.f g.f meets y.f meaning : Give constraints self.m <--> g.m self.m.giver <--> n.m self.m.recipient <--> y.m
A model of context is key to learning • The context model makes it possible for the learning model to: • learn new constructions using contextually available information • learn argument-structure constructions in pro-drop languages
Events + Utterances Understanding an utterance in context Transcripts Schemas + Constructions Context Model Recency Model Analysis + Resolution Context Fitting Simulation Semantic Specification
Context model: Events + Utterances Setting participants, entities, & relations Start Event Event DS Sub-Event Sub-Event
Entities and Relations are instantiated Setting CHI, MOT (incl. body parts) livingroom (incl. ground, ceiling, chair, etc), lotion ds04 admonishing05 speaker = MOT addressee = CHI forcefulness = normal caused_motion01 forceful_motion motion Start apply02 applier = CHI substance = lotion surface = face(CHI) translational_motion03 mover = lotion spg = SPG
The context model is updated dynamically Events • Extended transcript annotation: speech acts & events • Simulator inserts events into context model & updates it with the effects • Some relations persists over time; some don’t. Context Model Recency Model Simulation
Competition-based analyzer finds the best analysis A-GIVE-B-X subj obj2 v obj1 Ref-Exp Give Ref-Exp Ref-Exp @Book @Man Give-Action @Woman Bill Mary book01 recipient giver theme • An analysis is made up of: • A constructional tree • A semantic specification • A set of resolutions Bill gave Mary the book
Combined score that determines best-fit • Syntactic Fit: • Constituency relations • Combine with preferences on non-local elements • Conditioned on syntactic context • Antecedent Fit: • Ability to find referents in the context • Conditioned on syntactic information, feature agreement • Semantic Fit: • Semantic bindings for frame roles • Frame roles’ fillers are scored
ni3 Addressee gei3 Give yi2 Aunt Context Fitting goes beyond resolution Form Meaning Context XIXI MOT addressee speaker giver meets giver Discourse Segment Give meets recipient recipient attentional-focus INV theme theme omitted Peach
Context Fitting, a.k.a. intention reading • Context Fitting takes resolution a step further • considers entire context model, ranked by recency • considers relations amongst entities • heuristically fits from top down, e.g. • discourse-related entities • complex processes • simple processes • other structured and unstructured entities • more heuristics for future events (e.g. in cases of commands or suggestions)
Adult grammar size • ~615 constructions total • ~100 abstract cxns (26 to capture lexical variants) • ~70 phrasal/clausal cxns • ~440 lexical cxns (~260 open class) • ~195 schemas (~120 open class, ~75 closed class)
Starter learner grammar size • No grammatical categories (except interjections) • Lexical items only • ~440 lexical constructions • ~260 open class: schema / ontology meanings • ~40 closed class: pronouns, negation markers, etc • ~60 function words: no meanings • ~195 schemas (~120 open class, ~75 closed class)
The process hierarchy defined in schemas Process Proto_Transitive State_Change State Action Complex_Process Intransitive_State Two_Participant_State Serial_Processes Mental_State Concurrent_Processes Cause_Effect Joint_Motion Caused_Motion
The process hierarchy defined in schemas Action Intransitive_Action Motion Translational_Motion Expression Self_Motion Translational_Self_Motion Force_Application Forceful_Motion Continuous_Force_Application Agentive_Impact
The process hierarchy defined in schemas Action Cause_Change Communication Obtainment Transfer Ingestion Perception Other_Transitive_Action
Events + Utterances Understanding an utterance in context Transcripts reorganize Schemas + Constructions Context Model Recency Model Analysis + Resolution Context Fitting Simulation reinforcement hypothesize Semantic Specification
Hypothesize & Reorganize • Hypothesize: • utterance-driven; • relies on the analysis (SemSpec & context) • operations: compose • Reorganize: • grammar-driven; • can be triggered by usage (to be determined) • operations: generalize
Composing new constructions ni3 Addressee gei3 Give • Compose operation: If roles from different constructions point to the same context element, propose a new construction and set up a meaning binding. Context MOT XIXI Peach giver recipient theme INV
Creating pivot constructions ni3 ni3 Addressee Addressee gei3 yi2 wo3 gei3 Give @Human @Aunt Give • Pivot generalization: Given a phrasal cxn, look for another cxn that shares 1+ constituents. Line up roles and bindings. Create new cxn category for the slot. meets meets giver giver meets meets recipient recipient
Resulting constructions construction ni3-gei3-cat01 constituents ni3, gei3, cat01 meaning : Give constraints self.m.recipient <--> g.m general construction cat01 subcase of Morpheme meaning: @Human construction wo3 subcase of cat01 meaning: @Human construction yi2 subcase of cat01 meaning: @Aunt
Pilot Results: Sample constructions learned • Composed: • Pivot Cxns:
Challenge #1: Non-compositional meaning you a cake Addressee @Cake bake Bake Context • Non-compositional meaning: Search for additional meaning schemas (in context or in general) that relate the meanings of the individual constructions Bake-Event Give-Event baker baked MOT CHI Cake