1 / 27

Modeling Long-Distance Dependencies in Double R July 2008

Modeling Long-Distance Dependencies in Double R July 2008. Jerry Ball Human Effectiveness Directorate Air Force Research Laboratory. Double R Model.

azura
Download Presentation

Modeling Long-Distance Dependencies in Double R July 2008

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Modeling Long-Distance Dependencies in Double R July 2008 Jerry Ball Human Effectiveness Directorate Air Force Research Laboratory

  2. Double R Model • Goal: Model the basic grammatical patterns of English to support development of cognitively plausible and functional language comprehension systems • Declaratives – “The man hit the ball” • Questions • Yes-No Questions – “Did the man hit the ball?” • Wh Questions – “Where did the man hit the ball?” • Imperatives – “Hit the ball!” • Relative Clauses – “The ball that the man hit” • Wh Clauses – “I know where the man hit the ball” • Passive constructions – “The ball was hit”

  3. Empirical Evidence • Basic grammatical patterns have been most extensively studied in generative grammar • The focus in generative grammar has been on studying the syntactic form of linguistic expressions in isolation frommeaningandprocessing • The “Simpler Syntax”of Culicover and Jackendoff (2005) is redressing the consideration of meaning and simplifying syntax as a side effect • O’Grady’s “Syntactic Carpentry” (2005) integrates processing as well (see also Hawkins, 2004) • Reference grammars (Huddleston & Pullum, 2002; Quirk et al., 1985) provide a wealth of examples which integrate form, function and meaning

  4. Long-Distance Dependencies • Long-distance dependencies are the sin qua non of modern linguistic theorizing • An important motivation for Chomsky’s transformational grammar – deep structures with arguments in place are mapped to surface structures with arguments “moved” by various transformations • Introduction of traces supported the collapsing of deep and surface structure – traces mark the original location • Construction specific transformations were generalized to Move asubject to universal, parameterized constraints • Many basic grammatical constructions involve long-distance dependencies • Wh questions, relative clauses, passive constructions… • Require retention of grammatical information for extended stretches of input

  5. Long-Distance Dependencies • Binding of pronouns and anaphors: • Anaphors (“himself”) vs. pronouns (“him”) • Johni kicked himselfi ( i = i) (Principle A of GB Theory) • Johnikicked himj ( i not = j) (Principle B of GB Theory) • Proper binding often requires use of semantic information (but considered syntactic in generative grammar) • Johni and Maryjwere talking. Shejtold himi… (gender) • Johni is reading a bookj. Itjis about… (animacy) • Johni is reading the comicsj. Theyjare… (number)

  6. Long-Distance Dependencies • Verb Control • Object Control: “HeipersuadedmejPROj to go” • PROj is an “implicit” pronoun (a trace without movement) • Subject Control: “HeipromisedmejPROito go” • Raising Verbs • “Heiseemsti to like me” • ti is a trace of a “raised” argument

  7. Long-Distance Dependencies • Passive Constructions • “The balliwas kickedti by the man” • The object is “raised” out of its normal position and the subject is pushed into an oblique complement position “by the man” • Wh Questions • “Whoi did JohnjdecidePROj to see ti” • Relative Clauses • “The ballithat the man kicked ti”

  8. Modeling Long-Distance Dependencies • An ontology of DM chunk types supports the grammatical distinctions • Productions match buffer elements at the appropriate level of the ontology given the function of the production, e.g. • Production matches pronoun “he…”  project nominal and put in subject buffer • Production matches predicate specifier (e.g. “…is…”)  project a declarative clause • Production matches declarative clause and a nominal in subject buffer (e.g. “he is…”)  integrate the nominal as the subject of the clause • Production matches transitive verb (e.g. “hitting”) functioning as clausal head (e.g. “he is hitting…”) and a nominal (e.g. “…the ball”)  integrate the nominal as the object of the verb

  9. Ontology of Situation Referring Expressions • Decl-sit-refer-expr • Yes-no-quest-sit-refer-expr • “Is he going?” • Wh-quest-sit-refer-expr • “Where did he go?” • Imp-sit-refer-expr • “Don’t go!” • Wh-sit-refer-expr • “I know where he went” • Rel-sit-refer-expr • “The book that you like” Note: Situation Referring Expression corresponds to Clause in other approaches What are the grammatical cues that trigger recognition of an expression type? These cues need to be accessible!

  10. Slots in Referring Expressions • Bind-indx (all referring expression types) • Identifier for referring expression • Parent (all chunk types) • Links child to parentchunk • Used to avoid multiply integrating chunk into other chunks • Token (all chunk types) • Distinguishes types from tokens (and type-tokens) • Grammatically relevant semantic info • Animate (all object referring expression types) • Gender (all animate referring expression types) • Number (all object referring expression types) • Person (all object referring expression types)

  11. Recognizing Wh-Quest and Wh-Situation Referring Expressions …where he went Where did he…? (p cog-process-obj-refer-expr--> project-wh-quest-sit-refer-expr =goal> isa process-obj-refer-expr =wh-focus> isa wh-refer-expr ;; “where” =most-recent-child-sre-head> isa operator-pred-spec ;; “did” =retrieval-2> isa obj-refer-expr ;; “he” =subject> isa nothing =context> isa context - sit-context "wh-quest-sit-refer-expr“ ==> project wh-quest-sit-refer-expr (p cog-process-pred-type project-wh-sit-refer-expr =goal> isa process-pred-type =wh-focus> isa wh-refer-expr ;; “where” =subject> isa refer-expr ;; “he” =retrieval-2> isa pred-type ;; “went” =context> isa context - sit-context "wh-sit-refer-expr" - sit-context "wh-quest-sit-refer-expr" ==> project wh-sit-refer-expr 1 1 2 2 3 3 4 Note: the more grammatical cues, the greater the likelihood of being correct! “Who kicked…?” “Where the heck is...?” “Why is there…?

  12. Modeling Long-Distance Dependencies • Model needs simultaneous access to multiple grammatical elements • Serial retrieval from DM is not a viable option • Buffers support simultaneous access – buffers on left-hand side of productionconstitute focus of attention – limited to ~4(Cowan, 2000) besides goal and context buffers • Can’t predict in advance of production selection which grammatical elements will be needed • Buffers and productions are functionally motivated – they are needed in the processing of various constructions • A model with fewer buffers (and productions) that handles a similar set of phenomena might be a better model, but a model with fewer buffers that handles fewer phenomena is not comparable (Ball, in preparation)

  13. Double R Buffers – Single Chunk • Subject – stores the subject • Wh-focus – stores the fronted wh expression • Rel-focus – stores the relative clause marker • Context – stores contextual information • Construct – buffer for constructing DM chunks • Dual path processing –construct chunk vs. retrieve chunk • Retrieval-2 – buffer for storing retrieved or constructed DM chunks • Retrieval buffer only used temporarily, retrieved chunk is copied into retrieval-2 for subsequent processing • Most-recent-loc-refer-expr –just the most recent • Supports locative fronting “On the table is the book”

  14. Most-recent-child- obj-refer-expr Most-recent-parent- obj-refer-expr Most-recent-grandparent- obj-refer-expr Most-recent-child- obj-refer-expr-head Most-recent-parent- obj-refer-expr-head Most-recent-grandparent- obj-refer-expr-head Double R Buffers – Multiple Chunk Obj-Refer-Expr buffers Obj-Refer-Expr-Head buffers • St-wm-1 • St-wm-2 • St-wm-3 • St-wm-4 Note: object referring expression corresponds to nominal in other approaches Four generic Short-Term Working Memory buffers

  15. Most-recent-child- sit-refer-expr Most-recent-parent- sit-refer-expr Most-recent-grandparent- sit-refer-expr Most-recent-child- sit-refer-expr-head Most-recent-parent- sit-refer-expr-head Most-recent-grandparent- sit-refer-expr-head Double R Buffers – Multiple Chunk Sit-Refer-Expr buffers Sit-Refer-Expr-Head buffers Note 1: with the introduction of obj-refer-expr and sit-refer-expr specific buffers, the short-term working memory buffers are infrequently used (primarily for conjunctions and adverbs) Note 2: child, parent and grandparent buffers are all directly accessible, whereas only st-wm-1 is directly accessible

  16. Long-Distance Dependencies I want to go Note: entire representation is not accessible at once! Infinitive sit-refer-expr has implied subj with trace bound to matrix subj Traces only occur in argument positions! Combination of“bind-indx” and “trace” needed to indicate long-distance dependency

  17. Long-Distance Dependencies He promised me to go Alternative view: antecedent & trace both bind to same object in situation model Subject control (verb): matrix clause subject binds to subject of infinitive situation complement – subject must be accessible

  18. Long-Distance Dependencies He persuaded me to go Objectcontrol (verb): matrix clause (indirect) object binds to subject of infinitive situation complement – object must be accessible

  19. Long-Distance Dependencies Who did he kick the ball to? Object of preposition is bound to fronted who-obj-refer-expr – wh-focus must be accessible

  20. Long-Distance Dependencies The man that I gave the book I-Obj Trace to Obj-Refer-Expr with animate or humanhead Rel-focus co-indexed with Obj-Refer-Expr rel-focus and subject must be accessible (rel-focus is optional)

  21. Long-Distance Dependencies The book that I gave the man Obj Trace to Obj-Refer-Expr with inanimate head Rel-focus co-indexed with Obj-Refer-Expr rel-focus and subjectmust be accessible (rel-focus is optional)

  22. Architectural Constraints • No hard architectural limit on the number of buffers • Buffers provide the context for production selection and execution • Highly context sensitive • Productions limited to accessing ~4 buffers on left-hand side (beside goal and context buffers) • Focus of attention (Cowan, 2000) • “Conscious activity corresponds to the manipulation of the contents of these buffers by production rules” (Anderson, 2007) • Can humans learn to buffer useful information? • Fronted Wh-expression buffer very useful in English, but not needed in in situ languages like Chinese

  23. Processing Constraints • A “mildly” deterministic, serial processing mechanism (selection and integration) operating over a parallel, probabilistic substrate (activation) • Interactive and non-autonomous processing (no distinctly syntactic representations exist) • Incremental processing with immediate determination of meaning – word by word • No algorithmic backtracking or lookahead – a mechanism of context accommodation (Ball et al. 2007) used instead • Forward chaining only • Declarative and explicit linguistic representations generated via implicit execution of productions • Operates in real-time on Marr’s algorithmic level (serial and parallel processing are relevant) • No slow down with length of input

  24. Summary • Additions to model are • motivated by functional considerations • driven by empirical evidence • constrained by well-established cognitive constraints on language processing • Goal is a large-scale, functional language comprehension system implemented in the ACT-R cognitive architecture • Model currently handles a fairly wide-range of grammatical constructions including numerous forms of long-distance dependency

  25. Questions?

  26. References Anderson, J. (2007). How Can the Human Mind Occur in the Physical Universe. Oxford: Oxford University Press. Ball, J. (in preparation). A Naturalistic Functional Approach to Modeling Language Comprehension. Ball, J., Heiberg, A. & Silber, R. (2007). Toward a Large-Scale Model of Language Comprehension in ACT-R 6. Proceedings of the 8th International Conference on Cognitive Modeling. Cowan, N. (2000). The magical number 4 in short-term memory: A reconsideration of mental storage capacity. Behavioral and Brain Sciences, 24, 87-185. Culicover, P. & Jackendoff, R. (2005). SimplerSyntax. Oxford: Oxford University Press. Hawkins, J. (2004). Efficiency and Complexity in Grammars. Oxford: Oxford University Press. Huddleston, R. & Pullum G.(2002). The Cambridge Grammar of the English Language. NY: Cambridge Unversity Press. O’Grady, William (2005). Syntactic Carpentry, an Emergentist Approach to Syntax. Mahway, NJ:LEA. Quirk, R., Greenbaum, S., Leech, G. & Svartvik, J. (1985). A Comprehensive Grammar of the English Language. Essex, UK: Pearson Education Limited.

  27. Long-Distance Dependencies The ball by the table was kickedby the man passive cue (be + V-ed or V-en) Subject co-indexed with Object subject must be accessible

More Related