SlideShare uma empresa Scribd logo
1 de 107
An Interpretation-Driven Model of Syntax Richard Caneba  canebr@rpi.edu RPI Cognitive Science Department Human-Level Intelligence Laboratory 5/2/2011
Introduction We start with a goal: Develop a system that can understand natural language. 5/2/2011
Introduction We start with a goal: Develop a system that can understand natural language. (Roughly) three sub-goals: Syntactic Parsing Semantic Representation Pragmatics/Discourse 5/2/2011
Introduction We start with a goal: Develop a system that can understand natural language. (Roughly) three sub-goals: Syntactic Parsing Semantic Representation Pragmatics/Discourse Stage 1: Syntax Why is syntax important for natural language understanding? 5/2/2011
Introduction: Syntax “The dog bit the man.” 5/2/2011
Introduction: Syntax “The dog bit the man.” 5/2/2011
Introduction: Syntax “The man bit the dog.” 5/2/2011
Introduction: Syntax “The man bit the dog.” 5/2/2011
Introduction: Syntax “I hit the man with my car.” 5/2/2011
Introduction: Syntax 5/2/2011 “[I] hit [the man] [with my car].”
Introduction: Syntax “[I] hit [the man] [with my car].” 5/2/2011 Option 1: “with my car” modifies “hit””I hit the man while driving my car.”
Introduction: Syntax “[I] hit [the man] [with my car].” 5/2/2011 Option 2: “with my car” modifies “the man””I hit the man who had my car.”
Introduction: Syntax Syntactic interpretation yields very distinct semantic interpretations. 5/2/2011
Introduction: Syntax Syntactic interpretation yields very distinct semantic interpretations. Helps identify the role words play in an utterance. 5/2/2011
Introduction: Syntax Thus, syntax plays a fundamental role in natural language understanding. 5/2/2011
Syntax Current grammar formalisms have number of shortcomings 5/2/2011
Syntax Current grammar formalisms have number of shortcomings From perspective of trying to develop a system to understand natural language 5/2/2011
Syntax Current grammar formalisms have number of shortcomings From perspective of trying to develop a system to understand natural language Focus on generative grammar (i.e. Chomskyan, HPSG [11], [13]) 5/2/2011
Syntax: Theory & Implementation Two classes of shortcomings: Theoretical: shortcoming in the way a theory of language represents the users mental knowledge of that language. Implementation: shortcoming in the way a theory implies or represents language processing in terms of computability and/or cognitive realism. 5/2/2011
Syntax: Theory & Implementation We will show that these two classes are so closely related, does not make sense to make a strong distinction 5/2/2011
Theoretical Shortcomings Shortcoming in the way a theory of language represents the users mental knowledge of that language 5/2/2011
Theoretical Shortcomings Interpretability 5/2/2011
Theoretical Shortcomings i.e. “Fido bit dog.” 5/2/2011
Theoretical Shortcomings i.e. “Fido bit dog.” cf. “Fido bit the dog.”  5/2/2011
Why Syntax? Why syntax at all, if we can still interpret ungrammaticality? 5/2/2011
Why Syntax? “Fido bit dog.” cf. “Dog bit Fido.”  5/2/2011 Fido Fido
Theoretical Shortcomings Phrasal Nesting 5/2/2011
Theoretical Shortcomings Phrasal Nesting example: “the man in boston with the hat is here.” 5/2/2011
5/2/2011 [S] [NP] [VP] [NP] is [PP] [PP] here man [NP] [DP] in the [PP] boston with [NP] hat [DP] “the man in boston with the hat is here.” the
5/2/2011 [S] [NP] [VP] [NP] is [PP] [PP] here man [NP] [DP] in the [PP] boston with [NP] hat [DP] “the man in boston with the hat is here.” the
5/2/2011 [S] [VP] [NP] is [PP] man [DP] [PP] [PP] here with [NP] [NP] in the hat [DP] boston the “the man in boston with the hat is here.”
5/2/2011 [S] [VP] [NP] is [PP] man [DP] [PP] [PP] here with [NP] [NP] in the hat [DP] boston the “the man in boston with the hat is here.”
Implementation Shortcomings Shortcoming in the way a theory implies or represents language processing in terms of computability and/or cognitive realism. 5/2/2011
Implementation Shortcomings Perceptual Ordering 5/2/2011
Implementation Shortcomings john 5/2/2011
Implementation Shortcomings john saw 5/2/2011
Implementation Shortcomings john saw mary 5/2/2011
Implementation Shortcomings [VP] [NP] 5/2/2011 john saw mary
Implementation Shortcomings [S] [VP] [NP] [NP] 5/2/2011 john saw mary
Implementation Shortcomings Excessive Structure 5/2/2011
Implementation Shortcomings 5/2/2011 the tall strong angry man
Implementation Shortcomings 5/2/2011 [NP] the tall strong angry man
Implementation Shortcomings 5/2/2011 [NP] [NP] the tall strong angry man
Implementation Shortcomings 5/2/2011 [NP] [NP] [NP] the tall strong angry man
Implementation Shortcomings 5/2/2011 [NP] [NP] [NP] [NP] the tall strong angry man
Implementation Shortcomings 5/2/2011 [NP] [NP] [NP] [NP] the tall strong angry man
Implementation Shortcomings 5/2/2011 [NP] [NP] [NP] [NP] [PP] the tall strong angry man in boston
Implementation Shortcomings 5/2/2011 [NP] [NP] [NP] [NP] [PP] the tall strong angry man in boston
Implementation Shortcomings 5/2/2011 [NP] [NP] [NP] [NP] [PP] the tall strong angry man in boston
Implementation Shortcomings 5/2/2011 [NP] [NP] [NP] [NP] [PP] the tall strong angry man in boston
Implementation Shortcomings 5/2/2011 [NP] [NP] [NP] [NP] [PP] the tall strong angry man in boston
Implementation Shortcomings 5/2/2011 [NP] [NP] [NP] [NP] [PP] the tall strong angry man in boston
Syntax: Statistical Approach Most currently successful parsing algorithms rely heavily on statistics. However, inferences that require notion of semantics difficult 5/2/2011
Syntax: Statistical Approach i.e. “The couple walked in the park.  He held her hand.” 5/2/2011
Syntax: Statistical Approach i.e. “The couple walked in the park.  He held her hand.” Statistics does not give anaphoric bind. 5/2/2011
Chomskyan Approach From [6], consider  “Joe has put those raw potatoes in the pot.” 5/2/2011
Chomskyan Approach 5/2/2011
Chomskyan Approach A simpler analysis is possible: 5/2/2011
Chomskyan Approach 5/2/2011
Chomskyan Approach Illustrates: way a theory is represented has very real computational concerns. 5/2/2011
Other Notable Work Other Cognitive Architecture based Parsers [8] R. Lewis et.al. developed a parser based on "immediate reasoning" in Soar [9] R. Lewis et.al. developed an activation based parser model in ACT-R [1][2][3] J.T. Ball et.al. developed a parser based on Double R Grammar Model, for "synthetic teammate" development in ACT-R 5/2/2011
Other Notable Work However, each of these theories suffers from the shortcomings we’ve already seen. 5/2/2011
Other Notable Work However, each of these theories suffers from the shortcomings we’ve already seen. Both parsers designed by Lewis rely on a CFG formalism Bell’s ACT-R parser does not deeply integrated with reasoning 5/2/2011
Other Notable Work However, each of these theories suffers from the shortcomings we’ve already seen. Both parsers designed by Lewis rely on a CFG formalism Bell’s ACT-R parser does not deeply integrated with reasoning These approaches are not well integrated with reasoning overall 5/2/2011
Motivating Principles To address these shortcomings from an interpretative perspective, four principles are motivated: 5/2/2011
Motivating Principles To address these shortcomings from an interpretative perspective, four principles are motivated:a) The existence of satellite structuresb) Feature structure unificationc) Feature structure aggregationd) Incrementality 5/2/2011
Satellite Structures When we hear words, infer existence of other structures related to that word 5/2/2011
Satellite Structures When we hear words, infer existence of other structures related to that word i.e. “bit” 5/2/2011 bit bit
Satellite Structures When we hear words, infer existence of other structures related to that word i.e. “bit” 5/2/2011 [NP] [NP] Obj Subj bit bit
Feature Structure Unification With the existence of satellite structures, we can unify observed structures together 5/2/2011
Feature Structure Unification With the existence of satellite structures, we can unify observed structures together 5/2/2011 [NP] [NP] Obj Subj john bit bit fido
Feature Structure Unification With the existence of satellite structures, we can unify observed structures together 5/2/2011 [NP] [NP] Obj Subj john bit bit fido
Feature Structure Aggregation Sequential unification multiple feature structures yields “aggregating” feature structure 5/2/2011
Feature Structure Aggregation i.e. “the big dog” 5/2/2011
Feature Structure Aggregation i.e. “the big dog” 5/2/2011 big the dog
Feature Structure Aggregation i.e. “the big dog” 5/2/2011 [NP] [NP] [NP] big the dog
Feature Structure Aggregation i.e. “the big dog” 5/2/2011 [NP] [NP] [NP] Unify big the dog
Feature Structure Aggregation i.e. “the big dog” 5/2/2011 [NP] [NP] [NP] Unify Unify big the dog
Feature Structure Aggregation i.e. “the big dog” 5/2/2011 Unify [NP] [NP] [NP] big the dog
Incrementality Human sentence comprehension is left-to-right, and incremental Incrementally generate an interpretation 5/2/2011
Using these principles, attempt to address the aforementioned shortcomingsin a model 5/2/2011 Architectural Implementation
Architectural Implementation Parser prototype implemented in the Polyscheme Cognitive Architecture. 5/2/2011
Architectural Implementation Parser prototype implemented in the Polyscheme Cognitive Architecture. Use the model to process and interpret natural language input. 5/2/2011
Architectural Implementation Parsing driven by pair-wise pattern matching 5/2/2011
Architectural Implementation 5/2/2011 Parsing driven by pair-wise pattern matching [XP] * [NP] [XP] * [NP] [DetP] [NP] [DetP] [D] [N] [D] [N] *Underspecified Feature Structure
Architectural Implementation 5/2/2011 Parsing driven by pair-wise pattern matching [XP] * [NP] [XP] * [NP] [DetP] [NP] [DetP] [D] [N] [D] [N] *Underspecified Feature Structure
Architectural Implementation 5/2/2011 [XP] [XP] [NP] [NP] [DetP] [NP] [DetP] [1] man the man the DetP [1]
Architectural Implementation 5/2/2011 IsA(?the, WordUtteranceEvent) IsA(?the, Determiner) PartOf(?d, ?dp) IsA(?dp, PhraseUtteranceEvent) IsA(?dp, Determiner) PartOf(?dp, ?np1) IsA(?np1, PhraseUtteranceEvent) IsA(?np1, Noun) Meets(?the, ?man) IsA(?man, WordUtteranceEvent) IsA(?man, CommonNoun) Specifier(?man, ?detSpr) IsA(?detSpr, Determiner) PartOf(?man, ?np2) IsA(?np2, PhraseUtteranceEvent) IsA(?np2, Noun)  Same(?dp, ?detSpr) Same(?np1, ?np2) [XP] [XP] [NP] [NP] [DetP] [NP] [DetP] [1] man the man the DetP [1]
5/2/2011 [DetP][N]  NP[DetP, N] [AdjP][N]  NP[AdjP, N] [XP] [XP] [XP] [XP] [XP] [NP] [NP] [NP] [DetP] [NP] [AdjP] [NP] [DetP] [AdjP] [D] [N] [D] [N] [Adj] [N] [Adj] [N] [DetP][AdjP]  NP[DetP, AdjP] [XP] [XP] [NP] [NP] [DetP] [AdjP] [AdjP] [DetP] [D] [Adj] [D] [Adj]
5/2/2011 [P][NP]  PP[P, NP] [V][NP]  VP[V, NP] [XP] [XP] [XP] [XP] [XP] [VP] [XP] [PP] [PP] [NP] [NP] [NP] [VP] [NP] [P] [X] [P] [X] [X] [X] [V] [V] [NP][V]  VP[NP, V] [XP] [XP] [XP] [VP] [NP] [VP] [NP] [N] [V] [N] [V]
"the tall man" 5/2/2011 Architectural Implementation
Architectural Implementation "the tall man" 5/2/2011 [NP] [DetP] the
Architectural Implementation "the tall man" 5/2/2011 [NP] [NP] [AdjP] [DetP] tall the
Architectural Implementation "the tall man" 5/2/2011 [NP] [NP] [AdjP] [DetP] tall the
Architectural Implementation "the tall man" 5/2/2011 [NP] [AdjP] [DetP] tall the
Architectural Implementation 5/2/2011 "the tall man" [XP] [NP] [AdjP] [DetP] [NP] tall the man DetP
Architectural Implementation 5/2/2011 "the tall man" [XP] [NP] [AdjP] [DetP] [NP] tall the man DetP
Architectural Implementation 5/2/2011 [XP] "the tall man" [NP] [AdjP] [DetP] [1] tall the man DetP [1]
Comparative Results Compare to previous incarnation of syntactic parser in Polyscheme: Loyal to CFG formalism Relies very heavily on search 5/2/2011
Comparative Results New model has significant benefits: Orders of magnitude faster (tens of minutes vs. seconds) Wider coverage of sentences 5/2/2011
Conclusions Introduced a syntactic theory based on the principles of Satellite Structure Positing Feature Structure Unification Feature Structure Aggregation Incrementality 5/2/2011
Conclusions Working implementation in Cognitive Architecture that is: Structurally Efficient Computationally fast Cognitively plausible 5/2/2011
Contribution We have presented a new grammatical formalism Implemented in a cognitive architecture Integrated with reasoning capabilities Computationally efficient, cognitively plausible Will lead towards a system that can understand natural language. 5/2/2011
Future Directions Many linguistic phenomena not mentioned here (not necessarily syntactic) Developing a new lexical representation theory to support interpretive grammar Integration with notions of pragmatics/discourse Integrate theory into working applications 5/2/2011
Questions? 5/2/2011
Special thanks to (in no particular order): Dr. Nick Cassimatis Perrin Bignoli JR Scally Soledad Vedovato John Borland Hiroyuki Uchida 5/2/2011
References 5/2/2011 Ball, J. T. (2004). A Cognitively Plausible Model of Language Comprehension. Proceedings of the 13th Conference on Behavior Representation in Modeling and Simulation. Ball, J., Rodgers, S., & Gluck, K. (2001). Integrating ACT-R and Cyc in a large-scale model of language comprehension for use in intelligent agents. Artificial Intelligence. Ball, J., Heiberg, A., & Silber, R. (2007). Toward a large-scale model of language comprehension in ACT-R 6. In R. L. Lewis, T. A. Polk, & J. E. Laird (Eds.), Proceedings of the 8th International Conference on Cognitive Modeling (pp. 163-168). Ball, Jerry T; Heiberg, Andrea; Silber, R. (2005). Toward a Large-Scale Model of Language Comprehension in ACT-R 6 Construction-Driven Language Processing. Language, 1. Ball, J. T. (2004). A Cognitively Plausible Model of Language Comprehension. Proceedings of the 13th Conference on Behavior Representation in Modeling and Simulation. Culicover, P. W., & Jackendoff, R. (2006). The simpler syntax hypothesis. Trends in cognitive sciences, 10(9), 413-8. doi: 10.1016/j.tics.2006.07.007. Lewis, R. L. (1993). An architecturally-based theory of human sentence comprehension. Proceedings of the fifteenth annual conference of the Cognitive Science Society June 18 to 21 1993 Institute of Cognitive Science University of Colorado Boulder (p. 108). Lawrence Erlbaum.  Lewis, R. L., Newell, A., & Polk, T. A. (1989). Toward a Soar theory of taking instructions for immediate reasoning. Proceedings of the Eleventh Annual Conference of the Cognitive Science Society (pp. 514-521). Erlbaum. Lewis, R. L., & Vasishth, S. (2005). An Activation-Based Model of Sentence Processing as Skilled Memory Retrieval. Cognitive Science, 29(3), 375-419. Psychology Press. doi: 10.1207/s15516709cog0000_25. Nivre, J. (2005). Dependency grammar and dependency parsing. MSI report, 5133(1959), 1-32. Citeseer.  Pollard, C., & Sag, I. (1994). Head Driven Phrase Structure Grammar. Studies in Contemporary Linguistics. University of Chicago Press. Pullman, Stepen G. (1991). Basic Parsing Techniques: an introductory survey. Sag, I. A., Wasow, T., & Bender, E. (2003). Syntactic Theory: A Formal Introduction (second edition). (I. A. Sag, Thomas Wasow, & Emily Bender, Eds.). CSLI Publications.

Mais conteúdo relacionado

Destaque

Understanding the mysteries of the CSS property value syntax
Understanding the mysteries of the CSS property value syntaxUnderstanding the mysteries of the CSS property value syntax
Understanding the mysteries of the CSS property value syntaxRuss Weakley
 
The Boundary between Syntax and Semantics - Prof. Fredreck J. Newmeyer
The Boundary between Syntax and Semantics - Prof. Fredreck J. NewmeyerThe Boundary between Syntax and Semantics - Prof. Fredreck J. Newmeyer
The Boundary between Syntax and Semantics - Prof. Fredreck J. NewmeyerPhoenix Tree Publishing Inc
 
Chomskyan linguistics 2
 Chomskyan linguistics 2 Chomskyan linguistics 2
Chomskyan linguistics 2Hina Honey
 
The Study of Language - Syntax - Theoretical Linguistics
The Study of Language - Syntax - Theoretical LinguisticsThe Study of Language - Syntax - Theoretical Linguistics
The Study of Language - Syntax - Theoretical LinguisticsMay F Ra'fat
 
Avram noam chomsky's services to syntax.
Avram noam chomsky's services to syntax.Avram noam chomsky's services to syntax.
Avram noam chomsky's services to syntax.Nabeela Taimur Ali
 
Syntax by George Yule
Syntax by George YuleSyntax by George Yule
Syntax by George YuleAsif Ali Raza
 
The Chomskyan Revolution - Prof. Fredreck J. Newmeyer
The Chomskyan Revolution - Prof. Fredreck J. NewmeyerThe Chomskyan Revolution - Prof. Fredreck J. Newmeyer
The Chomskyan Revolution - Prof. Fredreck J. NewmeyerPhoenix Tree Publishing Inc
 
Chomskyan linguistics lec 3
Chomskyan linguistics lec 3Chomskyan linguistics lec 3
Chomskyan linguistics lec 3Hina Honey
 
The Propaganda Model (Chomsky) -ZK
The Propaganda Model (Chomsky) -ZKThe Propaganda Model (Chomsky) -ZK
The Propaganda Model (Chomsky) -ZKZareen Khan
 
BIT (UCSC) Final Year Project - Microfinance Loan Management System
BIT (UCSC) Final Year Project - Microfinance Loan Management SystemBIT (UCSC) Final Year Project - Microfinance Loan Management System
BIT (UCSC) Final Year Project - Microfinance Loan Management SystemThiwanka Makumburage
 
Introduction to syntax
Introduction to syntaxIntroduction to syntax
Introduction to syntaxFarjana Ela
 

Destaque (19)

Space syntax
Space syntaxSpace syntax
Space syntax
 
Understanding the mysteries of the CSS property value syntax
Understanding the mysteries of the CSS property value syntaxUnderstanding the mysteries of the CSS property value syntax
Understanding the mysteries of the CSS property value syntax
 
Noam Chomsky
Noam ChomskyNoam Chomsky
Noam Chomsky
 
The Boundary between Syntax and Semantics - Prof. Fredreck J. Newmeyer
The Boundary between Syntax and Semantics - Prof. Fredreck J. NewmeyerThe Boundary between Syntax and Semantics - Prof. Fredreck J. Newmeyer
The Boundary between Syntax and Semantics - Prof. Fredreck J. Newmeyer
 
Noam Chomsky
Noam ChomskyNoam Chomsky
Noam Chomsky
 
Syntax
SyntaxSyntax
Syntax
 
Chomskyan linguistics 2
 Chomskyan linguistics 2 Chomskyan linguistics 2
Chomskyan linguistics 2
 
The Study of Language - Syntax - Theoretical Linguistics
The Study of Language - Syntax - Theoretical LinguisticsThe Study of Language - Syntax - Theoretical Linguistics
The Study of Language - Syntax - Theoretical Linguistics
 
Avram noam chomsky's services to syntax.
Avram noam chomsky's services to syntax.Avram noam chomsky's services to syntax.
Avram noam chomsky's services to syntax.
 
Syntax by George Yule
Syntax by George YuleSyntax by George Yule
Syntax by George Yule
 
The Chomskyan Revolution - Prof. Fredreck J. Newmeyer
The Chomskyan Revolution - Prof. Fredreck J. NewmeyerThe Chomskyan Revolution - Prof. Fredreck J. Newmeyer
The Chomskyan Revolution - Prof. Fredreck J. Newmeyer
 
Chomskyan linguistics lec 3
Chomskyan linguistics lec 3Chomskyan linguistics lec 3
Chomskyan linguistics lec 3
 
Linguistics -Syntax
Linguistics -SyntaxLinguistics -Syntax
Linguistics -Syntax
 
The Propaganda Model (Chomsky) -ZK
The Propaganda Model (Chomsky) -ZKThe Propaganda Model (Chomsky) -ZK
The Propaganda Model (Chomsky) -ZK
 
Noam Chomsky
Noam ChomskyNoam Chomsky
Noam Chomsky
 
BIT (UCSC) Final Year Project - Microfinance Loan Management System
BIT (UCSC) Final Year Project - Microfinance Loan Management SystemBIT (UCSC) Final Year Project - Microfinance Loan Management System
BIT (UCSC) Final Year Project - Microfinance Loan Management System
 
Introduction to syntax
Introduction to syntaxIntroduction to syntax
Introduction to syntax
 
SYNTAX - head and modifiers
SYNTAX - head and modifiersSYNTAX - head and modifiers
SYNTAX - head and modifiers
 
ENGLISH SYNTAX
ENGLISH SYNTAXENGLISH SYNTAX
ENGLISH SYNTAX
 

Semelhante a Issues W2011 Final

Towards Linked Ontologies and Data on the Semantic Web
Towards Linked Ontologies and Data on the Semantic WebTowards Linked Ontologies and Data on the Semantic Web
Towards Linked Ontologies and Data on the Semantic WebJie Bao
 
STRESS TEST FOR BERT AND DEEP MODELS: PREDICTING WORDS FROM ITALIAN POETRY
STRESS TEST FOR BERT AND DEEP MODELS: PREDICTING WORDS FROM ITALIAN POETRYSTRESS TEST FOR BERT AND DEEP MODELS: PREDICTING WORDS FROM ITALIAN POETRY
STRESS TEST FOR BERT AND DEEP MODELS: PREDICTING WORDS FROM ITALIAN POETRYkevig
 
STRESS TEST FOR BERT AND DEEP MODELS: PREDICTING WORDS FROM ITALIAN POETRY
STRESS TEST FOR BERT AND DEEP MODELS: PREDICTING WORDS FROM ITALIAN POETRYSTRESS TEST FOR BERT AND DEEP MODELS: PREDICTING WORDS FROM ITALIAN POETRY
STRESS TEST FOR BERT AND DEEP MODELS: PREDICTING WORDS FROM ITALIAN POETRYkevig
 
Formalization and implementation of BFO 2 with a focus on the OWL implementation
Formalization and implementation of BFO 2 with a focus on the OWL implementationFormalization and implementation of BFO 2 with a focus on the OWL implementation
Formalization and implementation of BFO 2 with a focus on the OWL implementationgolpedegato2
 
ESSLLI2016 DTS Lecture Day 5-1: Introduction to day 5
ESSLLI2016 DTS Lecture Day 5-1: Introduction to day 5ESSLLI2016 DTS Lecture Day 5-1: Introduction to day 5
ESSLLI2016 DTS Lecture Day 5-1: Introduction to day 5Daisuke BEKKI
 
ELABORATE LEXICON EXTENDED LANGUAGE WITH A LOT OF CONCEPTUAL INFORMATION
ELABORATE LEXICON EXTENDED LANGUAGE WITH A LOT OF CONCEPTUAL INFORMATION ELABORATE LEXICON EXTENDED LANGUAGE WITH A LOT OF CONCEPTUAL INFORMATION
ELABORATE LEXICON EXTENDED LANGUAGE WITH A LOT OF CONCEPTUAL INFORMATION IJCSEA Journal
 
Elaborate Lexicon Extended Language with a Lot of Conceptual Information
Elaborate Lexicon Extended Language with a Lot of Conceptual InformationElaborate Lexicon Extended Language with a Lot of Conceptual Information
Elaborate Lexicon Extended Language with a Lot of Conceptual InformationIJCSEA Journal
 
ELABORATE LEXICON EXTENDED LANGUAGE WITH A LOT OF CONCEPTUAL INFORMATION
ELABORATE LEXICON EXTENDED LANGUAGE WITH A LOT OF CONCEPTUAL INFORMATION ELABORATE LEXICON EXTENDED LANGUAGE WITH A LOT OF CONCEPTUAL INFORMATION
ELABORATE LEXICON EXTENDED LANGUAGE WITH A LOT OF CONCEPTUAL INFORMATION IJCSEA Journal
 
ELABORATE LEXICON EXTENDED LANGUAGE WITH A LOT OF CONCEPTUAL INFORMATION
ELABORATE LEXICON EXTENDED LANGUAGE WITH A LOT OF CONCEPTUAL INFORMATIONELABORATE LEXICON EXTENDED LANGUAGE WITH A LOT OF CONCEPTUAL INFORMATION
ELABORATE LEXICON EXTENDED LANGUAGE WITH A LOT OF CONCEPTUAL INFORMATIONIJCSEA Journal
 
The Onion Patch: Migration in Open Source Ecosystems
The Onion Patch: Migration in Open Source EcosystemsThe Onion Patch: Migration in Open Source Ecosystems
The Onion Patch: Migration in Open Source EcosystemsPatrick Wagstrom
 
Evolution of minds and languages: What evolved first and develops first in ch...
Evolution of minds and languages: What evolved first and develops first in ch...Evolution of minds and languages: What evolved first and develops first in ch...
Evolution of minds and languages: What evolved first and develops first in ch...Aaron Sloman
 
Classification of serialverb constructions
Classification of serialverb constructionsClassification of serialverb constructions
Classification of serialverb constructionsijaia
 
G2 pil a grapheme to-phoneme conversion tool for the italian language
G2 pil a grapheme to-phoneme conversion tool for the italian languageG2 pil a grapheme to-phoneme conversion tool for the italian language
G2 pil a grapheme to-phoneme conversion tool for the italian languageijnlc
 
A importância dos dados em sua arquitetura... uma visão muito além do SQL Ser...
A importância dos dados em sua arquitetura... uma visão muito além do SQL Ser...A importância dos dados em sua arquitetura... uma visão muito além do SQL Ser...
A importância dos dados em sua arquitetura... uma visão muito além do SQL Ser...Alexandre Porcelli
 
Cross-lingual ontology lexicalisation, translation and information extraction...
Cross-lingual ontology lexicalisation, translation and information extraction...Cross-lingual ontology lexicalisation, translation and information extraction...
Cross-lingual ontology lexicalisation, translation and information extraction...Tobias Wunner
 

Semelhante a Issues W2011 Final (20)

Towards Linked Ontologies and Data on the Semantic Web
Towards Linked Ontologies and Data on the Semantic WebTowards Linked Ontologies and Data on the Semantic Web
Towards Linked Ontologies and Data on the Semantic Web
 
Universal grammar
Universal grammarUniversal grammar
Universal grammar
 
STRESS TEST FOR BERT AND DEEP MODELS: PREDICTING WORDS FROM ITALIAN POETRY
STRESS TEST FOR BERT AND DEEP MODELS: PREDICTING WORDS FROM ITALIAN POETRYSTRESS TEST FOR BERT AND DEEP MODELS: PREDICTING WORDS FROM ITALIAN POETRY
STRESS TEST FOR BERT AND DEEP MODELS: PREDICTING WORDS FROM ITALIAN POETRY
 
STRESS TEST FOR BERT AND DEEP MODELS: PREDICTING WORDS FROM ITALIAN POETRY
STRESS TEST FOR BERT AND DEEP MODELS: PREDICTING WORDS FROM ITALIAN POETRYSTRESS TEST FOR BERT AND DEEP MODELS: PREDICTING WORDS FROM ITALIAN POETRY
STRESS TEST FOR BERT AND DEEP MODELS: PREDICTING WORDS FROM ITALIAN POETRY
 
Formalization and implementation of BFO 2 with a focus on the OWL implementation
Formalization and implementation of BFO 2 with a focus on the OWL implementationFormalization and implementation of BFO 2 with a focus on the OWL implementation
Formalization and implementation of BFO 2 with a focus on the OWL implementation
 
AICOL2015_paper_16
AICOL2015_paper_16AICOL2015_paper_16
AICOL2015_paper_16
 
ESSLLI2016 DTS Lecture Day 5-1: Introduction to day 5
ESSLLI2016 DTS Lecture Day 5-1: Introduction to day 5ESSLLI2016 DTS Lecture Day 5-1: Introduction to day 5
ESSLLI2016 DTS Lecture Day 5-1: Introduction to day 5
 
ELABORATE LEXICON EXTENDED LANGUAGE WITH A LOT OF CONCEPTUAL INFORMATION
ELABORATE LEXICON EXTENDED LANGUAGE WITH A LOT OF CONCEPTUAL INFORMATION ELABORATE LEXICON EXTENDED LANGUAGE WITH A LOT OF CONCEPTUAL INFORMATION
ELABORATE LEXICON EXTENDED LANGUAGE WITH A LOT OF CONCEPTUAL INFORMATION
 
Elaborate Lexicon Extended Language with a Lot of Conceptual Information
Elaborate Lexicon Extended Language with a Lot of Conceptual InformationElaborate Lexicon Extended Language with a Lot of Conceptual Information
Elaborate Lexicon Extended Language with a Lot of Conceptual Information
 
ELABORATE LEXICON EXTENDED LANGUAGE WITH A LOT OF CONCEPTUAL INFORMATION
ELABORATE LEXICON EXTENDED LANGUAGE WITH A LOT OF CONCEPTUAL INFORMATION ELABORATE LEXICON EXTENDED LANGUAGE WITH A LOT OF CONCEPTUAL INFORMATION
ELABORATE LEXICON EXTENDED LANGUAGE WITH A LOT OF CONCEPTUAL INFORMATION
 
ELABORATE LEXICON EXTENDED LANGUAGE WITH A LOT OF CONCEPTUAL INFORMATION
ELABORATE LEXICON EXTENDED LANGUAGE WITH A LOT OF CONCEPTUAL INFORMATIONELABORATE LEXICON EXTENDED LANGUAGE WITH A LOT OF CONCEPTUAL INFORMATION
ELABORATE LEXICON EXTENDED LANGUAGE WITH A LOT OF CONCEPTUAL INFORMATION
 
The Onion Patch: Migration in Open Source Ecosystems
The Onion Patch: Migration in Open Source EcosystemsThe Onion Patch: Migration in Open Source Ecosystems
The Onion Patch: Migration in Open Source Ecosystems
 
Evolution of minds and languages: What evolved first and develops first in ch...
Evolution of minds and languages: What evolved first and develops first in ch...Evolution of minds and languages: What evolved first and develops first in ch...
Evolution of minds and languages: What evolved first and develops first in ch...
 
Classification of serialverb constructions
Classification of serialverb constructionsClassification of serialverb constructions
Classification of serialverb constructions
 
Barbiers iclave-fr
Barbiers iclave-frBarbiers iclave-fr
Barbiers iclave-fr
 
STI Summit 2011 - Linked Data & Ontologies
STI Summit 2011 - Linked Data & OntologiesSTI Summit 2011 - Linked Data & Ontologies
STI Summit 2011 - Linked Data & Ontologies
 
Generative grammar
Generative grammarGenerative grammar
Generative grammar
 
G2 pil a grapheme to-phoneme conversion tool for the italian language
G2 pil a grapheme to-phoneme conversion tool for the italian languageG2 pil a grapheme to-phoneme conversion tool for the italian language
G2 pil a grapheme to-phoneme conversion tool for the italian language
 
A importância dos dados em sua arquitetura... uma visão muito além do SQL Ser...
A importância dos dados em sua arquitetura... uma visão muito além do SQL Ser...A importância dos dados em sua arquitetura... uma visão muito além do SQL Ser...
A importância dos dados em sua arquitetura... uma visão muito além do SQL Ser...
 
Cross-lingual ontology lexicalisation, translation and information extraction...
Cross-lingual ontology lexicalisation, translation and information extraction...Cross-lingual ontology lexicalisation, translation and information extraction...
Cross-lingual ontology lexicalisation, translation and information extraction...
 

Issues W2011 Final

  • 1. An Interpretation-Driven Model of Syntax Richard Caneba canebr@rpi.edu RPI Cognitive Science Department Human-Level Intelligence Laboratory 5/2/2011
  • 2. Introduction We start with a goal: Develop a system that can understand natural language. 5/2/2011
  • 3. Introduction We start with a goal: Develop a system that can understand natural language. (Roughly) three sub-goals: Syntactic Parsing Semantic Representation Pragmatics/Discourse 5/2/2011
  • 4. Introduction We start with a goal: Develop a system that can understand natural language. (Roughly) three sub-goals: Syntactic Parsing Semantic Representation Pragmatics/Discourse Stage 1: Syntax Why is syntax important for natural language understanding? 5/2/2011
  • 5. Introduction: Syntax “The dog bit the man.” 5/2/2011
  • 6. Introduction: Syntax “The dog bit the man.” 5/2/2011
  • 7. Introduction: Syntax “The man bit the dog.” 5/2/2011
  • 8. Introduction: Syntax “The man bit the dog.” 5/2/2011
  • 9. Introduction: Syntax “I hit the man with my car.” 5/2/2011
  • 10. Introduction: Syntax 5/2/2011 “[I] hit [the man] [with my car].”
  • 11. Introduction: Syntax “[I] hit [the man] [with my car].” 5/2/2011 Option 1: “with my car” modifies “hit””I hit the man while driving my car.”
  • 12. Introduction: Syntax “[I] hit [the man] [with my car].” 5/2/2011 Option 2: “with my car” modifies “the man””I hit the man who had my car.”
  • 13. Introduction: Syntax Syntactic interpretation yields very distinct semantic interpretations. 5/2/2011
  • 14. Introduction: Syntax Syntactic interpretation yields very distinct semantic interpretations. Helps identify the role words play in an utterance. 5/2/2011
  • 15. Introduction: Syntax Thus, syntax plays a fundamental role in natural language understanding. 5/2/2011
  • 16. Syntax Current grammar formalisms have number of shortcomings 5/2/2011
  • 17. Syntax Current grammar formalisms have number of shortcomings From perspective of trying to develop a system to understand natural language 5/2/2011
  • 18. Syntax Current grammar formalisms have number of shortcomings From perspective of trying to develop a system to understand natural language Focus on generative grammar (i.e. Chomskyan, HPSG [11], [13]) 5/2/2011
  • 19. Syntax: Theory & Implementation Two classes of shortcomings: Theoretical: shortcoming in the way a theory of language represents the users mental knowledge of that language. Implementation: shortcoming in the way a theory implies or represents language processing in terms of computability and/or cognitive realism. 5/2/2011
  • 20. Syntax: Theory & Implementation We will show that these two classes are so closely related, does not make sense to make a strong distinction 5/2/2011
  • 21. Theoretical Shortcomings Shortcoming in the way a theory of language represents the users mental knowledge of that language 5/2/2011
  • 23. Theoretical Shortcomings i.e. “Fido bit dog.” 5/2/2011
  • 24. Theoretical Shortcomings i.e. “Fido bit dog.” cf. “Fido bit the dog.” 5/2/2011
  • 25. Why Syntax? Why syntax at all, if we can still interpret ungrammaticality? 5/2/2011
  • 26. Why Syntax? “Fido bit dog.” cf. “Dog bit Fido.” 5/2/2011 Fido Fido
  • 28. Theoretical Shortcomings Phrasal Nesting example: “the man in boston with the hat is here.” 5/2/2011
  • 29. 5/2/2011 [S] [NP] [VP] [NP] is [PP] [PP] here man [NP] [DP] in the [PP] boston with [NP] hat [DP] “the man in boston with the hat is here.” the
  • 30. 5/2/2011 [S] [NP] [VP] [NP] is [PP] [PP] here man [NP] [DP] in the [PP] boston with [NP] hat [DP] “the man in boston with the hat is here.” the
  • 31. 5/2/2011 [S] [VP] [NP] is [PP] man [DP] [PP] [PP] here with [NP] [NP] in the hat [DP] boston the “the man in boston with the hat is here.”
  • 32. 5/2/2011 [S] [VP] [NP] is [PP] man [DP] [PP] [PP] here with [NP] [NP] in the hat [DP] boston the “the man in boston with the hat is here.”
  • 33. Implementation Shortcomings Shortcoming in the way a theory implies or represents language processing in terms of computability and/or cognitive realism. 5/2/2011
  • 37. Implementation Shortcomings john saw mary 5/2/2011
  • 38. Implementation Shortcomings [VP] [NP] 5/2/2011 john saw mary
  • 39. Implementation Shortcomings [S] [VP] [NP] [NP] 5/2/2011 john saw mary
  • 41. Implementation Shortcomings 5/2/2011 the tall strong angry man
  • 42. Implementation Shortcomings 5/2/2011 [NP] the tall strong angry man
  • 43. Implementation Shortcomings 5/2/2011 [NP] [NP] the tall strong angry man
  • 44. Implementation Shortcomings 5/2/2011 [NP] [NP] [NP] the tall strong angry man
  • 45. Implementation Shortcomings 5/2/2011 [NP] [NP] [NP] [NP] the tall strong angry man
  • 46. Implementation Shortcomings 5/2/2011 [NP] [NP] [NP] [NP] the tall strong angry man
  • 47. Implementation Shortcomings 5/2/2011 [NP] [NP] [NP] [NP] [PP] the tall strong angry man in boston
  • 48. Implementation Shortcomings 5/2/2011 [NP] [NP] [NP] [NP] [PP] the tall strong angry man in boston
  • 49. Implementation Shortcomings 5/2/2011 [NP] [NP] [NP] [NP] [PP] the tall strong angry man in boston
  • 50. Implementation Shortcomings 5/2/2011 [NP] [NP] [NP] [NP] [PP] the tall strong angry man in boston
  • 51. Implementation Shortcomings 5/2/2011 [NP] [NP] [NP] [NP] [PP] the tall strong angry man in boston
  • 52. Implementation Shortcomings 5/2/2011 [NP] [NP] [NP] [NP] [PP] the tall strong angry man in boston
  • 53. Syntax: Statistical Approach Most currently successful parsing algorithms rely heavily on statistics. However, inferences that require notion of semantics difficult 5/2/2011
  • 54. Syntax: Statistical Approach i.e. “The couple walked in the park. He held her hand.” 5/2/2011
  • 55. Syntax: Statistical Approach i.e. “The couple walked in the park. He held her hand.” Statistics does not give anaphoric bind. 5/2/2011
  • 56. Chomskyan Approach From [6], consider “Joe has put those raw potatoes in the pot.” 5/2/2011
  • 58. Chomskyan Approach A simpler analysis is possible: 5/2/2011
  • 60. Chomskyan Approach Illustrates: way a theory is represented has very real computational concerns. 5/2/2011
  • 61. Other Notable Work Other Cognitive Architecture based Parsers [8] R. Lewis et.al. developed a parser based on "immediate reasoning" in Soar [9] R. Lewis et.al. developed an activation based parser model in ACT-R [1][2][3] J.T. Ball et.al. developed a parser based on Double R Grammar Model, for "synthetic teammate" development in ACT-R 5/2/2011
  • 62. Other Notable Work However, each of these theories suffers from the shortcomings we’ve already seen. 5/2/2011
  • 63. Other Notable Work However, each of these theories suffers from the shortcomings we’ve already seen. Both parsers designed by Lewis rely on a CFG formalism Bell’s ACT-R parser does not deeply integrated with reasoning 5/2/2011
  • 64. Other Notable Work However, each of these theories suffers from the shortcomings we’ve already seen. Both parsers designed by Lewis rely on a CFG formalism Bell’s ACT-R parser does not deeply integrated with reasoning These approaches are not well integrated with reasoning overall 5/2/2011
  • 65. Motivating Principles To address these shortcomings from an interpretative perspective, four principles are motivated: 5/2/2011
  • 66. Motivating Principles To address these shortcomings from an interpretative perspective, four principles are motivated:a) The existence of satellite structuresb) Feature structure unificationc) Feature structure aggregationd) Incrementality 5/2/2011
  • 67. Satellite Structures When we hear words, infer existence of other structures related to that word 5/2/2011
  • 68. Satellite Structures When we hear words, infer existence of other structures related to that word i.e. “bit” 5/2/2011 bit bit
  • 69. Satellite Structures When we hear words, infer existence of other structures related to that word i.e. “bit” 5/2/2011 [NP] [NP] Obj Subj bit bit
  • 70. Feature Structure Unification With the existence of satellite structures, we can unify observed structures together 5/2/2011
  • 71. Feature Structure Unification With the existence of satellite structures, we can unify observed structures together 5/2/2011 [NP] [NP] Obj Subj john bit bit fido
  • 72. Feature Structure Unification With the existence of satellite structures, we can unify observed structures together 5/2/2011 [NP] [NP] Obj Subj john bit bit fido
  • 73. Feature Structure Aggregation Sequential unification multiple feature structures yields “aggregating” feature structure 5/2/2011
  • 74. Feature Structure Aggregation i.e. “the big dog” 5/2/2011
  • 75. Feature Structure Aggregation i.e. “the big dog” 5/2/2011 big the dog
  • 76. Feature Structure Aggregation i.e. “the big dog” 5/2/2011 [NP] [NP] [NP] big the dog
  • 77. Feature Structure Aggregation i.e. “the big dog” 5/2/2011 [NP] [NP] [NP] Unify big the dog
  • 78. Feature Structure Aggregation i.e. “the big dog” 5/2/2011 [NP] [NP] [NP] Unify Unify big the dog
  • 79. Feature Structure Aggregation i.e. “the big dog” 5/2/2011 Unify [NP] [NP] [NP] big the dog
  • 80. Incrementality Human sentence comprehension is left-to-right, and incremental Incrementally generate an interpretation 5/2/2011
  • 81. Using these principles, attempt to address the aforementioned shortcomingsin a model 5/2/2011 Architectural Implementation
  • 82. Architectural Implementation Parser prototype implemented in the Polyscheme Cognitive Architecture. 5/2/2011
  • 83. Architectural Implementation Parser prototype implemented in the Polyscheme Cognitive Architecture. Use the model to process and interpret natural language input. 5/2/2011
  • 84. Architectural Implementation Parsing driven by pair-wise pattern matching 5/2/2011
  • 85. Architectural Implementation 5/2/2011 Parsing driven by pair-wise pattern matching [XP] * [NP] [XP] * [NP] [DetP] [NP] [DetP] [D] [N] [D] [N] *Underspecified Feature Structure
  • 86. Architectural Implementation 5/2/2011 Parsing driven by pair-wise pattern matching [XP] * [NP] [XP] * [NP] [DetP] [NP] [DetP] [D] [N] [D] [N] *Underspecified Feature Structure
  • 87. Architectural Implementation 5/2/2011 [XP] [XP] [NP] [NP] [DetP] [NP] [DetP] [1] man the man the DetP [1]
  • 88. Architectural Implementation 5/2/2011 IsA(?the, WordUtteranceEvent) IsA(?the, Determiner) PartOf(?d, ?dp) IsA(?dp, PhraseUtteranceEvent) IsA(?dp, Determiner) PartOf(?dp, ?np1) IsA(?np1, PhraseUtteranceEvent) IsA(?np1, Noun) Meets(?the, ?man) IsA(?man, WordUtteranceEvent) IsA(?man, CommonNoun) Specifier(?man, ?detSpr) IsA(?detSpr, Determiner) PartOf(?man, ?np2) IsA(?np2, PhraseUtteranceEvent) IsA(?np2, Noun)  Same(?dp, ?detSpr) Same(?np1, ?np2) [XP] [XP] [NP] [NP] [DetP] [NP] [DetP] [1] man the man the DetP [1]
  • 89. 5/2/2011 [DetP][N]  NP[DetP, N] [AdjP][N]  NP[AdjP, N] [XP] [XP] [XP] [XP] [XP] [NP] [NP] [NP] [DetP] [NP] [AdjP] [NP] [DetP] [AdjP] [D] [N] [D] [N] [Adj] [N] [Adj] [N] [DetP][AdjP]  NP[DetP, AdjP] [XP] [XP] [NP] [NP] [DetP] [AdjP] [AdjP] [DetP] [D] [Adj] [D] [Adj]
  • 90. 5/2/2011 [P][NP]  PP[P, NP] [V][NP]  VP[V, NP] [XP] [XP] [XP] [XP] [XP] [VP] [XP] [PP] [PP] [NP] [NP] [NP] [VP] [NP] [P] [X] [P] [X] [X] [X] [V] [V] [NP][V]  VP[NP, V] [XP] [XP] [XP] [VP] [NP] [VP] [NP] [N] [V] [N] [V]
  • 91. "the tall man" 5/2/2011 Architectural Implementation
  • 92. Architectural Implementation "the tall man" 5/2/2011 [NP] [DetP] the
  • 93. Architectural Implementation "the tall man" 5/2/2011 [NP] [NP] [AdjP] [DetP] tall the
  • 94. Architectural Implementation "the tall man" 5/2/2011 [NP] [NP] [AdjP] [DetP] tall the
  • 95. Architectural Implementation "the tall man" 5/2/2011 [NP] [AdjP] [DetP] tall the
  • 96. Architectural Implementation 5/2/2011 "the tall man" [XP] [NP] [AdjP] [DetP] [NP] tall the man DetP
  • 97. Architectural Implementation 5/2/2011 "the tall man" [XP] [NP] [AdjP] [DetP] [NP] tall the man DetP
  • 98. Architectural Implementation 5/2/2011 [XP] "the tall man" [NP] [AdjP] [DetP] [1] tall the man DetP [1]
  • 99. Comparative Results Compare to previous incarnation of syntactic parser in Polyscheme: Loyal to CFG formalism Relies very heavily on search 5/2/2011
  • 100. Comparative Results New model has significant benefits: Orders of magnitude faster (tens of minutes vs. seconds) Wider coverage of sentences 5/2/2011
  • 101. Conclusions Introduced a syntactic theory based on the principles of Satellite Structure Positing Feature Structure Unification Feature Structure Aggregation Incrementality 5/2/2011
  • 102. Conclusions Working implementation in Cognitive Architecture that is: Structurally Efficient Computationally fast Cognitively plausible 5/2/2011
  • 103. Contribution We have presented a new grammatical formalism Implemented in a cognitive architecture Integrated with reasoning capabilities Computationally efficient, cognitively plausible Will lead towards a system that can understand natural language. 5/2/2011
  • 104. Future Directions Many linguistic phenomena not mentioned here (not necessarily syntactic) Developing a new lexical representation theory to support interpretive grammar Integration with notions of pragmatics/discourse Integrate theory into working applications 5/2/2011
  • 106. Special thanks to (in no particular order): Dr. Nick Cassimatis Perrin Bignoli JR Scally Soledad Vedovato John Borland Hiroyuki Uchida 5/2/2011
  • 107. References 5/2/2011 Ball, J. T. (2004). A Cognitively Plausible Model of Language Comprehension. Proceedings of the 13th Conference on Behavior Representation in Modeling and Simulation. Ball, J., Rodgers, S., & Gluck, K. (2001). Integrating ACT-R and Cyc in a large-scale model of language comprehension for use in intelligent agents. Artificial Intelligence. Ball, J., Heiberg, A., & Silber, R. (2007). Toward a large-scale model of language comprehension in ACT-R 6. In R. L. Lewis, T. A. Polk, & J. E. Laird (Eds.), Proceedings of the 8th International Conference on Cognitive Modeling (pp. 163-168). Ball, Jerry T; Heiberg, Andrea; Silber, R. (2005). Toward a Large-Scale Model of Language Comprehension in ACT-R 6 Construction-Driven Language Processing. Language, 1. Ball, J. T. (2004). A Cognitively Plausible Model of Language Comprehension. Proceedings of the 13th Conference on Behavior Representation in Modeling and Simulation. Culicover, P. W., & Jackendoff, R. (2006). The simpler syntax hypothesis. Trends in cognitive sciences, 10(9), 413-8. doi: 10.1016/j.tics.2006.07.007. Lewis, R. L. (1993). An architecturally-based theory of human sentence comprehension. Proceedings of the fifteenth annual conference of the Cognitive Science Society June 18 to 21 1993 Institute of Cognitive Science University of Colorado Boulder (p. 108). Lawrence Erlbaum. Lewis, R. L., Newell, A., & Polk, T. A. (1989). Toward a Soar theory of taking instructions for immediate reasoning. Proceedings of the Eleventh Annual Conference of the Cognitive Science Society (pp. 514-521). Erlbaum. Lewis, R. L., & Vasishth, S. (2005). An Activation-Based Model of Sentence Processing as Skilled Memory Retrieval. Cognitive Science, 29(3), 375-419. Psychology Press. doi: 10.1207/s15516709cog0000_25. Nivre, J. (2005). Dependency grammar and dependency parsing. MSI report, 5133(1959), 1-32. Citeseer. Pollard, C., & Sag, I. (1994). Head Driven Phrase Structure Grammar. Studies in Contemporary Linguistics. University of Chicago Press. Pullman, Stepen G. (1991). Basic Parsing Techniques: an introductory survey. Sag, I. A., Wasow, T., & Bender, E. (2003). Syntactic Theory: A Formal Introduction (second edition). (I. A. Sag, Thomas Wasow, & Emily Bender, Eds.). CSLI Publications.

Notas do Editor

  1. Consider this sentence “the dog bit the man”
  2. Our interpretation of the described event is like this.
  3. Now consider a rearrangement of the words: “The man bit the dog.”
  4. We get a complete different interpretation.
  5. Now it may seem as though word order is all that syntax is. However, consider this example: “I hit the man with my car.”In particular, we can focus on the prepositional phrase at the end, “with my car.”
  6. Now it may seem as though word order is all that syntax is. However, consider this example: “I hit the man with my car.”In particular, we can focus on the prepositional phrase at the end, “with my car.”
  7. We have the first reading here, where the preposition modifies the VP “hit”.In this instance, we get the interpretation that “I hit the man WHILE DRIVING my car.”However, a second reading does exist…
  8. If “with my car” modifies the NP “the man,” we get an interpretation that is “I hit the man WHO HAD my car.”Thus you can see that, with these two reasons, syntactic interpretation is a lot more than simply word order. On top of that, it is important to us as human beings, as the syntactic interpretation of “I hit the man with my car” has very important legal ramifications.
  9. Moving towards a theory of grammar that is useful for interpreting natural language, we realized that current grammar formalisms have a number of shortcomings.
  10. ->We will show that, in fact, theoretical and implementation considerations are so closely related, that it doesn’t make sense to make a strong distinction between them.->That is to say, we can deal with many important shortcomings both theoretically and implementation-wise in a fairly elegant manner, using a unified framework for addressing both.
  11. Generative grammar not concerned with interpretability.Concerned with over-generationInterpretability extends beyond simply grammatical generation.
  12. Generative grammar formalisms do not give us an account of how we can interpret sentences. Generative grammar, in this case, only goes so far as saying that this sentence is ungrammatical, but has no concern for the fact that it is still interpretable.
  13. Generative grammar formalisms do not give us an account of how we can interpret sentences. Generative grammar, in this case, only goes so far as saying that this sentence is ungrammatical, but has no concern for the fact that it is still interpretable.
  14. We use our knowledge of grammaticality to interpret an utterance, even if it is ungrammatical. So, there is a similarity between “Fido bit dog” and “Fido bit the dog.” Our knowledge of grammar allows us to get this meaning.
  15. We use our knowledge of grammaticality to interpret an utterance, even if it is ungrammatical. Our knowledge of grammar allows us to get this meaning.Thus, traditional grammar formalism is not concerned with how grammatical knowledge is actually used, but rather only the nature of grammatical knowledge itself.
  16. Many grammar formalisms require a binary branching in order to preserve the recursive aspect of the rule applications. However, this leads to spurious parse trees that are a result of the phrasal nesting. In many cases, it’s more accurate to deliver a flat parse tree.i.e. Consider the sentence “the man in boston with the hat is here.”
  17. Many grammar formalisms require a binary branching in order to preserve the recursive aspect of the rule applications. However, this leads to spurious parse trees that are a result of the phrasal nesting. In many cases, it’s more accurate to deliver a flat parse tree.i.e. Consider the sentence “the man in boston with the hat is here.”
  18. Many grammar formalisms require a binary branching in order to preserve the recursive aspect of the rule applications. However, this leads to spurious parse trees that are a result of the phrasal nesting. In many cases, it’s more accurate to deliver a flat parse tree.i.e. Consider the sentence “the man in boston with the hat is here.”In particular, let’s focus on the prepositional phrasal nesting that occurs here:
  19. Many grammar formalisms require a binary branching in order to preserve the recursive aspect of the rule applications. However, this leads to spurious parse trees that are a result of the phrasal nesting. In many cases, it’s more accurate to deliver a flat parse tree.i.e. Consider the sentence “the man in boston with the hat is here.”In particular, let’s focus on the prepositional phrasal nesting that occurs here:This seems to imply that that “with the hat” is modifying the NP headed by “boston,” as if the city of Boston has a hat. In fact, we are suggesting that it is “the man” that is the one with the hat. A more correct parse would be:
  20. Many grammar formalisms require a binary branching in order to preserve the recursive aspect of the rule applications. However, this leads to spurious parse trees that are a result of the phrasal nesting. In many cases, it’s more accurate to deliver a flat parse tree.i.e. Consider the sentence “the man in boston with the hat is here.”In particular, let’s focus on the prepositional phrasal nesting that occurs here:This seems to imply that that “with the hat” is modifying the NP headed by “boston,” as if the city of Boston has a hat. In fact, we are suggesting that it is “the man” that is the one with the hat. A more correct parse would be:
  21. Many grammar formalisms require a binary branching in order to preserve the recursive aspect of the rule applications. However, this leads to spurious parse trees that are a result of the phrasal nesting. In many cases, it’s more accurate to deliver a flat parse tree.i.e. Consider the sentence “the man in boston with the hat is here.”In particular, let’s focus on the prepositional phrasal nesting that occurs here:This seems to imply that that “with the hat” is modifying the NP headed by “boston,” as if the city of Boston has a hat. In fact, we are suggesting that it is “the man” that is the one with the hat. A more correct parse would be:Notice the 4-ary branching that occurs at this NP. This is more realistic and closer to what is actually intended by the utterance.Hence, you’ll notice that syntactic interpretation is more than just a question of word order. It is in fact highly integrated with a lot of reasoning that would yield this interpretation, as opposed to the previous one.
  22. Many theories of grammar are not concerned with the dynamic nature of syntactic parsing, and how syntax is generated on-line and incrementally in a left-right fashion.i.e. Complement selection before specifier selection.
  23. Many theories of grammar are not concerned with the dynamic nature of syntactic parsing, and how syntax is generated on-line and incrementally in a left-right fashion.i.e. Complement selection before specifier selection.
  24. Many theories of grammar are not concerned with the dynamic nature of syntactic parsing, and how syntax is generated on-line and incrementally in a left-right fashion.i.e. Complement selection before specifier selection.
  25. Many theories of grammar are not concerned with the dynamic nature of syntactic parsing, and how syntax is generated on-line and incrementally in a left-right fashion.i.e. Complement selection before specifier selection.
  26. Many theories of grammar are not concerned with the dynamic nature of syntactic parsing, and how syntax is generated on-line and incrementally in a left-right fashion.i.e. Complement selection before specifier selection.
  27. Many theories of grammar are not concerned with the dynamic nature of syntactic parsing, and how syntax is generated on-line and incrementally in a left-right fashion.i.e. Complement selection before specifier selection.You can see in this case that, before we could identify the subject of the sentence (“john”), we had to identify the object (“mary”). However, this misses the crucial intuition that, when we hear “john saw…” we’ve already identified the fact that “john” is the subject of the sentence.
  28. Many theories of grammar posit an unnecessary amount of structure in order to preserve the recursiveness of their rule applications. Consider this case:“the tall strong angry man.”
  29. Many theories of grammar posit an unnecessary amount of structure in order to preserve the recursiveness of their rule applications. Consider this case:“the tall strong angry man.”
  30. Many theories of grammar posit an unnecessary amount of structure in order to preserve the recursiveness of their rule applications. Consider this case:“the tall strong angry man.”
  31. Many theories of grammar posit an unnecessary amount of structure in order to preserve the recursiveness of their rule applications. Consider this case:“the tall strong angry man.”
  32. Many theories of grammar posit an unnecessary amount of structure in order to preserve the recursiveness of their rule applications. Consider this case:“the tall strong angry man.”
  33. Many theories of grammar posit an unnecessary amount of structure in order to preserve the recursiveness of their rule applications. Consider this case:“the tall strong angry man.”As the parse develops, we can see the needless addition of NP constituency structures. It would be ideal if we could obviate this need for unnecessary structure, as it is computationally inefficient, and increases the load on working memory.
  34. Another problem with this is that, potentially, there are a number of nodes here that are open for modificational attachment. In particular, consider the possibility of a prepositional phrase following, like “in boston.”
  35. Another problem with this is that, potentially, there are a number of nodes here that are open for modificational attachment. In particular, consider the possibility of a prepositional phrase following, like “in boston.”
  36. Another problem with this is that, potentially, there are a number of nodes here that are open for modificational attachment. In particular, consider the possibility of a prepositional phrase following, like “in boston.”
  37. Another problem with this is that, potentially, there are a number of nodes here that are open for modificational attachment. In particular, consider the possibility of a prepositional phrase following, like “in boston.”
  38. Another problem with this is that, potentially, there are a number of nodes here that are open for modificational attachment. In particular, consider the possibility of a prepositional phrase following, like “in boston.”
  39. Another problem with this is that, potentially, there are a number of nodes here that are open for modificational attachment. In particular, consider the possibility of a prepositional phrase following, like “in boston.”
  40. Another problem with this is that, potentially, there are a number of nodes here that are open for modificational attachment. In particular, consider the possibility of a prepositional phrase following, like “in boston.”So almost immediately, there is a high degree of attachment ambiguity, for what should be an unproblematic parse.
  41. Thus, if you start with the goal of natural language “understanding,” in any deep sense of understanding, statistics falters where representation beyond simply the words are required. So in this case, there has to be a notion of what it we mean by “the couple” in order to draw the link between the couple and the two pronouns in the second sentence.
  42. “joe has put those raw potatoes in the pot.”
  43. “joe has put those raw potatoes in the pot.”
  44. “joe has put those raw potatoes in the pot.”
  45. “joe has put those raw potatoes in the pot.”
  46. The way a theory is represented has very real concerns in terms of computability. Simply, the generation of a parse for what could potentially be a simple procedure (that is to some extent empirically supported by the fact that we don’t struggle heavily with interpreting these sentences) is a very expensive procedure, perhaps prohibitively so.
  47. -However, each of these theories suffers from the shortcomings we’ve already seen.
  48. -However, each of these theories suffers from the shortcomings we’ve already seen.-Both of Lewis' parsers rely on a context-free grammar formalism, and so are subject to those limitations.-Ball's parser: non-monotonic, no backtracking. We are monotonic, and use backtracking, which corresponds more strongly to human sentence comprehension. we disagree on this particular theoretical point: we believe that human cognition utilizes monotonicity, and backtracking, while they make the implicit assumption that this is not the case. -These approaches are not well integrated with reasoning.
  49. -However, each of these theories suffers from the shortcomings we’ve already seen.-Both of Lewis' parsers rely on a context-free grammar formalism, and so are subject to those limitations.-Ball's parser: non-monotonic, no backtracking. We are monotonic, and use backtracking, which corresponds more strongly to human sentence comprehension. we disagree on this particular theoretical point: we believe that human cognition utilizes monotonicity, and backtracking, while they make the implicit assumption that this is not the case. -These approaches are not well integrated with reasoning. They suffer from the shortcomings that we’ve already identified in CFG This poor-integration with reasoning can be seen as very related to the shortcomings we’ve identified.
  50. I know that there are satellite structures to the bit verb, that it takes an NP Subject, and an NP Object.
  51. Suppose we observe the existence of two NP’s, one that precedes the verb, and one that follows the verb.
  52. Thus, syntactic interpretation can be thought as the UNIFICATION of feature structures that already exist.This kind of unification is driven by two ideas:That words have satellite structures (in this case, “bit” has a subject NP and an object NP)Many words and phrases are also parts of other “satellite” centers (in this case, “fido” and “john” are taken by “bit” and treat it as a satellite center) With these ideas, we can unify these structures.So far then, we’ve observed that, we do not NEED to posit additional structures during a parse, so long as we know that they exist in some form. With this knowledge, unification of feature structures allows us to avoid the problem of excessive structure positing that occurs in many CFG formalisms.
  53. -Sequential unification of multiple feature structures yields an "aggregating" feature structure-Builds up the internal representation of a feature structure through aggregation across multiply posited structures
  54. -Sequential unification of multiple feature structures yields an "aggregating" feature structure-Builds up the internal representation of a feature structure through aggregation across multiply posited structures
  55. -Sequential unification of multiple feature structures yields an "aggregating" feature structure-Builds up the internal representation of a feature structure through aggregation across multiply posited structures
  56. -Sequential unification of multiple feature structures yields an "aggregating" feature structure-Builds up the internal representation of a feature structure through aggregation across multiply posited structures
  57. -Sequential unification of multiple feature structures yields an "aggregating" feature structure-Builds up the internal representation of a feature structure through aggregation across multiply posited structures
  58. -Sequential unification of multiple feature structures yields an "aggregating" feature structure-Builds up the internal representation of a feature structure through aggregation across multiply posited structures
  59. -Computationally, once again, we are avoiding the positing of excessive structure through the sequential application of unification of multiple feature structures. -Thus, reasoned unification applied in sequence generates these “aggregate” structures that are composed of the internal features of a number of different structures.
  60. This is fairly clear, that humans process sentences in an incremental manner, and do so (mostly) one word at a time, and as a sequential process.
  61. Using those points as motivations, we’ve developed a parser prototype that has been implemented in the Polyscheme cognitive architecture.
  62. The intention of the model is to process and interpret natural language input.
  63. Pair-wise pattern matching:Have a native set of pair-wise matching constraints that generate a parse from right to left. For instance, if you have a Determiner before a Noun, you can apply a rule that unifies structures and aggregates the NP.Pair-wise pattern matching is very fast, as it limits the relevant details to a decision point to a very small number of identifiable features of the current state of the parse.
  64. Pair-wise pattern matching:Have a native set of pair-wise matching constraints that generate a parse from right to left. For instance, if you have a Determiner before a Noun, you can apply a rule that unifies structures and aggregates the NP.Pair-wise pattern matching is very fast, as it limits the relevant details to a decision point to a very small number of identifiable features of the current state of the parse.
  65. Pair-wise pattern matching:Have a native set of pair-wise matching constraints that generate a parse from right to left. For instance, if you have a Determiner before a Noun, you can apply a rule that unifies structures and aggregates the NP.Pair-wise pattern matching is very fast, as it limits the relevant details to a decision point to a very small number of identifiable features of the current state of the parse.
  66. As you can see, the unification mechanism is actually very simple from a high level view like this. The backbone of feature structure unification is the Same() atom, which simply expresses the fact that two object’s are the same, and thus inherit each other’s attributes, similar to Leibniz’s law.
  67. As you can see, the unification mechanism is actually very simple from a high level view like this. The backbone of feature structure unification is the Same() atom, which simply expresses the fact that two object’s are the same, and thus inherit each other’s attributes.
  68. Here are just a small subset of the number of pair-wise pattern matching rules we can use.
  69. Here are just a small subset of the number of pair-wise pattern matching rules we can use.Focus on the two ways of attaching the complements. I’m still on the fence about which one I prefer, but I’m leaning towards the first way.
  70. At time step 1, we hear “the”.
  71. At this point, we’ve heard two words, “the” and “tall”.So we’ve incrementally added the next word in the utterance.We have an underspecified node at the top of phrasal chain of “tall.” What we can do is unify these structures with a rule.
  72. At this point, we’ve heard two words, “the” and “tall”.We have an underspecified node at the top of phrasal chain of “tall.” What we can do is unify these structures with a rule.
  73. At this point, we’ve heard two words, “the” and “tall”.We have an underspecified node at the top of phrasal chain of “tall.” What we can do is unify these structures with a rule.What has happened here is that the NP has been aggregated to the right, by positing that the NP whose existence was inferred at time step 1 is the SAME as the underspecified node inferred to exist at time step 2. So what this amounts to, in this case is that we have generated an aggregate grouping that, at this point, contains the first NP and the next XP.
  74. Now add the next word: “man”Once again, we can apply a rule to incrementally grow our parse, to incorporate the new word “man” into the growing syntactic parse tree.
  75. Now add the next word: “man”Once again, we can apply a rule to incrementally grow our parse, to incorporate the new word “man” into the growing syntactic parse tree.
  76. Now add the next word: “man”Once again, we can apply a rule to incrementally grow our parse, to incorporate the new word “man” into the growing syntactic parse tree.Thus, you can see that we’ve once again aggregated the NP structure, and inherited more information that helps to define in greater detail the internal structure of the aggregating NP.Thus, we’ve shown a method parsing that is both incremental left to right to grow a parse, that doesn’t posit excessive structure through the notion of unification of unification and aggregation.We’ve also identified the satellite structure that satisfies the obligatory specifier argument in the internal structure of “man”.
  77. So in a sense, it’s also more cognitively plausible, in that it reflects two key facts about human performance:Human sentence comprehension occurs in a finite time. We cannot take tens of minutes to converge upon an interpretation. So quicker sentence processing is closer to human level intelligence.Wider coverage of sentence phenomena is also moving closer to human-level intelligence, as the more sentences we can cover, the more human performance we’ve captured. So, high performance is, to a large degree, a move towards cognitive plausibility.