Modelling Human Syntax by Means of Probabilistic Dependency Grammars
Center for Language & Speech Processing(CLSP), JHU via YouTube
Overview
Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Learn about computational approaches to modeling human syntactic structures through probabilistic dependency grammars in this lecture by Matthias Buch-Kromann from Copenhagen Business School's Center for Computational Modelling of Language. Explore how dependency grammar frameworks can be enhanced with probabilistic methods to better capture the statistical patterns found in natural language syntax. Discover the theoretical foundations of dependency parsing and how probabilistic models can account for syntactic ambiguity and variation in human language processing. Examine practical applications of these computational models in natural language processing tasks and their implications for understanding how humans parse and generate syntactic structures. Gain insights into the intersection of computational linguistics, cognitive science, and statistical modeling as applied to syntactic analysis.
Syllabus
Matthias Buch-Kromann: Modelling human syntax by means of probabilistic dependency grammars
Taught by
Center for Language & Speech Processing(CLSP), JHU