UTCS Artificial Intelligence
courses
talks/events
demos
people
projects
publications
software/data
labs
areas
admin
Weakly-Supervised Grammar-Informed Bayesian CCG Parser Learning (2015)
Dan Garrette
, Chris Dyer, Jason Baldridge, Noah A. Smith
Combinatory Categorial Grammar (CCG) is a lexicalized grammar formalism in which words are associated with categories that, in combination with a small universal set of rules, specify the syntactic configurations in which they may occur. Previous work has shown that learning sequence models for CCG tagging can be improved by using priors that are sensitive to the formal properties of CCG as well as cross-linguistic universals. We extend this approach to the task of learning a full CCG parser from weak supervision. We present a Bayesian formulation for CCG parser induction that assumes only supervision in the form of an incomplete tag dictionary mapping some word types to sets of potential categories. Our approach outperforms a baseline model trained with uniform priors by exploiting universal, intrinsic properties of the CCG formalism to bias the model toward simpler, more cross-linguistically common categories.
View:
PDF
Citation:
In
Proceedings of the Twenty-Ninth AAAI Conference on Artificial Intelligence (AAAI-15)
, Austin, TX, January 2015.
Bibtex:
@inproceedings{garrette:aaai15, title={Weakly-Supervised Grammar-Informed Bayesian CCG Parser Learning}, author={Dan Garrette and Chris Dyer and Jason Baldridge and Noah A. Smith}, booktitle={Proceedings of the Twenty-Ninth AAAI Conference on Artificial Intelligence (AAAI-15)}, month={January}, address={Austin, TX}, url="http://www.cs.utexas.edu/users/ai-labpub-view.php?PubID=127485", year={2015} }
Presentation:
Slides (PDF)
People
Dan Garrette
Ph.D. Alumni
dhg [at] cs utexas edu
Areas of Interest
Machine Learning
Natural Language Processing
Semi-Supervised Learning
Labs
Machine Learning