skip to main content
10.3115/1119355.1119367dlproceedingsArticle/Chapter ViewAbstractPublication PagesemnlpConference Proceedingsconference-collections
Article
Free Access

Using LTAG based features in parse reranking

Published:11 July 2003Publication History

ABSTRACT

We propose the use of Lexicalized Tree Adjoining Grammar (LTAG) as a source of features that are useful for reranking the output of a statistical parser. In this paper, we extend the notion of a tree kernel over arbitrary sub-trees of the parse to the derivation trees and derived trees provided by the LTAG formalism, and in addition, we extend the original definition of the tree kernel, making it more lexicalized and more compact. We use LTAG based features for the parse reranking task and obtain labeled recall and precision of 89.7%/90.0% on WSJ section 23 of Penn Treebank for sentences of length ≤ 100 words. Our results show that the use of LTAG based tree kernel gives rise to a 17% relative difference in f-score improvement over the use of a linear kernel without LTAG based features.

References

  1. E. Black, F. Jelinek, J. Lafferty, Magerman D. M., R. Mercer, and S. Roukos. 1993. Towards history-based grammars: Using richer models for probabilistic parsing. In Proc. of the ACL 1993. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. R. Bod. 2003. An Efficient Implementation of a New DOP Model. In Proc. of EACL 2003, Budapest. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. J. Chen and K. Vijay-Shanker. 2000. Automated Extraction of TAGs from the Penn Treebank. In Proc. of the 6th IWPT.Google ScholarGoogle Scholar
  4. D. Chiang. 2000. Statistical Parsing with an Automatically-Extracted Tree Adjoining Grammar. In Proc. of ACL-2000. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. M. Collins and N. Duffy. 2001. Convolution kernels for natural language. In Proc. of the 14th NIPS.Google ScholarGoogle Scholar
  6. M. Collins and N. Duffy. 2002. New ranking algorithms for parsing and tagging: Kernels over discrete structures, and the voted perceptron. In Proc. of ACL 2002. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. M. Collins. 1999. Head-Driven Statistical Models for Natural Language Parsing. Ph.D. thesis, University of Pennsylvania. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. M. Collins. 2000. Discriminative reranking for natural language parsing. In Proc. of 7th ICML. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. M. Collins. 2001. Parameter estimation for statistical parsing models: Theory and practice of distribution-free methods. In Proc. of IWPT 2001. Invited Talk at IWPT 2001.Google ScholarGoogle Scholar
  10. R. Herbrich. 2002. Learning Kernel Classifiers: Theory and Algorithms. MIT Press. Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. A. K. Joshi and Y. Schabes. 1997. Tree-adjoining grammars. In G. Rozenberg and A. Salomaa, editors, Handbook of Formal Languages, volume 3, pages 69 - 124. Springer. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. L. Shen and A. K. Joshi. 2003. An SVM based voting algorithm with application to parse reranking. In Proc. of CoNLL 2003. Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. V. N. Vapnik. 1999. The Nature of Statistical Learning Theory. Springer, 2nd edition. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. F. Xia. 2001. Investigating the Relationship between Grammars and Treebanks for Natural Languages. Ph.D. thesis, University of Pennsylvania, Philadelphia, PA.Google ScholarGoogle Scholar

Recommendations

Comments

Login options

Check if you have access through your login credentials or your institution to get full access on this article.

Sign in
  • Published in

    cover image DL Hosted proceedings
    EMNLP '03: Proceedings of the 2003 conference on Empirical methods in natural language processing
    July 2003
    224 pages

    Publisher

    Association for Computational Linguistics

    United States

    Publication History

    • Published: 11 July 2003

    Qualifiers

    • Article

    Acceptance Rates

    Overall Acceptance Rate73of234submissions,31%

PDF Format

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader