Abstract
The high-dimensionality of lexical features in parsing can be memory consuming and cause over-fitting problems. We propose a general framework to replace all lexical feature templates by low-dimensional features induced from word embeddings. Applied to a near state-of-the-art dependency parser (Huang et al., 2012), our method improves the baseline, performs better than using cluster bit string features, and outperforms a recent neural network based parser. A further analysis shows that our framework has the effect hypothesized by Andreas and Klein (2014), namely (i) connecting unseen words to known ones, and (ii) encouraging common behaviors among invocabulary words.
Original language | English |
---|---|
Pages | 106-113 |
Number of pages | 8 |
Publication status | Published - 2015 Jan 1 |
Event | 29th Pacific Asia Conference on Language, Information and Computation, PACLIC 2015 - Shanghai, China Duration: 2015 Oct 30 → 2015 Nov 1 |
Other
Other | 29th Pacific Asia Conference on Language, Information and Computation, PACLIC 2015 |
---|---|
Country/Territory | China |
City | Shanghai |
Period | 15/10/30 → 15/11/1 |
ASJC Scopus subject areas
- Artificial Intelligence
- Human-Computer Interaction
- Linguistics and Language