Proceedings of the Eighteenth Conference on Computational Natural Language Learning 2014
DOI: 10.3115/v1/w14-1613
|View full text |Cite
|
Sign up to set email alerts
|

Distributed Word Representation Learning for Cross-Lingual Dependency Parsing

Abstract: This paper proposes to learn languageindependent word representations to address cross-lingual dependency parsing, which aims to predict the dependency parsing trees for sentences in the target language by training a dependency parser with labeled sentences from a source language. We first combine all sentences from both languages to induce real-valued distributed representation of words under a deep neural network architecture, which is expected to capture semantic similarities of words not only within the sa… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

4
63
0

Year Published

2016
2016
2024
2024

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 68 publications
(67 citation statements)
references
References 15 publications
4
63
0
Order By: Relevance
“…Previous work has shown its effectiveness across a wide range of multilingual transfer tasks including tagging (Kim et al, 2015), syntactic parsing (Xiao and Guo, 2014;Guo et al, 2015;Durrett et al, 2012), and machine translation (Zou et al, 2013;Mikolov et al, 2013b). However, these approaches commonly require parallel sentences or bilingual lexicon to learn multilingual embeddings.…”
Section: Multilingual Word Embeddingsmentioning
confidence: 99%
“…Previous work has shown its effectiveness across a wide range of multilingual transfer tasks including tagging (Kim et al, 2015), syntactic parsing (Xiao and Guo, 2014;Guo et al, 2015;Durrett et al, 2012), and machine translation (Zou et al, 2013;Mikolov et al, 2013b). However, these approaches commonly require parallel sentences or bilingual lexicon to learn multilingual embeddings.…”
Section: Multilingual Word Embeddingsmentioning
confidence: 99%
“…In addition to having a direct application in inherently crosslingual tasks like machine translation (Zou et al, 2013) and crosslingual entity linking (Tsai and Roth, 2016), they provide an excellent mechanism for transfer learning, where a model trained in a resource-rich language is transferred to a less-resourced one, as shown with part-of-speech tagging , parsing (Xiao and Guo, 2014) and document classification (Klementiev et al, 2012).…”
Section: Introductionmentioning
confidence: 99%
“…Bilingual word representations could serve as an useful source knowledge for problems in cross-lingual information retrieval (Levow, Oard, & Resnik, 2005;Vulić, De Smet, & Moens, 2013), statistical machine translation (Wu, Wang, & Zong, 2008), document classification (Ni, Sun, Hu, & Chen, 2011;Klementiev et al, 2012;Hermann & Blunsom, 2014b;Chandar, Lauly, Larochelle, Khapra, Ravindran, Raykar, & Saha, 2014;Vulić, De Smet, Tang, & Moens, 2015), bilingual lexicon extraction (Tamura, Watanabe, & Sumita, 2012;Vulić & Moens, 2013a), or knowledge transfer and annotation projection from resource-rich to resource-poor languages for a myriad of NLP tasks such as dependency parsing, POS tagging, semantic role labeling or selectional preferences (Yarowsky & Ngai, 2001;Padó & Lapata, 2009;Peirsman & Padó, 2010;Das & Petrov, 2011;Täckström, Das, Petrov, McDonald, & Nivre, 2013;Ganchev & Das, 2013;Tiedemann, Agić, & Nivre, 2014;Xiao & Guo, 2014). Other interesting application domains are machine translation (e.g., Zou, Socher, Cer, & Manning, 2013;Wu, Dong, Hu, Yu, He, Wu, Wang, & Liu, 2014;Zhang, Liu, Li, Zhou, & Zong, 2014) and cross-lingual information retrieval (e.g., .…”
Section: Bilingual Word Embeddingsmentioning
confidence: 99%
“…We may cluster the current work in three different groups: (1) the models that rely on hard word alignments obtained from parallel data to constrain the learning of BWEs (Klementiev et al, 2012;Zou et al, 2013;Wu et al, 2014); (2) the models that use the alignment of parallel data at the sentence level (Kočiský, Hermann, & Blunsom, 2014;Hermann & Blunsom, 2014aChandar et al, 2014;Shi, Liu, Liu, & Sun, 2015;; (3) the models that critically require readily available bilingual lexicons (Mikolov et al, 2013b;Faruqui & Dyer, 2014;Xiao & Guo, 2014). The main disadvantage of all these models is the limited availability of parallel data and bilingual lexicons, resources which are scarce and/or domain-restricted for plenty of language pairs.…”
Section: Bilingual Word Embeddingsmentioning
confidence: 99%