Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing 2016
DOI: 10.18653/v1/d16-1157
|View full text |Cite
|
Sign up to set email alerts
|

Charagram: Embedding Words and Sentences via Character n-grams

Abstract: We present CHARAGRAM embeddings, a simple approach for learning character-based compositional models to embed textual sequences. A word or sentence is represented using a character n-gram count vector, followed by a single nonlinear transformation to yield a low-dimensional embedding. We use three tasks for evaluation: word similarity, sentence similarity, and part-of-speech tagging. We demonstrate that CHARAGRAM embeddings outperform more complex architectures based on character-level recurrent and convolutio… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
121
0

Year Published

2017
2017
2022
2022

Publication Types

Select...
5
5

Relationship

0
10

Authors

Journals

citations
Cited by 141 publications
(121 citation statements)
references
References 40 publications
(41 reference statements)
0
121
0
Order By: Relevance
“…These representations are robust to out-of-vocabulary items, often producing improved results. Our pretraining procedure is reminiscent of several recent papers (Wieting et al, 2016, inter alia) who aim for general purpose character n-gram embeddings. In contrast, we pretrain all model parameters on automatic but in-domain paraphrase data.…”
Section: Related Workmentioning
confidence: 99%
“…These representations are robust to out-of-vocabulary items, often producing improved results. Our pretraining procedure is reminiscent of several recent papers (Wieting et al, 2016, inter alia) who aim for general purpose character n-gram embeddings. In contrast, we pretrain all model parameters on automatic but in-domain paraphrase data.…”
Section: Related Workmentioning
confidence: 99%
“…Other ways of representing words with embeddings have to be experimented with, especially the ones where word and character-level representations are mixed, like Charagram (Wieting et al, 2016). It is also interesting to see, whether this metric can be used for hill-climbing and system development.…”
Section: Discussionmentioning
confidence: 99%
“…There is a large literature on exploiting characters, morphology, and composition for embedding models (Chen et al, 2015;Ling et al, 2015a;Qiu et al, 2014;Wieting et al, 2016;Lazaridou et al, 2013), and a comparison with these different models may be interesting future work.…”
Section: Morphological Word Embeddingsmentioning
confidence: 99%