Proceedings of the First Workshop on Subword and Character Level Models in NLP 2017
DOI: 10.18653/v1/w17-4101
|View full text |Cite
|
Sign up to set email alerts
|

Character and Subword-Based Word Representation for Neural Language Modeling Prediction

Abstract: Most of neural language models use different kinds of embeddings for word prediction. While word embeddings can be associated to each word in the vocabulary or derived from characters as well as factored morphological decomposition, these word representations are mainly used to parametrize the input, i.e. the context of prediction. This work investigates the effect of using subword units (character and factored morphological decomposition) to build output representations for neural language modeling. We presen… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2018
2018
2025
2025

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 10 publications
(4 citation statements)
references
References 36 publications
0
4
0
Order By: Relevance
“…The structure of these subword sequences is especially important towards the representation of nonverbal dynamics. In addition, modeling subword information has become essential for various tasks in natural language processing (Faruqui et al 2017), including language modeling (Labeau and Allauzen 2017;), learning word representations for different languages (Peters et al 2018;Oh et al 2018;Bojanowski et al 2016), and machine translation (Kudo 2018;Sennrich, Haddow, and Birch 2015). However, many of these previous works in understanding and modeling multimodal language has ignored the role of subword analysis.…”
Section: Introductionmentioning
confidence: 99%
“…The structure of these subword sequences is especially important towards the representation of nonverbal dynamics. In addition, modeling subword information has become essential for various tasks in natural language processing (Faruqui et al 2017), including language modeling (Labeau and Allauzen 2017;), learning word representations for different languages (Peters et al 2018;Oh et al 2018;Bojanowski et al 2016), and machine translation (Kudo 2018;Sennrich, Haddow, and Birch 2015). However, many of these previous works in understanding and modeling multimodal language has ignored the role of subword analysis.…”
Section: Introductionmentioning
confidence: 99%
“…Secondly, we would like to introduce characterlevel information in our models, based on (Wieting et al, 2016;Labeau and Allauzen, 2017), in order to overcome the problem of out-of-vocabulary (OOV) words and learn syntactic and stylistic features (Peters et al, 2018), which are highly indicative of emotions and their intensity.…”
Section: Discussionmentioning
confidence: 99%
“…However, this naive and standard way ignores the language information of the text. In order to effectively resort to linguistic information for scene text recognition, we incorporate subword [75] tokenization mechanism in NLP [76] into the text recognition method. Subword tokenization algorithms aim to decompose rare words into meaningful subwords and remain frequently used words, so that the grammatical information of word has already been captured in the subwords.…”
Section: Multi-granularity Predictionsmentioning
confidence: 99%