Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Confere 2015
DOI: 10.3115/v1/p15-1162
|View full text |Cite
|
Sign up to set email alerts
|

Deep Unordered Composition Rivals Syntactic Methods for Text Classification

Abstract: Many existing deep learning models for natural language processing tasks focus on learning the compositionality of their inputs, which requires many expensive computations. We present a simple deep neural network that competes with and, in some cases, outperforms such models on sentiment analysis and factoid question answering tasks while taking only a fraction of the training time. While our model is syntactically-ignorant, we show significant improvements over previous bag-of-words models by deepening our ne… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

13
590
3
2

Year Published

2017
2017
2024
2024

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 696 publications
(628 citation statements)
references
References 22 publications
13
590
3
2
Order By: Relevance
“…Upon word order, Four models in 'complex' group use the combination of neural networks to capture more complex information from textual data. Compared with neural models such as CNNs and RNNs, BOW models have the advantages of being efficient and robust, but are inferior to those complex models in accuracies [7]. NBSVM introduces supervised weighting schemes into bag-of-ngrams representation and provides strong baselines.…”
Section: Comparisons Of State-of-the-art Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Upon word order, Four models in 'complex' group use the combination of neural networks to capture more complex information from textual data. Compared with neural models such as CNNs and RNNs, BOW models have the advantages of being efficient and robust, but are inferior to those complex models in accuracies [7]. NBSVM introduces supervised weighting schemes into bag-of-ngrams representation and provides strong baselines.…”
Section: Comparisons Of State-of-the-art Methodsmentioning
confidence: 99%
“…Instead of constructing complex compositions upon word embeddings, these models basically ignore order and syntax information. Representative neural bagof-words models include Deep Averaging Network (DAN) [7] and Paragraph Vector (PV) [11]. DAN firstly takes the average of word embeddings as the inputs and then constructs multiple neural layers upon them.…”
Section: Neural Bag-of-words Modelsmentioning
confidence: 99%
“…There are also other works for regularizing classifiers by adding random noise to the data, such as dropout (Srivastava et al, 2014) and its variant for NLP tasks, word dropout (Iyyer et al, 2015). Xie et al (2017) discusses various data noising techniques for language models.…”
Section: Related Workmentioning
confidence: 99%
“…This feature network takes as input the average value of pretrained embeddings 3 for the tokens in the character description (we remove stopwords 4 ). This initial vector is passed through hidden layers to yield the feature embedding e (reminiscent of deep averaging networks by Iyyer et al (2015)). …”
Section: Action and Char Language Modelsmentioning
confidence: 99%