Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing 2017
DOI: 10.18653/v1/d17-1201
|View full text |Cite
|
Sign up to set email alerts
|

Initializing Convolutional Filters with Semantic Features for Text Classification

Abstract: Convolutional Neural Networks (CNNs) are widely used in NLP tasks. This paper presents a novel weight initialization method to improve the CNNs for text classification. Instead of randomly initializing the convolutional filters, we encode semantic features into them, which helps the model focus on learning useful features at the beginning of the training. Experiments demonstrate the effectiveness of the initialization technique on seven text classification tasks, including sentiment analysis and topic classifi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
27
0

Year Published

2018
2018
2021
2021

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 50 publications
(28 citation statements)
references
References 21 publications
1
27
0
Order By: Relevance
“…They supposed words derived from an external resource, which can be divided into N group. And this is similar to Li's parameter initialization through clustering [29]. Our model is consistent with its purpose.…”
Section: Related Worksupporting
confidence: 69%
See 2 more Smart Citations
“…They supposed words derived from an external resource, which can be divided into N group. And this is similar to Li's parameter initialization through clustering [29]. Our model is consistent with its purpose.…”
Section: Related Worksupporting
confidence: 69%
“…Johnson et al [28] proposed a deep pyramid CNNs, they utilized the downsampling to decrease computational cost while efficiently representing long range associations in text. Li et al [29] employ initializing convolutional filters with features that computed by K-means rather than initializing filters randomly, which enable features to be efficiently captured. Zhang et al [30] proposed a grouped weight sharing way to instead of word embeddings.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…We compare our models with several state-of-the-art baseline models, including Bow+SVM (Chris Manning et al 2012), CNN (Kim 2014), CharCNN (Zhang, Zhao, and Le-Cun 2015), CNN-non-static+UNI (Li et al 2017), KPCNN (Wang et al 2017).…”
Section: Baseline Methodsmentioning
confidence: 99%
“…In this process, the n-grams are extracted from the train data and clustered by k-means. The experimental results show that the method effectively reduces the error caused by parameter adjustment [11]. Johnson et al propose a deep pyramid convolutional neural network model, which improves the classification accuracy by deepening the hierarchical structure of convolutional neural networks.…”
Section: Text Classification In General Domainmentioning
confidence: 99%