2019 International Conference on Asian Language Processing (IALP) 2019
DOI: 10.1109/ialp48816.2019.9037668
|View full text |Cite
|
Sign up to set email alerts
|

Using Convolutional Neural Network with BERT for Intent Determination

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
15
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
6
3
1

Relationship

0
10

Authors

Journals

citations
Cited by 23 publications
(15 citation statements)
references
References 12 publications
0
15
0
Order By: Relevance
“…For some NLP applications, however, a language model by itself is not sufficient for accomplishing a given downstream task, and it becomes necessary to expand the language model's overall architecture by stacking it with another form of neural network, for example using a convolutional neural network for language models targeting classification NLP tasks. For such application scenarios, the combination of the BERT language model and deep learning models such as recurrent neural networks or convolutional neural networks were shown to be effective in recent studies for capturing meaningful features from the available data [8,[38][39][40]. We utilize a similar approach for our ClaimsBERT classifier.…”
Section: Related Workmentioning
confidence: 99%
“…For some NLP applications, however, a language model by itself is not sufficient for accomplishing a given downstream task, and it becomes necessary to expand the language model's overall architecture by stacking it with another form of neural network, for example using a convolutional neural network for language models targeting classification NLP tasks. For such application scenarios, the combination of the BERT language model and deep learning models such as recurrent neural networks or convolutional neural networks were shown to be effective in recent studies for capturing meaningful features from the available data [8,[38][39][40]. We utilize a similar approach for our ClaimsBERT classifier.…”
Section: Related Workmentioning
confidence: 99%
“…There have been plenty of researches on conventional neural network methods in the last few decades (Xu and Sarikaya 2013;Liu and Lane 2016;Haihong et al 2019;Wang et al 2020a;Gerz et al 2021). During recent years, with the rapid development of computing power, pre-trained models such as BERT (Devlin et al 2018) are employed for intent detection frequently (Castellucci et al 2019;He et al 2019;Zhang, Zhang, and Chen 2019;Athiwaratkun et al 2020;Gong et al 2021). Confidence Calibration.…”
Section: Related Workmentioning
confidence: 99%
“…CNN is used instead of other typical deep neural networks such as LSTM [32], Bi-LSTM [33], and GRU [34] since it is currently the most successful model for addressing short text classification tasks [35]. The convolution and pooling techniques of CNN aid in the extraction of the main concepts and keywords of the text as features, resulting in a significant improvement in the performance of the classification model.…”
Section: Existing Hsd Modelsmentioning
confidence: 99%