2021
DOI: 10.1016/j.neunet.2021.05.012
|View full text |Cite
|
Sign up to set email alerts
|

Anti-transfer learning for task invariance in convolutional neural networks for speech processing

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4

Citation Types

1
17
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
7
1
1

Relationship

1
8

Authors

Journals

citations
Cited by 14 publications
(18 citation statements)
references
References 51 publications
1
17
0
Order By: Relevance
“…A well-established solution to overcome the data scarcity in SER is transfer learning by weight initialization: network weights are initialized with values from a network that was pretrained with a different task, possibly on a different (usually large) dataset. Many variants of this method have been shown to improve the performance of SER models in limited-data scenarios and even when the task is rather distant from speech emotion [33]- [35]. Also, various data augmentation strategies have been successfully adopted for the same purpose, e.g.…”
Section: Introductionmentioning
confidence: 99%
“…A well-established solution to overcome the data scarcity in SER is transfer learning by weight initialization: network weights are initialized with values from a network that was pretrained with a different task, possibly on a different (usually large) dataset. Many variants of this method have been shown to improve the performance of SER models in limited-data scenarios and even when the task is rather distant from speech emotion [33]- [35]. Also, various data augmentation strategies have been successfully adopted for the same purpose, e.g.…”
Section: Introductionmentioning
confidence: 99%
“…A well-established solution to overcome the data scarcity in SER is transfer learning by weight initialization: network weights are initialized with values from a network that was pretrained with a different task, possibly on a different (usually large) dataset. Many variants of this method have been shown to improve the performance of SER models in limited-data scenarios and even when the task is rather distant from speech emotion [30]- [32]. Also, various data augmentation strategies have been successfully adopted for the same purpose, e.g.…”
Section: Introductionmentioning
confidence: 99%
“…During the heading and fowering stages, wheat has no evident changes in features other than the heading rate, which rapidly changes. Tus, better feature extraction and expression capacity of the regression model are required [26][27][28]. Convolutional neural network (CNN) has a high capacity for feature extraction and expression and has been widely used in image processing, video analysis, and other felds [29][30][31].…”
Section: Introductionmentioning
confidence: 99%