2021
DOI: 10.1016/j.cie.2021.107531
|View full text |Cite
|
Sign up to set email alerts
|

An end-to-end framework for remaining useful life prediction of rolling bearing based on feature pre-extraction mechanism and deep adaptive transformer model

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
21
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 71 publications
(21 citation statements)
references
References 44 publications
0
21
0
Order By: Relevance
“…To solve this problem, Hochreiter and Schmidhuber proposed a new network structure called LSTM (Hochreiter and Schmidhuber, 1997). Based on RNN, LSTM adds an information memory pathway and performs information update with the computation of three gates (forgetting gate, input gate and output gate) (Su et al , 2021). This novel structure effectively captures long-term dependencies by additionally introducing a flow of memory information.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…To solve this problem, Hochreiter and Schmidhuber proposed a new network structure called LSTM (Hochreiter and Schmidhuber, 1997). Based on RNN, LSTM adds an information memory pathway and performs information update with the computation of three gates (forgetting gate, input gate and output gate) (Su et al , 2021). This novel structure effectively captures long-term dependencies by additionally introducing a flow of memory information.…”
Section: Methodsmentioning
confidence: 99%
“…FCLNN and RNCAM do not require feature extraction and directly use the raw vibration signal of the bearing as the input, and the size of the input data is 2,560 × 1. The third model is the attention transformer (AT) proposed by (Su et al , 2021). For the first time, AT uses LSTM for adaptive location coding and uses the self-attentive mechanism for feature extraction.…”
Section: Experimental Verificationmentioning
confidence: 99%
“…Ma et al [53] created a variation of the G-Transformer model architecture that uses the encoder from traditional Transformer models as it is applied to natural language processing for sampling and extracting features for PM. The Adaptive Transformer (AT) is a modified deep attention architecture that handles temporal data of the low-level features related to RUL to minimize the recurring problem of vanishing gradients in prediction [117].…”
Section: ) Transformermentioning
confidence: 99%
“…[26,27,38,55,58,59,65,92,93,94,95,96,97,98,99,100,101,102,103,104,105,106,107,108,109,110,111,112,113,114,115,116,117,118,119,120,121,122,123,124,125,126,127,128] …”
mentioning
confidence: 99%
“…Ma et al [53] created a variation of the G-Transformer model architecture that uses the encoder from traditional Transformer models as it is applied to natural language processing for sampling and extracting features for PM. The Adaptive Transformer (AT) is a modified deep attention architecture that handles temporal data of the low-level features related to RUL to minimize the recurring problem of vanishing gradients in prediction [132].…”
Section: ) Transformermentioning
confidence: 99%