2023
DOI: 10.1007/s11042-023-14727-0
|View full text |Cite
|
Sign up to set email alerts
|

A lung sound recognition model to diagnoses the respiratory diseases by using transfer learning

Abstract: Respiratory disease is one of the leading causes of death in the world. Through advances in Artificial Intelligence, it appears possible for the days of misdiagnosis and treatment of respiratory disease symptoms rather than their root cause to move behind us. The traditional convolutional neural network cannot extract the temporal features of lung sounds. To solve the problem, a lung sounds recognition algorithm based on VGGish- stacked BiGRU is proposed which combines the VGGish network with the stacked bidir… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
4
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 10 publications
(4 citation statements)
references
References 22 publications
0
4
0
Order By: Relevance
“…Specifically, it computes the gradients of the target concept (i.e., the class output) with respect to the feature maps of the last convolutional layer. By pooling these gradients over the spatial dimensions, Grad-CAM produces a coarse localization map that highlights the parts of the image that have the greatest influence on CNN’s decision [ 34 , 38 ]. The architecture explaining the Grad-CAM technique is shown in Figure S2 .…”
Section: Resultsmentioning
confidence: 99%
See 3 more Smart Citations
“…Specifically, it computes the gradients of the target concept (i.e., the class output) with respect to the feature maps of the last convolutional layer. By pooling these gradients over the spatial dimensions, Grad-CAM produces a coarse localization map that highlights the parts of the image that have the greatest influence on CNN’s decision [ 34 , 38 ]. The architecture explaining the Grad-CAM technique is shown in Figure S2 .…”
Section: Resultsmentioning
confidence: 99%
“…Finally, the logarithm of the amplitude values (in dB) is taken to mimic the human ear’s logarithmic perception of loudness. The resulting LMS displays the frequency content in mels on one axis and time on the other, with the amplitude represented by a logarithmically scaled color map [ 38 ]. In this research, LMS representations were computed using 128 ms (2048 samples) window lengths and 32 ms (512 samples) hop lengths for the STFT, with examples provided in the referenced Figure 4 .…”
Section: Methodsmentioning
confidence: 99%
See 2 more Smart Citations