2020
DOI: 10.3390/ijgi9040194
|View full text |Cite
|
Sign up to set email alerts
|

Landslide Image Captioning Method Based on Semantic Gate and Bi-Temporal LSTM

Abstract: When a landslide happens, it is important to recognize the hazard-affected bodies surrounding the landslide for the risk assessment and emergency rescue. In order to realize the recognition, the spatial relationship between landslides and other geographic objects such as residence, roads and schools needs to be defined. Comparing with semantic segmentation and instance segmentation that can only recognize the geographic objects separately, image captioning can provide richer semantic information including the … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
16
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
8

Relationship

0
8

Authors

Journals

citations
Cited by 15 publications
(16 citation statements)
references
References 69 publications
0
16
0
Order By: Relevance
“…Thus, Bidirectional-LSTM was employed to learn the factor data of forward and backward information. This Bi-LSTM aimed to model the inter-relationship among the factors from sequential observed data [ 44 , 45 ].…”
Section: Methodsmentioning
confidence: 99%
“…Thus, Bidirectional-LSTM was employed to learn the factor data of forward and backward information. This Bi-LSTM aimed to model the inter-relationship among the factors from sequential observed data [ 44 , 45 ].…”
Section: Methodsmentioning
confidence: 99%
“…Furthermore, J. Aneja [21] and S. J. Rennie [25] adapted the attention mechanism to generate caption. For vision part of image captioning Vgg16 were used by most of the papers [2], [11], [24], [25], [27], [30] as CNN but some of them also used YOLO [9], Inception V3 [6], [31], AlexNet [24], [30] ResNet [11], [18], [24] or Unet [4] as CNN for feature extraction. Concurrently, LSTM [6], [9], [11], [17], [31] was used by most of the papers for generating the next word in the sequence.…”
Section: Related Workmentioning
confidence: 99%
“…Concurrently, LSTM [6], [9], [11], [17], [31] was used by most of the papers for generating the next word in the sequence. However, some of the researcher also utilized RNN [19] or BiLSTM [4], [30]. Moreover, P. Blandfort et al [32] systematically characterize diverse image captions that appear "in the wild" in order to understand how people caption images naturally.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Similarly, a stacked LSTM model was developed by stacking LSTM layers to forecast the soil movements (Xing et al, 2019). Beyond some of these attempts for forecasting soil movements, there have been some attempts at developing RNN models for the time series forecasting problems across different domains (Huang et al, 2015;Behera et al, 2018Behera et al, , 2021bQiu et al, 2018;Zhang et al, 2019;Barzegar et al, 2020;Cui et al, 2020;Singh et al, 2020). For example, convolutional LSTM (Conv-LSTM), bidirectional LSTM (Bi-LSTM), and CNN-LSTM models have been developed in the natural language processing (NLP), crowd time series forecasting, software reliability assessment, and water quality variable forecasting (Huang et al, 2015;Behera et al, 2018Behera et al, , 2021bQiu et al, 2018;Zhang et al, 2019;Barzegar et al, 2020;Cui et al, 2020;Singh et al, 2020).…”
Section: Introductionmentioning
confidence: 99%