2018
DOI: 10.1007/978-3-030-02698-1_49
|View full text |Cite
|
Sign up to set email alerts
|

Intensive Positioning Network for Remote Sensing Image Captioning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
8
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 7 publications
(8 citation statements)
references
References 8 publications
0
8
0
Order By: Relevance
“…In Attention mechanism helps the encoderdecoder algorithm in concentrating on important parts of an image so the generated caption will be capturing main aspect of the image. This approach for caption generation is followed in [15,49,50,[61][62][63][64].…”
Section: Encoder-decoder Approachmentioning
confidence: 99%
See 2 more Smart Citations
“…In Attention mechanism helps the encoderdecoder algorithm in concentrating on important parts of an image so the generated caption will be capturing main aspect of the image. This approach for caption generation is followed in [15,49,50,[61][62][63][64].…”
Section: Encoder-decoder Approachmentioning
confidence: 99%
“…To avoid information loss while captioning an aerial image due to higher pixels and smaller target size, Intensive Positioning Network (IPN) is proposed in [61]. The algorithmic view of this model is shown in Figure 5.…”
Section: Encoder-decoder Approachmentioning
confidence: 99%
See 1 more Smart Citation
“…RSI captioning has been an interesting topic in RSI processing combined with the natural language processing [1], [11]- [13], [20], [28], [29]. According to the way of generating sentences, the main methods can be divided into three categories: templatebased methods, retrieval-based methods, and RNN-based methods.…”
Section: A Rsi Captioningmentioning
confidence: 99%
“…Considering the scale ambiguity, category ambiguity, and rotation ambiguity of RSI, Lu et al [13] explore the RSI captioning task based on both the hand-crafted features and deep features. A captioning method is proposed by Wang et al [28], which can provide more accurate object location via adding additional location and category tags to images. The information from the fully connected layer of CNN is treated as attribute information to improve the performance of captioning method based on convolutional features [29].…”
Section: A Rsi Captioningmentioning
confidence: 99%