Proceedings of the 3rd International Conference on Machine Learning and Soft Computing 2019
DOI: 10.1145/3310986.3311002
|View full text |Cite
|
Sign up to set email alerts
|

The Role of Attention Mechanism and Multi-Feature in Image Captioning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(2 citation statements)
references
References 20 publications
0
2
0
Order By: Relevance
“…From the overall hidden states of the recurrent layer, they derive variable specific hidden representations over time, which can be flexibility used for g-forecasting and temporal-variable level attentions. In his master's thesis, Lee [14] and Na et al [15,16] proposed a bidirectional Encoder-Decoder with dual-stage attention model that slightly modified a dual-stage attention-based recurrent neural network proposed by Qin and colleagues for multivariate time series prediction. In addition, he used the stock price transaction data of companies included in KODEX 200 to evaluate the performance of the proposed model.…”
Section: Introductionmentioning
confidence: 99%
“…From the overall hidden states of the recurrent layer, they derive variable specific hidden representations over time, which can be flexibility used for g-forecasting and temporal-variable level attentions. In his master's thesis, Lee [14] and Na et al [15,16] proposed a bidirectional Encoder-Decoder with dual-stage attention model that slightly modified a dual-stage attention-based recurrent neural network proposed by Qin and colleagues for multivariate time series prediction. In addition, he used the stock price transaction data of companies included in KODEX 200 to evaluate the performance of the proposed model.…”
Section: Introductionmentioning
confidence: 99%
“…For generating sentence descriptions for images, it is adapted to identify only image features relevant to generating words at each time step of the LSTM word generation sequence. Dang et al (2019) explored the importance of the attention mechanism in their work using two pretrained CNNs for multi-feature leaning. Comparing different architectures to test the effect of the attention mechanism, the results indicated that the attention mechanism improved performance significantly as the architecture with the attention layer performed better than the one without in terms of evaluation metrics.…”
Section: Attention Based Image Captioningmentioning
confidence: 99%