2021
DOI: 10.3390/app12010317
|View full text |Cite
|
Sign up to set email alerts
|

Att-BiL-SL: Attention-Based Bi-LSTM and Sequential LSTM for Describing Video in the Textual Formation

Abstract: With the advancement of the technological field, day by day, people from around the world are having easier access to internet abled devices, and as a result, video data is growing rapidly. The increase of portable devices such as various action cameras, mobile cameras, motion cameras, etc., can also be considered for the faster growth of video data. Data from these multiple sources need more maintenance to process for various usages according to the needs. By considering these enormous amounts of video data, … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
2
0
1

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 12 publications
(3 citation statements)
references
References 53 publications
0
2
0
1
Order By: Relevance
“…Building on Bi-LSTM networks, integrating attention mechanisms 40 can further enhance the identification and modelling of critical features in time series (Fig. 2 ), thereby improving the temporal prediction task performance.…”
Section: Methodsmentioning
confidence: 99%
“…Building on Bi-LSTM networks, integrating attention mechanisms 40 can further enhance the identification and modelling of critical features in time series (Fig. 2 ), thereby improving the temporal prediction task performance.…”
Section: Methodsmentioning
confidence: 99%
“…LSTM memiliki kelemahan hanya memproses kata satu satu arah saja, yaitu dari awal sampai akhir [18]. Kami dalam penelitian ini mempertimbangkan untuk menggunakan bidirectional LSTM agar dapat membaca kata dari 2 arah, yaitu awal sampai akhir dan akhir sampai awal [19]. Sesuai arstitektur bidirectional LSTM pada gambar 4, bidirectional LSTM memilki 2 lapisan LSTM yang terpisah satu untuk maju dan satu untuk mundur sesuai dengan persamaan 7.…”
Section: Bidirectional Lstmunclassified
“…The disadvantage of these methods is the low efficiency of generating real and accurate sentences and the poor ability to generate structurally novel sentences. In recent research [7][8][9][10][11], visual and language information has been embedded into a common space via recurrent neural networks (RNNs) initially. Convolutional neural networks (CNNs) were then embedded within the visual space and combined with long short-term memory (LSTM) to produce more effective results.…”
Section: Introductionmentioning
confidence: 99%