2021 IEEE International Conference on Robotics, Automation, Artificial-Intelligence and Internet-of-Things (RAAICON) 2021
DOI: 10.1109/raaicon54709.2021.9929536
|View full text |Cite
|
Sign up to set email alerts
|

Fine-Grained Image Generation from Bangla Text Description using Attentional Generative Adversarial Network

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
4
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 7 publications
(5 citation statements)
references
References 15 publications
0
4
0
Order By: Relevance
“…Although there have been many studies on text-to-image generation in English, very few have been applied to other languages. In [54], the use of Attn-GAN was proposed for generating fine-grained images based on descriptions in Bangla text. It is capable of integrating the most exact details in various subregions of the image, with a specific emphasis on the pertinent terms mentioned in the natural language description.…”
Section: A Text To Image Generation Using Gansmentioning
confidence: 99%
“…Although there have been many studies on text-to-image generation in English, very few have been applied to other languages. In [54], the use of Attn-GAN was proposed for generating fine-grained images based on descriptions in Bangla text. It is capable of integrating the most exact details in various subregions of the image, with a specific emphasis on the pertinent terms mentioned in the natural language description.…”
Section: A Text To Image Generation Using Gansmentioning
confidence: 99%
“…Attentional Generative Adversarial Network (AttnGAN) enables multi-stage, attention-driven image generation from textual description [7], [8]. AttnGAN begins with a rudimentary low-resolution image which it then refines in multiple phases to produce a final image from the natural language description.…”
Section: A Gansmentioning
confidence: 99%
“…Their proposed model showed substantial improvements in Text to image synthesis. Later on, the paper [7] suggested the first Bangla language-based Text-to-image generation method AttnGAN that analyzed Deep Attentional Multimodal Similarity Model and Attentional GAN to generate improved and realistic highresolution images from Bangla text description surpassing the state-of-the-art (SOTA) image synthesis GAN models by an ideal inception score of 3.58 ± .06.…”
Section: B Text To Image Synthesismentioning
confidence: 99%
“…Reed et al [8] suggested a method of translating singlesentenced text descriptions directly into image pixels by introducing a deep convolutional GAN based on text description embedding compressed using a fully connected layer and leaky-ReLU activation and text features used to perform feedforward inference by the generator and discriminator network fundamentally demonstrating enhanced text to image synthesis. In Paper [9], they proposed the use of the Bangla Attentional Generative Adversarial Network (AttnGAN) to generate highquality images from texts through multi-staged processing and incorporation of specific details in distinct parts of images, achieving an enhanced inception score on the CUB dataset.…”
Section: A Text To Image Generationmentioning
confidence: 99%