Recognizing Geographical Locations using a GAN-Based Text-To-Image Approach
dc.contributor.author | Ibrahim, Dina M. | |
dc.contributor.author | Al-Shargabi, Amal A. | |
dc.date.acceptance | 2024-09-30 | |
dc.date.accessioned | 2024-10-30T14:46:42Z | |
dc.date.available | 2024-10-30T14:46:42Z | |
dc.date.issued | 2025 | |
dc.description | open access article | |
dc.description.abstract | Generating photo-realistic images that align with the text descriptions is the goal of the text-to-image generation (T2I) model. They can assist in visualizing the descriptions thanks to advancements in Machine Learning Algorithms. Using text as a source, Generative Adversarial Networks (GANs) can generate a series of pictures that serve as descriptions. Recent GANs have allowed oldest T2I models to achieve remarkable gains. However, they have some limitations. The main target of this study is to address these limitations to enhance the text-to-image generation models to enhance location services. To produce high-quality photos utilizing a multi-step approach, we build an attentional generating network called AttnGAN. The fine-grained image-text matching loss needed to train the AttnGAN’s generator is computed using our multimodal similarity model. With an inception score of 4.81 on the PatternNet dataset, our AttnGAN model achieves an impressive R-precision value of 70.61 percent. Because the PatternNet dataset comprises photographs, we’ve added verbal descriptions to each one to make it a text-based dataset instead. Many experiments have shown that AttnGAN’s proposed attention procedures, which are critical for text-to-image production in complex circumstances, are effective. | |
dc.funder | Other external funder (please detail below) | |
dc.funder.other | Qassim University | |
dc.identifier.citation | Ibrahim, Dina M. and Al-Shargabi, Amal A. (2025) Recognizing Geographical Locations using a GAN-Based Text-To-Image Approach. Indonesian Journal of Electrical Engineering and Computer Science | |
dc.identifier.issn | 2502-4752 | |
dc.identifier.uri | https://hdl.handle.net/2086/24437 | |
dc.language.iso | en | |
dc.peerreviewed | Yes | |
dc.projectid | 2023-SDG-1-BSRC36794 | |
dc.publisher | Institute of Advanced Engineering and Science (IAES) | |
dc.researchinstitute.institute | Institute of Digital Research, Communication and Responsible Innovation | |
dc.rights | Attribution 4.0 International | en |
dc.rights.uri | http://creativecommons.org/licenses/by/4.0/ | |
dc.subject | Text-to-Image | |
dc.subject | Text reading | |
dc.subject | Generative Adversarial Network; GANs | |
dc.subject | AttnGAN model | |
dc.subject | Location-based services | |
dc.subject | Road infrastructure | |
dc.subject | Deep learning | |
dc.title | Recognizing Geographical Locations using a GAN-Based Text-To-Image Approach | |
dc.type | Article |
Files
Original bundle
1 - 1 of 1
Loading...
- Name:
- Recognizing_Geographical_Locations_using_a_GAN_Based_Text_To_Image_Approach.pdf
- Size:
- 2.81 MB
- Format:
- Adobe Portable Document Format
License bundle
1 - 1 of 1
No Thumbnail Available
- Name:
- license.txt
- Size:
- 4.2 KB
- Format:
- Item-specific license agreed upon to submission
- Description: