<p>According to the World Health Organization (WHO), more than hundreds of million people have some type of visual disability today. Vision impairment has a personal impact with lifelong consequences because more than 80\% of our perception, cognition, learning, and daily activities are mediated through vision. As a result, visually impaired and blind people stagnate at work and in education due to inaccessible technologies. In this regard, we present a speech-based assistive system for visually impaired and blind people to interact with ChatGPT. The system employs automatic speech recognition (ASR), text-to-speech (TTS), and a messenger bot (Telegram) to enable a natural interface to ChatGPT for people with visual disabilities. Our preliminary experiments showed the feasibility of the proposed system as an assistive technology. We have made the source code publicly available at https://github.com/IS2AI/telegram-bot-chatgpt to stimulate the development of assistive systems for the visually impaired and blind.</p>
<p>According to the World Health Organization (WHO), more than hundreds of million people have some type of visual disability today. Vision impairment has a personal impact with lifelong consequences because more than 80\% of our perception, cognition, learning, and daily activities are mediated through vision. As a result, visually impaired and blind people stagnate at work and in education due to inaccessible technologies. In this regard, we present a speech-based assistive system for visually impaired and blind people to interact with ChatGPT. The system employs automatic speech recognition (ASR), text-to-speech (TTS), and a messenger bot (Telegram) to enable a natural interface to ChatGPT for people with visual disabilities. Our preliminary experiments showed the feasibility of the proposed system as an assistive technology. We have made the source code publicly available at https://github.com/IS2AI/telegram-bot-chatgpt to stimulate the development of assistive systems for the visually impaired and blind.</p>
<p>Visually impaired and blind people often face a range of socioeconomic problems that can make it difficult for them to live independently and participate fully in society. Advances in machine learning pave new venues to implement assistive devices for the visually impaired and blind. In this work, we combined image captioning and text-to-speech technologies to create an assistive device for the visually impaired and blind. Our system can provide the user with descriptive auditory feedback in the Kazakh language on a scene acquired in real-time by a head-mounted camera. The image captioning model for the Kazakh language provided satisfactory results in both quantitative metrics and subjective evaluation. Finally, experiments with a visually unimpaired blindfolded participant demonstrated the feasibility of our approach.</p>
<p>Visually impaired and blind people often face a range of socioeconomic problems that can make it difficult for them to live independently and participate fully in society. Advances in machine learning pave new venues to implement assistive devices for the visually impaired and blind. In this work, we combined image captioning and text-to-speech technologies to create an assistive device for the visually impaired and blind. Our system can provide the user with descriptive auditory feedback in the Kazakh language on a scene acquired in real-time by a head-mounted camera. The image captioning model for the Kazakh language provided satisfactory results in both quantitative metrics and subjective evaluation. Finally, experiments with a visually unimpaired blindfolded participant demonstrated the feasibility of our approach.</p>
<p>Visually impaired and blind people often face a range of socioeconomic problems that can make it difficult for them to live independently and participate fully in society. Advances in machine learning pave new venues to implement assistive devices for the visually impaired and blind. In this work, we combined image captioning and text-to-speech technologies to create an assistive device for the visually impaired and blind. Our system can provide the user with descriptive auditory feedback in the Kazakh language on a scene acquired in real-time by a head-mounted camera. The image captioning model for the Kazakh language provided satisfactory results in both quantitative metrics and subjective evaluation. Finally, experiments with a visually unimpaired blindfolded participant demonstrated the feasibility of our approach.</p>
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
customersupport@researchsolutions.com
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
Copyright © 2025 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.