EMBEDDED VOICE ASSISTANCE MODULE FOR THE VISUALLY IMPAIRED BASED ON VISUAL SCENE ANALYSIS

Authors

DOI:

https://doi.org/10.35546/kntu2078-4481.2025.3.2.3

Keywords:

Raspberry Pi, YOLOv8, Tesseract, OpenCV, OCR, pyttsx3, intelligent module, visually impaired, hardware implementation, public transport

Abstract

The article discusses a hardware module designed to assist visually impaired individuals at public transport stops. The relevance of this research is determined by the increasing number of people with visual impairments in Ukraine, caused both by global trends and by the consequences of military actions that have led to eye injuries among both military personnel and civilians. Since urban infrastructure is predominantly oriented toward visual perception, visually impaired individuals face significant difficulties in orientation, especially at public transport stops. The use of an autonomous device that does not depend on a smartphone or Internet access is of great importance for ensuring mobility and safety in urban environments.The purpose of the study is to develop and analyze the effectiveness of a hardware-software complex based on Raspberry Pi 5, which provides automatic recognition of public transport types (bus, trolleybus, tram) and their route numbers, followed by real-time voice output of the results. The hardware part includes a Raspberry Pi 5 single-board computer with active cooling, a Logitech Web Camera Carl Zeiss Tessar 2.0/3.7 2MP connected via USB 3.0, a Logitech G Pro X external sound card for audio output through headphones, and a portable power bank to ensure autonomous operation.Experimental studies have confirmed high detection accuracy of public transport in daytime conditions and acceptable performance on the embedded platform. At the same time, under low-visibility conditions (twilight, fog, image blurring), a decrease in classification accuracy of transport types and recognition of route numbers was observed. It has been determined that further system development may include expanding the dataset with images collected under real weather conditions, improving video preprocessing algorithms (low-light enhancement and denoising), as well as integrating with mobile applications and GPS navigation. The obtained results demonstrate the practical applicability of the developed solution and its potential for scaling to improve the inclusiveness and mobility of visually impaired individuals.

References

Нечітайло О. В., Гаврашенко А. О. Інтелектуальний модуль навігації для реабілітації осіб з вадами зору на зупинках громадського транспорту. Сучасні напрями розвитку інформаційно-комунікаційних технологій та засобів управління : тези доп. 15-ї міжнар. наук.-техн. конф. (24–25 квіт. 2025 р., Баку – Харків – Жиліна). Т. 3. Харків ; Баку ; Жиліна, 2025. С. 123.

Barkovska O., Serdechnyi V. Development of a universal model of an intellectual system to help people with visual impairments. Information Technology and Security. 2024. Vol. 12, No. 2. P. 47–56.

Kumar A., Ahmed S., Lee J. MR.NAVI: Mixed Reality Navigation Assistance for Visually Impaired Using Computer Vision and Natural Language Processing. 2025.

Shah P., Patel K., Reddy S. Smart Real-Time Object Detection for Blind Individuals Using YOLOv8 and OCR. Journal of Emerging Technologies. 2024. Vol. 18, No. 3. P. 115–123.

Mahmud R., Zhang T., Li Y. VISA: A Vision-Based Spatial Awareness System for Indoor Navigation of Visually Impaired. Journal of Imaging. 2024. Vol. 11, No. 1. P. 9.

Rafiq M., Hussain A., Khan N. Navigation Technologies for Visually Impaired: A Systematic Review (2019–2024). Sensors. 2024. Vol. 25, No. 7. P. 2213.

Published

2025-11-28