This thesis presents the development of a mobile application entitled »VIDA« (Visual Interpretation and Direction Assistant), designed as an assistive tool for blind and visually impaired individuals to enhance their independence and spatial orientation in daily life. The primary objective was to create a user-friendly and technically robust application that operates offline, supports the Slovenian language, and incorporates key real-time environment recognition functionalities.
The theoretical part of the thesis provides an in-depth analysis of existing assistive technologies, identifying critical limitations in terms of accessibility, functionality, and linguistic support. Based on these insights, »VIDA« was designed and implemented with integrated modules for object detection (YOLOv8), scene classification (ResNet), optical character recognition (OCR), and text-to-speech synthesis (TTS), all functioning in real time. We trained a custom computer vision model capable of reliable performance under diverse conditions, achieving a mean Average Precision (mAP50) of 0.48.
The application was implemented for the Android platform using the Model–View–ViewModel (MVVM) architecture and Jetpack Compose UI framework, providing high modularity, responsiveness, and accessibility for users with visual impairments. The solution was tested in collaboration with a blind user, confirming its usability, intuitiveness, and stability in real-world scenarios.
The results of this work demonstrate that modern computer vision and artificial intelligence approaches can be effectively applied to develop high-quality mobile solutions that significantly contribute to the inclusion and autonomy of blind and visually impaired individuals. The conclusion outlines future development directions, particularly in the integration of additional assistive functionalities.
|