Visually impaired people are handicapped in perceiving and acting within their environment based on limited contextual information. Conventional aid tools like white canes and GPS devices give incomplete solutions but cannot offer real-time descriptive directions. With advances in multimodal artificial intelligence (Al), especially vision-language models, it is now feasible to develop smart, chat-like systems that combine vision, language, and speech. This paper introduces VISOR - Guiding Shield for Vision, a multimodal AI-driven assistive system that processes visual input with computer vision, narrates scenes based on vision-language models, and provides natural audio output through speech synthesis. With its integration of object detection, visual question answering, and speech interfaces, VIBE provides context-aware, real-time assistance to visually impaired users. The architecture is lightweight, portable, and modular, which enables the cost-effective solution with scalability for the future.
VISOR: An AI-Powered Guiding Shield for Vision
Shreya Shinde,Sneh Patel,Supriya V. Mahadevkar,Archana Y. Chaudhari,Anilkumar Gupta
Published 2025 in 2025 Modern Electronics Devices and Intelligent Communication Systems (MEDCOM)
ABSTRACT
PUBLICATION RECORD
- Publication year
2025
- Venue
2025 Modern Electronics Devices and Intelligent Communication Systems (MEDCOM)
- Publication date
2025-12-11
- Fields of study
Not labeled
- Identifiers
- External record
- Source metadata
Semantic Scholar
CITATION MAP
EXTRACTION MAP
CLAIMS
- No claims are published for this paper.
CONCEPTS
- No concepts are published for this paper.
REFERENCES
Showing 1-8 of 8 references · Page 1 of 1
CITED BY
- No citing papers are available for this paper.
Showing 0-0 of 0 citing papers · Page 1 of 1