Drishti Vaani - (Blind Assistance App)

Abstract

Blind individuals face significant challenges in navigating their surroundings independently. To address this, we propose an intelligent Blind Assistance System that integrates computer vision, machine learning, and IoT to provide real-time object recognition and environmental awareness through audio feedback. The system consists of a wearable camera that captures live video, an ML model that performs object detection and scene recognition, and an NLP-based voice assistant that converts detected objects into speech output. Additionally, an Arduino-based ultrasonic sensor detects nearby obstacles and triggers a buzzer for proximity alerts. The system is further enhanced by a Flutter-based mobile application, which utilizes a TensorFlow Lite (TFLite) MobileNet SSD model for live object detection and offers voice-controlled interactions. To improve accessibility, a Flask/FastAPI server hosts the scene detection model, allowing seamless integration into the mobile app. This low-cost, AI-powered assistive solution empowers visually impaired individuals by enhancing their spatial awareness, improving mobility, and promoting independent living.

Country : India

1 Shubham Shinde2 Neha Dubey3 Sakshi Gorave4 Arya More5 Prof. Manisha Hatkar

  1. Student, Smt. Indira Gandhi College of Engineering, Ghansoli, New Mumbai, Maharashtra, India
  2. Student, Smt. Indira Gandhi College of Engineering, Ghansoli, New Mumbai, Maharashtra, India
  3. Student, Smt. Indira Gandhi College of Engineering, Ghansoli, New Mumbai, Maharashtra, India
  4. Student, Smt. Indira Gandhi College of Engineering, Ghansoli, New Mumbai, Maharashtra, India
  5. Professor, Dept. of AI & ML, Smt. Indira Gandhi College of Engineering, Ghansoli, New Mumbai, Maharashtra, India

IRJIET, Volume 9, Issue 4, April 2025 pp. 27-32

doi.org/10.47001/IRJIET/2025.904005

References

  1. F. Catherine, Shiri Azenkot, Maya Cakmak, “Designing a Robot Guide for Blind People in Indoor Environments,” ACM/IEEE International Conference on Human-Robot Interaction Extended Abstracts, 2015.
  2. H. E. Chen, Y. Y. Lin, C. H. Chen, I. F. Wang, “Blindnavi: a mobile navigation app specially designed for the visually impaired,” ACM Conference Extended Abstracts on Human Factors in Computing Systems, 2015.
  3. K. W. Chen, C. H. Wang, X. Wei, Q. Liang, C. S. Chen, M. H. Yang, and Y. P. Hung, “Vision-based positioning for Internet-of-Vehicles,” IEEE Transactions on Intelligent Transportation Systems, vol. 18, no.2, pp. 364–376, 2016.
  4. M. Cordts, M. Omran, S. Ramos, T. Rehfeld, M. Enzweiler, R. Benenson, U. Franke,S. Roth, and B. Schiele, “The Cityscapes Dataset for Semantic Urban Scene Understanding,” IEEE Conference on Computer Vision and Pattern Recognition, 2016.
  5. Choi D., and Kim M. ’, “Trends on Object Detection Techniques Based on Deep Learning,” Electronics and Telecommunications Trends, Vol. 33, No. 4, pp. 23- 32, Aug. 2018.
  6. D. P. Khairnar, R. B. Karad, A. Kapse, G. Kale and P. Jadhav, "PARTHA: A Visually Impaired Assistance System," 2020 3rd International Conference on Communication System, Computing and IT Applications (CSCITA), 2020, pp. 32-37.