DRISHTI: Visual Navigation Assistant for Visually Impaired
- URL: http://arxiv.org/abs/2303.07451v1
- Date: Mon, 13 Mar 2023 20:10:44 GMT
- Title: DRISHTI: Visual Navigation Assistant for Visually Impaired
- Authors: Malay Joshi, Aditi Shukla, Jayesh Srivastava, Manya Rastogi
- Abstract summary: Blind and visually impaired (BVI) people face challenges because they need manual support to prompt information about their environment.
In this work, we took our first step towards developing an affordable and high-performing eye wearable assistive device, DRISHTI.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In today's society, where independent living is becoming increasingly
important, it can be extremely constricting for those who are blind. Blind and
visually impaired (BVI) people face challenges because they need manual support
to prompt information about their environment. In this work, we took our first
step towards developing an affordable and high-performing eye wearable
assistive device, DRISHTI, to provide visual navigation assistance for BVI
people. This system comprises a camera module, ESP32 processor, Bluetooth
module, smartphone and speakers. Using artificial intelligence, this system is
proposed to detect and understand the nature of the users' path and obstacles
ahead of the user in that path and then inform BVI users about it via audio
output to enable them to acquire directions by themselves on their journey.
This first step discussed in this paper involves establishing a
proof-of-concept of achieving the right balance of affordability and
performance by testing an initial software integration of a currency detection
algorithm on a low-cost embedded arrangement. This work will lay the foundation
for our upcoming works toward achieving the goal of assisting the maximum of
BVI people around the globe in moving independently.
Related papers
- NaVIP: An Image-Centric Indoor Navigation Solution for Visually Impaired People [12.230718190579724]
NaVIP aims to create an image-centric indoor navigation and exploration solution for inclusiveness.
We start by curating large-scale phone camera data in a four-floor research building, with 300K images.
Every image is labelled with precise 6DoF camera poses, details of indoor PoIs, and descriptive captions.
arXiv Detail & Related papers (2024-10-08T21:16:50Z) - Floor extraction and door detection for visually impaired guidance [78.94595951597344]
Finding obstacle-free paths in unknown environments is a big navigation issue for visually impaired people and autonomous robots.
New devices based on computer vision systems can help impaired people to overcome the difficulties of navigating in unknown environments in safe conditions.
In this work it is proposed a combination of sensors and algorithms that can lead to the building of a navigation system for visually impaired people.
arXiv Detail & Related papers (2024-01-30T14:38:43Z) - Camera-Radar Perception for Autonomous Vehicles and ADAS: Concepts,
Datasets and Metrics [77.34726150561087]
This work aims to carry out a study on the current scenario of camera and radar-based perception for ADAS and autonomous vehicles.
Concepts and characteristics related to both sensors, as well as to their fusion, are presented.
We give an overview of the Deep Learning-based detection and segmentation tasks, and the main datasets, metrics, challenges, and open questions in vehicle perception.
arXiv Detail & Related papers (2023-03-08T00:48:32Z) - AVLEN: Audio-Visual-Language Embodied Navigation in 3D Environments [60.98664330268192]
We present AVLEN -- an interactive agent for Audio-Visual-Language Embodied Navigation.
The goal of AVLEN is to localize an audio event via navigating the 3D visual world.
To realize these abilities, AVLEN uses a multimodal hierarchical reinforcement learning backbone.
arXiv Detail & Related papers (2022-10-14T16:35:06Z) - Detect and Approach: Close-Range Navigation Support for People with
Blindness and Low Vision [13.478275180547925]
People with blindness and low vision (pBLV) experience significant challenges when locating final destinations or targeting specific objects in unfamiliar environments.
We develop a novel wearable navigation solution to provide real-time guidance for a user to approach a target object of interest efficiently and effectively in unfamiliar environments.
arXiv Detail & Related papers (2022-08-17T18:38:20Z) - Augmented reality navigation system for visual prosthesis [67.09251544230744]
We propose an augmented reality navigation system for visual prosthesis that incorporates a software of reactive navigation and path planning.
It consists on four steps: locating the subject on a map, planning the subject trajectory, showing it to the subject and re-planning without obstacles.
Results show how our augmented navigation system help navigation performance by reducing the time and distance to reach the goals, even significantly reducing the number of obstacles collisions.
arXiv Detail & Related papers (2021-09-30T09:41:40Z) - VisBuddy -- A Smart Wearable Assistant for the Visually Challenged [0.0]
VisBuddy is a voice-based assistant, where the user can give voice commands to perform specific tasks.
It uses the techniques of image captioning for describing the user's surroundings, optical character recognition (OCR) for reading the text in the user's view, object detection to search and find the objects in a room and web scraping to give the user the latest news.
arXiv Detail & Related papers (2021-08-17T17:15:23Z) - Deep Learning for Embodied Vision Navigation: A Survey [108.13766213265069]
"Embodied visual navigation" problem requires an agent to navigate in a 3D environment mainly rely on its first-person observation.
This paper attempts to establish an outline of the current works in the field of embodied visual navigation by providing a comprehensive literature survey.
arXiv Detail & Related papers (2021-07-07T12:09:04Z) - Active Visual Information Gathering for Vision-Language Navigation [115.40768457718325]
Vision-language navigation (VLN) is the task of entailing an agent to carry out navigational instructions inside photo-realistic environments.
One of the key challenges in VLN is how to conduct a robust navigation by mitigating the uncertainty caused by ambiguous instructions and insufficient observation of the environment.
This work draws inspiration from human navigation behavior and endows an agent with an active information gathering ability for a more intelligent VLN policy.
arXiv Detail & Related papers (2020-07-15T23:54:20Z) - DeFINE: Delayed Feedback based Immersive Navigation Environment for
Studying Goal-Directed Human Navigation [10.7197371210731]
Delayed Feedback based Immersive Navigation Environment (DeFINE) is a framework that allows for easy creation and administration of navigation tasks.
DeFINE has a built-in capability to provide performance feedback to participants during an experiment.
arXiv Detail & Related papers (2020-03-06T11:00:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.