Assisting Blind People Using Object Detection with Vocal Feedback
- URL: http://arxiv.org/abs/2401.01362v1
- Date: Mon, 18 Dec 2023 19:28:23 GMT
- Title: Assisting Blind People Using Object Detection with Vocal Feedback
- Authors: Heba Najm, Khirallah Elferjani and Alhaam Alariyibi
- Abstract summary: The proposed approach suggests detection of objects in real-time video by using a web camera.
The OpenCV libraries of Python is used to implement the software program.
Image recognition results are transferred to the visually impaired users in audible form by means of Google text-to-speech library.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: For visually impaired people, it is highly difficult to make independent
movement and safely move in both indoors and outdoors environment. Furthermore,
these physically and visually challenges prevent them from in day-today live
activities. Similarly, they have problem perceiving objects of surrounding
environment that may pose a risk to them. The proposed approach suggests
detection of objects in real-time video by using a web camera, for the object
identification, process. You Look Only Once (YOLO) model is utilized which is
CNN-based real-time object detection technique. Additionally, The OpenCV
libraries of Python is used to implement the software program as well as deep
learning process is performed. Image recognition results are transferred to the
visually impaired users in audible form by means of Google text-to-speech
library and determine object location relative to its position in the screen.
The obtaining result was evaluated by using the mean Average Precision (mAP),
and it was found that the proposed approach achieves excellent results when it
compared to previous approaches.
Related papers
- Accelerating Object Detection with YOLOv4 for Real-Time Applications [0.276240219662896]
Convolutional Neural Network (CNN) have emerged as a powerful tool for recognizing image content and in computer vision approach for most problems.
This paper introduces the brief introduction of deep learning and object detection framework like Convolutional Neural Network(CNN)
arXiv Detail & Related papers (2024-10-17T17:44:57Z) - Visual Context-Aware Person Fall Detection [52.49277799455569]
We present a segmentation pipeline to semi-automatically separate individuals and objects in images.
Background objects such as beds, chairs, or wheelchairs can challenge fall detection systems, leading to false positive alarms.
We demonstrate that object-specific contextual transformations during training effectively mitigate this challenge.
arXiv Detail & Related papers (2024-04-11T19:06:36Z) - SalienDet: A Saliency-based Feature Enhancement Algorithm for Object
Detection for Autonomous Driving [160.57870373052577]
We propose a saliency-based OD algorithm (SalienDet) to detect unknown objects.
Our SalienDet utilizes a saliency-based algorithm to enhance image features for object proposal generation.
We design a dataset relabeling approach to differentiate the unknown objects from all objects in training sample set to achieve Open-World Detection.
arXiv Detail & Related papers (2023-05-11T16:19:44Z) - Active Visual Search in the Wild [12.354788629408933]
We propose a system where a user can enter target commands using free-form language.
We call this system Active Visual Search in the Wild (AVSW)
AVSW detects and plans to search for a target object inputted by a user through a semantic grid map represented by static landmarks.
arXiv Detail & Related papers (2022-09-19T07:18:46Z) - SHOP: A Deep Learning Based Pipeline for near Real-Time Detection of
Small Handheld Objects Present in Blurry Video [0.0]
We present SHOP (Small Handheld Object Pipeline), a pipeline that reliably interprets blurry images containing handheld objects.
The specific models used in each stage of the pipeline are flexible and can be changed based on performance requirements.
We also present a subset of MS COCO consisting solely of handheld objects that can be used to continue the development of handheld object detection methods.
arXiv Detail & Related papers (2022-03-29T04:31:30Z) - Object Manipulation via Visual Target Localization [64.05939029132394]
Training agents to manipulate objects, poses many challenges.
We propose an approach that explores the environment in search for target objects, computes their 3D coordinates once they are located, and then continues to estimate their 3D locations even when the objects are not visible.
Our evaluations show a massive 3x improvement in success rate over a model that has access to the same sensory suite.
arXiv Detail & Related papers (2022-03-15T17:59:01Z) - Siamese Network Training Using Sampled Triplets and Image Transformation [0.0]
The device used in this work detects the objects over the surface of the water using two thermal cameras.
To avoid the obstacle collision autonomously, it is required to track the objects in real-time.
A Machine Learning (ML) approach for Computer Vision (CV) was used using as the high-level programming environment in Python.
arXiv Detail & Related papers (2021-06-13T14:47:52Z) - Analysis of voxel-based 3D object detection methods efficiency for
real-time embedded systems [93.73198973454944]
Two popular voxel-based 3D object detection methods are studied in this paper.
Our experiments show that these methods mostly fail to detect distant small objects due to the sparsity of the input point clouds at large distances.
Our findings suggest that a considerable part of the computations of existing methods is focused on locations of the scene that do not contribute with successful detection.
arXiv Detail & Related papers (2021-05-21T12:40:59Z) - Scale Normalized Image Pyramids with AutoFocus for Object Detection [75.71320993452372]
A scale normalized image pyramid (SNIP) is generated that, like human vision, only attends to objects within a fixed size range at different scales.
We propose an efficient spatial sub-sampling scheme which only operates on fixed-size sub-regions likely to contain objects.
The resulting algorithm is referred to as AutoFocus and results in a 2.5-5 times speed-up during inference when used with SNIP.
arXiv Detail & Related papers (2021-02-10T18:57:53Z) - POMP: Pomcp-based Online Motion Planning for active visual search in
indoor environments [89.43830036483901]
We focus on the problem of learning an optimal policy for Active Visual Search (AVS) of objects in known indoor environments with an online setup.
Our POMP method uses as input the current pose of an agent and a RGB-D frame.
We validate our method on the publicly available AVD benchmark, achieving an average success rate of 0.76 with an average path length of 17.1.
arXiv Detail & Related papers (2020-09-17T08:23:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.