SANIP: Shopping Assistant and Navigation for the visually impaired
- URL: http://arxiv.org/abs/2209.03570v1
- Date: Thu, 8 Sep 2022 05:35:03 GMT
- Title: SANIP: Shopping Assistant and Navigation for the visually impaired
- Authors: Shubham Deshmukh, Favin Fernandes, Amey Chavan, Monali Ahire, Devashri
Borse, Jyoti Madake
- Abstract summary: The proposed model consists of three python models i.e. Custom Object Detection, Text Detection and Barcode detection.
For object detection of the hand held object, we have created our own custom dataset that comprises daily goods such as Parle-G, Tide, and Lays.
For the other 2 models proposed the text and barcode information retrieved is converted from text to speech and relayed to the Blind person.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: The proposed shopping assistant model SANIP is going to help blind persons to
detect hand held objects and also to get a video feedback of the information
retrieved from the detected and recognized objects. The proposed model consists
of three python models i.e. Custom Object Detection, Text Detection and Barcode
detection. For object detection of the hand held object, we have created our
own custom dataset that comprises daily goods such as Parle-G, Tide, and Lays.
Other than that we have also collected images of Cart and Exit signs as it is
essential for any person to use a cart and also notice the exit sign in case of
emergency. For the other 2 models proposed the text and barcode information
retrieved is converted from text to speech and relayed to the Blind person. The
model was used to detect objects that were trained on and was successful in
detecting and recognizing the desired output with a good accuracy and
precision.
Related papers
- Fine-Grained Open-Vocabulary Object Detection with Fined-Grained Prompts: Task, Dataset and Benchmark [6.93847426808971]
3F-OVD is a novel task that extends supervised fine-grained object detection to the open-vocabulary setting.
We benchmark state-of-the-art object detectors on our dataset for both settings.
arXiv Detail & Related papers (2025-03-19T03:41:46Z) - Uncertainty Estimation for 3D Object Detection via Evidential Learning [63.61283174146648]
We introduce a framework for quantifying uncertainty in 3D object detection by leveraging an evidential learning loss on Bird's Eye View representations in the 3D detector.
We demonstrate both the efficacy and importance of these uncertainty estimates on identifying out-of-distribution scenes, poorly localized objects, and missing (false negative) detections.
arXiv Detail & Related papers (2024-10-31T13:13:32Z) - Intent3D: 3D Object Detection in RGB-D Scans Based on Human Intention [86.39271731460927]
3D intention grounding is a new task in 3D object detection employing RGB-D, based on human intention, such as "I want something to support my back"
We introduce the new Intent3D dataset, consisting of 44,990 intention texts associated with 209 fine-grained classes from 1,042 scenes of the ScanNet dataset.
We also propose IntentNet, our unique approach, designed to tackle this intention-based detection problem.
arXiv Detail & Related papers (2024-05-28T15:48:39Z) - Latent Object Characteristics Recognition with Visual to Haptic-Audio Cross-modal Transfer Learning [9.178588671620963]
This work aims to recognise the latent unobservable object characteristics.
vision is commonly used for object recognition by robots, but it is ineffective for detecting hidden objects.
We propose a cross-modal transfer learning approach from vision to haptic-audio.
arXiv Detail & Related papers (2024-03-15T21:18:14Z) - SeSame: Simple, Easy 3D Object Detection with Point-Wise Semantics [0.7373617024876725]
In autonomous driving, 3D object detection provides more precise information for downstream tasks, including path planning and motion estimation.
We propose SeSame: a method aimed at enhancing semantic information in existing LiDAR-only based 3D object detection.
Experiments demonstrate the effectiveness of our method with performance improvements on the KITTI object detection benchmark.
arXiv Detail & Related papers (2024-03-11T08:17:56Z) - PatchContrast: Self-Supervised Pre-training for 3D Object Detection [14.603858163158625]
We introduce PatchContrast, a novel self-supervised point cloud pre-training framework for 3D object detection.
We show that our method outperforms existing state-of-the-art models on three commonly-used 3D detection datasets.
arXiv Detail & Related papers (2023-08-14T07:45:54Z) - SalienDet: A Saliency-based Feature Enhancement Algorithm for Object
Detection for Autonomous Driving [160.57870373052577]
We propose a saliency-based OD algorithm (SalienDet) to detect unknown objects.
Our SalienDet utilizes a saliency-based algorithm to enhance image features for object proposal generation.
We design a dataset relabeling approach to differentiate the unknown objects from all objects in training sample set to achieve Open-World Detection.
arXiv Detail & Related papers (2023-05-11T16:19:44Z) - Open-Vocabulary Point-Cloud Object Detection without 3D Annotation [62.18197846270103]
The goal of open-vocabulary 3D point-cloud detection is to identify novel objects based on arbitrary textual descriptions.
We develop a point-cloud detector that can learn a general representation for localizing various objects.
We also propose a novel de-biased triplet cross-modal contrastive learning to connect the modalities of image, point-cloud and text.
arXiv Detail & Related papers (2023-04-03T08:22:02Z) - Detect Only What You Specify : Object Detection with Linguistic Target [0.0]
We propose Language-Targeted Detector (LTD) for the targeted detection based on a recently proposed Transformer-based detector.
LTD is a encoder-decoder architecture and our conditional decoder allows the model to reason about the encoded image with the textual input as the linguistic context.
arXiv Detail & Related papers (2022-11-18T07:28:47Z) - Detect-and-describe: Joint learning framework for detection and
description of objects [0.0]
We present a new approach to simultaneously detect objects and infer their attributes, we call it Detect and Describe (DaD) framework.
DaD is a deep learning-based approach that extends object detection to object attribute prediction as well.
We achieve 97.0% in Area Under the Receiver Operating Characteristic Curve (AUC) for object attributes prediction on aPascal test set.
arXiv Detail & Related papers (2022-04-19T11:57:30Z) - Learning Object Detection from Captions via Textual Scene Attributes [70.90708863394902]
We argue that captions contain much richer information about the image, including attributes of objects and their relations.
We present a method that uses the attributes in this "textual scene graph" to train object detectors.
We empirically demonstrate that the resulting model achieves state-of-the-art results on several challenging object detection datasets.
arXiv Detail & Related papers (2020-09-30T10:59:20Z) - DOPS: Learning to Detect 3D Objects and Predict their 3D Shapes [54.239416488865565]
We propose a fast single-stage 3D object detection method for LIDAR data.
The core novelty of our method is a fast, single-pass architecture that both detects objects in 3D and estimates their shapes.
We find that our proposed method achieves state-of-the-art results by 5% on object detection in ScanNet scenes, and it gets top results by 3.4% in the Open dataset.
arXiv Detail & Related papers (2020-04-02T17:48:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.