AccessLens: Auto-detecting Inaccessibility of Everyday Objects
- URL: http://arxiv.org/abs/2401.15996v2
- Date: Fri, 23 Feb 2024 17:06:14 GMT
- Title: AccessLens: Auto-detecting Inaccessibility of Everyday Objects
- Authors: Nahyun Kwon, Qian Lu, Muhammad Hasham Qazi, Joanne Liu, Changhoon Oh,
Shu Kong, Jeeeun Kim
- Abstract summary: We introduce AccessLens, an end-to-end system designed to identify inaccessible interfaces in daily objects.
Our approach involves training a detector using the novel AccessDB dataset designed to automatically recognize 21 distinct Inaccessibility Classes.
AccessMeta serves as a robust way to build a comprehensive dictionary linking these accessibility classes to open-source 3D augmentation designs.
- Score: 17.269659576368536
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In our increasingly diverse society, everyday physical interfaces often
present barriers, impacting individuals across various contexts. This
oversight, from small cabinet knobs to identical wall switches that can pose
different contextual challenges, highlights an imperative need for solutions.
Leveraging low-cost 3D-printed augmentations such as knob magnifiers and
tactile labels seems promising, yet the process of discovering unrecognized
barriers remains challenging because disability is context-dependent. We
introduce AccessLens, an end-to-end system designed to identify inaccessible
interfaces in daily objects, and recommend 3D-printable augmentations for
accessibility enhancement. Our approach involves training a detector using the
novel AccessDB dataset designed to automatically recognize 21 distinct
Inaccessibility Classes (e.g., bar-small and round-rotate) within 6 common
object categories (e.g., handle and knob). AccessMeta serves as a robust way to
build a comprehensive dictionary linking these accessibility classes to
open-source 3D augmentation designs. Experiments demonstrate our detector's
performance in detecting inaccessible objects.
Related papers
- Open3DTrack: Towards Open-Vocabulary 3D Multi-Object Tracking [73.05477052645885]
We introduce open-vocabulary 3D tracking, which extends the scope of 3D tracking to include objects beyond predefined categories.
We propose a novel approach that integrates open-vocabulary capabilities into a 3D tracking framework, allowing for generalization to unseen object classes.
arXiv Detail & Related papers (2024-10-02T15:48:42Z) - V3Det Challenge 2024 on Vast Vocabulary and Open Vocabulary Object Detection: Methods and Results [142.5704093410454]
The V3Det Challenge 2024 aims to push the boundaries of object detection research.
The challenge consists of two tracks: Vast Vocabulary Object Detection and Open Vocabulary Object Detection.
We aim to inspire future research directions in vast vocabulary and open-vocabulary object detection.
arXiv Detail & Related papers (2024-06-17T16:58:51Z) - UniTeam: Open Vocabulary Mobile Manipulation Challenge [4.523096223190858]
This report introduces our UniTeam agent - an improved baseline for the "HomeRobot: Open Vocabulary Mobile Manipulation" challenge.
The challenge poses problems of navigation in unfamiliar environments, manipulation of novel objects, and recognition of open-vocabulary object classes.
This challenge aims to facilitate cross-cutting research in embodied AI using recent advances in machine learning, computer vision, natural language, and robotics.
arXiv Detail & Related papers (2023-12-14T02:24:29Z) - Fine-grained 3D object recognition: an approach and experiments [0.0]
Three-dimensional (3D) object recognition technology is being used as a core technology in advanced technologies such as autonomous driving of automobiles.
There are two sets of approaches for 3D object recognition: (i) hand-crafted approaches like Global Orthographic Object Descriptor (GOOD), and (ii) deep learning-based approaches such as MobileNet and VGG.
In this paper, we first implemented an offline 3D object recognition system that takes an object view as input and generates category labels as output.
In the offline stage, instance-based learning (IBL) is used to form a new
arXiv Detail & Related papers (2023-06-28T04:48:21Z) - View-to-Label: Multi-View Consistency for Self-Supervised 3D Object
Detection [46.077668660248534]
We propose a novel approach to self-supervise 3D object detection purely from RGB sequences alone.
Our experiments on KITTI 3D dataset demonstrate performance on par with state-of-the-art self-supervised methods.
arXiv Detail & Related papers (2023-05-29T09:30:39Z) - Universal Object Detection with Large Vision Model [79.06618136217142]
This study focuses on the large-scale, multi-domain universal object detection problem.
To address these challenges, we introduce our approach to label handling, hierarchy-aware design, and resource-efficient model training.
Our method has demonstrated remarkable performance, securing a prestigious second-place ranking in the object detection track of the Robust Vision Challenge 2022.
arXiv Detail & Related papers (2022-12-19T12:40:13Z) - Integrating Accessibility in a Mobile App Development Course [0.0]
The course introduced three accessibility-related topics using various interventions: Accessibility Awareness (a guest lecture by a legal expert), Technical Knowledge (lectures on Android accessibility guidelines and testing practices), and Empathy (an activity that required students to blindfold themselves and interact with their phones using a screen-reader)
All students could correctly identify at least one accessibility issue in the user interface of a real-world app given its screenshot, and 90% of them could provide a correct solution to fix it.
arXiv Detail & Related papers (2022-10-12T12:44:33Z) - Continual Object Detection via Prototypical Task Correlation Guided
Gating Mechanism [120.1998866178014]
We present a flexible framework for continual object detection via pRotOtypical taSk corrElaTion guided gaTingAnism (ROSETTA)
Concretely, a unified framework is shared by all tasks while task-aware gates are introduced to automatically select sub-models for specific tasks.
Experiments on COCO-VOC, KITTI-Kitchen, class-incremental detection on VOC and sequential learning of four tasks show that ROSETTA yields state-of-the-art performance.
arXiv Detail & Related papers (2022-05-06T07:31:28Z) - Multi-modal Transformers Excel at Class-agnostic Object Detection [105.10403103027306]
We argue that existing methods lack a top-down supervision signal governed by human-understandable semantics.
We develop an efficient and flexible MViT architecture using multi-scale feature processing and deformable self-attention.
We show the significance of MViT proposals in a diverse range of applications.
arXiv Detail & Related papers (2021-11-22T18:59:29Z) - Learnable Online Graph Representations for 3D Multi-Object Tracking [156.58876381318402]
We propose a unified and learning based approach to the 3D MOT problem.
We employ a Neural Message Passing network for data association that is fully trainable.
We show the merit of the proposed approach on the publicly available nuScenes dataset by achieving state-of-the-art performance of 65.6% AMOTA and 58% fewer ID-switches.
arXiv Detail & Related papers (2021-04-23T17:59:28Z) - ORBIT: A Real-World Few-Shot Dataset for Teachable Object Recognition [21.594641488685376]
We present the ORBIT dataset and benchmark, grounded in a real-world application of teachable object recognizers for people who are blind/low vision.
The dataset contains 3,822 videos of 486 objects recorded by people who are blind/low-vision on their mobile phones.
The benchmark reflects a realistic, highly challenging recognition problem, providing a rich playground to drive research in robustness to few-shot, high-variation conditions.
arXiv Detail & Related papers (2021-04-08T15:32:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.