Empowering Visually Impaired Individuals: A Novel Use of Apple Live
Photos and Android Motion Photos
- URL: http://arxiv.org/abs/2309.08022v1
- Date: Thu, 14 Sep 2023 20:46:35 GMT
- Title: Empowering Visually Impaired Individuals: A Novel Use of Apple Live
Photos and Android Motion Photos
- Authors: Seyedalireza Khoshsirat, Chandra Kambhamettu
- Abstract summary: We advocate for the use of Apple Live Photos and Android Motion Photos technologies.
Our findings reveal that both Live Photos and Motion Photos outperform single-frame images in common visual assisting tasks.
- Score: 3.66237529322911
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Numerous applications have been developed to assist visually impaired
individuals that employ a machine learning unit to process visual input.
However, a critical challenge with these applications is the sub-optimal
quality of images captured by the users. Given the complexity of operating a
camera for visually impaired individuals, we advocate for the use of Apple Live
Photos and Android Motion Photos technologies. In this study, we introduce a
straightforward methodology to evaluate and contrast the efficacy of
Live/Motion Photos against traditional image-based approaches. Our findings
reveal that both Live Photos and Motion Photos outperform single-frame images
in common visual assisting tasks, specifically in object classification and
VideoQA. We validate our results through extensive experiments on the ORBIT
dataset, which consists of videos collected by visually impaired individuals.
Furthermore, we conduct a series of ablation studies to delve deeper into the
impact of deblurring and longer temporal crops.
Related papers
- Low-Light Enhancement Effect on Classification and Detection: An Empirical Study [48.6762437869172]
We evaluate the impact of Low-Light Image Enhancement (LLIE) methods on high-level vision tasks.
Our findings suggest a disconnect between image enhancement for human visual perception and for machine analysis.
This insight is crucial for the development of LLIE techniques that align with the needs of both human and machine vision.
arXiv Detail & Related papers (2024-09-22T14:21:31Z) - Freeview Sketching: View-Aware Fine-Grained Sketch-Based Image Retrieval [85.73149096516543]
We address the choice of viewpoint during sketch creation in Fine-Grained Sketch-Based Image Retrieval (FG-SBIR)
A pilot study highlights the system's struggle when query-sketches differ in viewpoint from target instances.
To reconcile this, we advocate for a view-aware system, seamlessly accommodating both view-agnostic and view-specific tasks.
arXiv Detail & Related papers (2024-07-01T21:20:44Z) - Foveation in the Era of Deep Learning [6.602118206533142]
We introduce an end-to-end differentiable foveated active vision architecture that leverages a graph convolutional network to process foveated images.
Our model learns to iteratively attend to regions of the image relevant for classification.
We find that our model outperforms a state-of-the-art CNN and foveated vision architectures of comparable parameters and a given pixel or budget.
arXiv Detail & Related papers (2023-12-03T16:48:09Z) - Real-Time Under-Display Cameras Image Restoration and HDR on Mobile
Devices [81.61356052916855]
The images captured by under-display cameras (UDCs) are degraded by the screen in front of them.
Deep learning methods for image restoration can significantly reduce the degradation of captured images.
We propose a lightweight model for blind UDC Image Restoration and HDR, and we also provide a benchmark comparing the performance and runtime of different methods on smartphones.
arXiv Detail & Related papers (2022-11-25T11:46:57Z) - Blind Users Accessing Their Training Images in Teachable Object
Recognizers [12.833745050235047]
MyCam is a mobile app that incorporates automatically estimated descriptors for non-visual access to the photos in the users' training sets.
We demonstrate that the real-time photo-level descriptors enabled blind users to reduce photos with cropped objects, and that participants could add more variations by iterating through and accessing the quality of their training sets.
arXiv Detail & Related papers (2022-08-16T21:59:48Z) - Exploring CLIP for Assessing the Look and Feel of Images [87.97623543523858]
We introduce Contrastive Language-Image Pre-training (CLIP) models for assessing both the quality perception (look) and abstract perception (feel) of images in a zero-shot manner.
Our results show that CLIP captures meaningful priors that generalize well to different perceptual assessments.
arXiv Detail & Related papers (2022-07-25T17:58:16Z) - FaIRCoP: Facial Image Retrieval using Contrastive Personalization [43.293482565385055]
Retrieving facial images from attributes plays a vital role in various systems such as face recognition and suspect identification.
Existing methods do so by comparing specific characteristics from the user's mental image against the suggested images.
We propose a method that uses the user's feedback to label images as either similar or dissimilar to the target image.
arXiv Detail & Related papers (2022-05-28T09:52:09Z) - Muscle Vision: Real Time Keypoint Based Pose Classification of Physical
Exercises [52.77024349608834]
3D human pose recognition extrapolated from video has advanced to the point of enabling real-time software applications.
We propose a new machine learning pipeline and web interface that performs human pose recognition on a live video feed to detect when common exercises are performed and classify them accordingly.
arXiv Detail & Related papers (2022-03-23T00:55:07Z) - Performance of object recognition in wearable videos [9.669942356088377]
This work studies the problem of object detection and localization on videos captured by this type of camera.
We present a study of the well known YOLO architecture, that offers an excellent trade-off between accuracy and speed.
arXiv Detail & Related papers (2020-09-10T15:20:17Z) - Self-Supervised Linear Motion Deblurring [112.75317069916579]
Deep convolutional neural networks are state-of-the-art for image deblurring.
We present a differentiable reblur model for self-supervised motion deblurring.
Our experiments demonstrate that self-supervised single image deblurring is really feasible.
arXiv Detail & Related papers (2020-02-10T20:15:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.