A Consumer-tier based Visual-Brain Machine Interface for Augmented
Reality Glasses Interactions
- URL: http://arxiv.org/abs/2308.15056v1
- Date: Tue, 29 Aug 2023 06:33:13 GMT
- Title: A Consumer-tier based Visual-Brain Machine Interface for Augmented
Reality Glasses Interactions
- Authors: Yuying Jiang, Fan Bai, Zicheng Zhang, Xiaochen Ye, Zheng Liu, Zhiping
Shi, Jianwei Yao, Xiaojun Liu, Fangkun Zhu, Junling Li Qian Guo, Xiaoan Wang,
Junwen Luo
- Abstract summary: We develop a consumer-tier Visual-Brain Machine Inteface(V-BMI) system specialized for Augmented Reality(AR) glasses interactions.
The developed system consists of a wearable hardware which takes advantages of fast set-up, reliable recording and comfortable wearable experience.
- Score: 7.608396231537724
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Objective.Visual-Brain Machine Interface(V-BMI) has provide a novel
interaction technique for Augmented Reality (AR) industries. Several
state-of-arts work has demonstates its high accuracy and real-time interaction
capbilities. However, most of the studies employ EEGs devices that are rigid
and difficult to apply in real-life AR glasseses application sceniraros. Here
we develop a consumer-tier Visual-Brain Machine Inteface(V-BMI) system
specialized for Augmented Reality(AR) glasses interactions. Approach. The
developed system consists of a wearable hardware which takes advantages of fast
set-up, reliable recording and comfortable wearable experience that
specificized for AR glasses applications. Complementing this hardware, we have
devised a software framework that facilitates real-time interactions within the
system while accommodating a modular configuration to enhance scalability. Main
results. The developed hardware is only 110g and 120x85x23 mm, which with 1
Tohm and peak to peak voltage is less than 1.5 uV, and a V-BMI based angry bird
game and an Internet of Thing (IoT) AR applications are deisgned, we
demonstrated such technology merits of intuitive experience and efficiency
interaction. The real-time interaction accuracy is between 85 and 96
percentages in a commercial AR glasses (DTI is 2.24s and ITR 65 bits-min ).
Significance. Our study indicates the developed system can provide an essential
hardware-software framework for consumer based V-BMI AR glasses. Also, we
derive several pivotal design factors for a consumer-grade V-BMI-based AR
system: 1) Dynamic adaptation of stimulation patterns-classification methods
via computer vision algorithms is necessary for AR glasses applications; and 2)
Algorithmic localization to foster system stability and latency reduction.
Related papers
- Real-Time Hand Gesture Recognition: Integrating Skeleton-Based Data Fusion and Multi-Stream CNN [0.0]
Hand Gesture Recognition (HGR) enables intuitive human-computer interactions in various real-world contexts.
Existing frameworks often struggle to meet the real-time requirements essential for practical HGR applications.
This study introduces a robust, skeleton-based framework for dynamic HGR that simplifies the recognition of dynamic hand gestures into a static image task.
arXiv Detail & Related papers (2024-06-21T09:30:59Z) - Random resistive memory-based deep extreme point learning machine for
unified visual processing [67.51600474104171]
We propose a novel hardware-software co-design, random resistive memory-based deep extreme point learning machine (DEPLM)
Our co-design system achieves huge energy efficiency improvements and training cost reduction when compared to conventional systems.
arXiv Detail & Related papers (2023-12-14T09:46:16Z) - Towards High-Frequency Tracking and Fast Edge-Aware Optimization [2.2662585107579165]
This dissertation advances the state of the art for AR/VR tracking systems by increasing the tracking frequency by orders of magnitude.
It proposes an efficient algorithm for the problem of edge-aware optimization.
arXiv Detail & Related papers (2023-09-02T01:20:34Z) - A Brain-Computer Interface Augmented Reality Framework with
Auto-Adaptive SSVEP Recognition [1.1674893622721483]
We propose a simple adaptive ensemble classification system that handles the inter-subject variability.
We also present a simple BCI-AR framework that supports the development of a wide range of SSVEP-based BCI-AR applications.
arXiv Detail & Related papers (2023-08-11T21:56:00Z) - ArK: Augmented Reality with Knowledge Interactive Emergent Ability [115.72679420999535]
We develop an infinite agent that learns to transfer knowledge memory from general foundation models to novel domains.
The heart of our approach is an emerging mechanism, dubbed Augmented Reality with Knowledge Inference Interaction (ArK)
We show that our ArK approach, combined with large foundation models, significantly improves the quality of generated 2D/3D scenes.
arXiv Detail & Related papers (2023-05-01T17:57:01Z) - Distributed On-Sensor Compute System for AR/VR Devices: A
Semi-Analytical Simulation Framework for Power Estimation [2.5696683295721883]
We show that a novel distributed on-sensor compute architecture can reduce the system power consumption compared to a centralized system.
We show that, in the case of the compute-intensive machine learning based Hand Tracking algorithm, the distributed on-sensor compute architecture can reduce the system power consumption.
arXiv Detail & Related papers (2022-03-14T20:18:24Z) - FPGA-optimized Hardware acceleration for Spiking Neural Networks [69.49429223251178]
This work presents the development of a hardware accelerator for an SNN, with off-line training, applied to an image recognition task.
The design targets a Xilinx Artix-7 FPGA, using in total around the 40% of the available hardware resources.
It reduces the classification time by three orders of magnitude, with a small 4.5% impact on the accuracy, if compared to its software, full precision counterpart.
arXiv Detail & Related papers (2022-01-18T13:59:22Z) - AEGIS: A real-time multimodal augmented reality computer vision based
system to assist facial expression recognition for individuals with autism
spectrum disorder [93.0013343535411]
This paper presents the development of a multimodal augmented reality (AR) system which combines the use of computer vision and deep convolutional neural networks (CNN)
The proposed system, which we call AEGIS, is an assistive technology deployable on a variety of user devices including tablets, smartphones, video conference systems, or smartglasses.
We leverage both spatial and temporal information in order to provide an accurate expression prediction, which is then converted into its corresponding visualization and drawn on top of the original video frame.
arXiv Detail & Related papers (2020-10-22T17:20:38Z) - Augment Yourself: Mixed Reality Self-Augmentation Using Optical
See-through Head-mounted Displays and Physical Mirrors [49.49841698372575]
Optical see-though head-mounted displays (OST HMDs) are one of the key technologies for merging virtual objects and physical scenes to provide an immersive mixed reality (MR) environment to its user.
We propose a novel concept and prototype system that combines OST HMDs and physical mirrors to enable self-augmentation and provide an immersive MR environment centered around the user.
Our system, to the best of our knowledge the first of its kind, estimates the user's pose in the virtual image generated by the mirror using an RGBD camera attached to the HMD and anchors virtual objects to the reflection rather
arXiv Detail & Related papers (2020-07-06T16:53:47Z) - ASFD: Automatic and Scalable Face Detector [129.82350993748258]
We propose a novel Automatic and Scalable Face Detector (ASFD)
ASFD is based on a combination of neural architecture search techniques as well as a new loss design.
Our ASFD-D6 outperforms the prior strong competitors, and our lightweight ASFD-D0 runs at more than 120 FPS with Mobilenet for VGA-resolution images.
arXiv Detail & Related papers (2020-03-25T06:00:47Z) - A Markerless Deep Learning-based 6 Degrees of Freedom PoseEstimation for
with Mobile Robots using RGB Data [3.4806267677524896]
We propose a method to deploy state of the art neural networks for real time 3D object localization on augmented reality devices.
We focus on fast 2D detection approaches which are extracting the 3D pose of the object fast and accurately by using only 2D input.
For the 6D annotation of 2D images, we developed an annotation tool, which is, to our knowledge, the first open source tool to be available.
arXiv Detail & Related papers (2020-01-16T09:13:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.