XRBench: An Extended Reality (XR) Machine Learning Benchmark Suite for
the Metaverse
- URL: http://arxiv.org/abs/2211.08675v2
- Date: Sat, 20 May 2023 00:16:23 GMT
- Title: XRBench: An Extended Reality (XR) Machine Learning Benchmark Suite for
the Metaverse
- Authors: Hyoukjun Kwon, Krishnakumar Nair, Jamin Seo, Jason Yik, Debabrata
Mohapatra, Dongyuan Zhan, Jinook Song, Peter Capak, Peizhao Zhang, Peter
Vajda, Colby Banbury, Mark Mazumder, Liangzhen Lai, Ashish Sirasao, Tushar
Krishna, Harshit Khaitan, Vikas Chandra, Vijay Janapa Reddi
- Abstract summary: Real-time multi-task multi-model (MTMM) workloads are emerging for applications areas like extended reality (XR) to support metaverse use cases.
These workloads combine user interactivity with computationally complex machine learning (ML) activities.
These workloads present unique difficulties and constraints.
- Score: 18.12263246913058
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Real-time multi-task multi-model (MTMM) workloads, a new form of deep
learning inference workloads, are emerging for applications areas like extended
reality (XR) to support metaverse use cases. These workloads combine user
interactivity with computationally complex machine learning (ML) activities.
Compared to standard ML applications, these ML workloads present unique
difficulties and constraints. Real-time MTMM workloads impose heterogeneity and
concurrency requirements on future ML systems and devices, necessitating the
development of new capabilities. This paper begins with a discussion of the
various characteristics of these real-time MTMM ML workloads and presents an
ontology for evaluating the performance of future ML hardware for XR systems.
Next, we present XRBENCH, a collection of MTMM ML tasks, models, and usage
scenarios that execute these models in three representative ways: cascaded,
concurrent, and cascaded-concurrent for XR use cases. Finally, we emphasize the
need for new metrics that capture the requirements properly. We hope that our
work will stimulate research and lead to the development of a new generation of
ML systems for XR use cases. XRBench is available as an open-source project:
https://github.com/XRBench
Related papers
- DeeR-VLA: Dynamic Inference of Multimodal Large Language Models for Efficient Robot Execution [114.61347672265076]
Development of MLLMs for real-world robots is challenging due to the typically limited computation and memory capacities available on robotic platforms.
We propose a Dynamic Early-Exit Framework for Robotic Vision-Language-Action Model (DeeR) that automatically adjusts the size of the activated MLLM.
DeeR demonstrates significant reductions in computational costs of LLM by 5.2-6.5x and GPU memory of LLM by 2-6x without compromising performance.
arXiv Detail & Related papers (2024-11-04T18:26:08Z) - xGen-MM (BLIP-3): A Family of Open Large Multimodal Models [157.44696790158784]
This report introduces xGen-MM, a framework for developing Large Multimodal Models (LMMs)
The framework comprises meticulously curated datasets, a training recipe, model architectures, and a resulting suite of LMMs.
Our models undergo rigorous evaluation across a range of tasks, including both single and multi-image benchmarks.
arXiv Detail & Related papers (2024-08-16T17:57:01Z) - MIBench: Evaluating Multimodal Large Language Models over Multiple Images [70.44423964171088]
We propose a new benchmark MIBench, to comprehensively evaluate fine-grained abilities of MLLMs in multi-image scenarios.
Specifically, MIBench categorizes the multi-image abilities into three scenarios: multi-image instruction (MII), multimodal knowledge-seeking (MKS) and multimodal in-context learning (MIC)
The results reveal that although current models excel in single-image tasks, they exhibit significant shortcomings when faced with multi-image inputs.
arXiv Detail & Related papers (2024-07-21T21:22:58Z) - RedCoast: A Lightweight Tool to Automate Distributed Training of LLMs on Any GPU/TPUs [32.01139974519813]
We present RedCoast, a tool crafted to automate distributed training and inference for large language models (LLMs)
We also propose a mechanism that allows for the customization of diverse ML pipelines through the definition of merely three functions.
As a result, Redco implementations exhibit significantly fewer lines of code compared to their official counterparts.
arXiv Detail & Related papers (2023-10-25T04:32:35Z) - Simultaneous Machine Translation with Large Language Models [51.470478122113356]
We investigate the possibility of applying Large Language Models to SimulMT tasks.
We conducted experiments using the textttLlama2-7b-chat model on nine different languages from the MUST-C dataset.
The results show that LLM outperforms dedicated MT models in terms of BLEU and LAAL metrics.
arXiv Detail & Related papers (2023-09-13T04:06:47Z) - In Situ Framework for Coupling Simulation and Machine Learning with
Application to CFD [51.04126395480625]
Recent years have seen many successful applications of machine learning (ML) to facilitate fluid dynamic computations.
As simulations grow, generating new training datasets for traditional offline learning creates I/O and storage bottlenecks.
This work offers a solution by simplifying this coupling and enabling in situ training and inference on heterogeneous clusters.
arXiv Detail & Related papers (2023-06-22T14:07:54Z) - DREAM: A Dynamic Scheduler for Dynamic Real-time Multi-model ML
Workloads [8.266680870089997]
We propose a new scheduler, DREAM, which effectively handles various dynamicity in RTMM workloads.
DREAM quantifies the unique requirements for RTMM workloads and utilizes the scores quantified to drive scheduling decisions.
In our evaluation of five scenarios of RTMM workload, DREAM reduces the overall UXCost by 32.2% and 50.0% in the mean geometric (up to 80.8% and 97.6%) compared to state-of-the-art baselines.
arXiv Detail & Related papers (2022-12-07T02:48:14Z) - BPMN4sML: A BPMN Extension for Serverless Machine Learning. Technology
Independent and Interoperable Modeling of Machine Learning Workflows and
their Serverless Deployment Orchestration [0.0]
Machine learning (ML) continues to permeate all layers of academia, industry and society.
Business Process Model and Notation (BPMN) is widely accepted and applied.
BPMN is short of specific support to represent machine learning.
We introduce BPMN4sML (BPMN for serverless machine learning)
arXiv Detail & Related papers (2022-08-02T10:36:00Z) - Few-Shot Class-Incremental Learning by Sampling Multi-Phase Tasks [59.12108527904171]
A model should recognize new classes and maintain discriminability over old classes.
The task of recognizing few-shot new classes without forgetting old classes is called few-shot class-incremental learning (FSCIL)
We propose a new paradigm for FSCIL based on meta-learning by LearnIng Multi-phase Incremental Tasks (LIMIT)
arXiv Detail & Related papers (2022-03-31T13:46:41Z) - Symphony: Composing Interactive Interfaces for Machine Learning [15.322027013779689]
Symphony is a framework for composing interactive ML interfaces with task-specific, data-driven components.
We developed Symphony through participatory design sessions with 10 teams (n=31), and discuss our findings from deploying Symphony to 3 production ML projects at Apple.
arXiv Detail & Related papers (2022-02-18T00:27:30Z) - Multi-Faceted Hierarchical Multi-Task Learning for a Large Number of
Tasks with Multi-dimensional Relations [10.326429525379181]
This work studies the "macro" perspective of shared learning network design and proposes a Multi-Faceted Hierarchical MTL model(MFH)
MFH exploits the multi-dimensional task relations with a nested hierarchical tree structure which maximizes the shared learning.
We evaluate MFH and SOTA models in a large industry video platform of 10 billion samples and results show that MFH outperforms SOTA MTL models significantly in both offline and online evaluations.
arXiv Detail & Related papers (2021-10-26T02:35:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.