MultiTest: Physical-Aware Object Insertion for Testing Multi-sensor
Fusion Perception Systems
- URL: http://arxiv.org/abs/2401.14314v1
- Date: Thu, 25 Jan 2024 17:03:02 GMT
- Title: MultiTest: Physical-Aware Object Insertion for Testing Multi-sensor
Fusion Perception Systems
- Authors: Xinyu Gao, Zhijie Wang, Yang Feng, Lei Ma, Zhenyu Chen, Baowen Xu
- Abstract summary: Multi-sensor fusion (MSF) is a key technique in addressing numerous safety-critical tasks and applications, e.g., self-driving cars and automated robotic arms.
Existing testing methods primarily concentrate on single-sensor perception systems.
We introduce MultiTest, a fitness-guided metamorphic testing method for complex MSF perception systems.
- Score: 23.460181958075566
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Multi-sensor fusion stands as a pivotal technique in addressing numerous
safety-critical tasks and applications, e.g., self-driving cars and automated
robotic arms. With the continuous advancement in data-driven artificial
intelligence (AI), MSF's potential for sensing and understanding intricate
external environments has been further amplified, bringing a profound impact on
intelligent systems and specifically on their perception systems. Similar to
traditional software, adequate testing is also required for AI-enabled MSF
systems. Yet, existing testing methods primarily concentrate on single-sensor
perception systems (e.g., image-/point cloud-based object detection systems).
There remains a lack of emphasis on generating multi-modal test cases for MSF
systems. To address these limitations, we design and implement MultiTest, a
fitness-guided metamorphic testing method for complex MSF perception systems.
MultiTest employs a physical-aware approach to synthesize realistic multi-modal
object instances and insert them into critical positions of background images
and point clouds. A fitness metric is designed to guide and boost the test
generation process. We conduct extensive experiments with five SOTA perception
systems to evaluate MultiTest from the perspectives of: (1) generated test
cases' realism, (2) fault detection capabilities, and (3) performance
improvement. The results show that MultiTest can generate realistic and
modality-consistent test data and effectively detect hundreds of diverse faults
of an MSF system under test. Moreover, retraining an MSF system on the test
cases generated by MultiTest can improve the system's robustness.
Related papers
- Using Causal Inference to Test Systems with Hidden and Interacting Variables: An Evaluative Case Study [2.1146241717926664]
Software systems with large parameter spaces, nondeterminism and high computational cost are challenging to test.
Recent software testing techniques based on causal inference have been successfully applied to systems that exhibit such characteristics.
arXiv Detail & Related papers (2025-04-23T08:50:24Z) - Optimizing Multispectral Object Detection: A Bag of Tricks and Comprehensive Benchmarks [49.84182981950623]
Multispectral object detection, utilizing RGB and TIR (thermal infrared) modalities, is widely recognized as a challenging task.
It requires not only the effective extraction of features from both modalities and robust fusion strategies, but also the ability to address issues such as spectral discrepancies.
We introduce an efficient and easily deployable multispectral object detection framework that can seamlessly optimize high-performing single-modality models.
arXiv Detail & Related papers (2024-11-27T12:18:39Z) - AI-Compass: A Comprehensive and Effective Multi-module Testing Tool for AI Systems [26.605694684145313]
In this study, we design and implement a testing tool, tool, to comprehensively and effectively evaluate AI systems.
The tool extensively assesses adversarial robustness, model interpretability, and performs neuron analysis.
Our research sheds light on a general solution for AI systems testing landscape.
arXiv Detail & Related papers (2024-11-09T11:15:17Z) - RADAR: Robust Two-stage Modality-incomplete Industrial Anomaly Detection [61.71770293720491]
We propose a novel two-stage Robust modAlity-imcomplete fusing and Detecting frAmewoRk, abbreviated as RADAR.
Our bootstrapping philosophy is to enhance two stages in MIIAD, improving the robustness of the Multimodal Transformer.
Our experimental results demonstrate that the proposed RADAR significantly surpasses conventional MIAD methods in terms of effectiveness and robustness.
arXiv Detail & Related papers (2024-10-02T16:47:55Z) - Towards Reliable AI: Adequacy Metrics for Ensuring the Quality of
System-level Testing of Autonomous Vehicles [5.634825161148484]
We introduce a set of black-box test adequacy metrics called "Test suite Instance Space Adequacy" (TISA) metrics.
The TISA metrics offer a way to assess both the diversity and coverage of the test suite and the range of bugs detected during testing.
We evaluate the efficacy of the TISA metrics by examining their correlation with the number of bugs detected in system-level simulation testing of AVs.
arXiv Detail & Related papers (2023-11-14T10:16:05Z) - DARTH: Holistic Test-time Adaptation for Multiple Object Tracking [87.72019733473562]
Multiple object tracking (MOT) is a fundamental component of perception systems for autonomous driving.
Despite the urge of safety in driving systems, no solution to the MOT adaptation problem to domain shift in test-time conditions has ever been proposed.
We introduce DARTH, a holistic test-time adaptation framework for MOT.
arXiv Detail & Related papers (2023-10-03T10:10:42Z) - TestLab: An Intelligent Automated Software Testing Framework [0.0]
TestLab is an automated software testing framework that attempts to gather a set of testing methods and automate them using Artificial Intelligence.
The first two modules aim to identify vulnerabilities from different perspectives, while the third module enhances traditional automated software testing by automatically generating test cases.
arXiv Detail & Related papers (2023-06-06T11:45:22Z) - Benchmarking Robustness of AI-Enabled Multi-sensor Fusion Systems:
Challenges and Opportunities [23.460181958075566]
Multi-Sensor Fusion (MSF) based perception systems have been the foundation in supporting many industrial applications and domains, such as self-driving cars, robotic arms, and unmanned aerial vehicles.
Over the past few years, the fast progress in data-driven artificial intelligence (AI) has brought a fast-increasing trend to empower MSF systems by deep learning techniques to further improve performance, especially on intelligent systems and their perception systems.
Although quite a few AI-enabled MSF perception systems and techniques have been proposed, up to the present, limited benchmarks that focus on MSF perception are publicly available.
arXiv Detail & Related papers (2023-06-06T07:17:56Z) - Interactive System-wise Anomaly Detection [66.3766756452743]
Anomaly detection plays a fundamental role in various applications.
It is challenging for existing methods to handle the scenarios where the instances are systems whose characteristics are not readily observed as data.
We develop an end-to-end approach which includes an encoder-decoder module that learns system embeddings.
arXiv Detail & Related papers (2023-04-21T02:20:24Z) - MMRNet: Improving Reliability for Multimodal Object Detection and
Segmentation for Bin Picking via Multimodal Redundancy [68.7563053122698]
We propose a reliable object detection and segmentation system with MultiModal Redundancy (MMRNet)
This is the first system that introduces the concept of multimodal redundancy to address sensor failure issues during deployment.
We present a new label-free multi-modal consistency (MC) score that utilizes the output from all modalities to measure the overall system output reliability and uncertainty.
arXiv Detail & Related papers (2022-10-19T19:15:07Z) - Testing Deep Learning Models: A First Comparative Study of Multiple
Testing Techniques [15.695048480513536]
Vision-based systems (VBS) are used in autonomous driving, robotic surgery, critical infrastructure surveillance, air and maritime traffic control, etc.
Deep Learning (DL) has revolutionized the capabilities of vision-based systems (VBS) in critical applications such as autonomous driving, robotic surgery, critical infrastructure surveillance, air and maritime traffic control, etc.
arXiv Detail & Related papers (2022-02-24T15:05:19Z) - Increasing the Confidence of Deep Neural Networks by Coverage Analysis [71.57324258813674]
This paper presents a lightweight monitoring architecture based on coverage paradigms to enhance the model against different unsafe inputs.
Experimental results show that the proposed approach is effective in detecting both powerful adversarial examples and out-of-distribution inputs.
arXiv Detail & Related papers (2021-01-28T16:38:26Z) - Testing the Safety of Self-driving Vehicles by Simulating Perception and
Prediction [88.0416857308144]
We propose an alternative to sensor simulation, as sensor simulation is expensive and has large domain gaps.
We directly simulate the outputs of the self-driving vehicle's perception and prediction system, enabling realistic motion planning testing.
arXiv Detail & Related papers (2020-08-13T17:20:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.