Diversity-guided Search Exploration for Self-driving Cars Test
Generation through Frenet Space Encoding
- URL: http://arxiv.org/abs/2401.14682v1
- Date: Fri, 26 Jan 2024 06:57:00 GMT
- Title: Diversity-guided Search Exploration for Self-driving Cars Test
Generation through Frenet Space Encoding
- Authors: Timo Blattner, Christian Birchler, Timo Kehrer, Sebastiano Panichella
- Abstract summary: The rise of self-driving cars (SDCs) presents important safety challenges to address in dynamic environments.
While field testing is essential, current methods lack diversity in assessing critical SDC scenarios.
We show that the likelihood of leading to an out-of-bound condition can be learned by the deep-learning vanilla transformer model.
- Score: 4.135985106933988
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The rise of self-driving cars (SDCs) presents important safety challenges to
address in dynamic environments. While field testing is essential, current
methods lack diversity in assessing critical SDC scenarios. Prior research
introduced simulation-based testing for SDCs, with Frenetic, a test generation
approach based on Frenet space encoding, achieving a relatively high percentage
of valid tests (approximately 50%) characterized by naturally smooth curves.
The "minimal out-of-bound distance" is often taken as a fitness function, which
we argue to be a sub-optimal metric. Instead, we show that the likelihood of
leading to an out-of-bound condition can be learned by the deep-learning
vanilla transformer model. We combine this "inherently learned metric" with a
genetic algorithm, which has been shown to produce a high diversity of tests.
To validate our approach, we conducted a large-scale empirical evaluation on a
dataset comprising over 1,174 simulated test cases created to challenge the
SDCs behavior. Our investigation revealed that our approach demonstrates a
substantial reduction in generating non-valid test cases, increased diversity,
and high accuracy in identifying safety violations during SDC test execution.
Related papers
- High-Dimensional Fault Tolerance Testing of Highly Automated Vehicles Based on Low-Rank Models [39.139025989575686]
Fault Injection (FI) testing is conducted to evaluate the safety level of HAVs.
To fully cover test cases, various driving scenarios and fault settings should be considered.
We propose to accelerate FI testing under the low-rank Smoothness Regularized Matrix Factorization framework.
arXiv Detail & Related papers (2024-07-28T14:27:13Z) - How does Simulation-based Testing for Self-driving Cars match Human
Perception? [5.742965094549775]
This study investigates the factors that determine how humans perceive self-driving cars test cases as safe, unsafe, realistic, or unrealistic.
Our findings indicate that the human assessment of the safety and realism of failing and passing test cases can vary based on different factors.
arXiv Detail & Related papers (2024-01-26T09:58:12Z) - Unsupervised Domain Adaptation for Self-Driving from Past Traversal
Features [69.47588461101925]
We propose a method to adapt 3D object detectors to new driving environments.
Our approach enhances LiDAR-based detection models using spatial quantized historical features.
Experiments on real-world datasets demonstrate significant improvements.
arXiv Detail & Related papers (2023-09-21T15:00:31Z) - From Static Benchmarks to Adaptive Testing: Psychometrics in AI Evaluation [60.14902811624433]
We discuss a paradigm shift from static evaluation methods to adaptive testing.
This involves estimating the characteristics and value of each test item in the benchmark and dynamically adjusting items in real-time.
We analyze the current approaches, advantages, and underlying reasons for adopting psychometrics in AI evaluation.
arXiv Detail & Related papers (2023-06-18T09:54:33Z) - On Pitfalls of Test-Time Adaptation [82.8392232222119]
Test-Time Adaptation (TTA) has emerged as a promising approach for tackling the robustness challenge under distribution shifts.
We present TTAB, a test-time adaptation benchmark that encompasses ten state-of-the-art algorithms, a diverse array of distribution shifts, and two evaluation protocols.
arXiv Detail & Related papers (2023-06-06T09:35:29Z) - A Comprehensive Survey on Test-Time Adaptation under Distribution Shifts [143.14128737978342]
Test-time adaptation, an emerging paradigm, has the potential to adapt a pre-trained model to unlabeled data during testing, before making predictions.
Recent progress in this paradigm highlights the significant benefits of utilizing unlabeled data for training self-adapted models prior to inference.
arXiv Detail & Related papers (2023-03-27T16:32:21Z) - Identifying and Explaining Safety-critical Scenarios for Autonomous
Vehicles via Key Features [5.634825161148484]
This paper uses Instance Space Analysis (ISA) to identify the significant features of test scenarios that affect their ability to reveal the unsafe behaviour of AVs.
ISA identifies the features that best differentiate safety-critical scenarios from normal driving and visualises the impact of these features on test scenario outcomes (safe/unsafe) in 2D.
To test the predictive ability of the identified features, we train five Machine Learning classifiers to classify test scenarios as safe or unsafe.
arXiv Detail & Related papers (2022-12-15T00:52:47Z) - Efficient and Effective Generation of Test Cases for Pedestrian
Detection -- Search-based Software Testing of Baidu Apollo in SVL [14.482670650074885]
This paper presents a study on testing pedestrian detection and emergency braking system of the Baidu Apollo autonomous driving platform within the SVL simulator.
We propose an evolutionary automated test generation technique that generates failure-revealing scenarios for Apollo in the SVL environment.
In order to demonstrate the efficiency and effectiveness of our approach, we also report the results from a baseline random generation technique.
arXiv Detail & Related papers (2021-09-16T13:11:53Z) - Generating and Characterizing Scenarios for Safety Testing of Autonomous
Vehicles [86.9067793493874]
We propose efficient mechanisms to characterize and generate testing scenarios using a state-of-the-art driving simulator.
We use our method to characterize real driving data from the Next Generation Simulation (NGSIM) project.
We rank the scenarios by defining metrics based on the complexity of avoiding accidents and provide insights into how the AV could have minimized the probability of incurring an accident.
arXiv Detail & Related papers (2021-03-12T17:00:23Z) - Noisy Adaptive Group Testing using Bayesian Sequential Experimental
Design [63.48989885374238]
When the infection prevalence of a disease is low, Dorfman showed 80 years ago that testing groups of people can prove more efficient than testing people individually.
Our goal in this paper is to propose new group testing algorithms that can operate in a noisy setting.
arXiv Detail & Related papers (2020-04-26T23:41:33Z) - Efficient statistical validation with edge cases to evaluate Highly
Automated Vehicles [6.198523595657983]
The widescale deployment of Autonomous Vehicles seems to be imminent despite many safety challenges that are yet to be resolved.
Existing standards focus on deterministic processes where the validation requires only a set of test cases that cover the requirements.
This paper presents a new approach to compute the statistical characteristics of a system's behaviour by biasing automatically generated test cases towards the worst case scenarios.
arXiv Detail & Related papers (2020-03-04T04:35:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.