Rare event modeling with self-regularized normalizing flows: what can we learn from a single failure?
- URL: http://arxiv.org/abs/2502.21110v1
- Date: Fri, 28 Feb 2025 14:47:52 GMT
- Title: Rare event modeling with self-regularized normalizing flows: what can we learn from a single failure?
- Authors: Charles Dawson, Van Tran, Max Z. Li, Chuchu Fan,
- Abstract summary: This paper introduces CalNF, a framework for posterior learning from limited data.<n>It achieves state-of-the-art performance on data-limited failure modeling and inverse problems.<n>It enables a first-of-a-kind case study into the root causes of the 2022 Southwest Airlines scheduling crisis.
- Score: 10.460029312784911
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Increased deployment of autonomous systems in fields like transportation and robotics have seen a corresponding increase in safety-critical failures. These failures can be difficult to model and debug due to the relative lack of data: compared to tens of thousands of examples from normal operations, we may have only seconds of data leading up to the failure. This scarcity makes it challenging to train generative models of rare failure events, as existing methods risk either overfitting to noise in the limited failure dataset or underfitting due to an overly strong prior. We address this challenge with CalNF, or calibrated normalizing flows, a self-regularized framework for posterior learning from limited data. CalNF achieves state-of-the-art performance on data-limited failure modeling and inverse problems and enables a first-of-a-kind case study into the root causes of the 2022 Southwest Airlines scheduling crisis.
Related papers
- Fed-Joint: Joint Modeling of Nonlinear Degradation Signals and Failure Events for Remaining Useful Life Prediction using Federated Learning [1.024113475677323]
We propose a new prognostic framework for RUL prediction using the joint modeling of nonlinear degradation signals and time-to-failure data.
The proposed method constructs a nonparametric degradation model using a federated multi-output Gaussian process and then employs a federated survival model to predict failure times and probabilities for in-service machinery.
arXiv Detail & Related papers (2025-03-17T17:34:34Z) - A probabilistic framework for learning non-intrusive corrections to long-time climate simulations from short-time training data [12.566163525039558]
We present a strategy for training neural network models to non-intrusively correct under-resolved long-time simulations of chaotic systems.
We demonstrate its ability to accurately predict the anisotropic statistics over time horizons more than 30 times longer than the data seen in training.
arXiv Detail & Related papers (2024-08-02T18:34:30Z) - Label-Free Model Failure Detection for Lidar-based Point Cloud Segmentation [15.779651238128562]
We introduce label-free model failure detection for lidar-based point cloud segmentation.
We leverage different data characteristics by training a supervised and self-supervised stream for the same task to detect failure modes.
We perform a large-scale qualitative analysis and present LidarCODA, the first publicly available dataset with labeled anomalies in real-world lidar data.
arXiv Detail & Related papers (2024-07-19T13:36:35Z) - Root Causing Prediction Anomalies Using Explainable AI [3.970146574042422]
We present a novel application of explainable AI (XAI) for root-causing performance degradation in machine learning models.
A single feature corruption can cause cascading feature, label and concept drifts.
We have successfully applied this technique to improve the reliability of models used in personalized advertising.
arXiv Detail & Related papers (2024-03-04T19:38:50Z) - Identifying and Mitigating Model Failures through Few-shot CLIP-aided
Diffusion Generation [65.268245109828]
We propose an end-to-end framework to generate text descriptions of failure modes associated with spurious correlations.
These descriptions can be used to generate synthetic data using generative models, such as diffusion models.
Our experiments have shown remarkable textbfimprovements in accuracy ($sim textbf21%$) on hard sub-populations.
arXiv Detail & Related papers (2023-12-09T04:43:49Z) - DeepFT: Fault-Tolerant Edge Computing using a Self-Supervised Deep
Surrogate Model [12.335763358698564]
We propose DeepFT to proactively avoid system overloads and their adverse effects.
DeepFT uses a deep surrogate model to accurately predict and diagnose faults in the system.
It offers a highly scalable solution as the model size scales by only 3 and 1 percent per unit increase in the number of active tasks and hosts.
arXiv Detail & Related papers (2022-12-02T16:51:58Z) - Ranking-Based Physics-Informed Line Failure Detection in Power Grids [66.0797334582536]
Real-time and accurate detecting of potential line failures is the first step to mitigating the extreme weather impact and activating emergency controls.
Power balance equations nonlinearity, increased uncertainty in generation during extreme events, and lack of grid observability compromise the efficiency of traditional data-driven failure detection methods.
This paper proposes a Physics-InformEd Line failure Detector (FIELD) that leverages grid topology information to reduce sample and time complexities and improve localization accuracy.
arXiv Detail & Related papers (2022-08-31T18:19:25Z) - Repairing Neural Networks by Leaving the Right Past Behind [23.78437548836594]
Prediction failures of machine learning models often arise from deficiencies in training data.
This work develops a generic framework for both identifying training examples that have given rise to the target failure, and fixing the model through erasing information about them.
arXiv Detail & Related papers (2022-07-11T12:07:39Z) - Distilling Model Failures as Directions in Latent Space [87.30726685335098]
We present a scalable method for automatically distilling a model's failure modes.
We harness linear classifiers to identify consistent error patterns, and induce a natural representation of these failure modes as directions within the feature space.
We demonstrate that this framework allows us to discover and automatically caption challenging subpopulations within the training dataset, and intervene to improve the model's performance on these subpopulations.
arXiv Detail & Related papers (2022-06-29T16:35:24Z) - Fast and Accurate Error Simulation for CNNs against Soft Errors [64.54260986994163]
We present a framework for the reliability analysis of Conal Neural Networks (CNNs) via an error simulation engine.
These error models are defined based on the corruption patterns of the output of the CNN operators induced by faults.
We show that our methodology achieves about 99% accuracy of the fault effects w.r.t. SASSIFI, and a speedup ranging from 44x up to 63x w.r.t.FI, that only implements a limited set of error models.
arXiv Detail & Related papers (2022-06-04T19:45:02Z) - Sufficiently Accurate Model Learning for Planning [119.80502738709937]
This paper introduces the constrained Sufficiently Accurate model learning approach.
It provides examples of such problems, and presents a theorem on how close some approximate solutions can be.
The approximate solution quality will depend on the function parameterization, loss and constraint function smoothness, and the number of samples in model learning.
arXiv Detail & Related papers (2021-02-11T16:27:31Z) - Robot Learning with Crash Constraints [37.685515446816105]
In robot applications where failing is undesired but not catastrophic, many algorithms struggle with leveraging data obtained from failures.
This is usually caused by (i) the failed experiment ending prematurely, or (ii) the acquired data being scarce or corrupted.
We consider failing behaviors as those that violate a constraint and address the problem of learning with crash constraints.
arXiv Detail & Related papers (2020-10-16T23:56:35Z) - TadGAN: Time Series Anomaly Detection Using Generative Adversarial
Networks [73.01104041298031]
TadGAN is an unsupervised anomaly detection approach built on Generative Adversarial Networks (GANs)
To capture the temporal correlations of time series, we use LSTM Recurrent Neural Networks as base models for Generators and Critics.
To demonstrate the performance and generalizability of our approach, we test several anomaly scoring techniques and report the best-suited one.
arXiv Detail & Related papers (2020-09-16T15:52:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.