A non-cooperative meta-modeling game for automated third-party
calibrating, validating, and falsifying constitutive laws with parallelized
adversarial attacks
- URL: http://arxiv.org/abs/2004.09392v1
- Date: Mon, 13 Apr 2020 18:43:28 GMT
- Title: A non-cooperative meta-modeling game for automated third-party
calibrating, validating, and falsifying constitutive laws with parallelized
adversarial attacks
- Authors: Kun Wang, WaiChing Sun, Qiang Du
- Abstract summary: The evaluation of models, especially for high-risk and high-regret engineering applications, requires efficient and rigorous third-party calibration, validation and falsification.
This work attempts to introduce concepts from game theory and machine learning techniques to overcome many of these existing difficulties.
We introduce an automated meta-modeling game where two competing AI agents generate experimental data to calibrate a given model and to explore its weakness, in order to improve experiment design and model robustness through competition.
- Score: 6.113400683524824
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The evaluation of constitutive models, especially for high-risk and
high-regret engineering applications, requires efficient and rigorous
third-party calibration, validation and falsification. While there are numerous
efforts to develop paradigms and standard procedures to validate models,
difficulties may arise due to the sequential, manual and often biased nature of
the commonly adopted calibration and validation processes, thus slowing down
data collections, hampering the progress towards discovering new physics,
increasing expenses and possibly leading to misinterpretations of the
credibility and application ranges of proposed models. This work attempts to
introduce concepts from game theory and machine learning techniques to overcome
many of these existing difficulties. We introduce an automated meta-modeling
game where two competing AI agents systematically generate experimental data to
calibrate a given constitutive model and to explore its weakness, in order to
improve experiment design and model robustness through competition. The two
agents automatically search for the Nash equilibrium of the meta-modeling game
in an adversarial reinforcement learning framework without human intervention.
By capturing all possible design options of the laboratory experiments into a
single decision tree, we recast the design of experiments as a game of
combinatorial moves that can be resolved through deep reinforcement learning by
the two competing players. Our adversarial framework emulates idealized
scientific collaborations and competitions among researchers to achieve a
better understanding of the application range of the learned material laws and
prevent misinterpretations caused by conventional AI-based third-party
validation.
Related papers
- Causality can systematically address the monsters under the bench(marks) [64.36592889550431]
Benchmarks are plagued by various biases, artifacts, or leakage.
Models may behave unreliably due to poorly explored failure modes.
causality offers an ideal framework to systematically address these challenges.
arXiv Detail & Related papers (2025-02-07T17:01:37Z) - Explainable data-driven modeling via mixture of experts: towards
effective blending of grey and black-box models [6.331947318187792]
We propose a comprehensive framework based on a "mixture of experts" rationale.
This approach enables the data-based fusion of diverse local models.
We penalize abrupt variations in the expert's combination to enhance interpretability.
arXiv Detail & Related papers (2024-01-30T15:53:07Z) - Two-Stage Surrogate Modeling for Data-Driven Design Optimization with
Application to Composite Microstructure Generation [1.912429179274357]
This paper introduces a novel two-stage machine learning-based surrogate modeling framework to address inverse problems in scientific and engineering fields.
In the first stage, a machine learning model termed the "learner" identifies a limited set of candidates within the input design space whose predicted outputs closely align with desired outcomes.
In the second stage, a separate surrogate model, functioning as an "evaluator," is employed to assess the reduced candidate space generated in the first stage.
arXiv Detail & Related papers (2024-01-04T00:25:12Z) - Leveraging World Model Disentanglement in Value-Based Multi-Agent
Reinforcement Learning [18.651307543537655]
We propose a novel model-based multi-agent reinforcement learning approach named Value Decomposition Framework with Disentangled World Model.
We present experimental results in Easy, Hard, and Super-Hard StarCraft II micro-management challenges to demonstrate that our method achieves high sample efficiency and exhibits superior performance in defeating the enemy armies compared to other baselines.
arXiv Detail & Related papers (2023-09-08T22:12:43Z) - Mimicking To Dominate: Imitation Learning Strategies for Success in
Multiagent Competitive Games [13.060023718506917]
We develop a new multi-agent imitation learning model for predicting next moves of the opponents.
We also present a new multi-agent reinforcement learning algorithm that combines our imitation learning model and policy training into one single training process.
Experimental results show that our approach achieves superior performance compared to existing state-of-the-art multi-agent RL algorithms.
arXiv Detail & Related papers (2023-08-20T07:30:13Z) - Exploring the Trade-off between Plausibility, Change Intensity and
Adversarial Power in Counterfactual Explanations using Multi-objective
Optimization [73.89239820192894]
We argue that automated counterfactual generation should regard several aspects of the produced adversarial instances.
We present a novel framework for the generation of counterfactual examples.
arXiv Detail & Related papers (2022-05-20T15:02:53Z) - Exploration and Exploitation: Two Ways to Improve Chinese Spelling
Correction Models [51.744357472072416]
We propose a method, which continually identifies the weak spots of a model to generate more valuable training instances.
Experimental results show that such an adversarial training method combined with the pretraining strategy can improve both the generalization and robustness of multiple CSC models.
arXiv Detail & Related papers (2021-05-31T09:17:33Z) - Congestion-aware Multi-agent Trajectory Prediction for Collision
Avoidance [110.63037190641414]
We propose to learn congestion patterns explicitly and devise a novel "Sense--Learn--Reason--Predict" framework.
By decomposing the learning phases into two stages, a "student" can learn contextual cues from a "teacher" while generating collision-free trajectories.
In experiments, we demonstrate that the proposed model is able to generate collision-free trajectory predictions in a synthetic dataset.
arXiv Detail & Related papers (2021-03-26T02:42:33Z) - Adversarial Training is Not Ready for Robot Learning [55.493354071227174]
Adversarial training is an effective method to train deep learning models that are resilient to norm-bounded perturbations.
We show theoretically and experimentally that neural controllers obtained via adversarial training are subjected to three types of defects.
Our results suggest that adversarial training is not yet ready for robot learning.
arXiv Detail & Related papers (2021-03-15T07:51:31Z) - Automated and Formal Synthesis of Neural Barrier Certificates for
Dynamical Models [70.70479436076238]
We introduce an automated, formal, counterexample-based approach to synthesise Barrier Certificates (BC)
The approach is underpinned by an inductive framework, which manipulates a candidate BC structured as a neural network, and a sound verifier, which either certifies the candidate's validity or generates counter-examples.
The outcomes show that we can synthesise sound BCs up to two orders of magnitude faster, with in particular a stark speedup on the verification engine.
arXiv Detail & Related papers (2020-07-07T07:39:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.