PIONM: A Generalized Approach to Solving Density-Constrained Mean-Field Games Equilibrium under Modified Boundary Conditions
- URL: http://arxiv.org/abs/2504.03209v1
- Date: Fri, 04 Apr 2025 06:46:09 GMT
- Title: PIONM: A Generalized Approach to Solving Density-Constrained Mean-Field Games Equilibrium under Modified Boundary Conditions
- Authors: Jinwei Liu, Wang Yao, Xiao Zhang,
- Abstract summary: Neural network-based methods are effective for solving equilibria in Mean-Field Games (MFGs)<n>We propose a generalized framework, PIONM, which leverages physics-informed neural operators to solve MFGs equations.<n>PIONM efficiently computes equilibria under varying boundary conditions, including obstacles, diffusion coefficients, initial densities, and terminal functions.
- Score: 6.738098801743817
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Neural network-based methods are effective for solving equilibria in Mean-Field Games (MFGs), particularly in high-dimensional settings. However, solving the coupled partial differential equations (PDEs) in MFGs limits their applicability since solving coupled PDEs is computationally expensive. Additionally, modifying boundary conditions, such as the initial state distribution or terminal value function, necessitates extensive retraining, reducing scalability. To address these challenges, we propose a generalized framework, PIONM (Physics-Informed Neural Operator NF-MKV Net), which leverages physics-informed neural operators to solve MFGs equations. PIONM utilizes neural operators to compute MFGs equilibria for arbitrary boundary conditions. The method encodes boundary conditions as input features and trains the model to align them with density evolution, modeled using discrete-time normalizing flows. Once trained, the algorithm efficiently computes the density distribution at any time step for modified boundary condition, ensuring efficient adaptation to different boundary conditions in MFGs equilibria. Unlike traditional MFGs methods constrained by fixed coefficients, PIONM efficiently computes equilibria under varying boundary conditions, including obstacles, diffusion coefficients, initial densities, and terminal functions. PIONM can adapt to modified conditions while preserving density distribution constraints, demonstrating superior scalability and generalization capabilities compared to existing methods.
Related papers
- Physics-Constrained Flow Matching: Sampling Generative Models with Hard Constraints [0.6990493129893112]
Deep generative models have recently been applied to physical systems governed by partial differential equations (PDEs)<n>Existing methods often rely on soft penalties or architectural biases that fail to guarantee hard constraints.<n>We propose Physics-Constrained Flow Matching, a zero-shot inference framework that enforces arbitrary nonlinear constraints in pretrained flow-based generative models.
arXiv Detail & Related papers (2025-06-04T17:12:37Z) - A Constraint-Preserving Neural Network Approach for Solving Mean-Field Games Equilibrium [6.168001746423872]
Neural network-based methods have demonstrated effectiveness in solving high-dimensional Mean-Field Games (MFG) equilibria.<n>This paper proposes the NF-MKV Net, a network approach that integrates neural networks to solve FBSDEs and their associated fixed-point formulations of MFG equilibria.
arXiv Detail & Related papers (2025-01-29T07:14:09Z) - PINN-FEM: A Hybrid Approach for Enforcing Dirichlet Boundary Conditions in Physics-Informed Neural Networks [1.1060425537315088]
Physics-Informed Neural Networks (PINNs) solve partial differential equations (PDEs)<n>We propose a hybrid approach, PINN-FEM, which combines PINNs with finite element methods (FEM) to impose strong Dirichlet boundary conditions via domain decomposition.<n>This method incorporates FEM-based representations near the boundary, ensuring exact enforcement without compromising convergence.
arXiv Detail & Related papers (2025-01-14T00:47:15Z) - Nonlinear Optimal Control of Electron Dynamics within Hartree-Fock Theory [0.8192907805418581]
We present a solution of this problem within time-dependent Hartree-Fock (TDHF) theory.<n>We frame this problem as an optimization problem constrained by the nonlinear TDHF equations.<n>For three molecular systems, we show that with very small neural network parametrizations of the control, our method yields solutions that achieve desired targets.
arXiv Detail & Related papers (2024-12-04T19:06:21Z) - Monte Carlo Neural PDE Solver for Learning PDEs via Probabilistic Representation [59.45669299295436]
We propose a Monte Carlo PDE solver for training unsupervised neural solvers.<n>We use the PDEs' probabilistic representation, which regards macroscopic phenomena as ensembles of random particles.<n>Our experiments on convection-diffusion, Allen-Cahn, and Navier-Stokes equations demonstrate significant improvements in accuracy and efficiency.
arXiv Detail & Related papers (2023-02-10T08:05:19Z) - Sampling with Mollified Interaction Energy Descent [57.00583139477843]
We present a new optimization-based method for sampling called mollified interaction energy descent (MIED)
MIED minimizes a new class of energies on probability measures called mollified interaction energies (MIEs)
We show experimentally that for unconstrained sampling problems our algorithm performs on par with existing particle-based algorithms like SVGD.
arXiv Detail & Related papers (2022-10-24T16:54:18Z) - Green, Quantized Federated Learning over Wireless Networks: An
Energy-Efficient Design [68.86220939532373]
The finite precision level is captured through the use of quantized neural networks (QNNs) that quantize weights and activations in fixed-precision format.
The proposed FL framework can reduce energy consumption until convergence by up to 70% compared to a baseline FL algorithm.
arXiv Detail & Related papers (2022-07-19T16:37:24Z) - Mean-field Variational Inference via Wasserstein Gradient Flow [8.05603983337769]
Variational inference, such as the mean-field (MF) approximation, requires certain conjugacy structures for efficient computation.
We introduce a general computational framework to implement MFal inference for Bayesian models, with or without latent variables, using the Wasserstein gradient flow (WGF)
We propose a new constraint-free function approximation method using neural networks to numerically realize our algorithm.
arXiv Detail & Related papers (2022-07-17T04:05:32Z) - A hybrid MGA-MSGD ANN training approach for approximate solution of
linear elliptic PDEs [0.0]
We introduce a hybrid "Modified Genetic-Multilevel Gradient Descent" (MGA-MSGD) training algorithm.
It considerably improves accuracy and efficiency of solving 3D mechanical problems described, in strong-form, by PDEs via ANNs.
arXiv Detail & Related papers (2020-12-18T10:59:07Z) - Amortized Conditional Normalized Maximum Likelihood: Reliable Out of
Distribution Uncertainty Estimation [99.92568326314667]
We propose the amortized conditional normalized maximum likelihood (ACNML) method as a scalable general-purpose approach for uncertainty estimation.
Our algorithm builds on the conditional normalized maximum likelihood (CNML) coding scheme, which has minimax optimal properties according to the minimum description length principle.
We demonstrate that ACNML compares favorably to a number of prior techniques for uncertainty estimation in terms of calibration on out-of-distribution inputs.
arXiv Detail & Related papers (2020-11-05T08:04:34Z) - Training Deep Energy-Based Models with f-Divergence Minimization [113.97274898282343]
Deep energy-based models (EBMs) are very flexible in distribution parametrization but computationally challenging.
We propose a general variational framework termed f-EBM to train EBMs using any desired f-divergence.
Experimental results demonstrate the superiority of f-EBM over contrastive divergence, as well as the benefits of training EBMs using f-divergences other than KL.
arXiv Detail & Related papers (2020-03-06T23:11:13Z) - Learning Likelihoods with Conditional Normalizing Flows [54.60456010771409]
Conditional normalizing flows (CNFs) are efficient in sampling and inference.
We present a study of CNFs where the base density to output space mapping is conditioned on an input x, to model conditional densities p(y|x)
arXiv Detail & Related papers (2019-11-29T19:17:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.