Computing the action ground state for the rotating nonlinear
Schr\"odinger equation
- URL: http://arxiv.org/abs/2203.06383v2
- Date: Wed, 26 Apr 2023 05:18:02 GMT
- Title: Computing the action ground state for the rotating nonlinear
Schr\"odinger equation
- Authors: Wei Liu, Yongjun Yuan, Xiaofei Zhao
- Abstract summary: We consider the computations of the action ground state for a rotating nonlinear Schr"odinger equation.
In the focusing case, we identify an equivalent formulation of the problem which simplifies the constraint.
In the defocusing case, we prove that the ground state can be obtained by the unconstrained minimization.
- Score: 6.6772808699409705
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We consider the computations of the action ground state for a rotating
nonlinear Schr\"odinger equation. It reads as a minimization of the action
functional under the Nehari constraint. In the focusing case, we identify an
equivalent formulation of the problem which simplifies the constraint. Based on
it, we propose a normalized gradient flow method with asymptotic Lagrange
multiplier and establish the energy-decaying property. Popular optimization
methods are also applied to gain more efficiency. In the defocusing case, we
prove that the ground state can be obtained by the unconstrained minimization.
Then the direct gradient flow method and unconstrained optimization methods are
applied. Numerical experiments show the convergence and accuracy of the
proposed methods in both cases, and comparisons on the efficiency are
discussed. Finally, the relation between the action and the energy ground
states are numerically investigated.
Related papers
- The inexact power augmented Lagrangian method for constrained nonconvex optimization [44.516958213972885]
This work introduces an unconventional augmented Lagrangian term, where the augmenting term is a Euclidean norm raised to a power.
We show that using lower powers for augmenting term to faster rate, albeit with a slower decrease in residual.
Our results further show that using lower powers for augmenting term to faster rate, albeit with a slower decrease in residual.
arXiv Detail & Related papers (2024-10-26T11:31:56Z) - An Adaptive Stochastic Gradient Method with Non-negative Gauss-Newton Stepsizes [17.804065824245402]
In machine learning applications, each loss function is non-negative and can be expressed as the composition of a square and its real-valued square root.
We show how to apply the Gauss-Newton method or the Levssquardt method to minimize the average of smooth but possibly non-negative functions.
arXiv Detail & Related papers (2024-07-05T08:53:06Z) - Efficient Low-rank Identification via Accelerated Iteratively Reweighted Nuclear Norm Minimization [8.879403568685499]
We introduce an adaptive updating strategy for smoothing parameters.
This behavior transforms the algorithm into one that effectively solves problems after a few iterations.
We prove the global proposed experiment, guaranteeing that every iteration is a critical one.
arXiv Detail & Related papers (2024-06-22T02:37:13Z) - Incremental Quasi-Newton Methods with Faster Superlinear Convergence
Rates [50.36933471975506]
We consider the finite-sum optimization problem, where each component function is strongly convex and has Lipschitz continuous gradient and Hessian.
The recently proposed incremental quasi-Newton method is based on BFGS update and achieves a local superlinear convergence rate.
This paper proposes a more efficient quasi-Newton method by incorporating the symmetric rank-1 update into the incremental framework.
arXiv Detail & Related papers (2024-02-04T05:54:51Z) - D4FT: A Deep Learning Approach to Kohn-Sham Density Functional Theory [79.50644650795012]
We propose a deep learning approach to solve Kohn-Sham Density Functional Theory (KS-DFT)
We prove that such an approach has the same expressivity as the SCF method, yet reduces the computational complexity.
In addition, we show that our approach enables us to explore more complex neural-based wave functions.
arXiv Detail & Related papers (2023-03-01T10:38:10Z) - Sampling with Mollified Interaction Energy Descent [57.00583139477843]
We present a new optimization-based method for sampling called mollified interaction energy descent (MIED)
MIED minimizes a new class of energies on probability measures called mollified interaction energies (MIEs)
We show experimentally that for unconstrained sampling problems our algorithm performs on par with existing particle-based algorithms like SVGD.
arXiv Detail & Related papers (2022-10-24T16:54:18Z) - Second-order flows for computing the ground states of rotating
Bose-Einstein condensates [5.252966797394752]
Some artificial evolutionary differential equations involving second-order time derivatives are considered to be first-order.
The proposed artificial dynamics are novel second-order hyperbolic partial differential equations with dissipation.
New algorithms are superior to the state-of-the-art numerical methods based on the gradient flow.
arXiv Detail & Related papers (2022-05-02T10:45:49Z) - Whiplash Gradient Descent Dynamics [2.0508733018954843]
We introduce the symplectic convergence analysis for the Whiplash system for convex functions.
We study the algorithm's performance for various costs and provide a practical methodology for analyzing convergence rates.
arXiv Detail & Related papers (2022-03-04T05:47:26Z) - Zeroth-Order Hybrid Gradient Descent: Towards A Principled Black-Box
Optimization Framework [100.36569795440889]
This work is on the iteration of zero-th-order (ZO) optimization which does not require first-order information.
We show that with a graceful design in coordinate importance sampling, the proposed ZO optimization method is efficient both in terms of complexity as well as as function query cost.
arXiv Detail & Related papers (2020-12-21T17:29:58Z) - Conditional gradient methods for stochastically constrained convex
minimization [54.53786593679331]
We propose two novel conditional gradient-based methods for solving structured convex optimization problems.
The most important feature of our framework is that only a subset of the constraints is processed at each iteration.
Our algorithms rely on variance reduction and smoothing used in conjunction with conditional gradient steps, and are accompanied by rigorous convergence guarantees.
arXiv Detail & Related papers (2020-07-07T21:26:35Z) - Cogradient Descent for Bilinear Optimization [124.45816011848096]
We introduce a Cogradient Descent algorithm (CoGD) to address the bilinear problem.
We solve one variable by considering its coupling relationship with the other, leading to a synchronous gradient descent.
Our algorithm is applied to solve problems with one variable under the sparsity constraint.
arXiv Detail & Related papers (2020-06-16T13:41:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.