版权所有:内蒙古大学图书馆 技术提供:维普资讯• 智图
内蒙古自治区呼和浩特市赛罕区大学西街235号 邮编: 010021
作者机构:Institut für Atmosphäre und Umwelt Goethe-Universität Frankfurt Frankfurt/Main60438 Germany Mathematical Institute for Machine Learning and Data Science KU Eichstätt–Ingolstadt Ingolstadt85049 Germany School of Engineering and Science Jacobs University Bremen28759 Germany
出 版 物:《arXiv》 (arXiv)
年 卷 期:2022年
核心收录:
摘 要:Optimal balance is a non-asymptotic numerical method for computing a point on an elliptic slow manifold for two-scale dynamical systems with strong gyroscopic forces. It works by solving a modified differential equation as a boundary value problem in time, where the nonlinear terms are adiabatically ramped up from zero to the fully nonlinear dynamics. A dedicated boundary value solver, however, is often not directly available. The most natural alternative is a nudging solver, where the problem is repeatedly solved forward and backward in time and the respective boundary conditions are restored whenever one of the temporal end points is visited. In this paper, we show quasi-convergence of this scheme in the sense that the termination residual of the nudging iteration is as small as the asymptotic error of the method itself, i.e., under appropriate assumptions exponentially small. This confirms that optimal balance in its nudging formulation is an effective algorithm. Further, it shows that the boundary value problem formulation of optimal balance is well posed up at most a residual error as small as the asymptotic error of the method itself. The key step in our proof is a careful two-component Gronwall *** Codes Primary 34E13, Secondary 34B15, 37M21 Copyright © 2022, The Authors. All rights reserved.