Abstract
This work focuses on developing and motivating a stochastic version of a wellknown inverse problem methodology. Specifically, we consider the iteratively regularized Gauss–Newton method, originally proposed by Bakushinskii for infinite-dimensional problems. Recent work have extended this method to handle sequential observations, rather than a single instance of the data, demonstrating notable improvements in reconstruction accuracy. In this paper, we further extend these methods to a stochastic framework through mini-batching, introducing a new algorithm, the stochastic iteratively regularized Gauss–Newton method (SIRGNM). Our algorithm is designed through the use randomized sketching. We provide an analysis for the SIRGNM, which includes a preliminary error decomposition and a convergence analysis, related to the residuals. We provide numerical experiments on a 2D elliptic partial differential equation example. This illustrates the effectiveness of the SIRGNM, through maintaining a similar level of accuracy while reducing on the computational time. © 2024 The Author(s).
Original language | English |
---|---|
Article number | 015005 |
Journal | Inverse Problems |
Volume | 41 |
Issue number | 1 |
Online published | 23 Dec 2024 |
DOIs | |
Publication status | Published - Jan 2025 |
Funding
N K C is supported by an EPSRC-UKRI AI for Net Zero Grant: ‘Enabling CO2 Capture And Storage Projects Using AI’, (Grant EP/Y006143/1). NKC is also supported by a City University of Hong Kong startup grant.
Research Keywords
- stochastic optimization
- inverse problems
- regularization
- Gauss–Newton method
- convergence analysis
- random projection
Publisher's Copyright Statement
- This full text is made available under CC-BY 4.0. https://creativecommons.org/licenses/by/4.0/