2000 character limit reached
Convergence rates of a dual gradient method for constrained linear ill-posed problems
Published 15 Jun 2022 in math.NA, cs.NA, and math.OC | (2206.07379v1)
Abstract: In this paper we consider a dual gradient method for solving linear ill-posed problems $Ax = y$, where $A : X \to Y$ is a bounded linear operator from a Banach space $X$ to a Hilbert space $Y$. A strongly convex penalty function is used in the method to select a solution with desired feature. Under variational source conditions on the sought solution, convergence rates are derived when the method is terminated by either an {\it a priori} stopping rule or the discrepancy principle. We also consider an acceleration of the method as well as its various applications.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.