2000 character limit reached
Convergence rates for an inertial algorithm of gradient type associated to a smooth nonconvex minimization
Published 1 Jul 2018 in math.FA | (1807.00387v3)
Abstract: We investigate an inertial algorithm of gradient type in connection with the minimization of a nonconvex differentiable function. The algorithm is formulated in the spirit of Nesterov's accelerated convex gradient method. We show that the generated sequences converge to a critical point of the objective function, if a regularization of the objective function satisfies the Kurdyka-{\L}ojasiewicz property. Further, we provide convergence rates for the generated sequences and the function values formulated in terms of the {\L}ojasiewicz exponent.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.