TY - JOUR
T1 - A stochastic use of the Kurdyka-Łojasiewicz property: Investigation of optimization algorithms behaviours in a non-convex differentiable framework
AU - Fest, Jean-Baptiste
AU - Repetti, Audrey
AU - Chouzenoux, Emilie
PY - 2025/8/18
Y1 - 2025/8/18
N2 - Asymptotic analysis of generic stochastic algorithms often relies on descent conditions. In a convex setting, some technical shortcuts can be considered to establish asymptotic convergence guarantees of the associated scheme. However, in a non-convex setting, obtaining similar guarantees is usually more complicated, and relies on the use of the Kurdyka-Lojasiewicz (KL) property. While this tool has become popular in the field of deterministic optimization, it is much less widespread in the stochastic context and the few works making use of it are essentially based on trajectory-by-trajectory approaches. In this paper, we propose a new framework for using the KL property in a non-convex stochastic setting based on conditioning theory. We show that this framework allows for deeper asymptotic investigations on stochastic schemes verifying some generic descent conditions. We further show that our methodology can be used to prove convergence of generic stochastic gradient descent (SGD) schemes, and unifies conditions investigated in multiple articles of the literature.
AB - Asymptotic analysis of generic stochastic algorithms often relies on descent conditions. In a convex setting, some technical shortcuts can be considered to establish asymptotic convergence guarantees of the associated scheme. However, in a non-convex setting, obtaining similar guarantees is usually more complicated, and relies on the use of the Kurdyka-Lojasiewicz (KL) property. While this tool has become popular in the field of deterministic optimization, it is much less widespread in the stochastic context and the few works making use of it are essentially based on trajectory-by-trajectory approaches. In this paper, we propose a new framework for using the KL property in a non-convex stochastic setting based on conditioning theory. We show that this framework allows for deeper asymptotic investigations on stochastic schemes verifying some generic descent conditions. We further show that our methodology can be used to prove convergence of generic stochastic gradient descent (SGD) schemes, and unifies conditions investigated in multiple articles of the literature.
U2 - 10.3934/fods.2025016
DO - 10.3934/fods.2025016
M3 - Article
SN - 2639-8001
JO - Foundations of Data Science
JF - Foundations of Data Science
ER -