Super Level Sets and Exponential Decay: A Synergistic Approach to Stable Neural Network Training
| dc.contributor.author | Chaudary, Jatin | |
| dc.contributor.author | Nidhi, Dipak | |
| dc.contributor.author | Heikkonen, Jukka | |
| dc.contributor.author | Merisaari, Harri | |
| dc.contributor.author | Kanth, Rajiv | |
| dc.contributor.organization | fi=data-analytiikka|en=Data-analytiikka| | |
| dc.contributor.organization | fi=kliiniset neurotieteet|en=Clinical Neurosciences| | |
| dc.contributor.organization-code | 1.2.246.10.2458963.20.68940835793 | |
| dc.contributor.organization-code | 1.2.246.10.2458963.20.74845969893 | |
| dc.converis.publication-id | 499842072 | |
| dc.converis.url | https://research.utu.fi/converis/portal/Publication/499842072 | |
| dc.date.accessioned | 2026-01-21T15:02:38Z | |
| dc.date.available | 2026-01-21T15:02:38Z | |
| dc.description.abstract | <p>This paper presents a theoretically grounded optimization framework for neural network training that integrates an Exponentially Decaying Learning Rate with Lyapunov-based stability analysis. We develop a dynamic learning rate algorithm and prove that it induces connected and stable descent paths through the loss landscape by maintaining the connectivity of super-level sets 𝑆𝜆={𝜃∈R𝑛:L(𝜃) ≥𝜆}. Under the condition that the Lyapunov function 𝑉(𝜃)=L(𝜃)satisfies∇𝑉(𝜃)·∇L(𝜃) ≥0, we establish that these super-level sets are not only connected but also equiconnected across epochs, providing uniform topological stability. We further derive convergence guarantees using a second-order Taylor expansion and demonstrate that our exponentially scheduled learning rate with gradient-based modulation leads to a monotonic decrease in loss. The proposed algorithm incorporates this schedule into a stability-aware update mechanism that adapts step sizes based on both curvature and energy-level geometry. This work formalizes the role of topological structure in convergence dynamics and introduces a provably stable optimization algorithm for high-dimensional, non-convex neural networks.</p> | |
| dc.identifier.eissn | 1943-5037 | |
| dc.identifier.jour-issn | 1076-9757 | |
| dc.identifier.olddbid | 214031 | |
| dc.identifier.oldhandle | 10024/197049 | |
| dc.identifier.uri | https://www.utupub.fi/handle/11111/56260 | |
| dc.identifier.url | https://doi.org/10.1613/jair.1.17272 | |
| dc.identifier.urn | URN:NBN:fi-fe202601216436 | |
| dc.language.iso | en | |
| dc.okm.affiliatedauthor | Chaudhary, Jatin | |
| dc.okm.affiliatedauthor | Nidhi, Dipak | |
| dc.okm.affiliatedauthor | Heikkonen, Jukka | |
| dc.okm.affiliatedauthor | Merisaari, Harri | |
| dc.okm.affiliatedauthor | Kanth, Rajeev | |
| dc.okm.discipline | 113 Computer and information sciences | en_GB |
| dc.okm.discipline | 113 Tietojenkäsittely ja informaatiotieteet | fi_FI |
| dc.okm.internationalcopublication | not an international co-publication | |
| dc.okm.internationality | International publication | |
| dc.okm.type | A1 ScientificArticle | |
| dc.publisher | AI Access Foundation | |
| dc.publisher.country | United States | en_GB |
| dc.publisher.country | Yhdysvallat (USA) | fi_FI |
| dc.publisher.country-code | US | |
| dc.relation.articlenumber | 21 | |
| dc.relation.doi | 10.1613/jair.1.17272 | |
| dc.relation.ispartofjournal | Journal of Artificial Intelligence Research | |
| dc.relation.volume | 83 | |
| dc.source.identifier | https://www.utupub.fi/handle/10024/197049 | |
| dc.title | Super Level Sets and Exponential Decay: A Synergistic Approach to Stable Neural Network Training | |
| dc.year.issued | 2025 |
Tiedostot
1 - 1 / 1