Super Level Sets and Exponential Decay: A Synergistic Approach to Stable Neural Network Training

dc.contributor.authorChaudary, Jatin
dc.contributor.authorNidhi, Dipak
dc.contributor.authorHeikkonen, Jukka
dc.contributor.authorMerisaari, Harri
dc.contributor.authorKanth, Rajiv
dc.contributor.organizationfi=data-analytiikka|en=Data-analytiikka|
dc.contributor.organizationfi=kliiniset neurotieteet|en=Clinical Neurosciences|
dc.contributor.organization-code1.2.246.10.2458963.20.68940835793
dc.contributor.organization-code1.2.246.10.2458963.20.74845969893
dc.converis.publication-id499842072
dc.converis.urlhttps://research.utu.fi/converis/portal/Publication/499842072
dc.date.accessioned2026-01-21T15:02:38Z
dc.date.available2026-01-21T15:02:38Z
dc.description.abstract<p>This paper presents a theoretically grounded optimization framework for neural network training that integrates an Exponentially Decaying Learning Rate with Lyapunov-based stability analysis. We develop a dynamic learning rate algorithm and prove that it induces connected and stable descent paths through the loss landscape by maintaining the connectivity of super-level sets 𝑆𝜆={𝜃∈R𝑛:L(𝜃) ≥𝜆}. Under the condition that the Lyapunov function 𝑉(𝜃)=L(𝜃)satisfies∇𝑉(𝜃)·∇L(𝜃) ≥0, we establish that these super-level sets are not only connected but also equiconnected across epochs, providing uniform topological stability. We further derive convergence guarantees using a second-order Taylor expansion and demonstrate that our exponentially scheduled learning rate with gradient-based modulation leads to a monotonic decrease in loss. The proposed algorithm incorporates this schedule into a stability-aware update mechanism that adapts step sizes based on both curvature and energy-level geometry. This work formalizes the role of topological structure in convergence dynamics and introduces a provably stable optimization algorithm for high-dimensional, non-convex neural networks.</p>
dc.identifier.eissn1943-5037
dc.identifier.jour-issn1076-9757
dc.identifier.olddbid214031
dc.identifier.oldhandle10024/197049
dc.identifier.urihttps://www.utupub.fi/handle/11111/56260
dc.identifier.urlhttps://doi.org/10.1613/jair.1.17272
dc.identifier.urnURN:NBN:fi-fe202601216436
dc.language.isoen
dc.okm.affiliatedauthorChaudhary, Jatin
dc.okm.affiliatedauthorNidhi, Dipak
dc.okm.affiliatedauthorHeikkonen, Jukka
dc.okm.affiliatedauthorMerisaari, Harri
dc.okm.affiliatedauthorKanth, Rajeev
dc.okm.discipline113 Computer and information sciencesen_GB
dc.okm.discipline113 Tietojenkäsittely ja informaatiotieteetfi_FI
dc.okm.internationalcopublicationnot an international co-publication
dc.okm.internationalityInternational publication
dc.okm.typeA1 ScientificArticle
dc.publisherAI Access Foundation
dc.publisher.countryUnited Statesen_GB
dc.publisher.countryYhdysvallat (USA)fi_FI
dc.publisher.country-codeUS
dc.relation.articlenumber21
dc.relation.doi10.1613/jair.1.17272
dc.relation.ispartofjournalJournal of Artificial Intelligence Research
dc.relation.volume83
dc.source.identifierhttps://www.utupub.fi/handle/10024/197049
dc.titleSuper Level Sets and Exponential Decay: A Synergistic Approach to Stable Neural Network Training
dc.year.issued2025

Tiedostot

Näytetään 1 - 1 / 1
Ladataan...
Name:
17272b.pdf
Size:
8.35 MB
Format:
Adobe Portable Document Format