This paper establishes a novel connection between local minimization principles for Sturm-Liouville equations and optimization techniques used in training neural networks. By interpreting the training of neural networks as a variational problem, we demonstrate how recent results on energy estimates for mixed boundary value problems in Sturm-Liouville theory can be adapted to analyze and improve neural network convergence. We present two main theorems: the first establishes conditions for guaranteed convergence to non-zero local minima in neural network training, and the second demonstrates the existence of multiple critical points with energy estimates. Our theoretical results are supported by experimental validation on benchmark datasets, showing improved performance in avoiding trivial solutions during training. This work bridges the gap between classical differential equation theory and modern machine learning optimization.

VARIATIONAL METHODS FOR NEURAL NETWORK TRAINING: APPLICATIONS OF STURM-LIOUVILLE ENERGY ESTIMATES / Ferrara, Massimiliano. - In: FAR EAST JOURNAL OF MATHEMATICAL SCIENCES: FJMS. - ISSN 0972-0871. - 142:3(2025), pp. 243-255. [10.17654/0972087125015]

VARIATIONAL METHODS FOR NEURAL NETWORK TRAINING: APPLICATIONS OF STURM-LIOUVILLE ENERGY ESTIMATES

Ferrara, Massimiliano
Conceptualization
2025-01-01

Abstract

This paper establishes a novel connection between local minimization principles for Sturm-Liouville equations and optimization techniques used in training neural networks. By interpreting the training of neural networks as a variational problem, we demonstrate how recent results on energy estimates for mixed boundary value problems in Sturm-Liouville theory can be adapted to analyze and improve neural network convergence. We present two main theorems: the first establishes conditions for guaranteed convergence to non-zero local minima in neural network training, and the second demonstrates the existence of multiple critical points with energy estimates. Our theoretical results are supported by experimental validation on benchmark datasets, showing improved performance in avoiding trivial solutions during training. This work bridges the gap between classical differential equation theory and modern machine learning optimization.
2025
variational methods, artificial neural networks, boundary value problems
File in questo prodotto:
File Dimensione Formato  
M.Ferrara_2025_FEJMS_Variational methods_editor.pdf

accesso aperto

Descrizione: Articolo
Tipologia: Versione Editoriale (PDF)
Licenza: Creative commons
Dimensione 1.09 MB
Formato Adobe PDF
1.09 MB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/20.500.12318/159486
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus ND
  • ???jsp.display-item.citation.isi??? ND
social impact