This research aims at focusing attention on Halpern iteration, a technique that can be exploited to define optimizers in neural network settings to outperform numerous current state-of-the-art methods. More specifically, we introduce HalpernSGD, an innovative network optimizer that leverages Halpern’s method to enhance the rate of convergence of the Stochastic Gradient Descent (SGD), reducing the carbon footprint associated with neural network training. HalpernSGD exhibits a quadratic rate of convergence compared to the one of SGD, without compromising the accuracy of the model. We compared it with SGD and Adam through experiments that demonstrate HalpernSGD’s superior efficiency by significantly reducing the number of epochs required for convergence, thereby lowering energy consumption and carbon emissions. The study also identifies potential improvements in Adam’s approach to stability and convergence, suggesting a future direction for developing combined optimizers.

HalpernSGD: A Halpern-Inspired Optimizer for Accelerated Neural Network Convergence and Reduced Carbon Footprint

Ritacco E.
2024-01-01

Abstract

This research aims at focusing attention on Halpern iteration, a technique that can be exploited to define optimizers in neural network settings to outperform numerous current state-of-the-art methods. More specifically, we introduce HalpernSGD, an innovative network optimizer that leverages Halpern’s method to enhance the rate of convergence of the Stochastic Gradient Descent (SGD), reducing the carbon footprint associated with neural network training. HalpernSGD exhibits a quadratic rate of convergence compared to the one of SGD, without compromising the accuracy of the model. We compared it with SGD and Adam through experiments that demonstrate HalpernSGD’s superior efficiency by significantly reducing the number of epochs required for convergence, thereby lowering energy consumption and carbon emissions. The study also identifies potential improvements in Adam’s approach to stability and convergence, suggesting a future direction for developing combined optimizers.
2024
9783031626999
9783031627002
File in questo prodotto:
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11390/1281224
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
  • ???jsp.display-item.citation.isi??? 0
social impact