Preconditioned Sharpness-Aware Minimization: Unifying Analysis and a Novel Learning Algorithm

Research output: Contribution to journalConference articlepeer-review

1 Scopus citations

Abstract

Targeting solutions over 'flat' regions of the loss landscape, sharpness-aware minimization (SAM) has emerged as a powerful tool to improve generalizability of deep neural network based learning. While several SAM variants have been developed to this end, a unifying approach that also guides principled algorithm design has been elusive. This contribution leverages preconditioning (pre) to unify SAM variants and provide not only unifying convergence analysis, but also valuable insights. Building upon preSAM, a novel algorithm termed infoSAM is introduced to address the so-called adversarial model degradation issue in SAM by adjusting gradients depending on noise estimates. Extensive numerical tests demonstrate the superiority of infoSAM across various benchmarks.

Bibliographical note

Publisher Copyright:
© 2025 Institute of Electrical and Electronics Engineers Inc.. All rights reserved.

Keywords

  • convergence analysis
  • generalizability
  • machine learning
  • preconditioning
  • sharpness-aware minimization

Fingerprint

Dive into the research topics of 'Preconditioned Sharpness-Aware Minimization: Unifying Analysis and a Novel Learning Algorithm'. Together they form a unique fingerprint.

Cite this