Trading-off static and dynamic regret in online least-squares and beyond

Jianjun Yuan, Andrew Lamperski

Research output: Chapter in Book/Report/Conference proceedingConference contribution

8 Scopus citations

Abstract

Recursive least-squares algorithms often use forgetting factors as a heuristic to adapt to non-stationary data streams. The first contribution of this paper rigorously characterizes the effect of forgetting factors for a class of online Newton algorithms. For exp-concave and strongly convex objectives, the algorithms achieve the dynamic regret of max{O(log T), O(√TV )}, where V is a bound on the path length of the comparison sequence. In particular, we show how classic recursive least-squares with a forgetting factor achieves this dynamic regret bound. By varying V , we obtain a trade-off between static and dynamic regret. In order to obtain more computationally efficient algorithms, our second contribution is a novel gradient descent step size rule for strongly convex functions. Our gradient descent rule recovers the order optimal dynamic regret bounds described above. For smooth problems, we can also obtain static regret of O(T1β) and dynamic regret of O(TβV ), where β ∈ (0, 1) and V is the path length of the sequence of minimizers. By varying β, we obtain a trade-off between static and dynamic regret.

Original languageEnglish (US)
Title of host publicationAAAI 2020 - 34th AAAI Conference on Artificial Intelligence
PublisherAAAI press
Pages6712-6719
Number of pages8
ISBN (Electronic)9781577358350
StatePublished - 2020
Event34th AAAI Conference on Artificial Intelligence, AAAI 2020 - New York, United States
Duration: Feb 7 2020Feb 12 2020

Publication series

NameAAAI 2020 - 34th AAAI Conference on Artificial Intelligence

Conference

Conference34th AAAI Conference on Artificial Intelligence, AAAI 2020
Country/TerritoryUnited States
CityNew York
Period2/7/202/12/20

Bibliographical note

Funding Information:
∗Work supported in part by NSF CMMI 1727096 and a University of MnDRIVE Informatics Graduate Fellowship. Copyright ©c 2020, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved.

Publisher Copyright:
Copyright © 2020, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved.

Fingerprint

Dive into the research topics of 'Trading-off static and dynamic regret in online least-squares and beyond'. Together they form a unique fingerprint.

Cite this