Lazy Lagrangians for Optimistic Learning With Budget Constraints
Daron Anderson (Trinity College Dublin)
George Iosifidis (TU Delft - Embedded Systems)
Douglas Leith (Trinity College Dublin)
More Info
expand_more
Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.
Abstract
We consider the general problem of online convex optimization with time-varying budget constraints in the presence of predictions for the next cost and constraint functions, that arises in a plethora of network resource management problems. A novel saddle-point algorithm is designed by combining a Follow-The-Regularized-Leader iteration with prediction-adaptive dynamic steps. The algorithm achieves O(T(3β/4) regret and O(T(1+β)/2) constraint violation bounds that are tunable via parameter β ∈ [1/2,1) and have constant factors that shrink with the predictions quality, achieving eventually O(1) regret for perfect predictions. Our work extends the seminal FTRL framework for this new OCO setting and outperforms the respective state-of-the-art greedy-based solutions which naturally cannot benefit from predictions, without imposing conditions on the (unknown) quality of predictions, the cost functions or the geometry of constraints, beyond convexity.