Slow Learners are Fast
Online learning algorithms have impressive convergence properties when it comes to risk minimization and convex games on very large problems. However, they are inherently sequential in their design which prevents them from taking advantage of modern multi-core architectures. In this paper we prove that online learning with delayed updates converges well, thereby facilitating parallel online learning.
š” Research Summary
The paper addresses a fundamental tension in modern machine learning: online learning algorithms enjoy strong theoretical guarantees for risk minimization and convex games, yet their classic designs are inherently sequential, preventing efficient exploitation of todayās multiācore and distributed hardware. The authors propose and rigorously analyze a ādelayedāupdateā variant of online learning, showing that allowing updates to be applied after a bounded or stochastic lag does not destroy the algorithmās convergence properties.
First, the authors formalize the delayed model. At iterationāÆt the learner receives a loss function ā_t and computes a subāgradient g_t, but the parameter vector w is updated only using a subāgradient from Ļ steps earlier: w_{t+1}=w_tāĪ·_tāÆg_{tāĻ}. Under the standard assumptions that each loss is convex and LāLipschitz, and that the delay Ļ is either a known constant or has a finite expectation, they derive a regret bound of the form
R_T ⤠(D²)/(2Ī·) + Ī·āÆL²āÆT + LāÆDāÆĻ,
where D is the diameter of the feasible set and Ī· is the learning rate. By choosing Ī·āD/(LāT) the bound simplifies to O(āTāÆ+āÆĻ). Consequently, even with a nonāzero delay, the average regret converges to zero at the same āT rate as the classic, instantaneousāupdate algorithm.
The analysis is then extended to stochastic delays. Assuming Ļ_t are independent random variables with E
Comments & Academic Discussion
Loading comments...
Leave a Comment