Online Learning of Noisy Data with Kernels

Online Learning of Noisy Data with Kernels
Notice: This research summary and analysis were automatically generated using AI technology. For absolute accuracy, please refer to the [Original Paper Viewer] below or the Original ArXiv Source.

We study online learning when individual instances are corrupted by adversarially chosen random noise. We assume the noise distribution is unknown, and may change over time with no restriction other than having zero mean and bounded variance. Our technique relies on a family of unbiased estimators for non-linear functions, which may be of independent interest. We show that a variant of online gradient descent can learn functions in any dot-product (e.g., polynomial) or Gaussian kernel space with any analytic convex loss function. Our variant uses randomized estimates that need to query a random number of noisy copies of each instance, where with high probability this number is upper bounded by a constant. Allowing such multiple queries cannot be avoided: Indeed, we show that online learning is in general impossible when only one noisy copy of each instance can be accessed.


💡 Research Summary

The paper addresses the problem of online learning when each incoming instance is corrupted by adversarially chosen random noise. Unlike prior work that assumes a known noise distribution (often Gaussian) or a fixed variance, this study only requires that the noise have zero mean and bounded variance, and it may change arbitrarily over time. The authors introduce a family of unbiased estimators for arbitrary analytic functions, which become the cornerstone of their algorithmic design.

The key technical contribution is a construction that, given any analytic function f (including kernel functions), produces a stochastic estimator (\hat{f}) such that (\mathbb{E}


Comments & Academic Discussion

Loading comments...

Leave a Comment