A Smoothing Stochastic Gradient Method for Composite Optimization

We consider the unconstrained optimization problem whose objective function is composed of a smooth and a non-smooth conponents where the smooth component is the expectation a random function. This ty

A Smoothing Stochastic Gradient Method for Composite Optimization

We consider the unconstrained optimization problem whose objective function is composed of a smooth and a non-smooth conponents where the smooth component is the expectation a random function. This type of problem arises in some interesting applications in machine learning. We propose a stochastic gradient descent algorithm for this class of optimization problem. When the non-smooth component has a particular structure, we propose another stochastic gradient descent algorithm by incorporating a smoothing method into our first algorithm. The proofs of the convergence rates of these two algorithms are given and we show the numerical performance of our algorithm by applying them to regularized linear regression problems with different sets of synthetic data.


💡 Research Summary

The paper addresses a class of unconstrained composite optimization problems that arise frequently in modern machine learning, where the objective function consists of a smooth stochastic component and a possibly nonsmooth regularizer:

  minₓ f(x) = E₍ξ₎


📜 Original Paper Content

🚀 Synchronizing high-quality layout from 1TB storage...