Material Detail

Convergence Rates of Inexact Proximal-Gradient Methods for Convex Optimization

Convergence Rates of Inexact Proximal-Gradient Methods for Convex Optimization

This video was recorded at 25th Annual Conference on Neural Information Processing Systems (NIPS), Granada 2011. We consider the problem of optimizing the sum of a smooth convex function and a non-smooth convex function using proximal-gradient methods, where an error is present in the calculation of the gradient of the smooth term or in the proximity operator with respect to the second term. We show that the basic proximal-gradient method, the basic proximal-gradient method with a strong convexity assumption, and the accelerated proximal-gradient method achieve the same convergence rates as in the error-free case, provided the errors decrease at an appropriate rate. Our experimental results on a structured sparsity problem indicate that sequences of errors with these appealing theoretical properties can lead to practical performance improvements.

Quality

  • User Rating
  • Comments
  • Learning Exercises
  • Bookmark Collections
  • Course ePortfolios
  • Accessibility Info

More about this material

Browse...

Disciplines with similar materials as Convergence Rates of Inexact Proximal-Gradient Methods for Convex Optimization

Comments

Log in to participate in the discussions or sign up if you are not already a MERLOT member.