What is the difference between projected gradient descent and ordinary gradient descent?

I just read about projected gradient descent but I did not see the intuition to use Projected one instead of normal gradient descent. Would you tell me the reason and preferable situations of projected gradient descent? What does that projection contribute?

Answer

At a basic level, projected gradient descent is just a more general method for solving a more general problem.

Gradient descent minimizes a function by moving in the negative gradient direction at each step. There is no constraint on the variable.
Problem 1:minxf(x)
xk+1=xktkf(xk)

On the other hand, projected gradient descent minimizes a function subject to a constraint. At each step we move in the direction of the negative gradient, and then “project” onto the feasible set.

Problem 2:minxf(x) subject to xC

yk+1=xktkf(xk)xk+1=argminxC

Attribution
Source : Link , Question Author : erogol , Answer Author : p.s.

Leave a Comment