Semi-gradients are commonly used for function approximation updates. While most deep learning methods follow some variant of ⛰️ Gradient Descent, reinforcement learning updates sometimes require bootstrapping—that is, our function estimate is part of the target, and thus updating the function with a standard gradient descent update gives us a semi-gradient instead of a true gradient.
For example, let’s approximate
Note that this update is exactly the standard gradient update for some target