I'm reading through the deepest descent method, and I'm struggling to understand a specific bound. I'll try to be as much clear as I can, but the notation used is a bit messy.
It is basically explained what the deepest descent is, as a generalization of the Gauss-Seidel method. It is also said that it's NOT a good algorithm, but it is worth keep looking into its properties. You can assume we have to solve the problem
$$ \min_{x \in \mathbb{R}^n} f(x) $$
where $f(x) : \mathbb{R}^n \rightarrow \mathbb{R}$.
Theorem 2.11. in this book states the following
the (i) and (iii) are the followings:
Where $q(t)$ is defined as $f(x_k + td_k)$, so it's a parametric version of the next iteration, and for given direction we wish to find the best "step". The theorem is proved, and it make sense, however after such proof (that basically converges to a minima, or it keeps going further down) it is stated something else about the "convergence speed" and there's specifically the following bound I don't quite understand:
I don't get the bound
$$ f(x) \leq f^* + \frac{1}{2}\text{dist}^2(x,X^*) $$
can anyone tell me why is that? My guess is probably related to the mean value theorem and lipshitz conditions, but I can't really figure out why that's the bound.



