I was reading Boyd & Vandenberghe's "Convex Optimization" (Ch10), where the following equality constrained convex program is considered (p526):
\begin{equation} \begin{array}{rl} \textrm{minimize} & f(x)\\ \textrm{subject to} & Ax=b, \end{array} \end{equation}
and Newton's method is employed to solve the optimality conditions, $$Ax=b, \qquad \nabla f(x)+A^Tw=0.$$
More specifically, given a feasible point $x$ (which satisfies $Ax=b$), we find the next feasible point, $x+\Delta x_{nt}$, by solving
$$\left[\begin{array}{cc}\nabla^2 f(x) & A^T\\A & 0\end{array}\right] \left[\begin{array}{c}\Delta x_{nt} \\ w\end{array}\right] = \left[\begin{array}{c}-\nabla f(x) \\ 0\end{array}\right].$$
This is straightforward, if the so-called KKT matrix,$\left[\begin{array}{cc}\nabla^2 f(x) & A^T\\A & 0\end{array}\right]$, is nonsingular (invertible). And it is stated in the book that the Newton step, $\Delta x_{nt}$, is defined only at points for which the KKT matrix is non-singular.
So what do we do when the KKT matrix happen to be singular during the iterations? Conceivably, if the above system of equations is consistent, we can just pick one solution as $\Delta x_{nt}$ and continue with the iterations, right?
However, if the above system of equations is not even consistent, it's not clear to me what a good strategy would be? Are there better ways than just picking another feasible initial point and restarting the Newton's method all over again, wishing that we may end up with a sequence of non-singular KKT matrices that lead to an optimal (acceptable) solution?
Does it make sense to use the projection of the negative gradient $-\nabla f(x)$ on the null space of matrix $A$ as $\Delta x_{nt}$? (This ensures that $x+\Delta x_{nt}$ remains feasible, but it's not clear if it's still a descent direction?)
I'd appreciate any pointers or comments. Thanks a lot!