In optimization, sometimes you encounter functions or difference equations defined in terms of $\text{argmin}$ or $\text{argmax}$
$p: X \to X, x \mapsto \text{argmin}_{y \in X} \phi(y) + \|x-y\|^2$
I am puzzled by the fact we need to solve an optimization algorithm in order to evaluate these functions.
For example, consider $p: X \to X, x \mapsto \text{argmin}_{y \in X} \phi(y) + \|x-y\|^2$ .
On the one hand, the interpretation of this function is clear. Given a vector $x \in X$, we map it to a vector that gives the minimum of $\phi(y) + \|x-y\|^2$.
But on the other hand, we have no idea of this map is even well defined, since it is well known in optimization this map $\text{argmin}$ could have one, multiple, or no solution. So given a sequence of $x_k$, $p(x_k)$ could yield something like $y_1, \{y_{21},y_{22}\}, \text{no solution}, y_4, ...$
Further, it is not at all clear how this minimizer $y$ is found, it is usually not stated as part of the function. I would be much happier if the function was defined as "$p: X \to X, x \mapsto \text{argmin}_{y \in X} \phi(y) + \|x-y\|^2$ where we solve for $y$ using ... algorithm which guarantees a unique minimizer"
How do people in optimization deal with this ambiguity.