If I do the derivative of the log-likelihood when the data follows a $Po(\lambda)$ distribution, we obtain the estimator $\hat \lambda =\bar X$. However, how do we prove that indeed this estimator does maximize the likelihood?
I've tried the 2nd order conditions, $\frac{\partial^2}{\partial \lambda^2}l(\lambda|\mathbf{x})=\frac{-\sum x_i}{\lambda^2}\leq 0$. If the inequality was strict, then we would be good. However, there's the not null chance that $\sum x_i=0$. In this case, we get $l(\lambda|\mathbf{x})=-\lambda n$, and would like to minimize the value of lambda, but however $\lambda>0$. How to proceed in this case?