I came across this question about a classic homework problem:
Let $(X_n)$ be i.i.d., positive random variables. Compute $$ E\left[\frac{\sum_{i=1}^k X_i}{\sum_{j=1}^n X_j}\right]$$ for $k \le n.$
In the question, the asker links to a previous answer to the same question and explains that their professor thought it was wrong, or inadequate in some way. I'm wondering why.
The answer, paraphrased, is
By linearity and symmetry, $$ E\left[\frac{\sum_{i=1}^k X_i}{\sum_{j=1}^n X_j}\right] = \sum_{i=1}^kE\left[\frac{X_i}{\sum_{j=1}^n X_j}\right] = kE\left[\frac{X_1}{\sum_{j=1}^n X_j}\right].$$ If $k=n,$ the answer is obviously $1,$ so we must have $$E\left[\frac{X_1}{\sum_{j=1}^n X_j}\right] = \frac{1}{n}.$$ Thus the answer is $\frac{k}{n}.$
They said their professor said this was not really a 'solution' and that they instead needed to condition on the denominator. Particularly, to define $M = \sum_{j=1}^nX_j$ and consider $$ E\left( \frac{\sum_{i=1}^k X_i}{M}\mid M = m\right)$$ where m is a positive integer. Then using the law of total probability or iterated expectation (as you can see in the two answers), the proof goes through much as before, using symmetry and linearity.
Nevermind the fact that there's no need for $M$ to be an integer. Let's take it for granted and assume that the $X_i$ are integer-valued RVs. (In fact the law of total probability answer only needs this assumption for convenience and the law of iterated expectations answer doesn't need it at all.)
What I don't get is how this conditioning improves the correctness of the solution at all. Can anyone think of a good reason?
Perhaps the professor forgot how linearity/numerators work and thought it was illegal to apply linearity until we'd pulled the denominator out of the conditional expectation? Or maybe there's some subtlety about the use of symmetry or linearity that I'm missing?