I came across André Nicolas' answer to this question: Determining Convergence of Power Series
I have this mental visual representation of what expected value (in this case the expected number of rolls of a die before getting a 6, for ex) looks like, and it's something like this:
The 1 rectangle is $\frac{1}{6}$ of the total area (the probability of getting a six in one throw), the 2 rectangle is $\frac{1}{6}$ of $\frac{5}{6}$ of the total area, etc...without bounds, meaning that the subdivisions keep going forever.
If we assume that the total area (of the sample space) is 1, then the expected number of throws before getting a six is going to be the sum of [each subrectangle's area times that subrectangle's value].
What I'm trying to do is reconcile this geometric/visual model with the formula $$ E(X) = P(success) \cdot 1 + P(failure) \cdot (1 + E(X))$$
The formula does make sense, don't get me wrong, but only superficially (for me); it's not crystal clear. At some point in the algebra I go "hmm...why does this make sense again?"
I guess what I'm asking is whether there's an obvious way to see for example that the sum of each rectangle times its area (minus rectangle 1), is $\frac{5}{6} (1 + E(X))$
