I am reading Markov chain and Mixing time and I do not understand the following two things:
A random mapping representation of a transition matrix P on a state $\Omega$ is a function $f:\Omega\times\Lambda \rightarrow \Omega$ along with a $\Lambda$-valued random variable $Z$, satisfying $$\textbf{P}\{f(x,Z)=y \}=P(x,y)$$ The reader should chech that if $Z_1,Z_2,\dots$ is a sequence of independent random variables, each having the same distribution as $Z$, and $X_0$ has distribution $\mu$, then the sequence $(X_0,X_1,\dots)$ defined by $$X_n=f(X_{n-1}, Z_n)$$ is a Markov chain with transition matrix $P$ and initial distribution $\mu$.
Here I missing something because I could not prove what the author asks. I ask for a calculation and to know where should the independence of the random variables $Z_1,Z_2,\cdots$ is used.
Thanks a lot!
EDIT: If $n=1$
$$\textbf{P}(X_1=y|X_0=x)=\textbf{P}(f(X_0,Z_1)=y|X_0=x)=\textbf{P}(f(x_0,Z)=y)=P(x,y)$$
and that
$$\textbf{P}(X_2=y|X_1=x_1,X_0=x)=\textbf{P}(f(X_1,Z_2)=y|X_1=x,X_0=x_0)\overset{?}{=}\textbf{P}(f(x,Z_2)=y|X_0=x_0)\overset{?}{=}P(x,y)$$
It the next I do the same: $$\textbf{P}(X_{t+1}=y|X_{t}=x,X_{t-1}=x_{t-1}\dots X_0=x_0)=\textbf{P}(f(X_t,Z_t)=y|X_{t}=x,X_{t-1}=x_{t-1}\dots X_0=x_0)=\textbf{P}(f(x,Z_t)=y|X_{t-1}=x_{t-1}\dots X_0=x_0)\overset{?}{=}\textbf{P}(X_{t+1}=y|X_{t}=x)\overset{?}{=}P(x,y)$$
But I don't know if this is right and if so why and where the independency is used