Here is an example that I have been experimenting with students for the same purpose, understanding covariance and correlation, two highly related concepts, on a simple example.
In fact, I will stress the correlation aspect for two reasons:
Consider four measures of a same quantity, modelized by four i.i.d. (independent identically distributed) random variables $X_1, X_2, X_3, X_4.$
Let us now define random variables as the (positive or negative) gap between consecutive measures:
$$\begin{cases}Y_1&=&X_2-X_1\\Y_2&=&X_3-X_2\\Y_3&=&X_4-X_3\end{cases}$$
What is the correlation matrix of $(Y_1,Y_2,Y_3) ?$
A first remark is that $corr(Y_1,Y_3)=0$ because $X_3,X_4$ being independent from $X_1,X_2$, $Y_3$ is independent from $Y_1$.
On the contrary, $Y_2$ shouldn't be independent from $Y_1$, but what kind of correlation do we await ?
Let us assume that the $X_k$ have a uniform distribution on $[0,1]$.
Running the following simulation program, written in Matlab, but understandable without any knowledge of this language:
n=1000000;
X=rand(4,n);
Y=diff(X);
corrcoef(Y')
where $X$ is a $4 \times n$ array of $n$ realizations of $(X_1,X_2,X_3,X_4)$, $Y$ is a $3 \times n$ array of $n$ realizations of $(Y_1,Y_2,Y_3)$ and corrcoef(Y') is the $ 3 \times 3$ "experimental" matrix of correlation coefficients $r_{ij}=corr(Y_i,Y_j)$. Here is such a matrix obtained by running this program:
$$\pmatrix{1.0000&-0.4998&0.0014\\-0.4998&1.0000&-0.5020\\0.0014&-0.5020&1.0000} \ \ \ \text{to be compared to :} \ \ \ \pmatrix{1&-1/2&0\\-1/2&1&-1/2\\0&-1/2&1}$$
which is the theoretical matrix, that we are going to justify. More precisely, it suffices to justify that
$$\tag{1}corr(Y_1,Y_2)=-\tfrac12.$$
A rigorous justification of (1) will be given in the Appendix.
Nevertheless, in the spirit of your question, let us show that there is an intuitive explanation in (1), in particular about the negative sign of correlation coefficient. Imagine that $Y_1=X_2-X_1$ is bigger that $1/2$, meaning that $X_2$ is closer to $1$ than to $0$; in such a case, it is more likely that $X_3$ is smaller than $X_2$, i.e., that $Y_2=X_3-X_2<0$. It means that the knowledge of $Y_1$ permits to prognose, to a certain extend, the sign of $Y_2$.
A global understanding of this negative trend is as well provided by the following graphical representation where each point has coordinates $(Y_1,Y_2)$ (think to the equivalent ellipse): the cases where $Y_1$ and $Y_2$ have different signs are dominant.

Appendix : Rigorous proof of relationship (1)
Let us first recall that, if $X$ is a uniformly distributed random variable on $[0,1]$, then
$$\tag{2}V(X)=\frac{1}{12}.$$
As a direct consequence, due to the mutual independence of $X_1$ and $X_2$:
$$V(Y_1)=V(X_2-X_1)=V(X_2)+(-1)^2 V(X_1)=\frac{1}{6}$$
Using bilinearity property of covariance function, and the fact that the $X_k$ are mutually independent:
$cov(Y_1,Y_2)=cov(X_2-X_1,X_3-X_2)$
$=cov(X_2,X_3)-cov(X_2,X_2)-cov(X_1,X_3)+cov(X_1,X_2)$
$=0-var(X_2)-0+0=-\frac{1}{12}$
Thus, using ($2$),
$$corr(Y_1,Y_2)=\frac{cov(Y_1,Y_2)}{\sqrt{V(Y_1)V(Y_2)}}=\frac{-1/12}{1/6}=-\frac12,$$
as expected.
Remark: A similar computation could be done for $X_1,X_2,X_3,X_4$ having a $N(0,1)$ distribution. We would obtain the same correlation matrix.