Assume $u$ is a solution of the wave equation in $\mathbb R^n \times [0,\infty)$ with $x \mapsto u(x,0)$ smooth with compact support. Why then must the support of $u(\cdot,t)$ be compact for all $t \geq 0$ ? Does this follow from the following theorem?
Domain of dependence: Assume suppose $u \in \mathcal C^2(\overline \Omega \times[0,T])$ solves the wave equation in $\Omega \times (0,T)$ and $u=g$ on $\Omega \times \{0\} \cup \partial \Omega \times[0,T)$ and $u_t(\cdot,0)=h$ on $\Omega$. Then if for some point $x_0 \in \Omega$ and $t_0 >0$ and $B_{t_0}(x_0) \subset \Omega$ we have
$u = u_t = 0$ on $B_{t_0}(x_0)\times\{0\}$, then we already have
$u=0$ on the cone $K = \{(x,t):0 \leq t \leq t_0, |x-x_0|\leq t_0-t\}$