0
$\begingroup$

Is there any official (i.e., to be found in probability books) metric for the distance between two probability measures, defined only on a subset of their support?

Take, for example, the total variation distance: $TV(\mu,\nu)=\sup_{A\in\mathcal{F}}|\mu(A)-\nu(A)|.$

If $X$ and $Y$ are two real positive continuous random variables with densities $f_X$ and $f_Y$, then their total variation distance is, if I understand correctly: $TV(\mu_X,\mu_Y)=\int_{0}^{\infty}|f_X(z)−f_Y(z)|dz.$

Would it make any sense to calculate a quantity, for $\tau>0$, let's call it partial distance, like this: $PV(\mu_X,\mu_Y;\tau)=\int_{\tau}^{\infty}|f_X(z)−f_Y(z)|dz\;\;\;?$

If this does not make any sense (sorry, I really cannot tell, as I am not that good with measure theory...), can anyone think of a measure that would make sense?

What I want to use this for is to compare the closeness of two PDFs (or other functions describing a distribution: CDF, CCDF...) $f_X(t)$, $f_Y(t)$ to a third one $f_Z(t)$. I know that both $f_X$ and $f_Y$ "eventually" ($t\to\infty$) converge to $f_Z$, but I would like to show that one of them gets closer, sooner than the other one...

3 Answers 3

0

Something like this is what I was looking for:

"A measure of discrimination between two residual life-time distributions and its applications" by Nader Ebrahimi and S.N.U.A. Kirmani from the Annals of the Institute of Statistical Mathematics, Volume 48, Number 2 (1996), 257-265

Can be downloaded from:

(official) http://www.springerlink.com/content/r186240907v85964/

(maybe official?) http://www.ism.ac.jp/editsec/aism/pdf/048_2_0257.pdf

EDIT: I realize the proposal in the paper is not a metric (because not symmetric), but just to give an example of what I have in mind... Would a similar adaptation be possible for other measures of distance which are metrics?

0

In statistics there are several different measures of distance that can be used depending on application. When determining whether or not two multivariate normal distributions with the same covariance matrix have the same mean vector the Mahalanobis distance is used. This can be applied in discriminant analysis or cluster analysis. Link:http://en.wikipedia.org/wiki/Mahalanobis_distance

From the perspective of information theory people often use the Kullback-Leiber distance. Link:http://en.wikipedia.org/wiki/Kullback%E2%80%93Leibler_divergence The following link gives these measures and several more:

http://en.wikipedia.org/wiki/Category:Statistical_distance_measures

  • 0
    I could see such a measure being useful for comparing symmetric distributions, normal vs t vs Cauchy and a one sided measure for distributions like the chi square, lognormal and F.2012-07-26
0

I gave you this same answer over at MathOverflow. Looking at your response to Michael Chernick, you probably do want to consult Dudley, as the Prohorov metric and its follow-up in Proposition 11.3.2 refer directly to metrics on random variables, which could be defined for the tail only as you request.

You may want to check out Real Analysis and Probability by R. M. Dudley (2002, Cambridge University Press). Chapters 9-11 discuss several metrics on probability measures and random variables (laws), and since restricting your support would be equivalent to some random variable on the measure, you should be able to use something like the metrics discussed in section 11.3 in particular.

  • 0
    Great! Thanks for the explanation, this is one thing that I was very curious about (the difference between integrating on the whole domain vs only part of it).2012-07-28