When.com Web Search

Search results

  1. Results From The WOW.Com Content Network
  2. Statistical distance - Wikipedia

    en.wikipedia.org/wiki/Statistical_distance

    In statistics, probability theory, and information theory, a statistical distance quantifies the distance between two statistical objects, which can be two random variables, or two probability distributions or samples, or the distance can be between an individual sample point and a population or a wider sample of points. A distance between ...

  3. Total variation distance of probability measures - Wikipedia

    en.wikipedia.org/wiki/Total_variation_distance...

    In probability theory, the total variation distance is a distance measure for probability distributions. It is an example of a statistical distance metric, and is sometimes called the statistical distance , statistical difference or variational distance .

  4. Integral probability metric - Wikipedia

    en.wikipedia.org/wiki/Integral_probability_metric

    In probability theory, integral probability metrics are types of distance functions between probability distributions, defined by how well a class of functions can distinguish the two distributions. Many important statistical distances are integral probability metrics, including the Wasserstein-1 distance and the total variation distance .

  5. Probabilistic metric space - Wikipedia

    en.wikipedia.org/wiki/Probabilistic_metric_space

    Download as PDF; Printable version; In other projects ... denotes a distance between means of X and Y. Example. For example if both probability distribution functions ...

  6. Divergence (statistics) - Wikipedia

    en.wikipedia.org/wiki/Divergence_(statistics)

    In information geometry, a divergence is a kind of statistical distance: a binary function which establishes the separation from one probability distribution to another on a statistical manifold. The simplest divergence is squared Euclidean distance (SED), and divergences can be viewed as generalizations of SED.

  7. Fréchet distance - Wikipedia

    en.wikipedia.org/wiki/Fréchet_distance

    4 As a distance between probability distributions (the FID score) 5 Variants. 6 Examples. 7 Applications. 8 See also. ... Two Selected Topics (PDF), pp. ...

  8. Hellinger distance - Wikipedia

    en.wikipedia.org/wiki/Hellinger_distance

    In probability and statistics, the Hellinger distance (closely related to, although different from, the Bhattacharyya distance) is used to quantify the similarity between two probability distributions. It is a type of f-divergence. The Hellinger distance is defined in terms of the Hellinger integral, which was introduced by Ernst Hellinger in 1909.

  9. Wasserstein metric - Wikipedia

    en.wikipedia.org/wiki/Wasserstein_metric

    In mathematics, the Wasserstein distance or Kantorovich–Rubinstein metric is a distance function defined between probability distributions on a given metric space. It is named after Leonid Vaseršteĭn .