Central limit theorems for Sinkhorn divergence between probability distributions on finite spaces and statistical applications

2017 
The notion of Sinkhorn divergence has recently gained popularity in machine learning and statistics, as it makes feasible the use of smoothed optimal transportation distances for data analysis. The Sinkhorn divergence allows the fast computation of an entropically regularized Wasserstein distance between two probability distributions supported on a finite metric space of (possibly) high-dimension. For data sampled from one or two unknown probability distributions, we derive central limit theorems for empirical Sinkhorn divergences. We also propose a bootstrap procedure which allows to obtain new test statistics for measuring the discrepancies between multivariate probability distributions. The strategy of proof uses the notions of directional Hadamard differentiability and delta-method in this setting. It is inspired by the results in the work of Sommerfeld and Munk (2016) on the asymptotic distribution of empirical Wasserstein distance on finite space using un-regularized transportation costs. Simulated and real datasets are used to illustrate our approach. A comparison with existing methods to measure the discrepancy between multivariate distributions is also proposed.
    • Correction
    • Source
    • Cite
    • Save
    • Machine Reading By IdeaReader
    10
    References
    19
    Citations
    NaN
    KQI
    []