Statistical entropy estimation
Statistical estimation of information and statistical distances
SMU has been conducting research into entropy estimation for a number of years. The entropy of a random variable is the expected information gained by observing a realisation of that random variable. To estimate the entropy of a random vector, the naive approach is to partition the sample space into a finite number of cells, estimate the density in each cell by the proportion of sample points falling into that cell, then estimate the entropy by that of the associated empirical density function. For non-uniform distributions, fixed partitions lead to low occupancy numbers in sparse regions and poor resolution in dense regions. In a seminal paper (Kozachenko, Leonenko, 1989) an alternative approach is proposed for the problem of entropy estimation, based on the expected distance between a point and its nearest neighbour in the sample. Nearest neighbour relations define an adaptive partition of the sample space, which allows control of the number of points in each spatial cell, and hence the computation time of the algorithm.
Relative entropy and mutual information extend the notion of entropy to two or more random variable. Evans (2008a) showed that estimators based only on nearest neighbour relations in the marginal space can be computationally expensive, and shows how computational efficiency can be maintained by considering nearest neighbour relations in the joint probability space. Leonenko et al. (2008) presented a more general class of estimators for Renyi entropy and divergence, and showed that these estimators satisfy a Strong Law of Large Numbers. Evans (2008b) showed that this also holds for a broad class of nearest-neighbour statistics. Whether such statistics also satisfy a Central Limit Theorem is currently a subject of active research at Cardiff.
Entropy and divergence (Shannon and Kullback-Leibler) estimation is a central problem in image processing, with many applications for image compression, segmentation, calibration, registration, etc. Mutual information, which is strongly related to Shannon entropy and Kullback-Leibler divergence, is a widely used measure of similarity between images. We study the statistical inference theory for entropies, epsilon-entropies and statistical distance.
Cardiff Investigators
- Prof. Nikolai Leonenko
- Prof. Anatoly Zhigljavsky
- Dr. Dafydd Evans
- Dr. Jonathan Gillard
Collaborators
Dr. Luc Pronzato,
Dr. Oleg Seleznev,
Dr. Denis Denisov.
Selected publications
- Kozachenko, L.F., Leonenko, N.N. (1987) Sample estimate of entropy of a random vector. Problems of Information Transmission 23, 95-101.
- Evans, D. (2008a) A computationally efficient estimator for mutual information. Proceedings of the Royal Society A 464 (2093), 1203-1215. (doi:10.1098/rspa.2007.0196)
- Leonenko, N.N., Pronzato, L., Savani, V. (2008) A class of Renyi information estimators for multidimensional densities. Annals of Statistics 36 (5), 2153-2182. (doi:10.1214/07-AOS539).
- Evans, D. (2008b) A law of large numbers for nearest neighbour statistics. Proceedings of the Royal Society A 464 (2100), 3175-3192. (doi:10.1098/rspa.2008.0235)
- Leonenko, N.N., Seleznev, O. (2010) Statistical inference for ?-entropy and quadratic Renyi entropy. Journal of Multivariate Analysis, 101, 1981-1994 (A).
- Evans D., Gillard J.W. (2011) Computationally efficient estimators of Renyi information and intrinsic dimension. In preparation.