Dudley's entropy integral is a mathematical concept in the field of probability theory that describes a relationship involving the entropy of certain metric spaces and the concentration of measure phenomenon. It is named after the mathematician R. M. Dudley, who introduced the integral as part of his work on the uniform central limit theorem.

Definition

The Dudley's entropy integral is defined for a metric space equipped with a probability measure . Given a set and an -covering, the entropy of is the logarithm of the minimum number of balls of radius required to cover . Dudley's entropy integral is then given by the formula:

where is the covering number, i.e. the minimum number of balls of radius with respect to the metric that cover the space .[1]

Mathematical background

Dudley's entropy integral arises in the context of empirical processes and Gaussian processes, where it is used to bound the supremum of a stochastic process. Its significance lies in providing a metric entropy measure to assess the complexity of a space with respect to a given probability distribution. More specifically, the expected supremum of a sub-gaussian process is bounded up to finite constants by the entropy integral. Additionally, function classes with a finite entropy integral satisfy a uniform central limit theorem.[2][1]

See also

References

  1. 1 2 Vershynin R. High-Dimensional Probability: An Introduction with Applications in Data Science. Cambridge University Press; 2018.
  2. Vaart AW van der. Asymptotic Statistics. Cambridge University Press; 1998.
This article is issued from Wikipedia. The text is licensed under Creative Commons - Attribution - Sharealike. Additional terms may apply for the media files.