The growth function, also called the shatter coefficient or the shattering number, measures the richness of a set family or class of function. It is especially used in the context of statistical learning theory, where it is used to study properties of statistical learning methods. The term 'growth function' was coined by Vapnik and Chervonenkis in their 1968 paper, where they also proved many of its properties.[1] It is a basic concept in machine learning.[2] [3]

Definitions

Set-family definition

Let be a set family (a set of sets) and a set. Their intersection is defined as the following set-family:

The intersection-size (also called the index) of with respect to is . If a set has elements then the index is at most . If the index is exactly 2m then the set is said to be shattered by , because contains all the subsets of , i.e.:

The growth function measures the size of as a function of . Formally:

Hypothesis-class definition

Equivalently, let be a hypothesis-class (a set of binary functions) and a set with elements. The restriction of to is the set of binary functions on that can be derived from :[3]:45

The growth function measures the size of as a function of :[3]:49

Examples

1. The domain is the real line . The set-family contains all the half-lines (rays) from a given number to positive infinity, i.e., all sets of the form for some . For any set of real numbers, the intersection contains sets: the empty set, the set containing the largest element of , the set containing the two largest elements of , and so on. Therefore: .[1]:Ex.1 The same is true whether contains open half-lines, closed half-lines, or both.

2. The domain is the segment . The set-family contains all the open sets. For any finite set of real numbers, the intersection contains all possible subsets of . There are such subsets, so . [1]:Ex.2

3. The domain is the Euclidean space . The set-family contains all the half-spaces of the form: , where is a fixed vector. Then , where Comp is the number of components in a partitioning of an n-dimensional space by m hyperplanes.[1]:Ex.3

4. The domain is the real line . The set-family contains all the real intervals, i.e., all sets of the form for some . For any set of real numbers, the intersection contains all runs of between 0 and consecutive elements of . The number of such runs is , so .

Polynomial or exponential

The main property that makes the growth function interesting is that it can be either polynomial or exponential - nothing in-between.

The following is a property of the intersection-size:[1]:Lem.1

  • If, for some set of size , and for some number , -
  • then, there exists a subset of size such that .

This implies the following property of the Growth function.[1]:Th.1 For every family there are two cases:

  • The exponential case: identically.
  • The polynomial case: is majorized by , where is the smallest integer for which .

Other properties

Trivial upper bound

For any finite :

since for every , the number of elements in is at most . Therefore, the growth function is mainly interesting when is infinite.

Exponential upper bound

For any nonempty :

I.e, the growth function has an exponential upper-bound.

We say that a set-family shatters a set if their intersection contains all possible subsets of , i.e. . If shatters of size , then , which is the upper bound.

Cartesian intersection

Define the Cartesian intersection of two set-families as:

.

Then:[2]:57

Union

For every two set-families:[2]:58

VC dimension

The VC dimension of is defined according to these two cases:

  • In the polynomial case, = the largest integer for which .
  • In the exponential case .

So if-and-only-if .

The growth function can be regarded as a refinement of the concept of VC dimension. The VC dimension only tells us whether is equal to or smaller than , while the growth function tells us exactly how changes as a function of .

Another connection between the growth function and the VC dimension is given by the Sauer–Shelah lemma:[3]:49

If , then:
for all :

In particular,

for all :
so when the VC dimension is finite, the growth function grows polynomially with .

This upper bound is tight, i.e., for all there exists with VC dimension such that:[2]:56

Entropy

While the growth-function is related to the maximum intersection-size, the entropy is related to the average intersection size:[1]:272–273

The intersection-size has the following property. For every set-family :

Hence:

Moreover, the sequence converges to a constant when .

Moreover, the random-variable is concentrated near .

Applications in probability theory

Let be a set on which a probability measure is defined. Let be family of subsets of (= a family of events).

Suppose we choose a set that contains elements of , where each element is chosen at random according to the probability measure , independently of the others (i.e., with replacements). For each event , we compare the following two quantities:

  • Its relative frequency in , i.e., ;
  • Its probability .

We are interested in the difference, . This difference satisfies the following upper bound:

which is equivalent to:[1]:Th.2

In words: the probability that for all events in , the relative-frequency is near the probability, is lower-bounded by an expression that depends on the growth-function of .

A corollary of this is that, if the growth function is polynomial in (i.e., there exists some such that ), then the above probability approaches 1 as . I.e, the family enjoys uniform convergence in probability.

References

  1. 1 2 3 4 5 6 7 8 Vapnik, V. N.; Chervonenkis, A. Ya. (1971). "On the Uniform Convergence of Relative Frequencies of Events to Their Probabilities". Theory of Probability & Its Applications. 16 (2): 264. doi:10.1137/1116025. This is an English translation, by B. Seckler, of the Russian paper: "On the Uniform Convergence of Relative Frequencies of Events to Their Probabilities". Dokl. Akad. Nauk. 181 (4): 781. 1968. The translation was reproduced as: Vapnik, V. N.; Chervonenkis, A. Ya. (2015). "On the Uniform Convergence of Relative Frequencies of Events to Their Probabilities". Measures of Complexity. p. 11. doi:10.1007/978-3-319-21852-6_3. ISBN 978-3-319-21851-9.
  2. 1 2 3 4 Mohri, Mehryar; Rostamizadeh, Afshin; Talwalkar, Ameet (2012). Foundations of Machine Learning. US, Massachusetts: MIT Press. ISBN 9780262018258., especially Section 3.2
  3. 1 2 3 4 Shalev-Shwartz, Shai; Ben-David, Shai (2014). Understanding Machine Learning – from Theory to Algorithms. Cambridge University Press. ISBN 9781107057135.
This article is issued from Wikipedia. The text is licensed under Creative Commons - Attribution - Sharealike. Additional terms may apply for the media files.