Natarajan dimension

From testwiki
Revision as of 13:26, 19 February 2024 by 192.114.105.254 (talk) (Definition)
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to navigation Jump to search

In the theory of Probably Approximately Correct Machine Learning, the Natarajan dimension characterizes the complexity of learning a set of functions, generalizing from the Vapnik-Chervonenkis dimension for boolean functions to multi-class functions. Originally introduced as the Generalized Dimension by Natarajan,[1] it was subsequently renamed the Natarajan Dimension by Haussler and Long.[2]

Definition

Let H be a set of functions from a set X to a set Y. H shatters a set CX if there exist two functions f0,f1H such that

  • For every xC,f0(x)f1(x).
  • For every BC, there exists a function hH such that

for all xB,h(x)=f0(x) and for all xCB,h(x)=f1(x).

The Natarajan dimension of H is the maximal cardinality of a set shattered by H.

It is easy to see that if |Y|=2, the Natarajan dimension collapses to the Vapnik Chervonenkis dimension.

Shalev-Shwartz and Ben-David [3] present comprehensive material on multi-class learning and the Natarajan dimension, including uniform convergence and learnability.

References