Title :
Some contributions to fixed-distribution learning theory
Author :
Vidyasagar, M. ; Kulkarni, Sanjeev R.
Author_Institution :
Centre for Artificial Intelligence & Robotics, Bangalore, India
fDate :
2/1/2000 12:00:00 AM
Abstract :
We consider some problems in learning with respect to a fixed distribution. We introduce two new notions of learnability; these are probably uniformly approximately correct (PUAC) learnability which is a stronger requirement than the widely studied PAC learnability, and minimal empirical risk (MER) learnability, which is a stronger requirement than the previously defined notions of “solid” or “potential” learnability. It is shown that, although the motivations for defining these two notions of learnability are entirely different, these two notions are in fact equivalent to each other and, in turn, equivalent to a property introduced here, referred to as the shrinking width property. It is further shown that if the function class to be learned has the property that empirical means converge uniformly to their true values, then all of these learnability properties hold. In the course of proving conditions for these forms of learnability, we also obtain a new estimate for the VC-dimension of a collection of sets obtained by performing Boolean operations on a given collection; this result is of independent interest. We consider both the case in which there is an underlying target function, as well as the case of “model-free” (or agnostic) learning. Finally, we consider the issue of representation of a collection of sets by its subcollection of equivalence classes. It is shown by example that, by suitably choosing representatives of each equivalence class, it is possible to affect the property of uniform convergence of empirical probabilities
Keywords :
convergence; equivalence classes; learning (artificial intelligence); probability; Boolean operations; VC-dimension; agnostic learning; empirical probabilities; fixed-distribution learning theory; function class; minimal empirical risk learnability; model-free learning; potential learnability; probably uniformly approximately correct learnability; shrinking width property; solid learnability; uniform convergence; Artificial intelligence; Computer science; Convergence; Intelligent robots; Machine learning; Mathematical model;
Journal_Title :
Automatic Control, IEEE Transactions on