Title :
Generalizing the PAC model: sample size bounds from metric dimension-based uniform convergence results
Author_Institution :
California Univ., Santa Cruz, CA, USA
fDate :
30 Oct-1 Nov 1989
Abstract :
The probably approximately correct (PAC) model of learning from examples is generalized. The problem of learning functions from a set X into a set Y is considered, assuming only that the examples are generated by independent draws according to an unknown probability measure on X×Y. The learner´s goal is to find a function in a given hypothesis space of functions from X into Y that on average give Y values that are close to those observed in random examples. The discrepancy is measured by a bounded real-valued loss function. The average loss is called the error of the hypothesis. A theorem on the uniform convergence of empirical error estimates to true error rates is given for certain hypothesis spaces, and it is shown how this implies learnability. A generalized notion of VC dimension that applies to classes of real-valued functions and a notion of capacity for classes of functions that map into a bounded metric space are given. These measures are used to bound the rate of convergence of empirical error estimates to true error rates, giving bounds on the sample size needed for learning using hypotheses in these classes. As an application, a distribution-independent uniform convergence result for certain classes of functions computed by feedforward neural nets is obtained. Distribution-specific uniform convergence results for classes of functions that are uniformly continuous on average are also obtained
Keywords :
computational complexity; convergence of numerical methods; learning systems; probability; PAC model; VC dimension; average loss; bounded metric space; bounded real-valued loss function; distribution specific uniform convergence; distribution-independent uniform convergence result; empirical error estimates; feedforward neural nets; hypothesis error; independent draws; learnability; learning from examples; learning functions; metric dimension-based uniform convergence results; probably approximately correct; real-valued functions; set; true error rates; unknown probability; Computational complexity; Convergence; Distributed computing; Error analysis; Extraterrestrial measurements; Fractals; Loss measurement; Machine learning; Size measurement; Virtual colonoscopy;
Conference_Titel :
Foundations of Computer Science, 1989., 30th Annual Symposium on
Conference_Location :
Research Triangle Park, NC
Print_ISBN :
0-8186-1982-1
DOI :
10.1109/SFCS.1989.63453