Title :
The Key Theorem of Learning Theory on Set-Valued Probability Space
Author :
Chen, Ji-qiang ; Ha, Ming-Hu ; Zheng, Li-fang
Author_Institution :
Hebei Univ., Baoding
Abstract :
Statistical Learning Theory based on random samples on probability space is considered as the best theory about small samples statistics learning at present and has become a new hot field in machine learning after neural networks. However, the theory can not handle the small samples statistical learning problems on set-valued probability space which widely exists in real world. This paper discussed statistical learning theory on a special kind of set-valued probability space. Firstly, we shall give the definition of random vectors and the definition of the distributed function and the expectation of random vectors, and then we will give the definition of the expected risk functional, the empirical risk functional and the definition of the consistency of the principle (method) of empirical risk minimization (ERM) on set-valued probability space. Finally, we will give and prove the key theorem of learning theory on set-valued probability space, which has laid the theoretical foundation for us to establish the statistical learning theory on probability space.
Keywords :
learning (artificial intelligence); minimisation; probability; random processes; statistical analysis; Hausdorff metric; empirical risk minimization; random vector; set-valued probability space; statistical learning theory; Convergence; Cybernetics; Educational institutions; Machine learning; Neural networks; Probability; Random variables; Risk management; Statistical learning; Support vector machines; Hausdorff metric; Set-valued probability; The key theorem; The principle of empirical risk minimization;
Conference_Titel :
Machine Learning and Cybernetics, 2007 International Conference on
Conference_Location :
Hong Kong
Print_ISBN :
978-1-4244-0973-0
Electronic_ISBN :
978-1-4244-0973-0
DOI :
10.1109/ICMLC.2007.4370620