Title :
Reducing the computational requirement of the orthogonal least squares algorithm
Author :
Chng, E.S. ; Chen, S. ; Mulgrew, B.
Author_Institution :
Dept. of Electr. Eng., Edinburgh Univ., UK
Abstract :
The orthogonal, least squares (OLS) algorithm is an efficient implementation of the forward regression procedure for subset model selection. The ability to find good subset parameters with only linear increase in computational complexity makes this method attractive for practical implementations. We examine the computation requirement of the OLS algorithm to reduce a model of K terms to a subset model of R terms when the number of training data available is N. We show that in the case where N≫K, we can reduce the computation requirement by introducing an unitary transformation on the problem
Keywords :
computational complexity; least squares approximations; parameter estimation; prediction theory; statistical analysis; OLS algorithm; computation requirement; computational complexity; forward regression procedure; nonlinear predictors; orthogonal least squares algorithm; subset model selection; subset parameters; training data; unitary transformation; Degradation; Equations; Least squares methods; Linear regression; Predictive models; Reflection; Systems engineering and theory; Testing; Training data; Vectors;
Conference_Titel :
Acoustics, Speech, and Signal Processing, 1994. ICASSP-94., 1994 IEEE International Conference on
Conference_Location :
Adelaide, SA
Print_ISBN :
0-7803-1775-0
DOI :
10.1109/ICASSP.1994.389973