DocumentCode :
2453900
Title :
Parallel Training of a Back-Propagation Neural Network Using CUDA
Author :
Sierra-Canto, Xavier ; Madera-Ramírez, Francisco ; Uc-Cetina, Víctor
Author_Institution :
Div. Ind., Univ. Tecnolgica Metropolitana, Merida, Mexico
fYear :
2010
fDate :
12-14 Dec. 2010
Firstpage :
307
Lastpage :
312
Abstract :
The Artificial Neural Networks (ANN) training represents a time-consuming process in machine learning systems. In this work we provide an implementation of the back-propagation algorithm on CUDA, a parallel computing architecture developed by NVIDIA. Using CUBLAS, a CUDA implementation of the Basic Linear Algebra Subprograms library (BLAS), the process is simplified, however, the use of kernels was necessary since CUBLAS does not have all the required operations. The implementation was tested with two standard benchmark data sets and the results show that the parallel training algorithm runs 63 times faster than its sequential version.
Keywords :
backpropagation; linear algebra; neural nets; operating system kernels; parallel architectures; CUBLAS; CUD A; NVIDIA; artificial neural network parallel training; backpropagation algorithm; linear algebra subprograms library; machine learning systems; parallel computing architecture; Artificial neural networks; Distance measurement; Graphics processing unit; Instruction sets; Kernel; Neurons; Training; CUDA; back-propagation; neural networks;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Machine Learning and Applications (ICMLA), 2010 Ninth International Conference on
Conference_Location :
Washington, DC
Print_ISBN :
978-1-4244-9211-4
Type :
conf
DOI :
10.1109/ICMLA.2010.52
Filename :
5708849
Link To Document :
بازگشت