Title :
Resource Efficient Arithmetic Effects on RBM Neural Network Solution Quality Using MNIST
Author :
Savich, Antony W. ; Moussa, Medhat
Author_Institution :
Sch. of Eng., Univ. of Guelph, Guelph, ON, Canada
fDate :
Nov. 30 2011-Dec. 2 2011
Abstract :
This paper presents a case study on the impact of using reduced precision arithmetic on learning in Restricted Boltzmann Machine (RBM) deep belief networks. FPGAs provide a hardware accelerator framework to speed up many algorithms, including the learning and recognition tasks of ever growing neural network topologies and problem complexities. Current FPGAs include DSP blocks - hard blocks that allow designers to roll in hardware otherwise built using significant quantity of reconfigurable logic (slices) and increase clock performance of arithmetic operations. Accelerators on FPGAs can take advantage of, in some products, thousands DSP blocks on a single chip to scale up the parallelism of designs. Conversely, IEEE floating point representation cannot be fully implemented in single DSP slices and requires a significant amount of general logic thus reducing the amount of resources available to breadth of parallelism in an accelerator design. Reduced precision fixed point format arithmetic can fit within a single DSP slice without external logic. It has been used successfully for training MLP-BP neural networks on small problems. The merit of reduced precision computation in RBM networks for sizable problems has not been evaluated. In this work, a three layer RBM network linked to one classification layer (1.6M weights) is used to learn the classic MNIST dataset over a set of common limited precisions used in FPGA designs. Issues of parameter saturation and a method to overcome inherent training difficulties is discussed. The results demonstrate that RBM can be trained successfully using resource-efficient fixed point formats commonly found in current FPGA devices.
Keywords :
Boltzmann machines; backpropagation; belief networks; field programmable gate arrays; floating point arithmetic; learning (artificial intelligence); multilayer perceptrons; DSP blocks; DSP slices; FPGA designs; IEEE floating point representation; MLP-BP neural network training; MNIST; RBM neural network solution quality; hardware accelerator framework; reconfigurable logic; reduced precision fixed point format arithmetic; resource efficient arithmetic effects; restricted Boltzmann machine deep belief networks; Clocks; Convergence; Digital signal processing; Field programmable gate arrays; Hardware; Neurons; Training; Arithmetic representation; DSP block; FPGA; Neural network; RBM;
Conference_Titel :
Reconfigurable Computing and FPGAs (ReConFig), 2011 International Conference on
Conference_Location :
Cancun
Print_ISBN :
978-1-4577-1734-5
DOI :
10.1109/ReConFig.2011.79