Title :
Finite precision error analysis of neural network hardware implementations
Author :
Holi, J.L. ; Hwang, Jenq-Neng
Author_Institution :
Adaptive Solutions Inc., Beaverton, OR, USA
fDate :
3/1/1993 12:00:00 AM
Abstract :
Through parallel processing, low precision fixed point hardware can be used to build a very high speed neural network computing engine where the low precision results in a drastic reduction in system cost. The reduced silicon area required to implement a single processing unit is taken advantage of by implementing multiple processing units on a single piece of silicon and operating them in parallel. The important question which arises is how much precision is required to implement neural network algorithms on this low precision hardware. A theoretical analysis of error due to finite precision computation was undertaken to determine the necessary precision for successful forward retrieving and back-propagation learning in a multilayer perceptron. This analysis can easily be further extended to provide a general finite precision analysis technique by which most neural network algorithms under any set of hardware constraints may be evaluated
Keywords :
error analysis; feedforward neural nets; neural chips; back-propagation learning; finite precision computation; forward retrieving; low precision; multilayer perceptron; neural network algorithms; neural network hardware; parallel processing; silicon area; system cost; Algorithm design and analysis; Computer networks; Concurrent computing; Costs; Engines; Error analysis; Neural network hardware; Neural networks; Parallel processing; Silicon;
Journal_Title :
Computers, IEEE Transactions on