Title :
A Parallel Digital VLSI Architecture for Integrated Support Vector Machine Training and Classification
Author :
Qian Wang ; Peng Li ; Yongtae Kim
Author_Institution :
Dept. of Electr. & Comput. Eng., Texas A&M Univ., College Station, TX, USA
Abstract :
This paper presents a parallel digital VLSI architecture for combined support vector machine (SVM) training and classification. For the first time, cascade SVM, a powerful training algorithm, is leveraged to significantly improve the scalability of hardware-based SVM training and develop an efficient parallel VLSI architecture. The presented architecture achieves excellent scalability by spreading the training workload of a given data set over multiple SVM processing units with minimal communication overhead. Hardware-friendly implementation of the cascade algorithm is employed to achieve low hardware overhead and allow for training over data sets of variable size. In the proposed parallel cascade architecture, a multilayer system bus and multiple distributed memories are used to fully exploit parallelism. In addition, the proposed architecture is rather flexible and can be tailored to realize hybrid use of hardware parallel processing and temporal reuse of processing resources, leading to good tradeoffs between throughput, silicon overhead and power dissipation. Several parallel cascade SVM processors have been designed with a commercial 90-nm CMOS technology, which provide up to a 561× training time speedup and a significant estimated 21 859× energy reduction compared with the software SVM algorithm running on a 45-nm commercial general-purpose CPU.
Keywords :
CMOS digital integrated circuits; VLSI; electronic engineering computing; multiprocessing systems; parallel processing; support vector machines; cascade algorithm; combined SVM classification; combined support vector machine training; commercial 90-nm CMOS technology; hardware parallel processing; hardware-based SVM training; hardware-friendly implementation; multilayer system bus; multiple SVM processing units; multiple distributed memories; parallel cascade SVM processors; parallel cascade architecture; parallel digital VLSI architecture; size 90 nm; training workload; Computer architecture; Hardware; Kernel; Optimization; Support vector machines; Training; Very large scale integration; Digital integrated circuits; multicore processing; parallel architectures; support vector machines; system buses;
Journal_Title :
Very Large Scale Integration (VLSI) Systems, IEEE Transactions on
DOI :
10.1109/TVLSI.2014.2343231