• DocumentCode
    24152
  • Title

    A Parallel Digital VLSI Architecture for Integrated Support Vector Machine Training and Classification

  • Author

    Qian Wang ; Peng Li ; Yongtae Kim

  • Author_Institution
    Dept. of Electr. & Comput. Eng., Texas A&M Univ., College Station, TX, USA
  • Volume
    23
  • Issue
    8
  • fYear
    2015
  • fDate
    Aug. 2015
  • Firstpage
    1471
  • Lastpage
    1484
  • Abstract
    This paper presents a parallel digital VLSI architecture for combined support vector machine (SVM) training and classification. For the first time, cascade SVM, a powerful training algorithm, is leveraged to significantly improve the scalability of hardware-based SVM training and develop an efficient parallel VLSI architecture. The presented architecture achieves excellent scalability by spreading the training workload of a given data set over multiple SVM processing units with minimal communication overhead. Hardware-friendly implementation of the cascade algorithm is employed to achieve low hardware overhead and allow for training over data sets of variable size. In the proposed parallel cascade architecture, a multilayer system bus and multiple distributed memories are used to fully exploit parallelism. In addition, the proposed architecture is rather flexible and can be tailored to realize hybrid use of hardware parallel processing and temporal reuse of processing resources, leading to good tradeoffs between throughput, silicon overhead and power dissipation. Several parallel cascade SVM processors have been designed with a commercial 90-nm CMOS technology, which provide up to a 561× training time speedup and a significant estimated 21 859× energy reduction compared with the software SVM algorithm running on a 45-nm commercial general-purpose CPU.
  • Keywords
    CMOS digital integrated circuits; VLSI; electronic engineering computing; multiprocessing systems; parallel processing; support vector machines; cascade algorithm; combined SVM classification; combined support vector machine training; commercial 90-nm CMOS technology; hardware parallel processing; hardware-based SVM training; hardware-friendly implementation; multilayer system bus; multiple SVM processing units; multiple distributed memories; parallel cascade SVM processors; parallel cascade architecture; parallel digital VLSI architecture; size 90 nm; training workload; Computer architecture; Hardware; Kernel; Optimization; Support vector machines; Training; Very large scale integration; Digital integrated circuits; multicore processing; parallel architectures; support vector machines; system buses;
  • fLanguage
    English
  • Journal_Title
    Very Large Scale Integration (VLSI) Systems, IEEE Transactions on
  • Publisher
    ieee
  • ISSN
    1063-8210
  • Type

    jour

  • DOI
    10.1109/TVLSI.2014.2343231
  • Filename
    6876212