Sökning: WFRF:(Lu Zhonghai) >
Accelerating Non-Ne...
Accelerating Non-Negative Matrix Factorization on Embedded FPGA with Hybrid Logarithmic Dot-Product Approximation
-
- Chen, Yizhi, 1995- (författare)
- KTH,Elektronik och inbyggda system
-
- Nevarez, Yarib (författare)
- University of Bremen, Institute of Electrodynamics and Microelectronics (ITEM.ids), Bremen, Germany
-
- Lu, Zhonghai (författare)
- KTH,Elektronik och inbyggda system
-
visa fler...
-
- Garcia-Ortiz, Alberto (författare)
- KTH,Skolan för elektroteknik och datavetenskap (EECS)
-
visa färre...
-
(creator_code:org_t)
- Institute of Electrical and Electronics Engineers (IEEE), 2022
- 2022
- Engelska.
-
Ingår i: Proceedings. - : Institute of Electrical and Electronics Engineers (IEEE). ; , s. 239-246
- Relaterad länk:
-
https://urn.kb.se/re...
-
visa fler...
-
https://doi.org/10.1...
-
visa färre...
Abstract
Ämnesord
Stäng
- Non-negative matrix factorization (NMF) is an ef-fective method for dimensionality reduction and sparse decom-position. This method has been of great interest to the scien-tific community in applications including signal processing, data mining, compression, and pattern recognition. However, NMF implies elevated computational costs in terms of performance and energy consumption, which is inadequate for embedded applications. To overcome this limitation, we implement the vector dot-product with hybrid logarithmic approximation as a hardware optimization approach. This technique accelerates floating-point computation, reduces energy consumption, and preserves accuracy. To demonstrate our approach, we employ a design exploration flow using high-level synthesis on an embedded FPGA. Compared with software solutions on ARM CPU, this hardware implementation accelerates the overall computation to decompose matrix by 5.597 × and reduces energy consumption by 69.323×. Log approximation NMF combined with KNN(k-nearest neighbors) has only 2.38% decreasing accuracy compared with the result of KNN processing the matrix after floating-point NMF on MNIST. Further on, compared with a dedicated floating-point accelerator, the logarithmic approximation approach achieves 3.718× acceleration and 8.345× energy reduction. Compared with the fixed-point approach, our approach has an accuracy degradation of 1.93% on MNIST and an accuracy amelioration of 28.2% on the FASHION MNIST data set without pre-knowledge of the data range. Thus, our approach has better compatibility with the input data range.
Ämnesord
- TEKNIK OCH TEKNOLOGIER -- Elektroteknik och elektronik -- Datorsystem (hsv//swe)
- ENGINEERING AND TECHNOLOGY -- Electrical Engineering, Electronic Engineering, Information Engineering -- Computer Systems (hsv//eng)
Nyckelord
- approximate computing
- embedded systems
- FPGA accelerator
- hard-ware/software co-design
- machine learning
- non-negative matrix factorization (NMF)
Publikations- och innehållstyp
- ref (ämneskategori)
- kon (ämneskategori)