SwePub
Sök i LIBRIS databas

  Utökad sökning

onr:"swepub:oai:DiVA.org:kth-333420"
 

Sökning: onr:"swepub:oai:DiVA.org:kth-333420" > Accelerating Non-Ne...

Accelerating Non-Negative Matrix Factorization on Embedded FPGA with Hybrid Logarithmic Dot-Product Approximation

Chen, Yizhi, 1995- (författare)
KTH,Elektronik och inbyggda system
Nevarez, Yarib (författare)
University of Bremen, Institute of Electrodynamics and Microelectronics (ITEM.ids), Bremen, Germany
Lu, Zhonghai (författare)
KTH,Elektronik och inbyggda system
visa fler...
Garcia-Ortiz, Alberto (författare)
KTH,Skolan för elektroteknik och datavetenskap (EECS)
visa färre...
 (creator_code:org_t)
Institute of Electrical and Electronics Engineers (IEEE), 2022
2022
Engelska.
Ingår i: Proceedings. - : Institute of Electrical and Electronics Engineers (IEEE). ; , s. 239-246
  • Konferensbidrag (refereegranskat)
Abstract Ämnesord
Stäng  
  • Non-negative matrix factorization (NMF) is an ef-fective method for dimensionality reduction and sparse decom-position. This method has been of great interest to the scien-tific community in applications including signal processing, data mining, compression, and pattern recognition. However, NMF implies elevated computational costs in terms of performance and energy consumption, which is inadequate for embedded applications. To overcome this limitation, we implement the vector dot-product with hybrid logarithmic approximation as a hardware optimization approach. This technique accelerates floating-point computation, reduces energy consumption, and preserves accuracy. To demonstrate our approach, we employ a design exploration flow using high-level synthesis on an embedded FPGA. Compared with software solutions on ARM CPU, this hardware implementation accelerates the overall computation to decompose matrix by 5.597 × and reduces energy consumption by 69.323×. Log approximation NMF combined with KNN(k-nearest neighbors) has only 2.38% decreasing accuracy compared with the result of KNN processing the matrix after floating-point NMF on MNIST. Further on, compared with a dedicated floating-point accelerator, the logarithmic approximation approach achieves 3.718× acceleration and 8.345× energy reduction. Compared with the fixed-point approach, our approach has an accuracy degradation of 1.93% on MNIST and an accuracy amelioration of 28.2% on the FASHION MNIST data set without pre-knowledge of the data range. Thus, our approach has better compatibility with the input data range.

Ämnesord

TEKNIK OCH TEKNOLOGIER  -- Elektroteknik och elektronik -- Datorsystem (hsv//swe)
ENGINEERING AND TECHNOLOGY  -- Electrical Engineering, Electronic Engineering, Information Engineering -- Computer Systems (hsv//eng)

Nyckelord

approximate computing
embedded systems
FPGA accelerator
hard-ware/software co-design
machine learning
non-negative matrix factorization (NMF)

Publikations- och innehållstyp

ref (ämneskategori)
kon (ämneskategori)

Till lärosätets databas

Hitta mer i SwePub

Av författaren/redakt...
Chen, Yizhi, 199 ...
Nevarez, Yarib
Lu, Zhonghai
Garcia-Ortiz, Al ...
Om ämnet
TEKNIK OCH TEKNOLOGIER
TEKNIK OCH TEKNO ...
och Elektroteknik oc ...
och Datorsystem
Artiklar i publikationen
Av lärosätet
Kungliga Tekniska Högskolan

Sök utanför SwePub

Kungliga biblioteket hanterar dina personuppgifter i enlighet med EU:s dataskyddsförordning (2018), GDPR. Läs mer om hur det funkar här.
Så här hanterar KB dina uppgifter vid användning av denna tjänst.

 
pil uppåt Stäng

Kopiera och spara länken för att återkomma till aktuell vy