SwePub
Sök i LIBRIS databas

  Extended search

onr:"swepub:oai:DiVA.org:kth-333420"
 

Search: onr:"swepub:oai:DiVA.org:kth-333420" > Accelerating Non-Ne...

  • 1 of 1
  • Previous record
  • Next record
  •    To hitlist

Accelerating Non-Negative Matrix Factorization on Embedded FPGA with Hybrid Logarithmic Dot-Product Approximation

Chen, Yizhi, 1995- (author)
KTH,Elektronik och inbyggda system
Nevarez, Yarib (author)
University of Bremen, Institute of Electrodynamics and Microelectronics (ITEM.ids), Bremen, Germany
Lu, Zhonghai (author)
KTH,Elektronik och inbyggda system
show more...
Garcia-Ortiz, Alberto (author)
KTH,Skolan för elektroteknik och datavetenskap (EECS)
show less...
 (creator_code:org_t)
Institute of Electrical and Electronics Engineers (IEEE), 2022
2022
English.
In: Proceedings. - : Institute of Electrical and Electronics Engineers (IEEE). ; , s. 239-246
  • Conference paper (peer-reviewed)
Abstract Subject headings
Close  
  • Non-negative matrix factorization (NMF) is an ef-fective method for dimensionality reduction and sparse decom-position. This method has been of great interest to the scien-tific community in applications including signal processing, data mining, compression, and pattern recognition. However, NMF implies elevated computational costs in terms of performance and energy consumption, which is inadequate for embedded applications. To overcome this limitation, we implement the vector dot-product with hybrid logarithmic approximation as a hardware optimization approach. This technique accelerates floating-point computation, reduces energy consumption, and preserves accuracy. To demonstrate our approach, we employ a design exploration flow using high-level synthesis on an embedded FPGA. Compared with software solutions on ARM CPU, this hardware implementation accelerates the overall computation to decompose matrix by 5.597 × and reduces energy consumption by 69.323×. Log approximation NMF combined with KNN(k-nearest neighbors) has only 2.38% decreasing accuracy compared with the result of KNN processing the matrix after floating-point NMF on MNIST. Further on, compared with a dedicated floating-point accelerator, the logarithmic approximation approach achieves 3.718× acceleration and 8.345× energy reduction. Compared with the fixed-point approach, our approach has an accuracy degradation of 1.93% on MNIST and an accuracy amelioration of 28.2% on the FASHION MNIST data set without pre-knowledge of the data range. Thus, our approach has better compatibility with the input data range.

Subject headings

TEKNIK OCH TEKNOLOGIER  -- Elektroteknik och elektronik -- Datorsystem (hsv//swe)
ENGINEERING AND TECHNOLOGY  -- Electrical Engineering, Electronic Engineering, Information Engineering -- Computer Systems (hsv//eng)

Keyword

approximate computing
embedded systems
FPGA accelerator
hard-ware/software co-design
machine learning
non-negative matrix factorization (NMF)

Publication and Content Type

ref (subject category)
kon (subject category)

To the university's database

  • 1 of 1
  • Previous record
  • Next record
  •    To hitlist

Find more in SwePub

By the author/editor
Chen, Yizhi, 199 ...
Nevarez, Yarib
Lu, Zhonghai
Garcia-Ortiz, Al ...
About the subject
ENGINEERING AND TECHNOLOGY
ENGINEERING AND ...
and Electrical Engin ...
and Computer Systems
Articles in the publication
By the university
Royal Institute of Technology

Search outside SwePub

Kungliga biblioteket hanterar dina personuppgifter i enlighet med EU:s dataskyddsförordning (2018), GDPR. Läs mer om hur det funkar här.
Så här hanterar KB dina uppgifter vid användning av denna tjänst.

 
pil uppåt Close

Copy and save the link in order to return to this view