SwePub
Sök i LIBRIS databas

  Utökad sökning

id:"swepub:oai:DiVA.org:kth-283870"
 

Sökning: id:"swepub:oai:DiVA.org:kth-283870" > Optimizing BCPNN Le...

Optimizing BCPNN Learning Rule for Memory Access

Yang, Yu (författare)
KTH,Elektronik och inbyggda system
Stathis, Dimitrios (författare)
KTH,Elektronik och inbyggda system
Jordao, Rodolfo (författare)
KTH,Elektronik och inbyggda system
visa fler...
Hemani, Ahmed, 1961- (författare)
KTH,Elektronik och inbyggda system
Lansner, Anders, Professor (författare)
Stockholms universitet,KTH,Beräkningsvetenskap och beräkningsteknik (CST),Department of Mathematics, Stockholm University, Stockholm, Sweden,Matematiska institutionen,KTH Royal Institute of Technology, Sweden
visa färre...
 (creator_code:org_t)
2020-08-31
2020
Engelska.
Ingår i: Frontiers in Neuroscience. - : Frontiers Media SA. - 1662-4548 .- 1662-453X. ; 14
  • Tidskriftsartikel (refereegranskat)
Abstract Ämnesord
Stäng  
  • Simulation of large scale biologically plausible spiking neural networks, e.g., Bayesian Confidence Propagation Neural Network (BCPNN), usually requires high-performance supercomputers with dedicated accelerators, such as GPUs, FPGAs, or even Application-Specific Integrated Circuits (ASICs). Almost all of these computers are based on the von Neumann architecture that separates storage and computation. In all these solutions, memory access is the dominant cost even for highly customized computation and memory architecture, such as ASICs. In this paper, we propose an optimization technique that can make the BCPNN simulation memory access friendly by avoiding a dual-access pattern. The BCPNN synaptic traces and weights are organized as matrices accessed both row-wise and column-wise. Accessing data stored in DRAM with a dual-access pattern is extremely expensive. A post-synaptic history buffer and an approximation function thus are introduced to eliminate the troublesome column update. The error analysis combining theoretical analysis and experiments suggests that the probability of introducing intolerable errors by such optimization can be bounded to a very small number, which makes it almost negligible. Derivation and validation of such a bound is the core contribution of this paper. Experiments on a GPU platform shows that compared to the previously reported baseline simulation strategy, the proposed optimization technique reduces the storage requirement by 33%, the global memory access demand by more than 27% and DRAM access rate by more than 5%; the latency of updating synaptic traces decreases by roughly 50%. Compared with the other similar optimization technique reported in the literature, our method clearly shows considerably better results. Although the BCPNN is used as the targeted neural network model, the proposed optimization method can be applied to other artificial neural network models based on a Hebbian learning rule.

Ämnesord

NATURVETENSKAP  -- Data- och informationsvetenskap -- Datavetenskap (hsv//swe)
NATURAL SCIENCES  -- Computer and Information Sciences -- Computer Sciences (hsv//eng)
NATURVETENSKAP  -- Data- och informationsvetenskap (hsv//swe)
NATURAL SCIENCES  -- Computer and Information Sciences (hsv//eng)

Nyckelord

Bayesian Confidence Propagation Neural Network (BCPNN)
neuromorphic computing
Hebbian learning
spiking neural networks
memory optimization
DRAM
cache
digital neuromorphic hardware

Publikations- och innehållstyp

ref (ämneskategori)
art (ämneskategori)

Hitta via bibliotek

Till lärosätets databas

Kungliga biblioteket hanterar dina personuppgifter i enlighet med EU:s dataskyddsförordning (2018), GDPR. Läs mer om hur det funkar här.
Så här hanterar KB dina uppgifter vid användning av denna tjänst.

 
pil uppåt Stäng

Kopiera och spara länken för att återkomma till aktuell vy