SwePub
Sök i SwePub databas

  Utökad sökning

Träfflista för sökning "WFRF:(Buchberger Andreas 1990) srt2:(2021)"

Sökning: WFRF:(Buchberger Andreas 1990) > (2021)

  • Resultat 1-2 av 2
Sortera/gruppera träfflistan
   
NumreringReferensOmslagsbildHitta
1.
  • Buchberger, Andreas, 1990, et al. (författare)
  • Learned Decimation for Neural Belief Propagation Decoders
  • 2021
  • Ingår i: ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings. - 1520-6149. ; 2021-June, s. 8273-8277
  • Konferensbidrag (refereegranskat)abstract
    • We introduce a two-stage decimation process to improve the performance of neural belief propagation (NBP), recently introduced by Nachmani et al., for short low-density parity-check (LDPC) codes. In the first stage, we build a list by iterating between a conventional NBP decoder and guessing the least reliable bit. The second stage iterates between a conventional NBP decoder and learned decimation, where we use a neural network to decide the decimation value for each bit. For a (128,64) LDPC code, the proposed NBP with decimation outperforms NBP decoding by 0.75dB and performs within 1dB from maximum-likelihood decoding at a block error rate of 10-4.
  •  
2.
  • Buchberger, Andreas, 1990, et al. (författare)
  • Pruning and Quantizing Neural Belief Propagation Decoders
  • 2021
  • Ingår i: IEEE Journal on Selected Areas in Communications. - 0733-8716 .- 1558-0008. ; 39:7, s. 1957-1966
  • Tidskriftsartikel (refereegranskat)abstract
    • We consider near maximum-likelihood (ML) decoding of short linear block codes. In particular, we propose a novel decoding approach based on neural belief propagation (NBP) decoding recently introduced by Nachmani et al. in which we allow a different parity-check matrix in each iteration of the algorithm. The key idea is to consider NBP decoding over an overcomplete parity-check matrix and use the weights of NBP as a measure of the importance of the check nodes (CNs) to decoding. The unimportant CNs are then pruned. In contrast to NBP, which performs decoding on a given fixed parity-check matrix, the proposed pruning-based neural belief propagation (PB-NBP) typically results in a different parity-check matrix in each iteration. For a given complexity in terms of CN evaluations, we show that PB-NBP yields significant performance improvements with respect to NBP. We apply the proposed decoder to the decoding of a Reed-Muller code, a short low-density parity-check (LDPC) code, and a polar code. PB-NBP outperforms NBP decoding over an overcomplete parity-check matrix by 0.27–0.31 dB while reducing the number of required CN evaluations by up to 97%. For the LDPC code, PB-NBP outperforms conventional belief propagation with the same number of CN evaluations by 0.52 dB. We further extend the pruning concept to offset min-sum decoding and introduce a pruning-based neural offset min-sum (PB-NOMS) decoder, for which we jointly optimize the offsets and the quantization of the messages and offsets. We demonstrate performance 0.5 dB from ML decoding with 5-bit quantization for the Reed-Muller code.
  •  
Skapa referenser, mejla, bekava och länka
  • Resultat 1-2 av 2
Typ av publikation
konferensbidrag (1)
tidskriftsartikel (1)
Typ av innehåll
refereegranskat (2)
Författare/redaktör
Schmalen, Laurent (2)
Graell i Amat, Alexa ... (2)
Buchberger, Andreas, ... (2)
Häger, Christian, 19 ... (2)
Pfister, Henry D. (2)
Lärosäte
Chalmers tekniska högskola (2)
Språk
Engelska (2)
Forskningsämne (UKÄ/SCB)
Teknik (2)
Naturvetenskap (1)
År

Kungliga biblioteket hanterar dina personuppgifter i enlighet med EU:s dataskyddsförordning (2018), GDPR. Läs mer om hur det funkar här.
Så här hanterar KB dina uppgifter vid användning av denna tjänst.

 
pil uppåt Stäng

Kopiera och spara länken för att återkomma till aktuell vy