SwePub
Sök i SwePub databas

  Utökad sökning

Träfflista för sökning "WFRF:(Daneshtalab Masoud.) "

Sökning: WFRF:(Daneshtalab Masoud.)

  • Resultat 1-10 av 168
Sortera/gruppera träfflistan
   
NumreringReferensOmslagsbildHitta
1.
  • Afsharmazayejani, R., et al. (författare)
  • HoneyWiN : Novel honeycomb-based wireless NoC architecture in many-core era
  • 2018
  • Ingår i: Lecture Notes in Computer Science. - Cham : Springer Verlag. - 0302-9743 .- 1611-3349. ; 10824 LNCS, s. 304-316
  • Tidskriftsartikel (refereegranskat)abstract
    • Although NoC-based systems with many cores are commercially available, their multi-hop nature has become a bottleneck on scaling performance and energy consumption parameters. Alternatively, hybrid wireless NoC provides a postern by exploiting single-hop express links for long-distance communications. Also, there is a common wisdom that grid-like mesh is the most stable topology in conventional designs. That is why almost all of the emerging architectures had been relying on this topology as well. In this paper, first we challenge the efficiency of the grid-like mesh in emerging systems. Then, we propose HoneyWiN, a hybrid reconfigurable wireless NoC architecture that relies on the honeycomb topology. The simulation results show that on average HoneyWiN saves 17% of energy consumption while increases the network throughput by 10% compared to its wireless mesh counterpart. 
  •  
2.
  • Ahmadilivani, M. H., et al. (författare)
  • A Systematic Literature Review on Hardware Reliability Assessment Methods for Deep Neural Networks
  • 2024
  • Ingår i: ACM Computing Surveys. - : ASSOC COMPUTING MACHINERY. - 0360-0300 .- 1557-7341. ; 56:6
  • Tidskriftsartikel (refereegranskat)abstract
    • Artificial Intelligence (AI) and, in particular, Machine Learning (ML), have emerged to be utilized in various applications due to their capability to learn how to solve complex problems. Over the past decade, rapid advances in ML have presented Deep Neural Networks (DNNs) consisting of a large number of neurons and layers. DNN Hardware Accelerators (DHAs) are leveraged to deploy DNNs in the target applications. Safety-critical applications, where hardware faults/errors would result in catastrophic consequences, also benefit from DHAs. Therefore, the reliability of DNNs is an essential subject of research. In recent years, several studies have been published accordingly to assess the reliability of DNNs. In this regard, various reliability assessment methods have been proposed on a variety of platforms and applications. Hence, there is a need to summarize the state-of-the-art to identify the gaps in the study of the reliability of DNNs. In this work, we conduct a Systematic Literature Review (SLR) on the reliability assessment methods of DNNs to collect relevant research works as much as possible, present a categorization of them, and address the open challenges. Through this SLR, three kinds of methods for reliability assessment of DNNs are identified, including Fault Injection (FI), Analytical, and Hybrid methods. Since the majority of works assess the DNN reliability by FI, we characterize different approaches and platforms of the FI method comprehensively. Moreover, Analytical and Hybrid methods are propounded. Thus, different reliability assessment methods for DNNs have been elaborated on their conducted DNN platforms and reliability evaluation metrics. Finally, we highlight the advantages and disadvantages of the identified methods and address the open challenges in the research area. We have concluded that Analytical and Hybrid methods are light-weight yet sufficiently accurate and have the potential to be extended in future research and to be utilized in establishing novel DNN reliability assessment frameworks.
  •  
3.
  • Ahmadilivani, M. H., et al. (författare)
  • Analysis and Improvement of Resilience for Long Short-Term Memory Neural Networks
  • 2023
  • Ingår i: Proc. IEEE Int. Symp. Defect Fault Toler. VLSI Nanotechnol. Syst., DFT. - : Institute of Electrical and Electronics Engineers Inc.. - 9798350315004
  • Konferensbidrag (refereegranskat)abstract
    • The reliability of Artificial Neural Networks (ANNs) has emerged as a prominent research topic due to their increasing utilization in safety-critical applications. Long Short-Term Memory (LSTM) ANNs have demonstrated significant advantages in healthcare applications, primarily attributed to their robust processing of time-series data and memory-facilitated capabilities. This paper, for the first time, presents a comprehensive and fine-grain analysis of the resilience of LSTM-based ANNs in the context of gait analysis using fault injection into weights. Additionally, we improve their resilience by replacing faulty weights with zero, enabling ANNs to withstand environments that are up to 20 times harsher while experiencing up to 7 times fewer critical faults than an unprotected ANN.
  •  
4.
  • Ahmadilivani, Mohammad Hasan, et al. (författare)
  • DeepVigor : VulnerabIlity Value RanGes and FactORs for DNNs' Reliability Assessment
  • 2023
  • Ingår i: 2023 IEEE EUROPEAN TEST SYMPOSIUM, ETS. - : IEEE. - 9798350336344
  • Konferensbidrag (refereegranskat)abstract
    • Deep Neural Networks (DNNs) and their accelerators are being deployed ever more frequently in safety-critical applications leading to increasing reliability concerns. A traditional and accurate method for assessing DNNs' reliability has been resorting to fault injection, which, however, suffers from prohibitive time complexity. While analytical and hybrid fault injection-/analyticalbased methods have been proposed, they are either inaccurate or specific to particular accelerator architectures. In this work, we propose a novel accurate, fine-grain, metric-oriented, and accelerator-agnostic method called DeepVigor that provides vulnerability value ranges for DNN neurons' outputs. An outcome of DeepVigor is an analytical model representing vulnerable and non-vulnerable ranges for each neuron that can be exploited to develop different techniques for improving DNNs' reliability. Moreover, DeepVigor provides reliability assessment metrics based on vulnerability factors for bits, neurons, and layers using the vulnerability ranges. The proposed method is not only faster than fault injection but also provides extensive and accurate information about the reliability of DNNs, independent from the accelerator. The experimental evaluations in the paper indicate that the proposed vulnerability ranges are 99.9% to 100% accurate even when evaluated on previously unseen test data. Also, it is shown that the obtained vulnerability factors represent the criticality of bits, neurons, and layers proficiently. DeepVigor is implemented in the PyTorch framework and validated on complex DNN benchmarks.
  •  
5.
  • Ahmadilivani, M. H., et al. (författare)
  • Enhancing Fault Resilience of QNNs by Selective Neuron Splitting
  • 2023
  • Ingår i: AICAS 2023 - IEEE International Conference on Artificial Intelligence Circuits and Systems, Proceeding. - : Institute of Electrical and Electronics Engineers Inc.. - 9798350332674
  • Konferensbidrag (refereegranskat)abstract
    • The superior performance of Deep Neural Networks (DNNs) has led to their application in various aspects of human life. Safety-critical applications are no exception and impose rigorous reliability requirements on DNNs. Quantized Neural Networks (QNNs) have emerged to tackle the complexity of DNN accelerators, however, they are more prone to reliability issues.In this paper, a recent analytical resilience assessment method is adapted for QNNs to identify critical neurons based on a Neuron Vulnerability Factor (NVF). Thereafter, a novel method for splitting the critical neurons is proposed that enables the design of a Lightweight Correction Unit (LCU) in the accelerator without redesigning its computational part.The method is validated by experiments on different QNNs and datasets. The results demonstrate that the proposed method for correcting the faults has a twice smaller overhead than a selective Triple Modular Redundancy (TMR) while achieving a similar level of fault resiliency. 
  •  
6.
  • Ahmadilivani, Mohammed. H., et al. (författare)
  • Special Session : Approximation and Fault Resiliency of DNN Accelerators
  • 2023
  • Ingår i: Proceedings of the IEEE VLSI Test Symposium. - : IEEE Computer Society. - 9798350346305
  • Konferensbidrag (refereegranskat)abstract
    • Deep Learning, and in particular, Deep Neural Network (DNN) is nowadays widely used in many scenarios, including safety-critical applications such as autonomous driving. In this context, besides energy efficiency and performance, reliability plays a crucial role since a system failure can jeopardize human life. As with any other device, the reliability of hardware architectures running DNNs has to be evaluated, usually through costly fault injection campaigns. This paper explores approximation and fault resiliency of DNN accelerators. We propose to use approximate (AxC) arithmetic circuits to agilely emulate errors in hardware without performing fault injection on the DNN. To allow fast evaluation of AxC DNN, we developed an efficient GPU-based simulation framework. Further, we propose a fine-grain analysis of fault resiliency by examining fault propagation and masking in networks.
  •  
7.
  • Akbari, N., et al. (författare)
  • A Customized Processing-in-Memory Architecture for Biological Sequence Alignment
  • 2018
  • Ingår i: Proceedings of the International Conference on Application-Specific Systems, Architectures and Processors. - : Institute of Electrical and Electronics Engineers Inc.. - 9781538674796
  • Konferensbidrag (refereegranskat)abstract
    • Sequence alignment is the most widely used operation in bioinformatics. With the exponential growth of the biological sequence databases, searching a database to find the optimal alignment for a query sequence (that can be at the order of hundreds of millions of characters long) would require excessive processing power and memory bandwidth. Sequence alignment algorithms can potentially benefit from the processing power of massive parallel processors due their simple arithmetic operations, coupled with the inherent fine-grained and coarse-grained parallelism that they exhibit. However, the limited memory bandwidth in conventional computing systems prevents exploiting the maximum achievable speedup. In this paper, we propose a processing-in-memory architecture as a viable solution for the excessive memory bandwidth demand of bioinformatics applications. The design is composed of a set of simple and lightweight processing elements, customized to the sequence alignment algorithm, integrated at the logic layer of an emerging 3D DRAM architecture. Experimental results show that the proposed architecture results in up to 2.4x speedup and 41% reduction in power consumption, compared to a processor-side parallel implementation.
  •  
8.
  • Aldinucci, Marco, et al. (författare)
  • Preface
  • 2017
  • Ingår i: The international journal of high performance computing applications. - : Sage Publications. - 1094-3420 .- 1741-2846. ; 31:3, s. 179-180
  • Tidskriftsartikel (refereegranskat)
  •  
9.
  • Amin, Yoosefi, et al. (författare)
  • Efficient On-device Transfer Learning using Activation Memory Reduction
  • 2023
  • Ingår i: Int. Conf. Fog Mob. Edge Comput., FMEC. - : Institute of Electrical and Electronics Engineers Inc.. - 9798350316971 ; , s. 210-215
  • Konferensbidrag (refereegranskat)abstract
    • On-device transfer learning suggests fine-tuning pretrained neural networks on new input data directly on edge devices. The memory limitation of edge devices necessitates using memory-efficient fine-tuning methods. Fine-tuning involves two primary phases: the forward-pass phase and the backwardpass phase. The forward-pass phase generates output activations, and the backward-pass phase computes gradients and updates the parameters accordingly. Although the forward-pass phase demands a temporary memory to store a layer’s input and output activations, the backward-pass phase may require storing the output activations from all layers to compute gradients. This fact introduces the memory cost of the backward-pass phase as the main contributor to the huge training memory demands of deep neural networks (DNNs), which has been the focus of many studies. However, little attention has been made to how the temporary activation memory involved in the forward-pass phase may also act as the memory bottleneck, which is the main focus of this paper. This paper aims to mitigate this memory bottleneck by pruning unimportant channels from layers that require significant temporary activation memory. Experimental results demonstrate how the proposed method effectively reduces peak activation memory and total memory costs of MobileNetV2 by 65% and 59%, respectively, at the cost of 3% accuracy drop.
  •  
10.
  • Anwar, Hassan, et al. (författare)
  • Exploring Spiking Neural Network on Coarse-Grain Reconfigurable Architectures
  • 2014
  • Ingår i: ACM International Conference Proceeding Series. - New York, NY, USA : ACM. - 9781450328227 ; , s. 64-67
  • Konferensbidrag (refereegranskat)abstract
    • Today, reconfigurable architectures are becoming increas- ingly popular as the candidate platforms for neural net- works. Existing works, that map neural networks on re- configurable architectures, only address either FPGAs or Networks-on-chip, without any reference to the Coarse-Grain Reconfigurable Architectures (CGRAs). In this paper we investigate the overheads imposed by implementing spiking neural networks on a Coarse Grained Reconfigurable Ar- chitecture (CGRAs). Experimental results (using point to point connectivity) reveal that up to 1000 neurons can be connected, with an average response time of 4.4 msec.
  •  
Skapa referenser, mejla, bekava och länka
  • Resultat 1-10 av 168
Typ av publikation
konferensbidrag (107)
tidskriftsartikel (47)
doktorsavhandling (4)
bokkapitel (4)
licentiatavhandling (3)
samlingsverk (redaktörskap) (1)
visa fler...
rapport (1)
annan publikation (1)
visa färre...
Typ av innehåll
refereegranskat (146)
övrigt vetenskapligt/konstnärligt (22)
Författare/redaktör
Daneshtalab, Masoud (164)
Sjödin, Mikael, 1971 ... (30)
Mubeen, Saad (20)
Plosila, Juha (20)
Tenhunen, Hannu (19)
Ashjaei, Seyed Moham ... (18)
visa fler...
Ebrahimi, Masoumeh (18)
Loni, Mohammad (14)
Hemani, Ahmed (10)
Rezaei, A (10)
Riazati, Mohammad (10)
Modarressi, M. (9)
Loni, Mohammad, PhD ... (9)
Majd, Amin (9)
Majd, A. (9)
Lisper, Björn (7)
Sinaei, Sima (7)
Zhao, D (7)
Modarressi, Mehdi (7)
Houtan, Bahar, 1989- (7)
Raik, J. (6)
Liljeberg, Pasi (6)
Palesi, Maurizio (6)
Plosila, J. (6)
Wang, X. (5)
Troubitsyna, Elena (5)
Mak, Terrence (5)
Zoljodi, Ali (5)
Fotouhi, Hossein (4)
Ahmadilivani, M. H. (4)
Taheri, M. (4)
Jenihhin, M. (4)
Mousavi, Hamid (4)
Jafri, Syed Mohammad ... (4)
Dytckov, Sergei (4)
Paul, Kolin (4)
Jafri, Syed M. A. H. (4)
Houtan, Bahar (4)
Sjödin, Mikael, Prof ... (4)
Forsberg, Håkan (3)
Raik, Jaan (3)
Jenihhin, Maksim (3)
Berisa, Aldin (3)
Nolin, Mikael, 1971- (3)
Bagherzadeh, Nader (3)
Palesi, M. (3)
Sarbazi-Azad, Hamid (3)
Yang, Mei (3)
Daneshtalab, Masoud, ... (3)
Wang, Xiaohang (3)
visa färre...
Lärosäte
Mälardalens universitet (104)
Kungliga Tekniska Högskolan (79)
RISE (7)
Linköpings universitet (1)
Språk
Engelska (168)
Forskningsämne (UKÄ/SCB)
Teknik (127)
Naturvetenskap (56)

År

Kungliga biblioteket hanterar dina personuppgifter i enlighet med EU:s dataskyddsförordning (2018), GDPR. Läs mer om hur det funkar här.
Så här hanterar KB dina uppgifter vid användning av denna tjänst.

 
pil uppåt Stäng

Kopiera och spara länken för att återkomma till aktuell vy