SwePub
Sök i LIBRIS databas

  Extended search

onr:"swepub:oai:research.chalmers.se:7161d705-057c-44bf-af73-c84f499647ac"
 

Search: onr:"swepub:oai:research.chalmers.se:7161d705-057c-44bf-af73-c84f499647ac" > Activation sparsity...

  • 1 of 1
  • Previous record
  • Next record
  •    To hitlist
  • Haberer, JanekChristian-Albrechts-Universität zu Kiel,University of Kiel (author)

Activation sparsity and dynamic pruning for split computing in edge AI

  • Article/chapterEnglish2022

Publisher, publication year, extent ...

  • 2022-12-06
  • New York, NY, USA :ACM,2022

Numbers

  • LIBRIS-ID:oai:research.chalmers.se:7161d705-057c-44bf-af73-c84f499647ac
  • https://doi.org/10.1145/3565010.3569066DOI
  • https://research.chalmers.se/publication/533970URI

Supplementary language notes

  • Language:English
  • Summary in:English

Part of subdatabase

Classification

  • Subject category:kon swepub-publicationtype
  • Subject category:ref swepub-contenttype

Notes

  • Deep neural networks are getting larger and, therefore, harder to deploy on constrained IoT devices. Split computing provides a solution by splitting a network and placing the first few layers on the IoT device. The output of these layers is transmitted to the cloud where inference continues. Earlier works indicate a degree of high sparsity in intermediate activation outputs, this paper analyzes and exploits activation sparsity to reduce the network communication overhead when transmitting intermediate data to the cloud. Specifically, we analyze the intermediate activations of two early layers in ResNet-50 on CIFAR-10 and ImageNet, focusing on sparsity to guide the process of choosing a splitting point. We employ dynamic pruning of activations and feature maps and find that sparsity is very dependent on the size of a layer, and weights do not correlate with activation sparsity in convolutional layers. Additionally, we show that sparse intermediate outputs can be compressed by a factor of 3.3X at an accuracy loss of 1.1% without any fine-tuning. When adding fine-tuning, the compression factor increases up to 14X at a total accuracy loss of 1%.

Subject headings and genre

Added entries (persons, corporate bodies, meetings, titles ...)

  • Landsiedel, Olaf,1979Christian-Albrechts-Universität zu Kiel,University of Kiel,Chalmers tekniska högskola,Chalmers University of Technology(Swepub:cth)olafl (author)
  • Christian-Albrechts-Universität zu KielChalmers tekniska högskola (creator_code:org_t)

Related titles

  • In:DistributedML 2022 - Proceedings of the 3rd International Workshop on Distributed Machine Learning, Part of CoNEXT 2022New York, NY, USA : ACM, s. 30-36

Internet link

To the university's database

  • 1 of 1
  • Previous record
  • Next record
  •    To hitlist

Find more in SwePub

By the author/editor
Haberer, Janek
Landsiedel, Olaf ...
About the subject
NATURAL SCIENCES
NATURAL SCIENCES
and Computer and Inf ...
and Computer Enginee ...
ENGINEERING AND TECHNOLOGY
ENGINEERING AND ...
and Electrical Engin ...
and Telecommunicatio ...
ENGINEERING AND TECHNOLOGY
ENGINEERING AND ...
and Electrical Engin ...
and Communication Sy ...
Articles in the publication
By the university
Chalmers University of Technology

Search outside SwePub

Kungliga biblioteket hanterar dina personuppgifter i enlighet med EU:s dataskyddsförordning (2018), GDPR. Läs mer om hur det funkar här.
Så här hanterar KB dina uppgifter vid användning av denna tjänst.

 
pil uppåt Close

Copy and save the link in order to return to this view