SwePub
Sök i LIBRIS databas

  Extended search

onr:"swepub:oai:DiVA.org:uu-507559"
 

Search: onr:"swepub:oai:DiVA.org:uu-507559" > End-to-end Multiple...

  • 1 of 1
  • Previous record
  • Next record
  •    To hitlist

End-to-end Multiple Instance Learning with Gradient Accumulation

Andersson, Axel (author)
Uppsala universitet,Bildanalys och människa-datorinteraktion,Avdelningen Vi3
Koriakina, Nadezhda, 1991- (author)
Uppsala universitet,Bildanalys och människa-datorinteraktion,Avdelningen Vi3
Sladoje, Nataša (author)
Uppsala universitet,Bildanalys och människa-datorinteraktion,Avdelningen Vi3
show more...
Lindblad, Joakim (author)
Uppsala universitet,Bildanalys och människa-datorinteraktion,Avdelningen Vi3
show less...
 (creator_code:org_t)
Institute of Electrical and Electronics Engineers (IEEE), 2022
2022
English.
In: 2022 IEEE International Conference on Big Data (Big Data). - : Institute of Electrical and Electronics Engineers (IEEE). - 9781665480451 - 9781665480468 ; , s. 2742-2746
  • Conference paper (peer-reviewed)
Abstract Subject headings
Close  
  • Being able to learn on weakly labeled data and provide interpretability are two of the main reasons why attention-based deep multiple instance learning (ABMIL) methods have become particularly popular for classification of histopathological images. Such image data usually come in the form of gigapixel-sized whole-slide-images (WSI) that are cropped into smaller patches (instances). However, the sheer volume of the data poses a practical big data challenge: All the instances from one WSI cannot fit the GPU memory of conventional deep-learning models. Existing solutions compromise training by relying on pre-trained models, strategic selection of instances, sub-sampling, or self-supervised pre-training. We propose a training strategy based on gradient accumulation that enables direct end-to-end training of ABMIL models without being limited by GPU memory. We conduct experiments on both QMNIST and Imagenette to investigate the performance and training time and compare with the conventional memory-expensive baseline as well as a recent sampled-based approach. This memory-efficient approach, although slower, reaches performance indistinguishable from the memory-expensive baseline.

Subject headings

TEKNIK OCH TEKNOLOGIER  -- Elektroteknik och elektronik -- Annan elektroteknik och elektronik (hsv//swe)
ENGINEERING AND TECHNOLOGY  -- Electrical Engineering, Electronic Engineering, Information Engineering -- Other Electrical Engineering, Electronic Engineering, Information Engineering (hsv//eng)

Keyword

Multiple Instance Learning
deep learning
memory management
big data
interpretability
Computerized Image Processing
Datoriserad bildbehandling

Publication and Content Type

ref (subject category)
kon (subject category)

Find in a library

To the university's database

  • 1 of 1
  • Previous record
  • Next record
  •    To hitlist

Find more in SwePub

By the author/editor
Andersson, Axel
Koriakina, Nadez ...
Sladoje, Nataša
Lindblad, Joakim
About the subject
ENGINEERING AND TECHNOLOGY
ENGINEERING AND ...
and Electrical Engin ...
and Other Electrical ...
Articles in the publication
2022 IEEE Intern ...
By the university
Uppsala University

Search outside SwePub

Kungliga biblioteket hanterar dina personuppgifter i enlighet med EU:s dataskyddsförordning (2018), GDPR. Läs mer om hur det funkar här.
Så här hanterar KB dina uppgifter vid användning av denna tjänst.

 
pil uppåt Close

Copy and save the link in order to return to this view