SwePub
Sök i SwePub databas

  Utökad sökning

Träfflista för sökning "WFRF:(Matuszewski Damian J.) srt2:(2019)"

Sökning: WFRF:(Matuszewski Damian J.) > (2019)

  • Resultat 1-2 av 2
Sortera/gruppera träfflistan
   
NumreringReferensOmslagsbildHitta
1.
  • Matuszewski, Damian J., 1988- (författare)
  • Image and Data Analysis for Biomedical Quantitative Microscopy
  • 2019
  • Doktorsavhandling (övrigt vetenskapligt/konstnärligt)abstract
    • This thesis presents automatic image and data analysis methods to facilitate and improve microscopy-based research and diagnosis. New technologies and computational tools are necessary for handling the ever-growing amounts of data produced in life science. The thesis presents methods developed in three projects with different biomedical applications.In the first project, we analyzed a large high-content screen aimed at enabling personalized medicine for glioblastoma patients. We focused on capturing drug-induced cell-cycle disruption in fluorescence microscopy images of cancer cell cultures. Our main objectives were to identify drugs affecting the cell-cycle and to increase the understanding of different drugs’ mechanisms of action.  Here we present tools for automatic cell-cycle analysis and identification of drugs of interest and their effective doses.In the second project, we developed a feature descriptor for image matching. Image matching is a central pre-processing step in many applications. For example, when two or more images must be matched and registered to create a larger field of view or to analyze differences and changes over time. Our descriptor is rotation-, scale-, and illumination-invariant and it has a short feature vector which makes it computationally attractive. The flexibility to combine it with any feature detector and the customization possibility make it a very versatile tool.In the third project, we addressed two general problems for bridging the gap between deep learning method development and their use in practical scenarios. We developed a method for convolutional neural network training using minimally annotated images. In many biomedical applications, the objects of interest cannot be accurately delineated due to their fuzzy shape, ambiguous morphology, image quality, or the expert knowledge and time it requires. The minimal annotations, in this case, consist of center-points or centerlines of target objects of approximately known size. We demonstrated our training method in a challenging application of a multi-class semantic segmentation of viruses in transmission electron microscopy images. We also systematically explored the influence of network architecture hyper-parameters on its size and performance and show the possibility to substantially reduce the size of a network without compromising its performance.All methods in this thesis were designed to work with little or no input from biomedical experts but of course, require fine-tuning for new applications. The usefulness of the tools has been demonstrated by collaborators and other researchers and has inspired further development of related algorithms.
  •  
2.
  • Matuszewski, Damian J., et al. (författare)
  • Reducing the U-Net size for practical scenarios : Virus recognition in electron microscopy images
  • 2019
  • Ingår i: Computer Methods and Programs in Biomedicine. - : ELSEVIER IRELAND LTD. - 0169-2607 .- 1872-7565. ; 178, s. 31-39
  • Tidskriftsartikel (refereegranskat)abstract
    • Background and objective: Convolutional neural networks (CNNs) offer human experts-like performance and in the same time they are faster and more consistent in their prediction. However, most of the proposed CNNs require an expensive state-of-the-art hardware which substantially limits their use in practical scenarios and commercial systems, especially for clinical, biomedical and other applications that require on-the-fly analysis. In this paper, we investigate the possibility of making CNNs lighter by parametrizing the architecture and decreasing the number of trainable weights of a popular CNN: U-Net. Methods: In order to demonstrate that comparable results can be achieved with substantially less trainable weights than the original U-Net we used a challenging application of a pixel-wise virus classification in Transmission Electron Microscopy images with minimal annotations (i.e. consisting only of the virus particle centers or centerlines). We explored 4 U-Net hyper-parameters: the number of base feature maps, the feature maps multiplier, the number of the encoding-decoding levels and the number of feature maps in the last 2 convolutional layers. Results: Our experiments lead to two main conclusions: 1) the architecture hyper-parameters are pivotal if less trainable weights are to be used, and 2) if there is no restriction on the trainable weights number using a deeper network generally gives better results. However, training larger networks takes longer, typically requires more data and such networks are also more prone to overfitting. Our best model achieved an accuracy of 82.2% which is similar to the original U-Net while using nearly 4 times less trainable weights (7.8 M in comparison to 31.0 M). We also present a network with < 2M trainable weights that achieved an accuracy of 76.4%. Conclusions: The proposed U-Net hyper-parameter exploration can be adapted to other CNNs and other applications. It allows a comprehensive CNN architecture designing with the aim of a more efficient trainable weight use. Making the networks faster and lighter is crucial for their implementation in many practical applications. In addition, a lighter network ought to be less prone to over-fitting and hence generalize better. (C) 2019 Published by Elsevier B.V.
  •  
Skapa referenser, mejla, bekava och länka
  • Resultat 1-2 av 2

Kungliga biblioteket hanterar dina personuppgifter i enlighet med EU:s dataskyddsförordning (2018), GDPR. Läs mer om hur det funkar här.
Så här hanterar KB dina uppgifter vid användning av denna tjänst.

 
pil uppåt Stäng

Kopiera och spara länken för att återkomma till aktuell vy