SwePub
Sök i SwePub databas

  Extended search

Träfflista för sökning "(WFRF:(Danelljan Martin)) srt2:(2020)"

Search: (WFRF:(Danelljan Martin)) > (2020)

  • Result 1-6 of 6
Sort/group result
   
EnumerationReferenceCoverFind
1.
  • Goutam, Bhat, et al. (author)
  • Learning What to Learn for Video Object Segmentation
  • 2020
  • In: Computer Vision. - Cham : Springer International Publishing. - 9783030585358 - 9783030585365 ; , s. 777-794
  • Conference paper (peer-reviewed)abstract
    • Video object segmentation (VOS) is a highly challengingproblem, since the target object is only defined by a first-frame refer-ence mask during inference. The problem of how to capture and utilizethis limited information to accurately segment the target remains a fun-damental research question. We address this by introducing an end-to-end trainable VOS architecture that integrates a differentiable few-shotlearner. Our learner is designed to predict a powerful parametric modelof the target by minimizing a segmentation error in the first frame. Wefurther go beyond the standard few-shot learning paradigm by learningwhat our target model should learn in order to maximize segmentationaccuracy. We perform extensive experiments on standard benchmarks.Our approach sets a new state-of-the-art on the large-scale YouTube-VOS 2018 dataset by achieving an overall score of 81.5, corresponding toa 2.6% relative improvement over the previous best result. The code andmodels are available at https://github.com/visionml/pytracking.
  •  
2.
  •  
3.
  • Gustafsson, Fredrik K., et al. (author)
  • Evaluating Scalable Bayesian Deep Learning Methods for Robust Computer Vision
  • 2020
  • In: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW 2020). - : IEEE Computer Society. - 9781728193601 ; , s. 1289-1298
  • Conference paper (peer-reviewed)abstract
    • While deep neural networks have become the go-to approach in computer vision, the vast majority of these models fail to properly capture the uncertainty inherent in their predictions. Estimating this predictive uncertainty can be crucial, for example in automotive applications. In Bayesian deep learning, predictive uncertainty is commonly decomposed into the distinct types of aleatoric and epistemic uncertainty. The former can be estimated by letting a neural network output the parameters of a certain probability distribution. Epistemic uncertainty estimation is a more challenging problem, and while different scalable methods recently have emerged, no extensive comparison has been performed in a real-world setting. We therefore accept this task and propose a comprehensive evaluation framework for scalable epistemic uncertainty estimation methods in deep learning. Our proposed framework is specifically designed to test the robustness required in real-world computer vision applications. We also apply this framework to provide the first properly extensive and conclusive comparison of the two current state-of-the-art scalable methods: ensembling and MC-dropout. Our comparison demonstrates that ensembling consistently provides more reliable and practically useful uncertainty estimates. Code is available at https://github.com/fregu856/evaluating_bdl.
  •  
4.
  •  
5.
  • Kristan, M., et al. (author)
  • The Eighth Visual Object Tracking VOT2020 Challenge Results
  • 2020
  • In: Computer Vision. - Cham : Springer International Publishing. - 9783030682378 ; , s. 547-601
  • Conference paper (peer-reviewed)abstract
    • The Visual Object Tracking challenge VOT2020 is the eighth annual tracker benchmarking activity organized by the VOT initiative. Results of 58 trackers are presented; many are state-of-the-art trackers published at major computer vision conferences or in journals in the recent years. The VOT2020 challenge was composed of five sub-challenges focusing on different tracking domains: (i) VOT-ST2020 challenge focused on short-term tracking in RGB, (ii) VOT-RT2020 challenge focused on “real-time” short-term tracking in RGB, (iii) VOT-LT2020 focused on long-term tracking namely coping with target disappearance and reappearance, (iv) VOT-RGBT2020 challenge focused on short-term tracking in RGB and thermal imagery and (v) VOT-RGBD2020 challenge focused on long-term tracking in RGB and depth imagery. Only the VOT-ST2020 datasets were refreshed. A significant novelty is introduction of a new VOT short-term tracking evaluation methodology, and introduction of segmentation ground truth in the VOT-ST2020 challenge – bounding boxes will no longer be used in the VOT-ST challenges. A new VOT Python toolkit that implements all these novelites was introduced. Performance of the tested trackers typically by far exceeds standard baselines. The source code for most of the trackers is publicly available from the VOT page. The dataset, the evaluation kit and the results are publicly available at the challenge website (http://votchallenge.net ). 
  •  
6.
  • Robinson, Andreas, 1975-, et al. (author)
  • Learning Fast and Robust Target Models for Video Object Segmentation
  • 2020
  • In: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). - : IEEE. - 9781728171685 ; , s. 7404-7413
  • Conference paper (peer-reviewed)abstract
    • Video object segmentation (VOS) is a highly challenging problem since the initial mask, defining the target object, is only given at test-time. The main difficulty is to effectively handle appearance changes and similar background objects, while maintaining accurate segmentation. Most previous approaches fine-tune segmentation networks on the first frame, resulting in impractical frame-rates and risk of overfitting. More recent methods integrate generative target appearance models, but either achieve limited robustness or require large amounts of training data. We propose a novel VOS architecture consisting of two network components. The target appearance model consists of a light-weight module, which is learned during the inference stage using fast optimization techniques to predict a coarse but robust target segmentation. The segmentation model is exclusively trained offline, designed to process the coarse scores into high quality segmentation masks. Our method is fast, easily trainable and remains highly effective in cases of limited training data. We perform extensive experiments on the challenging YouTube-VOS and DAVIS datasets. Our network achieves favorable performance, while operating at higher frame-rates compared to state-of-the-art. Code and trained models are available at https://github.com/andr345/frtm-vos.
  •  
Skapa referenser, mejla, bekava och länka
  • Result 1-6 of 6
Type of publication
conference paper (6)
Type of content
peer-reviewed (6)
Author/Editor
Gustafsson, Fredrik ... (3)
Schön, Thomas B., Pr ... (3)
Timofte, Radu (2)
Chen, S. (1)
Chen, Y. (1)
Jiang, Y. (1)
show more...
Li, B. (1)
Li, H. (1)
Li, Y. (1)
Liu, K. (1)
Peng, H. (1)
Wang, F. (1)
Yu, J. (1)
Zhang, H. (1)
Zhang, L. (1)
Zhang, X. (1)
Zhang, Z. (1)
Yao, Y. (1)
Li, J. (1)
Chen, G. (1)
Choi, S. (1)
Wu, Z. (1)
Wang, D. (1)
Wang, Y. (1)
Zhu, X. (1)
Wang, Z. (1)
Wang, L (1)
Yang, X. (1)
Zhang, P (1)
Lee, J. (1)
Yang, J. (1)
Wang, N. (1)
Wang, Q. (1)
Xu, J (1)
Tang, Z. (1)
Zhao, S (1)
Fernandez, G (1)
Gu, Y. (1)
Cheng, L (1)
Lu, W (1)
Fan, H (1)
Zhao, H (1)
Yu, K (1)
Lu, H (1)
Ye, Y. (1)
Xu, T. (1)
Ma, Z (1)
Zhou, W. (1)
Gustafsson, F. (1)
Lee, Y (1)
show less...
University
Uppsala University (3)
Linköping University (3)
Language
English (6)
Research subject (UKÄ/SCB)
Natural sciences (5)
Engineering and Technology (1)
Year

Kungliga biblioteket hanterar dina personuppgifter i enlighet med EU:s dataskyddsförordning (2018), GDPR. Läs mer om hur det funkar här.
Så här hanterar KB dina uppgifter vid användning av denna tjänst.

 
pil uppåt Close

Copy and save the link in order to return to this view