SwePub
Sök i SwePub databas

  Utökad sökning

Booleska operatorer måste skrivas med VERSALER

Träfflista för sökning "AMNE:(NATURAL SCIENCES) AMNE:(Computer and Information Sciences) AMNE:(Computer Vision and Robotics) srt2:(2020-2024)"

Sökning: AMNE:(NATURAL SCIENCES) AMNE:(Computer and Information Sciences) AMNE:(Computer Vision and Robotics) > (2020-2024)

  • Resultat 1-10 av 1306
Sortera/gruppera träfflistan
   
NumreringReferensOmslagsbildHitta
1.
  • Lindgren, Helena, Professor, et al. (författare)
  • The wasp-ed AI curriculum : A holistic curriculum for artificial intelligence
  • 2023
  • Ingår i: INTED2023 Proceedings. - : IATED. - 9788409490264 ; , s. 6496-6502
  • Konferensbidrag (refereegranskat)abstract
    • Efforts in lifelong learning and competence development in Artificial Intelligence (AI) have been on the rise for several years. These initiatives have mostly been applied to Science, Technology, Engineering and Mathematics (STEM) disciplines. Even though there has been significant development in Digital Humanities to incorporate AI methods and tools in higher education, the potential for such competences in Arts, Humanities and Social Sciences is far from being realised. Furthermore, there is an increasing awareness that the STEM disciplines need to include competences relating to AI in humanity and society. This is especially important considering the widening and deepening of the impact of AI on society at large and individuals. The aim of the presented work is to provide a broad and inclusive AI Curriculum that covers the breadth of the topic as it is seen today, which is significantly different from only a decade ago. It is important to note that with the curriculum we mean an overview of the subject itself, rather than a particular education program. The curriculum is intended to be used as a foundation for educational activities in AI to for example harmonize terminology, compare different programs, and identify educational gaps to be filled. An important aspect of the curriculum is the ethical, legal, and societal aspects of AI and to not limit the curriculum to the STEM subjects, instead extending to a holistic, human-centred AI perspective. The curriculum is developed as part of the national research program WASP-ED, the Wallenberg AI and transformative technologies education development program. 
  •  
2.
  • Blanch, Krister, 1991 (författare)
  • Beyond-application datasets and automated fair benchmarking
  • 2023
  • Licentiatavhandling (övrigt vetenskapligt/konstnärligt)abstract
    • Beyond-application perception datasets are generalised datasets that emphasise the fundamental components of good machine perception data. When analysing the history of perception datatsets, notable trends suggest that design of the dataset typically aligns with an application goal. Instead of focusing on a specific application, beyond-application datasets instead look at capturing high-quality, high-volume data from a highly kinematic environment, for the purpose of aiding algorithm development and testing in general. Algorithm benchmarking is a cornerstone of autonomous systems development, and allows developers to demonstrate their results in a comparative manner. However, most benchmarking systems allow developers to use their own hardware or select favourable data. There is also little focus on run time performance and consistency, with benchmarking systems instead showcasing algorithm accuracy. By combining both beyond-application dataset generation and methods for fair benchmarking, there is also the dilemma of how to provide the dataset to developers for this benchmarking, as the result of a high-volume, high-quality dataset generation is a significant increase in dataset size when compared to traditional perception datasets. This thesis presents the first results of attempting the creation of such a dataset. The dataset was built using a maritime platform, selected due to the highly dynamic environment presented on water. The design and initial testing of this platform is detailed, as well as as methods of sensor validation. Continuing, the thesis then presents a method of fair benchmarking, by utilising remote containerisation in a way that allows developers to present their software to the dataset, instead of having to first locally store a copy. To test this dataset and automatic online benchmarking, a number of reference algorithms were required for initial results. Three algorithms were built, using the data from three different sensors captured on the maritime platform. Each algorithm calculates vessel odometry, and the automatic benchmarking system was utilised to show the accuracy and run-time performance of these algorithms. It was found that the containerised approach alleviated data management concerns, prevented inflated accuracy results, and demonstrated precisely how computationally intensive each algorithm was.
  •  
3.
  • Suchan, Jakob, et al. (författare)
  • Commonsense Visual Sensemaking for Autonomous Driving : On Generalised Neurosymbolic Online Abduction Integrating Vision and Semantics
  • 2021
  • Ingår i: Artificial Intelligence. - : Elsevier. - 0004-3702 .- 1872-7921. ; 299
  • Tidskriftsartikel (refereegranskat)abstract
    • We demonstrate the need and potential of systematically integrated vision and semantics solutions for visual sensemaking in the backdrop of autonomous driving. A general neurosymbolic method for online visual sensemaking using answer set programming (ASP) is systematically formalised and fully implemented. The method integrates state of the art in visual computing, and is developed as a modular framework that is generally usable within hybrid architectures for realtime perception and control. We evaluate and demonstrate with community established benchmarks KITTIMOD, MOT-2017, and MOT-2020. As use-case, we focus on the significance of human-centred visual sensemaking —e.g., involving semantic representation and explainability, question-answering, commonsense interpolation— in safety-critical autonomous driving situations. The developed neurosymbolic framework is domain-independent, with the case of autonomous driving designed to serve as an exemplar for online visual sensemaking in diverse cognitive interaction settings in the backdrop of select human-centred AI technology design considerations.
  •  
4.
  • Ali, Muhaddisa Barat, 1986 (författare)
  • Deep Learning Methods for Classification of Gliomas and Their Molecular Subtypes, From Central Learning to Federated Learning
  • 2023
  • Doktorsavhandling (övrigt vetenskapligt/konstnärligt)abstract
    • The most common type of brain cancer in adults are gliomas. Under the updated 2016 World Health Organization (WHO) tumor classification in central nervous system (CNS), identification of molecular subtypes of gliomas is important. For low grade gliomas (LGGs), prediction of molecular subtypes by observing magnetic resonance imaging (MRI) scans might be difficult without taking biopsy. With the development of machine learning (ML) methods such as deep learning (DL), molecular based classification methods have shown promising results from MRI scans that may assist clinicians for prognosis and deciding on a treatment strategy. However, DL requires large amount of training datasets with tumor class labels and tumor boundary annotations. Manual annotation of tumor boundary is a time consuming and expensive process. The thesis is based on the work developed in five papers on gliomas and their molecular subtypes. We propose novel methods that provide improved performance.  The proposed methods consist of a multi-stream convolutional autoencoder (CAE)-based classifier, a deep convolutional generative adversarial network (DCGAN) to enlarge the training dataset, a CycleGAN to handle domain shift, a novel federated learning (FL) scheme to allow local client-based training with dataset protection, and employing bounding boxes to MRIs when tumor boundary annotations are not available. Experimental results showed that DCGAN generated MRIs have enlarged the original training dataset size and have improved the classification performance on test sets. CycleGAN showed good domain adaptation on multiple source datasets and improved the classification performance. The proposed FL scheme showed a slightly degraded performance as compare to that of central learning (CL) approach while protecting dataset privacy. Using tumor bounding boxes showed to be an alternative approach to tumor boundary annotation for tumor classification and segmentation, with a trade-off between a slight decrease in performance and saving time in manual marking by clinicians. The proposed methods may benefit the future research in bringing DL tools into clinical practice for assisting tumor diagnosis and help the decision making process.
  •  
5.
  • Jacobsson, Martin, 1976-, et al. (författare)
  • A Drone-mounted Depth Camera-based Motion Capture System for Sports Performance Analysis
  • 2023
  • Ingår i: Artificial Intelligence in HCI. - : Springer Nature. - 9783031358937 ; , s. 489-503
  • Konferensbidrag (refereegranskat)abstract
    • Video is the most used tool for sport performance analysis as it provides a common reference point for the coach and the athlete. The problem with video is that it is a subjective tool. To overcome this, motion capture systems can used to get an objective 3D model of a per- son’s posture and motion, but only in laboratory settings. Unfortunately, many activities, such as most outdoor sports, cannot be captured in a lab without compromising the activity. In this paper, we propose to use an aerial drone system equipped with depth cameras, AI-based marker- less motion capture software to perform automatic skeleton tracking and real-time sports performance analysis of athletes. We experiment with off-the-shelf drone systems, miniaturized depth cameras, and commer- cially available skeleton tracking software to build a system for analyzing sports-related performance of athletes in their real settings. To make this a fully working system, we have conducted a few initial experiments and identified many issues that still needs to be addressed.
  •  
6.
  • Frid, Emma, 1988-, et al. (författare)
  • Perceptual Evaluation of Blended Sonification of Mechanical Robot Sounds Produced by Emotionally Expressive Gestures : Augmenting Consequential Sounds to Improve Non-verbal Robot Communication
  • 2021
  • Ingår i: International Journal of Social Robotics. - : Springer Nature. - 1875-4791 .- 1875-4805.
  • Tidskriftsartikel (refereegranskat)abstract
    • This paper presents two experiments focusing on perception of mechanical sounds produced by expressive robot movement and blended sonifications thereof. In the first experiment, 31 participants evaluated emotions conveyed by robot sounds through free-form text descriptions. The sounds were inherently produced by the movements of a NAO robot and were not specifically designed for communicative purposes. Results suggested no strong coupling between the emotional expression of gestures and how sounds inherent to these movements were perceived by listeners; joyful gestures did not necessarily result in joyful sounds. A word that reoccurred in text descriptions of all sounds, regardless of the nature of the expressive gesture, was “stress”. In the second experiment, blended sonification was used to enhance and further clarify the emotional expression of the robot sounds evaluated in the first experiment. Analysis of quantitative ratings of 30 participants revealed that the blended sonification successfully contributed to enhancement of the emotional message for sound models designed to convey frustration and joy. Our findings suggest that blended sonification guided by perceptual research on emotion in speech and music can successfully improve communication of emotions through robot sounds in auditory-only conditions.
  •  
7.
  • Koriakina, Nadezhda, 1991-, et al. (författare)
  • Deep multiple instance learning versus conventional deep single instance learning for interpretable oral cancer detection
  • 2024
  • Ingår i: PLOS ONE. - : Public Library of Science (PLoS). - 1932-6203. ; 19:4 April
  • Tidskriftsartikel (refereegranskat)abstract
    • The current medical standard for setting an oral cancer (OC) diagnosis is histological examination of a tissue sample taken from the oral cavity. This process is time-consuming and more invasive than an alternative approach of acquiring a brush sample followed by cytological analysis. Using a microscope, skilled cytotechnologists are able to detect changes due to malignancy; however, introducing this approach into clinical routine is associated with challenges such as a lack of resources and experts. To design a trustworthy OC detection system that can assist cytotechnologists, we are interested in deep learning based methods that can reliably detect cancer, given only per-patient labels (thereby minimizing annotation bias), and also provide information regarding which cells are most relevant for the diagnosis (thereby enabling supervision and understanding). In this study, we perform a comparison of two approaches suitable for OC detection and interpretation: (i) conventional single instance learning (SIL) approach and (ii) a modern multiple instance learning (MIL) method. To facilitate systematic evaluation of the considered approaches, we, in addition to a real OC dataset with patient-level ground truth annotations, also introduce a synthetic dataset—PAP-QMNIST. This dataset shares several properties of OC data, such as image size and large and varied number of instances per bag, and may therefore act as a proxy model of a real OC dataset, while, in contrast to OC data, it offers reliable per-instance ground truth, as defined by design. PAP-QMNIST has the additional advantage of being visually interpretable for non-experts, which simplifies analysis of the behavior of methods. For both OC and PAP-QMNIST data, we evaluate performance of the methods utilizing three different neural network architectures. Our study indicates, somewhat surprisingly, that on both synthetic and real data, the performance of the SIL approach is better or equal to the performance of the MIL approach. Visual examination by cytotechnologist indicates that the methods manage to identify cells which deviate from normality, including malignant cells as well as those suspicious for dysplasia. We share the code as open source.
  •  
8.
  • Lv, Zhihan, Dr. 1984-, et al. (författare)
  • Editorial : 5G for Augmented Reality
  • 2022
  • Ingår i: Mobile Networks and Applications. - : Springer. - 1383-469X .- 1572-8153.
  • Tidskriftsartikel (refereegranskat)
  •  
9.
  • Norlund, Tobias, 1991, et al. (författare)
  • Transferring Knowledge from Vision to Language: How to Achieve it and how to Measure it?
  • 2021
  • Ingår i: Proceedings of the Fourth BlackboxNLP Workshop on Analyzing and Interpreting Neural Networks for NLP, pp. 149-162, Punta Cana, Dominican Republic. - : Association for Computational Linguistics.
  • Konferensbidrag (refereegranskat)abstract
    • Large language models are known to suffer from the hallucination problem in that they are prone to output statements that are false or inconsistent, indicating a lack of knowledge. A proposed solution to this is to provide the model with additional data modalities that complements the knowledge obtained through text. We investigate the use of visual data to complement the knowledge of large language models by proposing a method for evaluating visual knowledge transfer to text for uni- or multimodal language models. The method is based on two steps, 1) a novel task querying for knowledge of memory colors, i.e. typical colors of well-known objects, and 2) filtering of model training data to clearly separate knowledge contributions. Additionally, we introduce a model architecture that involves a visual imagination step and evaluate it with our proposed method. We find that our method can successfully be used to measure visual knowledge transfer capabilities in models and that our novel model architecture shows promising results for leveraging multimodal knowledge in a unimodal setting.
  •  
10.
  • Ghielmetti, N., et al. (författare)
  • Real-time semantic segmentation on FPGAs for autonomous vehicles with hls4ml
  • 2022
  • Ingår i: Machine Learning - Science and Technology. - : IOP Publishing. - 2632-2153. ; 3:4
  • Tidskriftsartikel (refereegranskat)abstract
    • In this paper, we investigate how field programmable gate arrays can serve as hardware accelerators for real-time semantic segmentation tasks relevant for autonomous driving. Considering compressed versions of the ENet convolutional neural network architecture, we demonstrate a fully-on-chip deployment with a latency of 4.9 ms per image, using less than 30% of the available resources on a Xilinx ZCU102 evaluation board. The latency is reduced to 3 ms per image when increasing the batch size to ten, corresponding to the use case where the autonomous vehicle receives inputs from multiple cameras simultaneously. We show, through aggressive filter reduction and heterogeneous quantization-aware training, and an optimized implementation of convolutional layers, that the power consumption and resource utilization can be significantly reduced while maintaining accuracy on the Cityscapes dataset.
  •  
Skapa referenser, mejla, bekava och länka
  • Resultat 1-10 av 1306
Typ av publikation
konferensbidrag (590)
tidskriftsartikel (556)
doktorsavhandling (63)
bokkapitel (24)
rapport (19)
forskningsöversikt (19)
visa fler...
licentiatavhandling (16)
annan publikation (12)
samlingsverk (redaktörskap) (3)
konstnärligt arbete (3)
bok (1)
patent (1)
visa färre...
Typ av innehåll
refereegranskat (1133)
övrigt vetenskapligt/konstnärligt (167)
populärvet., debatt m.m. (1)
Författare/redaktör
Khan, Fahad (37)
Khan, Salman (34)
Nikolakopoulos, Geor ... (32)
Liwicki, Marcus (24)
Kragic, Danica, 1971 ... (21)
Oskarsson, Magnus (17)
visa fler...
Svensson, Lennart, 1 ... (17)
Felsberg, Michael, 1 ... (17)
Åström, Kalle (16)
Sattler, Torsten, 19 ... (16)
Pollefeys, Marc (16)
Zach, Christopher, 1 ... (16)
Kanellakis, Christof ... (15)
Khan, Fahad Shahbaz, ... (15)
Andreasson, Henrik, ... (14)
Kahl, Fredrik, 1972 (14)
Sladoje, Nataša (14)
Anwer, Rao Muhammad (14)
Felsberg, Michael (14)
Heyden, Anders (13)
Shah, Mubarak (12)
Lv, Zhihan, Dr. 1984 ... (12)
Larsson, Viktor (12)
Karayiannidis, Yiann ... (12)
Lindblad, Joakim (12)
Ho, Luis C. (11)
Danelljan, Martin (11)
Wymeersch, Henk, 197 ... (10)
Loutfi, Amy, 1978- (10)
Dobnik, Simon, 1977 (10)
Stricker, Didier (10)
Conway, John, 1963 (10)
Berger, Christian, 1 ... (10)
Servin, Martin (10)
Bekiroglu, Yasemin, ... (10)
Cholakkal, Hisham (10)
Afzal, Muhammad Zesh ... (9)
Britzen, Silke (9)
Broderick, Avery E. (9)
Chen, Yongjun (9)
Cui, Yuzhu (9)
Fromm, Christian M. (9)
Galison, Peter (9)
Georgiev, Boris (9)
James, David J. (9)
Jeter, Britton (9)
Jensfelt, Patric, 19 ... (9)
Palmieri, Luigi (9)
Arras, Kai O. (9)
Björkman, Mårten, 19 ... (9)
visa färre...
Lärosäte
Chalmers tekniska högskola (314)
Kungliga Tekniska Högskolan (282)
Linköpings universitet (164)
Lunds universitet (138)
Luleå tekniska universitet (101)
Uppsala universitet (91)
visa fler...
Örebro universitet (89)
Umeå universitet (67)
Göteborgs universitet (59)
Högskolan i Halmstad (27)
Blekinge Tekniska Högskola (25)
Mittuniversitetet (21)
Högskolan i Skövde (20)
RISE (18)
Stockholms universitet (14)
Mälardalens universitet (14)
Linnéuniversitetet (12)
Karolinska Institutet (12)
Sveriges Lantbruksuniversitet (11)
Jönköping University (10)
Högskolan Väst (4)
Malmö universitet (3)
Högskolan Dalarna (2)
Stockholms konstnärliga högskola (2)
Försvarshögskolan (1)
VTI - Statens väg- och transportforskningsinstitut (1)
IVL Svenska Miljöinstitutet (1)
visa färre...
Språk
Engelska (1302)
Svenska (4)
Forskningsämne (UKÄ/SCB)
Naturvetenskap (1305)
Teknik (418)
Medicin och hälsovetenskap (71)
Samhällsvetenskap (41)
Humaniora (24)
Lantbruksvetenskap (23)

År

Kungliga biblioteket hanterar dina personuppgifter i enlighet med EU:s dataskyddsförordning (2018), GDPR. Läs mer om hur det funkar här.
Så här hanterar KB dina uppgifter vid användning av denna tjänst.

 
pil uppåt Stäng

Kopiera och spara länken för att återkomma till aktuell vy