SwePub
Sök i SwePub databas

  Extended search

Boolean operators must be entered wtih CAPITAL LETTERS

Träfflista för sökning "hsv:(NATURAL SCIENCES) hsv:(Computer and Information Sciences) "

Search: hsv:(NATURAL SCIENCES) hsv:(Computer and Information Sciences)

  • Result 1-50 of 67535
Sort/group result
   
EnumerationReferenceCoverFind
1.
  • Wilhelmsson, Kenneth, 1976 (author)
  • Huvudansatser för parsningsmetoder. Om programutvecklingens förutsättningar i en svensk kontext
  • 2016
  • Reports (other academic/artistic)abstract
    • Syftet med denna text var att ge en inblick i området (syntaktisk) parsning. Tanken var att ge en bild av utvecklingen som var 1) fri från alltför tekniska detaljer, då området är programmeringstekniskt, och 2) beskriven ur ett svenskt perspektiv. Bakgrunden till valet av ämne till texten, som var tänkt att finnas med i antologin Text och kontext, var att parsning är relativt okänt för många personer verksamma inom närliggande områden, samtidigt som det är ett absolut nyckelbegrepp för den som ägnar sig åt datorlingvistik eller språkteknologi. Målet var alltså att ge en ganska allmän utifrånblick på några centrala sidor av utvecklingen, samtidigt som det tydligt är så att den som själv arbetat med utveckling kan ha starka åsikter och preferenser rörande metodval, något som i ärlighetens namn kanske inte heller denna text är lösgjord från. Hur ska det göras? Konsten att utveckla automatisk syntaxanalys av naturlig text kan läras ut från ett flertal perspektiv. Det kan t.ex. ske med fokus på användandet av en viss grammatikformalism, med fokus på beräkningssnabbhet, med fokus på entydiggörande av möjliga ambiguiteter. Tolkningsval kan göras med hjälp av antingen handskrivna regler eller inhämtad statistik. En sorts huvudtema i denna text är hur metoder för parsning på senare år uppvisar förändringar som kanske kan förklaras med att programmen har fått andra användningsområden och att metoderna har anpassats därefter (en annan tolkning är att flera senare system inte längre gör parsning i strikt mening). När detta tänkta ”kapitel” var färdigt fick det kommentaren att det inte var anpassat för antologins målgrupp. Det fick skrivas en annan kapiteltext, men det kom samtidigt ett förslag att publicera texten om parsning här som denna rapport.
  •  
2.
  • Wilhelmsson, Kenneth, 1976 (author)
  • Autentiska och artificiella frågor till svensk text Automatisk frågegenerering jämfört med användares frågor för informationsåtkomst : Authentic and artificial questions to Swedish text Automatically generated questions versus user-generated questions for information access
  • 2015
  • Other publication (other academic/artistic)abstract
    • Informationssökning mot ostrukturerade datakällor som fri text är ett av de områden där användargränssnitt med fri formulering i naturligt språk har tagits fram. I ett sådant, eventuellt AI-betonat, system kan några grundläggande svårigheter från användarperspektivet märkas. En sådan svårighet är att en användare inte känner till huruvida en fråga som hon avser att ställa egentligen kan besvaras av den aktuella texten. Denna svårighet, tillsammans med andra, som de kraftiga variationsmöjligheterna för formen för ett giltigt svar på en ställd fråga, riskerar att leda till att användarintrycken av systemtypen blir negativa. De moment som behöver ingå i ett sådant frågebaserat informationssystems funktionssätt måste på något sätt inbegripa en mappning av frågeled i frågan (t.ex. när) till den form och grammatisk funktion som svaret i texten måste ha (för frågan när normalt ett tidsadverbial). Bland annat denna iakttagelse inbjuder till användning av automatisk frågegenerering (question generation, QG). Frågegenerering innebär att frågor som en naturlig text besvarar initialt utvinns av ett program som samlar in dem i explicit form. Tanken för användning i informationssökning är att en användare i gränssnittet enbart ska kunna ställa just dessa frågor, vilka faktiskt besvaras av texten. Denna studie gäller just de frågor som ett automatiskt frågegenereringssystem för svenska kan, och genom vidare utveckling, skulle kunna generera för godtycklig digital svensk text. Även om mängden automatiskt genererade frågor och frågeformuleringar kan bli mycket stor, utrymmesmässigt många gånger större än ursprungstexten, så är det tydligt att den beskrivna metoden för frågegenerering för svenska inte kan och troligen inte heller kommer att kunna förmås att skapa alla de frågor och frågeformuleringar som en vanlig användare skulle anse att en viss text besvarar. Men hur väl fungerar då automatiskt genererade frågor i detta sammanhang? Denna uppsats kretsar kring en användarundersökning där undersökningsdeltagare har ombetts att formulera frågor som texter besvarar, och som anses vara relevanta frågor. Den resulterande samlingen frågor undersöktes och kategoriserades. Resultatet av undersökningens huvudfråga visar att bara 20-25 % av användarnas frågeformuleringar skulle kunna genereras direkt automatiskt med aktuell ansats – utan vissa informationstekniska förbättringar. Uppsatsen föreslår viss ny terminologi för detta outforskade område, bl.a. för att skilja mellan de olika grader av processkrav som generering av olika frågeslag från text kräver.
  •  
3.
  •  
4.
  • Chatterjee, Bapi, 1982 (author)
  • Lock-free Concurrent Search
  • 2017
  • Doctoral thesis (other academic/artistic)abstract
    • The contemporary computers typically consist of multiple computing cores with high compute power. Such computers make excellent concurrent asynchronous shared memory system. On the other hand, though many celebrated books on data structure and algorithm provide a comprehensive study of sequential search data structures, unfortunately, we do not have such a luxury if concurrency comes in the setting. The present dissertation aims to address this paucity. We describe novel lock-free algorithms for concurrent data structures that target a variety of search problems. (i) Point search (membership query, predecessor query, nearest neighbour query) for 1-dimensional data: Lock-free linked-list; lock-free internal and external binary search trees (BST). (ii) Range search for 1-dimensional data: A range search method for lock-free ordered set data structures - linked-list, skip-list and BST. (iii) Point search for multi-dimensional data: Lock-free kD-tree, specially, a generic method for nearest neighbour search. We prove that the presented algorithms are linearizable i.e. the concurrent data structure operations intuitively display their sequential behaviour to an observer of the concurrent system. The lock-freedom in the introduced algorithms guarantee overall progress in an asynchronous shared memory system. We present the amortized analysis of lock-free data structures to show their efficiency. Moreover, we provide sample implementations of the algorithms and test them over extensive micro-benchmarks. Our experiments demonstrate that the implementations are scalable and perform well when compared to related existing alternative implementations on common multi-core computers. Our focus is on propounding the generic methodologies for efficient lock-free concurrent search. In this direction, we present the notion of help-optimality, which captures the optimization of amortized step complexity of the operations. In addition to that, we explore the language-portable design of lock-free data structures that aims to simplify an implementation from programmer’s point of view. Finally, our techniques to implement lock-free linearizable range search and nearest neighbour search are independent of the underlying data structures and thus are adaptive to similar data structures.
  •  
5.
  • Yun, Yixiao, 1987, et al. (author)
  • Maximum-Likelihood Object Tracking from Multi-View Video by Combining Homography and Epipolar Constraints
  • 2012
  • In: 6th ACM/IEEE Int'l Conf on Distributed Smart Cameras (ICDSC 12), Oct 30 - Nov.2, 2012, Hong Kong. - 9781450317726 ; , s. 6 pages-
  • Conference paper (peer-reviewed)abstract
    • This paper addresses problem of object tracking in occlusion scenarios, where multiple uncalibrated cameras with overlapping fields of view are used. We propose a novel method where tracking is first done independently for each view and then tracking results are mapped between each pair of views to improve the tracking in individual views, under the assumptions that objects are not occluded in all views and move uprightly on a planar ground which may induce a homography relation between each pair of views. The tracking results are mapped by jointly exploiting the geometric constraints of homography, epipolar and vertical vanishing point. Main contributions of this paper include: (a) formulate a reference model of multi-view object appearance using region covariance for each view; (b) define a likelihood measure based on geodesics on a Riemannian manifold that is consistent with the destination view by mapping both the estimated positions and appearances of tracked object from other views; (c) locate object in each individual view based on maximum likelihood criterion from multi-view estimations of object position. Experiments have been conducted on videos from multiple uncalibrated cameras, where targets experience long-term partial or full occlusions. Comparison with two existing methods and performance evaluations are also made. Test results have shown effectiveness of the proposed method in terms of robustness against tracking drifts caused by occlusions.
  •  
6.
  •  
7.
  • Liu, Yuanhua, 1971, et al. (author)
  • Considering the importance of user profiles in interface design
  • 2009
  • In: User Interfaces. ; , s. 23-
  • Book chapter (other academic/artistic)abstract
    • User profile is a popular term widely employed during product design processes by industrial companies. Such a profile is normally intended to represent real users of a product. The ultimate purpose of a user profile is actually to help designers to recognize or learn about the real user by presenting them with a description of a real user’s attributes, for instance; the user’s gender, age, educational level, attitude, technical needs and skill level. The aim of this chapter is to provide information on the current knowledge and research about user profile issues, as well as to emphasize the importance of considering these issues in interface design. In this chapter, we mainly focus on how users’ difference in expertise affects their performance or activity in various interaction contexts. Considering the complex interaction situations in practice, novice and expert users’ interactions with medical user interfaces of different technical complexity will be analyzed as examples: one focuses on novice and expert users’ difference when interacting with simple medical interfaces, and the other focuses on differences when interacting with complex medical interfaces. Four issues will be analyzed and discussed: (1) how novice and expert users differ in terms of performance during the interaction; (2) how novice and expert users differ in the perspective of cognitive mental models during the interaction; (3) how novice and expert users should be defined in practice; and (4) what are the main differences between novice and expert users’ implications for interface design. Besides describing the effect of users’ expertise difference during the interface design process, we will also pinpoint some potential problems for the research on interface design, as well as some future challenges that academic researchers and industrial engineers should face in practice.
  •  
8.
  • Lu, Zhihan, et al. (author)
  • Multimodal Hand and Foot Gesture Interaction for Handheld Devices
  • 2014
  • In: ACM Transactions on Multimedia Computing, Communications, and Applications (TOMCCAP). - : Association for Computing Machinery (ACM). - 1551-6857 .- 1551-6865. ; 11:1
  • Journal article (peer-reviewed)abstract
    • We present a hand-and-foot-based multimodal interaction approach for handheld devices. Our method combines input modalities (i.e., hand and foot) and provides a coordinated output to both modalities along with audio and video. Human foot gesture is detected and tracked using contour-based template detection (CTD) and Tracking-Learning-Detection (TLD) algorithm. 3D foot pose is estimated from passive homography matrix of the camera. 3D stereoscopic and vibrotactile are used to enhance the immersive feeling. We developed a multimodal football game based on the multimodal approach as a proof-of-concept. We confirm our systems user satisfaction through a user study.
  •  
9.
  • Rumman, Nadine Abu, et al. (author)
  • Skin deformation methods for interactive character animation
  • 2017
  • In: Communications in Computer and Information Science. - Cham : Springer International Publishing. - 1865-0937 .- 1865-0929. ; 693, s. 153-174, s. 153-174
  • Conference paper (peer-reviewed)abstract
    • Character animation is a vital component of contemporary computer games, animated feature films and virtual reality applications. The problem of creating appealing character animation can best be described by the title of the animation bible: “The Illusion of Life”. The focus is not on completing a given motion task, but more importantly on how this motion task is performed by the character. This does not necessarily require realistic behavior, but behavior that is believable. This of course includes the skin deformations when the character is moving. In this paper, we focus on the existing research in the area of skin deformation, ranging from skeleton-based deformation and volume preserving techniques to physically based skinning methods. We also summarize the recent contributions in deformable and soft body simulations for articulated characters, and discuss various geometric and example-based approaches. © Springer International Publishing AG 2017.
  •  
10.
  • Norlund, Tobias, 1991, et al. (author)
  • Transferring Knowledge from Vision to Language: How to Achieve it and how to Measure it?
  • 2021
  • In: Proceedings of the Fourth BlackboxNLP Workshop on Analyzing and Interpreting Neural Networks for NLP, pp. 149-162, Punta Cana, Dominican Republic. - : Association for Computational Linguistics.
  • Conference paper (peer-reviewed)abstract
    • Large language models are known to suffer from the hallucination problem in that they are prone to output statements that are false or inconsistent, indicating a lack of knowledge. A proposed solution to this is to provide the model with additional data modalities that complements the knowledge obtained through text. We investigate the use of visual data to complement the knowledge of large language models by proposing a method for evaluating visual knowledge transfer to text for uni- or multimodal language models. The method is based on two steps, 1) a novel task querying for knowledge of memory colors, i.e. typical colors of well-known objects, and 2) filtering of model training data to clearly separate knowledge contributions. Additionally, we introduce a model architecture that involves a visual imagination step and evaluate it with our proposed method. We find that our method can successfully be used to measure visual knowledge transfer capabilities in models and that our novel model architecture shows promising results for leveraging multimodal knowledge in a unimodal setting.
  •  
11.
  • Lidstrom, D, et al. (author)
  • Agent based match racing simulations : Starting practice
  • 2022
  • In: SNAME 24th Chesapeake Sailing Yacht Symposium, CSYS 2022. - : Society of Naval Architects and Marine Engineers.
  • Conference paper (peer-reviewed)abstract
    • Match racing starts in sailing are strategically complex and of great importance for the outcome of a race. With the return of the America's Cup to upwind starts and the World Match Racing Tour attracting young and development sailors, the tactical skills necessary to master the starts could be trained and learned by means of computer simulations to assess a large range of approaches to the starting box. This project used game theory to model the start of a match race, intending to develop and study strategies using Monte-Carlo tree search to estimate the utility of a player's potential moves throughout a race. Strategies that utilised the utility estimated in different ways were defined and tested against each other through means of simulation and with an expert advice on match racing start strategy from a sailor's perspective. The results show that the strategies that put greater emphasis on what the opponent might do, perform better than those that did not. It is concluded that Monte-Carlo tree search can provide a basis for decision making in match races and that it has potential for further use. 
  •  
12.
  • Strannegård, Claes, 1962, et al. (author)
  • Ecosystem Models Based on Artificial Intelligence
  • 2022
  • In: 34th Workshop of the Swedish Artificial Intelligence Society, SAIS 2022. - : IEEE.
  • Conference paper (peer-reviewed)abstract
    • Ecosystem models can be used for understanding general phenomena of evolution, ecology, and ethology. They can also be used for analyzing and predicting the ecological consequences of human activities on specific ecosystems, e.g., the effects of agriculture, forestry, construction, hunting, and fishing. We argue that powerful ecosystem models need to include reasonable models of the physical environment and of animal behavior. We also argue that several well-known ecosystem models are unsatisfactory in this regard. Then we present the open-source ecosystem simulator Ecotwin, which is built on top of the game engine Unity. To model a specific ecosystem in Ecotwin, we first generate a 3D Unity model of the physical environment, based on topographic or bathymetric data. Then we insert digital 3D models of the organisms of interest into the environment model. Each organism is equipped with a genome and capable of sexual or asexual reproduction. An organism dies if it runs out of some vital resource or reaches its maximum age. The animal models are equipped with behavioral models that include sensors, actions, reward signals, and mechanisms of learning and decision-making. Finally, we illustrate how Ecotwin works by building and running one terrestrial and one marine ecosystem model.
  •  
13.
  • Aramrattana, Maytheewat, 1988-, et al. (author)
  • Team Halmstad Approach to Cooperative Driving in the Grand Cooperative Driving Challenge 2016
  • 2018
  • In: IEEE transactions on intelligent transportation systems (Print). - Piscataway, N.J. : Institute of Electrical and Electronics Engineers Inc.. - 1524-9050 .- 1558-0016. ; 19:4, s. 1248-1261
  • Journal article (peer-reviewed)abstract
    • This paper is an experience report of team Halmstad from the participation in a competition organised by the i-GAME project, the Grand Cooperative Driving Challenge 2016. The competition was held in Helmond, The Netherlands, during the last weekend of May 2016. We give an overview of our car’s control and communication system that was developed for the competition following the requirements and specifications of the i-GAME project. In particular, we describe our implementation of cooperative adaptive cruise control, our solution to the communication and logging requirements, as well as the high level decision making support. For the actual competition we did not manage to completely reach all of the goals set out by the organizers as well as ourselves. However, this did not prevent us from outperforming the competition. Moreover, the competition allowed us to collect data for further evaluation of our solutions to cooperative driving. Thus, we discuss what we believe were the strong points of our system, and discuss post-competition evaluation of the developments that were not fully integrated into our system during competition time. © 2000-2011 IEEE.
  •  
14.
  • Bainomugisha, Engineer, et al. (author)
  • Message from Chairs of SEiA 2018
  • 2018
  • In: Proceedings - International Conference on Software Engineering. - New York, NY, USA : ACM. - 0270-5257. ; 2018, s. x-xi
  • Conference paper (other academic/artistic)
  •  
15.
  • David, I., et al. (author)
  • Blended modeling in commercial and open-source model-driven software engineering tools: A systematic study
  • 2023
  • In: Software and Systems Modeling. - : Springer Science and Business Media LLC. - 1619-1366 .- 1619-1374. ; 22, s. 415-447
  • Journal article (peer-reviewed)abstract
    • Blended modeling aims to improve the user experience of modeling activities by prioritizing the seamless interaction with models through multiple notations over the consistency of the models. Inconsistency tolerance, thus, becomes an important aspect in such settings. To understand the potential of current commercial and open-source modeling tools to support blended modeling, we have designed and carried out a systematic study. We identify challenges and opportunities in the tooling aspect of blended modeling. Specifically, we investigate the user-facing and implementation-related characteristics of existing modeling tools that already support multiple types of notations and map their support for other blended aspects, such as inconsistency tolerance, and elevated user experience. For the sake of completeness, we have conducted a multivocal study, encompassing an academic review, and grey literature review. We have reviewed nearly 5000 academic papers and nearly 1500 entries of grey literature. We have identified 133 candidate tools, and eventually selected 26 of them to represent the current spectrum of modeling tools.
  •  
16.
  • Hujainah, Fadhl Mohammad Omar, 1987, et al. (author)
  • SRPTackle: A semi-automated requirements prioritisation technique for scalable requirements of software system projects
  • 2021
  • In: Information and Software Technology. - : Elsevier BV. - 0950-5849. ; 131
  • Journal article (peer-reviewed)abstract
    • Context Requirement prioritisation (RP) is often used to select the most important system requirements as perceived by system stakeholders. RP plays a vital role in ensuring the development of a quality system with defined constraints. However, a closer look at existing RP techniques reveals that these techniques suffer from some key challenges, such as scalability, lack of quantification, insufficient prioritisation of participating stakeholders, overreliance on the participation of professional expertise, lack of automation and excessive time consumption. These key challenges serve as the motivation for the present research. Objective This study aims to propose a new semiautomated scalable prioritisation technique called ‘SRPTackle’ to address the key challenges. Method SRPTackle provides a semiautomated process based on a combination of a constructed requirement priority value formulation function using a multi-criteria decision-making method (i.e. weighted sum model), clustering algorithms (K-means and K-means++) and a binary search tree to minimise the need for expert involvement and increase efficiency. The effectiveness of SRPTackle is assessed by conducting seven experiments using a benchmark dataset from a large actual software project. Results Experiment results reveal that SRPTackle can obtain 93.0% and 94.65% as minimum and maximum accuracy percentages, respectively. These values are better than those of alternative techniques. The findings also demonstrate the capability of SRPTackle to prioritise large-scale requirements with reduced time consumption and its effectiveness in addressing the key challenges in comparison with other techniques. Conclusion With the time effectiveness, ability to scale well with numerous requirements, automation and clear implementation guidelines of SRPTackle, project managers can perform RP for large-scale requirements in a proper manner, without necessitating an extensive amount of effort (e.g. tedious manual processes, need for the involvement of experts and time workload).
  •  
17.
  • Laaber, C., et al. (author)
  • An Evaluation of Open-Source Software Microbenchmark Suites for Continuous Performance Assessment
  • 2018
  • In: MSR '18 Proceedings of the 15th International Conference on Mining Software Repositories. - New York, NY, USA : ACM Digital Library. - 9781450357166 ; , s. 119-130, s. 119-130
  • Book chapter (other academic/artistic)abstract
    • Continuous integration (CI) emphasizes quick feedback to developers. This is at odds with current practice of performance testing, which predominantely focuses on long-running tests against entire systems in production-like environments. Alternatively, software microbenchmarking attempts to establish a performance baseline for small code fragments in short time. This paper investigates the quality of microbenchmark suites with a focus on suitability to deliver quick performance feedback and CI integration. We study ten open-source libraries written in Java and Go with benchmark suite sizes ranging from 16 to 983 tests, and runtimes between 11 minutes and 8.75 hours. We show that our study subjects include benchmarks with result variability of 50% or higher, indicating that not all benchmarks are useful for reliable discovery of slow-downs. We further artificially inject actual slowdowns into public API methods of the study subjects and test whether test suites are able to discover them. We introduce a performance-test quality metric called the API benchmarking score (ABS). ABS represents a benchmark suite's ability to find slowdowns among a set of defined core API methods. Resulting benchmarking scores (i.e., fraction of discovered slowdowns) vary between 10% and 100% for the study subjects. This paper's methodology and results can be used to (1) assess the quality of existing microbenchmark suites, (2) select a set of tests to be run as part of CI, and (3) suggest or generate benchmarks for currently untested parts of an API.
  •  
18.
  • Laaber, C., et al. (author)
  • Applying test case prioritization to software microbenchmarks
  • 2021
  • In: Empirical Software Engineering. - : Springer Science and Business Media LLC. - 1382-3256 .- 1573-7616. ; 26:6
  • Journal article (peer-reviewed)abstract
    • Regression testing comprises techniques which are applied during software evolution to uncover faults effectively and efficiently. While regression testing is widely studied for functional tests, performance regression testing, e.g., with software microbenchmarks, is hardly investigated. Applying test case prioritization (TCP), a regression testing technique, to software microbenchmarks may help capturing large performance regressions sooner upon new versions. This may especially be beneficial for microbenchmark suites, because they take considerably longer to execute than unit test suites. However, it is unclear whether traditional unit testing TCP techniques work equally well for software microbenchmarks. In this paper, we empirically study coverage-based TCP techniques, employing total and additional greedy strategies, applied to software microbenchmarks along multiple parameterization dimensions, leading to 54 unique technique instantiations. We find that TCP techniques have a mean APFD-P (average percentage of fault-detection on performance) effectiveness between 0.54 and 0.71 and are able to capture the three largest performance changes after executing 29% to 66% of the whole microbenchmark suite. Our efficiency analysis reveals that the runtime overhead of TCP varies considerably depending on the exact parameterization. The most effective technique has an overhead of 11% of the total microbenchmark suite execution time, making TCP a viable option for performance regression testing. The results demonstrate that the total strategy is superior to the additional strategy. Finally, dynamic-coverage techniques should be favored over static-coverage techniques due to their acceptable analysis overhead; however, in settings where the time for prioritzation is limited, static-coverage techniques provide an attractive alternative.
  •  
19.
  • Mahmood, Wardah, 1992, et al. (author)
  • Effects of variability in models: a family of experiments
  • 2022
  • In: Empirical Software Engineering. - : Springer Science and Business Media LLC. - 1382-3256 .- 1573-7616. ; 27:3
  • Journal article (peer-reviewed)abstract
    • The ever-growing need for customization creates a need to maintain software systems in many different variants. To avoid having to maintain different copies of the same model, developers of modeling languages and tools have recently started to provide implementation techniques for such variant-rich systems, notably variability mechanisms, which support implementing the differences between model variants. Available mechanisms either follow the annotative or the compositional paradigm, each of which have dedicated benefits and drawbacks. Currently, language and tool designers select the used variability mechanism often solely based on intuition. A better empirical understanding of the comprehension of variability mechanisms would help them in improving support for effective modeling. In this article, we present an empirical assessment of annotative and compositional variability mechanisms for three popular types of models. We report and discuss findings from a family of three experiments with 164 participants in total, in which we studied the impact of different variability mechanisms during model comprehension tasks. We experimented with three model types commonly found in modeling languages: class diagrams, state machine diagrams, and activity diagrams. We find that, in two out of three experiments, annotative technique lead to better developer performance. Use of the compositional mechanism correlated with impaired performance. For all three considered tasks, the annotative mechanism was preferred over the compositional one in all experiments. We present actionable recommendations concerning support of flexible, tasks-specific solutions, and the transfer of established best practices from the code domain to models.
  •  
20.
  • Mallozzi, Piergiuseppe, 1990, et al. (author)
  • A runtime monitoring framework to enforce invariants on reinforcement learning agents exploring complex environments
  • 2019
  • In: RoSE 2019, IEEE/ACM 2nd International Workshop on Robotics Software Engineering, p.5-12. - : IEEE. - 9781728122496
  • Conference paper (peer-reviewed)abstract
    • © 2019 IEEE. Without prior knowledge of the environment, a software agent can learn to achieve a goal using machine learning. Model-free Reinforcement Learning (RL) can be used to make the agent explore the environment and learn to achieve its goal by trial and error. Discovering effective policies to achieve the goal in a complex environment is a major challenge for RL. Furthermore, in safety-critical applications, such as robotics, an unsafe action may cause catastrophic consequences in the agent or in the environment. In this paper, we present an approach that uses runtime monitoring to prevent the reinforcement learning agent to perform 'wrong' actions and to exploit prior knowledge to smartly explore the environment. Each monitor is de?ned by a property that we want to enforce to the agent and a context. The monitors are orchestrated by a meta-monitor that activates and deactivates them dynamically according to the context in which the agent is learning. We have evaluated our approach by training the agent in randomly generated learning environments. Our results show that our approach blocks the agent from performing dangerous and safety-critical actions in all the generated environments. Besides, our approach helps the agent to achieve its goal faster by providing feedback and shaping its reward during learning.
  •  
21.
  • Memedi, Mevludin, 1983-, et al. (author)
  • Automatic Spiral Analysis for Objective Assessment of Motor Symptoms in Parkinson's Disease
  • 2015
  • In: Sensors. - : MDPI AG. - 1424-8220. ; 15:9, s. 23727-23744
  • Journal article (peer-reviewed)abstract
    • A challenge for the clinical management of advanced Parkinson's disease (PD) patients is the emergence of fluctuations in motor performance, which represents a significant source of disability during activities of daily living of the patients. There is a lack of objective measurement of treatment effects for in-clinic and at-home use that can provide an overview of the treatment response. The objective of this paper was to develop a method for objective quantification of advanced PD motor symptoms related to off episodes and peak dose dyskinesia, using spiral data gathered by a touch screen telemetry device. More specifically, the aim was to objectively characterize motor symptoms (bradykinesia and dyskinesia), to help in automating the process of visual interpretation of movement anomalies in spirals as rated by movement disorder specialists. Digitized upper limb movement data of 65 advanced PD patients and 10 healthy (HE) subjects were recorded as they performed spiral drawing tasks on a touch screen device in their home environment settings. Several spatiotemporal features were extracted from the time series and used as inputs to machine learning methods. The methods were validated against ratings on animated spirals scored by four movement disorder specialists who visually assessed a set of kinematic features and the motor symptom. The ability of the method to discriminate between PD patients and HE subjects and the test-retest reliability of the computed scores were also evaluated. Computed scores correlated well with mean visual ratings of individual kinematic features. The best performing classifier (Multilayer Perceptron) classified the motor symptom (bradykinesia or dyskinesia) with an accuracy of 84% and area under the receiver operating characteristics curve of 0.86 in relation to visual classifications of the raters. In addition, the method provided high discriminating power when distinguishing between PD patients and HE subjects as well as had good test-retest reliability. This study demonstrated the potential of using digital spiral analysis for objective quantification of PD-specific and/or treatment-induced motor symptoms.
  •  
22.
  • Penzenstadler, Birgit, 1981, et al. (author)
  • Bots in Software Engineering
  • 2022
  • In: IEEE Software. - 1937-4194 .- 0740-7459. ; 39:5, s. 101-104
  • Research review (peer-reviewed)
  •  
23.
  • Pir Muhammad, Amna, 1990 (author)
  • Managing Human Factors and Requirements in Agile Development of Automated Vehicles: An Exploration
  • 2022
  • Licentiate thesis (other academic/artistic)abstract
    • Context: Automated Vehicle (AV) technology has evolved significantly in complexity and impact; it is expected to ultimately change urban transporta- tion. However, research shows that vehicle automation can only live up to this expectation if it is defined with human capabilities and limitations in mind. Therefore, it is necessary to bring human factors knowledge to AV developers. Objective: This thesis aims to empirically study how we can effectively bring the required human factors knowledge into large-scale agile AV develop- ment. The research goals are 1) to explore requirements engineering and human factors in agile AV development, 2) to investigate the problems of requirements engineering, human factors, and agile way of working in AV development, and 3) to demonstrate initial solutions to existing problems in agile AV development. Method: We conducted this research in close collaboration with industry, using different empirical methodologies to collect data—including interviews, workshops, and document analysis. To gain in-depth insights, we did a qualita- tive exploratory study to investigate the problem and used a design science approach to develop initial solution in several iterations. Findings and Conclusions: We found that applying human factors knowledge effectively is one of the key problem areas that need to be solved in agile development of artificial intelligence (AI)-intense systems. This motivated us to do an in-depth interview study on how to manage human factors knowl- edge during AV development. From our data, we derived a working definition of human factors for AV development, discovered the relevant properties of agile and human factors, and defined implications for agile ways of working, managing human factors knowledge, and managing requirements. The design science approach allowed us to identify challenges related to agile requirements engineering in three case companies in iterations. Based on these three case studies, we developed a solution strategy to resolve the RE challenges in agile AV development. Moreover, we derived building blocks and described guide- lines for the creation of a requirements strategy, which should describe how requirements are structured, how work is organized, and how RE is integrated into the agile work and feature flow. Future Outlook: In future work, I plan to define a concrete requirement strategy for human factors knowledge in large-scale agile AV development. It could help establishing clear communication channels and practices for incorporating explicit human factors knowledge into AI-based large-scale agile AV development.
  •  
24.
  • Samoaa, Hazem Peter, et al. (author)
  • A systematic mapping study of source code representation for deep learning in software engineering
  • 2022
  • In: Iet Software. - : Institution of Engineering and Technology (IET). - 1751-8806 .- 1751-8814. ; 16:4, s. 351-385
  • Journal article (peer-reviewed)abstract
    • The usage of deep learning (DL) approaches for software engineering has attracted much attention, particularly in source code modelling and analysis. However, in order to use DL, source code needs to be formatted to fit the expected input form of DL models. This problem is known as source code representation. Source code can be represented via different approaches, most importantly, the tree-based, token-based, and graph-based approaches. We use a systematic mapping study to investigate i detail the representation approaches adopted in 103 studies that use DL in the context of software engineering. Thus, studies are collected from 2014 to 2021 from 14 different journals and 27 conferences. We show that each way of representing source code can provide a different, yet orthogonal view of the same source code. Thus, different software engineering tasks might require different (combinations of) code representation approaches, depending on the nature and complexity of the task. Particularly, we show that it is crucial to define whether the DL approach requires lexical, syntactical, or semantic code information. Our analysis shows that a wide range of different representations and combinations of representations (hybrid representations) are used to solve a wide range of common software engineering problems. However, we also observe that current research does not generally attempt to transfer existing representations or models to other studies even though there are other contexts in which these representations and models may also be useful. We believe that there is potential for more reuse and the application of transfer learning when applying DL to software engineering tasks.
  •  
25.
  • Sandklef, Henrik, 1971, et al. (author)
  • Programming with Java
  • 2017
  • Book (other academic/artistic)abstract
    • Course literature in programming with Java
  •  
26.
  • Sundell, Håkan, 1968, et al. (author)
  • NOBLE: non-blocking programming support via lock-free shared abstract data types
  • 2009
  • In: SIGARCH Computer Architecture News. - : ACM, Association for Computing Machinery, Inc.. - 0163-5964 .- 1943-5851. ; 36:5, s. 80-87
  • Journal article (peer-reviewed)abstract
    • An essential part of programming for multi-core and multi-processor includes ef cient and reliable means for sharing data. Lock-free data structures are known as very suitable for this purpose, although experienced to be very complex to design. In this paper, we present a software library of non-blocking abstract data types that have been designed to facilitate lock-free programming for non-experts. The system provides: i) ef cient implementations of the most commonly used data types in concurrent and sequential software design, ii) a lock-free memory management system, and iii) a run time-system. The library provides clear semantics that are at least as strong as those of corresponding lock-based implementations of the respective data types. Our software library can be used for facilitating lockfree programming; its design enables the programmer to: i) replace lock-based components of sequential or parallel code easily and ef ciently , ii) use well-tuned concurrent algorithms inside a software or hardware transactional system. In the paper we describe the design and functionality of the system. We also provide experimental results that show that the library can considerably improve the performance of software systems.
  •  
27.
  • Lindén, Joakim, et al. (author)
  • Evaluating the Robustness of ML Models to Out-of-Distribution Data Through Similarity Analysis
  • 2023
  • In: Commun. Comput. Info. Sci.. - : Springer Science and Business Media Deutschland GmbH. - 9783031429408 ; , s. 348-359, s. 348-359
  • Conference paper (peer-reviewed)abstract
    • In Machine Learning systems, several factors impact the performance of a trained model. The most important ones include model architecture, the amount of training time, the dataset size and diversity. We present a method for analyzing datasets from a use-case scenario perspective, detecting and quantifying out-of-distribution (OOD) data on dataset level. Our main contribution is the novel use of similarity metrics for the evaluation of the robustness of a model by introducing relative Fréchet Inception Distance (FID) and relative Kernel Inception Distance (KID) measures. These relative measures are relative to a baseline in-distribution dataset and are used to estimate how the model will perform on OOD data (i.e. estimate the model accuracy drop). We find a correlation between our proposed relative FID/relative KID measure and the drop in Average Precision (AP) accuracy on unseen data.
  •  
28.
  • Huhnstock, Nikolas Alexander, 1988-, et al. (author)
  • An Infinite Replicated Softmax Model for Topic Modeling
  • 2019
  • In: Modeling Decisions for Artificial Intelligence. - Cham : Springer. - 9783030267728 - 9783030267735 ; , s. 307-318
  • Conference paper (peer-reviewed)abstract
    • In this paper, we describe the infinite replicated Softmax model (iRSM) as an adaptive topic model, utilizing the combination of the infinite restricted Boltzmann machine (iRBM) and the replicated Softmax model (RSM). In our approach, the iRBM extends the RBM by enabling its hidden layer to adapt to the data at hand, while the RSM allows for modeling low-dimensional latent semantic representation from a corpus. The combination of the two results is a method that is able to self-adapt to the number of topics within the document corpus and hence, renders manual identification of the correct number of topics superfluous. We propose a hybrid training approach to effectively improve the performance of the iRSM. An empirical evaluation is performed on a standard data set and the results are compared to the results of a baseline topic model. The results show that the iRSM adapts its hidden layer size to the data and when trained in the proposed hybrid manner outperforms the base RSM model.
  •  
29.
  • Kucher, Kostiantyn, et al. (author)
  • Visual Analysis of Online Social Media to Open Up the Investigation of Stance Phenomena
  • 2016
  • In: Information Visualization. - : Sage Publications. - 1473-8716 .- 1473-8724. ; 15:2, s. 93-116
  • Journal article (peer-reviewed)abstract
    • Online social media are a perfect text source for stance analysis. Stance in human communication is concerned with speaker attitudes, beliefs, feelings and opinions. Expressions of stance are associated with the speakers' view of what they are talking about and what is up for discussion and negotiation in the intersubjective exchange. Taking stance is thus crucial for the social construction of meaning. Increased knowledge of stance can be useful for many application fields such as business intelligence, security analytics, or social media monitoring. In order to process large amounts of text data for stance analyses, linguists need interactive tools to explore the textual sources as well as the processed data based on computational linguistics techniques. Both original texts and derived data are important for refining the analyses iteratively. In this work, we present a visual analytics tool for online social media text data that can be used to open up the investigation of stance phenomena. Our approach complements traditional linguistic analysis techniques and is based on the analysis of utterances associated with two stance categories: sentiment and certainty. Our contributions include (1) the description of a novel web-based solution for analyzing the use and patterns of stance meanings and expressions in human communication over time; and (2) specialized techniques used for visualizing analysis provenance and corpus overview/navigation. We demonstrate our approach by means of text media on a highly controversial scandal with regard to expressions of anger and provide an expert review from linguists who have been using our tool.
  •  
30.
  • Palmquist, Adam, 1983, et al. (author)
  • AUTOMATON : A Gamification Machine Learning Project
  • 2023
  • In: Encyclopedia of Data Science and Machine Learning. - : IGI Global. - 9781799892205 - 1799892204 - 9781799892212 ; , s. 3090-3101
  • Book chapter (peer-reviewed)abstract
    • This article displays a design ethnographic case study on an ongoing machine learning project at a Scandinavian gamification start-up company. From late 2020 until early 2021, the project produced a machine learning proof of concept, later implemented in the gamification start-up´s application programming interface to offer smart gamification. The initial results show promise in using prediction models to automate the cluster model selection affording more functional, autonomous, and scalable user segments that are faster to implement. The finding provides opportunities for gamification (e.g., in learning analytics and health informatics). An identified challenge was performance; the neural networks required hyperparameter fine-tuning, which is time-consuming and limits scalability. Interesting further investigations should consider the neural network fine-tuning process, but also attempt to verify the effectiveness of the cluster models selection compared with a control group.
  •  
31.
  •  
32.
  • Gu, Irene Yu-Hua, 1953, et al. (author)
  • Grassmann Manifold Online Learning and Partial Occlusion Handling for Visual Object Tracking under Bayesian Formulation
  • 2012
  • In: Proceedings - International Conference on Pattern Recognition. - 1051-4651. - 9784990644109 ; , s. 1463-1466
  • Conference paper (peer-reviewed)abstract
    • This paper addresses issues of online learning and occlusion handling in video object tracking. Although manifold tracking is promising, large pose changes and long term partial occlusions of video objects remain challenging.We propose a novel manifold tracking scheme that tackles such problems, with the following main novelties: (a) Online estimation of object appearances on Grassmann manifolds; (b) Optimal criterion-based occlusion handling during online learning; (c) Nonlinear dynamic model for appearance basis matrix and its velocity; (b) Bayesian formulations separately for the tracking and the online learning process. Two particle filters are employed: one is on the manifold for generating appearance particles and another on the linear space for generating affine box particles. Tracking and online updating are performed in alternative fashion to mitigate the tracking drift. Experiments on videos have shown robust tracking performance especially when objects contain significantpose changes accompanied with long-term partial occlusions. Evaluations and comparisons with two existing methods provide further support to the proposed method.
  •  
33.
  • Bergström, Gustav, et al. (author)
  • Evaluating the layout quality of UML class diagrams using machine learning
  • 2022
  • In: Journal of Systems and Software. - : Elsevier BV. - 0164-1212. ; 192
  • Journal article (peer-reviewed)abstract
    • UML is the de facto standard notation for graphically representing software. UML diagrams are used in the analysis, construction, and maintenance of software systems. Mostly, UML diagrams capture an abstract view of a (piece of a) software system. A key purpose of UML diagrams is to share knowledge about the system among developers. The quality of the layout of UML diagrams plays a crucial role in their comprehension. In this paper, we present an automated method for evaluating the layout quality of UML class diagrams. We use machine learning based on features extracted from the class diagram images using image processing. Such an automated evaluator has several uses: (1) From an industrial perspective, this tool could be used for automated quality assurance for class diagrams (e.g., as part of a quality monitor integrated into a DevOps toolchain). For example, automated feedback can be generated once a UML diagram is checked in the project repository. (2) In an educational setting, the evaluator can grade the layout aspect of student assignments in courses on software modeling, analysis, and design. (3) In the field of algorithm design for graph layouts, our evaluator can assess the layouts generated by such algorithms. In this way, this evaluator opens up the road for using machine learning to learn good layouting algorithms. Approach.: We use machine learning techniques to build (linear) regression models based on features extracted from the class diagram images using image processing. As ground truth, we use a dataset of 600+ UML Class Diagrams for which experts manually label the quality of the layout. Contributions.: This paper makes the following contributions: (1) We show the feasibility of the automatic evaluation of the layout quality of UML class diagrams. (2) We analyze which features of UML class diagrams are most strongly related to the quality of their layout. (3) We evaluate the performance of our layout evaluator. (4) We offer a dataset of labeled UML class diagrams. In this dataset, we supply for every diagram the following information: (a) a manually established ground truth of the quality of the layout, (b) an automatically established value for the layout-quality of the diagram (produced by our classifier), and (c) the values of key features of the layout of the diagram (obtained by image processing). This dataset can be used for replication of our study and others to build on and improve on this work. Editor's note: Open Science material was validated by the Journal of Systems and Software Open Science Board.
  •  
34.
  • Brunetta, Carlo, 1992 (author)
  • Cryptographic Tools for Privacy Preservation
  • 2021
  • Doctoral thesis (other academic/artistic)abstract
    • Data permeates every aspect of our daily life and it is the backbone of our digitalized society. Smartphones, smartwatches and many more smart devices measure, collect, modify and share data in what is known as the Internet of Things. Often, these devices don’t have enough computation power/storage space thus out-sourcing some aspects of the data management to the Cloud. Outsourcing computation/storage to a third party poses natural questions regarding the security and privacy of the shared sensitive data. Intuitively, Cryptography is a toolset of primitives/protocols of which security prop- erties are formally proven while Privacy typically captures additional social/legislative requirements that relate more to the concept of “trust” between people, “how” data is used and/or “who” has access to data. This thesis separates the concepts by introducing an abstract model that classifies data leaks into different types of breaches. Each class represents a specific requirement/goal related to cryptography, e.g. confidentiality or integrity, or related to privacy, e.g. liability, sensitive data management and more. The thesis contains cryptographic tools designed to provide privacy guarantees for different application scenarios. In more details, the thesis: (a) defines new encryption schemes that provide formal privacy guarantees such as theoretical privacy definitions like Differential Privacy (DP), or concrete privacy-oriented applications covered by existing regulations such as the European General Data Protection Regulation (GDPR); (b) proposes new tools and procedures for providing verifiable computation’s guarantees in concrete scenarios for post-quantum cryptography or generalisation of signature schemes; (c) proposes a methodology for utilising Machine Learning (ML) for analysing the effective security and privacy of a crypto-tool and, dually, proposes a secure primitive that allows computing specific ML algorithm in a privacy-preserving way; (d) provides an alternative protocol for secure communication between two parties, based on the idea of communicating in a periodically timed fashion.
  •  
35.
  • Dodig-Crnkovic, Gordana, 1955 (author)
  • Cognitive Architectures Based on Natural Info-Computation
  • 2022
  • In: Studies in Applied Philosophy, Epistemology and Rational Ethics. - Cham : Springer. - 2192-6255 .- 2192-6263. ; , s. 3-13, s. 3-13
  • Book chapter (peer-reviewed)abstract
    • At the time when the first models of cognitive architectures have been proposed, some forty years ago, understanding of cognition, embodiment and evolution was substantially different from today’s. So was the state of the art of information physics, information chemistry, bioinformatics, neuroinformatics, computational neuroscience, complexity theory, self-organization, theory of evolution, as well as the basic concepts of information and computation. Novel developments support a constructive interdisciplinary framework for cognitive architectures based on natural morphological computing, where interactions between constituents at different levels of organization of matter-energy and their corresponding time-dependent dynamics, lead to complexification of agency and increased cognitive capacities of living organisms that unfold through evolution. Proposed info-computational framework for naturalizing cognition considers present updates (generalizations) of the concepts of information, computation, cognition, and evolution in order to attain an alignment with the current state of the art in corresponding research fields. Some important open questions are suggested for future research with implications for further development of cognitive and intelligent technologies.
  •  
36.
  • Ferro, Nicola, et al. (author)
  • PROMISE Retreat Report Prospects and Opportunities for Information Access Evaluation
  • 2013
  • In: ACM SIGIR Forum. - : Association for Computing Machinery (ACM). - 0163-5840 .- 1558-0229. ; 46:2, s. 60-84
  • Journal article (other academic/artistic)abstract
    • The PROMISE network of excellence organized a two-days brainstorming workshop on 30th and 31st May 2012 in Padua, Italy, to discuss and envisage future directions and perspectives for the evaluation of information access and retrieval systems in multiple languages and multiple media. This document reports on the outcomes of this event and provides details about the six envisaged research lines: search applications; contextual evaluation; challenges in test collection design and exploitation; component-based evaluation; ongoing evaluation; and signal-aware evaluation. The ultimate goal of the PROMISE retreat is to stimulate and involve the research community along these research lines and to provide funding agencies with effective and scientifically sound ideas for coordinating and supporting information access research.
  •  
37.
  • Fröberg, Rikard, 1971, et al. (author)
  • Introduction to Databases
  • 2018
  • Book (other academic/artistic)abstract
    • Kurlitteratur för bland annat 7,5 hp databaser (delkurs i TIG058 programmeringsteknik och databaser). Kurlitteraturen (läromedlet) består av texter, exempel, övningar, föreläsningspresentationer, videoföreläsningar: Antal slidesidor193 Tid per slidesida i snitt (sv):0:00:56 Antal slides:15 Total tid eng. video:1:24:44 Total tid sv. video:2:46:39 Totalt antal övningar/frågor85 Totalt antal A4-sidor med övningar69 Totalt antal A4-sidor med text och exempel73 Totalt antal A4-sidor med text och övningar142 Materialet underhålls och utvecklas kontinuerligt. Senaste upplaga är från 2018.
  •  
38.
  • Scheuner, Joel, 1991, et al. (author)
  • Performance Benchmarking of Infrastructure-as-a-Service (IaaS) Clouds with CloudWorkBench
  • 2019
  • In: ICPE 2019 - Companion of the 2019 ACM/SPEC International Conference on Performance Engineering. - New York, NY, USA : ACM. ; , s. 53-56
  • Conference paper (peer-reviewed)abstract
    • The continuing growth of the cloud computing market has led to an unprecedented diversity of cloud services with different performance characteristics. To support service selection, researchers and practitioners conduct cloud performance benchmarking by measuring and objectively comparing the performance of different providers and configurations (e.g., instance types in different data center regions). In this tutorial, we demonstrate how to write performance tests for IaaS clouds using the Web-based benchmarking tool Cloud WorkBench (CWB). We will motivate and introduce benchmarking of IaaS cloud in general, demonstrate the execution of a simple benchmark in a public cloud environment, summarize the CWB tool architecture, and interactively develop and deploy a more advanced benchmark together with the participants.
  •  
39.
  •  
40.
  • Sweidan, Dirar, et al. (author)
  • Predicting Customer Churn in Retailing
  • 2022
  • In: Proceedings 21st IEEE International Conference on Machine Learning and Applications ICMLA 2022. - : IEEE. - 9781665462839 - 9781665462846 ; , s. 635-640
  • Conference paper (peer-reviewed)abstract
    • Customer churn is one of the most challenging problems for digital retailers. With significantly higher costs for acquiring new customers than retaining existing ones, knowledge about which customers are likely to churn becomes essential. This paper reports a case study where a data-driven approach to churn prediction is used for predicting churners and gaining insights about the problem domain. The real-world data set used contains approximately 200 000 customers, describing each customer using more than 50 features. In the pre-processing, exploration, modeling and analysis, attributes related to recency, frequency, and monetary concepts are identified and utilized. In addition, correlations and feature importance are used to discover and understand churn indicators. One important finding is that the churn rate highly depends on the number of previous purchases. In the segment consisting of customers with only one previous purchase, more than 75% will churn, i.e., not make another purchase in the coming year. For customers with at least four previous purchases, the corresponding churn rate is around 25%. Further analysis shows that churning customers in general, and as expected, make smaller purchases and visit the online store less often. In the experimentation, three modeling techniques are evaluated, and the results show that, in particular, Gradient Boosting models can predict churners with relatively high accuracy while obtaining a good balance between precision and recall. 
  •  
41.
  • Tsaloli, Georgia, 1993 (author)
  • Secure and Privacy-Preserving Cloud-Assisted Computing
  • 2022
  • Doctoral thesis (other academic/artistic)abstract
    • Smart devices such as smartphones, wearables, and smart appliances collect significant amounts of data and transmit them over the network forming the Internet of Things (IoT). Many applications in our daily lives (e.g., health, smart grid, traffic monitoring) involve IoT devices that often have low computational capabilities. Subsequently, powerful cloud servers are employed to process the data collected from these devices. Nevertheless, security and privacy concerns arise in cloud-assisted computing settings. Collected data can be sensitive, and it is essential to protect their confidentiality. Additionally, outsourcing computations to untrusted cloud servers creates the need to ensure that servers perform the computations as requested and that any misbehavior can be detected, safeguarding security. Cryptographic primitives and protocols are the foundation to design secure and privacy-preserving solutions that address these challenges. This thesis focuses on providing privacy and security guarantees when outsourcing heavy computations on sensitive data to untrusted cloud servers. More concretely, this work: (a)  provides solutions for outsourcing the secure computation of the sum and the product functions in the multi-server, multi-client setting, protecting the sensitive data of the data owners, even against potentially untrusted cloud servers; (b)  provides integrity guarantees for the proposed protocols, by enabling anyone to verify the correctness of the computed function values. More precisely, the employed servers or the clients (depending on the proposed solution) provide specific values which are the proofs that the computed results are correct; (c)  designs decentralized settings, where multiple cloud servers are employed to perform the requested computations as opposed to relying on a single server that might fail or lose connection; (d)  suggests ways to protect individual privacy and provide integrity. More pre- cisely, we propose a verifiable differentially private solution that provides verifiability and avoids any leakage of information regardless of the participa- tion of some individual’s sensitive data in the computation or not.
  •  
42.
  • Dobnik, Simon, 1977 (author)
  • Coordinating spatial perspective in discourse
  • 2012
  • In: Proceedings of the Workshop on Vision and Language 2012 (VL'12): The 2nd Annual Meeting of the EPSRC Network on Vision and Language.
  • Conference paper (other academic/artistic)abstract
    • We present results of an on-line data collection experiment where we investigate the assignment and coordination of spatial perspective between a pair of dialogue participants situated in a constrained virtual environment.
  •  
43.
  • Håkansson, Maria, et al. (author)
  • Facilitating Mobile Music Sharing and Social Interaction with Push!Music
  • 2007
  • In: Proceedings of the 40th Hawaii International Conference on System Sciences. - Los Alamitos, Calif. : IEEE Computer Society Washington. - 1530-1605. - 0769527558 ; , s. 87-
  • Conference paper (peer-reviewed)abstract
    • Push!Music is a novel mobile music listening and sharing system, where users automatically receive songs that have autonomously recommended themselves from nearby players depending on similar listening behaviour and music history. Push!Music also enables users to wirelessly send songs between each other as personal recommendations. We conducted a two-week preliminary user study of Push!Music, where a group of five friends used the application in their everyday life. We learned for example that the shared music in Push!Music became a start for social interaction and that received songs in general were highly appreciated and could be looked upon as 'treats'.
  •  
44.
  • Picazo-Sanchez, Pablo, 1985, et al. (author)
  • Are chrome extensions compliant with the spirit of least privilege?
  • 2022
  • In: International Journal of Information Security. - : Springer Science and Business Media LLC. - 1615-5262 .- 1615-5270. ; 21:6, s. 1283-1297
  • Journal article (peer-reviewed)abstract
    • Extensions are small applications installed by users and enrich the user experience of browsing the Internet. Browsers expose a set of restricted APIs to extensions. To be used, extensions need to list the permissions associated with these APIs in a mandatory extension file named manifest. In particular, Chrome’s permission ecosystem was designed in the spirit of the least privilege. Yet, this paper demonstrates that 39.8% of the analyzed extensions provided by the official Web Store are compliant with the spirit of least privilege. Also, we develop: (1) a browser extension to make aware regular users of the permissions the extensions they install; (2) a web app where extensions developers can check whether their extensions are compliant with the spirit of the least privileged; and (3) a set of scripts that can be part of the vendors’ acceptance criteria such that when developers upload their extensions to the official repositories, the scripts automatically analyze the extensions and generate a report about the permissions and the usage.
  •  
45.
  • Suchan, Jakob, et al. (author)
  • Commonsense Visual Sensemaking for Autonomous Driving : On Generalised Neurosymbolic Online Abduction Integrating Vision and Semantics
  • 2021
  • In: Artificial Intelligence. - : Elsevier. - 0004-3702 .- 1872-7921. ; 299
  • Journal article (peer-reviewed)abstract
    • We demonstrate the need and potential of systematically integrated vision and semantics solutions for visual sensemaking in the backdrop of autonomous driving. A general neurosymbolic method for online visual sensemaking using answer set programming (ASP) is systematically formalised and fully implemented. The method integrates state of the art in visual computing, and is developed as a modular framework that is generally usable within hybrid architectures for realtime perception and control. We evaluate and demonstrate with community established benchmarks KITTIMOD, MOT-2017, and MOT-2020. As use-case, we focus on the significance of human-centred visual sensemaking —e.g., involving semantic representation and explainability, question-answering, commonsense interpolation— in safety-critical autonomous driving situations. The developed neurosymbolic framework is domain-independent, with the case of autonomous driving designed to serve as an exemplar for online visual sensemaking in diverse cognitive interaction settings in the backdrop of select human-centred AI technology design considerations.
  •  
46.
  • Casado, Lander, 1985, et al. (author)
  • ContikiSec: A Secure Network Layer for Wireless Sensor Networks under the Contiki Operating System
  • 2009
  • In: Proceedings of the 14th Nordic Conference on Secure IT Systems (NordSec 2009), Lecture Notes in Computer Science. - 1611-3349. - 9783642047657 ; 5838, s. 133 - 147
  • Conference paper (peer-reviewed)abstract
    • In this paper we introduce ContikiSec, a secure network layer forwireless sensor networks, designed for the Contiki Operating System. ContikiSechas a configurable design, providing three security modes starting fromconfidentiality and integrity, and expanding to confidentiality, authentication,and integrity. ContikiSec has been designed to balance low energy consumptionand security while conforming to a small memory footprint. Our design wasbased on performance evaluation of existing security primitives and is part ofthe contribution of this paper. Our evaluation was performed in the ModularSensor Board hardware platform for wireless sensor networks, running Contiki.Contiki is an open source, highly portable operating system for wireless sensornetworks (WSN) that is widely used in WSNs.
  •  
47.
  • Chatterjee, Bapi, 1982 (author)
  • Efficient Implementation of Concurrent Data Structures on Multi-core and Many-core Architectures
  • 2015
  • Licentiate thesis (other academic/artistic)abstract
    • Synchronization of concurrent threads is the central problem in order to design efficient concurrent data-structures. The compute systems widely available in market are increasingly becoming heterogeneous involving multi-core Central Processing Units (CPUs) and many-core Graphics Processing Units (GPUs). This thesis contributes to the research of efficient synchronization in concurrent data-structures in more than one way. It is divided into two parts. In the first part, a novel design of a Set Abstract Data Type (ADT) based on an efficient lock-free Binary Search Tree (BST) with improved amortized bounds of the time complexity of set operations - Add, Remove and Contains, is presented. In the second part, a comprehensive evaluation of concurrent Queue implementations on multi-core CPUs as well as many-core GPUs are presented. Efficient Lock-free BST -To the best of our knowledge, the lock-free BST presented in this thesis is the first to achieve an amortized complexity of O(H(n)+c) for all Set operations where H(n) is the height of a BST on n nodes and c is the contention measure. Also, the presented lock-free algorithm of BST comes with an improved disjoint-access-parallelism compared to the previously existing concurrent BST algorithms. This algorithm uses single-word compare-and-swap (CAS) primitives. The presented algorithm is linearizable. We implemented the algorithm in Java and it shows good scalability. Evaluation of concurrent data-structures - We have evaluated the performance of a number of concurrent FIFO Queue algorithms on multi-core CPUs and many-core GPUs. We studied the portability of existing design of concurrent Queues from CPUs to GPUs which are inherently designed for SIMD programs. We observed that in general concurrent queues offer them to efficient implementation on GPUs with faster cache memory and better performance support for atomic synchronization primitives such as CAS. To the best of our knowledge, this is the first attempt to evaluate a concurrent data-structure on GPUs.
  •  
48.
  • Ha, Phuong, 1976, et al. (author)
  • The Synchronization Power of Coalesced Memory Accesses
  • 2008
  • In: Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics). - Berlin, Heidelberg : Springer Berlin Heidelberg. - 1611-3349 .- 0302-9743. - 9783540877783 ; 5218, s. 320-334
  • Conference paper (peer-reviewed)abstract
    • Multicore processor architectures have established themselves as the new generation of processor architectures. As part of the one core to many cores evolution, memory access mechanisms have advanced rapidly. Several new memory access mechanisms have been implemented in many modern commodity multicore processors. Memory access mechanisms, by devising how processing cores access the shared memory, directly influence the synchronization capabilities of the multicore processors. Therefore, it is crucial to investigate the synchronization power of these new memory access mechanisms.This paper investigates the synchronization power of coalesced memory accesses, a family of memory access mechanisms introduced in recent large multicore architectures like the CUDA graphics processors. We first design three memory access models to capture the fundamental features of the new memory access mechanisms. Subsequently, we prove the exact synchronization power of these models in terms of their consensus numbers. These tight results show that the coalesced memory access mechanisms can facilitate strong synchronization between the threads of multicore processors, without the need of synchronization primitives other than reads and writes. In the case of the contemporary CUDA processors, our results imply that the coalesced memory access mechanisms have consensus numbers up to sixteen.
  •  
49.
  • Ha, Phuong, 1976, et al. (author)
  • Wait-free Programming for General Purpose Computations on Graphics Processors
  • 2008
  • In: the Proceedings of the 22th International Parallel and Distributed Symposium (IPDPS 2008). - 1530-2075. - 9781424416936 ; , s. 1-12
  • Conference paper (peer-reviewed)abstract
    • The fact that graphics processors (GPUs) are today’s most powerful computational hardware for the dollar has motivated researchers to utilize the ubiquitous and powerful GPUs for general-purpose computing. Recent GPUs feature the single-program multiple-data (SPMD) multicore architecture instead of the single-instruction multiple-data (SIMD). However, unlike CPUs, GPUs devote their transistors mainly to data processing rather than data caching and flow control, and consequently most of the powerful GPUs with many cores do not support any synchronization mechanisms between their cores. This prevents GPUs from being deployed more widely for general-purpose computing. This paper aims at bridging the gap between the lack of synchronization mechanisms in recent GPU architectures and the need of synchronization mechanisms in parallel applications. Based on the intrinsic features of recent GPU architectures, we construct strong synchronization objects like wait-free and t-resilient read-modify-write objects for a general model of recent GPU architectures without strong hardware synchronization primitives like test-and-set and compare-and-swap. Accesses to the wait-free objects have time complexity O(N), whether N is the number of processes. Our result demonstrates that it is possible to construct wait-free synchronization mechanisms for GPUs without the need of strong synchronization primitives in hardware and that wait-free programming is possible for GPUs.
  •  
50.
  • Petersson, Jesper, 1974, et al. (author)
  • Off the record: The invisibility work of doctors in a patient-accessible electronic health record information service.
  • 2021
  • In: Sociology of health & illness. - : Wiley. - 1467-9566 .- 0141-9889. ; 43:5, s. 1270-1285
  • Journal article (peer-reviewed)abstract
    • In this article, we draw on Michael Lipsky's work on street-level bureaucrats and discretion to analyse a real case setting comprising an interview study of 30 Swedish doctors regarding their experiences of changes in clinical work following patients being given access to medical records information online. We introduce the notion of invisibility work to capture how doctors exercise discretion to preserve the invisibility of their work, in contrast to the well-established notion of invisible work, which denotes work made invisible by parties other than those performing it. We discuss three main forms of invisibility work in relation to records: omitting information, cryptic writing and parallel note writing. We argue that invisibility work is a way for doctors to resolve professional tensions arising from the political decision to provide patients with online access to record information. Although invisibility work is understood by doctors as a solution to government-initiated visibility, we highlight how it can create difficulties for doctors concerning accountability towards patients, peers and authorities.
  •  
Skapa referenser, mejla, bekava och länka
  • Result 1-50 of 67535
Type of publication
conference paper (34690)
journal article (18836)
book chapter (3286)
reports (3229)
doctoral thesis (2625)
licentiate thesis (1653)
show more...
other publication (1098)
editorial proceedings (626)
research review (496)
book (481)
editorial collection (340)
patent (108)
artistic work (71)
review (57)
show less...
Type of content
peer-reviewed (52379)
other academic/artistic (14217)
pop. science, debate, etc. (907)
Author/Editor
Vyatkin, Valeriy (317)
Bosch, Jan, 1967 (306)
Voigt, Thiemo (298)
Staron, Miroslaw, 19 ... (234)
Sandkuhl, Kurt, 1963 ... (232)
Tsigas, Philippas, 1 ... (224)
show more...
Runeson, Per (220)
Vasilakos, Athanasio ... (218)
Wohlin, Claes (217)
Weyns, Danny (210)
Lennartson, Bengt, 1 ... (209)
Ekenberg, Love (199)
Kragic, Danica (192)
Svensson, Arne, 1955 (188)
Davidsson, Paul (187)
Karlgren, Jussi (184)
Borin, Lars, 1957 (181)
Gu, Irene Yu-Hua, 19 ... (175)
Lundberg, Lars (168)
Stenström, Per, 1957 (167)
Papatriantafilou, Ma ... (161)
Peng, Zebo, 1958- (160)
Boström, Henrik (159)
Johannesson, Paul (156)
Risch, Tore (155)
Taheri, Javid (154)
Zdravkovic, Jelena (152)
Kassler, Andreas, 19 ... (149)
Lundell, Björn (149)
Hansen, Preben (148)
Johnsson, Lennart (147)
Nivre, Joakim, 1962- (146)
Nilsson, Anders G. (145)
Fabian, Martin, 1960 (142)
Johansson, Ulf (140)
Lambrix, Patrick (139)
Lilienthal, Achim J. ... (139)
Andersson, Karl, 197 ... (138)
Doherty, Patrick, 19 ... (135)
Dobnik, Simon, 1977 (133)
Höst, Martin (132)
Beskow, Jonas (130)
De Raedt, Luc, 1964- (130)
Larsson, Staffan, 19 ... (130)
Jönsson, Bodil (129)
Balkenius, Christian (129)
Eles, Petru Ion, 195 ... (128)
Grahn, Håkan (127)
Torra, Vicenç (127)
Fischer-Hübner, Simo ... (127)
show less...
University
Royal Institute of Technology (11828)
Chalmers University of Technology (10348)
Uppsala University (6930)
Linköping University (5545)
University of Gothenburg (4733)
Stockholm University (4443)
show more...
RISE (4053)
Lund University (3660)
Blekinge Institute of Technology (3179)
Umeå University (2860)
Luleå University of Technology (2439)
Karlstad University (2235)
Örebro University (2210)
Linnaeus University (2132)
University of Skövde (1873)
Mälardalen University (1674)
Mid Sweden University (1458)
Jönköping University (1383)
Halmstad University (1215)
University of Borås (698)
Malmö University (644)
Södertörn University (472)
Karolinska Institutet (451)
Högskolan Dalarna (419)
University of Gävle (380)
Swedish University of Agricultural Sciences (359)
University West (293)
Kristianstad University College (203)
Stockholm School of Economics (123)
VTI - The Swedish National Road and Transport Research Institute (68)
Swedish National Defence College (51)
Red Cross University College (42)
University College of Arts, Crafts and Design (30)
Royal College of Music (24)
Swedish Museum of Natural History (14)
The Institute for Language and Folklore (13)
IVL Swedish Environmental Research Institute (9)
Stockholm University of the Arts (7)
Swedish Environmental Protection Agency (3)
Swedish National Heritage Board (3)
Sophiahemmet University College (3)
The Nordic Africa Institute (2)
The Swedish School of Sport and Health Sciences (2)
Marie Cederschiöld högskola (2)
Swedish Agency for Marine and Water Management (1)
University College Stockholm (1)
The Royal Institute of Art (1)
show less...
Language
English (65471)
Swedish (1899)
German (53)
French (19)
Portuguese (18)
Japanese (15)
show more...
Spanish (13)
Norwegian (8)
Undefined language (8)
Russian (5)
Finnish (5)
Italian (4)
Danish (3)
Greek, Modern (3)
Estonian (2)
Hungarian (2)
Persian (2)
Mongolian (2)
Chinese (1)
Turkish (1)
Slovenian (1)
show less...
Research subject (UKÄ/SCB)
Natural sciences (67507)
Engineering and Technology (8466)
Social Sciences (4605)
Humanities (2380)
Medical and Health Sciences (1546)
Agricultural Sciences (216)

Year

Kungliga biblioteket hanterar dina personuppgifter i enlighet med EU:s dataskyddsförordning (2018), GDPR. Läs mer om hur det funkar här.
Så här hanterar KB dina uppgifter vid användning av denna tjänst.

 
pil uppåt Close

Copy and save the link in order to return to this view