SwePub
Sök i SwePub databas

  Utökad sökning

Träfflista för sökning "WFRF:(Källström Johan 1976 ) "

Sökning: WFRF:(Källström Johan 1976 )

  • Resultat 1-16 av 16
Sortera/gruppera träfflistan
   
NumreringReferensOmslagsbildHitta
1.
  • Domova, Veronika, 1987-, et al. (författare)
  • Improving Usability of Search and Rescue Decision Support Systems : WARA-PS Case Study
  • 2020
  • Ingår i: Proceedings 25th IEEE International Conference on Emerging Technologies and Factory Automation, ETFA 2020. - Vienna, Austria : Institute of Electrical and Electronics Engineers (IEEE). - 9781728189574 - 9781728189567 ; 2020-September, s. 1251-1254
  • Konferensbidrag (refereegranskat)abstract
    • Novel autonomous search and rescue systems, although powerful, still require a human decision-maker involvement. In this project, we focus on the human aspect of one such novel autonomous SAR system. Relying on the knowledge gained in a field study, as well as through the literature, we introduced several extensions to the system that allowed us to achieve a more user-centered interface. In the evaluation session with a rescue service specialist, we received positive feedback and defined potential directions for future work.
  •  
2.
  • Hayes, Conor F., et al. (författare)
  • A Brief Guide to Multi-Objective Reinforcement Learning and Planning
  • 2023
  • Ingår i: Proceedings of the 22nd International Conference on Autonomous Agents and Multiagent Systems (AAMAS). - 9781450394321 ; , s. 1988-1990
  • Konferensbidrag (refereegranskat)abstract
    • Real-world sequential decision-making tasks are usually complex, and require trade-offs between multiple–often conflicting–objectives. However, the majority of research in reinforcement learning (RL) and decision-theoretic planning assumes a single objective, or that multiple objectives can be handled via a predefined weighted sum over the objectives. Such approaches may oversimplify the underlying problem, and produce suboptimal results. This extended abstract outlines the limitations of using a semi-blind iterative process to solve multi-objective decision making problems. Our extended paper [4], serves as a guide for the application of explicitly multi-objective methods to difficult problems.
  •  
3.
  • Hayes, Conor F., et al. (författare)
  • A practical guide to multi-objective reinforcement learning and planning
  • 2022
  • Ingår i: Autonomous Agents and Multi-Agent Systems. - New York, NY, United States : Springer. - 1387-2532 .- 1573-7454. ; 36:1
  • Tidskriftsartikel (refereegranskat)abstract
    • Real-world sequential decision-making tasks are generally complex, requiring trade-offs between multiple, often conflicting, objectives. Despite this, the majority of research in reinforcement learning and decision-theoretic planning either assumes only a single objective, or that multiple objectives can be adequately handled via a simple linear combination. Such approaches may oversimplify the underlying problem and hence produce suboptimal results. This paper serves as a guide to the application of multi-objective methods to difficult problems, and is aimed at researchers who are already familiar with single-objective reinforcement learning and planning methods who wish to adopt a multi-objective perspective on their research, as well as practitioners who encounter multi-objective decision problems in practice. It identifies the factors that may influence the nature of the desired solution, and illustrates by example how these influence the design of multi-objective decision-making systems for complex problems.
  •  
4.
  • Källström, Johan, 1976- (författare)
  • Adaptive Agent-Based Simulation for Individualized Training
  • 2020
  • Ingår i: Proceedings of the 19th International Conference on Autonomous Agents and Multiagent Systems (AAMAS). - : The International Foundation for Autonomous Agents and Multiagent Systems (IFAAMAS). - 9781450375184 ; , s. 2193-2195
  • Konferensbidrag (refereegranskat)abstract
    • Agent-based simulation can be used for efficient and effective training of human operators and decision-makers. However, constructing realistic behavior models for the agents is challenging and time-consuming, especially for subject matter experts, who may not have expertise in artificial intelligence. In this work, we investigate how machine learning can be used to adapt simulation contents to the current needs of individual trainees. Our initial results demonstrate that multi-objective multi-agent reinforcement learning is a promising approach for creating agents with diverse and adaptive characteristics, which can stimulate humans in training.
  •  
5.
  • Källström, Johan, 1976-, et al. (författare)
  • Agent Coordination in Air Combat Simulation using Multi-Agent Deep Reinforcement Learning
  • 2020
  • Ingår i: 2020 IEEE International Conference on Systems, Man, and Cybernetics (SMC). - : IEEE conference proceedings. - 9781728185279 - 9781728185262 ; , s. 2157-2164
  • Konferensbidrag (refereegranskat)abstract
    • Simulation-based training has the potential to significantly improve training value in the air combat domain. However, synthetic opponents must be controlled by high-quality behavior models, in order to exhibit human-like behavior. Building such models by hand is recognized as a very challenging task. In this work, we study how multi-agent deep reinforcement learning can be used to construct behavior models for synthetic pilots in air combat simulation. We empirically evaluate a number of approaches in two air combat scenarios, and demonstrate that curriculum learning is a promising approach for handling the high-dimensional state space of the air combat domain, and that multi-objective learning can produce synthetic agents with diverse characteristics, which can stimulate human pilots in training.
  •  
6.
  • Källström, Johan, 1976-, et al. (författare)
  • Design of Simulation-Based Pilot Training Systems using Machine Learning Agents
  • 2021
  • Ingår i: Proceedings of the 32nd Congress of the International Council of Aeronautical Sciences (ICAS). - Bonn : The International Council of the Aeronautical Sciences. - 9783932182914
  • Konferensbidrag (refereegranskat)abstract
    • The high operational cost of aircraft, limited availability of air space, and strict safety regulations make training of fighter pilots increasingly challenging. By integrating Live, Virtual, and Constructive simulation resources, efficiency and effectiveness can be improved. In particular, if constructive simulations, which provide synthetic agents operating synthetic vehicles, were used to a higher degree, complex training scenarios could be realized at low cost, the need for support personnel could be reduced, and training availability could be improved. In this work, inspired by the recent improvements of techniques for artificial intelligence, we take a user perspective and investigate how intelligent, learning agents could help build future training systems. Through a domain analysis, a user study, and practical experiments, we identify important agent capabilities and characteristics, and then discuss design approaches and solution concepts for training systems to utilize learning agents for improved training value.
  •  
7.
  • Källström, Jonatan, 1976, et al. (författare)
  • Effects of allopurinol on 6-mercaptopurine metabolism in unselected patients with pediatric acute lymphoblastic leukemia: a prospective phase II study.
  • 2024
  • Ingår i: Haematologica. - 1592-8721.
  • Tidskriftsartikel (refereegranskat)abstract
    • Allopurinol can be used in maintenance therapy (MT) for pediatric acute lymphoblastic leukemia (ALL) to mitigate hepatic toxicity in patients with skewed 6- mercaptopurine metabolism. Allopurinol increases the erythrocyte levels of thioguanine nucleotides (e-TGN), which is the proposed main mediator of the antileukemic effect and decreases methyl mercaptopurine (e-MeMP) levels, associated with hepatotoxicity. We investigated the effects of allopurinol in thiopurine methyltransferase (TPMT) wild-type patients without previous clinical signs of skewed 6MP metabolism. Fifty-one patients from Sweden and Finland were enrolled in this prospective beforeafter trial during ALL MT. Mean e-TGN increased from 280 nmol/mmol Hb after 12 weeks of standard MT to 440 after 12 weeks of MT with addition of allopurinol 50 mg/m2 (p.
  •  
8.
  • Källström, Johan, 1976-, et al. (författare)
  • Learning Agents for Improved Efficiency and Effectiveness in Simulation-Based Training
  • 2020
  • Ingår i: Poceedings of the 32nd annual workshop of the Swedish Artificial Intelligence Society (SAIS). ; , s. 1-2
  • Konferensbidrag (övrigt vetenskapligt/konstnärligt)abstract
    • Team training in complex domains often requires a substantial amount of resources, e.g., instructors, role-players and vehicles. For this reason, it may be difficult to realize efficient and effective training scenarios in a real-world setting. Instead, intelligent agents can be used to construct synthetic, simulationbased training environments. However, building behavior models for such agents is challenging, especially for the end-users of the training systems, who typically do not have expertise in artificial intelligence. In this PhD project, we study how machine learning can be used to simplify the process of constructing agents for simulation-based training. As a case study we use a simulation-based air combat training system. By constructing smarter synthetic agents the dependency on human training providers can be reduced, and the availability as well as the quality of training can be improved.
  •  
9.
  • Källström, Johan, 1976-, et al. (författare)
  • Model-Based Actor-Critic for Multi-Objective Reinforcement Learning with Dynamic Utility Functions
  • 2023
  • Ingår i: Proceedings of the 22nd International Conference on Autonomous Agents and Multiagent Systems (AAMAS). - 9781450394321 ; , s. 2818-2820
  • Konferensbidrag (refereegranskat)abstract
    • Many real-world problems require a trade-off between multiple conflicting objectives. Decision-makers’ preferences over solutions to such problems are determined by their utility functions, which convert multi-objective values to scalars. In some settings, utility functions change over time, and the goal is to find methods that can efficiently adapt an agent’s policy to changes in utility. Previous work on learning with dynamic utility functions has focused on model-free methods, which often suffer from poor sample efficiency. In this work, we instead propose a model-based actor-critic, which explores with diverse utility functions through imagined rollouts within a learned world model between interactions with the real environment. An experimental evaluation on Minecart, a well-known benchmark for multi-objective reinforcement learning, shows that by learning a model of the environment the quality of the agent’s policy is improved compared to model-free algorithms.
  •  
10.
  • Källström, Johan, 1976-, et al. (författare)
  • Model-Based Multi-Objective Reinforcement Learning with Dynamic Utility Functions
  • 2023
  • Ingår i: Proceedings of the Adaptive and Learning Agents Workshop (ALA) at AAMAS 2023. ; , s. 1-9
  • Konferensbidrag (refereegranskat)abstract
    • Many real-world problems require a trade-off between multiple conflicting objectives. Decision-makers’ preferences over solutions to such problems are determined by their utility functions, which convert multi-objective values to scalars. In some settings, utility functions change over time, and the goal is to find methods that can efficiently adapt an agent’s policy to changes in utility. Previous work on learning with dynamic utility functions has focused on model-free methods, which often suffer from poor sample efficiency. In this work, we instead propose a model-based actor-critic, which explores with diverse utility functions through imagined rollouts within a learned world model between interactions with the real environment. An experimental evaluation shows that by learning a model of the environment the performance of the agent can be improved compared to model-free algorithms.
  •  
11.
  • Källström, Johan, 1976-, et al. (författare)
  • Multi-Agent Multi-Objective Deep Reinforcement Learning for Efficient and Effective Pilot Training
  • 2019
  • Ingår i: Proceedings of the 10th Aerospace Technology Congress (FT). - : Linköping University Electronic Press. - 9789175190068 ; , s. 101-111
  • Konferensbidrag (refereegranskat)abstract
    • The tactical systems and operational environment of modern fighter aircraft are becoming increasingly complex. Creating a realistic and relevant environment for pilot training using only live aircraft is difficult, impractical and highly expensive. The Live, Virtual and Constructive (LVC) simulation paradigm aims to address this challenge. LVC simulation means linking real aircraft, ground-based systems and soldiers (Live), manned simulators (Virtual) and computer controlled synthetic entities (Constructive). Constructive simulation enables realization of complex scenarios with a large number of autonomous friendly, hostile and neutral entities, which interact with each other as well as manned simulators and real systems. This reduces the need for personnel to act as role-players through operation of e.g. live or virtual aircraft, thus lowering the cost of training. Constructive simulation also makes it possible to improve the availability of training by embedding simulation capabilities in live aircraft, making it possible to train anywhere, anytime. In this paper we discuss how machine learning techniques can be used to automate the process of constructing advanced, adaptive behavior models for constructive simulations, to improve the autonomy of future training systems. We conduct a number of initial experiments, and show that reinforcement learning, in particular multi-agent and multi-objective deep reinforcement learning, allows synthetic pilots to learn to cooperate and prioritize among conflicting objectives in air combat scenarios. Though the results are promising, we conclude that further algorithm development is necessary to fully master the complex domain of air combat simulation.
  •  
12.
  • Källström, Johan, 1976-, et al. (författare)
  • Reinforcement Learning for Computer Generated Forces using Open-Source Software
  • 2019
  • Ingår i: Proceedings of the 2019 Interservice/Industry Training, Simulation, and Education Conference (IITSEC). ; , s. 1-11
  • Konferensbidrag (refereegranskat)abstract
    • The creation of behavior models for computer generated forces (CGF) is a challenging and time-consuming task, which often requires expertise in programming of complex artificial intelligence algorithms. This makes it difficult for a subject matter expert with knowledge about the application domain and the training goals to build relevant scenarios and keep the training system in pace with training needs. In recent years, machine learning has shown promise as a method for building advanced decision-making models for synthetic agents. Such agents have been able to beat human champions in complex games such as poker, Go and StarCraft. There is reason to believe that similar achievements are possible in the domain of military simulation. However, in order to efficiently apply these techniques, it is important to have access to the right tools, as well as knowledge about the capabilities and limitations of algorithms.   This paper discusses efficient applications of deep reinforcement learning, a machine learning technique that allows synthetic agents to learn how to achieve their goals by interacting with their environment. We begin by giving an overview of available open-source frameworks for deep reinforcement learning, as well as libraries with reference implementations of state-of-the art algorithms. We then present an example of how these resources were used to build a reinforcement learning environment for a CGF software intended to support training of fighter pilots. Finally, based on our exploratory experiments in the presented environment, we discuss opportunities and challenges related to the application of reinforcement learning techniques in the domain of air combat training systems, with the aim to efficiently construct high quality behavior models for computer generated forces.
  •  
13.
  • Källström, Johan, 1976- (författare)
  • Reinforcement Learning for Improved Utility of Simulation-Based Training
  • 2023
  • Doktorsavhandling (övrigt vetenskapligt/konstnärligt)abstract
    • Team training in complex domains often requires a substantial number of resources, e.g. vehicles, machines, and role-players. For this reason, it may be difficult to realise efficient and effective training scenarios in a real-world setting. Instead, part of the training can be conducted in synthetic, computer-generated environments. In these environments trainees can operate simulators instead of real vehicles, while synthetic actors can replace human role-players to increase the complexity of the simulated scenario at low operating cost. However, constructing behaviour models for synthetic actors is challenging, especially for the end users, who typically do not have expertise in artificial intelligence. In this dissertation, we study how machine learning can be used to simplify the construction of intelligent agents for simulation-based training. A simulation-based air combat training system is used as case study. The contributions of the dissertation are divided into two parts. The first part aims at improving the understanding of reinforcement learning in the domain of simulation-based training. First, a user-study is conducted to identify important capabilities and characteristics of learning agents that are intended to support training of fighter pilots. It is identified that one of the most important capabilities of learning agents in the context of simulation-based training is that their behaviour can be adapted to different phases of training, as well as to the training needs of individual human trainees. Second, methods for learning how to coordinate with other agents are studied in simplified training scenarios, to investigate how the design of the agent’s observation space, action space, and reward signal affects the performance of learning. It is identified that temporal abstractions and hierarchical reinforcement learning can improve the efficiency of learning, while also providing support for modelling of doctrinal behaviour. In more complex settings, curriculum learning and related methods are expected to help find novel tactics even when sparse, abstract reward signals are used. Third, based on the results from the user study and the practical experiments, a system concept for a user-adaptive training system is developed to support further research. The second part of the contributions focuses on methods for utility-based multi-objective reinforcement learning, which incorporates knowledge of the user’s utility function in the search for policies that balance multiple conflicting objectives. Two new agents for multi-objective reinforcement learning are proposed: the Tunable Actor (T-Actor) and the Multi-Objective Dreamer (MO-Dreamer). T-Actor provides decision support to instructors by learning a set of Pareto optimal policies, represented by a single neural network conditioned on objective preferences. This enables tuning of the agent’s behaviour to fit trainees’ current training needs. Experimental evaluations in gridworlds and in the target system show that T-Actor reduces the number of training steps required for learning. MO-Dreamer adapts online to changes in users’ utility, e.g. changes in training needs. It does so by learning a model of the environment, which it can use for anticipatory rollouts with a diverse set of utility functions to explore which policy to follow to optimise the return for a given set of objective preferences. An experimental evaluation shows that MO-Dreamer outperforms prior model-free approaches in terms of experienced regret, for frequent as well as sparse changes in utility. Overall, the research conducted in this dissertation contributes to improved knowledge about how to apply machine learning methods to construction of simulation-based training environments. While our focus was on air combat training, the results are general enough to be applicable in other domains. 
  •  
14.
  • Källström, Johan, 1976-, et al. (författare)
  • Tunable Dynamics in Agent-Based Simulation using Multi-Objective Reinforcement Learning
  • 2019
  • Ingår i: Proceedings of the 2019 Adaptive and Learning Agents Workshop (ALA), 2019. ; , s. 1-7
  • Konferensbidrag (refereegranskat)abstract
    • Agent-based simulation is a powerful tool for studying complex systems of interacting agents. To achieve good results, the behavior models used for the agents must be of high quality. Traditionally these models have been handcrafted by domain experts. This is a difficult, expensive and time consuming process. In contrast, reinforcement learning allows agents to learn how to achieve their goals by interacting with the environment. However, after training the behavior of such agents is often static, i.e. it can no longer be affected by a human. This makes it difficult to adapt agent behavior to specific user needs, which may vary among different runs of the simulation. In this paper we address this problem by studying how multi-objective reinforcement learning can be used as a framework for building tunable agents, whose characteristics can be adjusted at runtime to promote adaptiveness and diversity in agent-based simulation. We propose an agent architecture that allows us to adapt popular deep reinforcement learning algorithms to multi-objective environments. We empirically show that our method allows us to train tunable agents that can approximate the policies of multiple species of agents.
  •  
15.
  • Vamplew, Peter, et al. (författare)
  • Scalar reward is not enough : a response to Silver, Singh, Precup and Sutton (2021)
  • 2022
  • Ingår i: Autonomous Agents and Multi-Agent Systems. - : Springer. - 1387-2532 .- 1573-7454. ; 36:2
  • Tidskriftsartikel (refereegranskat)abstract
    • The recent paper “Reward is Enough” by Silver, Singh, Precup and Sutton posits that the concept of reward maximisation is sufficient to underpin all intelligence, both natural and artificial, and provides a suitable basis for the creation of artificial general intelligence. We contest the underlying assumption of Silver et al. that such reward can be scalar-valued. In this paper we explain why scalar rewards are insufficient to account for some aspects of both biological and computational intelligence, and argue in favour of explicitly multi-objective models of reward maximisation. Furthermore, we contend that even if scalar reward functions can trigger intelligent behaviour in specific cases, this type of reward is insufficient for the development of human-aligned artificial general intelligence due to unacceptable risks of unsafe or unethical behaviour.
  •  
16.
  • Vamplew, Peter, et al. (författare)
  • Scalar Reward is Not Enough
  • 2023
  • Ingår i: Proceedings of the 22nd International Conference on Autonomous Agents and Multiagent Systems (AAMAS). - 9781450394321 ; , s. 839-841
  • Konferensbidrag (refereegranskat)abstract
    • Silver et al.[14] posit that scalar reward maximisation is sufficient to underpin all intelligence and provides a suitable basis for artificial general intelligence (AGI). This extended abstract summarises the counter-argument from our JAAMAS paper [19].
  •  
Skapa referenser, mejla, bekava och länka
  • Resultat 1-16 av 16

Kungliga biblioteket hanterar dina personuppgifter i enlighet med EU:s dataskyddsförordning (2018), GDPR. Läs mer om hur det funkar här.
Så här hanterar KB dina uppgifter vid användning av denna tjänst.

 
pil uppåt Stäng

Kopiera och spara länken för att återkomma till aktuell vy