SwePub
Sök i LIBRIS databas

  Extended search

id:"swepub:oai:DiVA.org:kth-341608"
 

Search: id:"swepub:oai:DiVA.org:kth-341608" > A state-based inver...

  • 1 of 1
  • Previous record
  • Next record
  •    To hitlist
  • Song, YuchenKTH,Byggvetenskap,School of Transportation, Southeast University, Nanjing 211189, Jiangsu, China, Jiangsu (author)

A state-based inverse reinforcement learning approach to model activity-travel choices behavior with reward function recovery

  • Article/chapterEnglish2024

Publisher, publication year, extent ...

  • Elsevier BV,2024
  • printrdacarrier

Numbers

  • LIBRIS-ID:oai:DiVA.org:kth-341608
  • https://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-341608URI
  • https://doi.org/10.1016/j.trc.2023.104454DOI

Supplementary language notes

  • Language:English
  • Summary in:English

Part of subdatabase

Classification

  • Subject category:ref swepub-contenttype
  • Subject category:art swepub-publicationtype

Notes

  • QC 20231227
  • Behaviorally oriented activity-travel choices (ATC) modeling is a principal part of travel demand analysis. Traditional econometric and rule-based methods require explicit model structures and complex domain knowledge. While several recent studies used machine learning models, especially adversarial inverse reinforcement learning (IRL) models, to learn potential ATC patterns with less expert-designed settings, they lack a clear representation of rational ATC behavior. In this study, we propose a data-driven IRL framework based on the maximum causal approach to minimize f-divergences between expert and agent state marginal distributions, which provides a more sample-efficient measurement. In addition, we specify a separate state-only reward function and derive an analytical gradient of the f-divergence objective with respect to reward parameters to ensure good convergences. The method can recover a stationary reward function, which assures the agent to get close to the expert behavior when training from scratch. We validate the proposed model using cellular signaling data from Chongqing, China by comparing with baseline models (behavior cloning, policy-based, and reward-based models) in aspects of policy performance comparison, reward recovery, and reward transfer tasks. The experiment results indicate that the proposed model outperforms existing methods and is relatively less sensitive to the number of expert demonstrations. Qualitative analyses are provided on the fundamental ATC preferences on different features given the reward function recovered from the observed mobility trajectories, and on the learning behaviors under different choices of f-divergence.

Subject headings and genre

Added entries (persons, corporate bodies, meetings, titles ...)

  • Li, DaweiSchool of Transportation, Southeast University, Nanjing 211189, Jiangsu, China, Jiangsu; Jiangsu Key Laboratory of Urban ITS, Southeast University, Nanjing 211189, Jiangsu, China, Jiangsu; Jiangsu Province Collaborative Innovation Center of Modern Urban Traffic Technologies, Nanjing 211189, Jiangsu, China, Jiangsu (author)
  • Ma, ZhenliangKTH,Transportplanering(Swepub:kth)u12zc8sm (author)
  • Liu, DongjieSchool of Transportation, Southeast University, Nanjing 211189, Jiangsu, China, Jiangsu (author)
  • Zhang, TongSchool of Transportation, Southeast University, Nanjing 211189, Jiangsu, China, Jiangsu (author)
  • KTHByggvetenskap (creator_code:org_t)

Related titles

  • In:Transportation Research Part C: Elsevier BV1580968-090X1879-2359

Internet link

Find in a library

To the university's database

  • 1 of 1
  • Previous record
  • Next record
  •    To hitlist

Find more in SwePub

By the author/editor
Song, Yuchen
Li, Dawei
Ma, Zhenliang
Liu, Dongjie
Zhang, Tong
About the subject
ENGINEERING AND TECHNOLOGY
ENGINEERING AND ...
and Civil Engineerin ...
and Transport System ...
Articles in the publication
Transportation R ...
By the university
Royal Institute of Technology

Search outside SwePub

Kungliga biblioteket hanterar dina personuppgifter i enlighet med EU:s dataskyddsförordning (2018), GDPR. Läs mer om hur det funkar här.
Så här hanterar KB dina uppgifter vid användning av denna tjänst.

 
pil uppåt Close

Copy and save the link in order to return to this view