Learning to Navigate Over Stochastic Transport Networks Using Multi-Armed Bandits: A Contextual Approach for Efficient Online Learning in Road Network Graphs with Multi-Armed Bandits to Minimize Long- Term Travel Time

dc.contributor.authorNilsson, Hannes
dc.contributor.authorJohansson, Rikard
dc.contributor.departmentChalmers tekniska högskola / Institutionen för data och informationstekniksv
dc.contributor.departmentChalmers University of Technology / Department of Computer Science and Engineeringen
dc.contributor.examinerDubhashi, Devdatt
dc.contributor.supervisorHaghir Chehreghani, Morteza
dc.date.accessioned2024-10-17T13:32:23Z
dc.date.available2024-10-17T13:32:23Z
dc.date.issued2024
dc.date.submitted
dc.description.abstractAs part of the ongoing process of phasing out fossil fuel vehicles, attempts have been made to extend the effective range and adoption rate of electric vehicles through navigation systems focused on energy consumption. One way to approach this problem is by viewing route selections as a multi-armed bandit problem. This allows the system to adapt and recommend better routes over time, to minimize energy consumption. For navigation systems to be useful in practice, guiding vehicles from one point to another in minimal time is crucial. Therefore, this project examines the effectiveness of multi-armed bandit algorithms for time-efficient navigation in complex real-world environments, without initial information. For this purpose, we adapt a previously studied online learning framework developed for energy efficiency, and extract road segment travel time distributions from the traffic simulation software SUMO. The framework is applied to the Luxembourg road network and our results demonstrate that contextual multi-armed bandits using tree ensembles are highly effective. More specifically, we show that TEUCB and TETS, which we implement using both XGBoost and random forest, outperform state-of-the-art contextual multi-armed bandits based on neural networks and linear models. Further, by additional comparison of TEUCB and TETS to other bandit algorithms based on tree models, we identify at least two properties to explain their high level of performance. First, tree ensemble methods appear to offer relatively accurate travel time predictions from the contextual information available in this problem. Second, the ability to generalize over different arms and infer the travel time on one road segment from observations gathered on other ones, based on similarities, seems highly advantageous for this problem.
dc.identifier.coursecodeDATX60
dc.identifier.urihttp://hdl.handle.net/20.500.12380/308925
dc.language.isoeng
dc.setspec.uppsokTechnology
dc.subjectdecision-making under uncertainty
dc.subjectonline learning
dc.subjectmachine learning
dc.subjectreinforcement learning
dc.subjectmulti-armed bandits
dc.subjectnavigation
dc.subjectshortest path problem
dc.titleLearning to Navigate Over Stochastic Transport Networks Using Multi-Armed Bandits: A Contextual Approach for Efficient Online Learning in Road Network Graphs with Multi-Armed Bandits to Minimize Long- Term Travel Time
dc.type.degreeExamensarbete för masterexamensv
dc.type.degreeMaster's Thesisen
dc.type.uppsokH
local.programmeData science and AI (MPDSC), MSc
Ladda ner
Original bundle
Visar 1 - 1 av 1
Hämtar...
Bild (thumbnail)
Namn:
CSE 24-27 HN RJ.pdf
Storlek:
13.07 MB
Format:
Adobe Portable Document Format
Beskrivning:
License bundle
Visar 1 - 1 av 1
Hämtar...
Bild (thumbnail)
Namn:
license.txt
Storlek:
2.35 KB
Format:
Item-specific license agreed upon to submission
Beskrivning: