Please use this identifier to cite or link to this item: https://www.um.edu.mt/library/oar/handle/123456789/117943
Full metadata record
DC FieldValueLanguage
dc.contributor.authorBajada, Josef-
dc.contributor.authorGrech, Joseph-
dc.contributor.authorBajada, Therese-
dc.date.accessioned2024-01-30T06:30:29Z-
dc.date.available2024-01-30T06:30:29Z-
dc.date.issued2023-
dc.identifier.citationBajada, J., Grech, J., & Bajada, T. (2023). Deep Reinforcement Learning of Autonomous Control Actions to Improve Bus-Service Regularity. In S. Nowaczyk, P. Biecek, N.C. Chung, M. Vallati, P. Skruch, J. Jaworek-Korjakowska,…V. Dimitrova (Eds.), European Conference on Artificial Intelligence (pp. 138-155). Cham: Springer Nature Switzerland.en_GB
dc.identifier.urihttps://www.um.edu.mt/library/oar/handle/123456789/117943-
dc.description.abstractBus Bunching is caused by irregularities in demand across the bus route, together with other factors such as traffic. The effect of this problem is that buses operating on the same route start to catch up with each other, severely impacting the regularity and the quality of the service. Control actions such as Bus Holding and Stop Skipping can be used to regulate the service and adjust the headway between two buses. Traditionally, this phenomenon is mitigated either reactively online through simple rule-based control, or preemptively through analytical scheduling solutions, such as mathematical optimization. Over time, both approaches degrade to an irregular service. In this work, we investigate the use of Deep Reinforcement Learning algorithms to train a policy that determines which actions should take place at specific control points to regularise the bus service. While prior studies are typically restricted to one control action, we consider both Bus Holding and Stop Skipping. We replicate benchmarks found in the latest literature, and also introduce traffic to increase the realism of the simulation. Furthermore, we also consider scenarios where the service is already unstable and buses are already bunched together, a first of this kind of study. We compare the performance of the RL-based policies with a no-control policy and a rule-based policy. The learnt policies not only keep a significantly lower headway variance and mean waiting time, but also recover from unstable scenarios and restore service regularity.en_GB
dc.language.isoenen_GB
dc.publisherSpringer Nature Switzerlanden_GB
dc.rightsinfo:eu-repo/semantics/restrictedAccessen_GB
dc.subjectBusesen_GB
dc.subjectReinforcement learningen_GB
dc.subjectAutonomous distributed systemsen_GB
dc.subjectBuses -- Service lifeen_GB
dc.titleDeep reinforcement learning of autonomous control actions to improve bus-service regularityen_GB
dc.title.alternativeEuropean conference on artificial intelligenceen_GB
dc.typebookParten_GB
dc.rights.holderThe copyright of this work belongs to the author(s)/publisher. The rights of this work are as defined by the appropriate Copyright Legislation or as modified by any successive legislation. Users may access this work and can make use of the information contained in accordance with the Copyright Legislation provided that the author must be properly acknowledged. Further distribution or reproduction in any format is prohibited without the prior permission of the copyright holderen_GB
dc.description.reviewedpeer-revieweden_GB
dc.identifier.doi10.1007/978-3-031-50396-2_8-
Appears in Collections:Scholarly Works - FacICTAI

Files in This Item:
File Description SizeFormat 
Deep_reinforcement_learning_of_autonomous_control_actions_to_improve_bus_service_regularity.pdf
  Restricted Access
1.43 MBAdobe PDFView/Open Request a copy


Items in OAR@UM are protected by copyright, with all rights reserved, unless otherwise indicated.