hig.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • harvard-cite-them-right
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • sv-SE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • de-DE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Continuous residual reinforcement learning for traffic signal control optimization
University of Gävle, Faculty of Engineering and Sustainable Development, Department of Industrial Development, IT and Land Management, Computer science.
University of Gävle, Faculty of Engineering and Sustainable Development, Department of Industrial Development, IT and Land Management, Computer science. Division of Visual Information and Interaction, Department of Information Technology, Uppsala University, Uppsala, Sweden.
Institute of Artificial Intelligence and Cognitive Engineering, University of Groningen, Groningen, the Netherlands.
2018 (English)In: Canadian journal of civil engineering (Print), ISSN 0315-1468, E-ISSN 1208-6029, Vol. 45, no 8, p. 690-702Article in journal (Refereed) Published
Abstract [en]

Traffic signal control can be naturally regarded as a reinforcement learning problem. Unfortunately, it is one of the most difficult classes of reinforcement learning problems owing to its large state space. A straightforward approach to address this challenge is to control traffic signals based on continuous reinforcement learning. Although they have been successful in traffic signal control, they may become unstable and fail to converge to near-optimal solutions. We develop adaptive traffic signal controllers based on continuous residual reinforcement learning (CRL-TSC) that is more stable. The effect of three feature functions is empirically investigated in a microscopic traffic simulation. Furthermore, the effects of departing streets, more actions, and the use of the spatial distribution of the vehicles on the performance of CRL-TSCs are assessed. The results show that the best setup of the CRL-TSC leads to saving average travel time by 15% in comparison to an optimized fixed-time controller.

Place, publisher, year, edition, pages
NRC Research Press , 2018. Vol. 45, no 8, p. 690-702
Keywords [en]
continuous state reinforcement learning, adaptive traffic signal control, microscopic traffic simulation
National Category
Other Engineering and Technologies
Identifiers
URN: urn:nbn:se:hig:diva-27845DOI: 10.1139/cjce-2017-0408ISI: 000440632100009Scopus ID: 2-s2.0-85051122432OAI: oai:DiVA.org:hig-27845DiVA, id: diva2:1245454
Available from: 2018-09-05 Created: 2018-09-05 Last updated: 2018-09-05Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopusFinal author's version

Authority records BETA

Aslani, MohammadSeipel, Stefan

Search in DiVA

By author/editor
Aslani, MohammadSeipel, Stefan
By organisation
Computer science
In the same journal
Canadian journal of civil engineering (Print)
Other Engineering and Technologies

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 41 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • harvard-cite-them-right
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • sv-SE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • de-DE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf