SAFETYLIT WEEKLY UPDATE

We compile citations and summaries of about 400 new articles every week.
RSS Feed

HELP: Tutorials | FAQ
CONTACT US: Contact info

Search Results

Journal Article

Citation

Hart F, Okhrin O, Treiber M. Transp. Res. C Emerg. Technol. 2024; 159: e104486.

Copyright

(Copyright © 2024, Elsevier Publishing)

DOI

10.1016/j.trc.2024.104486

PMID

unavailable

Abstract

One of the biggest challenges in the development of learning-driven automated driving technologies remains the handling of uncommon, rare events that may have not been encountered in training. Especially when training a model with real driving data, unusual situations, such as emergency brakings, may be underrepresented, resulting in a model that lacks robustness in rare events. This study focuses on car-following based on reinforcement learning and demonstrates that existing approaches, trained with real driving data, fail to handle safety-critical situations. Since collecting data representing all kinds of possible car-following events, including safety-critical situations, is challenging, we propose a training environment that harnesses stochastic processes to generate diverse and challenging scenarios. Our experiments show that training with real data can lead to models that collide in safety-critical situations, whereas the proposed model exhibits excellent performance and remains accident-free, comfortable, and string-stable even in extreme scenarios, such as full-braking by the leading vehicle. Its robustness is demonstrated by simulating car-following scenarios for various reward function parametrizations and a diverse range of artificial and real leader data that were not included in training and were qualitatively different from the learning data. We further show that conventional reward designs can encourage aggressive behavior when approaching other vehicles. Additionally, we compared the proposed model with classical car-following models and found it to achieve equal or superior results.


Language: en

Keywords

Car-following model; Generalization capabilities; Reinforcement learning; String stability; Validation

NEW SEARCH


All SafetyLit records are available for automatic download to Zotero & Mendeley
Print