The increasing complexity of automated driving functions (ADFs) necessitates efficient data collection strategies due to the significant resources required for real-world data acquisition. Traditional methods rely on real-world vehicles, which implies costly and time-consuming collection. On the other hand, modern synthetic data generation methods do not explicitly account for the variation in complexity between the scenario classes and the data balance in relation to the training performance of the model. This paper presents a novel approach for generating synthetic automotive driving scenario datasets using a Reinforcement Learning (RL) -based framework. The system is designed to optimize data distribution dynamically at runtime, thus enhancing the stability of the training procedure. The proposed architecture iteratively guides the scenario generation process by identifying the most difficult scenarios in terms of model training F1 score performance. Preliminary experimental results demonstrate that this approach effectively stabilizes the training procedure for different scenario classes by introducing a controlled data imbalance. The study provides promising insights into the potential of RL-driven data generation for learning modules, also highlighting areas for future research to further enhance the capabilities of the system.

RL-Based Generation of a Synthetic Automotive Driving Scenario Dataset

Berta, R.;Bellotti, F.;Cossu, M.;Forneris, L.;Lazzaroni, L.;Pighetti, A.
2025-01-01

Abstract

The increasing complexity of automated driving functions (ADFs) necessitates efficient data collection strategies due to the significant resources required for real-world data acquisition. Traditional methods rely on real-world vehicles, which implies costly and time-consuming collection. On the other hand, modern synthetic data generation methods do not explicitly account for the variation in complexity between the scenario classes and the data balance in relation to the training performance of the model. This paper presents a novel approach for generating synthetic automotive driving scenario datasets using a Reinforcement Learning (RL) -based framework. The system is designed to optimize data distribution dynamically at runtime, thus enhancing the stability of the training procedure. The proposed architecture iteratively guides the scenario generation process by identifying the most difficult scenarios in terms of model training F1 score performance. Preliminary experimental results demonstrate that this approach effectively stabilizes the training procedure for different scenario classes by introducing a controlled data imbalance. The study provides promising insights into the potential of RL-driven data generation for learning modules, also highlighting areas for future research to further enhance the capabilities of the system.
2025
9783031840999
File in questo prodotto:
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11567/1273961
 Attenzione

Attenzione! I dati visualizzati non sono stati sottoposti a validazione da parte dell'ateneo

Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
  • ???jsp.display-item.citation.isi??? ND
social impact