driving style
LLM-based Human-like Traffic Simulation for Self-driving Tests
Li, Wendi, Wu, Hao, Gao, Han, Mao, Bing, Xu, Fengyuan, Zhong, Sheng
Ensuring realistic traffic dynamics is a prerequisite for simulation platforms to evaluate the reliability of self-driving systems before deployment in the real world. Because most road users are human drivers, reproducing their diverse behaviors within simulators is vital. Existing solutions, however, typically rely on either handcrafted heuristics or narrow data-driven models, which capture only fragments of real driving behaviors and offer limited driving style diversity and interpretability. To address this gap, we introduce HDSim, an HD traffic generation framework that combines cognitive theory with large language model (LLM) assistance to produce scalable and realistic traffic scenarios within simulation platforms. The framework advances the state of the art in two ways: (i) it introduces a hierarchical driver model that represents diverse driving style traits, and (ii) it develops a Perception-Mediated Behavior Influence strategy, where LLMs guide perception to indirectly shape driver actions. Experiments reveal that embedding HDSim into simulation improves detection of safety-critical failures in self-driving systems by up to 68% and yields realism-consistent accident interpretability.
An Evolutionary Game-Theoretic Merging Decision-Making Considering Social Acceptance for Autonomous Driving
Liu, Haolin, Guo, Zijun, Chen, Yanbo, Chen, Jiaqi, Yu, Huilong, Xi, Junqiang
--Highway on-ramp merging is of great challenge for autonomous vehicles (A Vs), since they have to proactively interact with surrounding vehicles to enter the main road safely within limited time. However, existing decision-making algorithms fail to adequately address dynamic complexities and social acceptance of A Vs, leading to suboptimal or unsafe merging decisions. T o address this, we propose an evolutionary game-theoretic (EGT) merging decision-making framework, grounded in the bounded rationality of human drivers, which dynamically balances the benefits of both A Vs and main-road vehicles (MVs). We formulate the cut-in decision-making process as an EGT problem with a multi-objective payoff function that reflects human-like driving preferences. By solving the replicator dynamic equation for the evolutionarily stable strategy (ESS), the optimal cut-in timing is derived, balancing efficiency, comfort, and safety for both A Vs and MVs. A real-time driving style estimation algorithm is proposed to adjust the game payoff function online by observing the immediate reactions of MVs. Empirical results demonstrate that we improve the efficiency, comfort and safety of both A Vs and MVs compared with existing game-theoretic and traditional planning approaches across multi-object metrics.
Multi-Objective Reinforcement Learning for Adaptable Personalized Autonomous Driving
Surmann, Hendrik, de Heuvel, Jorge, Bennewitz, Maren
Human drivers exhibit individual preferences regarding driving style. Adapting autonomous vehicles to these preferences is essential for user trust and satisfaction. However, existing end-to-end driving approaches often rely on predefined driving styles or require continuous user feedback for adaptation, limiting their ability to support dynamic, context-dependent preferences. We propose a novel approach using multi-objective reinforcement learning (MORL) with preference-driven optimization for end-to-end autonomous driving that enables runtime adaptation to driving style preferences. Preferences are encoded as continuous weight vectors to modulate behavior along interpretable style objectives$\unicode{x2013}$including efficiency, comfort, speed, and aggressiveness$\unicode{x2013}$without requiring policy retraining. Our single-policy agent integrates vision-based perception in complex mixed-traffic scenarios and is evaluated in diverse urban environments using the CARLA simulator. Experimental results demonstrate that the agent dynamically adapts its driving behavior according to changing preferences while maintaining performance in terms of collision avoidance and route completion.
- Europe > Germany > North Rhine-Westphalia > Cologne Region > Bonn (0.04)
- Asia > Middle East > Republic of Türkiye > Karaman Province > Karaman (0.04)
- Transportation > Ground > Road (1.00)
- Automobiles & Trucks (1.00)
Distributional Soft Actor-Critic with Diffusion Policy
Liu, Tong, Wang, Yinuo, Song, Xujie, Zou, Wenjun, Chen, Liangfa, Wang, Likun, Shuai, Bin, Duan, Jingliang, Li, Shengbo Eben
Reinforcement learning has been proven to be highly effective in handling complex control tasks. Traditional methods typically use unimodal distributions, such as Gaussian distributions, to model the output of value distributions. However, unimodal distribution often and easily causes bias in value function estimation, leading to poor algorithm performance. This paper proposes a distributional reinforcement learning algorithm called DSAC-D (Distributed Soft Actor Critic with Diffusion Policy) to address the challenges of estimating bias in value functions and obtaining multimodal policy representations. A multimodal distributional policy iteration framework that can converge to the optimal policy was established by introducing policy entropy and value distribution function. A diffusion value network that can accurately characterize the distribution of multi peaks was constructed by generating a set of reward samples through reverse sampling using a diffusion model. Based on this, a distributional reinforcement learning algorithm with dual diffusion of the value network and the policy network was derived. MuJoCo testing tasks demonstrate that the proposed algorithm not only learns multimodal policy, but also achieves state-of-the-art (SOTA) performance in all 9 control tasks, with significant suppression of estimation bias and total average return improvement of over 10% compared to existing mainstream algorithms. The results of real vehicle testing show that DSAC-D can accurately characterize the multimodal distribution of different driving styles, and the diffusion policy network can characterize multimodal trajectories.
- Transportation > Ground > Road (0.91)
- Automobiles & Trucks (0.91)
CHARMS: A Cognitive Hierarchical Agent for Reasoning and Motion Stylization in Autonomous Driving
Wang, Jingyi, Chu, Duanfeng, Deng, Zejian, Lu, Liping, Wang, Jinxiang, Sun, Chen
To address the limitations of these approaches, we propose CHARMS, a decision-making model based on Level-k game theory [20]. The distinction between our approach and the existing methods is illustrated in Figure 1. CHARMS incorporates cognitive hierarchy theory to model diverse reasoning depths among agents, coupled with Social V alue Orientation (SVO) to capture individual preferences in driving behavior. We employ a two-stage training process consisting of reinforcement learning pretraining and supervised fine-tuning (SFT) to generate decision-making models that exhibit a wide range of human-like driving styles. Additionally, we integrate Poisson cognitive hierarchy (PCH) theory to enable CHARMS to generate more complex simulation scenarios with diverse vehicle styles. The main contributions of this paper can be summarized as follows. A behavior model integrating Level-k reasoning and SVO is proposed to simulate cognitively diverse driving styles. A two-stage training scheme (DRL + SFT) ensures both style distinctiveness and behavioral realism. A scenario generation method based on PCH theory is used to control driving style distributions, with the aim of creating more realistic and behaviorally diverse simulation scenarios.
- Asia > China > Hubei Province > Wuhan (0.05)
- Asia > China > Hong Kong (0.04)
- Asia > Middle East > Republic of Türkiye > Karaman Province > Karaman (0.04)
- Asia > China > Jiangsu Province > Nanjing (0.04)
- Transportation > Ground > Road (1.00)
- Automobiles & Trucks (1.00)
Exploiting Prior Knowledge in Preferential Learning of Individualized Autonomous Vehicle Driving Styles
Theiner, Lukas, Hirt, Sebastian, Steinke, Alexander, Findeisen, Rolf
Trajectory planning for automated vehicles commonly employs optimization over a moving horizon - Model Predictive Control - where the cost function critically influences the resulting driving style. However, finding a suitable cost function that results in a driving style preferred by passengers remains an ongoing challenge. We employ preferential Bayesian optimization to learn the cost function by iteratively querying a passenger's preference. Due to increasing dimensionality of the parameter space, preference learning approaches might struggle to find a suitable optimum with a limited number of experiments and expose the passenger to discomfort when exploring the parameter space. We address these challenges by incorporating prior knowledge into the preferential Bayesian optimization framework. Our method constructs a virtual decision maker from real-world human driving data to guide parameter sampling. In a simulation experiment, we achieve faster convergence of the prior-knowledge-informed learning procedure compared to existing preferential Bayesian optimization approaches and reduce the number of inadequate driving styles sampled.
- Europe > Germany > Hesse > Darmstadt Region > Darmstadt (0.05)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- Transportation > Ground > Road (1.00)
- Automobiles & Trucks (1.00)
Discrete Contrastive Learning for Diffusion Policies in Autonomous Driving
Kujanpää, Kalle, Baimukashev, Daulet, Munir, Farzeen, Azam, Shoaib, Kucner, Tomasz Piotr, Pajarinen, Joni, Kyrki, Ville
Learning to perform accurate and rich simulations of human driving behaviors from data for autonomous vehicle testing remains challenging due to human driving styles' high diversity and variance. We address this challenge by proposing a novel approach that leverages contrastive learning to extract a dictionary of driving styles from pre-existing human driving data. We discretize these styles with quantization, and the styles are used to learn a conditional diffusion policy for simulating human drivers. Our empirical evaluation confirms that the behaviors generated by our approach are both safer and more human-like than those of the machine-learning-based baseline methods. We believe this has the potential to enable higher realism and more effective techniques for evaluating and improving the performance of autonomous vehicles.
- North America > United States (0.28)
- Europe > Finland (0.05)
- Europe > Italy > Calabria > Catanzaro Province > Catanzaro (0.04)
Quantifying and Modeling Driving Styles in Trajectory Forecasting
Zheng, Laura, Araghi, Hamidreza Yaghoubi, Wu, Tony, Thalapanane, Sandeep, Zhou, Tianyi, Lin, Ming C.
Trajectory forecasting has become a popular deep learning task due to its relevance for scenario simulation for autonomous driving. Specifically, trajectory forecasting predicts the trajectory of a short-horizon future for specific human drivers in a particular traffic scenario. Robust and accurate future predictions can enable autonomous driving planners to optimize for low-risk and predictable outcomes for human drivers around them. Although some work has been done to model driving style in planning and personalized autonomous polices, a gap exists in explicitly modeling human driving styles for trajectory forecasting of human behavior. Human driving style is most certainly a correlating factor to decision making, especially in edge-case scenarios where risk is nontrivial, as justified by the large amount of traffic psychology literature on risky driving. So far, the current real-world datasets for trajectory forecasting lack insight on the variety of represented driving styles. While the datasets may represent real-world distributions of driving styles, we posit that fringe driving style types may also be correlated with edge-case safety scenarios. In this work, we conduct analyses on existing real-world trajectory datasets for driving and dissect these works from the lens of driving styles, which is often intangible and non-standardized.
- Transportation > Ground > Road (1.00)
- Automobiles & Trucks (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning > Clustering (0.93)
- Information Technology > Artificial Intelligence > Robots > Autonomous Vehicles (0.87)
DISC: Dataset for Analyzing Driving Styles In Simulated Crashes for Mixed Autonomy
Kumar, Sandip Sharan Senthil, Thalapanane, Sandeep, Peethambari, Guru Nandhan Appiya Dilipkumar, SriHari, Sourang, Zheng, Laura, Lin, Ming C.
Handling pre-crash scenarios is still a major challenge for self-driving cars due to limited practical data and human-driving behavior datasets. We introduce DISC (Driving Styles In Simulated Crashes), one of the first datasets designed to capture various driving styles and behaviors in pre-crash scenarios for mixed autonomy analysis. DISC includes over 8 classes of driving styles/behaviors from hundreds of drivers navigating a simulated vehicle through a virtual city, encountering rare-event traffic scenarios. This dataset enables the classification of pre-crash human driving behaviors in unsafe conditions, supporting individualized trajectory prediction based on observed driving patterns. By utilizing a custom-designed VR-based in-house driving simulator, TRAVERSE, data was collected through a driver-centric study involving human drivers encountering twelve simulated accident scenarios. This dataset fills a critical gap in human-centric driving data for rare events involving interactions with autonomous vehicles. It enables autonomous systems to better react to human drivers and optimize trajectory prediction in mixed autonomy environments involving both human-driven and self-driving cars. In addition, individual driving behaviors are classified through a set of standardized questionnaires, carefully designed to identify and categorize driving behavior traits. We correlate data features with driving behaviors, showing that the simulated environment reflects real-world driving styles. DISC is the first dataset to capture how various driving styles respond to accident scenarios, offering significant potential to enhance autonomous vehicle safety and driving behavior analysis in mixed autonomy environments.
- North America > United States > Iowa (0.05)
- North America > United States > Montana (0.04)
- North America > United States > Maryland > Prince George's County > College Park (0.04)
- (3 more...)
- Transportation > Ground > Road (1.00)
- Automobiles & Trucks (1.00)
Anti-bullying Adaptive Cruise Control: A proactive right-of-way protection approach
Hu, Jia, Lian, Zhexi, Wang, Haoran, Zhang, Zihan, Qian, Ruoxi, Li, Duo, Jaehyun, null, So, null, Zheng, Junnian
The current Adaptive Cruise Control (ACC) systems are vulnerable to "road bully" such as cut-ins. This paper proposed an Anti-bullying Adaptive Cruise Control (AACC) approach with proactive right-of-way protection ability. It bears the following features: i) with the enhanced capability of preventing bullying from cut-ins; ii) optimal but not unsafe; iii) adaptive to various driving styles of cut-in vehicles; iv) with real-time field implementation capability. The proposed approach can identify other road users' driving styles online and conduct game-based motion planning for right-of-way protection. A detailed investigation of the simulation results shows that the proposed approach can prevent bullying from cut-ins and be adaptive to different cut-in vehicles' driving styles. The proposed approach is capable of enhancing travel efficiency by up to 29.55% under different cut-in gaps and can strengthen driving safety compared with the current ACC controller. The proposed approach is flexible and robust against traffic congestion levels. It can improve mobility by up to 11.93% and robustness by 8.74% in traffic flow. Furthermore, the proposed approach can support real-time field implementation by ensuring less than 50 milliseconds computation time.
- Asia > China > Shanghai > Shanghai (0.05)
- North America > United States > Massachusetts > Suffolk County > Boston (0.04)
- Europe > United Kingdom > England > Tyne and Wear > Newcastle (0.04)
- Asia > South Korea > Gyeonggi-do > Suwon (0.04)
- Transportation > Passenger (1.00)
- Transportation > Ground > Road (1.00)
- Consumer Products & Services > Travel (1.00)
- Automobiles & Trucks (1.00)