🤖 AI Summary
To address the joint lateral and longitudinal decision-making challenge for connected and autonomous vehicles (CAVs) in multi-vehicle cooperative scenarios, this paper proposes a value-driven parallel-update Monte Carlo Tree Search (MCTS) method, formulated within a finite-horizon discounted-time multi-agent Markov game framework. We innovatively design a partially steady traffic-flow-guided parallel action analysis mechanism to enable efficient action pruning and synergistic depth-width optimization. Compared with conventional MCTS, state-of-the-art reinforcement learning, and heuristic approaches, our method significantly improves hazardous maneuver detection efficiency and policy robustness under large-scale stochastic traffic flows. Specifically, it increases throughput in cooperative zones by 12.7% and reduces collision rates by 63%. Both safety and efficiency metrics surpass those of human drivers, demonstrating superior comprehensive performance in real-world CAV coordination tasks.
📝 Abstract
To solve the problem of lateral and logitudinal joint decision-making of multi-vehicle cooperative driving for connected and automated vehicles (CAVs), this paper proposes a Monte Carlo tree search (MCTS) method with parallel update for multi-agent Markov game with limited horizon and time discounted setting. By analyzing the parallel actions in the multi-vehicle joint action space in the partial-steady-state traffic flow, the parallel update method can quickly exclude potential dangerous actions, thereby increasing the search depth without sacrificing the search breadth. The proposed method is tested in a large number of randomly generated traffic flow. The experiment results show that the algorithm has good robustness and better performance than the SOTA reinforcement learning algorithms and heuristic methods. The vehicle driving strategy using the proposed algorithm shows rationality beyond human drivers, and has advantages in traffic efficiency and safety in the coordinating zone.