My research focuses on enabling AI agents to operate efficiently in practical and valuable scenarios, with an emphasis on allowing foundation models to learn and improve during run-time rather than solely at design-time. My prior work centered on addressing complex interaction challenges in autonomous driving. I am passionate about the future of artificial general intelligence (AGI) and excited to contribute to these rapidly advancing fields.
Large Language Models (LLMs), despite their advancements, are fundamentally limited by their static parametric knowledge, hindering performance on tasks requiring open-domain up-to-date information. While enabling LLMs to interact with external knowledge environments is a promising solution, current efforts primarily address closed-end problems. Open-ended questions, which characterized by lacking a standard answer or providing non-unique and diverse answers, remain underexplored. To bridge this gap, we present O2-Searcher, a novel search agent leveraging reinforcement learning to effectively tackle both open-ended and closed-ended questions in the open domain. O2-Searcher leverages an efficient, locally simulated search environment for dynamic knowledge acquisition, effectively decoupling the external world knowledge from model’s sophisticated reasoning processes. It employs a unified training mechanism with meticulously designed reward functions, enabling the agent to identify problem types and adapt different answer generation strategies. Furthermore, to evaluate performance on complex open-ended tasks, we construct O-QA, a high-quality benchmark featuring 300 manually curated, multi-domain open-ended questions with associated web page caches. Extensive experiments show that O2-Searcher, using only a 3B model, significantly surpasses leading LLM agents on O-QA. It also achieves SOTA results on various closed-ended QA benchmarks against similarly-sized models, while performing on par with much larger ones.
@article{mei2025o2searchersearchingbasedagentmodel,title={O$^2$-Searcher: A Searching-based Agent Model for Open-Domain Open-Ended Question Answering},author={Mei, Jianbiao and Hu, Tao and Fu, Daocheng and Wen, Licheng and Yang, Xuemeng and Wu, Rong and Cai, Pinlong and Cai, Xinyu and Gao, Xing and Yang, Yu and Xie, Chengjun and Shi, Botian and Liu, Yong and Qiao, Yu},journal={arXiv preprint arXiv:2505.16582},year={2025},}
DriveArena: A Closed-loop Generative Simulation Platform for Autonomous Driving
Xuemeng Yang*, Licheng Wen*, Yukai Ma*, and 11 more authors
2025 IEEE/CVF International Conference on Computer Vision (ICCV), 2025
This paper presetns DriveArena, the first high-fidelity closed-loop simulation system designed for driving agents navigating in real scenarios. DriveArena features a flexible, modular architecture, allowing for the seamless interchange of its core components: Traffic Manager, a traffic simulator capable of generating realistic traf- fic flow on any worldwide street map, and World Dreamer, a high-fidelity conditional generative model with infinite autoregression. This powerful synergy empowers any driving agent capable of processing real-world images to navigate in DriveArena simulated environment. The agent perceives its surroundings through images generated by World Dreamer and output trajectories; then these trajectories are fed into Traffic Manager, achieving realistic interactions with other vehicles and producing a new scene lay- out. Finally, the latest scene layout is relayed back into World Dreamer, perpetuating the simulation cycle. This iterative process fosters closed-loop exploration within a highly realistic environment, providing a valuable platform for developing and evaluating driving agents across diverse and challenging scenarios. DriveArena signifies a substantial leap forward in leveraging generative image data for the driving simulatior, opening insights for closed-loop autonomous driving.
@article{yang2024drivearena,title={DriveArena: A Closed-loop Generative Simulation Platform for Autonomous Driving},author={Yang*, Xuemeng and Wen*, Licheng and Ma*, Yukai and Mei*, Jianbiao and Li*, Xin and Wei*, Tiantian and Lei, Wenjie and Fu, Daocheng and Cai, Pinlong and Dou, Min and Shi, Botian and He, Liang and Liu, Yong and Qiao, Yu},journal={2025 IEEE/CVF International Conference on Computer Vision (ICCV)},year={2025},demo={https://pjlab-adg.github.io/DriveArena/},}
2024
ICLR
DiLu: A Knowledge-Driven Approach to Autonomous Driving with Large Language Models
Licheng Wen*, Daocheng Fu*, Xin Li*, and 7 more authors
In The Eleventh International Conference on Learning Representations (ICLR), 2024
Recent advancements in autonomous driving have relied on data-driven approaches, which are widely adopted but face challenges including dataset bias, overfitting, and uninterpretability. Drawing inspiration from the knowledge-driven nature of human driving, we explore the question of how to instill similar capabilities into autonomous driving systems and summarize a paradigm that integrates an interactive environment, a driver agent, as well as a memory component to address this question. Leveraging large language models with emergent abilities, we propose the DiLu framework, which combines a Reasoning and a Reflection module to enable the system to perform decision-making based on common-sense knowledge and evolve continuously. Extensive experiments prove DiLu’s capability to accumulate experience and demonstrate a significant advantage in generalization ability over reinforcement learning-based methods. Moreover, DiLu is able to directly acquire experiences from real-world datasets which highlights its potential to be deployed on practical autonomous driving systems. To the best of our knowledge, we are the first to instill knowledge-driven capability into autonomous driving systems from the perspective of how humans drive.
@inproceedings{wen2023dilu,title={DiLu: A Knowledge-Driven Approach to Autonomous Driving with Large Language Models},author={Wen*, Licheng and Fu*, Daocheng and Li*, Xin and Cai, Xinyu and Ma, Tao and Cai, Pinlong and Dou, Min and Shi, Botian and He, Liang and Qiao, Yu},booktitle={The Eleventh International Conference on Learning Representations (ICLR)},year={2024},demo={https://pjlab-adg.github.io/DiLu/},}
On the Road with GPT-4V(ision): Early Explorations of Visual-Language Model on Autonomous Driving
Licheng Wen*, Xuemeng Yang*, Daocheng Fu*, and 14 more authors
In ICLR 2024 Workshop on Large Language Model (LLM) Agents, 2024
The development of autonomous driving technology depends on merging perception, decision, and control systems. Traditional strategies have struggled to understand complex driving environments and other road users’ intent. This bottleneck, especially in constructing common sense reasoning and nuanced scene understanding, affects the safe and reliable operations of autonomous vehicles. The introduction of Visual Language Models (VLM) opens up possibilities for fully autonomous driving. This report evaluates the potential of GPT-4V(ision), the latest state-of-the-art VLM, as an autonomous driving agent. The evaluation primarily assesses the model’s ultimate ability to act as a driving agent under varying conditions, while also considering its capacity to understand driving scenes and make decisions. Findings show that GPT-4V outperforms existing systems in scene understanding and causal reasoning. It has the potential in handling unexpected scenarios, understanding intentions, and making informed decisions. However, limitations remain in direction determination, traffic light recognition, vision grounding, and spatial reasoning tasks, highlighting the need for further research.
@inproceedings{wen2024road,title={On the Road with GPT-4V(ision): Early Explorations of Visual-Language Model on Autonomous Driving},author={Wen*, Licheng and Yang*, Xuemeng and Fu*, Daocheng and Wang*, Xiaofeng and Cai, Pinlong and Li, Xin and Ma, Tao and Li, Yingxuan and Xu, Linran and Shang, Dengke and Zhu, Zheng and Sun, Shaoyan and Bai, Yeqi and Cai, Xinyu and Dou, Min and Hu, Shuanglu and Shi, Botian},booktitle={ICLR 2024 Workshop on Large Language Model (LLM) Agents},year={2024},}
2023
ITSC
LimSim: A Long-term Interactive Multi-scenario Traffic Simulator
Licheng Wen*, Daocheng Fu*, Song Mao, and 3 more authors
In 2023 IEEE 26th International Conference on Intelligent Transportation Systems (ITSC), 2023
With the growing popularity of digital twin and autonomous driving in transportation, the demand for simulation systems capable of generating high-fidelity and reliable scenarios is increasing. Existing simulation systems suffer from a lack of support for different types of scenarios, and the vehicle models used in these systems are too simplistic. Thus, such systems fail to represent driving styles and multi-vehicle interactions, and struggle to handle corner cases in the dataset. In this paper, we propose LimSim, the Long-term Interactive Multi-scenario traffic Simulator, which aims to provide a long-term continuous simulation capability under the urban road network. LimSim can simulate fine-grained dynamic scenarios and focus on the diverse interactions between multiple vehicles in the traffic flow. This paper provides a detailed introduction to the framework and features of the LimSim, and demonstrates its performance through case studies and experiments. LimSim is now open source on GitHub.
@inproceedings{wen2023limsim,title={LimSim: A Long-term Interactive Multi-scenario Traffic Simulator},author={Wen*, Licheng and Fu*, Daocheng and Mao, Song and Cai, Pinlong and Dou, Min and Li, Yikang},year={2023},booktitle={2023 IEEE 26th International Conference on Intelligent Transportation Systems (ITSC)},video={https://www.youtube.com/playlist?list=PLNeNtm096CAyYD1JJnkQ4gMaoFSdFLn2y},}
2022
CL-MAPF: Multi-Agent Path Finding for Car-Like robots with kinematic and spatiotemporal constraints
Multi-Agent Path Finding has been widely studied in the past few years due to its broad application in the field of robotics and AI. However, previous solvers rely on several simplifying assumptions. This limits their applicability in numerous real-world domains that adopt nonholonomic car-like agents rather than holonomic ones. In this paper, we give a mathematical formalization of the Multi-Agent Path Finding for Car-Like robots (CL-MAPF) problem. We propose a novel hierarchical search-based solver called Car-Like Conflict-Based Search to address this problem. It applies a body conflict tree to address collisions considering the shapes of the agents. We introduce a new algorithm called Spatiotemporal Hybrid-State A* as the single-agent planner to generate agents’ paths satisfying both kinematic and spatiotemporal constraints. We also present a sequential planning version of our method, sacrificing a small amount of solution quality to achieve a significant reduction in runtime. We compare our method with two baseline algorithms on a dedicated benchmark and validate it in real-world scenarios. The experiment results show that the planning success rate of both baseline algorithms is below 50% for all six scenarios, while our algorithm maintains that of over 98%. It also gives clear evidence that our algorithm scales well to 100 agents in 300 m × 300 m scenario and is able to produce solutions that can be directly applied to Ackermann-steering robots in the real world. The benchmark and source code are released in https://github.com/APRIL-ZJU/CL-CBS. The video of the experiments can be found on YouTube.
@article{10.1016/j.robot.2021.103997,title={CL-MAPF: Multi-Agent Path Finding for Car-Like robots with kinematic and spatiotemporal constraints},journal={Robotics and Autonomous Systems},volume={150},pages={103997},year={2022},issn={0921-8890},doi={https://doi.org/10.1016/j.robot.2021.103997},url={https://www.sciencedirect.com/science/article/pii/S0921889021002530},author={Wen, Licheng and Liu, Yong and Li, Hongliang},keywords={Multi-agent systems, Path planning, Mobile robots},video={https://www.youtube.com/watch?v=KThsX04ABvc},}
Talks
On October 31st, 2024, I had the honor of presenting a talk titled "Empowering Automated Driving with LLMs: A Knowledge-driven Paradigm" to SAE International as part of their AI Webinar series.
Email is the best way to reach me: wenlicheng [at] pjlab dot org dot cn