Receive, Reason, and React: Drive as You Say with Large Language Models
in Autonomous Vehicles
- URL: http://arxiv.org/abs/2310.08034v1
- Date: Thu, 12 Oct 2023 04:56:01 GMT
- Title: Receive, Reason, and React: Drive as You Say with Large Language Models
in Autonomous Vehicles
- Authors: Can Cui, Yunsheng Ma, Xu Cao, Wenqian Ye and Ziran Wang
- Abstract summary: We propose a novel framework that leverages Large Language Models (LLMs) to enhance the decision-making process in autonomous vehicles.
Our research includes experiments in HighwayEnv, a collection of environments for autonomous driving and tactical decision-making tasks.
We also examine real-time personalization, demonstrating how LLMs can influence driving behaviors based on verbal commands.
- Score: 13.102404404559428
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: The fusion of human-centric design and artificial intelligence (AI)
capabilities has opened up new possibilities for next-generation autonomous
vehicles that go beyond transportation. These vehicles can dynamically interact
with passengers and adapt to their preferences. This paper proposes a novel
framework that leverages Large Language Models (LLMs) to enhance the
decision-making process in autonomous vehicles. By utilizing LLMs' linguistic
and contextual understanding abilities with specialized tools, we aim to
integrate the language and reasoning capabilities of LLMs into autonomous
vehicles. Our research includes experiments in HighwayEnv, a collection of
environments for autonomous driving and tactical decision-making tasks, to
explore LLMs' interpretation, interaction, and reasoning in various scenarios.
We also examine real-time personalization, demonstrating how LLMs can influence
driving behaviors based on verbal commands. Our empirical results highlight the
substantial advantages of utilizing chain-of-thought prompting, leading to
improved driving decisions, and showing the potential for LLMs to enhance
personalized driving experiences through ongoing verbal feedback. The proposed
framework aims to transform autonomous vehicle operations, offering
personalized support, transparent decision-making, and continuous learning to
enhance safety and effectiveness. We achieve user-centric, transparent, and
adaptive autonomous driving ecosystems supported by the integration of LLMs
into autonomous vehicles.
Related papers
- Large Language Models for Autonomous Driving (LLM4AD): Concept, Benchmark, Simulation, and Real-Vehicle Experiment [15.52530518623987]
Large Language Models (LLMs) have the potential to enhance various aspects of autonomous driving systems.
This paper introduces novel concepts and approaches to designing LLMs for autonomous driving (LLM4AD)
arXiv Detail & Related papers (2024-10-20T04:36:19Z) - Probing Multimodal LLMs as World Models for Driving [72.18727651074563]
We look at the application of Multimodal Large Language Models (MLLMs) in autonomous driving.
Despite advances in models like GPT-4o, their performance in complex driving environments remains largely unexplored.
arXiv Detail & Related papers (2024-05-09T17:52:42Z) - Hybrid Reasoning Based on Large Language Models for Autonomous Car Driving [14.64475022650084]
Large Language Models (LLMs) have garnered significant attention for their ability to understand text and images, generate human-like text, and perform complex reasoning tasks.
We investigate how well LLMs can adapt and apply a combination of arithmetic and common-sense reasoning, particularly in autonomous driving scenarios.
arXiv Detail & Related papers (2024-02-21T08:09:05Z) - DriveMLM: Aligning Multi-Modal Large Language Models with Behavioral
Planning States for Autonomous Driving [69.82743399946371]
DriveMLM is a framework that can perform close-loop autonomous driving in realistic simulators.
We employ a multi-modal LLM (MLLM) to model the behavior planning module of a module AD system.
This model can plug-and-play in existing AD systems such as Apollo for close-loop driving.
arXiv Detail & Related papers (2023-12-14T18:59:05Z) - Evaluation of Large Language Models for Decision Making in Autonomous
Driving [4.271294502084542]
One strategy of using Large Language Models (LLMs) for autonomous driving involves inputting surrounding objects as text prompts to the LLMs.
When using LLMs for such purposes, capabilities such as spatial recognition and planning are essential.
This study quantitatively evaluated these two abilities of LLMs in the context of autonomous driving.
arXiv Detail & Related papers (2023-12-11T12:56:40Z) - Empowering Autonomous Driving with Large Language Models: A Safety Perspective [82.90376711290808]
This paper explores the integration of Large Language Models (LLMs) into Autonomous Driving systems.
LLMs are intelligent decision-makers in behavioral planning, augmented with a safety verifier shield for contextual safety learning.
We present two key studies in a simulated environment: an adaptive LLM-conditioned Model Predictive Control (MPC) and an LLM-enabled interactive behavior planning scheme with a state machine.
arXiv Detail & Related papers (2023-11-28T03:13:09Z) - LLM4Drive: A Survey of Large Language Models for Autonomous Driving [62.10344445241105]
Large language models (LLMs) have demonstrated abilities including understanding context, logical reasoning, and generating answers.
In this paper, we systematically review a research line about textitLarge Language Models for Autonomous Driving (LLM4AD).
arXiv Detail & Related papers (2023-11-02T07:23:33Z) - LanguageMPC: Large Language Models as Decision Makers for Autonomous
Driving [87.1164964709168]
This work employs Large Language Models (LLMs) as a decision-making component for complex autonomous driving scenarios.
Extensive experiments demonstrate that our proposed method not only consistently surpasses baseline approaches in single-vehicle tasks, but also helps handle complex driving behaviors even multi-vehicle coordination.
arXiv Detail & Related papers (2023-10-04T17:59:49Z) - DriveGPT4: Interpretable End-to-end Autonomous Driving via Large Language Model [84.29836263441136]
This study introduces DriveGPT4, a novel interpretable end-to-end autonomous driving system based on multimodal large language models (MLLMs)
DriveGPT4 facilitates the interpretation of vehicle actions, offers pertinent reasoning, and effectively addresses a diverse range of questions posed by users.
arXiv Detail & Related papers (2023-10-02T17:59:52Z) - DiLu: A Knowledge-Driven Approach to Autonomous Driving with Large
Language Models [30.23228092898916]
We propose the DiLu framework, which combines a Reasoning and a Reflection module to enable the system to perform decision-making based on common-sense knowledge.
Extensive experiments prove DiLu's capability to accumulate experience and demonstrate a significant advantage in generalization ability.
To the best of our knowledge, we are the first to leverage knowledge-driven capability in decision-making for autonomous vehicles.
arXiv Detail & Related papers (2023-09-28T09:41:35Z) - Drive as You Speak: Enabling Human-Like Interaction with Large Language
Models in Autonomous Vehicles [13.102404404559428]
We present a novel framework that leverages Large Language Models (LLMs) to enhance autonomous vehicles' decision-making processes.
The proposed framework holds the potential to revolutionize the way autonomous vehicles operate, offering personalized assistance, continuous learning, and transparent decision-making.
arXiv Detail & Related papers (2023-09-19T00:47:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.