TrafficSafetyGPT: Tuning a Pre-trained Large Language Model to a
Domain-Specific Expert in Transportation Safety
- URL: http://arxiv.org/abs/2307.15311v1
- Date: Fri, 28 Jul 2023 05:17:11 GMT
- Title: TrafficSafetyGPT: Tuning a Pre-trained Large Language Model to a
Domain-Specific Expert in Transportation Safety
- Authors: Ou Zheng, Mohamed Abdel-Aty, Dongdong Wang, Chenzhu Wang, Shengxuan
Ding
- Abstract summary: Large Language Models (LLMs) have shown remarkable effectiveness in various general-domain natural language processing (NLP) tasks.
We introduce TrafficSafetyGPT, a novel LLAMA-based model, which has undergone supervised fine-tuning using TrafficSafety-2K dataset.
- Score: 2.1906688755530968
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Large Language Models (LLMs) have shown remarkable effectiveness in various
general-domain natural language processing (NLP) tasks. However, their
performance in transportation safety domain tasks has been suboptimal,
primarily attributed to the requirement for specialized transportation safety
expertise in generating accurate responses [1]. To address this challenge, we
introduce TrafficSafetyGPT, a novel LLAMA-based model, which has undergone
supervised fine-tuning using TrafficSafety-2K dataset which has human labels
from government produced guiding books and ChatGPT-generated instruction-output
pairs. Our proposed TrafficSafetyGPT model and TrafficSafety-2K train dataset
are accessible at https://github.com/ozheng1993/TrafficSafetyGPT.
Related papers
- What Makes and Breaks Safety Fine-tuning? A Mechanistic Study [64.9691741899956]
Safety fine-tuning helps align Large Language Models (LLMs) with human preferences for their safe deployment.
We design a synthetic data generation framework that captures salient aspects of an unsafe input.
Using this, we investigate three well-known safety fine-tuning methods.
arXiv Detail & Related papers (2024-07-14T16:12:57Z) - Refuse Whenever You Feel Unsafe: Improving Safety in LLMs via Decoupled Refusal Training [67.30423823744506]
This study addresses a critical gap in safety tuning practices for Large Language Models (LLMs)
We introduce a novel approach, Decoupled Refusal Training (DeRTa), designed to empower LLMs to refuse compliance to harmful prompts at any response position.
DeRTa incorporates two novel components: (1) Maximum Likelihood Estimation with Harmful Response Prefix, which trains models to recognize and avoid unsafe content by appending a segment of harmful response to the beginning of a safe response, and (2) Reinforced Transition Optimization (RTO), which equips models with the ability to transition from potential harm to safety refusal consistently throughout the harmful
arXiv Detail & Related papers (2024-07-12T09:36:33Z) - Digital Twin-based Driver Risk-Aware Intelligent Mobility Analytics for Urban Transportation Management [18.015270631863665]
We propose a Digital Twin-based Driver Risk-Aware Intelligent Mobility Analytics system.
The DT-DIMA system integrates real-time traffic information from pan-tilt-cameras.
The system predicts network-wide mobility and safety risks in real time.
arXiv Detail & Related papers (2024-07-03T01:44:22Z) - BjTT: A Large-scale Multimodal Dataset for Traffic Prediction [49.93028461584377]
Traditional traffic prediction methods rely on historical traffic data to predict traffic trends.
In this work, we explore how generative models combined with text describing the traffic system can be applied for traffic generation.
We propose ChatTraffic, the first diffusion model for text-to-traffic generation.
arXiv Detail & Related papers (2024-03-08T04:19:56Z) - TransGPT: Multi-modal Generative Pre-trained Transformer for
Transportation [19.184173455587263]
This paper presents TransGPT, a novel (multi-modal) large language model for the transportation domain.
It consists of two independent variants: TransGPT-SM for single-modal data and TransGPT-MM for multi-modal data.
This work advances the state-of-the-art of NLP in the transportation domain and provides a useful tool for ITS researchers and practitioners.
arXiv Detail & Related papers (2024-02-11T15:50:35Z) - Empowering Autonomous Driving with Large Language Models: A Safety Perspective [82.90376711290808]
This paper explores the integration of Large Language Models (LLMs) into Autonomous Driving systems.
LLMs are intelligent decision-makers in behavioral planning, augmented with a safety verifier shield for contextual safety learning.
We present two key studies in a simulated environment: an adaptive LLM-conditioned Model Predictive Control (MPC) and an LLM-enabled interactive behavior planning scheme with a state machine.
arXiv Detail & Related papers (2023-11-28T03:13:09Z) - CAT: Closed-loop Adversarial Training for Safe End-to-End Driving [54.60865656161679]
Adversarial Training (CAT) is a framework for safe end-to-end driving in autonomous vehicles.
Cat aims to continuously improve the safety of driving agents by training the agent on safety-critical scenarios.
Cat can effectively generate adversarial scenarios countering the agent being trained.
arXiv Detail & Related papers (2023-10-19T02:49:31Z) - Generating and Explaining Corner Cases Using Learnt Probabilistic Lane
Graphs [5.309950889075669]
We introduce Probabilistic Lane Graphs (PLGs) to describe a finite set of lane positions and directions in which vehicles might travel.
The structure of PLGs is learnt directly from historic traffic data.
We use reinforcement learning techniques to modify this policy to generate realistic and explainable corner case scenarios.
arXiv Detail & Related papers (2023-08-25T20:17:49Z) - Towards formalization and monitoring of microscopic traffic parameters
using temporal logic [1.3706331473063877]
We develop specification-based monitoring for the analysis of traffic networks using the formal language Signal Temporal Logic.
We develop monitors that identify safety-related behavior such as conforming to speed limits and maintaining appropriate headway.
This work can be utilized by traffic management centers to study the traffic stream properties, identify possible hazards, and provide valuable feedback for automating the traffic monitoring systems.
arXiv Detail & Related papers (2021-10-12T17:59:26Z) - Cautious Adaptation For Reinforcement Learning in Safety-Critical
Settings [129.80279257258098]
Reinforcement learning (RL) in real-world safety-critical target settings like urban driving is hazardous.
We propose a "safety-critical adaptation" task setting: an agent first trains in non-safety-critical "source" environments.
We propose a solution approach, CARL, that builds on the intuition that prior experience in diverse environments equips an agent to estimate risk.
arXiv Detail & Related papers (2020-08-15T01:40:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.