Skip to content

Official implementation of the paper "Spatial-Temporal Large Language Model for Traffic Prediction"

License

Notifications You must be signed in to change notification settings

ChenxiLiu-HNU/ST-LLM

Repository files navigation

Spatial-Temporal Large Language Model for Traffic Prediction

This repository contains the code of ST-LLM "Spatial-Temporal Large Language Model for Traffic Prediction" paper

Abstract

Traffic prediction, an essential component for intelligent transportation systems, endeavours to use historical data to foresee future traffic features at specific locations. Although existing traffic prediction models often emphasize developing complex neural network structures, their accuracy has not improved. Recently, large language models have shown outstanding capabilities in time series analysis. Differing from existing models, LLMs progress mainly through parameter expansion and extensive pretraining while maintaining their fundamental structures. Motivated by these developments, we propose a Spatial-Temporal Large Language Model (ST-LLM) for traffic prediction. In the ST-LLM, we define timesteps at each location as tokens and design a spatial-temporal embedding to learn the spatial location and global temporal patterns of these tokens. Additionally, we integrate these embeddings by a fusion convolution to each token for a unified spatial-temporal representation. Furthermore, we innovate a partially frozen attention strategy to adapt the LLM to capture global spatial-temporal dependencies for traffic prediction. Comprehensive experiments on real traffic datasets offer evidence that ST-LLM is a powerful spatial-temporal learner that outperforms state-of-the-art models. Notably, the ST-LLM also exhibits robust performance in both few-shot and zero-shot prediction scenarios.

image

Dependencies

  • Python 3.8.19
  • PyTorch 2.4.1
  • cuda 11.7
  • torchvision 0.19.1
> conda env create -f env_ubuntu.yaml

Datasets

We provide preprocessed datasets, which you can access here.
If you need the original datasets, please refer to the ESG.

Training

CUDA_VISIBLE_DEVICES=0
nohup python train.py --data taxi_pick --device cuda:0  > your_log_name.log &

BibTeX

If you find our work useful in your research. Please consider giving a star ⭐ and citation 📚.

@inproceedings{liu2024spatial,
  title={Spatial-temporal large language model for traffic prediction},
  author={Liu, Chenxi and Yang, Sun and Xu, Qianxiong and Li, Zhishuai and Long, Cheng and Li, Ziyue and Zhao, Rui},
  booktitle={MDM},
  year={2024}
}

Acknowledgement

Our implementation adapts OFA as the code base and has extensively modified it for our purposes. We are grateful to the authors for providing their implementations and related resources.

About

Official implementation of the paper "Spatial-Temporal Large Language Model for Traffic Prediction"

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages