UniFuture

UniFuture: A 4D Driving World Model for Future Generation and Perception

Official Hugging Face repository for UniFuture.

Model description

UniFuture is a driving world model that jointly models future scene appearance and geometry within a unified framework.
Given the current image as input, it predicts future RGB-depth pairs for future generation and perception.

According to the project description, UniFuture introduces:

  • Dual-Latent Sharing for shared latent learning between image and depth sequences
  • Multi-scale Latent Interaction for bidirectional refinement between image and depth features across spatial scales

Paper

UniFuture: A 4D Driving World Model for Future Generation and Perception
ICRA 2026

Repository contents

  • unifuture.safetensors: pretrained model weights

Task

This repository is intended for research use on driving world modeling, including:

  • future scene generation
  • future depth prediction
  • future perception

Dataset

The paper reports experiments on nuScenes.

Usage

Please refer to the official GitHub repository for training, evaluation, and detailed usage instructions:

https://github.com/dk-liang/UniFuture

License

Apache-2.0

Citation

@inproceedings{liang2026UniFuture,
  title={UniFuture: A 4D Driving World Model for Future Generation and Perception},
  author={Liang, Dingkang and Zhang, Dingyuan and Zhou, Xin and Tu, Sifan and Feng, Tianrui and Li, Xiaofan and Zhang, Yumeng and Du, Mingyang and Tan, Xiao and Bai, Xiang},
  booktitle={IEEE International Conference on Robotics and Automation},
  year={2026}
}
Downloads last month
10
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support