Skip to the content.

:wave: Welcome to the Large-scale 3D Semantic Occupancy Prediction Challenge organized at :wave: WACV 2026

WACV 2026 Workshop / Challenge Info: TBA (date, venue, online link)


NuPlan-Occ overview figure
OpenScene vs NuPlan-Occ comparison

:page_facing_up: Paper

:tv: Video

Coming soon…


πŸ“Œ Overview

Semantic occupancy prediction is a cornerstone of 3D scene understanding in autonomous driving, providing dense, voxel-level semantic and geometric representations of dynamic environments. Despite recent advances, large-scale, high-resolution occupancy datasets remain scarce, limiting the development and evaluation of robust models.

This challenge introduces NuPlan-Occ, the largest publicly available semantic occupancy dataset to date, featuring 3.6 million frames with high-resolution voxel annotations (400 Γ— 400 Γ— 32). Derived from the widely adopted NuPlan benchmark, NuPlan-Occ enables scalable training and evaluation of both discriminative and generative models for 3D scene understanding.

We invite researchers to develop and submit models for 3D semantic occupancy prediction, with the goal of advancing state-of-the-art performance in accuracy, scalability, and generalization.

Dataset comparison table

🎯 Task

Participants are tasked with predicting 3D semantic occupancy grids from multi-view camera images.

The occupancy grid is defined over a predefined spatial volume with 9 semantic classes including:

vehicle, pedestrian, bicycle, traffic_cone, barrier, czone_sign, generic_object, background, empty

Two tracks are supported:


πŸ“Š Dataset: NuPlan-Occ

Access

Dense reconstruction and labeling pipeline

βš™οΈ Evaluation

Primary Metric: mean Intersection-over-Union (mIoU) across all semantic classes.

Secondary Metrics:

Evaluation code is released in the GitHub repository under the SOP/ directory:
https://github.com/Arlo0o/UniScene-Unified-Occupancy-centric-Driving-Scene-Generation/tree/v2/SOP/monoscene#3-evaluation

πŸ“ Evaluation Guidelines


πŸ† Baseline & Benchmark

We provide a reproduced baseline using MonoScene trained on NuPlan-Occ miniset:

Metric Value
Precision 48.99
Recall 42.54
IoU 29.49
mIoU 9.36

Per-class IoU:
background: 29.0124, vehicle: 17.6694, bicycle: 0.4056, pedestrian: 6.4708, traffic_cone: 1.8878, barrier: 2.8173, czone_sign: 2.7924, generic_object: 13.8316

Baseline Code & Pretrained Model:
Released in the GitHub repository under the SOP/ directory.
https://github.com/Arlo0o/UniScene-Unified-Occupancy-centric-Driving-Scene-Generation/tree/v2/SOP/monoscene


πŸ… Awards & Recognition


πŸ“š Recommended Readings & Citations

Participants are encouraged to cite the following works:

```bibtex @inproceedings{li2025uniscene, title={Uniscene: Unified occupancy-centric driving scene generation}, author={Li, Bohan and Guo, Jiazhe and Liu, Hongsi and Zou, Yin…u and Tan, Feiyang and Zhang, Chi and Wang, Tiancai and others}, booktitle={Proceedings of the Computer Vision and Pattern Recognition Conference}, pages={11971–11981}, year={2025} }

@article{li2025scaling, title={Scaling Up Occupancy-centric Driving Scene Generation: Dataset and Method}, author={Li, Bohan and Jin, Xin and Zhu, Hu and Liu, Hongsi and… Kaiwen and Ma, Chao and Jin, Yueming and Zhao, Hao and others}, journal={arXiv preprint arXiv:2510.22973}, year={2025} }