Multi-View Pedestrian Occupancy Prediction with a Novel Synthetic Dataset

Authors

  • Sithu Aung Korea Institute of Science and Technology
  • Min-Cheol Sagong Korea Institute of Science and Technology
  • Junghyun Cho Korea Institute of Science and Technology AI-Robotics, KIST School, University of Science and Technology Yonsei-KIST Convergence Research Institute, Yonsei University

DOI:

https://doi.org/10.1609/aaai.v39i2.32172

Abstract

We address an advanced challenge of predicting pedestrian occupancy as an extension of multi-view pedestrian detection in urban traffic. To support this, we have created a new synthetic dataset called MVP-Occ, designed for dense pedestrian scenarios in large-scale scenes. Our dataset provides detailed representations of pedestrians using voxel structures, accompanied by rich semantic scene understanding labels, facilitating visual navigation and insights into pedestrian spatial information. Furthermore, we present a robust baseline model, termed OmniOcc, capable of predicting both the voxel occupancy state and panoptic labels for the entire scene from multi-view images. Through in-depth analysis, we identify and evaluate the key elements of our proposed model, highlighting their specific contributions and importance.

Published

2025-04-11

How to Cite

Aung, S., Sagong, M.-C., & Cho, J. (2025). Multi-View Pedestrian Occupancy Prediction with a Novel Synthetic Dataset. Proceedings of the AAAI Conference on Artificial Intelligence, 39(2), 1782–1790. https://doi.org/10.1609/aaai.v39i2.32172

Issue

Section

AAAI Technical Track on Computer Vision I