Dynamic EventNeRF: Reconstructing General Dynamic Scenes from Multi-view RGB and Event Streams

DOI

Volumetric reconstruction of dynamic scenes is an important problem in computer vision. It is especially challenging in poor lighting and with fast motion. This is partly due to limitations of RGB cameras: To capture frames under low lighting, the exposure time needs to be increased, which leads to more motion blur. In contrast, event cameras, which record changes in pixel brightness asynchronously, are much less dependent on lighting, making them more suitable for recording fast motion. We hence propose the first method to spatiotemporally reconstruct a scene from sparse multi-view event streams and sparse RGB frames. We train a sequence of cross-faded time-conditioned NeRF models, one per short recording segment. The individual segments are supervised with a set of event- and RGB-based losses and sparse-view regularisation. We assemble a real-world multi-view camera rig with six static event cameras around the object and record a benchmark multi-view event stream dataset of challenging motions. Our work outperforms RGB-based baselines, producing state-of-the-art results, and opens up the topic of multi-view event-based reconstruction as a new path for fast scene capture beyond RGB cameras. This dataset contains: 1) Synthetic scenes 2) Raw aedat4 event+RGB recordings and calibration 3) Processed event+RGB scenes to be used with model code

Identifier
DOI https://doi.org/10.17617/3.AD2LQB
Metadata Access https://edmond.mpg.de/api/datasets/export?exporter=dataverse_json&persistentId=doi:10.17617/3.AD2LQB
Provenance
Creator Rudnev, Viktor; Fox, Gereon; Elgharib, Mohamed; Theobalt, Christian; Golyanik, Vladislav
Publisher Edmond
Publication Year 2025
OpenAccess true
Contact vrudnev(at)mpi-inf.mpg.de
Representation
Language English
Resource Type Dataset
Version 1
Discipline Other