EventNeRF: Neural Radiance Fields from a Single Colour Event Camera
Abstract
Asynchronously operating event cameras find many applications due to their high dynamic range, no motion blur, low latency and low data bandwidth. The field has seen remarkable progress during the last few years, and existing event-based 3D reconstruction approaches recover sparse point clouds of the scene. However, such sparsity is a limiting factor in many cases, especially in computer vision and graphics, that has not been addressed satisfactorily so far. Accordingly, this paper proposes the first approach for 3D-consistent, dense and photorealistic novel view synthesis using just a single colour event stream as input. At the core of our method is a neural radiance field trained entirely in a self-supervised manner from events while preserving the original resolution of the colour event channels. Next, our ray sampling strategy is tailored to events and allows for data-efficient training. At test, our method produces results in the RGB space at unprecedented quality. We evaluate our method qualitatively and quantitatively on several challenging synthetic and real scenes and show that it produces significantly denser and more visually appealing renderings than the existing methods. We also demonstrate robustness in challenging scenarios with fast motion and under low lighting conditions. We will release our dataset and our source code to facilitate the research field.
Baseline
Synthetic data results
E2VID+NeRF | Our EventNeRF | |||||
---|---|---|---|---|---|---|
Scene | PSNR ↑ | SSIM ↑ | LPIPS ↓ | PSNR ↑ | SSIM ↑ | LPIPS ↓ |
Drums | 19.71 | 0.85 | 0.22 | 27.43 | 0.91 | 0.07 |
Lego | 20.17 | 0.82 | 0.24 | 25.84 | 0.89 | 0.13 |
Chair | 24.12 | 0.92 | 0.12 | 30.62 | 0.94 | 0.05 |
Ficus | 24.97 | 0.92 | 0.10 | 31.94 | 0.94 | 0.05 |
Mic | 23.08 | 0.94 | 0.09 | 31.78 | 0.96 | 0.03 |
Hotdog | 24.38 | 0.93 | 0.12 | 30.26 | 0.94 | 0.04 |
Materials | 22.01 | 0.92 | 0.13 | 24.10 | 0.94 | 0.07 |
Average | 22.64 | 0.90 | 0.15 | 28.85 | 0.93 | 0.06 |
Real data capture setup
Real data results
Comparison to E2VID/ssl-E2VID+NeRF
Comparison to Deblur-NeRF
Ablation studies computed on the Drums
Method | PSNR ↑ | SSIM ↑ | LPIPS ↓ |
---|---|---|---|
Fixed 50 ms win. | 27.32 | 0.90 | 0.09 |
W/o neg. smpl. | 26.48 | 0.87 | 0.16 |
Full EventNeRF | 27.43 | 0.91 | 0.07 |
Data efficiency
Extracted mesh
Real-time Demo based on torch-ngp
Our EventNeRF | Our Real-time Implementation | |||||
---|---|---|---|---|---|---|
Scene | PSNR ↑ | SSIM ↑ | LPIPS ↓ | PSNR ↑ | SSIM ↑ | LPIPS ↓ |
Drums | 27.43 | 0.91 | 0.07 | 26.03 | 0.91 | 0.07 |
Lego | 25.84 | 0.89 | 0.13 | 22.82 | 0.89 | 0.08 |
Chair | 30.62 | 0.94 | 0.05 | 27.97 | 0.94 | 0.05 |
Ficus | 31.94 | 0.94 | 0.05 | 26.77 | 0.92 | 0.12 |
Mic | 31.78 | 0.96 | 0.03 | 28.34 | 0.95 | 0.04 |
Hotdog | 30.26 | 0.94 | 0.04 | 23.99 | 0.93 | 0.10 |
Materials | 24.10 | 0.94 | 0.07 | 26.05 | 0.93 | 0.07 |
Average | 28.85 | 0.93 | 0.06 | 25.99 | 0.92 | 0.07 |
Narrated Video with Experiments
Downloads
Citation
@InProceedings{rudnev2023eventnerf, title={EventNeRF: Neural Radiance Fields from a Single Colour Event Camera}, author={Viktor Rudnev and Mohamed Elgharib and Christian Theobalt and Vladislav Golyanik}, booktitle={Computer Vision and Pattern Recognition (CVPR)}, year={2023} }
Contact
For questions, clarifications, please get in touch with:Viktor Rudnev
vrudnev@mpi-inf.mpg.de
Vladislav Golyanik
golyanik@mpi-inf.mpg.de
Mohamed Elgharib
elgharib@mpi-inf.mpg.de