Title
IBRNet: Learning Multi-View Image-Based Rendering
Abstract
We present a method that synthesizes novel views of complex scenes by interpolating a sparse set of nearby views. The core of our method is a network architecture that includes a multilayer perceptron and a ray transformer that estimates radiance and volume density at continuous 5D locations (3D spatial locations and 2D viewing directions), drawing appearance information on the fly from multiple source views. By drawing on source views at render time, our method hearkens back to classic work on image-based rendering (IBR), and allows us to render high-resolution imagery. Unlike neural scene representation work that optimizes per-scene functions for rendering, we learn a generic view interpolation function that generalizes to novel scenes. We render images using classic volume rendering, which is fully differentiable and allows us to train using only multi-view posed images as supervision. Experiments show that our method outperforms recent novel view synthesis methods that also seek to generalize to novel scenes. Further, if fine-tuned on each scene, our method is competitive with state-of-the-art single-scene neural rendering methods.
Year
DOI
Venue
2021
10.1109/CVPR46437.2021.00466
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021
DocType
ISSN
Citations 
Conference
1063-6919
2
PageRank 
References 
Authors
0.35
0
9
Name
Order
Citations
PageRank
Qianqian Wang121.71
Zhicheng Wang217617.00
Kyle Genova3293.83
Srinivasan, P.P.48110.65
Howard Zhou520.35
Jonathan T. Barron688139.55
Ricardo Martin-Brualla720.35
Noah Snavely84262197.04
Thomas A. Funkhouser97616475.01