Submitted:
14 January 2024
Posted:
15 January 2024
You are already at the latest version
Abstract
Keywords:
1. Introduction

- We propose a tensorial able relighting radiance field for single-tensor scene representation.
- We achieve visual quality comparable to the previous state-of-the-art model at a rendering and relighting speed is about 1.5× faster.
- Our model training does not rely on complex geometric surface properties.
2. Releate Work
2.1. Representations for Novel View Synthesis
2.2. NeRF with Reflectance Equation
3. Method
3.1. Scene Representation
3.2. Incident Light Field
- the direct light from the light source in the scene;
- the direct light received by the surface point;
- the indirect light reflected from other surface points [35].
3.2.1. Direct Light
3.2.2. The Transmittance Function
3.2.3. Indirect Light
3.3. The Rendering Equation
3.4. The BRDF Decoder
4. Results
4.1. Datasets
4.2. Comparisons
4.3. Ablation Studies
5. Discussion
Author Contributions
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Mildenhall, B.; Srinivasan, P.P.; Tancik, M.; Barron, J.T.; Ramamoorthi, R.; Ng, R. Nerf: Representing scenes as neural radiance fields for view synthesis. Communications of the ACM 2021, 65, 99–106. [Google Scholar] [CrossRef]
- Barron, J.T.; Mildenhall, B.; Tancik, M.; Hedman, P.; Martin-Brualla, R.; Srinivasan, P.P. Mip-nerf: A multiscale representation for anti-aliasing neural radiance fields. In Proceedings of the Proceedings of the IEEE/CVF International Conference on Computer Vision; 2021; pp. 5855–5864. [Google Scholar]
- Fridovich-Keil, S.; Yu, A.; Tancik, M.; Chen, Q.; Recht, B.; Kanazawa, A. Plenoxels: Radiance fields without neural networks. In Proceedings of the Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition; 2022; pp. 5501–5510. [Google Scholar]
- Zhang, K.; Riegler, G.; Snavely, N.; Koltun, V. Nerf++: Analyzing and improving neural radiance fields. arXiv 2020, arXiv:2010.07492 2020. [Google Scholar]
- Sun, C.; Sun, M.; Chen, H.T. Direct voxel grid optimization: Super-fast convergence for radiance fields reconstruction. In Proceedings of the Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition; 2022; pp. 5459–5469. [Google Scholar]
- Liu, L.; Gu, J.; Zaw Lin, K.; Chua, T.S.; Theobalt, C. Neural sparse voxel fields. Advances in Neural Information Processing Systems 2020, 33, 15651–15663. [Google Scholar]
- Yu, A.; Li, R.; Tancik, M.; Li, H.; Ng, R.; Kanazawa, A. Plenoctrees for real-time rendering of neural radiance fields. In Proceedings of the Proceedings of the IEEE/CVF International Conference on Computer Vision; 2021; pp. 5752–5761. [Google Scholar]
- Hedman, P.; Srinivasan, P.P.; Mildenhall, B.; Barron, J.T.; Debevec, P. Baking neural radiance fields for real-time view synthesis. In Proceedings of the Proceedings of the IEEE/CVF International Conference on Computer Vision; 2021; pp. 5875–5884. [Google Scholar]
- Garbin, S.J.; Kowalski, M.; Johnson, M.; Shotton, J.; Valentin, J. Fastnerf: High-fidelity neural rendering at 200fps. In Proceedings of the Proceedings of the IEEE/CVF International Conference on Computer Vision; 2021; pp. 14346–14355. [Google Scholar]
- Chen, A.; Xu, Z.; Geiger, A.; Yu, J.; Su, H. Tensorf: Tensorial radiance fields. In Proceedings of the European Conference on Computer Vision. Springer; 2022; pp. 333–350. [Google Scholar]
- Jin, H.; Liu, I.; Xu, P.; Zhang, X.; Han, S.; Bi, S.; Zhou, X.; Xu, Z.; Su, H. TensoIR: Tensorial Inverse Rendering. In Proceedings of the Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition; 2023; pp. 165–174. [Google Scholar]
- Haber, T.; Fuchs, C.; Bekaer, P.; Seidel, H.P.; Goesele, M.; Lensch, H.P. Relighting objects from image collections. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition. IEEE; 2009; pp. 627–634. [Google Scholar]
- Chen, Z.; Chen, A.; Zhang, G.; Wang, C.; Ji, Y.; Kutulakos, K.N.; Yu, J. A neural rendering framework for free-viewpoint relighting. In Proceedings of the Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition; 2020; pp. 5599–5610. [Google Scholar]
- Kajiya, J.T. The rendering equation. In Proceedings of the Proceedings of the 13th annual conference on Computer graphics and interactive techniques; 1986; pp. 143–150. [Google Scholar]
- Burley, B.; Studios, W.D.A. Physically-based shading at disney. In Proceedings of the Acm Siggraph 2012, Volume 2012, 1–7. [Google Scholar]
- Davis, A.; Levoy, M.; Durand, F. Unstructured light fields. In Proceedings of the Computer Graphics Forum. Wiley Online Library; 2012; Vol. 31, pp. 305–314. [Google Scholar]
- Bi, S.; Xu, Z.; Srinivasan, P.; Mildenhall, B.; Sunkavalli, K.; Hašan, M.; Hold-Geoffroy, Y.; Kriegman, D.; Ramamoorthi, R. Neural reflectance fields for appearance acquisition. arXiv, arXiv:2008.03824 2020.
- Bi, S.; Xu, Z.; Sunkavalli, K.; Hašan, M.; Hold-Geoffroy, Y.; Kriegman, D.; Ramamoorthi, R. Deep reflectance volumes: Relightable reconstructions from multi-view photometric images. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28 2020; Proceedings, Part III 16. Springer, 2020; pp. 294–311. [Google Scholar]
- Wang, J.; Ren, P.; Gong, M.; Snyder, J.; Guo, B. All-frequency rendering of dynamic, spatially-varying reflectance. In ACM SIGGRAPH Asia 2009 papers; 2009; pp. 1–10.
- Kolda, T.G.; Bader, B.W. Tensor decompositions and applications. SIAM review 2009, 51, 455–500. [Google Scholar] [CrossRef]
- Srinivasan, P.P.; Deng, B.; Zhang, X.; Tancik, M.; Mildenhall, B.; Barron, J.T. Nerv: Neural reflectance and visibility fields for relighting and view synthesis. In Proceedings of the Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition; 2021; pp. 7495–7504. [Google Scholar]
- Boss, M.; Braun, R.; Jampani, V.; Barron, J.T.; Liu, C.; Lensch, H. Nerd: Neural reflectance decomposition from image collections. In Proceedings of the Proceedings of the IEEE/CVF International Conference on Computer Vision; 2021; pp. 12684–12694. [Google Scholar]
- Boss, M.; Jampani, V.; Braun, R.; Liu, C.; Barron, J.; Lensch, H. Neural-pil: Neural pre-integrated lighting for reflectance decomposition. Advances in Neural Information Processing Systems 2021, 34, 10691–10704. [Google Scholar]
- Dong, Y.; Chen, G.; Peers, P.; Zhang, J.; Tong, X. Appearance-from-motion: Recovering spatially varying surface reflectance under unknown lighting. ACM Transactions on Graphics (TOG) 2014, 33, 1–12. [Google Scholar] [CrossRef]
- Zhang, X.; Srinivasan, P.P.; Deng, B.; Debevec, P.; Freeman, W.T.; Barron, J.T. Nerfactor: Neural factorization of shape and reflectance under an unknown illumination. ACM Transactions on Graphics (ToG) 2021, 40, 1–18. [Google Scholar] [CrossRef]
- Zhang, K.; Luan, F.; Wang, Q.; Bala, K.; Snavely, N. Physg: Inverse rendering with spherical gaussians for physics-based material editing and relighting. In Proceedings of the Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition; 2021; pp. 5453–5462. [Google Scholar]
- Yao, Y.; Zhang, J.; Liu, J.; Qu, Y.; Fang, T.; McKinnon, D.; Tsin, Y.; Quan, L. Neilf: Neural incident light field for physically-based material estimation. In Proceedings of the European Conference on Computer Vision; Springer, 2022; pp. 700–716. [Google Scholar]
- Müller, T.; Evans, A.; Schied, C.; Keller, A. Instant neural graphics primitives with a multiresolution hash encoding. ACM Transactions on Graphics (ToG) 2022, 41, 1–15. [Google Scholar] [CrossRef]
- Levin, A.; Durand, F. Linear view synthesis using a dimensionality gap light field prior. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition; IEEE, 2010; pp. 1831–1838. [Google Scholar]
- Levoy, M.; Hanrahan, P. Light field rendering. In Seminal Graphics Papers: Pushing the Boundaries, Volume 2; 2023; pp. 441–452. [Google Scholar]
- Shi, L.; Hassanieh, H.; Davis, A.; Katabi, D.; Durand, F. Light field reconstruction using sparsity in the continuous fourier domain. ACM Transactions on Graphics (TOG) 2014, 34, 1–13. [Google Scholar] [CrossRef]
- Azinovic, D.; Li, T.M.; Kaplanyan, A.; Nießner, M. Inverse path tracing for joint material and lighting estimation. In Proceedings of the Proceedings of the IEEE/CVF conference on computer vision and pattern recognition; 2019; pp. 2447–2456. [Google Scholar]
- Ramamoorthi, R.; Hanrahan, P. An efficient representation for irradiance environment maps. In Proceedings of the Proceedings of the 28th annual conference on Computer graphics and interactive techniques; 2001; pp. 497–500. [Google Scholar]
- Curless, B.; Levoy, M. A volumetric method for building complex models from range images. In Proceedings of the Proceedings of the 23rd annual conference on Computer graphics and interactive techniques; 1996; pp. 303–312. [Google Scholar]
- Zhang, Y.; Sun, J.; He, X.; Fu, H.; Jia, R.; Zhou, X. Modeling indirect illumination for inverse rendering. In Proceedings of the Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition; 2022; pp. 18643–18652. [Google Scholar]
- Yan, L.Q.; Zhou, Y.; Xu, K.; Wang, R. Accurate translucent material rendering under spherical gaussian lights. In Proceedings of the Computer Graphics Forum. Wiley Online Library; 2012; Vol. 31, pp. 2267–2276. [Google Scholar]
- Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: from error visibility to structural similarity. IEEE transactions on image processing 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
- Zhang, R.; Isola, P.; Efros, A.A.; Shechtman, E.; Wang, O. The unreasonable effectiveness of deep features as a perceptual metric. In Proceedings of the Proceedings of the IEEE conference on computer vision and pattern recognition; 2018; pp. 586–595. [Google Scholar]



| Sence | Method | Nomal | Novel View Synthesis | Relighting | Training | ||||
|---|---|---|---|---|---|---|---|---|---|
| MAE↓ | PSNR↑ | SSIM↑ | LPIPS↓ | PSNR↑ | SSIM↑ | LPIPS↓ | Time↓ | ||
| Ficus | NeRFactor | 6.328 | 21.688 | 0.925 | 0.101 | 20.932 | 0.903 | 0.110 | 83.0hrs |
| TensoIR | 4.452 | 29.114 | 0.963 | 0.052 | 24.183 | 0.941 | 0.071 | 3.2hrs | |
| ours | 4.192 | 29.693 | 0.968 | 0.048 | 25.323 | 0.944 | 0.069 | 2.2hrs | |
| Lego | NeRFactor | 9.892 | 26.088 | 0.873 | 0.156 | 23.505 | 0.855 | 0.158 | 84.0hrs |
| TensoIR | 5.515 | 34.652 | 0.956 | 0.045 | 28.433 | 0.912 | 0.084 | 3.5hrs | |
| ours | 5.700 | 35.084 | 0.973 | 0.044 | 27.142 | 0.882 | 0.082 | 1.8hrs | |
| Armadillo | NeRFactor | 3.455 | 26.584 | 0.946 | 0.093 | 26.723 | 0.941 | 0.110 | 73.0hrs |
| TensoIR | 2.063 | 38.143 | 0.980 | 0.047 | 34.402 | 0.975 | 0.046 | 2.8hrs | |
| ours | 1.853 | 40.136 | 0.983 | 0.045 | 31.779 | 0.955 | 0.050 | 1.6hrs | |
| Hotdog | NeRFactor | 5.681 | 23.366 | 0.922 | 0.144 | 22.651 | 0.903 | 0.161 | 77.0hrs |
| TensoIR | 4.120 | 35.298 | 0.969 | 0.054 | 27.898 | 0.928 | 0.122 | 2.7hrs | |
| ours | 4.761 | 34.421 | 0.946 | 0.061 | 27.702 | 0.911 | 0.138 | 2.2hrs | |
| Average | NeRFactor | 6.339 | 24.432 | 0.917 | 0.124 | 23.453 | 0.901 | 0.135 | 79.3hrs |
| TensoIR | 4.038 | 34.302 | 0.967 | 0.050 | 28.729 | 0.939 | 0.081 | 3.1hrs | |
| ours | 4.127 | 34.834 | 0.968 | 0.049 | 27.987 | 0.923 | 0.085 | 2.0hrs | |
| Novel View Synthesis | Relighting | Training | ||||
|---|---|---|---|---|---|---|
| PSNR↑ | SSIM↑ | PSNR↑ | SSIM↑ | Time↓ | ||
| Ours decoder w/ E.F. | 28.864 | 0.952 | 26.578 | 0.949 | 5.6hrs | |
| Ours decoder w/ MLP | 29.693 | 0.968 | 25.323 | 0.944 | 2.2hrs | |
| Ours decoder w/ LSTM | 29.784 | 0.968 | 25.489 | 0.947 | 12.7hrs | |
| Novel View Synthesis | Relighting | Training | ||||
|---|---|---|---|---|---|---|
| PSNR↑ | SSIM↑ | PSNR↑ | SSIM↑ | Time↓ | ||
| Ours w/ A.M.F tensor | 29.231 | 0.962 | 25.217 | 0.942 | 4.3hrs | |
| Ours w/ A.S.F tensor | 29.622 | 0.968 | 25.339 | 0.944 | 3.2 hrs | |
| Ours | 29.693 | 0.968 | 25.323 | 0.944 | 2.2hrs | |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).