Yuchen Li yuchen.li@durham.ac.uk
PGR Student Doctor of Philosophy
SID-NERF: Few-Shot Nerf Based on Scene Information Distribution
Li, Yuchen; Wan, Fan; Long, Yang
Authors
Fan Wan fan.wan@durham.ac.uk
PGR Student Doctor of Philosophy
Dr Yang Long yang.long@durham.ac.uk
Associate Professor
Abstract
The novel view synthesis from a limited set of images is a significant research focus. Traditional NeRF methods, relying mainly on color supervision, struggle with accurate scene geometry reconstruction when faced with sparse input images, leading to suboptimal rendering. We propose a Few-shot NeRF Based on Scene Information Distribution(Sid-NeRF) to address this by integrating geometric and color supervision, enhancing the model’s understanding of scene geometry. We also implement a data selector during training to identify and utilize the most accurate geometric data, thus improving training efficiency. Additionally, a residual module is introduced to counteract any optimization biases from the selector. Our method was tested on three datasets and showed excellent performance in various environments with limited images. Notably, compared to other novel view synthesis methods based on fewer views, our method does not require any prior knowledge and thus does not incur additional computational and storage costs.
Citation
Li, Y., Wan, F., & Long, Y. (2024, July). SID-NERF: Few-Shot Nerf Based on Scene Information Distribution. Presented at 2024 IEEE International Conference on Multimedia and Expo (ICME), Niagara Falls, ON, Canada
Presentation Conference Type | Conference Paper (published) |
---|---|
Conference Name | 2024 IEEE International Conference on Multimedia and Expo (ICME) |
Start Date | Jul 15, 2024 |
End Date | Jul 19, 2024 |
Acceptance Date | May 1, 2024 |
Online Publication Date | Jul 15, 2024 |
Publication Date | Jul 15, 2024 |
Deposit Date | Dec 12, 2024 |
Publicly Available Date | Dec 12, 2024 |
Publisher | Institute of Electrical and Electronics Engineers |
Peer Reviewed | Peer Reviewed |
Pages | 1-6 |
Book Title | 2024 IEEE International Conference on Multimedia and Expo (ICME) |
DOI | https://doi.org/10.1109/icme57554.2024.10687533 |
Public URL | https://durham-repository.worktribe.com/output/3215940 |
Files
Accepted Conference Paper
(3.9 Mb)
PDF
You might also like
CTNeRF: Cross-time Transformer for dynamic neural radiance field from monocular video
(2024)
Journal Article
DS-Depth: Dynamic and Static Depth Estimation via a Fusion Cost Volume
(2023)
Journal Article
EfficientTDNN: Efficient Architecture Search for Speaker Recognition
(2022)
Journal Article
Kernelized distance learning for zero-shot recognition
(2021)
Journal Article
A plug-in attribute correction module for generalized zero-shot learning
(2020)
Journal Article
Downloadable Citations
About Durham Research Online (DRO)
Administrator e-mail: dro.admin@durham.ac.uk
This application uses the following open-source libraries:
SheetJS Community Edition
Apache License Version 2.0 (http://www.apache.org/licenses/)
PDF.js
Apache License Version 2.0 (http://www.apache.org/licenses/)
Font Awesome
SIL OFL 1.1 (http://scripts.sil.org/OFL)
MIT License (http://opensource.org/licenses/mit-license.html)
CC BY 3.0 ( http://creativecommons.org/licenses/by/3.0/)
Powered by Worktribe © 2024
Advanced Search