Xingyu Miao xingyu.miao@durham.ac.uk
PGR Student Doctor of Philosophy
CTNeRF: Cross-time Transformer for dynamic neural radiance field from monocular video
Miao, Xingyu; Bai, Yang; Duan, Haoran; Wan, Fan; Huang, Yawen; Long, Yang; Zheng, Yefeng
Authors
Yang Bai
Haoran Duan haoran.duan@durham.ac.uk
PGR Student Doctor of Philosophy
Fan Wan fan.wan@durham.ac.uk
PGR Student Doctor of Philosophy
Yawen Huang
Dr Yang Long yang.long@durham.ac.uk
Associate Professor
Yefeng Zheng
Abstract
The goal of our work is to generate high-quality novel views from monocular videos of complex and dynamic scenes. Prior methods, such as DynamicNeRF, have shown impressive performance by leveraging time-varying dynamic radiation fields. However, these methods have limitations when it comes to accurately modeling the motion of complex objects, which can lead to inaccurate and blurry renderings of details. To address this limitation, we propose a novel approach that builds upon a recent generalization NeRF, which aggregates nearby views onto new viewpoints. However, such methods are typically only effective for static scenes. To overcome this challenge, we introduce a module that operates in both the time and frequency domains to aggregate the features of object motion. This allows us to learn the relationship between frames and generate higher-quality images. Our experiments demonstrate significant improvements over state-of-the-art methods on dynamic scene datasets. Specifically, our approach outperforms existing methods in terms of both the accuracy and visual quality of the synthesized views. Our code is available on https://github.com/xingy038/CTNeRF.
Citation
Miao, X., Bai, Y., Duan, H., Wan, F., Huang, Y., Long, Y., & Zheng, Y. (2024). CTNeRF: Cross-time Transformer for dynamic neural radiance field from monocular video. Pattern Recognition, 156, Article 110729. https://doi.org/10.1016/j.patcog.2024.110729
Journal Article Type | Article |
---|---|
Acceptance Date | Jun 24, 2024 |
Online Publication Date | Jul 1, 2024 |
Publication Date | 2024-12 |
Deposit Date | Jul 31, 2024 |
Publicly Available Date | Aug 1, 2024 |
Journal | Pattern Recognition |
Print ISSN | 0031-3203 |
Publisher | Elsevier |
Peer Reviewed | Peer Reviewed |
Volume | 156 |
Article Number | 110729 |
DOI | https://doi.org/10.1016/j.patcog.2024.110729 |
Public URL | https://durham-repository.worktribe.com/output/2641933 |
Files
Accepted Journal Article
(38.1 Mb)
PDF
You might also like
DS-Depth: Dynamic and Static Depth Estimation via a Fusion Cost Volume
(2023)
Journal Article
EfficientTDNN: Efficient Architecture Search for Speaker Recognition
(2022)
Journal Article
Wearable-based behaviour interpolation for semi-supervised human activity recognition
(2024)
Journal Article
Dynamic Unary Convolution in Transformers
(2023)
Journal Article
Downloadable Citations
About Durham Research Online (DRO)
Administrator e-mail: dro.admin@durham.ac.uk
This application uses the following open-source libraries:
SheetJS Community Edition
Apache License Version 2.0 (http://www.apache.org/licenses/)
PDF.js
Apache License Version 2.0 (http://www.apache.org/licenses/)
Font Awesome
SIL OFL 1.1 (http://scripts.sil.org/OFL)
MIT License (http://opensource.org/licenses/mit-license.html)
CC BY 3.0 ( http://creativecommons.org/licenses/by/3.0/)
Powered by Worktribe © 2024
Advanced Search