We present CD-NGP, which is a fast and scalable representation for 3D reconstruction and novel view synthesis in dynamic scenes. Inspired by continual learning, our method first segments input videos into multiple chunks, followed by training the model chunk by chunk, and finally, fuses features of the first branch and subsequent branches. Experiments on the prevailing DyNeRF dataset demonstrate that our proposed novel representation reaches a great balance between memory consumption, model size, training speed, and rendering quality. Specifically, our method consumes $85\%$ less training memory ($<14$GB) than offline methods and requires significantly lower streaming bandwidth ($<0.4$MB/frame) than other online alternatives.
翻译:本文提出CD-NGP,这是一种面向动态场景三维重建与新视角合成的快速可扩展表征方法。受持续学习启发,本方法首先将输入视频分割为多个片段,随后逐片段训练模型,最终融合首分支与后续分支的特征。在主流DyNeRF数据集上的实验表明,我们提出的新型表征在内存消耗、模型规模、训练速度与渲染质量之间达到了良好平衡。具体而言,本方法相比离线方法减少85%训练内存(<14GB),相比其他在线方案显著降低传输带宽需求(<0.4MB/帧)。