Atmospheric turbulence in long-range imaging significantly degrades the quality and fidelity of captured scenes due to random variations in both spatial and temporal dimensions. These distortions present a formidable challenge across various applications, from surveillance to astronomy, necessitating robust mitigation strategies. While model-based approaches achieve good results, they are very slow. Deep learning approaches show promise in image and video restoration but have struggled to address these spatiotemporal variant distortions effectively. This paper proposes a new framework that combines geometric restoration with an enhancement module. Random perturbations and geometric distortion are removed using a pyramid architecture with deformable 3D convolutions, resulting in aligned frames. These frames are then used to reconstruct a sharp, clear image via a multi-scale architecture of 3D Swin Transformers. The proposed framework demonstrates superior performance over the state of the art for both synthetic and real atmospheric turbulence effects, with reasonable speed and model size.
翻译:远距离成像中的大气湍流由于空间和时间维度的随机变化,会显著降低捕获场景的质量与保真度。这些畸变给从监控到天文学等多种应用带来了严峻挑战,亟需鲁棒的抑制策略。基于模型的方法虽能取得良好效果,但处理速度极慢。深度学习方法在图像与视频复原领域展现出潜力,却难以有效处理此类时空变化的畸变。本文提出一种结合几何复原与增强模块的新框架。通过采用具有可变形3D卷积的金字塔结构消除随机扰动与几何畸变,实现帧对齐。随后利用这些对齐帧,通过多尺度3D Swin Transformers架构重建出清晰锐利的图像。所提框架在合成与真实大气湍流效应上均展现出超越现有技术的性能,同时具备合理的速度与模型规模。