欢迎访问中国科学院大学学报,今天是

中国科学院大学学报

• • 上一篇    下一篇

注意力引导的光学-SAR多模态互补信息融合去云方法*

吴浩田, 郭擎   

  1. 中国科学院空天信息创新研究院, 北京 100094;
    中国科学院大学 电子电气与通信工程学院, 北京 100049
  • 收稿日期:2026-02-12 修回日期:2026-03-25
  • 通讯作者: †E-mail: guoqing@aircas.ac.cn
  • 基金资助:
    * 国家自然科学基金(61771470 )资助

Cloud removal method based on attention-guided optical-SAR multimodal complementary information fusion

WU Haotian, GUO Qing   

  1. Aerospace Information Research Institute,Chinese Academy of Sciences,Beijing 100094,China;
    School of Electronic,Electrical and Communication Engineering,University of Chinese Academy of Sciences,Beijing 100049,China
  • Received:2026-02-12 Revised:2026-03-25

摘要: 光学遥感成像容易受到云的干扰,使得遥感图像中部分信息改变或丢失,降低数据的可用性。但合成孔径雷达(SAR)可以全天时全天候成像,围绕多模态遥感图像在复杂场景下的高质量利用需求,本文开展光学与SAR互补信息融合的去云方法研究。构建一种注意力引导的多模态融合模型,通过门控卷积结构与多层注意力机制协同建模,实现多尺度特征提取与跨模态全局依赖关系刻画,增强特征对齐与信息交互能力。利用交叉注意力引导SAR模态弥补云遮挡缺失的信息,并引入多模态去云单元整合深度特征,有效抑制云并强化地物表达,重建无云光学图像。实验结果表明,本文方法在去云精度、图像细节保持及结构一致性方面均取得明显提升,验证了多模态互补融合在遥感图像去云任务中的有效性。

关键词: 光学遥感图像, 图像去云, 合成孔径雷达, 注意力, 多模态融合

Abstract: Optical remote sensing imaging is highly susceptible to clouds, which leads to partial information degradation or loss and significantly limits data availability. In contrast, synthetic aperture radar (SAR) is capable of all-weather and all-day imaging, providing stable structural information unaffected by cloud cover. To address the demand for high-quality utilization of multimodal remote sensing data in complex cloud-covered scenarios, this paper investigates an optical-SAR complementary information fusion approach for cloud removal. An attention-guided multimodal fusion framework is proposed, in which gated convolutional structures and multi-level attention mechanisms are jointly employed to enable multi-scale feature extraction and global cross-modal dependency modeling, thereby enhancing feature alignment and information interaction between optical and SAR modalities. Specifically, a cross-attention mechanism is introduced to guide SAR features in compensating for the information missing in cloud-covered optical regions. Furthermore, a multimodal cloud removal unit is designed to integrate deep features and map them back to the image space, effectively suppressing cloud artifacts while strengthening ground object representation to reconstruct cloud-free optical images. Experimental results demonstrate that the proposed method achieves notable improvements in cloud removal accuracy, detail preservation, and structural consistency compared with existing methods, and validates the effectiveness of multimodal complementary fusion for optical remote sensing image cloud removal.

Key words: optical remote sensing images, cloud removal, SAR, attention mechanism, multimodal fusion

中图分类号: