结合语义分割与跨模态差分特征补偿的红外与可见光图像融合方法
DOI:
作者:
作者单位:

1.北京信息科技大学;2.北京工业大学

作者简介:

通讯作者:

中图分类号:

基金项目:

北京市教委科技计划一般项目(KM202011232007)、高校学科人才引进计划(D17021)、北京信息科技内涵发展项目(2019KYNH204)资助


An Infrared and Visible Image Fusion Method Integrating Semantic Segmentation and Cross-Modality Differential Feature Compensation
Author:
Affiliation:

Fund Project:

  • 摘要
  • |
  • 图/表
  • |
  • 访问统计
  • |
  • 参考文献
  • |
  • 相似文献
  • |
  • 引证文献
  • |
  • 资源附件
  • |
  • 文章评论
    摘要:

    针对现有红外与可见光图像融合模型在深层特征提取时细节信息丢失、显著目标轮廓模糊的问题,提出一种结合语义分割与跨模态差分特征补偿 (Cross-Modality Differential Feature Compensation, CMDFC) 的红外与可见光图像融合方法。通过具有卷积注意力机制 (Convolutional Block Attention Module, CBAM) 的跨模态差分特征补偿模块,叠加不同模态的互补特征信息至原始特征中进行深层特征提取,引入语义分割网络对融合图像进行像素级别的分类操作构造语义损失来约束融合网络,并使用解码器重构融合图像。在公开数据集上进行融合实验的结果表明,相较于对照模型中的最优指标,所选的五种指标均有不同程度的提高,其中互信息 (Mutual Information, MI) 和视觉信息保真度 (Visual Information Fidelity, VIF) 分别提高了4.41%和4.25%,说明本文所提出的模型生成的融合图像更清晰,与源图像相关性更强,该方法有效缓解了红外与可见光图像融合过程中特征细节信息丢失的问题,增强了生成图像的视觉效果和对比度。

    Abstract:

    To address the issues of detail information loss and blurred salient target contours in existing infrared and visible image fusion models during deep feature extraction, we propose an infrared and visible image fusion method that combines semantic segmentation with Cross-Modality Differential Feature Compensation (CMDFC). By incorporating a cross-modality differential feature compensation module with a Convolutional Block Attention Module (CBAM), complementary features from different modalities are integrated into the original features for deep feature extraction. Additionally, a semantic segmentation network is introduced to perform pixel-level classification on the fused image, constructing a semantic loss to constrain the fusion network, and a decoder is used to reconstruct the fused image. Experimental results on public datasets show that compared to the best metrics of the reference models, the proposed model achieves various degrees of improvement in five selected metrics, with Mutual Information (MI) and Visual Information Fidelity (VIF) increased of 4.41% and 4.25%, respectively. These results indicate that the proposed model generates clearer fused images with stronger correlation to the source images, effectively mitigating the issue of feature detail loss during the fusion process and enhancing the visual quality and contrast of the generated images.

    参考文献
    相似文献
    引证文献
引用本文
分享
文章指标
  • 点击次数:
  • 下载次数:
  • HTML阅读次数:
  • 引用次数:
历史
  • 收稿日期:2024-01-12
  • 最后修改日期:2024-05-16
  • 录用日期:2024-05-17
  • 在线发布日期:
  • 出版日期: