基于双支路特征提取和语义引导的偏振图像融合网络
DOI:
CSTR:
作者:
作者单位:

长春理工大学电子信息工程学院长春130022

作者简介:

通讯作者:

中图分类号:

TN919.81;TP391.4

基金项目:

国家自然科学基金重大仪器专项项目(62127813)资助


Polarization image fusion based on dual-branch feature extraction and semantic guidance
Author:
Affiliation:

School of Electronic Information Engineering, Changchun University of Science and Technology, Changchun 130022,China

Fund Project:

  • 摘要
  • |
  • 图/表
  • |
  • 访问统计
  • |
  • 参考文献
  • |
  • 相似文献
  • |
  • 引证文献
  • |
  • 资源附件
  • |
  • 文章评论
    摘要:

    为解决当前偏振图像融合技术着重于融合结果的视觉质量与统计指标,忽略了融合图像在后续高级视觉任务中的应用这一问题,提出一种由语义分割引导的偏振图像双支路特征提取网络结构。融合网络包括编码器、融合层和解码器。在编码器中,构建2个由梯度残差密集块GRDB和SwinTransformer组成的双支路特征提取器,用于提取源图像的局部偏振特征与全局强度信息;在融合层内,采用可逆神经网络INN建立两类特征相关性,用以无损增强偏振特征并进行融合;在解码器中,使用Restormer作为基本单元,恢复和保留融合特征中的高频特征和场景细节,以提升图像清晰度并获得融合图像。为了使融合结果包含丰富的语义信息,本文在训练阶段将融合网络与分割网络级联,利用语义分割损失引导高级语义信息回流并指导融合网络训练,提高融合图像在高级视觉任务中的应用性能。实验结果表明,提出的融合网络,其融合结果在主观评价和语义分割任务中均取得最优值,并在客观评价指标中信息熵EN和结构相似性指数SSIM分别比其他融合方法提升了27%和16.8%。

    Abstract:

    To address the current limitation in polarization image fusion technology—where the focus is predominantly on the visual quality and statistical metrics of the fused output while neglecting its applicability to subsequent high-level vision tasks—this paper proposes a dual-branch feature extraction architecture for polarization image fusion, guided by semantic segmentation. The fusion network comprises an encoder, a fusion layer, and a decoder. In the encoder, a dual-branch feature extractor—composed of GRDB and Swin Transformers—is constructed to extract local polarization features and global intensity information from the source images. Within the fusion layer, an INN is employed to model the inter-feature correlations, enabling lossless enhancement and effective fusion of the polarization characteristics. In the decoder, Restormer serves as the core building block to reconstruct and preserve high-frequency details and structural scene information from the fused features, thereby enhancing image clarity and generating the final fused result. To enrich the fused output with task-relevant semantics, the fusion network is cascaded with a segmentation network during training. The semantic segmentation loss is leveraged to guide the backpropagation of high-level semantic information, thereby optimizing the fusion network and improving the utility of the fused images for advanced vision tasks. Experimental results demonstrate that the proposed network achieves superior performance in both subjective visual assessment and downstream semantic segmentation tasks. Moreover, it outperforms existing fusion methods in objective metrics, with notable improvements of 27% in EN and 16.8% in SSIM.

    参考文献
    相似文献
    引证文献
引用本文

陈广秋,代宇航,段锦,黄丹丹.基于双支路特征提取和语义引导的偏振图像融合网络[J].电子测量与仪器学报,2026,40(2):55-66

复制
分享
相关视频

文章指标
  • 点击次数:
  • 下载次数:
  • HTML阅读次数:
  • 引用次数:
历史
  • 收稿日期:
  • 最后修改日期:
  • 录用日期:
  • 在线发布日期: 2026-04-30
  • 出版日期:
文章二维码
×
《电子测量与仪器学报》
关于防范虚假编辑部邮件的郑重公告