留言板

尊敬的读者、作者、审稿人, 关于本刊的投稿、审稿、编辑和出版的任何问题, 您可以本页添加留言。我们将尽快给您答复。谢谢您的支持!

姓名
邮箱
手机号码
标题
留言内容
验证码

视觉显著性增强的双鉴别器红外与可见光图像融合

陈永 周方春 董珂

陈永,周方春,董珂. 视觉显著性增强的双鉴别器红外与可见光图像融合[J]. 北京航空航天大学学报,2026,52(4):1107-1115
引用本文: 陈永,周方春,董珂. 视觉显著性增强的双鉴别器红外与可见光图像融合[J]. 北京航空航天大学学报,2026,52(4):1107-1115
CHEN Y,ZHOU F C,DONG K. Dual discriminator fusion of infrared and visible light images for visual saliency enhancement[J]. Journal of Beijing University of Aeronautics and Astronautics,2026,52(4):1107-1115 (in Chinese)
Citation: CHEN Y,ZHOU F C,DONG K. Dual discriminator fusion of infrared and visible light images for visual saliency enhancement[J]. Journal of Beijing University of Aeronautics and Astronautics,2026,52(4):1107-1115 (in Chinese)

视觉显著性增强的双鉴别器红外与可见光图像融合

doi: 10.13700/j.bh.1001-5965.2024.0072
基金项目: 

国家自然科学基金(62462043,61963023);甘肃省自然科学基金(26JRRA589)

详细信息
    通讯作者:

    E-mail:edukeylab@126.com

  • 中图分类号: TP391

Dual discriminator fusion of infrared and visible light images for visual saliency enhancement

Funds: 

National Natural Science Foundation of China (62462043,61963023); Gansu Provincial Nature Science Foundation (26JRRA589)

More Information
  • 摘要:

    针对红外与可见光图像融合中边缘不清晰、细节缺失等问题,提出了一种视觉显著性增强的双鉴别器融合方法。采用局部自适应对可见光图像进行增强,并采用各向异性扩散对红外与可见光图像分解;通过视觉显著性检测对分解后的细节层图像和基础层图像进行视觉增强;设计密集连接DenseNet生成器模型对视觉增强后图像进行特征学习;通过与双鉴别器博弈对抗得到融合结果。在公开数据集中与10种融合方法进行对比,实验结果表明:所提方法具有更清晰的细节信息,在主客观评估上均优于对比方法,客观评价指标较FusionGAN方法在信息熵、空间频率、结构相似性和标准偏差上分别提高了7.4%、58.8%、25.5%和35.7%。

     

  • 图 1  网络总体框架

    Figure 1.  Overall network framework

    图 2  生成器结构框架

    Figure 2.  Network structure of generator

    图 3  图像分解层面和视觉显著性增强角度融合方法比较

    Figure 3.  Comparison of image decomposition and visual saliency enhancement fusion methods

    图 4  生成对抗与深度学习融合方法比较

    Figure 4.  Comparison of generative adversarial and deep learning fusion methods

    图 5  消融实验结果

    Figure 5.  Results of ablation experiments

    表  1  TNO客观对比值

    Table  1.   Objective comparison value of TNO

    方法 信息熵 互信息 SSIM 标准偏差 空间频率 视觉保真度
    ADF[10] 7.1617 0.6883 0.5030 38.4506 0.0318 0.7075
    VSMWLS[11] 7.1748 0.6590 0.6248 40.0635 0.0323 0.7401
    GFCE[13] 7.1613 0.6883 0.5030 36.4036 0.0436 0.7241
    MGF[12] 7.2553 0.6932 0.6123 41.847 0.0311 0.4924
    BEMD[23] 7.3241 0.7623 0.6538 48.0324 0.03539 0.5624
    GANMcC[20] 7.3960 0.7725 0.6053 42.5729 0.0395 0.5853
    DDcGAN[21] 6.6324 0.5674 0.4423 31.4367 0.0256 0.4640
    GAN-FM[22] 7.6654 0.4525 0.4365 52.5466 0.0383 0.7206
    DenseFuse[18] 7.1360 0.6770 0.5474 35.75 0.0309 0.6521
    FusionGAN[19] 7.3374 0.6424 0.5370 40.3546 0.0311 0.5305
    本文 7.8787 0.6938 0.6740 54.7573 0.0494 0.7358
    下载: 导出CSV

    表  2  消融实验客观指标

    Table  2.   Objective indicators of ablation experiment

    模块 SSIM 平均梯度
    GAN 0.6944 2.084
    ACE+GAN 0.7572 2.719
    AD+ACE+GAN 0.8140 3.104
    AD+ACE+DenseGAN 0.8386 3.592
    下载: 导出CSV

    表  3  计算量和运算时间

    Table  3.   Computational amount and processing time

    方法 浮点运算速度/109 s−1 运行时间/s
    ADF[10] 2.874 0.786
    VSMWLS[11] 3.024 0.793
    DenseFuse[18] 3.592 0.881
    FusionGAN[19] 3.648 0.867
    本文 3.153 0.853
    下载: 导出CSV
  • [1] 陈永, 王镇, 卢晨涛, 等. 红外弱光下多特征融合与注意力增强铁路异物检测[J]. 北京航空航天大学学报, 2023, 49(8): 1884-1894.

    CHEN Y, WANG Z, LU C T, et al. Detection of railway object intrusion under infrared low light based on multi-feature and attention enhancement network[J]. Journal of Beijing University of Aeronautics and Astronautics, 2023, 49(8): 1884-1894(in Chinese).
    [2] GONZÁLEZ J R, DAMIÃO C, MORAN M, et al. A computational study on the role of parameters for identification of thyroid nodules by infrared images (and comparison with real data)[J]. Sensors, 2021, 21(13): 4459.
    [3] MA J Y, MA Y, LI C. Infrared and visible image fusion methods and applications: a survey[J]. Information Fusion, 2019, 45: 153-178.
    [4] TU Z Z, LI Z, LI C L, et al. Multi-interactive dual-decoder for RGB-thermal salient object detection[J]. IEEE Transactions on Image Processing, 2021, 30: 5678-5691.
    [5] XIAO W X, ZHANG Y F, WANG H B, et al. Heterogeneous knowledge distillation for simultaneous infrared-visible image fusion and super-resolution[J]. IEEE Transactions on Instrumentation and Measurement, 2022, 71: 5004015.
    [6] DINH P H. Combining Gabor energy with equilibrium optimizer algorithm for multi-modality medical image fusion[J]. Biomedical Signal Processing and Control, 2021, 68: 102696.
    [7] 陈永, 张娇娇, 王镇. 多尺度密集连接注意力的红外与可见光图像融合[J]. 光学精密工程, 2022, 30(18): 2253-2266.

    CHEN Y, ZHANG J J, WANG Z. Infrared and visible image fusion based on multi-scale dense attention connection network[J]. Optics and Precision Engineering, 2022, 30(18): 2253-2266(in Chinese).
    [8] WANG Z S, XU J W, JIANG X L, et al. Infrared and visible image fusion via hybrid decomposition of NSCT and morphological sequential toggle operator[J]. Optik, 2020, 201: 163497.
    [9] 娄熙承, 冯鑫. 潜在低秩表示框架下基于卷积神经网络结合引导滤波的红外与可见光图像融合[J]. 光子学报, 2021, 50(3): 0310004.

    LOU X C, FENG X. Infrared and visible image fusion in latent low rank representation framework based on convolution neural network and guided filtering[J]. Acta Photonica Sinica, 2021, 50(3): 0310004(in Chinese).
    [10] VASU G T, PALANISAMY P. Multi-focus image fusion using anisotropic diffusion filter[J]. Soft Computing, 2022, 26(24): 14029-14040.
    [11] MA J L, ZHOU Z Q, WANG B, et al. Infrared and visible image fusion based on visual saliency map and weighted least square optimization[J]. Infrared Physics & Technology, 2017, 82: 8-17.
    [12] BAVIRISETTI D P, XIAO G, ZHAO J H, et al. Multi-scale guided image and video fusion: a fast and efficient approach[J]. Circuits, Systems, and Signal Processing, 2019, 38(12): 5576-5605.
    [13] ZHOU Z Q, DONG M J, XIE X Z, et al. Fusion of infrared and visible images for night-vision context enhancement[J]. Applied Optics, 2016, 55(23): 6480-6490.
    [14] SUN C Q, ZHANG C, XIONG N X. Infrared and visible image fusion techniques based on deep learning: a review[J]. Electronics, 2020, 9(12): 2162.
    [15] LI J, ZHU J M, LI C, et al. CGTF: convolution-guided Transformer for infrared and visible image fusion[J]. IEEE Transactions on Instrumentation and Measurement, 2022, 71: 5012314.
    [16] WANG Z S, WANG J Y, WU Y Y, et al. UNFusion: a unified multi-scale densely connected network for infrared and visible image fusion[J]. IEEE Transactions on Circuits and Systems for Video Technology, 2022, 32(6): 3360-3374.
    [17] LI H, WU X J, KITTLER J. RFN-Nest: an end-to-end residual fusion network for infrared and visible images[J]. Information Fusion, 2021, 73: 72-86.
    [18] LI H, WU X J. DenseFuse: a fusion approach to infrared and visible images[J]. IEEE Transactions on Image Processing, 2019, 28(5): 2614-2623.
    [19] MA J Y, YU W, LIANG P W, et al. FusionGAN: a generative adversarial network for infrared and visible image fusion[J]. Information Fusion, 2019, 48: 11-26.
    [20] MA J Y, ZHANG H, SHAO Z F, et al. GANMcC: a generative adversarial network with multiclassification constraints for infrared and visible image fusion[J]. IEEE Transactions on Instrumentation and Measurement, 2020, 70: 5005014.
    [21] MA J Y, XU H, JIANG J J, et al. DDcGAN: a dual-discriminator conditional generative adversarial network for multi-resolution image fusion[J]. IEEE Transactions on Image Processing, 2020, 29: 4980-4995.
    [22] ZHANG H, YUAN J T, TIAN X, et al. GAN-FM: infrared and visible image fusion using GAN with full-scale skip connection and dual Markovian discriminators[J]. IEEE Transactions on Computational Imaging, 2021, 7: 1134-1147.
    [23] 崔晓荣, 沈涛, 黄建鲁, 等. 基于BEMD改进的视觉显著性红外和可见光图像融合[J]. 红外技术, 2020, 42(11): 1061-1071.

    CUI X R, SHEN T, HUANG J L, et al. Infrared and visible image fusion based on BEMD and improved visual saliency[J]. Infrared Technology, 2022, 42(11): 1061-1071(in Chinese).
    [24] YE X Y, GAO S L, LI F M. ACE-STDN: an infrared small target detection network with adaptive contrast enhancement[J]. Journal of Infrared Millim Waves, 2023, 42(5): 701-708.
    [25] JOO M G, LEE K T, SANG P, et al. Laser-generated focused ultrasound transmitters with frequency-tuned outputs over sub-10-MHz range[J]. Applied Physics Letters, 2019, 115(15): 154103.
  • 加载中
图(5) / 表(3)
计量
  • 文章访问数:  302
  • HTML全文浏览量:  142
  • PDF下载量:  13
  • 被引次数: 0
出版历程
  • 收稿日期:  2024-01-31
  • 录用日期:  2024-05-28
  • 网络出版日期:  2024-08-29
  • 整期出版日期:  2026-04-30

目录

    /

    返回文章
    返回
    常见问答