首页|基于信息增强和掩码损失的红外与可见光图像融合方法

基于信息增强和掩码损失的红外与可见光图像融合方法

扫码查看
针对低光场景下红外与可见光融合图像中存在的细节弱化和边缘模糊等问题,提出一种基于信息增强和掩码损失的图像融合方法。首先,采用引导滤波增强可见光图像的纹理细节和红外图像的边缘梯度;其次,构建双分支特征提取网络提取不同模态图像的特征信息,并设计交互增强模块以渐进交互的方式集成不同特征分支的互补信息,增强特征的细节表示;然后,在融合阶段设计注意力引导模块从空间和通道维度上关注特征信息,提升网络对关键特征的感知能力;最后,提出一种掩码损失以指导融合网络有针对性地保留源图像信息,提升融合质量。为验证所提方法的融合性能,在MSRS、TNO和LLVIP公开数据集上与9种主流的融合算法进行实验对比。结果表明,所提方法在定性和定量评估上均优于其它对比算法,生成的融合图像具有丰富的纹理细节、清晰的显著性目标和良好的视觉感知。
Infrared and Visible Image Fusion Method Based on Information Enhancement and Mask Loss
Aiming at the problems of detail weakening and edge blurring in infrared and visible fusion images in low-light scenes,an image fusion method based on information enhancement and mask loss is proposed.Firstly,considering the information degradation of the source image in the low-light scene,the guided filtering is employed to enhance the texture details of the visible image and the edge information of the infrared image before fusion.Secondly,in order to fully extract and effectively fuse the feature information of different modal images,a two-branch network is constructed to extract image features.Based on the dual-branch feature extraction network,an interactive enhancement module based on guided filtering is designed to integrate the complementary information of different feature branches in a progressive interactive way to enhance the detail representation of texture and salient targets in the features.During the fusion stage,an attention guidance module based on spatial and channel dimensions is constructed.In the attention mechanism,the maximum and average scales are used to focus on the feature information.By combining the attention guidance of different dimensions and different scales,the key information in the feature is amplified and the redundant information is filtered out to improve the perceptual capacity of the network for crucial features.In terms of loss function,a method of generating infrared mask is proposed,and a mask loss is designed based on infrared mask to guide the fusion network to retain salient targets,texture details and structure information in the target and background regions.In addition,in the training phase,in order to improve the adaptability of the fusion network and reduce the risk of over-fitting,1 083 pairs of images in the MSRS dataset are selected for clipping and expansion.The clipping size is 120 × 120,and the moving step size is 120.The obtained 21 660 pairs of image blocks are used as the training data of the model.In the test phase,to comprehensively evaluate the fusion performance of the method,comparative experiments are performed on three public datasets:MSRS,TNO,and LLVIP.This paper selects nine state-of-the-art fusion methods for qualitative comparison,including CNN-based SDNet,GAN-based FusionGAN and GANMcC,AE-based DenseFuse,RFN-Nest,and PIAFusion,visual tasks-based SeaFusion and IRFS,as well as Transformer-based SwinFusion.Five evaluation indexes are selected for quantitative comparison,namely information entropy,spatial frequency,average gradient,standard deviation,and visual fidelity.The experimental results show that the proposed method is superior to other comparison algorithms in both qualitative and quantitative evaluation on three public datasets.The generated fusion image exhibits rich texture details,clear saliency targets,and excellent visual perception.Finally,in order to verify the effectiveness of the proposed module,ablation experiments are conducted on the image pre-enhancement processing,interactive enhancement module,mask loss and attention guidance module,respectively.The qualitative and quantitative comparison results of the ablation experiments confirm the effectiveness of the proposed module in the fusion algorithm.

Image fusionInformation enhancementInfrared maskGuided filterAttention guidance

张晓东、王硕、高绍姝、王鑫瑞、张龙

展开 >

中国石油大学(华东)计算机科学与技术学院,青岛 226580

图像融合 信息增强 红外掩码 引导滤波 注意力引导

2024

光子学报
中国光学学会 中国科学院西安光学精密机械研究所

光子学报

CSTPCD北大核心
影响因子:0.948
ISSN:1004-4213
年,卷(期):2024.53(9)