首页|结合双视通路与尺度信息融合的轮廓检测方法

结合双视通路与尺度信息融合的轮廓检测方法

扫码查看
目的 考虑到图像信息在视觉通路中的表征是多尺度的,为了实现自然场景下多对比度分布图像的轮廓检测任务,提出了一种基于双视通路尺度信息融合的轮廓检测新方法.方法 首先构建对亮度信息敏感和对颜色信息敏感的大细胞(M)、小细胞(P)并行通路,构建不同尺度的感受野模拟神经节细胞对刺激的模糊和精细感知,使用亮度对比度和色差信息指导不同尺度感受野响应的自适应融合,使其能够充分提取亮度轮廓和颜色轮廓.其次结合外膝体(lateral geniculate nucleus,LGN)多尺度方向差异编码与多尺度朝向选择性抑制方法,构建显著轮廓提取模型,实现轮廓区域的增强以及背景纹理的抑制.最后将加工后的亮度轮廓和颜色轮廓前馈至初级视皮层(V1)区,构建双通道响应权重调节模型整合M、P通路所得信息,进一步丰富轮廓.结果 本文使用BSDS500(berkeley segmentation data set)图像库和NYUD(New York University-depth)图像库对提出的算法进行验证,其中在BSDS500图像库的最优平均准确率(average precision,AP)指标为 0.74,相对于 SCSI(subfield-based center-surround inhibition)、BAR(bilateral asymmetric receptive)和 SED(surround-modulated edge detection)等基于生物视觉机制的检测方法有4%~13%的提升,所得结果轮廓图也更为连续、准确.结论 本文利用M、P双通路机制以及亮度信息和颜色信息在前端视觉通路中的编码过程实现轮廓信息的加工与提取,可以有效实现自然图像的轮廓检测,尤其是对于图像中的细微轮廓边缘的检测,也为研究更高级皮层中视觉信息机制提供新的思路.
A contour detection method combing dual visual pathway and scale information fusion
Objective The extraction and utilization of contour information,as a low-level visual feature of the target sub-ject,contribute to the efficient execution of advanced visual tasks,such as object detection and image segmentation.When processing complex images,contour detection based on biological vision mechanisms can quickly extract object contour information.However,the perception of primary contour information is currently based on a single scale receptive field template or a simple fusion of multiple scale receptive field templates,which ignores the dynamic characteristics of recep-tive field scales and makes it difficult to accurately extract contours in complex scenes.Considering the serial parallel trans-mission and integration mechanism of visual information in the magnocellular(M)and parvocellular(P)dual vision path-ways,we propose a new contour detection method based on the fusion of dual vision pathway scale information.Method First,we introduce Lab,a color system that is close to human visual physiological characteristics,to extract color differ-ence and brightness information from an image.Compared with conventional RGB color systems,Lab is more in line with the way the human eye perceives visual information.Considering that the scale of the receptive field of ganglion cells varies with the size of local stimuli to adapt to different visual task requirements across various scenes,a smaller scale of the receptive field corresponds to a more refined perception of detailed information.We then simulate the fuzzy and fine percep-tion of the stimuli by ganglion cells using two different scale receptive fields,and we use color difference and brightness contrast information to guide the adaptive fusion of large-and small-scale receptive field responses and highlight the contour details.Second,considering the differences in the perception of orientation information among receptive fields at different scales of the lateral geniculate body,we introduce the standard deviation of the optimal orientation obtained from percep-tion at multiple scales as the encoding weight for the direction difference,thereby achieving a modulation of texture region suppression weight information.We also combine local contrast information to guide the lateral inhibition intensity of non-classical receptive fields based on the difference between the central and peripheral directions.Through the collaborative integration of these two,we successfully enhance the contour regions and suppress the background textures.Finally,to simulate the complementary fusion mechanism of color and brightness information in the primary visual cortex(V1)region,we propose a weight association model integrating contrast information.Based on the fusion weight coefficients obtained from the local color contrast and brightness contrast,we achieve a complementary fusion of information flows in the M and P paths,thereby enriching the contour details.Result We compared our model with three biological-vision-based mecha-nisms(SCSI,SED,and BAR)and one deep-learning-based model(PiDiNet).On the BSDS500 dataset,we used several quantitative evaluation indicators,including optimal dataset scale(ODS),optimal image scale(OIS),average precision(AP)indicators,and precision-recall(PR)curves,and selected five images to compare the detection performance of each method.Experimental results show that our model has a better overall performance than the other models.Compared with SCSI,SED,and BAR,our model obtains 4.45%,2.94%,and 4.45%higher ODS index,2.82%,5.80%,and 8.96%higher OIS index,and 7.25%,4.23%,and 5.71%higher AP index,respectively.While the PiDiNet model based on deep learning has some shortcomings compared with various indicators,this model does not require a pre-training of data,has biological interpretability,and has a small computational power requirement.We further extracted four images from the NYUD dataset to visually compared the false detection rate,missed detection rate,and overall performance of the models.We also conducted a series of ablation experiments to demonstrate the contribution of each module in the model to its over-all performance.Conclusion In this paper,we use the M and P dual-path mechanism and the encoding process of lumi-nance and color information in the front-end visual path to realize contour information processing and extraction.Our pro-posed approach can effectively realize a contour detection of natural images,especially for subtle contour edge detection in images,and provide novel insights for studying visual information mechanism in the higher-level cortex.

countour detectiondual visual pathwaymulti-scaleadaptive fusiondirectional difference

杜仕荣、范影乐、蔡哲飞、房涛

展开 >

杭州电子科技大学模式识别与图像处理实验室,杭州 310018

杭州电子科技大学电子信息学院,杭州 310018

轮廓检测 双视通路 多尺度 自适应融合 方向差异

2024

中国图象图形学报
中国科学院遥感应用研究所,中国图象图形学学会 ,北京应用物理与计算数学研究所

中国图象图形学报

CSTPCD北大核心
影响因子:1.111
ISSN:1006-8961
年,卷(期):2024.29(12)
  • 1