首页|Robust multi-sensor image matching based on normalized self-similarity region descriptor
Robust multi-sensor image matching based on normalized self-similarity region descriptor
扫码查看
点击上方二维码区域,可以放大扫码查看
原文链接
NETL
NSTL
万方数据
维普
Multi-modal image matching is crucial in aerospace applications because it can fully exploit the complementary and valuable information contained in the amount and diversity of remote sensing images.However,it remains a challenging task due to significant non-linear radio-metric,geometric differences,and noise across different sensors.To improve the performance of heterologous image matching,this paper proposes a normalized self-similarity region descriptor to extract consistent structural information.We first construct the pointwise self-similarity region descriptor based on the Euclidean distance between adjacent image blocks to reflect the structural properties of multi-modal images.Then,a linear normalization approach is used to form Modality Independent Region Descriptor(MIRD),which can effectively distinguish structural features such as points,lines,corners,and flat between multi-modal images.To further improve the matching accuracy,the included angle cosine similarity metric is adopted to exploit the directional vector information of multi-dimensional feature descriptors.The experimental results show that the pro-posed MIRD has better matching accuracy and robustness for various multi-modal image matching than the state-of-the-art methods.MIRD can effectively extract consistent geometric structure fea-tures and suppress the influence of SAR speckle noise using non-local neighboring image blocks operation,effectively applied to various multi-modal image matching.