中国航空学报(英文版)2024,Vol.37Issue(7) :439-449.DOI:10.1016/j.cja.2024.03.007

Filtering and regret network for spacecraft component segmentation based on gray images and depth maps

Xiang LIU Hongyuan WANG Zijian WANG Xinlong CHEN Weichun CHEN Zhengyou XIE
中国航空学报(英文版)2024,Vol.37Issue(7) :439-449.DOI:10.1016/j.cja.2024.03.007

Filtering and regret network for spacecraft component segmentation based on gray images and depth maps

Xiang LIU 1Hongyuan WANG 1Zijian WANG 1Xinlong CHEN 2Weichun CHEN 2Zhengyou XIE2
扫码查看

作者信息

  • 1. Institute of Aeronautics,Harbin Institute of Technology,Harbin 150001,China
  • 2. Qian Xuesen Laboratory of Space Technology,China Academy of Space Technology,Beijing 100080,China
  • 折叠

Abstract

Identifying and segmenting spacecraft components is vital in many on-orbit space mis-sions,such as on-orbit maintenance and component recovery.Integrating depth maps with visual images has been proven effective in improving segmentation accuracy.However,existing methods ignore the noise and fallacy in collected depth maps,which interfere with the network to extract representative features,decreasing the final segmentation accuracy.To this end,this paper proposes a Filtering and Regret Network(FRNet)for spacecraft component segmentation.The FRNet incorporates filtering and regret mechanisms to suppress the abnormal depth response in shallow layers and selectively reuses the filtered cues in deep layers,avoiding the detrimental effects of low-quality depth information while preserving the semantic context inherent in depth maps.Fur-thermore,a two-stage feature fusion module is proposed,which involves information interaction and aggregation.This module effectively explores the feature correlation and unifies the multi-modal features into a comprehensive representation.Finally,a large-scale spacecraft component recognition dataset is constructed for training and evaluating spacecraft component segmentation algorithms.Experimental results demonstrate that the FRNet achieves a state-of-the-art perfor-mance with a mean Intersection Over Union(mIOU)of 84.13%and an average inference time of 133.2 ms when tested on an NVIDIA RTX 2080 SUPER GPU.

Key words

Spacecraft component recognition/Multi-modal feature fusion/Satellite dataset/Intelligent systems/Deep learning

引用本文复制引用

出版年

2024
中国航空学报(英文版)
中国航空学会

中国航空学报(英文版)

CSTPCDEI
影响因子:0.847
ISSN:1000-9361
段落导航相关论文