Mobile monocular depth estimation based on multi-scale feature fusion
The current depth estimation model based on depth learning has a large number of param-eters,which is difficult to adapt to mobile devices.To address this issue,a lightweight depth estimation method with multi-scale feature fusion that can be deployed on mobile devices is proposed.Firstly,Mo-bileNetV2 is used as the backbone to extract features of four scales.Then,by constructing skip connec-tion paths from the encoder to the decoder,the features of the four scales are fused,fully utilizing the combined positional information from lower layers and semantic information from higher layers.Final-ly,the fused features are processed through convolutional layers to produce high-precision depth images.After training and testing on NYU Depth Dataset V2,the experimental results show that the proposed model achieves advanced performance with an evaluation index of δ1 up to 0.812 while only having 1.6×106 parameters numbers.Additionally,it only takes 0.094 seconds to infer a single image on the Kirin 980 CPU of a mobile device,demonstrating its practical application value.
deep learningdepth estimationmulti-scale featurelightweight networkmobile terminal model