Author
Listed:
- Xiaotong Guo
(School of Instrumentation and Optoelectronic Engineering, Key Laboratory of Precision Opto-Mechatronics Technology, Ministry of Education, Beihang University, Beijing 100191, China
Qingdao Research Institute, Beihang University, Qingdao 266104, China)
- Huijie Zhao
(School of Artificial Intelligence, Beihang University, Beijing 100191, China
Aerospace Optical-Microwave Integrated Precision Intelligent Sensing, Key Laboratory of Ministry of Industry and Information Technology, Beihang University, Beijing 100191, China)
- Shuwei Shao
(School of Automation Science and Electrical Engineering, Beihang University, Beijing 100191, China)
- Xudong Li
(School of Instrumentation and Optoelectronic Engineering, Key Laboratory of Precision Opto-Mechatronics Technology, Ministry of Education, Beihang University, Beijing 100191, China
Qingdao Research Institute, Beihang University, Qingdao 266104, China)
- Baochang Zhang
(School of Artificial Intelligence, Beihang University, Beijing 100191, China
Hangzhou Research Institute, Beihang University, Hangzhou 310051, China
Nanchang Institute of Technology, Nanchang 330044, China)
- Na Li
(School of Artificial Intelligence, Beihang University, Beijing 100191, China
Aerospace Optical-Microwave Integrated Precision Intelligent Sensing, Key Laboratory of Ministry of Industry and Information Technology, Beihang University, Beijing 100191, China)
Abstract
Due to the existence of low-textured areas in indoor scenes, some self-supervised depth estimation methods have specifically designed sparse photometric consistency losses and geometry-based losses. However, some of the loss terms cannot supervise all the pixels, which limits the performance of these methods. Some approaches introduce an additional optical flow network to provide dense correspondences supervision, but overload the loss function. In this paper, we propose to perform depth self-propagation based on feature self-similarities, where high-accuracy depths are propagated from supervised pixels to unsupervised ones. The enhanced self-supervised indoor monocular depth estimation network is called SPDepth. Since depth self-similarities are significant in a local range, a local window self-attention module is embedded at the end of the network to propagate depths in a window. The depth of a pixel is weighted using the feature correlation scores with other pixels in the same window. The effectiveness of self-propagation mechanism is demonstrated in the experiments on the NYU Depth V2 dataset. The root-mean-squared error of SPDepth is 0.585 and the δ 1 accuracy is 77.6%. Zero-shot generalization studies are also conducted on the 7-Scenes dataset and provide a more comprehensive analysis about the application characteristics of SPDepth.
Suggested Citation
Xiaotong Guo & Huijie Zhao & Shuwei Shao & Xudong Li & Baochang Zhang & Na Li, 2024.
"SPDepth: Enhancing Self-Supervised Indoor Monocular Depth Estimation via Self-Propagation,"
Future Internet, MDPI, vol. 16(10), pages 1-14, October.
Handle:
RePEc:gam:jftint:v:16:y:2024:i:10:p:375-:d:1500088
Download full text from publisher
Corrections
All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:gam:jftint:v:16:y:2024:i:10:p:375-:d:1500088. See general information about how to correct material in RePEc.
If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.
We have no bibliographic references for this item. You can help adding them by using this form .
If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.
For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: MDPI Indexing Manager (email available below). General contact details of provider: https://www.mdpi.com .
Please note that corrections may take a couple of weeks to filter through
the various RePEc services.