Depth Estimation and Dense Reconstruction with the Monocular Camera (Tutor: Tianjun Zhang, wechat: z619850002)

Introduction

3D reconstruction based on the monocular vision is a classic task in the field of computer vision. In this project, we will make a simplification: camera poses corresponding to each frame will be known, so only the depth estimation and the dense map construction need to be focused.

https://img-blog.csdnimg.cn/20200920003021642.png?x-oss-process=image/watermark,type_ZmFuZ3poZW5naGVpdGk,shadow_10,text_aHR0cHM6Ly9ibG9nLmNzZG4ubmV0L2h4eGp4dw==,size_16,color_FFFFFF,t_70

                                                    Fig. 1: Depth Estimation

 

https://www.pianshen.com/images/534/98df7c946aaca511341da07a5b6d2c1e.png

                                             Fig. 2: Dense Scene Reconstruction

 

   Fig. 3: Scene structure reconstruction incrementally using a single agent

 

Requirements

1)  The dataset, which is actually a video sequence of images captured by a monocular camera and corresponding camera poses, will be provided. Depth maps of key frames should be firstly recovered by any proper algorithms. Then the dense map of the scene can be constructed incrementally.

2) When the video stream is input, your system should be able to construct a dense map in real-time or quasi real-time, rather than offline. GPU is allowed.

3) We will offer a depth-estimation system as a baseline. You can both modify it or implement your own system.
 

Reference Materials

  1. Traditional depth estimation method, SGM: https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=4359315

  2. Deep learning based method, DPT: https://www.sciencedirect.com/science/article/abs/pii/S0950705122007821

  3. Demo pipeline of SGM: https://github.com/z619850002/DepthEstimation-SGM

  4. Official implementation of DPT: https://github.com/isl-org/DPT

  5. Testing dataset link: http://www.doc.ic.ac.uk/~ahanda/HighFrameRateTracking/downloads.html

 

Created on: Nov. 09, 2023