US 12,106,499 B2
Image depth estimation method and device, readable storage medium and electronic equipment
Hongmei Zhu, Beijing (CN); Qian Zhang, Beijing (CN); and Wei Sui, Beijing (CN)
Assigned to Beijing Horizon Information Technology Co., Ltd., Beijing (CN)
Filed by Beijing Horizon Information Technology Co., Ltd., Beijing (CN)
Filed on Nov. 13, 2021, as Appl. No. 17/525,903.
Claims priority of application No. 202011267769.1 (CN), filed on Nov. 13, 2020.
Prior Publication US 2022/0156957 A1, May 19, 2022
Int. Cl. G06T 7/55 (2017.01); G06T 7/70 (2017.01)
CPC G06T 7/55 (2017.01) [G06T 7/70 (2017.01); G06T 2207/20081 (2013.01); G06T 2207/20084 (2013.01)] 16 Claims
OG exemplary drawing
 
1. An image depth estimation method, including:
obtaining a first image frame and a second image frame collected in a movement process of an electronic apparatus;
determining a first feature map corresponding to the first image frame and a second feature map corresponding to the second image frame;
determining a scaled inter-frame geometrical relationship between the first image frame and the second image frame;
determining a reconstruction error between the first feature map and the second feature map based on the inter-frame geometrical relationship; and
determining a depth map corresponding to the first image frame based on the reconstruction error,
wherein the inter-frame geometrical relationship includes a translation distance and a rotation matrix;
wherein the determining a first feature map corresponding to the first image frame and a second feature map corresponding to the second image frame includes respectively performing feature extraction on the first image frame and the second image frame by utilizing a feature extraction branch in a first neural network to obtain the first feature map and the second feature map;
wherein the determining a scaled inter-frame geometrical relationship between the first image frame and the second image frame includes performing pose estimation on the first feature map and the second feature map by utilizing a pose estimation branch in the first neural network to obtain a translation distance and a rotation matrix of image collecting equipment between the first image frame and the second image frame,
wherein the determining a reconstruction error between the first feature map and the second feature map based on the inter-frame geometrical relationship includes calculating the reconstruction error from the second feature map to the first feature map based on the translation distance and the rotation matrix output by the pose estimation branch; and
the method further includes determining a cost volume from the second feature map to the first feature map based on the reconstruction error.