Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

inconsistent and incorrect scale in predicted depth #48

Open
ChauChorHim opened this issue Mar 15, 2022 · 5 comments
Open

inconsistent and incorrect scale in predicted depth #48

ChauChorHim opened this issue Mar 15, 2022 · 5 comments

Comments

@ChauChorHim
Copy link

Hi there, thanks for sharing your codes. I am trying to train and predict my own dataset.
The predicted depth jpeg image looks fine in eye. However, when I try to reconstruct the depth image into pointcloud in camera coordinate, the points are inconsistent in the same object and in the whole scene.

For example, the upper pole on the right are stretched, which is not obvious in the jpeg image. Also, the scale of the car is not normal.

I think one of the reason is that the scale of depth is not correct locally and globally.

Have your group noticed these issues? I am really appreciated if you could give me some advices!

RGB image

1645427110 701078

Depth image
1645427110 701078_disp_None

Pointcloud - the poles are twisted
Screenshot from 2022-03-15 11-39-43

Pointcloud - the car is streched
Screenshot from 2022-03-15 11-43-58

@daniyar-niantic
Copy link
Collaborator

Hi. Do you use single images for prediction, or multiple images?

@iariav
Copy link

iariav commented Apr 25, 2022

having a similar issue.
single image prediction.

@daniyar-niantic
Copy link
Collaborator

With single image prediction this is probably acceptable quality of depth maps. You have different intrinsics then at training time, and there are no additional images to resolve depth ambiguities.

@iariav
Copy link

iariav commented Apr 25, 2022

thank you for your reply.
But why do you assume different intrinsics?
if I train and test on images generated from the same camera rig, it's the same intrinsics, right?

@daniyar-niantic
Copy link
Collaborator

Oh, sorry I thought this is trained on KITTI. You are right, the intrinsics would be the same. I would suggest to make a multi-frame prediction and see if local/global scale is correct. If it too bad, then there may be a problem with data loading/processing or the model has not converged.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants