Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Error: principal point is in upper left, not setup for this right now #11

Open
rmadan2 opened this issue Nov 28, 2024 · 5 comments
Open

Comments

@rmadan2
Copy link

rmadan2 commented Nov 28, 2024

I am running the streetlearn model on stereo images from the KITTI dataset. I have slightly modified demo.py such that it accepts stereo pairs. Furthermore, I have made the following modifications to the intrinsics accepted by the script:
if "matterport" in args.ckpt: intrinsics = np.stack([np.array([[517.97, 517.97, 320, 240], [517.97, 517.97, 320, 240]])]).astype(np.float32) else: intrinsics = np.stack([np.array([[128,128,128,128], [128,128,128,128]])]).astype(np.float32) # intrinsics = np.stack([np.array([[9.786977e+02, 9.717435e+02, 6.900000e+02, 2.497222e+02], [9.786977e+02, 9.717435e+02, 6.900000e+02, 2.497222e+02]])]).astype(np.float32) intrinsics = torch.from_numpy(intrinsics).cuda()
However, The following error message seems to popup:
principal point is in upper left, not setup for this right now
I have tried using both the intrinsics ([128,128,128,128]) and [9.786977e+02, 9.717435e+02, 6.900000e+02, 2.497222e+02]. Is there a way to resolve this?

@crockwell
Copy link
Owner

Ah, I see. I responded to your other comment in issue #1. I'd recommend moving to DUSt3R. If you want to compare specifically to this model, you may just use Matterport intrinsics. Note the Matterport model is trained only on indoor scenes, so it probably won't generalize well to outdoor scenes.

@rmadan2
Copy link
Author

rmadan2 commented Nov 29, 2024

Thank you, I will look further into DUSt3R

@rmadan2
Copy link
Author

rmadan2 commented Nov 29, 2024

One thing I do mean to ask, does the network have any sort of memory between the pairs of images fed to it?

@crockwell
Copy link
Owner

If you're asking how it can process both images jointly, the network concatenates them before passing into the network. So it would scale similarly to passing two "halves" of an image into it. If you're talking about multiple pairs, it has no memory beyond learning from each pair of images via backprop during training.

@rmadan2
Copy link
Author

rmadan2 commented Dec 1, 2024

I suppose the reason I ask this is because I only see the error when I give the network a series of stereo pairs (from the KITTI dataset) in a loop. I don't see it when I simply input a stereo pair to demo.py.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants