-
Notifications
You must be signed in to change notification settings - Fork 27
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Where to find the camera intrinsics? #4
Comments
Futhermore, I would like to ask where to find the raw depth image in the raw data (27TB). Thank you very much! |
Thanks, I have found the code in the droid/scripts/convert/svo_to_mp4.py where I can trasform the original .svo file into rgb image, depth and pointcloud. However, i find that there are a lot of "nan" in depth and pointcloud. Could you please give me some advice about this? |
Hi Nimolty: Thanks for the questions, I created a draft PR here that should help answer most of your questions and show how you can work with the provided depth information in DROID. To summarize:
|
Thanks for the answer! |
Thank you for your detailed advice! |
You can try something like this. This is a lower quality open-source version of an internal model that's been working pretty well for us at TRI. |
Thanks for your great work! |
The extrinsics information is published as part of the metadata in the droid_raw subset, see here: https://droid-dataset.github.io/droid/the-droid-dataset.html#accessing-raw-data (in the |
@kpertsch @ashwin-balakrishna96 This is an example of the metadata of an epoch in the raw dataset. If I understand correctly, And thank you for propose #5 so we can get depth and intrinsics from the svo files of the raw dataset! If possible, please add the depth, intrinsics and extrinsics to the RLDS dataset!
|
Re adding things to RLDS: I will try to add the extrinsic info when I recompile the RLDS data! Depth increases dataset size a lot since it's not compressible and needs to be stored as float32 tensors (and thereby makes loading during training slower), so we intentionally didn't include it, but try to provide utilities for people to compute it themselves if they would like to! |
I can understand! But would you like to also add intrinsics to the RLDS data? @kpertsch I want to train a policy that can adapt to any intrinsics / extrinsics. My plan is to include such info in the policy input... |
Got it -- yeah we can add intrinsic info too |
Great thanks! |
Hi, have you finished this RLDS recompiling? I also need it! |
Hi @kpertsch |
@StarCycle I meet the same problem. Did you solve it? @kpertsch what is the wrist_cam_extrinsics in the metadata.json mean? Is that stand for wrist camera pose at the first frame in episodes? Is it possible to multiply the wrist_cam_extrinsics with the gripper pose of the initial frame to obtain cam2gripper, and then compute wrist_cam_extrinsics for each frame by applying matmul(cam2gripper, gripper_pose) for every frame?
|
Hello! Thanks for opening-source this great works!
I am trying to analyze the raw data (27TB). However, I was wondering where the camera intrinsics are. Could you please provide some information.
The text was updated successfully, but these errors were encountered: