Code Monkey home page Code Monkey logo

robi_toolkit's Introduction

ROBI_Toolkit

Python scripts to work on the ROBI dataset - a multi-view dataset for reflective objects in robotic bin-picking.
ROBI dataset is available at: https://www.trailab.utias.utoronto.ca/robi. ROBI

Requirements

Python >=3.5
opencv-python >= 3.1
numpy
ruamel.yaml

Baseline Methods

We provide the evaluation results on three object pose estimators (PPF [1], Line2D [2], AAE [3]), reasearchers are welcome to compare them against their our methods. The raw evaluation results can be downloaded here. Please run "eval_baselines.py" to load the the estimated object poses from these baseline approaches.

Code

  • "read_scene_imgs.py": A script to load test images (with 6D camera poses and the ground truth 6D object poses).
  • "read_pattern_imgs.py": A script to load stereo pattern images and disparity maps.
  • "eval_baselines.py": A script to load ground truth and the estimated object poses from provided baseline methods.

Ground Truth Depth Map

Please note that, we capture the ground truth depth map with only Ensenso camera (no Realsense data).

  • For Scene 4, 5, 8, 9, each viewpoint image has the corresponding GT depth map (in "GT_Depth" folder).
  • For Scene 1, 2, 3, 6, 7, the GT depth maps were captured only for a subset of viewpoints in Ensenso data folder:
    • Scene 1-3: DEPTH_view_{71-87}
    • Scene 6-7: DEPTH_view_{12-14, 16, 18, 22, 24-41}

Author

Jun Yang
[email protected]
Institute for Aerospace Studies, University of Toronto

References

[1] Drost, Bertram, et al. "Model globally, match locally: Efficient and robust 3D object recognition." 2010 IEEE computer society conference on computer vision and pattern recognition. Ieee, 2010.
[2] Hinterstoisser, Stefan, et al. "Model based training, detection and pose estimation of texture-less 3d objects in heavily cluttered scenes." Asian conference on computer vision. Springer, Berlin, Heidelberg, 2012.
[3] Sundermeyer, Martin, et al. "Implicit 3d orientation learning for 6d object detection from rgb images." Proceedings of the european conference on computer vision (ECCV). 2018.

Citation

If you find ROBI dataset useful in your work, please consider citing:

@inproceedings{yang2021robi,
  title={ROBI: A Multi-View Dataset for Reflective Objects in Robotic Bin-Picking},
  author={Yang, Jun and Gao, Yizhou and Li, Dong and Waslander, Steven L},
  booktitle={2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)},
  pages={9788--9795},
  year={2021},
  organization={IEEE}
}

@article{yang2021probabilistic,
  title={Probabilistic Multi-View Fusion of Active Stereo Depth Maps for Robotic Bin-Picking},
  author={Yang, Jun and Li, Dong and Waslander, Steven L},
  journal={IEEE Robotics and Automation Letters},
  volume={6},
  number={3},
  pages={4472--4479},
  year={2021},
  publisher={IEEE}
}

robi_toolkit's People

Contributors

junyang224 avatar

Stargazers

 avatar Yu Matsumura avatar  avatar  avatar Bharath Raj M avatar Linpeng Peng avatar  avatar  avatar xinxin avatar  avatar AlexYiningLiu avatar Juan avatar sunhan avatar  avatar  avatar  avatar

Watchers

 avatar  avatar

robi_toolkit's Issues

A question about disparity

Hi~ I notice the following in your code:
pix_right_x = pix_left_x + disparity

So how about the relationship between pix_right_y and pix_left_y?

 

Question about the 'GT_Depth'

Hi,

Thank you for your work.

I have a question about ROBI dataset. I noticed folder 'GT_Depth', the index of images in this folder is not matched with the index in the sensor's folders. Is there any relationship between images from GT_Depth and images from the sensor's folders? What is the function of it?

Regards,
Rui

visibility score

Sorry to interrupt, where to download the associated visibility score of visible objects.

About roi

hello,I want to ask how to get the box mak,like this,do you have sample way? Thank you!
image

question

Thanks for your dataset.I have three quetion.
1.In test/Ensenso_LEFT.yml cx: 379.326874 that’s right?In test/Ensenso_DEPTH.yml depth_unit: 0.03125 that‘s in mm?cv2.imread(DEPTH_path, cv2.IMREAD_ANYDEPTH) * depth_unit get the true depth and in mm
2.read_pattern_imgs.py what task is it used for?
3.the cam_pose is the pose from world to cam?the GT picture is processed by you and that's better than origin?

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.