Code Monkey home page Code Monkey logo

pointgoal-navigation's People

Contributors

dobosevych avatar erikwijmans avatar mathfac avatar rpartsey avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar

Forkers

ram81 lienbo

pointgoal-navigation's Issues

Settings for reproducing the visual odometry model performance

Hi, thank you for sharing your impressive work!

I'm trying to use your visual odometry model on the habitat simulator (with mp3d, gibson environments), but I'm struggling with reproducing the model performance.

Here are my settings, based on your codes

VO model input
- rgbd channel order : [source_rgbd, target_rgbd]
- input action : moveforward=0, turn_left=1, turn_right=2
- img size : (180, 320)
- hfov : 70 degrees
- rgb normalization : rgb(uint8)/255.
- depth normalization : depth/10. (np.where(rgbd['depth_sensor'] < 10, rgbd['depth_sensor']/10., 1.))

Agent, Sensor settings
- sensor height : 0.88
- actions : (move forward 0.25m, turn left, right 30 degrees)

I tested the provided pretrained model no. 10 (the biggest model trained on gibson), 11 (mp3d finetuned) on the gibson and mp3d validation dataset with above settings.
Unfortunately, I got the total translation MAE of 13cm, rotation mae of 0.075 rad which is far inferior to the supplementary results of the paper. I measured these with 1 step error, not the cumulated error of the trajectory. The qualitative results of the estimated trajectories are also far worse than the results on the paper.

Is there anything wrong with the above settings? Or are there any tips for using your pretrained models?
Thank you in advance.

can find the agent.py file to evaluate agent

export CHALLENGE_CONFIG_FILE=config_files/challenge_pointnav2021.local.rgbd.yaml python agent.py \ --agent-type PPOAgentV2 \ --input-type depth \ --evaluation local \ --ddppo-checkpoint-path <path-to-policy-checkpoint> \ --ddppo-config-path config_files/ddppo/ddppo_pointnav_2021.yaml \ --vo-config-path <path-to-vo-config> \ --vo-checkpoint-path <path-to-vo-checkpoint> \ --pth-gpu-id 0 \ --rotation-regularization-on \ --vertical-flip-on

I can not find the agent.py to evaluate agent(navigation policy + visual odometry), and I only find a agent.py in trajectory-sampling/agent/. There is no main function in trajectory-sampling/agent/agent.py.

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.