Code Monkey home page Code Monkey logo

kaanakan / stretchbev Goto Github PK

View Code? Open in Web Editor NEW
41.0 2.0 2.0 160 KB

Official implementation of our ECCV paper "StretchBEV: Stretching Future Instance Prediction Spatially and Temporally"

Home Page: https://kuis-ai.github.io/stretchbev/

License: MIT License

Python 100.00%
autonomous-driving autonomous-vehicles deep-learning eccv2022 future-prediction generative-models instance-segmentation latent-variable-models pytorch segmentation

stretchbev's Introduction

StretchBEV: Stretching Future Instance Prediction Spatially and Temporally (ECCV 2022)

report report report report

StretchBEV: Stretching Future Instance Prediction Spatially and Temporally,
Adil Kaan Akan, Fatma Guney,
European Conference on Computer Vision (ECCV), 2022

Features

StretchBEV is a future instance prediction network in Bird's-eye view representation. It earns temporal dynamics in a latent space through stochastic residual updates at each time step. By sampling from a learned distribution at each time step, we obtain more diverse future predictions that are also more accurate compared to previous work, especially stretching both spatially further regions in the scene and temporally over longer time horizons

Requirements

All models were trained with Python 3.7.10 and PyTorch 1.7.0

A list of required Python packages is available in the environment.yml file.

Datasets

For preparations of datasets, we followed FIERY. Please follow this link below if you want to construct the datasets.

Training

To train the model on NuScenes:

  • First, you need to download static_lift_splat_setting.ckpt and copy it to this directory.
  • Run python train.py --config fiery/configs/baseline.yml DATASET.DATAROOT ${NUSCENES_DATAROOT}.

This will train the model on 4 GPUs, each with a batch of size 2. To train on single GPU add the flag GPUS 1, and to change the batch size use the flag BATCHSIZE ${DESIRED_BATCHSIZE}.

Evaluation

To evaluate a trained model on NuScenes:

  • Download pre-trained weights.
  • Run python evaluate.py --checkpoint ${CHECKPOINT_PATH} --dataroot ${NUSCENES_DATAROOT}.

Pretrained weights

You can download the pretrained weights from the releases of this repository or the links below.

Normal setting weight

Fishing setting weight

How to Cite

Please cite the paper if you benefit from our paper or the repository:

@InProceedings{Akan2022ECCV,
            author    = {Akan, Adil Kaan and G\"uney, Fatma},
            title     = {StretchBEV: Stretching Future Instance Prediction Spatially and Temporally},
            journal = {European Conference on Computer Vision (ECCV)},
            year      = {2022},
            }

Acknowledgments

We would like to thank FIERY and SRVP authors for making their repositories public. This repository contains several code segments from FIERY's repository and SRVP's repository. We appreciate the efforts by Berkay Ugur Senocak for cleaning the code before release.

stretchbev's People

Contributors

kaanakan avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar

Forkers

githubgsq whuhxb

stretchbev's Issues

Config of StretchBEV-P and StretchBEV

Hi, thanks for your work.

File 'stretchbev.yml' is the config of strethbev or stretchbev-p? Could you provide both config files of strethbev and stretchbev-p?

Question about the inference pipeline

Hi,

I evaluated the pretrained model provided in this repo, whose result is similar to the StretchBEV-P model in the paper.

However, it seems that ground truth labels for history and current timeframe are involved in the evaluation process. In inference_srvp_generate() function, hx_z contains feature generated from label input.

Under this setting, the comparison with fiery in the paper seems unfair, and the extra data usage in the evaluation doesn't match the caption of table 1 in the paper

... the two versions of our model with (StretchBEV-P) and without (StretchBEV) using the labels for the output modalities in the posterior while learning the temporal dynamics ...

Reproduce the results on Nuscenes

Thanks for your great work.
I reproduced the results on Nuscenes dataset with your pre-trained checkpoint, following your instructions in README.md, and got the following outputs:
iou
57.5 & 52.1
pq
51.5 & 47.7
sq
75.5 & 73.1
rq
68.3 & 65.3
Is this result expected? Is there any error? Is this the result of StretchBEV-P?
Looking forward to your response. Thank you in advance.

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.