Code Monkey home page Code Monkey logo

csd-ssd's People

Contributors

soo89 avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar

csd-ssd's Issues

Training on GPUs

Hello,

Thanks for your nice work. I just have few questions about training on GPUs.

  1. Do you train your CSD on a single GPU or multiple GPUs? If trained on single GPU, do you mind sharing the series code(what type of GPU)? I am asking this because I am not able to train CSD with batch size of 32 on a single 2080 Ti and I'm not sure whether the problem is on my server or the nature of the code.

  2. What's your estimated training time on VOC(07 labeled and 12 unlabeled) in your configuration? It looks like training a baseline SSD on a single GPU would take me several days in my case so it would be nice if you could share the stats....

Thanks for your time and enjoy your day!

consistency_conf_loss

in paper, the consistency_conf_loss is JSD. but in the code, the consistency_conf_loss is sum of 2 KL
image
How do I understand this consistency_conf_loss_a and consistency_conf_loss_a.
JSD should be like this
image

thanks

ramp-down function for loss weight

Hi, thanks for providing such nice work!
I have a minor question about the loss weight schedule.
For the consistency loss that you've designed, is there any experimental results that ramp-down weight scheduling affects the performance? Ramp-up seems plausible for stable training but I'm not sure why ramp-down needed.

Understanding of Unlabeled Losses

Hi,

Thanks for sharing your code. I would like to inquire more about the JSD loss for classification consistency. specifically this line Link

Why is .detach() specifically used here for computing the losses?

And, I didn't understand how it impacts when you are detaching either output in each line? If I print requires_grad on those tensors it shows True.

The failure of download code

Hello, Thanks for your nice work!

When I downloaded your code, I tried many times but failed.
I thought maybe beacause of your "weight" folder that had the 70 M pretrained model.

Could you please delete the weight folder and offer another pretrained model link?

Thank you very much!

why flip twice?

Thanks for you great work!
read the code, I find some code which I can't to understand, as below,
image
why it need to flip append loc and append conf?


by the way, we compare loc and flip loc in

consistency_loc_loss_x = torch.mean(torch.pow(loc_sampled[:, 0] + loc_sampled_flip[:, 0], exponent=2))
...

but, how we guarantee each item between loc_sampled and loc_sampled_flip can match correctly?

Hope your kindly response, much thanks.

The setting of --dataset

Hi Jisoo, thanks for your nice work.
could you tell me what --dataset is? Does it mean using the different configuration settings for different sizes of inputs? For example, VOC300 refers to the setting for an ssd300 model while VOC512 refers to ssd512?

CSD in YOLO

Hi, have you ever validate your CSD method on YOLO detection?

consistency loss on fcos

thank you for your work. when i use consistence loss on focs model with pascalvoc dataset, the accuracy always decrease, do you have any suggestion ?

pytorch version

I'm interested in this paper.
So I'd like to train this model on custom datasets, but failed to train.
Could you tell me which version of pytorch and cuda you use?
thank you!

How do I train this model on a custom dataset?

Hi @soo89,

Wonderful work.

Could you please help understand how can I train this semi-supervised model on my own dataset. I only have images and xml files. I have not worked with VOC/COCO datasets before. Could you please give some suggestions about where should I start? Do I need to split my dataset into labelled and unlabelled datasets to perform semi-supervised learning?

I appreciate your response.

Localization consistency loss in csd.py

Hi author,
Thanks for your wonderful work and your released code. I am confused about the code for the Localization consistency loss in https://github.com/soo89/CSD-SSD/blob/30b184c86a87b0fc2d301cd8157b11d8cfe7da1e/train_csd.py#L297.

consistency_loc_loss_x = torch.mean(torch.pow(loc_sampled[:, 0] + loc_sampled_flip[:, 0], exponent=2))
  
consistency_loc_loss_y = torch.mean(torch.pow(loc_sampled[:, 1] - loc_sampled_flip[:, 1], exponent=2))
                
consistency_loc_loss_w = torch.mean(torch.pow(loc_sampled[:, 2] - loc_sampled_flip[:, 2], exponent=2))
                
consistency_loc_loss_h = torch.mean(torch.pow(loc_sampled[:, 3] - loc_sampled_flip[:, 3], exponent=2))

For y, w, and h you use '-' when calculating the mean squared loss. I am confused about why you use '+' for x when calculating the mean squared loss.

Single batch: Supervised or Unsupervised

hi,
in the experiment section, you mentioned that labeled and unlabeled data are randomly shuffled and selected. I wanted to know why it's necessary to have each mini-batch containing both labeled as well as unlabeled data?

What happens if one of mini-batch contains only unlabeled data (since the majority of data is unlabeled)? Can we handle that case? What would be the value of loss?

I'm new to semi-supervised.
thanks.

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.