cccorn / ap-loss Goto Github PK
View Code? Open in Web Editor NEWThe implementation of "Towards accurate one-stage object detection with AP-loss".
License: MIT License
The implementation of "Towards accurate one-stage object detection with AP-loss".
License: MIT License
Hi!
glad to meet you. a few days ago, I reviewed your paper and code.
At that time, I found that definition of anchor makes default boxes with real pixel coordinate (i.e, each default box is consist of large scale pixel coordinate).
Why didn't it perform that with normalization coordinate? (e.g, [0.01, 0.01, 0.05, 0.05] means that this box has x_min, y_min, width and height respectively as relative of size of image)
Normally, SSD has used this way.
Thanks!
Hi,
I am planning to use different backbones (resnext-101, resnet-152) different than that you've already provided. As you said in readme, the backbone models are converted from caffe version. Could you please provide the converter tool that you used?
Thank you for your time.
Baris.
Hi,
I appreciate your good work. Read me says that test-dev2017 folder under images should be used. But source code doesn't refer those images. Is it necessary to have test-dev2017 images and what for?
I have read your paper.How do you calculate score?Which formula can you point out?
Hi,
Thank you for the method and the code.
Currently, I am running the code with the default setting (2 GPUs and 8 images/GPU) on 2 V100 GPUs and approximately it will take 7-10 days, which seemed a bit long for me. So, can you specify your training time with your default configuration for me to validate olease?
Kemal
I'm wondering if there is a mistake in line 253 in AP-loss/lib/dataloader/dataloader.py.
Maybe it should be :
valid_boxes.append([x1,y1,x2,y2,cls])
Did I miss anything?
Hi,
I noticed that while in the paper in Equation 5 during AP derivation, you add 1 both to the nominator and to the denominator following simply the formulation of AP.
But in this line of the code, you add 0.5 to the variable a and then estimate the precision by a/(a+b), which implies 0.5 is added to nominator and denominator. I understand all of your code and match with the paper, but I only could not see why you add 0.5 instead of 1 here. Am I missing something?
Thanks for your time
Kemal
Thank your for share great project. Some question not solved and need help. as follow:
# get global max prec in one batch
max_prec = 0
for i in batch:
current_prec = 1 - L(ij)
max_prec =max_prec if max_prec>current_prec else current_prec
# same as your code
for ii in order:
........
if (max_prec>=current_prec):
tmp2*=((1-max_prec)/(1-current_prec))
valid_bg_grad+=tmp2
prec[ii]=max_prec
Hope your reply.
Hi,
As I see test_img_size parameter in the config file determines the size of the shorter side to be set to test_img_size[0] by ensuring that the longer side does not exceed test_img_size[1], so I think defining a test_img_size=[400,500,600,...] will not evaluate multi-scale testing.
Is there a way to have the results with multi-scale testing or the code does not allow it?
Kemal
Nice work!
But I have a question about the ranking samples, I will be appreciated if you could help me about that.
Are those samples belonging to one ground thruth or just all positive samples in one image, or all positive samples in a batch.
Hi, long time no see. sir!
with congratulating our second meet...
the last time, thank for your reply.
Above, following the code, I have a new question for regression(localization) formulation.
Would you please refer the each line 166, 168, 174 and 176.
why did you limit upper bound of regression_diff_abs by using torch.le()
and if regression_diff_abs don't satisfy the condition, why dose this subtract - 0.5/1.0 or meet torch.sign() ?
Thank you.
Hi, I am reading your repo and some of your code really confused me.
Anchor scale
Line 16 in 79ba97c
Epoch here
Line 21 in 79ba97c
Input size here
Line 12 in 79ba97c
could you explain it kindly ?
Just......I'm curious. What are the differences between this work and your 2019 CVPR AP-Loss?
A declarative, efficient, and flexible JavaScript library for building user interfaces.
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google ❤️ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.