Comments (9)
Thank you for your help. With lr_drop=40, lr 0.0002, lw_saliency=4, we finally reproduce the paper results.
from qd-detr.
I couldn't reproduce the results reported in the paper on the Charades dataset with the SF+CLIP features and default hyperparameters.
from qd-detr.
Hello! Sorry for being late due to our busy schedule for the rebuttal period.
Actually, we are very sorry that we currently only have the config file for QVHighlights for QD-DETR.
We suggest you try tuning the ['learning rate \in 1e-4, 2e-4', 'saliency loss ratio \in 1, 4'].
We also find that the results are not consistent on different machines due to some reason with our codebase for somewhat reason (we haven't figured out why).
So we recommend trying different parameters on your specific machine.
from qd-detr.
Thanks for getting back to me in your busy schedule,
I've tried the parameter you mentioned, but the model could only converge to 45 at [email protected], which is far from the records in the paper. However, when I tried I3D features instead of SF+CLIP, the model could converge to 53 at [email protected], which is close to the records in the paper.
I guess I need to change some parameters. It would be better if you could help me find the opt.json file after your busy rebuttal period. Thanks.
from qd-detr.
@wjun0830 @zxccade
Hi, I also cannot reproduce the experiments with Charades-STA with the default parameter.
The results are here, [email protected]=0.5 and far from 0.57 in the paper. Could you share the hyper-parameters?
"[email protected]": 51.88,
"[email protected]": 27.93,
"MR-full-mAP": 30.26,
"[email protected]": 60.92,
"[email protected]": 25.31
from qd-detr.
We are again very sorry that our codebase is not very robust to different server settings.
Have you tried changing params as above?
You can also refer to Params for succeeding paper can be found in the appendix of https://arxiv.org/abs/2311.08835.
from qd-detr.
Thank you for reply! I will try the hyper-parameters used in CG-DETR, let me take several hours to test it...
from qd-detr.
@wjun0830
Sorry for bothering you again. They are hyper-parameters in CG-DETR. Except for CG-DETR-specific parameters, you used these parameters? Several differences exist from QD-DETR. For example, enc_layers=3, dec_layers=3, --lr 0.0002, --lw_saliency=4 are set (In QD-DETR, enc_layers=2, dec_layers=2, lr=0.0001, lw_saliency=1.0).
#### training
bsz=32
eval_bsz=32
num_dummies=45
num_prompts=2
total_prompts=10
lr_drop=400
enc_layers=3
dec_layers=3
t2v_layers=2
dummy_layers=2
moment_layers=1
sent_layers=1
PYTHONPATH=$PYTHONPATH:. python cg_detr/train.py \
--dset_name ${dset_name} \
--ctx_mode ${ctx_mode} \
--train_path ${train_path} \
--eval_path ${eval_path} \
--eval_split_name ${eval_split_name} \
--v_feat_dirs ${v_feat_dirs[@]} \
--v_feat_dim ${v_feat_dim} \
--t_feat_dir ${t_feat_dir} \
--t_feat_dim ${t_feat_dim} \
--bsz ${bsz} \
--results_root ${results_root} \
--exp_id ${exp_id} \
--max_v_l -1 \
--clip_length 1 \
--lr 0.0002 \
--lr_drop ${lr_drop} \
--n_epoch 200 \
--contrastive_align_loss_coef 0.002 \
--lw_saliency 4 \
--enc_layers ${enc_layers} \
--dec_layers ${dec_layers} \
--t2v_layers ${t2v_layers} \
--moment_layers ${moment_layers} \
--dummy_layers ${dummy_layers} \
--sent_layers ${sent_layers} \
--eval_bsz ${eval_bsz} \
--num_dummies ${num_dummies} \
--num_prompts ${num_prompts} \
--total_prompts ${total_prompts} \
${@:1}
from qd-detr.
I remember that we havent changed the number of layers in QD detr.
Modification of number layers were implemented only in cg detr following works in iccv.
I remember lwsaliency and lr are the changes we have tuned for charades
from qd-detr.
Related Issues (20)
- The implement of rank-aware contrastive loss HOT 2
- What does parameter "use_tef" mean? HOT 2
- At training, RuntimeError: The size of tensor a (148) must match the size of tensor b (150) at non-signleton dimesnion 1
- Confusion about the code HOT 1
- TypeError: unsupported operand type(s) for /: 'dict' and 'float' HOT 4
- The I3D features about Charades-STA HOT 3
- Training Machine Question HOT 1
- Charades dataset feature
- Training on Charades-STA dataset with VGG backbone HOT 10
- Fail to download TVsum dataset and could you please provide a new link? HOT 2
- TVSUM Result HOT 2
- TVSUM data issue HOT 4
- SharePoint: That didn't work - user cannot be found in the directory HOT 1
- About run_on_videos HOT 6
- Use videoonly.ckpt HOT 5
- official feature files for QVHighlights dataset HOT 3
- About ablation study HOT 1
- Pretraining Modules with Contrastive Learning? HOT 1
- With the same seed, the set of eval_epoch can really influence the performance of model! Why? HOT 5
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from qd-detr.