Code Monkey home page Code Monkey logo

Comments (17)

zdou0830 avatar zdou0830 commented on August 17, 2024

See #6. Thanks!

from meter.

wanng-ide avatar wanng-ide commented on August 17, 2024

@zdou0830 Thanks!

I have another question about the codes.

I tried your pretrained model to finetune VQA v2 with default setting.
However, the score of val is only around 72.55.

It should be more than 80..

Could you share your experiment settings of finetuning VQA v2 and pretraining tasks?

from meter.

zdou0830 avatar zdou0830 commented on August 17, 2024

The command should be

python run.py with data_root=$DATA_DIR num_gpus=8 num_nodes=1 task_finetune_vqa_clip_bert per_gpu_batchsize=4 clip16 text_roberta image_size=576 clip_randaug load_path=meter_clip16_288_roberta_pretrain.ckpt

Hope this thread would help #7.

from meter.

wanng-ide avatar wanng-ide commented on August 17, 2024

@zdou0830

This setting is the same as mine ...

from meter.

wanng-ide avatar wanng-ide commented on August 17, 2024

If I use your finetuned model to do the test only task of VQA v2, the result is 77.66.
This model works.

from meter.

wanng-ide avatar wanng-ide commented on August 17, 2024

image

from meter.

zdou0830 avatar zdou0830 commented on August 17, 2024

You can try testing the last checkpoint and submitting the resulting json file to evalai.

from meter.

wanng-ide avatar wanng-ide commented on August 17, 2024

The result from evalai is 71.53.
Should I finetune the model with more step?

from meter.

zdou0830 avatar zdou0830 commented on August 17, 2024

The VQA dataset can be downloaded here: https://drive.google.com/file/d/1qT7YWHpLg-fAL43daKlOsYx2EbbQk--d/view?usp=sharing.

The training command is

python run.py with data_root=$DATA_DIR num_gpus=8 num_nodes=1 task_finetune_vqa_clip_bert per_gpu_batchsize=4 clip16 text_roberta image_size=576 clip_randaug load_path=meter_clip16_288_roberta_pretrain.ckpt

The testing command is

python run.py with data_root=$DATA_DIR num_gpus=8 num_nodes=1 test_only=True task_finetune_vqa_clip_bert per_gpu_batchsize=4 clip16 text_roberta image_size=576 load_path=last.ckpt

The provided VQA-finetuned checkpoint is trained in this way, so if you follow these steps correctly, you should be able to get a score of ~77.6 on test-dev. I didn't look at the dev scores and the number of training epochs was set to 10 as in config.py. The fine-tuning took about 2 days on 8 V100s for reference.

from meter.

wanng-ide avatar wanng-ide commented on August 17, 2024

Ok, I will have a try! Thank you for your patient.

from meter.

wanng-ide avatar wanng-ide commented on August 17, 2024

I found a problem.
If I use only one node to finetune the pretrained model, the result will be better than two nodes (around 5% in VQA v2 Val).

That might be the reason.

May you share your pretraining log?
I will pretrain the model in two nodes.
I want to know the different between one node and two nodes.

from meter.

zdou0830 avatar zdou0830 commented on August 17, 2024

I didn't save the logs, but I did pre-train the models with 1/2/4 nodes and there were no significant differences, so I'd suggest you to debug your multi-node training settings.

from meter.

jiyt17 avatar jiyt17 commented on August 17, 2024

I also met the same problem, which probably results from multi-node training settings. I use slurm to multi-node train. May you share your train bash file, if you also use slurm.

from meter.

zdou0830 avatar zdou0830 commented on August 17, 2024

I didn't use slurm, but I uploaded the running file for distributed training on Microsoft machines (https://github.com/zdou0830/METER/blob/main/azure_distributed_run.py). Not sure if this is helpful.

from meter.

jiyt17 avatar jiyt17 commented on August 17, 2024

ok, thank u~

from meter.

mactavish91 avatar mactavish91 commented on August 17, 2024

I also met the same problem, which probably results from multi-node training settings. I use slurm to multi-node train. May you share your train bash file, if you also use slurm.

@jiyt17 Hello, have you solved the problems you encountered before?

from meter.

mactavish91 avatar mactavish91 commented on August 17, 2024

@zdou0830 Thanks!

I have another question about the codes.

I tried your pretrained model to finetune VQA v2 with default setting. However, the score of val is only around 72.55.

It should be more than 80..

Could you share your experiment settings of finetuning VQA v2 and pretraining tasks?

@wanng-ide Hello, have you solved the problems you encountered before?

from meter.

Related Issues (20)

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.