Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

run the code on coco, but can not get the same results shown in the paper #8

Closed
rock4you opened this issue Jul 9, 2021 · 14 comments
Closed

Comments

@rock4you
Copy link

rock4you commented Jul 9, 2021

updetr

@rock4you rock4you changed the title run the code on coco, but can not get the same result in the paper run the code on coco, but can not get the same results shown in the paper Jul 9, 2021
@dddzg
Copy link
Owner

dddzg commented Jul 9, 2021

May I ask you for the detail config of training coco (number of gpus and etc)?
There is a log in our experiments:https://drive.google.com/file/d/1DQqveOZnMc2VaBhMzl9VilMxdeniiWXo/view?usp=sharing
You can compare your log with it.

@rock4you
Copy link
Author

rock4you commented Jul 9, 2021

GPU using 8 cards of V100 , and the commands are the same as your provided in the github.
Is there anything has to be modified before running the train program?

@dddzg
Copy link
Owner

dddzg commented Jul 9, 2021

I check it again. There is a mistake of my script. I am so sorry. The lr_backbone should set to 5e-5 instead of 5e-4. I will update the README. Thanks a lot! I will keep the issue open until you get the right result.

@dddzg
Copy link
Owner

dddzg commented Jul 14, 2021

Hi @rock4you , may I ask for some new progress?

@rock4you
Copy link
Author

Still running, it looks good this time.
training

@rock4you
Copy link
Author

The AP of coco val2017 with 300 epochs in Table 2 of the paper is 42.8,
is this result get from a certain training process or the mean value of several times ?

@dddzg
Copy link
Owner

dddzg commented Jul 15, 2021

As coco dataset is large, the result is reported at the last training epoch without serveral times (I guess the result variance is small on coco). BTW, may I ask for your result?

@rock4you
Copy link
Author

Still running, the AP around epoch 240 is 0.430.
The training speed is about 40 epochs / day

@dddzg
Copy link
Owner

dddzg commented Jul 15, 2021

Glad to hear the result. As far as I observe, the open-source pre-trained model is a little better than paper report.

@rock4you
Copy link
Author

👍🏻 👍🏻

@rock4you
Copy link
Author

0.435

updetr

@dddzg
Copy link
Owner

dddzg commented Jul 19, 2021

Nice to hear the result. Could you attach more detailed COCO style evaluation result (such like https://gist.github.com/dddzg/cd0957c5643f5656f6cdc979da4d6db1)?

@rock4you
Copy link
Author

rock4you commented Jul 19, 2021

The last epoch:

IoU metric: bbox
Average Precision (AP) @[ IoU=0.50:0.95 | area= all | maxDets=100 ] = 0.432
Average Precision (AP) @[ IoU=0.50 | area= all | maxDets=100 ] = 0.632
Average Precision (AP) @[ IoU=0.75 | area= all | maxDets=100 ] = 0.458
Average Precision (AP) @[ IoU=0.50:0.95 | area= small | maxDets=100 ] = 0.209
Average Precision (AP) @[ IoU=0.50:0.95 | area=medium | maxDets=100 ] = 0.475
Average Precision (AP) @[ IoU=0.50:0.95 | area= large | maxDets=100 ] = 0.622
Average Recall (AR) @[ IoU=0.50:0.95 | area= all | maxDets= 1 ] = 0.342
Average Recall (AR) @[ IoU=0.50:0.95 | area= all | maxDets= 10 ] = 0.550
Average Recall (AR) @[ IoU=0.50:0.95 | area= all | maxDets=100 ] = 0.589
Average Recall (AR) @[ IoU=0.50:0.95 | area= small | maxDets=100 ] = 0.311
Average Recall (AR) @[ IoU=0.50:0.95 | area=medium | maxDets=100 ] = 0.650
Average Recall (AR) @[ IoU=0.50:0.95 | area= large | maxDets=100 ] = 0.816
Training time 7 days, 12:19:43


The highest at epoch 288:

IoU metric: bbox
Average Precision (AP) @[ IoU=0.50:0.95 | area= all | maxDets=100 ] = 0.435
Average Precision (AP) @[ IoU=0.50 | area= all | maxDets=100 ] = 0.633
Average Precision (AP) @[ IoU=0.75 | area= all | maxDets=100 ] = 0.464
Average Precision (AP) @[ IoU=0.50:0.95 | area= small | maxDets=100 ] = 0.211
Average Precision (AP) @[ IoU=0.50:0.95 | area=medium | maxDets=100 ] = 0.477
Average Precision (AP) @[ IoU=0.50:0.95 | area= large | maxDets=100 ] = 0.626
Average Recall (AR) @[ IoU=0.50:0.95 | area= all | maxDets= 1 ] = 0.342
Average Recall (AR) @[ IoU=0.50:0.95 | area= all | maxDets= 10 ] = 0.550
Average Recall (AR) @[ IoU=0.50:0.95 | area= all | maxDets=100 ] = 0.589
Average Recall (AR) @[ IoU=0.50:0.95 | area= small | maxDets=100 ] = 0.315
Average Recall (AR) @[ IoU=0.50:0.95 | area=medium | maxDets=100 ] = 0.653
Average Recall (AR) @[ IoU=0.50:0.95 | area= large | maxDets=100 ] = 0.818

@dddzg dddzg closed this as completed Jul 19, 2021
dddzg added a commit that referenced this issue Jul 19, 2021
add the link with #8
@xiaoerlaigeid
Copy link

I check it again. There is a mistake of my script. I am so sorry. The lr_backbone should set to 5e-5 instead of 5e-4. I will update the README. Thanks a lot! I will keep the issue open until you get the right result.

Hi you mean in the finetune stage or pretrain stage ? Why in the pretrain stage the backbone should freeze ?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants