Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

@foralliance Hi!I think we should speak English because your questions may help the people in other countries. #8

Open
defei-coder opened this issue Jul 12, 2019 · 3 comments

Comments

@defei-coder
Copy link

@foralliance Hi!I think we should speak English because your questions may help the people in other countries.

  1. No. When you train the model with large input image size (e.g., 800x1300), the batch-size will reduce to 1-2 due to the limited GPU memory. Then the effect of BN will be constrained. If so, please replace BN with GN. GN does not care batch-size.

  2. This question is very complex. I can just answer your question that they are not relevant.

Originally posted by @KimSoybean in #3 (comment)

@defei-coder
Copy link
Author

@KimSoybean Hi
I read the answer above,you mean one GPU can not reach the large batch_size, I think 128 means accum_batch_size, we can use one GPU read 4 batch_size by 32 iter_size. So the batch_size of one GPU(such as 4) will Influence the effect of BN? Or why you recommend GN.

@defei-coder
Copy link
Author

Or B of BN means batch_size that one gpu can reach.Looking forward to your answer。

@KimSoybean
Copy link
Owner

BN is evaluated by the statistics in a batch, so batchsize will impact the perfoemance. I have tried GN on mmdetection on ScratchDet, and get 1mAP lower result. I haven't found the reason.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants