Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Data Augmentation #41

Open
csmailis opened this issue Mar 4, 2018 · 1 comment
Open

Data Augmentation #41

csmailis opened this issue Mar 4, 2018 · 1 comment

Comments

@csmailis
Copy link

csmailis commented Mar 4, 2018

I would like to ask how Data Augmentation is performed in the case of the Baseline RCNN that only uses groundtruth ROIs as Primary Regions.

More specifically within the paper, you mention that:

Rather than limiting training to the ground-truth person
locations, we use all regions that overlap more than 0.5 with
a ground-truth box. This condition serves as a form of data
augmentation. For every primary region, we randomly select
N regions from the set of candidate secondary regions.
N is a function of the GPU memory limit (we use a Nvidia
K40 GPU) and the batch size.
We fine-tune our network starting with a model trained
on ImageNet-1K for the image classification task. We tie
the weights of the fully connected primary and secondary
layers (fc6, fc7), but not for the final scoring models. We set
the learning rate to 0.0001, the batch size to 30 and consider
2 images per batch. We pick N = 10 and train for 10K
iterations. Larger learning rates prevented fine-tuning from
converging.

Thus for the case of the simple RCNN baseline that uses only primary regions and no secondary regions, this means that each batch contains 2 images and 30 ROIs for the ROI-Pooling layer.

Assuming the aforementioned assumption holds, in case the two images contain only 1 primary region each, with what do you fill the rest of the batch (as there should be 28 positions left empty) ?

Since the number of primary regions is not fixed per image, do you enforce the number of data augmentation samples to be balanced per class somehow?

Would it be possible to share the results you achieve without using data augmentation?

@gkioxari
Copy link
Owner

gkioxari commented Mar 5, 2018

I believe this part of the code answers your questions

https://github.com/gkioxari/RstarCNN/blob/master/lib/data_layer/minibatch.py#L78

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants