-
Notifications
You must be signed in to change notification settings - Fork 29
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Voxel Size #56
Comments
Hi @Steven-m2ai , The overfitting issues are strange from my point of view, 2000 image should be enough. Can you share your config, train/val metrics, some info about dataset (indoor/outdoor, number of classes, ...)? I think number of voxels and the voxel size are important for model quality, but it is probably not connected with overfitting. Smaller voxels may lead to better accuracy, but much more memory for 3d convolutions. Are you sure your projection matrices are fine? You can somehow visualize that your 3d object centers are projected to 2d object centers here. You can also try smaller model to prevent overfitting, e.g. ResNet50 -> ResNet18. |
Hello @filaPro, Thank you for your response. A little about the dataset: Okay I understand that the voxel size is important for quality, since the 3D convolution size relies on them. The projection matrices should be fine since I plotted all the ground truths of my dataset and they seem to be good. I haven't tried playing with the learning rate, or smaller model yet. Maybe this is a good road to take. The Config File
|
I think yes, if there is something more to share.
Have you plotted with our visualization functions? SUN RGB-D axis order or smth may differ from your coordinate system...
You should set So, what are your metrics on train / val? Are you able to achieve 100% accuracy on small subset of train? |
Yes, I will send you an email with more information. Yes, I plotted using by using the same method that you use. I simply modify your visualization function to show ground truth instead of predictions. Yes, I have set n_classes to 1 I am able to achieve a super high mAP (0.8) on a test set that is a subset of the training, but when using a new unseen set, i get very low (0.24) mAP. I will send you more information for extra visualizations via email. |
But in your config it is 10. However it is not important. I saw your images. It must be much better on test for 2000 images in train. I thing something is wrong with the training or inference. Can you check without |
oh yes what i mean is I just changed it to 1 in my config when you pointed it out. Yes I believe that wouldn't solve the overfitting but thanks for catching that. i am training without Do you think maybe the issue is the low diversity in the dataset? i.e. the frames do not look very different from each other, thus maybe there becomes a data imbalance? |
Hard to help you here :( Does reducing the model size help? |
Hello, Yes i guess this is a hard problem to debug. I will keep thinking about this. Conceptually, is it true that for the indoor head, there is no concept of "anchor boxes"? Rather each voxel acts as a center point, in which we use the ground truth min/max in each dimension to get the delta(x,y,z)? Then these deltas are the targets the model tries to predict? Perhaps I can open a separate issue for conceptual questions, i am very interested in your work. i would love to really understand the pipeline implemented here. |
Hello,
I am experimenting with a custom dataset on ImVoxelNet.
My dataset is ~2000 images, and i am running into extreme overfitting issues. For example the prediction on the validation image is in the same pattern as some of the training image predictions.
I was looking through what could be the case, I guess I could try playing with the lr and scheduler. however, I was also looking into voxel size and number. Do you think this could have any affect on the outcome? Any other advice? Thanks!
The text was updated successfully, but these errors were encountered: