Replies: 6 comments 10 replies
-
Hi @Can-Zhao, could you please help share some comments on this discussion? |
Beta Was this translation helpful? Give feedback.
-
Hi @apourchot Thank you for reaching out. I do not think you will need to worry about negative samples. If I understand correctly, you were worrying that there is no negative sample, so the network tends to predict positive samples (boxes) everywhere. In detection, positive sample means a box. But in an image, most of the voxels are not associated to a box, and they are all negative samples. If during inference, a model predicts boxes everywhere on the image, e.g., 10000 boxes on an image, it will give a high false positive rate, and it is reflected in metrics. Hope it helps. Please feel free to let me know if needed. |
Beta Was this translation helpful? Give feedback.
-
@apourchot Apologies for any misunderstanding caused. The initial dataset splits were generated using the |
Beta Was this translation helpful? Give feedback.
-
Upon reviewing the feedback, we have initiated a thorough examination of our data processing procedures and found that the images without nodules were omitted in data split json files unintentionally. We are in consultation with my colleagues to gain a comprehensive understanding of every step involved in our data handling. It is important to note that any errors in our data were not intentional. We are dedicated to rectifying any inaccuracies and implementing measures to prevent similar issues in the future. We will generate correct data split json files and recompute the metrics soon. We value the feedback from our community and are grateful for the opportunity to improve. We welcome any further suggestions or insights that could assist us in this process. |
Beta Was this translation helpful? Give feedback.
-
Thank you very much for acknowledging the problem. Looking forward to the updated splits and metrics 🙂 |
Beta Was this translation helpful? Give feedback.
-
Hello @apourchot, _________ EDIT ________ (Maybe related to my recent post here on the GitHub of nnDetection: MIC-DKFZ/nnDetection#272) I observe something like you I think. When I include "healthy" patients, the FROC score is decreased. I was explaining this like you, but just by hand, based on the fact that with no object to detect, only FP are possible, thus pushing scores toward lower values for fixed FP/scan. But a colleague challenged this explanation saying that for example if we have a score of Se 80% @ 2FP/scan in average, if we add negative healthy patients, sensitivity will remain the same, and we could expect 2FP/scan in average still on the new patients. Sounds also good to me... So I am a bit lost since I definitly observe this phenomenum. Do you have elements of answer or explanation of this phenomenum please. Thank you in advance. Best, |
Beta Was this translation helpful? Give feedback.
-
Hi,
I've been going through the dataset split files from the luna16 tutorial and noticed there aren't any negative cases, i.e. scans without bounding boxes. I get why they might be left out in training, but they seem to be missing in the validation datasets too. Could you help me understand a couple of things?
When did this approach start? It's surprising that the original challenge creators would skip negative cases, but maybe I'm missing something. I looked over the nnDetection scripts for making the splits and didn't find anything about filtering out scans based on the number of bounding boxes. But, again, I might have missed something.
Can you confirm if the metrics on the tutorial page include or exclude negative cases? If they're excluded, it seems like the metrics might be off, most likely over-estimated?
Thanks in advance for any info you can share!
Beta Was this translation helpful? Give feedback.
All reactions