-
Notifications
You must be signed in to change notification settings - Fork 9
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Some questions about the restored results. #6
Comments
Your problem arises mainly because of this: (1)Low-light enhancement is a ill-posed problem where there are multiple reasonable understandings of different brightnesses for one low-light image. For a better comparison, many SOTA methods in the current LOL benchmark (such as the second-ranked LLFLOW) adjust the brightness of the image output by the network to be consistent with GT to compare the restoration of texture details. (2)In the code, I have adjusted the image output by the network to the brightness of the GT image according to the average value (this process will not affect the generated texture details, it is just the simplest method to adjust the overall illumination, and it is also easy to adjust according to user preferences in practical applications). To sum up, in your case, we will adjust the image output by the network to the brightness of low light, so the situation you said will appear (that is, low light is GT). For details, please refer to the code: PyDIff/PyDiff/pydiff/models/pydiff_model.py Line 216 in 4144b89
Resolution: When testing on your own dataset, set 'use_kind_align' in 'infer.yaml' to false. The same goes for training. |
Thanks for your quickly reply |
I have another question, based on your reply, so do we just need to use normal-light images captured in any scene as "GT" when testing on the unpaired data? |
What you said is also a solution. But you don't actually need the normal-light image as a reference. you can inference in two steps. |
Thanks, according to my reproduced results, by setting "gt_root" to the same value as "input_root" would cause similar results with input, I will try the first one. |
Hello, have you found a solution to this problem. I tried to set 'use_kind_align' in 'infer.yaml' to false and Set "gt_root" to the same value as "input_root",but the results are not good. |
Hi, I'm encountering some confusions in reproducing the results on the LOL dataset, when I set the “gt_root” and “input_root” of the dataset in the "infer.yaml" to be the same, I find that the illumination of the output results is not enhanced, whereas when the ”gt_root” is set to be the path of reference images, the results turn out to be normal. Is it necessary to have a reference image in the inference stage to get the restored normal-light result? But in real scenes we can only get low-light images, so how to test it on the unpaired dataset?
The text was updated successfully, but these errors were encountered: