-
Notifications
You must be signed in to change notification settings - Fork 4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Compare DeepSpeech (w/Dropout) vs DeepSpeech (w/o Dropout) + BatchNorm #373
Comments
@kdavis-mozilla any progress on this ? What is is reason behind dropout of current layer being (1- dropout) of previous layer ? |
The minus one has nothing to do with this issue. TensorFlow uses keep probabilities not dropout rates, thus the minus one. This issue asks how does performance change when dropout is exchanged for batch norm. |
@reuben : I have few doubts for this one.
Please correct me if my understanding is incorrect. |
|
Closing for lack of activity. |
This thread has been automatically locked since there has not been any recent activity after it was closed. Please open a new issue for related bugs. |
A bonus, no more optimizing dropout rates.
The text was updated successfully, but these errors were encountered: