-
Notifications
You must be signed in to change notification settings - Fork 986
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Handling dead Sentinel slaves #669
Comments
Thanks a lot. Host |
Lettuce now considers Sentinel messages on channels +sdown, -sdown and +slave as signals to refresh topology. This change allows to add new nodes during runtime and close connections to nodes that are not available (connectable). Temporary failures to connect a slave result in closing the client connection until the node is reachable again.
Lettuce now considers Sentinel messages on channels +sdown, -sdown and +slave as signals to refresh topology. This change allows to add new nodes during runtime and close connections to nodes that are not available (connectable). Temporary failures to connect a slave result in closing the client connection until the node is reachable again.
Lettuce now considers Sentinel messages on channels +sdown, -sdown and +slave as signals to refresh topology. This change allows to add new nodes during runtime and close connections to nodes that are not available (connectable). Temporary failures to connect a slave result in closing the client connection until the node is reachable again.
Lettuce now considers Sentinel messages on channels +sdown, -sdown and +slave as signals to refresh topology. This change allows to add new nodes during runtime and close connections to nodes that are not available (connectable). Temporary failures to connect a slave result in closing the client connection until the node is reachable again.
The issue is caused by not considering messages on |
@mp911de Tried this version:
It did not work!. Logs exploded (notice the timestamp) with this error messages (ip of dead slaves):
topology was not updated, and there is endless error spam in logs. |
From the stack traces, it looks that the connected sentinel did not recognize the slave is actually down. You could trace the issue down yourself by connecting to all sentinels with |
lettuce
5.0.2.BUILD-SNAPSHOT
in Kubernetes in1.8.0
. Sentinel keeps unreachable nodes, which is normal. After master pod kill, lettuce unable to update topology (sentinel worked fine), trying ping dead node and retries forever.BEFORE taking down master (note: there already 2 dead slave nodes when I started application and lettuce is working fine):
AFTER:
And lettuce trying to ping
10.233.101.45
node, unable to refresh topology.The text was updated successfully, but these errors were encountered: