-
Notifications
You must be signed in to change notification settings - Fork 3.8k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
cli: cockroach quit
should only succeed as long as ranges will not become unavailable
#43412
Comments
cc @johnrk @andy-kimball for triage |
This issue will need discussion about how to handle cases where all nodes are brought down at once. Does |
Andy's question mirrors a common problem we see with decommissioning nodes - often users will be confused about why a node they are decommissioning doesn't appear to be draining. After some investigation, we'll realize the problem is that they are trying to decommission a node with a 3x replication factor and only 3 nodes. The decommission starts and then hangs since there isn't a location for the replicas to go. We should think about how to warn users about this for |
Update on the original issue at top:
Update on the specific questions about
|
The remaining work to be done (identified by my first bullet point above) is exactly the work described in the top description for #44206, so I'll close this issue in favor of that one. |
Is your feature request related to a problem? Please describe.
Currently, today users have to look at several different places to make sure that taking a node down in a Cockroach cluster will not create unavailable ranges. Namely today, a user would have to check
system.replication_stats.under_replicated_ranges
to make sure there are no under replicated ranges and check to make sure all nodes show up asis_live
incrdb_internal.gossip_nodes
.Describe the solution you'd like
Instead, the
cockroach quit
command should not return successfully if taking a node down would make data unavailable.The text was updated successfully, but these errors were encountered: