-
Notifications
You must be signed in to change notification settings - Fork 2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
token revocation 400s: panic: runtime error: invalid memory address or nil pointer dereference #6540
Comments
Since this cluster is a development cluster, I'm going to leave it non-functional for the moment. If you have questions or would like more information, please let me know. |
So sorry you hit this @dekimsey, but thanks for reporting it! Obviously it's a top priority for us to address. Is this a fresh cluster or did you upgrade from 0.9 or a 0.10 prerelease? Do note that if you upgraded from 0.10.0-beta1 you'll need to set |
I think I did a rolling update from a 0.8.4. No other changes were made to configs. I think the vault token may have expired (had jobs that couldn't get their secrets) and I went and re-issued those after the upgrade. |
I was able to provision a new job earlier today that worked just fine. The only difference I can imagine here was I was running run on a job that had existed prior to the upgrade. Note: I've updated the report to include server configuration settings. |
Hi @dekimsey, could you try running this version that @schmichael built with his patch and see if it solves the problem you are having? |
Thanks @lgfa29, I've gone home for the day. But I'll give it a try tomorrow morning! Thank you all for the quick turnaround. |
That resolved it, the cluster started and appears to be healthy! |
I'm going to lock this issue because it has been closed for 120 days ⏳. This helps our maintainers find and focus on the active issues. |
Nomad version
Operating system and Environment details
CentOS 7
Issue
I'm not really sure to be honest. I was trying to do a 'nomad run' a job, servers crashed and lost leadership. Looks like it was trying (and failing) to revoke a token. Logs don't tell me what the token accessor is so I cannot check the Vault audit logs to see what might be failing. Vault appears to be operating normally (I do see successful accesses).
At this time I'm unable to restore the cluster leadership, servers won't start. Not sure how to recover. From what I can tell the leader that wins election is unable to process the raft log and crashes few seconds later a new leader is elected and then crashes. After that the last node cannot elect a new leader and is stuck running but unable to complete an election.
Reproduction steps
Job file (if appropriate)
https://gist.github.com/dekimsey/1c7e761936bd3cfdfb749b65f43c11f7
Nomad Client logs (if appropriate)
Nomad Server logs (if appropriate)
Attempted to restart the node:
Server Configuration
Environment:
configuration:
The text was updated successfully, but these errors were encountered: