-
Notifications
You must be signed in to change notification settings - Fork 6.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Pods faiiling after restart of VM #8850
Comments
HI , @dimakyriakov . I try it at ubuntu 20.04.4 LTS. Everything is OK. |
Hello, @yankay |
same problem here on a set of ubuntu 20.04 desktop VMs with kubespray commit c24a3a3 Install went fine with the follwing command on a setup with a single master node and 2 worker nodes.
Reboot is CHAOS !
I think it may be caused by the fact that cgroupv2 are disabled !
etcd is running fine. |
enabling cgroupv2 makes reboot of master nodes possible ... :(
The kubernetes cluster reboot but I still have the coredns pods that do not restart :
Logs of coredns on the worker node that did not restart shows :
I still need this tweaks on all nodes :
|
issue with coredns seems to be linked to #5835
I use KVM... |
|
using iptables instead of ipvs does not solve the core dns crash loop after reboot. |
From the https://coredns.io/plugins/loop/#troubleshooting link it seems that disabling systemd-resolved is worth a shot ... Hourra it works ! |
Great if you found a way to fix your coredns issue 👍 |
@floryut do you know is the test suite include a simple restart of the cluster ? |
They do not, it would be possible to add ones though. But come to think of your issue it is strange that you have to disable systemd-resolved, to have coredns work pointing the resolv.conf to |
The issues with coredns appeared only after restarting. The initial install went fine. |
I'll try to spinup a ubuntu cluster to see if I can reproduce this behavior but as you're the first to report it, it would be strange if this was a bug in our codebase |
I had a similar problem |
same problem here on debian
not solve the issue... |
The Kubernetes project currently lacks enough contributors to adequately respond to all issues. This bot triages un-triaged issues according to the following rules:
You can:
Please send feedback to sig-contributor-experience at kubernetes/community. /lifecycle stale |
The Kubernetes project currently lacks enough active contributors to adequately respond to all issues. This bot triages un-triaged issues according to the following rules:
You can:
Please send feedback to sig-contributor-experience at kubernetes/community. /lifecycle rotten |
The Kubernetes project currently lacks enough active contributors to adequately respond to all issues and PRs. This bot triages issues according to the following rules:
You can:
Please send feedback to sig-contributor-experience at kubernetes/community. /close not-planned |
@k8s-triage-robot: Closing this issue, marking it as "Not Planned". In response to this:
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. |
Environment:
hardware configuration:
OS Ubuntu 20.04.4 LTS
Version of Ansible: ansible 2.10.15
Version of Python3: Python 3.8.10
Kubespray version (commit): 2cc5f04
Full inventory with variables:
Command used to invoke ansible:
ansible-playbook -i inventory/newCluster/hosts.yaml --become --become-user=root cluster.yml
Output of ansible run:
Anything else do we need to know:
After I rebooted my VM where I was installed master k8s node - all pods can't up because this error "Error response from daemon: cgroup-parent for systemd cgroup should be a valid slice named as "xxx.slice""
The text was updated successfully, but these errors were encountered: