-
Notifications
You must be signed in to change notification settings - Fork 39.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Node NotReady
status with "Kubelet stopped posting node status error"
#34314
Comments
This issue came up again. I've tried debugging with
to view logs. Nothing out of the ordinary. Also fine here:
How can I debug this? |
Ok this was related to changing the |
I had the same issue. I'm using GKE (google container) |
I just ran into this - on GKE 1.5.1 with alpha features turned on The problem appeared when the cluster auto-scaled. The first node went to status NotReady The node was non-responsive - I could not ssh into it. Restarting the node cleared the status |
The same problem on CoreOS. k8s 1.5.2. After recreating /var/lib/kubelet directory and re-registering master node I get this repeating messages in the log:
Node status becomes "NotReady" and pods, created by ReplicationContorllers with NodeSelector value of this node, get status "Pending", reason: "MatchNodeSelector". Reboot does not make sence. |
I just had the same problem k8s 1.4.7 stable. Very little was running on my cluster (1 master, 2 workers) other than Deis, running on AWS launched by Kops. Both workers had the same problems as above. AWS CloudWatch reported everything was fine on all servers.
|
Problem solved by applying changes to kubelet configuration (/etc/systemd/system/kubelet.service) according to latest version of reference page on CoreOS: https://coreos.com/kubernetes/docs/latest/deploy-master.html |
I am also facing same issue. I see some issue after deploy the app. App is deployed success But running is 0. root@kubernetes: Can you suggest and give me the appropriate suggestion. root@kubernetes:~# kubectl get nodes Thanks |
@axsuul were you able to resolve the issue, can you share the details. I also encountered the same issue where the master and worker is on the same node (one node cluster). |
@viveksinghggits Sorry I ended up moving to Docker Swarm and I don't remember the details anymore, sorry |
I'm having the same issue on EKS with Kubernetes 1.12. Minimal steps to reproduce:
Rebooting the EC2 instance doesn't help. |
I'm having the same issue. Is the issue resolved ?. If yes, can anyone provide step by step instructions on resolving the issue? |
Happens to me as well in AWS EKS.
Can't log into the instance to inspect kubelet. Seems the instance is frozen or something Edit: Follow up here awslabs/amazon-eks-ami#79 |
Keep the powershell scripts running. Don't shutdown the windows machine or close the powershell window. If you had to restart the machine. Run the join script again.
Sent from Yahoo Mail on Android
On Thu, 4 Jul 2019 at 7:57 PM, Jean-Michael Cyr<[email protected]> wrote:
Happens to me as well.
Any hint ?
Conditions:
Type Status LastHeartbeatTime LastTransitionTime Reason Message
---- ------ ----------------- ------------------ ------ -------
OutOfDisk Unknown Thu, 04 Jul 2019 10:12:19 -0400 Thu, 04 Jul 2019 10:13:04 -0400 NodeStatusUnknown Kubelet stopped posting node status.
MemoryPressure Unknown Thu, 04 Jul 2019 10:12:19 -0400 Thu, 04 Jul 2019 10:13:04 -0400 NodeStatusUnknown Kubelet stopped posting node status.
DiskPressure Unknown Thu, 04 Jul 2019 10:12:19 -0400 Thu, 04 Jul 2019 10:13:04 -0400 NodeStatusUnknown Kubelet stopped posting node status.
PIDPressure False Thu, 04 Jul 2019 10:12:19 -0400 Thu, 04 Jul 2019 08:26:42 -0400 KubeletHasSufficientPID kubelet has sufficient PID available
Ready Unknown Thu, 04 Jul 2019 10:12:19 -0400 Thu, 04 Jul 2019 10:13:04 -0400 NodeStatusUnknown Kubelet stopped posting node status.
—
You are receiving this because you commented.
Reply to this email directly, view it on GitHub, or mute the thread.
|
Happen to me as well, started to happen when I was running the stress test against the services running inside cluster |
For me first you find any HPA that is exceeding resources delete the HPA will work |
Is there any solution to this issue? I have observed it when my cluster got autoscaled. The first worker node became Not Ready and its in the same status till now (After 8 hours). |
same issue here using minikube |
CreationTimestamp: Sun, 26 Jul 2020 18:41:43 +0700 MemoryPressure Unknown Sun, 26 Jul 2020 19:42:21 +0700 Sun, 26 Jul 2020 21:26:06 +0700 NodeStatusUnknown Kubelet stopped posting node status. |
Hi, the issue is closed. But does anyone has a solution for it? |
Maybe this thread helps you, you probably need to reserve resources for host daemons using kubelet args: rancher/rancher#29997 (comment) |
Thanks @immanuelfodor . |
The problem is likely to be, the memory and processing resources within the clusters don't match the workload. That is you have exhausted the cluster resources and you need to deploy more worker nodes. Restart all the instances all at once. Give them some time to reboot and restart all the Kubernetes resources on the cluster. Worked for me on AWS |
I got the same issue we have debuged using below commands it works really KUBERNETES:
[root@cerebro05 etc]# kubeadm join 10.87.208.94:6443 --token eah77w.1yfl82ahipkdr1da --discovery-token-ca-cert-hash sha256:15e3637fa73615d30b97c162e610709384c8a395755dd6bba7982cde1a458da8 |
Check swap on or off ---> free -m |
On k8s 1.4 and used
kubeadm
to provision cluster:I have node and master on same server. Suddenly by node is posting a
NotReady
status. Running areturns
I've tried restarting the server to no success. How would I debug this? Thanks
The text was updated successfully, but these errors were encountered: