You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository has been archived by the owner on Jul 26, 2022. It is now read-only.
Hi, I was wondering if there is any advice on a weird issue we are seeing. I have used external secrets in other jobs and I've never seen instability like this before so it must be something to do with the setup here, but I'm not sure what to look for.
Randomly, the external secrets pod will simply stop syncing secrets. There's not always an error in the logs, but if you kubectl logs -f <podname> you can see the logs are not moving at all. If I delete the pod and let the cluster spin up a new one, all of a sudden things kick back into gear again. I tried to do some digging and found this error on one of the external secrets deployments:
{"level":30,"message_time":"2021-06-17T09:24:10.720Z","pid":17,"hostname":"external-secrets-fd64d899d-qcr8w","msg":"starting poller for monitoring/prometheus-alertmanager-config"}
failed to watch file "/var/lib/docker/containers/8a6570089510d5eb9d1d8e79365fde1cfaa6f18c20f444466be3220311cc86e4/8a6570089510d5eb9d1d8e79365fde1cfaa6f18c20f444466be3220311cc86e4-json.log": no space left on device%
However it seemed like at the time the secrets were indeed syncing ok because when checking a random secret in a namespace with kubectl get externalsecrets i could see the sync time was 10s and the status was SUCCESS. When it gets stuck, the sync status field is blank.
Following on from the error above which appears to be about disk space, i checked the space on the pod which was in the 80-90% range, so not completely full. I then checked the actual node which the pod was running on which I will put below. But long story short, i couldn't see any disk space issues on there.
Hi, I was wondering if there is any advice on a weird issue we are seeing. I have used external secrets in other jobs and I've never seen instability like this before so it must be something to do with the setup here, but I'm not sure what to look for.
Randomly, the external secrets pod will simply stop syncing secrets. There's not always an error in the logs, but if you
kubectl logs -f <podname>
you can see the logs are not moving at all. If I delete the pod and let the cluster spin up a new one, all of a sudden things kick back into gear again. I tried to do some digging and found this error on one of the external secrets deployments:However it seemed like at the time the secrets were indeed syncing ok because when checking a random secret in a namespace with
kubectl get externalsecrets
i could see the sync time was 10s and the status wasSUCCESS
. When it gets stuck, the sync status field is blank.Following on from the error above which appears to be about disk space, i checked the space on the pod which was in the 80-90% range, so not completely full. I then checked the actual node which the pod was running on which I will put below. But long story short, i couldn't see any disk space issues on there.
Node Details
I also checked the inode status as well but I must admit I'm not very clued up on troubleshooting issues caused by inode allocations:
Inode info
Any advice would be appreciated!
The text was updated successfully, but these errors were encountered: