-
Notifications
You must be signed in to change notification settings - Fork 103
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add retries to NNCP deployment #853
Add retries to NNCP deployment #853
Conversation
Skipping CI for Draft Pull Request. |
Build failed (check pipeline). Post https://review.rdoproject.org/zuul/buildset/cc15ac0dc5fa4477ae03b86426dbcbf8 ✔️ openstack-k8s-operators-content-provider SUCCESS in 45m 51s |
Build failed (check pipeline). Post https://review.rdoproject.org/zuul/buildset/e8039ea769ce4835aac615ced8f4491a ✔️ openstack-k8s-operators-content-provider SUCCESS in 2h 53m 35s |
recheck |
|
||
while true; do | ||
make nncp && break | ||
make nncp_cleanup |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
i feel like this is not a stable pattern
when we delete and recreate the resource even if its the same content its going to be a separate object at the k8s level so without an exponential backoof i.e. waiting longer each time we call make nncp i think its possible that we ould be in a loop just because we kept restarting before it could make progress.
should we increase the NNCP_TIMEOUT slightly on each retry?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
i feel like this is not a stable pattern
It's not ideal but it's a worst case recovery, we already give the NNCP CR 240s seconds to complete and I haven't seen any jobs failing due to timeout.
i think its possible that we ould be in a loop just because we kept restarting before it could make progress.
I haven't seen any job fail due to NNCP CR taking over 240s to apply. This loop is limited to 5 iterations so max it would be stuck for is 20mins and this would only be if there is a real issue with the NNCP CR that needs to be fixed, aka a real bug.
should we increase the NNCP_TIMEOUT slightly on each retry?
No I don't believe we should, I believe it's high enough already unless you have seen jobs failing to apply the NNCP in under 240 seconds?
Build failed (check pipeline). Post https://review.rdoproject.org/zuul/buildset/429261c567a74daa88e729d1ee8e15fa ❌ openstack-k8s-operators-content-provider POST_FAILURE in 3h 03m 41s |
recheck unrelated failure in post job collect logs failed:
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
If it helps bring some additional stability to CI jobs, then lgtm.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I had problem with nncp on my local environment and this pr fixed it so I'm giving LGTM. However I would be also interested why we need this oc apply retires in the first place.
[APPROVALNOTIFIER] This PR is APPROVED This pull-request has been approved by: bshephar, ciecierski, fao89, lewisdenny The full list of commands accepted by this bot can be found here. The pull request process is described here
Needs approval from an approver in each of these files:
Approvers can indicate their approval by writing |
3c1d04d
into
openstack-k8s-operators:main
Sometimes applying the NNCP CR fails due to once of the kubernetes-nmstate probes failing, usually the api-server.
There are two ideas considered to fix the issue:
Results:
[1] nmstate/kubernetes-nmstate#831
[2] https://github.com/nmstate/kubernetes-nmstate/blob/272b29459ca99a9de62af24f509afe9f3deb058f/pkg/probe/probes.go#L59-L71JIRA: https://issues.redhat.com/browse/OSPRH-7605