-
Notifications
You must be signed in to change notification settings - Fork 1
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
MonoVertex pods often unhealthy in e2e test #314
Comments
Hey @chandankumar4 - I unassigned this from you. Instead, I'll try running it again and since Sidhant has now run our e2e himself locally, he could be the one to look at it if it's occurring. |
Just re-ran this locally. It's after we update the MonoVertexRollout that the Monovertex is in a crash loop with this error:
|
Hey @dpadhiar - not super high priority, but would be good to fix the e2e test so that after updating MonoVertexRollout, the MonoVertex Pod is not in a crash loop (see log above) |
I see, looks like the version I change the upgrade to (from stable to 0.6.0) causes an issue. Will change that soon. |
Describe the bug
I'm not sure that this is any issue on our side, but would be worth investigating. Ultimately, could be something to hand over to Numaflow team to look at after some analysis on our side.
I was seeing that the MonoVertex pod was in a crash loop at the very end of the e2e test. I'm not sure if it's consistent or not, but I've seen it more than once. (Perhaps it's okay and it eventually fixes itself?)
This is the CI log from the test I ran locally:
ci.log.txt
These are the outputs from
tests/e2e/outputs
directory:output.zip
If you look at
outputs/resources/monovertexrollouts/pods
you can see many Pods in there, which seems to indicate that the Pods restarted a lot.To Reproduce
Steps to reproduce the behavior:
I assume this also happens for DATA_LOSS_PREVENTION=false, but I didn't try it.
Message from the maintainers:
Impacted by this bug? Give it a 👍. We often sort issues this way to know what to prioritize.
The text was updated successfully, but these errors were encountered: