-
Notifications
You must be signed in to change notification settings - Fork 748
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Container Restarts on 1.6.1 (EKS 1.15) #1054
Comments
Hi @InAnimaTe, I wonder if you could get throttled by EC2? That would cause the check in |
Gotcha, so you're suggesting going to 1.6.3 which incorporates those changes, and then tuning up the EDIT: Actually, I can just wait until the next release that includes #1028 I suppose. Let me know your thoughts (and when that might come out) |
@InAnimaTe Hey if this was a test cluster, would you be interesting in testing the latest release candidate v1.7.2-rc1 to see if this resolves the issue you saw here? In particular, #1186 changes the start up behavior by actually waiting for iptables to be available to update. |
@InAnimaTe Hi, have you tried with v1.7.2 or later versions? Are you still seeing this restart issue? |
Hey @mogren apologies for the delays. I've been on vacation and incredibly busy with some other work things lately. I have not tried 1.7.2 yet but I plan to start testing it over the next couple weeks. I'm going to close this issue for now and if I see the problem again, we can re-open. I strongly suspect that #1186 and other changes (#1028) your team has made will probably solve these issues for us. Thanks so much for your continued support with these issues and the great work to make the AWS CNI better. |
I'm seeing two types of errors from readiness and liveness probes in our Kubernetes event stream (provided by Datadog):
Unhealthy: Liveness probe failed: OCI runtime exec failed: exec failed: container_linux.go:349: starting container process caused "read init-p: connection reset by peer": unknown
Unhealthy: Liveness probe failed: OCI runtime exec failed: exec failed: container_linux.go:349: starting container process caused "process_linux.go:101: executing setns process caused \"exit status 1\"": unknown
This Pod,
aws-node-wfx2t
is currently running and shows 8 restarts. Here's its Describe output with information on the Last State:I was inspired to take a look at our instances by seeing #1038 although I don't think this is the same issue.
The text was updated successfully, but these errors were encountered: