-
Notifications
You must be signed in to change notification settings - Fork 2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Segfault in node:10-alpine(3.10) when running in AWS and unprivileged only #1158
Comments
I see a similar issue. Switching to |
Same issue, same mitigation as OP running on AWS Kops 1.14 instances. Our other clusters with lower Kops version are not affected so it seems to be something tied to OS. |
@schlippi curious about the kops thing there. kops 1.14 with k8s 1.13 also fails? |
Me too... I'm using But it seems this commit: Since then, the tag I guess this Segfault in node:10-alpine issue happens a lot on AWS this week. Two options to fix:
|
I would add its not happening only in AWS. Same issue started to appear in my on-premise environment specially with services using node-rdkafka library. Changing services to us node:10-alpine3.9 fixed the issue. |
@olemarkus |
re #1158 (comment): The stack size issue crash from #813 (comment) is fixed in Node in nodejs/node@5539d7e (I ran a bisect), which is included in v13, but not v10. A related libuv fix (libuv/libuv@d89bd15) is also included in Node v13. That crash reproduces on both |
We have encountered the same issue when using the Skylake cpu platform on GPC while trying to run the grpc package |
fwiw I ran into this recently with node:10-alpine 3.10 and 3.11. I can get node:10-alpine 3.10 and 3.11 to work if I copy Docker's default seccomp profile, add membarrier to the whitelisted syscalls and use that |
@tomelliff could you help to disclose the spec of the AWS VM, docker version, etc? Since I failed to reproduce this. Thank you! |
Slightly confused by what's happening here but we have a Docker image that was built with node:10-alpine and when this was rebuilt with Alpine 3.10 instead of 3.9 after this was moved to 3.10 we had the container enter a crash loop with a segfault during the startup of the application.
Rolling back to the previous image and also rolling forward with things pinned to node:10-alpine3.9 seems to make this go away. More weirdly I can't reproduce this on non AWS instances but can reliably reproduce on multiple AWS instances. I also noticed that when the container is running with
--privileged
then it works fine.Looking at the non debug build core dump it looks to be an issue in musl but I don't yet know what's triggering that without debug symbols:
I'm also very confused why it wouldn't be segfaulting like this when it's ran outside of AWS or when the container is ran as privileged.
Any ideas on how I can debug this further?
The text was updated successfully, but these errors were encountered: