You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
deploy a local minikube (with metrics server) or crc (with monitoring enabled), run the same set of tests we do in our pipeline, and check the API pods' resources utilization
Actually, the API pods were not crashing but getting restarted because of liveness probe failures.
After doing some tests and modifying the show_logs.sh script from CI I could see:
Warning Unhealthy 4m15s (x5 over 5m35s) kubelet Liveness probe failed: Get "http://10.244.0.9:24817/pulp/api/v3/status/": dial tcp 10.244.0.9:24817: connect: connection refused
Checking the logs from a restarted pod, it was running migration tasks when the container got terminated.
As a workaround, we can increase the default failure threshold for the liveness probe while #991 is not implemented.
git-hyagi
added a commit
to git-hyagi/pulp-operator
that referenced
this issue
Jul 5, 2023
Verify if the API pods crashing is related to a memory leak (https://discourse.pulpproject.org/t/api-server-memory-leak/851/12).
Some ideas to help with the investigation:
kubectl logs -p
) in theshow_logs.sh
script to see if we can get more informationhttps://github.com/pulp/pulp-operator/blob/main/.github/workflows/scripts/show_logs.sh#L30
The text was updated successfully, but these errors were encountered: