-
Notifications
You must be signed in to change notification settings - Fork 590
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
KIC looping forever on Context Deadline exceeded #2440
Comments
Adding some detail: KIC seem so be looping due to some evaluated diff on upstream which are default based (since I didn't configured them) Note the minus sign in the trace below
Once I add a KongIngress in order to add the "missing" values the loop on the upstream ended and started looping on Services items with same "diff" logic:
|
Hi there, I manage to update to kong 2.8.1 + KIC 2.3.1 and all loops/traces totally disappear. After a week of struggling I really think that kong 2.7.* + KIC 2.2.* should be considered an unstable release and removed from the available downloads! |
Increasing We do not remove older images because doing so would break instances that use them unexpectedly if, say, the scheduler tried to start a replica on a worker without a cached copy of that image. Not all users are equally affected by all issues. Many did run fine despite the unnecessary updates--those were a fairly long-standing issue that we only recently acquired the means to fix effectively. |
Is there an existing issue for this?
Current Behavior
I've a kong 2.7.1 + KIC 2.2.1 setup con GKE
I've ~50 namespace with 2 ingress resources each resulting in ~30 rules per namespaces so a total amount of ~1500 rules.
Some services aslo has plugins configured.
Once deployed KIC, starts to populate kong, but it periopdicaly fails on error such:
{"error":"2 errors occurred: failed to sync all entities: context deadline exceeded while processing event: {Update} upstream *******.8080.svc failed: making HTTP request: Put "http://kong-kong-admin.kong-system:8001/upstreams/1e7b3b64-fdbf-4f51-9c9e-2cb43172d92b": context deadline exceeded ", "level":"error", "msg":"could not update kong admin", "subsystem":"proxy-cache-resolver"}
on kong admin side matching error is:
[error] 1097#0: *54395 [lua] events.lua:364: post_local(): worker-events: dropping event; waiting for event data timed out
I've split my deployment in 3 standalone kong proxy/admin pods + 3 KIC pods using the kong services to spread load across the 3 admin nodes.
I've also tried to scale up kong pods up to 5 and reduce KIC concurrency to 1 but still error persists (the number in error log is usually concurrency number+1 )
One KIC enters such error it continues looping on different services names flooding admin with requests.
Kong pods have no resources limit and each kong pod is allocated on a different VM, both CPU and RAM are NOT saturated.
Is there any other configuration tuning I can apply to avoid such loop?
Expected Behavior
KIC should be able to send all configuration to kong and stop looping
Steps To Reproduce
No response
Kong Ingress Controller version
No response
Kubernetes version
No response
Anything else?
both kong and KIC are deployed via helm with such setup (concurrency has been lowered up to 1):
and such custom setting in nginx template:
such setups also suffers from #2422, I'm waiting for next release to upgrade to 2.8
The text was updated successfully, but these errors were encountered: