-
Notifications
You must be signed in to change notification settings - Fork 543
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Co-scheduled pods never get scheduled #416
Comments
Below message seems to be incorrect:
I only have 8 pods but it reports |
@asm582 can you try v0.23.10? |
How many pods speciy pgName=rag-pg? |
8 pods in a podgroup |
@Huang-Wei We see same issue with version
|
@asm582 could you describe the detailed reproducing steps from scratch? |
That message means |
ok, thanks for the confirmation, so the message has a bug it reports pods instead of nodes. |
well, it's not a bug, if you look at the regular message, they all follow the pattern:
|
Got it, thanks! [node count] [reason message] is the format, I am a new user, and may it's just not obvious to me. adding something more descriptive could help I think |
@Huang-Wei I confirm that the performance issues are gone but there is an issue wrt to events that co-scheduler fires. if the cluster has all nodes tainted and pods do not have tolerations then co-scheduler generates a message : |
More of a UX improvement. Why you saw it is b/c some of the pods in the PodGroup are "half-completed", i.e., scheduling constraints for them have been satisfied, so they're internally waiting for other sibling pods to be completed scheduling. However, other portion are not schedulable, so in the PostFilter, the optimization is to cancel waiting for those "half-completed" ones as they cannot succefuflly coscheduled, so release the resource they're holding is not a bad idea. That's why you saw different msgs for different pods - the pods can be scheduled (i.e., only waiting for PodGroup constraint), and the pods cannot be scheduled (i.e., due to lack of resources). |
Thanks for coming back :)! No pods were actually scheduled. The Test minikube environment has only one node which is tainted, and the pods scheduled have no tolerations. In such a scenario co-scheduler does not create |
How can pods without toleration be scheduled onto a tainted Node?
I'm a bit confused. If the pods mapping to a PodGroup doesn't meet the quorum, it should report an error like:
While if the quorum is met, but other scheduling constraints (like res) are not met, it reports the specific scheduling error like:
|
ok, got it, thanks, I don't have the setup ready yet to reproduce the issue. I did learn new things from the comments in the post. Please feel free to close this issue and I can always re-open the issue later. |
Sure, feel free to reopen if you can reproduce or create issues for new topics. |
Hello,
I have a cluster with 10 nodes, I am co-scheduling 8 pods but they are always pending, below are the events seen in one of the pods:
Co-scheduler version
k8s.gcr.io/scheduler-plugins/kube-scheduler:v0.22.6
The text was updated successfully, but these errors were encountered: