-
Notifications
You must be signed in to change notification settings - Fork 927
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Centralized HPA architecture #3379
Comments
cc @Garrybest @XiShanYongYe-Chang @chaunceyjiang what do you think? |
By the way,Will the new version adopt a centralized architecture ? When is the next version scheduled to be released? |
Hi @chenwei01 , The decentralized HPA has a higher priority. BTW, what's your use case? @chenwei01 |
@jwcesign got it, thanks, my scenario basically matches the use cases in the proposal. |
This feature is finished, let's close this issue |
What would you like to be added:
Decentralized HPA:
The concept of Decentralized HPA can be complex to comprehend and configure, leading to confusion among users. Additionally, it may result in duplicate jobs with the scheduler.
But it does have one advantage, though: if the control plane goes down, it can still handle burst requests.
Centralized HPA
Centralized HPA is quite easy to understand and configure, there are three kinds of architecture:

K8s native metrics server(CPU/Memory):
Customer metrics server:

By implementing this architecture, the HPA will solely focus on scaling deployments while the scheduler schedules replicas based on priority policies.
However, a drawback of this approach is that in case of control plane failure, it cannot manage sudden surge requests.
Why is this needed:
In the current era of single clusters, Horizontal Pod Autoscaling (HPA) is commonly used to dynamically scale workload replicas to handle incoming requests and enhance resource utilization. Therefore, in the era of multi-clusters, HPA should still work and be capable of scaling the workloads in different member clusters, to resolve resource limitations for a single cluster and improve scaling fail recovery capabilities.
The text was updated successfully, but these errors were encountered: