-
Notifications
You must be signed in to change notification settings - Fork 312
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[BUG] deviceshare plugin not handle AddPod\RemovePod correctly #1959
Comments
buptcozy
pushed a commit
to buptcozy/koordinator
that referenced
this issue
Mar 18, 2024
buptcozy
pushed a commit
to buptcozy/koordinator
that referenced
this issue
Mar 18, 2024
buptcozy
pushed a commit
to buptcozy/koordinator
that referenced
this issue
Mar 18, 2024
buptcozy
pushed a commit
to buptcozy/koordinator
that referenced
this issue
Mar 18, 2024
buptcozy
pushed a commit
to buptcozy/koordinator
that referenced
this issue
Mar 18, 2024
buptcozy
pushed a commit
to buptcozy/koordinator
that referenced
this issue
Mar 18, 2024
buptcozy
pushed a commit
to buptcozy/koordinator
that referenced
this issue
Mar 18, 2024
buptcozy
pushed a commit
to buptcozy/koordinator
that referenced
this issue
Mar 18, 2024
buptcozy
pushed a commit
to buptcozy/koordinator
that referenced
this issue
Mar 18, 2024
buptcozy
pushed a commit
to buptcozy/koordinator
that referenced
this issue
Mar 20, 2024
buptcozy
pushed a commit
to buptcozy/koordinator
that referenced
this issue
Mar 20, 2024
…1959) Signed-off-by: xingbao.zy <[email protected]>
Problem DescriptionLet's analysize by following example.
Suggested ProposalLet's design by the following examples. Example1
Example2
Scheduling Interpretability
|
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
What happened:
in generally, when we execute AddPod logic here, the pod may be in scheduling status,it won't exist in nodeDeviceCache's used map, so there is a bug that when the framework execute RunFilterPluginsWithNominatedPods with AddPod for high priority pods, the plugin can't reserve resource for hese high priority pods, In RDMA\VF\nv-switch scenario, it can cause high priority pods assign fail due to some resources is assigned by low priority pods. So we reused the "Reserve" logic to generate an assign placement and save it in nominator. We will clear the nominator cache In "Reserve" and "UnReserve", which means we will do clean job no matter assign success or not, this is the same process of the origin k8s framework nominate process.
What you expected to happen:
How to reproduce it (as minimally and precisely as possible):
Anything else we need to know?:
Environment:
kubectl version
):The text was updated successfully, but these errors were encountered: