-
Notifications
You must be signed in to change notification settings - Fork 8.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
YARN-9509: Added a configuration for admins to be able to capped per-container cpu usage based on a multiplier #766
base: trunk
Are you sure you want to change the base?
Conversation
…pu usage based on a multiplier
💔 -1 overall
This message was automatically generated. |
🎊 +1 overall
This message was automatically generated. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Hi @ashangit , thanks for your patch. But I have some suggestions for the style nits.
@@ -138,13 +139,18 @@ void initConfig() throws IOException { | |||
.NM_LINUX_CONTAINER_CGROUPS_STRICT_RESOURCE_USAGE, | |||
YarnConfiguration | |||
.DEFAULT_NM_LINUX_CONTAINER_CGROUPS_STRICT_RESOURCE_USAGE); | |||
this.cappedMultiplier = |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
this.cappedMultiplier = | |
this.cappedMultiplier = conf.getFloat( |
@@ -138,13 +139,18 @@ void initConfig() throws IOException { | |||
.NM_LINUX_CONTAINER_CGROUPS_STRICT_RESOURCE_USAGE, | |||
YarnConfiguration | |||
.DEFAULT_NM_LINUX_CONTAINER_CGROUPS_STRICT_RESOURCE_USAGE); | |||
this.cappedMultiplier = | |||
conf |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
conf |
@@ -138,13 +139,18 @@ void initConfig() throws IOException { | |||
.NM_LINUX_CONTAINER_CGROUPS_STRICT_RESOURCE_USAGE, | |||
YarnConfiguration | |||
.DEFAULT_NM_LINUX_CONTAINER_CGROUPS_STRICT_RESOURCE_USAGE); | |||
this.cappedMultiplier = | |||
conf | |||
.getFloat( |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
.getFloat( |
🎊 +1 overall
This message was automatically generated. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
+1 (non-binding)
💔 -1 overall
This message was automatically generated. |
🎊 +1 overall
This message was automatically generated. |
🎊 +1 overall
This message was automatically generated. |
💔 -1 overall
This message was automatically generated. |
🎊 +1 overall
This message was automatically generated. |
💔 -1 overall
This message was automatically generated. |
💔 -1 overall
This message was automatically generated. |
💔 -1 overall
This message was automatically generated. |
🎊 +1 overall
This message was automatically generated. |
💔 -1 overall
This message was automatically generated. |
💔 -1 overall
This message was automatically generated. |
💔 -1 overall
This message was automatically generated. |
💔 -1 overall
This message was automatically generated. |
Add a multiplier configuration on strict resource usage to authorize container to use spare cpu up to a limit.
Currently with strict resource usage you can't get more than what you request which is sometime not good for jobs that doesn't have a constant usage of cpu (for ex. spark jobs with multiple stages).
But without strict resource usage we have seen some bad behaviour from our users that don't tune at all their needs and it leads to some containers requesting 2 vcore but constantly using 20.
The idea here is to still authorize containers to get more cpu than what they request if some are free but also to avoid too big differencies so SLA on jobs is not breached if the cluster is full (at least increase of runtime is contain)