-
Notifications
You must be signed in to change notification settings - Fork 25k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[ML] Update the number of allocations per nlp process #86277
Conversation
Pinging @elastic/ml-core (Team:ML) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks good. Just a minor suggestion for a potential method extraction.
TimeValue timeout, | ||
ActionListener<ThreadSettings> listener | ||
) { | ||
var processContext = getProcessContext(task, listener::onFailure); |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
There seems to be another method which we could extract here that is something like executePyTorchAction
and takes in an AbstractPyTorchAction
. Then we could reuse it when we fire either the inference action or the control message action and it does the getting of the process context and the try-catch of running the action.
I might be missing something that makes it impossible to do this. In any case, just a thought.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
++ yeah there was an opportunity to refactor here
d439525
to
dee9ed1
Compare
dee9ed1
to
f950f4c
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
During elastic#86277 an error was introduced in parsing of pytorch thread settings. This commit fixes the issue.
During #86277 an error was introduced in parsing of pytorch thread settings. This commit fixes the issue.
During elastic#86277 an error was introduced in parsing of pytorch thread settings. This commit fixes the issue.
This is the Java side of elastic/ml-cpp#2258 which causes internal breakages while the PRs are out of sync due to naming changes.
Adds a method to
DeploymentManager
to update the number of allocations per process as implemented in elastic/ml-cpp#2258.Also
PyTorchResults
now has an error type rather than the error being a special case of the inference result and reverts the test mutes in #86263