-
Notifications
You must be signed in to change notification settings - Fork 34
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Make workers configurable #254
Comments
It sounds possible. The transform can load the config and tweak the worker field in the task |
Note that transforms will work if each worker pool has a consistent value. If we have multiple workers in the same pool that require different values, we won't know which ones are correct until runtime. IMO, in an ideal world, the workers would set GPUS/WORKSPACE themselves, and the tasks would inherit these values (since they are a property of the worker). I'm not sure at this point if that's possible. If not, runtime detection in a script is necessary for this case. |
Yes, ideally those values should depend on a worker. Related to #253 |
We talked about this a bit on Zoom today. We agreed that in the short, and maybe medium, term that we would keep worker pools consistent as far as their type and number of GPUs go. With this in mind we should be able to define the When we have other worker pools we want to train on, we'll just need to adjust those entries and the worker provisioner/worker-type in
When the on prem machines are available, this entry would change to something like:
(The |
Currently, it requires changing each step if we say, want to train the whole pipeline on a different worker pool. For example, we might want to experiment with different GPU types or run the pipeline on the on-prem cluster and set the specific GPU we want to use.
Can we configure workers from the training config? I think it should be possible with transforms.
The text was updated successfully, but these errors were encountered: