-
Notifications
You must be signed in to change notification settings - Fork 1
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
KilledWorker error in the deployment #5
Comments
@paigem, is it possible that any of the variables added in #4 are fields which, even when divided into 5 subsets
are still > 800 MB in size? For example, if any of the variables added were >= 5 GB in size, then with our current subset inputs, each subset will be >= 1 GB in size, which could easily kill a worker. If this is the case, the way to solve it would be to increase Also, just noting that dealing with these troublesome infrastructure concerns is one of many problems which the work scoped in pangeo-forge/pangeo-forge-recipes#256 should solve once complete. |
Thanks @cisaacstern for explaining this. Each individual netCDF file (1 year of daily output) is just under 1 GB. Since I'm stringing together 9 years of data, then the full time series for each variable would indeed be roughly 9 GB, which is > 800MB per chunk. I decided to go for the |
Yes, |
It looks like the most recent recipe run failed again, this time with a
KilledWorker
error. This appears to be a different failure than we saw in this previous recipe run, which did not result in an explicit error, but only processed a single chunk in time.@cisaacstern Any thoughts on how to get passed a
KilledWorker
error?The text was updated successfully, but these errors were encountered: