-
Notifications
You must be signed in to change notification settings - Fork 274
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Response payload is not completed
when reading a file
#786
Comments
At first glance it looks like a pretty low level network error. Are you there is no intermittent condition on your connection? |
This is happening on a coiled cluster. While I can't guarantee it, I believe our network there is pretty stable. Retrying the task sufficiently often works but I would expect such an error to be retried further down the stack, definitely not on dask level |
|
Well, the traceback is blaming L2360 (2023.6.0) in async def _inner_fetch(fs, bucket, key, version_id, start, end, req_kw=None):
resp = await fs._call_s3(
"get_object",
Bucket=bucket,
Key=key,
Range="bytes=%i-%i" % (start, end - 1),
**version_id_kw(version_id),
**req_kw,
)
return await resp["Body"].read() which is the I haven't checked the logs, yet, but the code does not look as if it would retry this exception at this position |
A similar issue was apparently fixed for other APIs already, see #601 |
This traceback is from a dask workload that is reading hd5 files. Unfortunately, I don't have a reliable reproducer and don't know how to trigger this.
However, based on this traceback and the exception message I assume that some data transfer is prematurely aborted, possibly because of an S3 blip, and I would expect something in this stack to retry such an exception OR raise an actionable error.
Does anybody know what is causing this?
The text was updated successfully, but these errors were encountered: