-
Notifications
You must be signed in to change notification settings - Fork 2.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Hang when downloading a large blob #25358
Comments
Thank you for your feedback. This has been routed to the support team for assistance. |
Hi @ktaebum, thanks for reaching out and sorry for the delay. A couple of follow-up questions/points.
import sys
import logging
from azure.storage.blob import BlobClient
# Set the logging level for the azure.storage.blob library
logger = logging.getLogger('azure.storage.blob')
logger.setLevel(logging.DEBUG)
# Direct logging output to stdout. Without adding a handler,
# no logging output is visible.
handler = logging.StreamHandler(stream=sys.stdout)
logger.addHandler(handler)
blob_client = BlobClient(..., logging_enable=True) Ultimately, this is likely a server-side issue but let's try and gather some more info before involving the service team. Thanks! |
@jalauzon-msft Thanks for the reply. I waited not much time (just a couple of minutes) and no error returned. However, I've checked that downloading does not fail if I set |
Unfortunately, |
Hi @ktaebum, apologies for the long delay. Read timeouts will be automatically retried by the SDK and it seems, from the screenshot you shared, this did help for that particular download as you see it time out and then continue. Changing the retry count will not eliminate read timeouts but will change the number of times a read timeout can be retried. Are you still seeing downloads not complete because of read timeouts? If they are completing after a read timeout, then the retry mechanism is working as expected and you should be good. I would recommend setting your If you've done all this and are still having trouble downloading blobs and experiencing so many read timeouts where the blob will not complete downloading, I would recommend opening a support ticket for your Storage account to have the service team investigate further. Thanks! |
Hi @ktaebum. Thank you for opening this issue and giving us the opportunity to assist. We believe that this has been addressed. If you feel that further discussion is needed, please add a comment with the text “ |
Hi @ktaebum, since you haven’t asked that we “ |
Describe the bug
I think this is related to #10572,
I am trying to download a blob whose size is 24GB.
I use
download_blob
ofBlobClient
and I setmax_concurrency
as 32 on my azure VM (VM size isStandard_D4ds_v5
).Expected behavior
I expect the downloading to be completed successfully.
Screenshots
Downloading is hanged as the following screenshot.
Ignore
MB
, it isB
(bytes)Additional context
This is a heisenbug. Sometimes downloading is finished successfully.
I've seen that the previous issue is fixed as https://github.com/Azure/azure-sdk-for-python/pull/18164/files.
However, I think it would be better if a user can configure
max_retry
which is fixed as 3 currently.The text was updated successfully, but these errors were encountered: