-
Notifications
You must be signed in to change notification settings - Fork 568
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
TotalPartsExceeded: exceeded total allowed configured MaxUploadParts #4117
Comments
Thanks for the detailed bug filing, @Jamesthegiantpeach To confirm is it the input, output, or both that is in the ~100GB range? Regardless, it certainly looks like you're right that the AWS SDK requires upping the |
I am not sure why I am hitting this now. I have uploaded files larger than 100GB with this tool before. |
This has been fixed by #4107, it'll be shipped in 1.9.6. |
@jdoliner @gabrielgrant “44fbb193200614326d680be667b6e14ec1cfa","data":[{"path":"/552660222","hash":"w2qH0p32RoSRlmU19CFm8gLBPFTDhstY3wQ3oGzzf7k="}],"ts":"2021-02-02T23:41:51.651013533Z","message":"failed processing datum: rpc error: code = Unknown desc = MultipartUpload: upload multipart failed\n\tupload id: k7wdTEYd8W8KxopAiqotOCuwUTWeHBkSWi6T4LC3QZfRsa4Kl5Q828Kr1EuoGC9EPHB4IvMEbrtpNDZFHqymtpr9lzJMyabM29icEtFplG1yzNcV73mjZO54txm6wu3k4WmJmEimZCGPf03pSDBiog--\ncaused by: TotalPartsExceeded: exceeded total allowed S3 limit MaxUploadParts (10000). Adjust PartSize to fit in this limit, retrying in 0s"}” However, I already set "--max-upload-parts 20000" during installation, and I can see value in secret file correctly, also I accessed into the docker container and export environment variables and I see. We are on 1.12.1, AWS EKS. |
@xubofei1983 - The upper limit for max upload parts is 10,000. If you try to set it higher than that, AWS ignores it and limits you to 10,000. What you need to do is adjust the PartSize, which defaults to 5 MB. By setting the PartSize higher, you'll have fewer parts. You just need to figure out how high you need to set it in order to get below the 10,000 part threshold. |
ah thanks @dgeorg42 . I think this should be better described in doc. this parameter does not seem very useful then. two more related questions:
|
What happened?:
Workload: ~100 GB datums at low frequency (maybe a few at a time). Each datum takes about 4 hours to run.
What happens: The logs show that all datums are processed, but the job progress from
pachctl list job
shows 0/num_datums marked as finished. The job fails after a few hours withWhat you expected to happen?:
I expected the files to be uploaded and merged since the datums were processed successfully.
How to reproduce it (as minimally and precisely as possible)?:
spec:
AWS servers: c5d.18xlarge
The script is just a simple shell script running this tool and
this tool.
Anything else we need to know?:
Error seems to come from the aws go dk:
https://github.com/aws/aws-sdk-go/blob/v1.20.3/service/s3/s3manager/upload.go#L574
See it here and here too
Example stats from one job
Environment?:
kubectl version
):1.15.0
pachctl version
):1.9.5
aws ec2
kope.io kops ami
The text was updated successfully, but these errors were encountered: