gsutil: gsutil rsync hangs on large files
Hi! We are using gsutil rsync
to upload our backups to coldline storage. After our files become larger (more than 30GB) rsync
starts hanging on each such file. The same situation is for gsutil cp
.
I’ve already sent a log file with gsutil -D rsync
to gs-team@google.com
gsutil version: 4.25
boto version: 2.42.0
python version: 2.7.12 (default, Nov 19 2016, 06:48:10) [GCC 5.4.0 20160609]
OS: Linux 4.4.0-77-generic
multiprocessing available: True
using cloud sdk: True
config path(s): /etc/boto.cfg
gsutil path: /usr/lib/google-cloud-sdk/platform/gsutil/gsutil
compiled crcmod: True
installed via package manager: False
editable install: False
About this issue
- Original URL
- State: open
- Created 7 years ago
- Reactions: 5
- Comments: 15
Hello, I’m trying to rsync from gcs to s3 buckets but some files are bigger than 5 GiB.
I receive this error: “exceeds the maximum gsutil-supported size for an S3 upload. S3 objects greater than 5 GiB in size require multipart uploads, which gsutil does not support.”
As you can see here https://docs.aws.amazon.com/AmazonS3/latest/dev/qfacts.html there is a limitation in S3 where you have to use multipart upload for files bigger than 5 GiB.
Please help me, how can I fix that?
Any help will be really appreciated! thanks @houglum for your time
Best Regards Fabio Rigato