From 67b9381825def0d850301ce95c626ad3f35adac1 Mon Sep 17 00:00:00 2001 From: Andrew Gaul Date: Wed, 3 Feb 2021 22:27:50 +0900 Subject: [PATCH] Remove limit for single-part objects AWS enforces a 5 GB limit for single-part objects but other S3 implementations like GCS support larger objects. Fixes #1542. --- src/s3fs.cpp | 7 +------ 1 file changed, 1 insertion(+), 6 deletions(-) diff --git a/src/s3fs.cpp b/src/s3fs.cpp index 812a576..e011024 100644 --- a/src/s3fs.cpp +++ b/src/s3fs.cpp @@ -726,12 +726,7 @@ int put_headers(const char* path, headers_t& meta, bool is_copy, bool update_mti // get_object_attribute() returns error with initializing buf. (void)get_object_attribute(path, &buf); - if(buf.st_size >= FIVE_GB){ - // multipart - if(nocopyapi || nomultipart){ - S3FS_PRN_WARN("Metadata update failed because the file is larger than 5GB and the options nocopyapi or nomultipart are set: [path=%s]", path); - return -EFBIG; // File too large - } + if(buf.st_size >= FIVE_GB && !nocopyapi && !nomultipart){ if(0 != (result = s3fscurl.MultipartHeadRequest(path, buf.st_size, meta, is_copy))){ return result; }