mirror of
https://github.com/s3fs-fuse/s3fs-fuse.git
synced 2024-11-19 02:35:13 +00:00
Merge pull request #213 from andrewgaul/rename-large-files
Parse ETag from copy multipart correctly
This commit is contained in:
commit
deb0e9eec3
33
src/curl.cpp
33
src/curl.cpp
@ -3127,12 +3127,35 @@ int S3fsCurl::CopyMultipartPostRequest(const char* from, const char* to, int par
|
|||||||
|
|
||||||
int result = RequestPerform();
|
int result = RequestPerform();
|
||||||
if(0 == result){
|
if(0 == result){
|
||||||
const char* start_etag= strstr(bodydata->str(), "ETag");
|
// parse ETag from response
|
||||||
const char* end_etag = strstr(bodydata->str(), "/ETag>");
|
xmlDocPtr doc;
|
||||||
|
if(NULL == (doc = xmlReadMemory(bodydata->str(), bodydata->size(), "", NULL, 0))){
|
||||||
partdata.etag.assign((start_etag + 11), (size_t)(end_etag - (start_etag + 11) - 7));
|
return result;
|
||||||
partdata.uploaded = true;
|
}
|
||||||
|
if(NULL == doc->children){
|
||||||
|
S3FS_XMLFREEDOC(doc);
|
||||||
|
return result;
|
||||||
|
}
|
||||||
|
for(xmlNodePtr cur_node = doc->children->children; NULL != cur_node; cur_node = cur_node->next){
|
||||||
|
if(XML_ELEMENT_NODE == cur_node->type){
|
||||||
|
string elementName = reinterpret_cast<const char*>(cur_node->name);
|
||||||
|
if(cur_node->children){
|
||||||
|
if(XML_TEXT_NODE == cur_node->children->type){
|
||||||
|
if(elementName == "ETag") {
|
||||||
|
string etag = reinterpret_cast<const char *>(cur_node->children->content);
|
||||||
|
if(etag.size() >= 2 && *etag.begin() == '"' && *etag.rbegin() == '"'){
|
||||||
|
etag.assign(etag.substr(1, etag.size() - 2));
|
||||||
|
}
|
||||||
|
partdata.etag.assign(etag);
|
||||||
|
partdata.uploaded = true;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
S3FS_XMLFREEDOC(doc);
|
||||||
}
|
}
|
||||||
|
|
||||||
delete bodydata;
|
delete bodydata;
|
||||||
bodydata = NULL;
|
bodydata = NULL;
|
||||||
delete headdata;
|
delete headdata;
|
||||||
|
@ -120,6 +120,7 @@ static bool is_s3fs_gid = false;// default does not set.
|
|||||||
static bool is_s3fs_umask = false;// default does not set.
|
static bool is_s3fs_umask = false;// default does not set.
|
||||||
static bool is_remove_cache = false;
|
static bool is_remove_cache = false;
|
||||||
static bool create_bucket = false;
|
static bool create_bucket = false;
|
||||||
|
static int64_t singlepart_copy_limit = FIVE_GB;
|
||||||
|
|
||||||
//-------------------------------------------------------------------
|
//-------------------------------------------------------------------
|
||||||
// Static functions : prototype
|
// Static functions : prototype
|
||||||
@ -1388,7 +1389,7 @@ static int s3fs_rename(const char* from, const char* to)
|
|||||||
// files larger than 5GB must be modified via the multipart interface
|
// files larger than 5GB must be modified via the multipart interface
|
||||||
if(S_ISDIR(buf.st_mode)){
|
if(S_ISDIR(buf.st_mode)){
|
||||||
result = rename_directory(from, to);
|
result = rename_directory(from, to);
|
||||||
}else if(!nomultipart && buf.st_size >= FIVE_GB){
|
}else if(!nomultipart && buf.st_size >= singlepart_copy_limit){
|
||||||
result = rename_large_object(from, to);
|
result = rename_large_object(from, to);
|
||||||
}else{
|
}else{
|
||||||
if(!nocopyapi && !norenameapi){
|
if(!nocopyapi && !norenameapi){
|
||||||
@ -4392,6 +4393,10 @@ static int my_fuse_opt_proc(void* data, const char* arg, int key, struct fuse_ar
|
|||||||
}
|
}
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
if(0 == STR2NCMP(arg, "singlepart_copy_limit=")){
|
||||||
|
singlepart_copy_limit = static_cast<int64_t>(s3fs_strtoofft(strchr(arg, '=') + sizeof(char))) * 1024;
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
if(0 == STR2NCMP(arg, "ahbe_conf=")){
|
if(0 == STR2NCMP(arg, "ahbe_conf=")){
|
||||||
string ahbe_conf = strchr(arg, '=') + sizeof(char);
|
string ahbe_conf = strchr(arg, '=') + sizeof(char);
|
||||||
if(!AdditionalHeader::get()->Load(ahbe_conf.c_str())){
|
if(!AdditionalHeader::get()->Load(ahbe_conf.c_str())){
|
||||||
|
@ -971,6 +971,10 @@ void show_help (void)
|
|||||||
" multipart_size (default=\"10\")\n"
|
" multipart_size (default=\"10\")\n"
|
||||||
" - part size, in MB, for each multipart request.\n"
|
" - part size, in MB, for each multipart request.\n"
|
||||||
"\n"
|
"\n"
|
||||||
|
" singlepart_copy_limit (default=\"5120\")\n"
|
||||||
|
" - maximum size, in MB, of a single-part copy before trying \n"
|
||||||
|
" multipart copy.\n"
|
||||||
|
"\n"
|
||||||
" fd_page_size (default=\"52428800\"(50MB))\n"
|
" fd_page_size (default=\"52428800\"(50MB))\n"
|
||||||
" - number of internal management page size for each file descriptor.\n"
|
" - number of internal management page size for each file descriptor.\n"
|
||||||
" For delayed reading and writing by s3fs, s3fs manages pages which \n"
|
" For delayed reading and writing by s3fs, s3fs manages pages which \n"
|
||||||
|
@ -301,7 +301,31 @@ then
|
|||||||
fi
|
fi
|
||||||
|
|
||||||
rm -f "/tmp/${BIG_FILE}"
|
rm -f "/tmp/${BIG_FILE}"
|
||||||
rm -f "${BIG_FILE}"
|
rm_test_file "${BIG_FILE}"
|
||||||
|
|
||||||
|
##########################################################
|
||||||
|
# Testing multi-part copy
|
||||||
|
##########################################################
|
||||||
|
# TODO: test disabled until S3Proxy 1.5.0 is released
|
||||||
|
if false
|
||||||
|
then
|
||||||
|
|
||||||
|
echo "Testing multi-part copy ..."
|
||||||
|
dd if=/dev/urandom of="/tmp/${BIG_FILE}" bs=$BIG_FILE_LENGTH count=1
|
||||||
|
dd if="/tmp/${BIG_FILE}" of="${BIG_FILE}" bs=$BIG_FILE_LENGTH count=1
|
||||||
|
mv "${BIG_FILE}" "${BIG_FILE}-copy"
|
||||||
|
|
||||||
|
# Verify contents of file
|
||||||
|
echo "Comparing test file"
|
||||||
|
if ! cmp "/tmp/${BIG_FILE}" "${BIG_FILE}-copy"
|
||||||
|
then
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
|
||||||
|
rm -f "/tmp/${BIG_FILE}"
|
||||||
|
rm_test_file "${BIG_FILE}-copy"
|
||||||
|
|
||||||
|
fi
|
||||||
|
|
||||||
##########################################################
|
##########################################################
|
||||||
# Testing special characters
|
# Testing special characters
|
||||||
|
@ -59,6 +59,7 @@ stdbuf -oL -eL $S3FS $TEST_BUCKET_1 $TEST_BUCKET_MOUNT_POINT_1 \
|
|||||||
-o createbucket \
|
-o createbucket \
|
||||||
-o passwd_file=$S3FS_CREDENTIALS_FILE \
|
-o passwd_file=$S3FS_CREDENTIALS_FILE \
|
||||||
-o sigv2 \
|
-o sigv2 \
|
||||||
|
-o singlepart_copy_limit=$((10 * 1024)) \
|
||||||
-o url=http://127.0.0.1:8080 \
|
-o url=http://127.0.0.1:8080 \
|
||||||
-o use_path_request_style -f -o f2 -d -d |& stdbuf -oL -eL sed -u "s/^/s3fs: /" &
|
-o use_path_request_style -f -o f2 -d -d |& stdbuf -oL -eL sed -u "s/^/s3fs: /" &
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user