From 6633366218588d0f24be343f8dfe9dd40ba97c71 Mon Sep 17 00:00:00 2001 From: Andriy Senkovych Date: Wed, 1 Oct 2014 13:42:39 +0300 Subject: [PATCH] Fix spelling errors --- README | 2 +- src/curl.cpp | 12 ++++++------ src/curl.h | 2 +- src/openssl_auth.cpp | 2 +- src/s3fs.cpp | 16 ++++++++-------- src/s3fs_util.cpp | 8 ++++---- 6 files changed, 21 insertions(+), 21 deletions(-) diff --git a/README b/README index fe8e641..77abd88 100644 --- a/README +++ b/README @@ -61,7 +61,7 @@ Known Issues: ------------- s3fs should be working fine with S3 storage. However, There are couple of limitations: -* Currently s3fs could hang the CPU if you have lots of time-outs. This is *NOT* a fault of s3fs but rather libcurl. This happends when you try to copy thousands of files in 1 session, it doesn't happend when you upload hundreds of files or less. +* Currently s3fs could hang the CPU if you have lots of time-outs. This is *NOT* a fault of s3fs but rather libcurl. This happens when you try to copy thousands of files in 1 session, it doesn't happen when you upload hundreds of files or less. * CentOS 4.x/RHEL 4.x users - if you use the kernel that shipped with your distribution and didn't upgrade to the latest kernel RedHat/CentOS gives, you might have a problem loading the "fuse" kernel. Please upgrade to the latest kernel (2.6.16 or above) and make sure "fuse" kernel module is compiled and loadable since FUSE requires this kernel module and s3fs requires it as well. * Moving/renaming/erasing files takes time since the whole file needs to be accessed first. A workaround could be to use s3fs's cache support with the use_cache option. diff --git a/src/curl.cpp b/src/curl.cpp index b6613a8..a698686 100644 --- a/src/curl.cpp +++ b/src/curl.cpp @@ -529,7 +529,7 @@ bool S3fsCurl::LocateBundle(void) if(0 == S3fsCurl::curl_ca_bundle.size()){ CURL_CA_BUNDLE = getenv("CURL_CA_BUNDLE"); if(CURL_CA_BUNDLE != NULL) { - // check for existance and readability of the file + // check for existence and readability of the file ifstream BF(CURL_CA_BUNDLE); if(!BF.good()){ DPRN("%s: file specified by CURL_CA_BUNDLE environment variable is not readable", program_name.c_str()); @@ -1905,7 +1905,7 @@ int S3fsCurl::GetIAMCredentials(void) // analizing response if(0 == result && !S3fsCurl::SetIAMCredentials(bodydata->str())){ - DPRN("Something error occured, could not get IAM credential."); + DPRN("Something error occurred, could not get IAM credential."); } delete bodydata; bodydata = NULL; @@ -3472,11 +3472,11 @@ bool AdditionalHeader::AddHeader(headers_t& meta, const char* path) const } int nPathLen = strlen(path); for(charcnt_list_t::const_iterator iter = charcntlist.begin(); iter != charcntlist.end(); ++iter){ - // get target charactor count + // get target character count if(nPathLen < (*iter)){ continue; } - // make target suffix(same charactor count) & find + // make target suffix(same character count) & find string suffix(&path[nPathLen - (*iter)]); if(addheader.end() == addheader.find(suffix)){ continue; @@ -3511,9 +3511,9 @@ bool AdditionalHeader::Dump(void) const if(!foreground2){ return true; } - // charactor count list + // character count list stringstream ssdbg; - ssdbg << "Charactor count list[" << charcntlist.size() << "] = {"; + ssdbg << "Character count list[" << charcntlist.size() << "] = {"; for(charcnt_list_t::const_iterator citer = charcntlist.begin(); citer != charcntlist.end(); ++citer){ ssdbg << " " << (*citer); } diff --git a/src/curl.h b/src/curl.h index 2896e63..f054460 100644 --- a/src/curl.h +++ b/src/curl.h @@ -351,7 +351,7 @@ class S3fsCurl // typedef std::map s3fscurlmap_t; typedef bool (*S3fsMultiSuccessCallback)(S3fsCurl* s3fscurl); // callback for succeed multi request -typedef S3fsCurl* (*S3fsMultiRetryCallback)(S3fsCurl* s3fscurl); // callback for failuer and retrying +typedef S3fsCurl* (*S3fsMultiRetryCallback)(S3fsCurl* s3fscurl); // callback for failure and retrying class S3fsMultiCurl { diff --git a/src/openssl_auth.cpp b/src/openssl_auth.cpp index 6108f7e..50f3117 100644 --- a/src/openssl_auth.cpp +++ b/src/openssl_auth.cpp @@ -131,7 +131,7 @@ static void s3fs_destroy_dyn_crypt_mutex(struct CRYPTO_dynlock_value* dyndata, c bool s3fs_init_crypt_mutex(void) { if(s3fs_crypt_mutex){ - FPRNNN("s3fs_crypt_mutex is not NULL, destory it."); + FPRNNN("s3fs_crypt_mutex is not NULL, destroy it."); if(!s3fs_destroy_crypt_mutex()){ DPRN("Failed to s3fs_crypt_mutex"); return false; diff --git a/src/s3fs.cpp b/src/s3fs.cpp index b4a1585..f624401 100644 --- a/src/s3fs.cpp +++ b/src/s3fs.cpp @@ -2040,7 +2040,7 @@ static int s3fs_release(const char* path, struct fuse_file_info* fi) // [NOTICE] // At first, we remove stats cache. - // Because fuse does not wait for responce from "release" function. :-( + // Because fuse does not wait for response from "release" function. :-( // And fuse runs next command before this function returns. // Thus we call deleting stats function ASSAP. // @@ -2562,7 +2562,7 @@ static bool is_truncated(xmlDocPtr doc) // return: the pointer to object name on allocated memory. // the pointer to "c_strErrorObjectName".(not allocated) -// NULL(a case of something error occured) +// NULL(a case of something error occurred) static char* get_object_name(xmlDocPtr doc, xmlNodePtr node, const char* path) { // Get full path @@ -2688,7 +2688,7 @@ static void s3fs_destroy(void*) { DPRN("destroy"); - // Destory curl + // Destroy curl if(!S3fsCurl::DestroyS3fsCurl()){ DPRN("Could not release curl library."); } @@ -2929,14 +2929,14 @@ static int s3fs_utility_mode(void) xmlDocPtr doc; if(NULL == (doc = xmlReadMemory(body.c_str(), static_cast(body.size()), "", NULL, 0))){ - DPRN("xmlReadMemory returns with error."); + DPRN("xmlReadMemory exited with error."); result = EXIT_FAILURE; }else{ // make working uploads list uncomp_mp_list_t list; if(!get_uncomp_mp_list(doc, list)){ - DPRN("get_uncomp_mp_list returns with error."); + DPRN("get_uncomp_mp_list exited with error."); result = EXIT_FAILURE; }else{ @@ -2944,7 +2944,7 @@ static int s3fs_utility_mode(void) print_uncomp_mp_list(list); // remove if(!abort_uncomp_mp_list(list)){ - DPRN("something error occured in removing process."); + DPRN("an error occurred during removal process."); result = EXIT_FAILURE; } } @@ -2952,7 +2952,7 @@ static int s3fs_utility_mode(void) } } - // Destory curl + // Destroy curl if(!S3fsCurl::DestroyS3fsCurl()){ DPRN("Could not release curl library."); } @@ -3900,7 +3900,7 @@ int main(int argc, char* argv[]) } } - // error checking of command line arguments for compatability + // error checking of command line arguments for compatibility if(S3fsCurl::IsPublicBucket() && S3fsCurl::IsSetAccessKeyId()){ fprintf(stderr, "%s: specifying both public_bucket and the access keys options is invalid\n", program_name.c_str()); diff --git a/src/s3fs_util.cpp b/src/s3fs_util.cpp index 0729e2a..6d8a19a 100644 --- a/src/s3fs_util.cpp +++ b/src/s3fs_util.cpp @@ -496,9 +496,9 @@ string get_username(uid_t uid) DPRNCRIT("failed to allocate memory."); return string(""); } - // get group infomation + // get group information if(0 != (result = getpwuid_r(uid, &pwinfo, pbuf, maxlen, &ppwinfo))){ - DPRNNN("could not get pw infomation."); + DPRNNN("could not get pw information."); free(pbuf); return string(""); } @@ -532,9 +532,9 @@ int is_uid_inculde_group(uid_t uid, gid_t gid) DPRNCRIT("failed to allocate memory."); return -ENOMEM; } - // get group infomation + // get group information if(0 != (result = getgrgid_r(gid, &ginfo, pbuf, maxlen, &pginfo))){ - DPRNNN("could not get group infomation."); + DPRNNN("could not get group information."); free(pbuf); return -result; }