Commit Graph

740 Commits

Author SHA1 Message Date
root
d5626fe595 Changed url encoding of space character to use %20 instead of the plus (+) sign. 2013-12-17 15:12:03 +01:00
ggtakec@gmail.com
d121bab3e5 Summary of Changes(1.73 -> 1.74)
==========================
List of Changes
==========================
1) Fixed bugs - r478, r480, r481, r482, r483, r484, r486, r489,
                r491, r492, r493, r494, r495, r496, r497, r498
   - Fixed wrong printf format.
   - Fixed a bug at changing umask.
   - Fixed a bug at changing utimens.
   - When changing object size, could not update stat cache.
   - Fixed some memory leak.
   - Fixed a bug about ftruncate(Issue 371).
   - Fixed a bug about multipart post request(Issue 371).
   - Fixed a bug about multipart post/head retrying(Issue 389).
   - Fixed public_bucket option bug
   - Fixed file size over flow on 32bit OS

2) Changed codes - r479, r480, r482, r485, r488, r490, r493
   - Calling curl_global_init function once.
   - Adds dynamic lock callback for OpenSSL.
   - Checking strict for s3fs command parameters.
   - Caches SSL session.
   - Changes parallel count 500 to 20 for multipart head req.
   - Adds --enable-nss-init configure option.
   - Changes updating timing for object cache(Issue 368).
   - Changes level for some debugging codes.
   - Adds casting for curl_ott_t variables.
   - Adds IAM role option.
   - Adds function on utility mode for removing failed multipart req.



git-svn-id: http://s3fs.googlecode.com/svn/trunk@499 df820570-a93a-0410-bd06-b72b767a4274
2013-11-25 12:04:13 -05:00
rrizun
9011a0bb73 .gitignore 2013-11-22 15:28:55 -05:00
ggtakec@gmail.com
74db6748dd Changes codes
1) Changed condition for retrying multipart error
   Changes condition that 404 is not retrying, other case is retrying,
   when multipart request failed.

2) file type wrong
   fdcache.h file type is wrong, so that fixed it.



git-svn-id: http://s3fs.googlecode.com/svn/trunk@498 df820570-a93a-0410-bd06-b72b767a4274
2013-11-19 01:48:53 +00:00
ggtakec@gmail.com
8acbaf7199 Fixed a bug.
* Fixed a bug
  Fixes a bug that is dead loop when s3fs listed in a directory with
  directory objects which was not object(no information).
  This bug made by r493, and reported by issue 389.




git-svn-id: http://s3fs.googlecode.com/svn/trunk@497 df820570-a93a-0410-bd06-b72b767a4274
2013-11-18 02:29:41 +00:00
ggtakec@gmail.com
40b9f0a408 Changes codes
1) Changed buffer size for file size
   Changes a internal buffer size from size_t to offt_t.
   It is a bug for 32 bit OS enviroment.



git-svn-id: http://s3fs.googlecode.com/svn/trunk@496 df820570-a93a-0410-bd06-b72b767a4274
2013-11-17 08:50:41 +00:00
ggtakec@gmail.com
882f13020e Fixed bugs(overflow)
1) Overflow
   About over 4GB file, when st_size which is member stat structure,
   the value is overflow.
   Fixed this bug and fixed like this bug in all sources. 

2) Changed retrying request
   If s3fs gets 500 HTTP status for multipart request, s3fs retry
   to send same request.





git-svn-id: http://s3fs.googlecode.com/svn/trunk@495 df820570-a93a-0410-bd06-b72b767a4274
2013-11-13 16:26:50 +00:00
ggtakec@gmail.com
c785be917f Changed a code
* Cut a #ifdef
  Cuts a code for clearing r493.




git-svn-id: http://s3fs.googlecode.com/svn/trunk@494 df820570-a93a-0410-bd06-b72b767a4274
2013-11-11 15:03:04 +00:00
ggtakec@gmail.com
09fc2593e3 Fixed bugs and Changed utility mode
1) Fixed bugs
 * Rename objects
   Fixes s3fs specifies wrong part number of multipart rename.
   And s3fs adds x-amz-acl and x-amz-server-side-encryption
   header when rename objects.

2) Changed retry logic for multipart uploading(and renaming)
   Sometimes, s3fs gets 400 HTTP response for one of part from
   S3 when s3fs uploads a large object by multipart.
   New logic retries uploading failed part until "retries"
   option count.

3) Added action on utility mode.
   s3fs have had utility mode for displaying the result of REST
   listing multipart uploading.
   Changed this row result(xml) to list, after that, s3fs starts
   conversation for removing it.
   Then you can remove the object which is failed uploading by
   multipart, and do not need to pay for that ever.




git-svn-id: http://s3fs.googlecode.com/svn/trunk@493 df820570-a93a-0410-bd06-b72b767a4274
2013-11-11 13:45:35 +00:00
ggtakec@gmail.com
1bae39e21f Fixed a bug.
* Fixed a bug
   Fixes a bug that the retrying multipart post request is not 
   complete, then it is fixed.
   This is reported by Issue 371#32.



git-svn-id: http://s3fs.googlecode.com/svn/trunk@492 df820570-a93a-0410-bd06-b72b767a4274
2013-10-09 01:44:56 +00:00
ggtakec@gmail.com
33431dec46 Fixed a bug.
* Fixed a bug
   Fixes a bug that is mis-coding about retrying Multipart Post.
   This is reported by Issue 371#28.




git-svn-id: http://s3fs.googlecode.com/svn/trunk@491 df820570-a93a-0410-bd06-b72b767a4274
2013-10-08 08:19:10 +00:00
ggtakec@gmail.com
99db6d13af Changes codes
1) Supported IAM role
   Supports IAM role by option, that is instead of AccessKeyID/
   SecretAccessKey.
   Adds new option "iam_role" which is specified as IAM role
   name.(like s3fs-c)



git-svn-id: http://s3fs.googlecode.com/svn/trunk@490 df820570-a93a-0410-bd06-b72b767a4274
2013-10-06 13:45:32 +00:00
ggtakec@gmail.com
44468ba00f Fixed a bug(public_bucket)
1) Fixed a bug(about public_bucket)
   Fixes a bug that public_bucket option does not work.




git-svn-id: http://s3fs.googlecode.com/svn/trunk@489 df820570-a93a-0410-bd06-b72b767a4274
2013-09-29 13:24:30 +00:00
ggtakec@gmail.com
e6038f74ed Fixed a bug
1) Fixed a bug(about curl_off_t)
   Fixes a bug which is not use cast to curl_off_t from off_t(ssize_t) 
   at calling curl_easy_setopt with CURLOPT_POSTFIELDSIZE and 
   CURLOPT_INFILESIZE_LARGE.
   Maybe this forgot cast occurred issue 471(failed multipart uploading).



git-svn-id: http://s3fs.googlecode.com/svn/trunk@488 df820570-a93a-0410-bd06-b72b767a4274
2013-09-27 07:39:07 +00:00
ggtakec@gmail.com
a6884f1c3a Changes codes
1) Changed debug message level
   Changes a level and format for a debugging message about 
   parallel multipart upload in curl.cpp



git-svn-id: http://s3fs.googlecode.com/svn/trunk@487 df820570-a93a-0410-bd06-b72b767a4274
2013-09-26 05:00:21 +00:00
ggtakec@gmail.com
3ba40a30cd Fixed a bug(Issue 371)
1) Fixed Issue 371
   Fixes a bug which is wrong return value in s3fs_truncate function(mis-coding).
 
   Issue 371: ftruncate failed



git-svn-id: http://s3fs.googlecode.com/svn/trunk@486 df820570-a93a-0410-bd06-b72b767a4274
2013-09-19 09:22:02 +00:00
ggtakec@gmail.com
41bf0e3558 * Re-Fixed a bug( Issue 368 )
1) Re-Fixed Issue 368
   Changes that s3fs checks always object stat information before opening it.
   Then the object updates other s3fs process or other client, s3fs can know
   it.

   (Issue 368)1.73: Updating existing file on server 'a' does not change length
   of file on server 'b'



git-svn-id: http://s3fs.googlecode.com/svn/trunk@485 df820570-a93a-0410-bd06-b72b767a4274
2013-09-17 05:16:30 +00:00
ggtakec@gmail.com
79152325ca Changes codes
1) fixed a bug
   fixes a code in s3fs.cpp.
   It freed memory twice because of careless.



git-svn-id: http://s3fs.googlecode.com/svn/trunk@484 df820570-a93a-0410-bd06-b72b767a4274
2013-09-17 01:26:26 +00:00
ggtakec@gmail.com
b231081aff Changes codes
1) fixed a bug
   fixes a code in curl.cpp.



git-svn-id: http://s3fs.googlecode.com/svn/trunk@483 df820570-a93a-0410-bd06-b72b767a4274
2013-09-14 21:53:30 +00:00
ggtakec@gmail.com
42b74c9d2e Changes codes
1) Changed codes about memory leak
   For memory leak, below codes are changed.
   * calls malloc_trim function
   * calls initializing NSS function, and adds configure
     option "--enable-nss-init".
     If libcurl is with NSS, s3fs initializes NSS manually.
     This initializing NSS is enabled by "--enable-nss-init"
     option at configure. if this option is specified, you
     need "nss-devel" package.
   * calls initializing libxml2(xmlInitParser).
   * BIO functions have memory leak, calls CRYPTO_free_ex_data.
   * changes cache structure.
   * changes cache out logic to LRU.
   * sets alignment for allcated memory in body data structure.
   * adds ssl session into share handle. and adds nosscache option.
   * deletes unused allocated memory.(bug)
   * changes defaule parallel count of head request in readdir
     (500->20)
   * fixes some bugs.



git-svn-id: http://s3fs.googlecode.com/svn/trunk@482 df820570-a93a-0410-bd06-b72b767a4274
2013-09-14 21:50:39 +00:00
ggtakec@gmail.com
d45f4707ea Fixed bugs( Issue 368 )
1) Fixed Issue 368
   Fixed a bug that s3fs could not update local cache.

   (Issue 368)1.73: Updating existing file on server 'a' does not change length of file on server 'b'




git-svn-id: http://s3fs.googlecode.com/svn/trunk@481 df820570-a93a-0410-bd06-b72b767a4274
2013-08-30 02:25:27 +00:00
ggtakec@gmail.com
3b393fe159 Fixed bugs
1) Fixed Issue 321
   Fixed a bug that a value of umask option is not reflected
   definitely.
   (Issue 321) no write permission for non-root user

2) Fixed a bug about utimens
   Fixed a bug that utimens function cloud not set value to the
   other user's object which does not allowed writable.

3) Strictly option check.
   Fixed checking s3fs options strictly.




git-svn-id: http://s3fs.googlecode.com/svn/trunk@480 df820570-a93a-0410-bd06-b72b767a4274
2013-08-29 06:23:51 +00:00
ggtakec@gmail.com
7ce8135fa9 Changes codes
1) Changed codes about initializing curl and openSSL
   s3fs before this change called curl_global_init() two times
   with curl_global_cleanup(). After reviewing this processing,
   s3fs calls curl_global_init() one time.
   Then s3fs_check_service function which checks user bucket
   existing is called after calling fuse.
   So this new processing does not have a problem, the codes
   updated.

   And about initializing openSSL(CRYPTO), old s3fs called only
   static locking callback function(ex. CRYPTO_set_locking_callback()).
   Added calling dynamic locking callback function for CRYPTO
   (ex. CRYPTO_set_dynlock_lock_callback()).



git-svn-id: http://s3fs.googlecode.com/svn/trunk@479 df820570-a93a-0410-bd06-b72b767a4274
2013-08-27 08:12:01 +00:00
ggtakec@gmail.com
8111edec61 Fixed codes for compiling(32)
1) Fixed codes
   Fixed compiling error on 32bit, which specified wrong dev_t format
   on 32bit.




git-svn-id: http://s3fs.googlecode.com/svn/trunk@478 df820570-a93a-0410-bd06-b72b767a4274
2013-08-26 07:18:05 +00:00
ggtakec@gmail.com
c3c6c0f572 Summary of Changes(1.72 -> 1.73)
==========================
List of Changes
==========================
1) Added ahbe_conf option - r465
  - Added ahbe_conf option means the configuration file path, and
    this file specifies additional HTTP header by file(object)
    extension.( Issue 292 )
  - Added sample configuration file in test directory for ahbe_conf
    option.

2) Changed mount point permission - r465
  - Not allow group/other permission for mount point when s3fs runs
    without allow_other.
  - Allow permission to all user for mount point when specified
    allow_other option.

3) Fixed bugs - r465, r466, r467, r468, r470, r471
  - Changed a code which s3fs returns error as soon as possible
    when user tries to change mount point.( Issue 229 )
  - Fixed mis-format for debugging print.
  - Changed request type to "virtual hosted-style" for checking
    bucket when s3fs is starting.( Issue 362 )
  - Fixed bug issue(bug), when s3fs retry a request at something
    error occurred, s3fs is wrong usage curl handle and fails
    retrying request.( Issue 343 )
  - Fixed mis-coding about fonction prototype.( Issue 360/Issue 361 )
  - Fixed a bug about umask.( Issue 321 )
  - Fixed a bug which s3fs exit though specified correct 
    $HOME/.passwd-s3fs.( Issue 365 )
  - Fixed a bug which deleting stat cache information is bad
    posission in s3fs_release function.( Issue 363 )

4) Added sample script - 472, r473
  - Added sample_delcache.sh in test directory for deleting
    cahce files.

5) Added debugging messages - r467, r474
  - Changed debugging level for prepare_url function.
  - Specified f2 option, s3s_getattr puts detail for file attar.
  - Added new option as curldbg for curl http(s) debugging.



git-svn-id: http://s3fs.googlecode.com/svn/trunk@475 df820570-a93a-0410-bd06-b72b767a4274
2013-08-23 17:24:47 +00:00
ggtakec@gmail.com
3dda0b20d4 Added debugging message
1) Added debugging message in s3fs_getattr
   If s3fs runs with "f2" option for deep debugging message, s3fs_getattr
   puts debugging message as file's uid/gid/mode.

2) Added curldbg option
   Added new option "curldbg" which is for debugging curl http/https
   information.
   It implements by CURLOPT_VERBOSE on curl_easy_setopt function.



git-svn-id: http://s3fs.googlecode.com/svn/trunk@474 df820570-a93a-0410-bd06-b72b767a4274
2013-08-23 16:28:50 +00:00
ggtakec@gmail.com
1a4e525465 Changed test/Makefile.am
1) Changed test/Makefile.am
   Changed test/Makefile.am because test/sample_delcache.sh is added
   by r472.




git-svn-id: http://s3fs.googlecode.com/svn/trunk@473 df820570-a93a-0410-bd06-b72b767a4274
2013-08-23 15:26:48 +00:00
ggtakec@gmail.com
3ed9a2c1e4 Added sample script
1) Added sample script for deleting cache
   Added sample script file which removes cache file and stats file
   by limiting disk space.
   This script is based DPeuscher posted codes, special thanks.

   Issue 364: Feature request: Control cache size



git-svn-id: http://s3fs.googlecode.com/svn/trunk@472 df820570-a93a-0410-bd06-b72b767a4274
2013-08-23 15:22:24 +00:00
ggtakec@gmail.com
b3682f87d2 Fixed bugs(Issue 363)
1) Fixed Issue 363
   Fixed a bug which has below reason.
   Fuse does not wait finishing "release file descriptor" function
   called by fuse, and fuse runs(calls) next processing(commands).
   Then s3fs could not clear stats cache information for that file
   before calling next processing, and s3fs uses old stats cache
   information.
   So that, s3fs clears stats cache in release function at first.

   And found two bad codes(but these codes do not influence normal
   movement) in fdcache.cpp and fixed these.

   Issue 363: make check failing inconsistently



git-svn-id: http://s3fs.googlecode.com/svn/trunk@471 df820570-a93a-0410-bd06-b72b767a4274
2013-08-22 09:36:16 +00:00
ggtakec@gmail.com
07a8caa250 Fixed bugs
1) Fixed Issue 321(#30)
   Fixed a bug(mis-coding).
     Issue 321(#30): no write permission for non-root user

2) Fixed Issue 365
   Fixed a bug(mis-coding).
     Issue 365: there is a logical error in s3fs-1.72 s3fs.cpp:2865




git-svn-id: http://s3fs.googlecode.com/svn/trunk@470 df820570-a93a-0410-bd06-b72b767a4274
2013-08-22 02:39:21 +00:00
ggtakec@gmail.com
2b3fb2d102 Fixed a bug(prototype and initializing enum)
1) Fixed a bug
   Fixed bugs(mis-coding) which is wrong prototype for md5hexsum, md5sum functions.
     Issue 361: complie time error after running #make
     Issue 360: 1.72 Will not compile on Ubuntu 12.04.2 (precise) i686

   And fixed a code for initializing enum member in S3fsCurl class. 



git-svn-id: http://s3fs.googlecode.com/svn/trunk@469 df820570-a93a-0410-bd06-b72b767a4274
2013-08-21 08:39:06 +00:00
ggtakec@gmail.com
171de649ef Fixed a bug(about retry request)
1) Fixed a bug
   s3fs was wrong for request retry processing so far.
   It was fixed.
   Probably, Issue 343(1.7 having curl 35 + other disconnect issue) is 
   occurred by this bug.



git-svn-id: http://s3fs.googlecode.com/svn/trunk@468 df820570-a93a-0410-bd06-b72b767a4274
2013-08-21 07:43:32 +00:00
ggtakec@gmail.com
7fa1e37a28 Changes codes
1) "virtual hosted-style request" for checking bucket
   Old version calls "path-style request" for checking bucket at 
   initializing, then after this revision s3fs requests "virtual 
   hosted-style request".
   This change is related to 
   "Operation not permitted - on any operation(Issue 362)".

2) Changed debugging message level
   Changed debugging message level in prepare_url() from DPRNNN
   to FPRNINFO.



git-svn-id: http://s3fs.googlecode.com/svn/trunk@467 df820570-a93a-0410-bd06-b72b767a4274
2013-08-20 07:16:12 +00:00
ggtakec@gmail.com
ee01c91e02 Fixed bugs for compiling
1) Fixed bugs
   Fixes below bugs( format error and undefined fund ).

   * 1.72 Will not compile on Ubuntu 12.04.2 (precise) i686(Issue 360)
   * complie time error after running #make(Issue 361)

   I'll close these Issue if I can confirm that these problem was solved.



git-svn-id: http://s3fs.googlecode.com/svn/trunk@466 df820570-a93a-0410-bd06-b72b767a4274
2013-08-19 06:29:24 +00:00
ggtakec@gmail.com
d7689151ab Fixed Issue 229 and Changes codes
1) Set metadata "Content-Encoding" automatically(Issue 292)
   For this issue, s3fs is added new option "ahbe_conf".

   New option means the configuration file path, and this file specifies
   additional HTTP header by file(object) extension.
   Thus you can specify any HTTP header for each object by extension.

   * ahbe_conf file format:
     -----------
     line                = [file suffix] HTTP-header [HTTP-header-values]
     file suffix         = file(object) suffix, if this field is empty, 
                           it means "*"(all object).
     HTTP-header         = additional HTTP header name
     HTTP-header-values  = additional HTTP header value
     -----------

   * Example:
     -----------
     .gz      Content-Encoding     gzip
     .Z       Content-Encoding     compress
              X-S3FS-MYHTTPHEAD    myvalue
     -----------
     A sample configuration file is uploaded in "test" directory.

   If ahbe_conf parameter is specified, s3fs loads it's configuration
   and compares extension(suffix) of object(file) when uploading
   (PUT/POST) it. If the extension is same, s3fs adds/sends specified
   HTTP header and value.

   A case of sample configuration file, if a object(it's extension is
   ".gz") which already has Content-Encoding HTTP header is renamed 
   to ".txt" extension, s3fs does not set Content-Encoding. Because
   ".txt" is not match any line in configuration file.
   So, s3fs matches the extension by each PUT/POST action.

   * Please take care about "Content-Encoding".
   This new option allows setting ANY HTTP header by object extension.
   For example, you can specify "Content-Encoding" for ".gz"/etc 
   extension in configuration. But this means that S3 always returns 
   "Content-Encoding: gzip" when a client requests with other 
   "Accept-Encoding:" header. It SHOULD NOT be good.
   Please see RFC 2616.

2) Changes about allow_other/uid/gid option for mount point
   I reviewed about mount point permission and allow_other/uid/gid
   options, and found bugs about these.
   s3fs is fixed bugs and changed to the following specifications.

   * s3fs only allows uid(gid) options as 0(root), when the effective 
     user is zero(root).
   * A mount point(directory) must have a permission to allow
     accessing by effective user/group.
   * If allow_other option is specified, the mount point permission
     is set 0777(all users allow all access).
     In another case, the mount point is set 0700(only allows 
     effective user).
   * When uid/gid option is specified, the mount point owner/group
     is set uid/gid option value.
     If uid/gid is not set, it is set effective user/group id.

   This changes maybe fixes some issue(321, 338).

3) Changes a logic about (Issue 229)
   The chmod command returns -EIO when changing the mount point.
   It is correct, s3fs can not changed owner/group/mtime for the
   mount point, but s3fs sends a request for changing the bucket.
   This revision does not send the request, and returns EIO as
   soon as possible.




git-svn-id: http://s3fs.googlecode.com/svn/trunk@465 df820570-a93a-0410-bd06-b72b767a4274
2013-08-16 19:24:01 +00:00
ggtakec@gmail.com
2361c83f69 Summary of Changes(1.71 -> 1.72)
==========================
List of Changes
==========================
1) Fixes bugs and changes codes - r448, r451, r452, r453, r454, 
                     r455, r460, r461
    - Fixed umask option which works correctly.( Issue 346 )
    - Added new S3fsCurl class for lapping curl functions.
    - Deleted YIKES macro which is not used.
    - Used memcpy instead of copying each bytes while downloading.
    - Fixed a bug, s3fs did not use servicepath when renaming.
    - Fixed and changed "use_sse"/"use_rrs" options( Issue 352 )
    - Fixed a memory leak in curl_slist_sort_insert() function.
    - Fixed a memory leak when multipart uploading with error. 
    - Supported mknod function.( Issue 355 )
    - Changed debugging macros for simple.

2) Changes codes for performance and adds "multireq_max" - r449
    Changed the order for checking directory objects.
    Added "multireq_max" option is maximum number of parallel
    request for listing objects.

3) Performance tuning - r456, r457, r458, r459
    - Changed for large object uploading/downloading by parallel
      requests.
    - Added "parallel_count"/"fd_page_size" option.
    - Not make temporary file when uploading large object by
      multipart uploading.
    - Changed about temporary file and local cache file.
      And added cache's status file for local cache file.
    - Use "Range" header for block downloading.



git-svn-id: http://s3fs.googlecode.com/svn/trunk@462 df820570-a93a-0410-bd06-b72b767a4274
2013-08-10 15:37:44 +00:00
ggtakec@gmail.com
02c3accb5b Changes codes
1) Changes macros for debugging
   Changed macros for debugging messages.



git-svn-id: http://s3fs.googlecode.com/svn/trunk@461 df820570-a93a-0410-bd06-b72b767a4274
2013-08-10 15:29:39 +00:00
ggtakec@gmail.com
b24c868417 Fixed Issue 355
1) Patch in support for special file and block device types( Issue 355 )
    Patched codes, and s3fs can make special files on S3.



git-svn-id: http://s3fs.googlecode.com/svn/trunk@460 df820570-a93a-0410-bd06-b72b767a4274
2013-07-30 07:27:22 +00:00
ggtakec@gmail.com
bf11a0444f Fixed a bug
1 ) Fixed a bug
    Since r458, s3fs uses stat files of cache files. But s3fs forgot removing these
    stat files when s3fs removed cache files.
    Fixed this bug.




git-svn-id: http://s3fs.googlecode.com/svn/trunk@459 df820570-a93a-0410-bd06-b72b767a4274
2013-07-29 08:20:19 +00:00
ggtakec@gmail.com
3274f58948 Changes codes for performance(part 3)
* Summay
   This revision includes big change about temporary file and local cache file. 
   By this big change, s3fs works with good performance when s3fs opens/
   closes/syncs/reads object.
   I made a big change about the handling about temporary file and local cache
   file to do this implementation.

* Detail
1) About temporary file(local file)
   s3fs uses a temporary file on local file system when s3fs does download/
   upload/open/seek object on S3.
   After this revision, s3fs calls ftruncate() function when s3fs makes the 
   temporary file.
   In this way s3fs can set a file size of precisely length without downloading.
   (Notice - ftruncate function is for XSI-compliant systems, so that possibly
    you have a problem on non-XSI-compliant systems.)

   By this change, s3fs can download a part of a object by requesting with 
   "Range" http header. It seems like downloading by each block unit.
   The default block(part) size is 50MB, it is caused the result which is default 
   parallel requests count(5) by default multipart upload size(10MB).
   If you need to change this block size, you can change by new option 
   "fd_page_size". This option can take from 1MB(1024 * 1024) to any bytes.

   So that, you have to take care about that fdcache.cpp(and fdcache.h) were 
   changed a lot.

2) About local cache
   Local cache files which are in directory specified by "use_cache" option do 
   not have always all of object data.
   This cause is that s3fs uses ftruncate function and reads(writes) each block 
   unit of a temporary file.
   s3fs manages each block unit's status which are "downloaded area" or "not".
   For this status, s3fs makes new temporary file in cache directory which is 
   specified by "use_cache" option. This status files is in a directory which is 
   named "<use_cache sirectory>/.<bucket_name>/".

   When s3fs opens this status file, s3fs locks this file for exclusive control by 
   calling flock function. You need to take care about this, the status files can 
   not be laid on network drive(like NFS).

   This revision changes about file open mode, s3fs always opens a local cache 
   file and each status file with writable mode.
   Last, this revision adds new option "del_cache", this option means that s3fs 
   deletes all local cache file when s3fs starts and exits.

3) Uploading
   When s3fs writes data to file descriptor through FUSE request, old s3fs 
   revision downloads all of the object. But new revision does not download all, 
   it downloads only small percial area(some block units) including writing data 
   area.
   And when s3fs closes or flushes the file descriptor, s3fs downloads other area 
   which is not downloaded from server. After that,  s3fs uploads all of data.
   Already r456 revision has parallel upload function, then this revision with 
   r456 and r457 are very big change for performance.

4) Downloading
   By changing a temporary file and a local cache file, when s3fs downloads a 
   object, it downloads only the required range(some block units). 
   And s3fs downloads units by parallel GET request, it is same as a case of 
   uploading. (Maximum parallel request count and each download size are 
   specified same parameters for uploading.)

   In the new revision, when s3fs opens file, s3fs returns file descriptor soon.
   Because s3fs only opens(makes) the file descriptor with no downloading 
   data. And when s3fs reads a data, s3fs downloads only some block unit 
   including specified area.
   This result is good for performance.

5) Changes option name
   The option "parallel_upload" which added at r456 is changed to new option 
   name as "parallel_count". This reason is this option value is not only used by 
   uploading object, but a uploading object also uses this option. (For a while, 
   you can use old option name "parallel_upload" for compatibility.)



git-svn-id: http://s3fs.googlecode.com/svn/trunk@458 df820570-a93a-0410-bd06-b72b767a4274
2013-07-23 16:01:48 +00:00
ggtakec@gmail.com
1c93dd30c1 Changes codes
1) For uploading performance(part 2)
   Changed a codes about uploading large object(multipart uploading).
   This revision does not make temporary file when s3fs uploads large object by multipart uploading.
   Before this revision, s3fs made temporary file(/tmp/s3fs.XXXXX) for multipart, but it was not good for performance.
   So that, new codes do not use those files, and s3fs reads directly large object from s3fs's cache file.

2) Some value to symbol
   Changed some value to symbol(define).



git-svn-id: http://s3fs.googlecode.com/svn/trunk@457 df820570-a93a-0410-bd06-b72b767a4274
2013-07-12 00:33:36 +00:00
ggtakec@gmail.com
1095b7bc52 Changes codes
1) For uploading performance(part 1)
   Changed a code for large object uploading.
   New codes makes s3fs send parallel requests when s3fs uploads large 
   object(20MB) by multipart post.

   And added new "parallel_upload" option, which limits parallel request 
   count which s3fs requests at once.
   This option's default value is "5", and you can change this value. But it
   is necessary to set this value depending on a CPU and a network band.
   s3fs became to work good performance by this option, please try to set 
   your value for this option.

2) Changes debugging messages
    Changed debugging message in s3fs.cpp.



git-svn-id: http://s3fs.googlecode.com/svn/trunk@456 df820570-a93a-0410-bd06-b72b767a4274
2013-07-10 06:24:06 +00:00
ggtakec@gmail.com
6e169f6bda Changes codes
1) Changes a code in PutRequest function
   Changed a code in S3fsCurl:: PutRequest function to duplicate file discriptor in 
   this function.

2) Changes debugging messages
    Changed debugging message's indent in curl.cpp functions. 



git-svn-id: http://s3fs.googlecode.com/svn/trunk@455 df820570-a93a-0410-bd06-b72b767a4274
2013-07-08 01:25:11 +00:00
ggtakec@gmail.com
0c630ba2d0 Fixed a bug
1) Fixed a bug
    When something error occured in multipart uploading process, s3fs forgets to free memory.
    (from r451)
    Fixed this bug.




git-svn-id: http://s3fs.googlecode.com/svn/trunk@454 df820570-a93a-0410-bd06-b72b767a4274
2013-07-05 06:36:11 +00:00
ggtakec@gmail.com
b1e1312cfb Fixed Issue 352(+)
1) Option syntax verbosity in doc (  Issue 352  )
    Mistook fixing codes, then re-check'in correctly code.



git-svn-id: http://s3fs.googlecode.com/svn/trunk@453 df820570-a93a-0410-bd06-b72b767a4274
2013-07-05 05:49:41 +00:00
ggtakec@gmail.com
d1a17cbe3d Fixed Issue 352 and bugs
1) Option syntax verbosity in doc ( Issue 352 )
    Before this revision(version), "use_rrs" option needs to set a parameter like "use_sse" option.
    But this option does not need a parameter, specified "use_rrs" option means enabled RRS.
    (because RRS is desabled by default.)
    After this revision, "use_rrs" option can be specified without a parameter, and "use_sse" too.
    Changed codes, man page and help page.
    Please notice, for old version "use_rrs"(and "use_sse") can be specified with a parameter("1" or "0") yet.

2) Fixes a bug about analizing "use_sse" option.
    Fixed a bug in r451, "use_sse" option is not worked because s3fs mistook to call function for "use_rrs".

3) Fixes a memory leak.
    Fixed a memory leak in r451.
    Fixed that the curl_slist_sort_insert() function forgot to free memory.



git-svn-id: http://s3fs.googlecode.com/svn/trunk@452 df820570-a93a-0410-bd06-b72b767a4274
2013-07-05 05:41:46 +00:00
ggtakec@gmail.com
ad19ffa458 Changes codes
1) Adds new S3fsCurl class
   Added new S3fsCurl class instead of directly calling curl function.
   This class is lapping curl function for s3fs(AWS S3 API).

2) Changes codes about adding S3fsCurl class
    Changed and deleted classes and structures which are related to curl in curl.cpp/curl.h.
    Changed codes which are calling S3 API with curl in s3fs.cpp.

3) Deletes YKIES macro
    Deleted YIKES macro, because this macro is used no more.

4) Changes a code
    s3fs does not get good performance because s3fs copies each byte while downloading.
    So that the codes is changed instead of memcpy, then s3fs performance not a little improves.

5) Fixes a bug
    When s3fs renames a file, s3fs does not use the value which is specified by servicepath option.
    Fixed this bug.



git-svn-id: http://s3fs.googlecode.com/svn/trunk@451 df820570-a93a-0410-bd06-b72b767a4274
2013-07-05 02:28:31 +00:00
ggtakec@gmail.com
5a035a33f0 Changes codes
1) Changes FdCache class(cleanup codes)
    The FdCache class is for caching file discriptor.
    This class is modified as adding reference count for file discriptor and 
    removing pid for each path.



git-svn-id: http://s3fs.googlecode.com/svn/trunk@450 df820570-a93a-0410-bd06-b72b767a4274
2013-06-21 06:07:22 +00:00
ggtakec@gmail.com
45950044f7 Changes codes
1) Changes codes for performance and request's costs
    s3fs gets object's  attributes by using HEAD request.
    Directory objects is following 4 type:
      a) name type is "dir", with meta information
      b) name type is "dir", without meta information(but has files in itself)
      c) name type is "dir/", with(out) meta information
      d) name type is "dir_$folder$", with(out) meta information
    The codes is changed to order checking directory object.
    So that, s3fs decreases requests for checking objects.

    Before version has a bug, that is s3fs can not be recognizable type-b) 
    definitely when s3fs check the object directly.(but s3fs can, when s3fs 
    check the object by listing)
    This change fixes this bug.

2) Adds "multireq_max" option
    Added "multireq_max" option is maximum number of parallel request 
    for listing objects.
    This changes is possible to solve CURLE_COULDNT_CONNECT.
    If this option can not solve it, this option will be useful for tuning 
    performance by each.




git-svn-id: http://s3fs.googlecode.com/svn/trunk@449 df820570-a93a-0410-bd06-b72b767a4274
2013-06-19 14:53:58 +00:00
ggtakec@gmail.com
b65c3e195e Fixed Issue 346
1) Not recognizing group permissions( Issue 346 )
    Fixed umask option which works correctly.



git-svn-id: http://s3fs.googlecode.com/svn/trunk@448 df820570-a93a-0410-bd06-b72b767a4274
2013-06-18 01:17:32 +00:00