Make man page and --help more consistent

Mostly small fixes but also some reordering.  References #929.  Found
via:

diff -u <(man --no-hyphenation doc/man/s3fs.1 | tr -s ' ' '\n' | sed '/^-o$/d' ) <(src/s3fs --help | tr -s ' ' '\n' | sed '/^-$/d')
This commit is contained in:
Andrew Gaul 2019-01-28 23:35:42 -08:00
parent e5e124b9aa
commit 8a18806a57
2 changed files with 98 additions and 85 deletions

View File

@ -6,7 +6,7 @@ S3FS \- FUSE-based file system backed by Amazon S3
.TP .TP
\fBs3fs bucket[:/path] mountpoint \fP [options] \fBs3fs bucket[:/path] mountpoint \fP [options]
.TP .TP
\fBs3fs mountpoint \fP [options(must specify bucket= option)] \fBs3fs mountpoint \fP [options (must specify bucket= option)]
.SS unmounting .SS unmounting
.TP .TP
\fBumount mountpoint \fBumount mountpoint
@ -14,9 +14,9 @@ For root.
.TP .TP
\fBfusermount -u mountpoint \fBfusermount -u mountpoint
For unprivileged user. For unprivileged user.
.SS utility mode ( remove interrupted multipart uploading objects ) .SS utility mode (remove interrupted multipart uploading objects)
.TP .TP
\fBs3fs --incomplete-mpu-list(-u) bucket \fBs3fs --incomplete-mpu-list (-u) bucket
.TP .TP
\fBs3fs --incomplete-mpu-abort[=all | =<expire date format>] bucket \fBs3fs --incomplete-mpu-abort[=all | =<expire date format>] bucket
.SH DESCRIPTION .SH DESCRIPTION
@ -59,7 +59,7 @@ All s3fs options must given in the form where "opt" is:
<option_name>=<option_value> <option_name>=<option_value>
.TP .TP
\fB\-o\fR bucket \fB\-o\fR bucket
if it is not specified bucket name(and path) in command line, must specify this option after \-o option for bucket name. if it is not specified bucket name (and path) in command line, must specify this option after \-o option for bucket name.
.TP .TP
\fB\-o\fR default_acl (default="private") \fB\-o\fR default_acl (default="private")
the default canned acl to apply to all written s3 objects, e.g., "private", "public-read". the default canned acl to apply to all written s3 objects, e.g., "private", "public-read".
@ -79,7 +79,7 @@ If this option is not specified, it will be created at runtime when the cache di
\fB\-o\fR del_cache - delete local file cache \fB\-o\fR del_cache - delete local file cache
delete local file cache when s3fs starts and exits. delete local file cache when s3fs starts and exits.
.TP .TP
\fB\-o\fR storage_class (default is standard) \fB\-o\fR storage_class (default="standard")
store object with specified storage class. store object with specified storage class.
this option replaces the old option use_rrs. this option replaces the old option use_rrs.
Possible values: standard, standard_ia, onezone_ia and reduced_redundancy. Possible values: standard, standard_ia, onezone_ia and reduced_redundancy.
@ -93,19 +93,19 @@ this option has been replaced by new storage_class option.
\fB\-o\fR use_sse (default is disable) \fB\-o\fR use_sse (default is disable)
Specify three type Amazon's Server-Site Encryption: SSE-S3, SSE-C or SSE-KMS. SSE-S3 uses Amazon S3-managed encryption keys, SSE-C uses customer-provided encryption keys, and SSE-KMS uses the master key which you manage in AWS KMS. Specify three type Amazon's Server-Site Encryption: SSE-S3, SSE-C or SSE-KMS. SSE-S3 uses Amazon S3-managed encryption keys, SSE-C uses customer-provided encryption keys, and SSE-KMS uses the master key which you manage in AWS KMS.
You can specify "use_sse" or "use_sse=1" enables SSE-S3 type (use_sse=1 is old type parameter). You can specify "use_sse" or "use_sse=1" enables SSE-S3 type (use_sse=1 is old type parameter).
Case of setting SSE-C, you can specify "use_sse=custom", "use_sse=custom:<custom key file path>" or "use_sse=<custom key file path>"(only <custom key file path> specified is old type parameter). Case of setting SSE-C, you can specify "use_sse=custom", "use_sse=custom:<custom key file path>" or "use_sse=<custom key file path>" (only <custom key file path> specified is old type parameter).
You can use "c" for short "custom". You can use "c" for short "custom".
The custom key file must be 600 permission. The file can have some lines, each line is one SSE-C key. The custom key file must be 600 permission. The file can have some lines, each line is one SSE-C key.
The first line in file is used as Customer-Provided Encryption Keys for uploading and changing headers etc. The first line in file is used as Customer-Provided Encryption Keys for uploading and changing headers etc.
If there are some keys after first line, those are used downloading object which are encrypted by not first key. If there are some keys after first line, those are used downloading object which are encrypted by not first key.
So that, you can keep all SSE-C keys in file, that is SSE-C key history. So that, you can keep all SSE-C keys in file, that is SSE-C key history.
If you specify "custom"("c") without file path, you need to set custom key by load_sse_c option or AWSSSECKEYS environment.(AWSSSECKEYS environment has some SSE-C keys with ":" separator.) If you specify "custom" ("c") without file path, you need to set custom key by load_sse_c option or AWSSSECKEYS environment. (AWSSSECKEYS environment has some SSE-C keys with ":" separator.)
This option is used to decide the SSE type. This option is used to decide the SSE type.
So that if you do not want to encrypt a object at uploading, but you need to decrypt encrypted object at downloading, you can use load_sse_c option instead of this option. So that if you do not want to encrypt a object at uploading, but you need to decrypt encrypted object at downloading, you can use load_sse_c option instead of this option.
For setting SSE-KMS, specify "use_sse=kmsid" or "use_sse=kmsid:<kms id>". For setting SSE-KMS, specify "use_sse=kmsid" or "use_sse=kmsid:<kms id>".
You can use "k" for short "kmsid". You can use "k" for short "kmsid".
If you san specify SSE-KMS type with your <kms id> in AWS KMS, you can set it after "kmsid:"(or "k:"). If you san specify SSE-KMS type with your <kms id> in AWS KMS, you can set it after "kmsid:" (or "k:").
If you specify only "kmsid"("k"), you need to set AWSSSEKMSID environment which value is <kms id>. If you specify only "kmsid" ("k"), you need to set AWSSSEKMSID environment which value is <kms id>.
You must be careful about that you can not use the KMS id which is not same EC2 region. You must be careful about that you can not use the KMS id which is not same EC2 region.
.TP .TP
\fB\-o\fR load_sse_c - specify SSE-C keys \fB\-o\fR load_sse_c - specify SSE-C keys
@ -119,12 +119,12 @@ AWSSSECKEYS environment is as same as this file contents.
specify the path to the password file, which which takes precedence over the password in $HOME/.passwd-s3fs and /etc/passwd-s3fs specify the path to the password file, which which takes precedence over the password in $HOME/.passwd-s3fs and /etc/passwd-s3fs
.TP .TP
\fB\-o\fR ahbe_conf (default="" which means disabled) \fB\-o\fR ahbe_conf (default="" which means disabled)
This option specifies the configuration file path which file is the additional HTTP header by file(object) extension. This option specifies the configuration file path which file is the additional HTTP header by file (object) extension.
The configuration file format is below: The configuration file format is below:
----------- -----------
line = [file suffix or regex] HTTP-header [HTTP-values] line = [file suffix or regex] HTTP-header [HTTP-values]
file suffix = file(object) suffix, if this field is empty, it means "reg:(.*)".(=all object). file suffix = file (object) suffix, if this field is empty, it means "reg:(.*)".(=all object).
regex = regular expression to match the file(object) path. this type starts with "reg:" prefix. regex = regular expression to match the file (object) path. this type starts with "reg:" prefix.
HTTP-header = additional HTTP header name HTTP-header = additional HTTP header name
HTTP-values = additional HTTP header value HTTP-values = additional HTTP header value
----------- -----------
@ -158,17 +158,17 @@ specify the maximum number of keys returned by S3 list object API. The default i
maximum number of entries in the stat cache maximum number of entries in the stat cache
.TP .TP
\fB\-o\fR stat_cache_expire (default is no expire) \fB\-o\fR stat_cache_expire (default is no expire)
specify expire time(seconds) for entries in the stat cache. This expire time indicates the time since stat cached. specify expire time (seconds) for entries in the stat cache. This expire time indicates the time since stat cached.
.TP .TP
\fB\-o\fR stat_cache_interval_expire (default is no expire) \fB\-o\fR stat_cache_interval_expire (default is no expire)
specify expire time(seconds) for entries in the stat cache. This expire time is based on the time from the last access time of the stat cache. specify expire time (seconds) for entries in the stat cache. This expire time is based on the time from the last access time of the stat cache.
This option is exclusive with stat_cache_expire, and is left for compatibility with older versions. This option is exclusive with stat_cache_expire, and is left for compatibility with older versions.
.TP .TP
\fB\-o\fR enable_noobj_cache (default is disable) \fB\-o\fR enable_noobj_cache (default is disable)
enable cache entries for the object which does not exist. enable cache entries for the object which does not exist.
s3fs always has to check whether file(or sub directory) exists under object(path) when s3fs does some command, since s3fs has recognized a directory which does not exist and has files or sub directories under itself. s3fs always has to check whether file (or sub directory) exists under object (path) when s3fs does some command, since s3fs has recognized a directory which does not exist and has files or sub directories under itself.
It increases ListBucket request and makes performance bad. It increases ListBucket request and makes performance bad.
You can specify this option for performance, s3fs memorizes in stat cache that the object(file or directory) does not exist. You can specify this option for performance, s3fs memorizes in stat cache that the object (file or directory) does not exist.
.TP .TP
\fB\-o\fR no_check_certificate (by default this option is disabled) \fB\-o\fR no_check_certificate (by default this option is disabled)
do not check ssl certificate. do not check ssl certificate.
@ -185,26 +185,29 @@ maximum number of parallel request for listing objects.
.TP .TP
\fB\-o\fR parallel_count (default="5") \fB\-o\fR parallel_count (default="5")
number of parallel request for uploading big objects. number of parallel request for uploading big objects.
s3fs uploads large object(default:over 20MB) by multipart post request, and sends parallel requests. s3fs uploads large object (over 20MB) by multipart post request, and sends parallel requests.
This option limits parallel request count which s3fs requests at once. This option limits parallel request count which s3fs requests at once.
It is necessary to set this value depending on a CPU and a network band. It is necessary to set this value depending on a CPU and a network band.
.TP .TP
\fB\-o\fR multipart_size(default="10"(10MB)) \fB\-o\fR multipart_size (default="10")
number of one part size in multipart uploading request. part size, in MB, for each multipart request.
The default size is 10MB(10485760byte), minimum value is 5MB(5242880byte). The minimum value is 5 MB and the maximum value is 5 GB.
Specify number of MB and over 5(MB).
.TP .TP
\fB\-o\fR ensure_diskfree(default 0) \fB\-o\fR ensure_diskfree (default 0)
sets MB to ensure disk free space. This option means the threshold of free space size on disk which is used for the cache file by s3fs. sets MB to ensure disk free space. This option means the threshold of free space size on disk which is used for the cache file by s3fs.
s3fs makes file for downloading, and uploading and caching files. s3fs makes file for downloading, uploading and caching files.
If the disk free space is smaller than this value, s3fs do not use diskspace as possible in exchange for the performance. If the disk free space is smaller than this value, s3fs do not use diskspace as possible in exchange for the performance.
.TP .TP
\fB\-o\fR singlepart_copy_limit (default="5120")
maximum size, in MB, of a single-part copy before trying
multipart copy.
.TP
\fB\-o\fR url (default="https://s3.amazonaws.com") \fB\-o\fR url (default="https://s3.amazonaws.com")
sets the url to use to access Amazon S3. If you want to use HTTP, then you can set "url=http://s3.amazonaws.com". sets the url to use to access Amazon S3. If you want to use HTTP, then you can set "url=http://s3.amazonaws.com".
If you do not use https, please specify the URL with the url option. If you do not use https, please specify the URL with the url option.
.TP .TP
\fB\-o\fR endpoint (default="us-east-1") \fB\-o\fR endpoint (default="us-east-1")
sets the endpoint to use. sets the endpoint to use on signature version 4.
If this option is not specified, s3fs uses "us-east-1" region as the default. If this option is not specified, s3fs uses "us-east-1" region as the default.
If the s3fs could not connect to the region specified by this option, s3fs could not run. If the s3fs could not connect to the region specified by this option, s3fs could not run.
But if you do not specify this option, and if you can not connect with the default region, s3fs will retry to automatically connect to the other region. But if you do not specify this option, and if you can not connect with the default region, s3fs will retry to automatically connect to the other region.
@ -217,28 +220,28 @@ sets signing AWS requests by using Signature Version 2.
sets umask for the mount point directory. sets umask for the mount point directory.
If allow_other option is not set, s3fs allows access to the mount point only to the owner. If allow_other option is not set, s3fs allows access to the mount point only to the owner.
In the opposite case s3fs allows access to all users as the default. In the opposite case s3fs allows access to all users as the default.
But if you set the allow_other with this option, you can control permissions of the mount point by this option like umask. But if you set the allow_other with this option, you can control the permissions of the mount point by this option like umask.
.TP .TP
\fB\-o\fR nomultipart - disable multipart uploads \fB\-o\fR nomultipart - disable multipart uploads
.TP .TP
\fB\-o\fR enable_content_md5 ( default is disable ) \fB\-o\fR enable_content_md5 (default is disable)
Allow S3 server to check data integrity of uploads via the Content-MD5 header. Allow S3 server to check data integrity of uploads via the Content-MD5 header.
This can add CPU overhead to transfers. This can add CPU overhead to transfers.
.TP .TP
\fB\-o\fR ecs ( default is disable ) \fB\-o\fR ecs (default is disable)
This option instructs s3fs to query the ECS container credential metadata address instead of the instance metadata address. This option instructs s3fs to query the ECS container credential metadata address instead of the instance metadata address.
.TP .TP
\fB\-o\fR iam_role ( default is no IAM role ) \fB\-o\fR iam_role (default is no IAM role)
This option requires the IAM role name or "auto". If you specify "auto", s3fs will automatically use the IAM role names that are set to an instance. If you specify this option without any argument, it is the same as that you have specified the "auto". This option requires the IAM role name or "auto". If you specify "auto", s3fs will automatically use the IAM role names that are set to an instance. If you specify this option without any argument, it is the same as that you have specified the "auto".
.TP .TP
\fB\-o\fR ibm_iam_auth ( default is not using IBM IAM authentication ) \fB\-o\fR ibm_iam_auth (default is not using IBM IAM authentication)
This option instructs s3fs to use IBM IAM authentication. In this mode, the AWSAccessKey and AWSSecretKey will be used as IBM's Service-Instance-ID and APIKey, respectively. This option instructs s3fs to use IBM IAM authentication. In this mode, the AWSAccessKey and AWSSecretKey will be used as IBM's Service-Instance-ID and APIKey, respectively.
.TP .TP
\fB\-o\fR ibm_iam_endpoint ( default is https://iam.bluemix.net ) \fB\-o\fR ibm_iam_endpoint (default is https://iam.bluemix.net)
Set the URL to use for IBM IAM authentication. Sets the URL to use for IBM IAM authentication.
.TP .TP
\fB\-o\fR use_xattr ( default is not handling the extended attribute ) \fB\-o\fR use_xattr (default is not handling the extended attribute)
Enable to handle the extended attribute(xattrs). Enable to handle the extended attribute (xattrs).
If you set this option, you can use the extended attribute. If you set this option, you can use the extended attribute.
For example, encfs and ecryptfs need to support the extended attribute. For example, encfs and ecryptfs need to support the extended attribute.
Notice: if s3fs handles the extended attribute, s3fs can not work to copy command with preserve=mode. Notice: if s3fs handles the extended attribute, s3fs can not work to copy command with preserve=mode.
@ -248,12 +251,12 @@ disable registering xml name space for response of ListBucketResult and ListVers
This option should not be specified now, because s3fs looks up xmlns automatically after v1.66. This option should not be specified now, because s3fs looks up xmlns automatically after v1.66.
.TP .TP
\fB\-o\fR nocopyapi - for other incomplete compatibility object storage. \fB\-o\fR nocopyapi - for other incomplete compatibility object storage.
For a distributed object storage which is compatibility S3 API without PUT(copy api). For a distributed object storage which is compatibility S3 API without PUT (copy api).
If you set this option, s3fs do not use PUT with "x-amz-copy-source"(copy api). Because traffic is increased 2-3 times by this option, we do not recommend this. If you set this option, s3fs do not use PUT with "x-amz-copy-source" (copy api). Because traffic is increased 2-3 times by this option, we do not recommend this.
.TP .TP
\fB\-o\fR norenameapi - for other incomplete compatibility object storage. \fB\-o\fR norenameapi - for other incomplete compatibility object storage.
For a distributed object storage which is compatibility S3 API without PUT(copy api). For a distributed object storage which is compatibility S3 API without PUT (copy api).
This option is a subset of nocopyapi option. The nocopyapi option does not use copy-api for all command(ex. chmod, chown, touch, mv, etc), but this option does not use copy-api for only rename command(ex. mv). This option is a subset of nocopyapi option. The nocopyapi option does not use copy-api for all command (ex. chmod, chown, touch, mv, etc), but this option does not use copy-api for only rename command (ex. mv).
If this option is specified with nocopyapi, then s3fs ignores it. If this option is specified with nocopyapi, then s3fs ignores it.
.TP .TP
\fB\-o\fR use_path_request_style (use legacy API calling style) \fB\-o\fR use_path_request_style (use legacy API calling style)
@ -264,7 +267,7 @@ Usually s3fs outputs of the User-Agent in "s3fs/<version> (commit hash <hash>; <
If this option is specified, s3fs suppresses the output of the User-Agent. If this option is specified, s3fs suppresses the output of the User-Agent.
.TP .TP
\fB\-o\fR cipher_suites \fB\-o\fR cipher_suites
Customize TLS cipher suite list. Expects a colon separated list of cipher suite names. Customize the list of TLS cipher suites. Expects a colon separated list of cipher suite names.
A list of available cipher suites, depending on your TLS engine, can be found on the CURL library documentation: A list of available cipher suites, depending on your TLS engine, can be found on the CURL library documentation:
https://curl.haxx.se/docs/ssl-ciphers.html https://curl.haxx.se/docs/ssl-ciphers.html
.TP .TP
@ -296,7 +299,7 @@ Unicode set.
Useful on clients not using utf-8 as their file system encoding. Useful on clients not using utf-8 as their file system encoding.
.TP .TP
\fB\-o\fR dbglevel (default="crit") \fB\-o\fR dbglevel (default="crit")
Set the debug message level. set value as crit(critical), err(error), warn(warning), info(information) to debug level. default debug level is critical. Set the debug message level. set value as crit (critical), err (error), warn (warning), info (information) to debug level. default debug level is critical.
If s3fs run with "-d" option, the debug level is set information. If s3fs run with "-d" option, the debug level is set information.
When s3fs catch the signal SIGUSR2, the debug level is bumpup. When s3fs catch the signal SIGUSR2, the debug level is bumpup.
.TP .TP
@ -307,10 +310,10 @@ Put the debug message from libcurl when this option is specified.
\fB\-u\fR or \fB\-\-incomplete\-mpu\-list\fR \fB\-u\fR or \fB\-\-incomplete\-mpu\-list\fR
Lists multipart incomplete objects uploaded to the specified bucket. Lists multipart incomplete objects uploaded to the specified bucket.
.TP .TP
\fB\-\-incomplete\-mpu\-abort\fR all or date format(default="24H") \fB\-\-incomplete\-mpu\-abort\fR all or date format (default="24H")
Delete the multipart incomplete object uploaded to the specified bucket. Delete the multipart incomplete object uploaded to the specified bucket.
If "all" is specified for this option, all multipart incomplete objects will be deleted. If "all" is specified for this option, all multipart incomplete objects will be deleted.
If you specify no argument as an option, objects older than 24 hours(24H) will be deleted(This is the default value). If you specify no argument as an option, objects older than 24 hours (24H) will be deleted (This is the default value).
You can specify an optional date format. You can specify an optional date format.
It can be specified as year, month, day, hour, minute, second, and it is expressed as "Y", "M", "D", "h", "m", "s" respectively. It can be specified as year, month, day, hour, minute, second, and it is expressed as "Y", "M", "D", "h", "m", "s" respectively.
For example, "1Y6M10D12h30m30s". For example, "1Y6M10D12h30m30s".

View File

@ -991,9 +991,9 @@ void show_help ()
"Usage:\n" "Usage:\n"
" mounting\n" " mounting\n"
" s3fs bucket[:/path] mountpoint [options]\n" " s3fs bucket[:/path] mountpoint [options]\n"
" s3fs mountpoint [options(must specify bucket= option)]\n" " s3fs mountpoint [options (must specify bucket= option)]\n"
"\n" "\n"
" umounting\n" " unmounting\n"
" umount mountpoint\n" " umount mountpoint\n"
"\n" "\n"
" General forms for s3fs and FUSE/mount options:\n" " General forms for s3fs and FUSE/mount options:\n"
@ -1001,7 +1001,7 @@ void show_help ()
" -o opt [-o opt] ...\n" " -o opt [-o opt] ...\n"
"\n" "\n"
" utility mode (remove interrupted multipart uploading objects)\n" " utility mode (remove interrupted multipart uploading objects)\n"
" s3fs --incomplete-mpu-list(-u) bucket\n" " s3fs --incomplete-mpu-list (-u) bucket\n"
" s3fs --incomplete-mpu-abort[=all | =<date format>] bucket\n" " s3fs --incomplete-mpu-abort[=all | =<date format>] bucket\n"
"\n" "\n"
"s3fs Options:\n" "s3fs Options:\n"
@ -1011,7 +1011,7 @@ void show_help ()
" <option_name>=<option_value>\n" " <option_name>=<option_value>\n"
"\n" "\n"
" bucket\n" " bucket\n"
" - if it is not specified bucket name(and path) in command line,\n" " - if it is not specified bucket name (and path) in command line,\n"
" must specify this option after -o option for bucket name.\n" " must specify this option after -o option for bucket name.\n"
"\n" "\n"
" default_acl (default=\"private\")\n" " default_acl (default=\"private\")\n"
@ -1022,14 +1022,14 @@ void show_help ()
" for the full list of canned acls\n" " for the full list of canned acls\n"
"\n" "\n"
" retries (default=\"5\")\n" " retries (default=\"5\")\n"
" - number of times to retry a failed s3 transaction\n" " - number of times to retry a failed S3 transaction\n"
"\n" "\n"
" use_cache (default=\"\" which means disabled)\n" " use_cache (default=\"\" which means disabled)\n"
" - local folder to use for local file cache\n" " - local folder to use for local file cache\n"
"\n" "\n"
" check_cache_dir_exist (default is disable)\n" " check_cache_dir_exist (default is disable)\n"
" - if use_cache is set, check if the cache directory exists.\n" " - if use_cache is set, check if the cache directory exists.\n"
" if this option is not specified, it will be created at runtime\n" " If this option is not specified, it will be created at runtime\n"
" when the cache directory does not exist.\n" " when the cache directory does not exist.\n"
"\n" "\n"
" del_cache (delete local file cache)\n" " del_cache (delete local file cache)\n"
@ -1045,10 +1045,10 @@ void show_help ()
" keys, SSE-C uses customer-provided encryption keys, and\n" " keys, SSE-C uses customer-provided encryption keys, and\n"
" SSE-KMS uses the master key which you manage in AWS KMS.\n" " SSE-KMS uses the master key which you manage in AWS KMS.\n"
" You can specify \"use_sse\" or \"use_sse=1\" enables SSE-S3\n" " You can specify \"use_sse\" or \"use_sse=1\" enables SSE-S3\n"
" type(use_sse=1 is old type parameter).\n" " type (use_sse=1 is old type parameter).\n"
" Case of setting SSE-C, you can specify \"use_sse=custom\",\n" " Case of setting SSE-C, you can specify \"use_sse=custom\",\n"
" \"use_sse=custom:<custom key file path>\" or\n" " \"use_sse=custom:<custom key file path>\" or\n"
" \"use_sse=<custom key file path>\"(only <custom key file path>\n" " \"use_sse=<custom key file path>\" (only <custom key file path>\n"
" specified is old type parameter). You can use \"c\" for\n" " specified is old type parameter). You can use \"c\" for\n"
" short \"custom\".\n" " short \"custom\".\n"
" The custom key file must be 600 permission. The file can\n" " The custom key file must be 600 permission. The file can\n"
@ -1058,9 +1058,9 @@ void show_help ()
" after first line, those are used downloading object which\n" " after first line, those are used downloading object which\n"
" are encrypted by not first key. So that, you can keep all\n" " are encrypted by not first key. So that, you can keep all\n"
" SSE-C keys in file, that is SSE-C key history.\n" " SSE-C keys in file, that is SSE-C key history.\n"
" If you specify \"custom\"(\"c\") without file path, you\n" " If you specify \"custom\" (\"c\") without file path, you\n"
" need to set custom key by load_sse_c option or AWSSSECKEYS\n" " need to set custom key by load_sse_c option or AWSSSECKEYS\n"
" environment.(AWSSSECKEYS environment has some SSE-C keys\n" " environment. (AWSSSECKEYS environment has some SSE-C keys\n"
" with \":\" separator.) This option is used to decide the\n" " with \":\" separator.) This option is used to decide the\n"
" SSE type. So that if you do not want to encrypt a object\n" " SSE type. So that if you do not want to encrypt a object\n"
" object at uploading, but you need to decrypt encrypted\n" " object at uploading, but you need to decrypt encrypted\n"
@ -1069,8 +1069,8 @@ void show_help ()
" For setting SSE-KMS, specify \"use_sse=kmsid\" or\n" " For setting SSE-KMS, specify \"use_sse=kmsid\" or\n"
" \"use_sse=kmsid:<kms id>\". You can use \"k\" for short \"kmsid\".\n" " \"use_sse=kmsid:<kms id>\". You can use \"k\" for short \"kmsid\".\n"
" If you san specify SSE-KMS type with your <kms id> in AWS\n" " If you san specify SSE-KMS type with your <kms id> in AWS\n"
" KMS, you can set it after \"kmsid:\"(or \"k:\"). If you\n" " KMS, you can set it after \"kmsid:\" (or \"k:\"). If you\n"
" specify only \"kmsid\"(\"k\"), you need to set AWSSSEKMSID\n" " specify only \"kmsid\" (\"k\"), you need to set AWSSSEKMSID\n"
" environment which value is <kms id>. You must be careful\n" " environment which value is <kms id>. You must be careful\n"
" about that you can not use the KMS id which is not same EC2\n" " about that you can not use the KMS id which is not same EC2\n"
" region.\n" " region.\n"
@ -1096,13 +1096,13 @@ void show_help ()
"\n" "\n"
" ahbe_conf (default=\"\" which means disabled)\n" " ahbe_conf (default=\"\" which means disabled)\n"
" - This option specifies the configuration file path which\n" " - This option specifies the configuration file path which\n"
" file is the additional HTTP header by file(object) extension.\n" " file is the additional HTTP header by file (object) extension.\n"
" The configuration file format is below:\n" " The configuration file format is below:\n"
" -----------\n" " -----------\n"
" line = [file suffix or regex] HTTP-header [HTTP-values]\n" " line = [file suffix or regex] HTTP-header [HTTP-values]\n"
" file suffix = file(object) suffix, if this field is empty,\n" " file suffix = file (object) suffix, if this field is empty,\n"
" it means \"reg:(.*)\".(=all object).\n" " it means \"reg:(.*)\".(=all object).\n"
" regex = regular expression to match the file(object) path.\n" " regex = regular expression to match the file (object) path.\n"
" this type starts with \"reg:\" prefix.\n" " this type starts with \"reg:\" prefix.\n"
" HTTP-header = additional HTTP header name\n" " HTTP-header = additional HTTP header name\n"
" HTTP-values = additional HTTP header value\n" " HTTP-values = additional HTTP header value\n"
@ -1136,18 +1136,25 @@ void show_help ()
" - maximum number of entries in the stat cache\n" " - maximum number of entries in the stat cache\n"
"\n" "\n"
" stat_cache_expire (default is no expire)\n" " stat_cache_expire (default is no expire)\n"
" - specify expire time(seconds) for entries in the stat cache.\n" " - specify expire time (seconds) for entries in the stat cache.\n"
" This expire time indicates the time since stat cached.\n" " This expire time indicates the time since stat cached.\n"
"\n" "\n"
" stat_cache_interval_expire (default is no expire)\n"
" - specify expire time (seconds) for entries in the stat cache.\n"
" This expire time is based on the time from the last access time\n"
" of the stat cache. This option is exclusive with\n"
" stat_cache_expire, and is left for compatibility with older\n"
" versions.\n"
"\n"
" enable_noobj_cache (default is disable)\n" " enable_noobj_cache (default is disable)\n"
" - enable cache entries for the object which does not exist.\n" " - enable cache entries for the object which does not exist.\n"
" s3fs always has to check whether file(or sub directory) exists \n" " s3fs always has to check whether file (or sub directory) exists \n"
" under object(path) when s3fs does some command, since s3fs has \n" " under object (path) when s3fs does some command, since s3fs has \n"
" recognized a directory which does not exist and has files or \n" " recognized a directory which does not exist and has files or \n"
" sub directories under itself. It increases ListBucket request \n" " sub directories under itself. It increases ListBucket request \n"
" and makes performance bad.\n" " and makes performance bad.\n"
" You can specify this option for performance, s3fs memorizes \n" " You can specify this option for performance, s3fs memorizes \n"
" in stat cache that the object(file or directory) does not exist.\n" " in stat cache that the object (file or directory) does not exist.\n"
"\n" "\n"
" no_check_certificate\n" " no_check_certificate\n"
" - server certificate won't be checked against the available \n" " - server certificate won't be checked against the available \n"
@ -1165,7 +1172,7 @@ void show_help ()
"\n" "\n"
" parallel_count (default=\"5\")\n" " parallel_count (default=\"5\")\n"
" - number of parallel request for uploading big objects.\n" " - number of parallel request for uploading big objects.\n"
" s3fs uploads large object(over 20MB) by multipart post request, \n" " s3fs uploads large object (over 20MB) by multipart post request, \n"
" and sends parallel requests.\n" " and sends parallel requests.\n"
" This option limits parallel request count which s3fs requests \n" " This option limits parallel request count which s3fs requests \n"
" at once. It is necessary to set this value depending on a CPU \n" " at once. It is necessary to set this value depending on a CPU \n"
@ -1173,9 +1180,12 @@ void show_help ()
"\n" "\n"
" multipart_size (default=\"10\")\n" " multipart_size (default=\"10\")\n"
" - part size, in MB, for each multipart request.\n" " - part size, in MB, for each multipart request.\n"
" The minimum value is 5 MB and the maximum value is 5 GB.\n"
"\n" "\n"
" ensure_diskfree (default 0)\n" " ensure_diskfree (default 0)\n"
" - sets MB to ensure disk free space. s3fs makes file for\n" " - sets MB to ensure disk free space. This option means the\n"
" threshold of free space size on disk which is used for the\n"
" cache file by s3fs. s3fs makes file for\n"
" downloading, uploading and caching files. If the disk free\n" " downloading, uploading and caching files. If the disk free\n"
" space is smaller than this value, s3fs do not use diskspace\n" " space is smaller than this value, s3fs do not use diskspace\n"
" as possible in exchange for the performance.\n" " as possible in exchange for the performance.\n"
@ -1217,7 +1227,7 @@ void show_help ()
" Allow S3 server to check data integrity of uploads via the\n" " Allow S3 server to check data integrity of uploads via the\n"
" Content-MD5 header. This can add CPU overhead to transfers.\n" " Content-MD5 header. This can add CPU overhead to transfers.\n"
"\n" "\n"
" ecs\n" " ecs (default is disable)\n"
" - This option instructs s3fs to query the ECS container credential\n" " - This option instructs s3fs to query the ECS container credential\n"
" metadata address instead of the instance metadata address.\n" " metadata address instead of the instance metadata address.\n"
"\n" "\n"
@ -1227,16 +1237,16 @@ void show_help ()
" to an instance. If you specify this option without any argument, it\n" " to an instance. If you specify this option without any argument, it\n"
" is the same as that you have specified the \"auto\".\n" " is the same as that you have specified the \"auto\".\n"
"\n" "\n"
" ibm_iam_auth\n" " ibm_iam_auth (default is not using IBM IAM authentication)\n"
" - This option instructs s3fs to use IBM IAM authentication.\n" " - This option instructs s3fs to use IBM IAM authentication.\n"
" In this mode, the AWSAccessKey and AWSSecretKey will be used as\n" " In this mode, the AWSAccessKey and AWSSecretKey will be used as\n"
" IBM's Service-Instance-ID and APIKey, respectively.\n" " IBM's Service-Instance-ID and APIKey, respectively.\n"
"\n" "\n"
" ibm_iam_endpoint (default is https://iam.bluemix.net)\n" " ibm_iam_endpoint (default is https://iam.bluemix.net)\n"
" - sets the url to use for IBM IAM authentication.\n" " - sets the URL to use for IBM IAM authentication.\n"
"\n" "\n"
" use_xattr (default is not handling the extended attribute)\n" " use_xattr (default is not handling the extended attribute)\n"
" Enable to handle the extended attribute(xattrs).\n" " Enable to handle the extended attribute (xattrs).\n"
" If you set this option, you can use the extended attribute.\n" " If you set this option, you can use the extended attribute.\n"
" For example, encfs and ecryptfs need to support the extended attribute.\n" " For example, encfs and ecryptfs need to support the extended attribute.\n"
" Notice: if s3fs handles the extended attribute, s3fs can not work to\n" " Notice: if s3fs handles the extended attribute, s3fs can not work to\n"
@ -1251,18 +1261,18 @@ void show_help ()
"\n" "\n"
" nocopyapi (for other incomplete compatibility object storage)\n" " nocopyapi (for other incomplete compatibility object storage)\n"
" For a distributed object storage which is compatibility S3\n" " For a distributed object storage which is compatibility S3\n"
" API without PUT(copy api).\n" " API without PUT (copy api).\n"
" If you set this option, s3fs do not use PUT with \n" " If you set this option, s3fs do not use PUT with \n"
" \"x-amz-copy-source\"(copy api). Because traffic is increased\n" " \"x-amz-copy-source\" (copy api). Because traffic is increased\n"
" 2-3 times by this option, we do not recommend this.\n" " 2-3 times by this option, we do not recommend this.\n"
"\n" "\n"
" norenameapi (for other incomplete compatibility object storage)\n" " norenameapi (for other incomplete compatibility object storage)\n"
" For a distributed object storage which is compatibility S3\n" " For a distributed object storage which is compatibility S3\n"
" API without PUT(copy api).\n" " API without PUT (copy api).\n"
" This option is a subset of nocopyapi option. The nocopyapi\n" " This option is a subset of nocopyapi option. The nocopyapi\n"
" option does not use copy-api for all command(ex. chmod, chown,\n" " option does not use copy-api for all command (ex. chmod, chown,\n"
" touch, mv, etc), but this option does not use copy-api for\n" " touch, mv, etc), but this option does not use copy-api for\n"
" only rename command(ex. mv). If this option is specified with\n" " only rename command (ex. mv). If this option is specified with\n"
" nocopyapi, then s3fs ignores it.\n" " nocopyapi, then s3fs ignores it.\n"
"\n" "\n"
" use_path_request_style (use legacy API calling style)\n" " use_path_request_style (use legacy API calling style)\n"
@ -1276,17 +1286,7 @@ void show_help ()
" If this option is specified, s3fs suppresses the output of the\n" " If this option is specified, s3fs suppresses the output of the\n"
" User-Agent.\n" " User-Agent.\n"
"\n" "\n"
" dbglevel (default=\"crit\")\n" " cipher_suites\n"
" Set the debug message level. set value as crit(critical), err\n"
" (error), warn(warning), info(information) to debug level.\n"
" default debug level is critical. If s3fs run with \"-d\" option,\n"
" the debug level is set information. When s3fs catch the signal\n"
" SIGUSR2, the debug level is bumpup.\n"
"\n"
" curldbg - put curl debug message\n"
" Put the debug message from libcurl when this option is specified.\n"
"\n"
" cipher_suites - customize TLS cipher suite list\n"
" Customize the list of TLS cipher suites.\n" " Customize the list of TLS cipher suites.\n"
" Expects a colon separated list of cipher suite names.\n" " Expects a colon separated list of cipher suite names.\n"
" A list of available cipher suites, depending on your TLS engine,\n" " A list of available cipher suites, depending on your TLS engine,\n"
@ -1329,6 +1329,16 @@ void show_help ()
" Unicode set.\n" " Unicode set.\n"
" Useful on clients not using utf-8 as their file system encoding.\n" " Useful on clients not using utf-8 as their file system encoding.\n"
"\n" "\n"
" dbglevel (default=\"crit\")\n"
" Set the debug message level. set value as crit (critical), err\n"
" (error), warn (warning), info (information) to debug level.\n"
" default debug level is critical. If s3fs run with \"-d\" option,\n"
" the debug level is set information. When s3fs catch the signal\n"
" SIGUSR2, the debug level is bumpup.\n"
"\n"
" curldbg - put curl debug message\n"
" Put the debug message from libcurl when this option is specified.\n"
"\n"
"FUSE/mount Options:\n" "FUSE/mount Options:\n"
"\n" "\n"
" Most of the generic mount options described in 'man mount' are\n" " Most of the generic mount options described in 'man mount' are\n"
@ -1345,12 +1355,12 @@ void show_help ()
" -u, --incomplete-mpu-list\n" " -u, --incomplete-mpu-list\n"
" Lists multipart incomplete objects uploaded to the specified\n" " Lists multipart incomplete objects uploaded to the specified\n"
" bucket.\n" " bucket.\n"
" --incomplete-mpu-abort(=all or =<date format>)\n" " --incomplete-mpu-abort (=all or =<date format>)\n"
" Delete the multipart incomplete object uploaded to the specified\n" " Delete the multipart incomplete object uploaded to the specified\n"
" bucket.\n" " bucket.\n"
" If \"all\" is specified for this option, all multipart incomplete\n" " If \"all\" is specified for this option, all multipart incomplete\n"
" objects will be deleted. If you specify no argument as an option,\n" " objects will be deleted. If you specify no argument as an option,\n"
" objects older than 24 hours(24H) will be deleted(This is the\n" " objects older than 24 hours (24H) will be deleted (This is the\n"
" default value). You can specify an optional date format. It can\n" " default value). You can specify an optional date format. It can\n"
" be specified as year, month, day, hour, minute, second, and it is\n" " be specified as year, month, day, hour, minute, second, and it is\n"
" expressed as \"Y\", \"M\", \"D\", \"h\", \"m\", \"s\" respectively.\n" " expressed as \"Y\", \"M\", \"D\", \"h\", \"m\", \"s\" respectively.\n"
@ -1374,7 +1384,7 @@ void show_help ()
void show_version() void show_version()
{ {
printf( printf(
"Amazon Simple Storage Service File System V%s(commit:%s) with %s\n" "Amazon Simple Storage Service File System V%s (commit:%s) with %s\n"
"Copyright (C) 2010 Randy Rizun <rrizun@gmail.com>\n" "Copyright (C) 2010 Randy Rizun <rrizun@gmail.com>\n"
"License GPL2: GNU GPL version 2 <https://gnu.org/licenses/gpl.html>\n" "License GPL2: GNU GPL version 2 <https://gnu.org/licenses/gpl.html>\n"
"This is free software: you are free to change and redistribute it.\n" "This is free software: you are free to change and redistribute it.\n"