mirror of
https://github.com/s3fs-fuse/s3fs-fuse.git
synced 2024-11-12 23:46:30 +00:00
5575 lines
192 KiB
C++
5575 lines
192 KiB
C++
/*
|
|
* s3fs - FUSE-based file system backed by Amazon S3
|
|
*
|
|
* Copyright(C) 2007 Randy Rizun <rrizun@gmail.com>
|
|
*
|
|
* This program is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU General Public License
|
|
* as published by the Free Software Foundation; either version 2
|
|
* of the License, or (at your option) any later version.
|
|
*
|
|
* This program is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with this program; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.
|
|
*/
|
|
|
|
#include <cstdio>
|
|
#include <cstdlib>
|
|
#include <errno.h>
|
|
#include <unistd.h>
|
|
#include <dirent.h>
|
|
#include <sys/types.h>
|
|
#include <getopt.h>
|
|
|
|
#include "common.h"
|
|
#include "s3fs.h"
|
|
#include "s3fs_logger.h"
|
|
#include "metaheader.h"
|
|
#include "fdcache.h"
|
|
#include "fdcache_auto.h"
|
|
#include "fdcache_stat.h"
|
|
#include "curl.h"
|
|
#include "curl_multi.h"
|
|
#include "s3objlist.h"
|
|
#include "cache.h"
|
|
#include "mvnode.h"
|
|
#include "addhead.h"
|
|
#include "sighandlers.h"
|
|
#include "s3fs_xml.h"
|
|
#include "string_util.h"
|
|
#include "s3fs_auth.h"
|
|
#include "s3fs_cred.h"
|
|
#include "s3fs_help.h"
|
|
#include "s3fs_util.h"
|
|
#include "mpu_util.h"
|
|
#include "threadpoolman.h"
|
|
#include "autolock.h"
|
|
|
|
//-------------------------------------------------------------------
|
|
// Symbols
|
|
//-------------------------------------------------------------------
|
|
#if !defined(ENOATTR)
|
|
#define ENOATTR ENODATA
|
|
#endif
|
|
|
|
enum dirtype {
|
|
DIRTYPE_UNKNOWN = -1,
|
|
DIRTYPE_NEW = 0,
|
|
DIRTYPE_OLD = 1,
|
|
DIRTYPE_FOLDER = 2,
|
|
DIRTYPE_NOOBJ = 3,
|
|
};
|
|
|
|
//-------------------------------------------------------------------
|
|
// Static variables
|
|
//-------------------------------------------------------------------
|
|
static uid_t mp_uid = 0; // owner of mount point(only not specified uid opt)
|
|
static gid_t mp_gid = 0; // group of mount point(only not specified gid opt)
|
|
static mode_t mp_mode = 0; // mode of mount point
|
|
static mode_t mp_umask = 0; // umask for mount point
|
|
static bool is_mp_umask = false;// default does not set.
|
|
static bool has_mp_stat = false;// whether the stat information file for mount point exists
|
|
static std::string mountpoint;
|
|
static S3fsCred* ps3fscred = NULL; // using only in this file
|
|
static std::string mimetype_file;
|
|
static bool nocopyapi = false;
|
|
static bool norenameapi = false;
|
|
static bool nonempty = false;
|
|
static bool allow_other = false;
|
|
static uid_t s3fs_uid = 0;
|
|
static gid_t s3fs_gid = 0;
|
|
static mode_t s3fs_umask = 0;
|
|
static bool is_s3fs_uid = false;// default does not set.
|
|
static bool is_s3fs_gid = false;// default does not set.
|
|
static bool is_s3fs_umask = false;// default does not set.
|
|
static bool is_remove_cache = false;
|
|
static bool is_use_xattr = false;
|
|
static off_t multipart_threshold = 25 * 1024 * 1024;
|
|
static int64_t singlepart_copy_limit = 512 * 1024 * 1024;
|
|
static bool is_specified_endpoint = false;
|
|
static int s3fs_init_deferred_exit_status = 0;
|
|
static bool support_compat_dir = false;// default does not support compatibility directory type
|
|
static int max_keys_list_object = 1000;// default is 1000
|
|
static off_t max_dirty_data = 5LL * 1024LL * 1024LL * 1024LL;
|
|
static bool use_wtf8 = false;
|
|
static off_t fake_diskfree_size = -1; // default is not set(-1)
|
|
static int max_thread_count = 5; // default is 5
|
|
static bool update_parent_dir_stat= false; // default not updating parent directory stats
|
|
|
|
//-------------------------------------------------------------------
|
|
// Global functions : prototype
|
|
//-------------------------------------------------------------------
|
|
int put_headers(const char* path, headers_t& meta, bool is_copy, bool use_st_size = true); // [NOTE] global function because this is called from FdEntity class
|
|
|
|
//-------------------------------------------------------------------
|
|
// Static functions : prototype
|
|
//-------------------------------------------------------------------
|
|
static bool is_special_name_folder_object(const char* path);
|
|
static int chk_dir_object_type(const char* path, std::string& newpath, std::string& nowpath, std::string& nowcache, headers_t* pmeta = NULL, dirtype* pDirType = NULL);
|
|
static int remove_old_type_dir(const std::string& path, dirtype type);
|
|
static int get_object_attribute(const char* path, struct stat* pstbuf, headers_t* pmeta = NULL, bool overcheck = true, bool* pisforce = NULL, bool add_no_truncate_cache = false);
|
|
static int check_object_access(const char* path, int mask, struct stat* pstbuf);
|
|
static int check_object_owner(const char* path, struct stat* pstbuf);
|
|
static int check_parent_object_access(const char* path, int mask);
|
|
static int get_local_fent(AutoFdEntity& autoent, FdEntity **entity, const char* path, int flags = O_RDONLY, bool is_load = false);
|
|
static bool multi_head_callback(S3fsCurl* s3fscurl, void* param);
|
|
static S3fsCurl* multi_head_retry_callback(S3fsCurl* s3fscurl);
|
|
static int readdir_multi_head(const char* path, const S3ObjList& head, void* buf, fuse_fill_dir_t filler);
|
|
static int list_bucket(const char* path, S3ObjList& head, const char* delimiter, bool check_content_only = false);
|
|
static int directory_empty(const char* path);
|
|
static int rename_large_object(const char* from, const char* to);
|
|
static int create_file_object(const char* path, mode_t mode, uid_t uid, gid_t gid);
|
|
static int create_directory_object(const char* path, mode_t mode, const struct timespec& ts_atime, const struct timespec& ts_mtime, const struct timespec& ts_ctime, uid_t uid, gid_t gid, const char* pxattrvalue);
|
|
static int rename_object(const char* from, const char* to, bool update_ctime);
|
|
static int rename_object_nocopy(const char* from, const char* to, bool update_ctime);
|
|
static int clone_directory_object(const char* from, const char* to, bool update_ctime, const char* pxattrvalue);
|
|
static int rename_directory(const char* from, const char* to);
|
|
static int update_mctime_parent_directory(const char* _path);
|
|
static int remote_mountpath_exists(const char* path);
|
|
static bool get_meta_xattr_value(const char* path, std::string& rawvalue);
|
|
static bool get_parent_meta_xattr_value(const char* path, std::string& rawvalue);
|
|
static bool get_xattr_posix_key_value(const char* path, std::string& xattrvalue, bool default_key);
|
|
static bool build_inherited_xattr_value(const char* path, std::string& xattrvalue);
|
|
static void free_xattrs(xattrs_t& xattrs);
|
|
static bool parse_xattr_keyval(const std::string& xattrpair, std::string& key, PXATTRVAL& pval);
|
|
static size_t parse_xattrs(const std::string& strxattrs, xattrs_t& xattrs);
|
|
static std::string raw_build_xattrs(const xattrs_t& xattrs);
|
|
static std::string build_xattrs(const xattrs_t& xattrs);
|
|
static int s3fs_check_service();
|
|
static bool set_mountpoint_attribute(struct stat& mpst);
|
|
static int set_bucket(const char* arg);
|
|
static int my_fuse_opt_proc(void* data, const char* arg, int key, struct fuse_args* outargs);
|
|
|
|
//-------------------------------------------------------------------
|
|
// fuse interface functions
|
|
//-------------------------------------------------------------------
|
|
static int s3fs_getattr(const char* path, struct stat* stbuf);
|
|
static int s3fs_readlink(const char* path, char* buf, size_t size);
|
|
static int s3fs_mknod(const char* path, mode_t mode, dev_t rdev);
|
|
static int s3fs_mkdir(const char* path, mode_t mode);
|
|
static int s3fs_unlink(const char* path);
|
|
static int s3fs_rmdir(const char* path);
|
|
static int s3fs_symlink(const char* from, const char* to);
|
|
static int s3fs_rename(const char* from, const char* to);
|
|
static int s3fs_link(const char* from, const char* to);
|
|
static int s3fs_chmod(const char* path, mode_t mode);
|
|
static int s3fs_chmod_nocopy(const char* path, mode_t mode);
|
|
static int s3fs_chown(const char* path, uid_t uid, gid_t gid);
|
|
static int s3fs_chown_nocopy(const char* path, uid_t uid, gid_t gid);
|
|
static int s3fs_utimens(const char* path, const struct timespec ts[2]);
|
|
static int s3fs_utimens_nocopy(const char* path, const struct timespec ts[2]);
|
|
static int s3fs_truncate(const char* path, off_t size);
|
|
static int s3fs_create(const char* path, mode_t mode, struct fuse_file_info* fi);
|
|
static int s3fs_open(const char* path, struct fuse_file_info* fi);
|
|
static int s3fs_read(const char* path, char* buf, size_t size, off_t offset, struct fuse_file_info* fi);
|
|
static int s3fs_write(const char* path, const char* buf, size_t size, off_t offset, struct fuse_file_info* fi);
|
|
static int s3fs_statfs(const char* path, struct statvfs* stbuf);
|
|
static int s3fs_flush(const char* path, struct fuse_file_info* fi);
|
|
static int s3fs_fsync(const char* path, int datasync, struct fuse_file_info* fi);
|
|
static int s3fs_release(const char* path, struct fuse_file_info* fi);
|
|
static int s3fs_opendir(const char* path, struct fuse_file_info* fi);
|
|
static int s3fs_readdir(const char* path, void* buf, fuse_fill_dir_t filler, off_t offset, struct fuse_file_info* fi);
|
|
static int s3fs_access(const char* path, int mask);
|
|
static void* s3fs_init(struct fuse_conn_info* conn);
|
|
static void s3fs_destroy(void*);
|
|
#if defined(__APPLE__)
|
|
static int s3fs_setxattr(const char* path, const char* name, const char* value, size_t size, int flags, uint32_t position);
|
|
static int s3fs_getxattr(const char* path, const char* name, char* value, size_t size, uint32_t position);
|
|
#else
|
|
static int s3fs_setxattr(const char* path, const char* name, const char* value, size_t size, int flags);
|
|
static int s3fs_getxattr(const char* path, const char* name, char* value, size_t size);
|
|
#endif
|
|
static int s3fs_listxattr(const char* path, char* list, size_t size);
|
|
static int s3fs_removexattr(const char* path, const char* name);
|
|
|
|
//-------------------------------------------------------------------
|
|
// Functions
|
|
//-------------------------------------------------------------------
|
|
static bool IS_REPLACEDIR(dirtype type)
|
|
{
|
|
return DIRTYPE_OLD == type || DIRTYPE_FOLDER == type || DIRTYPE_NOOBJ == type;
|
|
}
|
|
|
|
static bool IS_RMTYPEDIR(dirtype type)
|
|
{
|
|
return DIRTYPE_OLD == type || DIRTYPE_FOLDER == type;
|
|
}
|
|
|
|
static bool IS_CREATE_MP_STAT(const char* path)
|
|
{
|
|
// [NOTE]
|
|
// "has_mp_stat" is set in get_object_attribute()
|
|
//
|
|
return (path && 0 == strcmp(path, "/") && !has_mp_stat);
|
|
}
|
|
|
|
static bool is_special_name_folder_object(const char* path)
|
|
{
|
|
if(!support_compat_dir){
|
|
// s3fs does not support compatibility directory type("_$folder$" etc) now,
|
|
// thus always returns false.
|
|
return false;
|
|
}
|
|
|
|
if(!path || '\0' == path[0]){
|
|
return false;
|
|
}
|
|
if(0 == strcmp(path, "/") && mount_prefix.empty()){
|
|
// the path is the mount point which is the bucket root
|
|
return false;
|
|
}
|
|
|
|
std::string strpath = path;
|
|
headers_t header;
|
|
|
|
if(std::string::npos == strpath.find("_$folder$", 0)){
|
|
if('/' == *strpath.rbegin()){
|
|
strpath.erase(strpath.length() - 1);
|
|
}
|
|
strpath += "_$folder$";
|
|
}
|
|
S3fsCurl s3fscurl;
|
|
if(0 != s3fscurl.HeadRequest(strpath.c_str(), header)){
|
|
return false;
|
|
}
|
|
header.clear();
|
|
S3FS_MALLOCTRIM(0);
|
|
return true;
|
|
}
|
|
|
|
// [Detail]
|
|
// This function is complicated for checking directory object type.
|
|
// Arguments is used for deleting cache/path, and remake directory object.
|
|
// Please see the codes which calls this function.
|
|
//
|
|
// path: target path
|
|
// newpath: should be object path for making/putting/getting after checking
|
|
// nowpath: now object name for deleting after checking
|
|
// nowcache: now cache path for deleting after checking
|
|
// pmeta: headers map
|
|
// pDirType: directory object type
|
|
//
|
|
static int chk_dir_object_type(const char* path, std::string& newpath, std::string& nowpath, std::string& nowcache, headers_t* pmeta, dirtype* pDirType)
|
|
{
|
|
dirtype TypeTmp = DIRTYPE_UNKNOWN;
|
|
int result = -1;
|
|
bool isforce = false;
|
|
dirtype* pType = pDirType ? pDirType : &TypeTmp;
|
|
|
|
// Normalize new path.
|
|
newpath = path;
|
|
if('/' != *newpath.rbegin()){
|
|
std::string::size_type Pos;
|
|
if(std::string::npos != (Pos = newpath.find("_$folder$", 0))){
|
|
newpath.erase(Pos);
|
|
}
|
|
newpath += "/";
|
|
}
|
|
|
|
// Always check "dir/" at first.
|
|
if(0 == (result = get_object_attribute(newpath.c_str(), NULL, pmeta, false, &isforce))){
|
|
// Found "dir/" cache --> Check for "_$folder$", "no dir object"
|
|
nowcache = newpath;
|
|
if(is_special_name_folder_object(newpath.c_str())){ // check support_compat_dir in this function
|
|
// "_$folder$" type.
|
|
(*pType) = DIRTYPE_FOLDER;
|
|
nowpath.erase(newpath.length() - 1);
|
|
nowpath += "_$folder$"; // cut and add
|
|
}else if(isforce){
|
|
// "no dir object" type.
|
|
(*pType) = DIRTYPE_NOOBJ;
|
|
nowpath = "";
|
|
}else{
|
|
nowpath = newpath;
|
|
if(!nowpath.empty() && '/' == *nowpath.rbegin()){
|
|
// "dir/" type
|
|
(*pType) = DIRTYPE_NEW;
|
|
}else{
|
|
// "dir" type
|
|
(*pType) = DIRTYPE_OLD;
|
|
}
|
|
}
|
|
}else if(support_compat_dir){
|
|
// Check "dir" when support_compat_dir is enabled
|
|
nowpath.erase(newpath.length() - 1);
|
|
if(0 == (result = get_object_attribute(nowpath.c_str(), NULL, pmeta, false, &isforce))){
|
|
// Found "dir" cache --> this case is only "dir" type.
|
|
// Because, if object is "_$folder$" or "no dir object", the cache is "dir/" type.
|
|
// (But "no dir object" is checked here.)
|
|
nowcache = nowpath;
|
|
if(isforce){
|
|
(*pType) = DIRTYPE_NOOBJ;
|
|
nowpath = "";
|
|
}else{
|
|
(*pType) = DIRTYPE_OLD;
|
|
}
|
|
}else{
|
|
// Not found cache --> check for "_$folder$" and "no dir object".
|
|
// (come here is that support_compat_dir is enabled)
|
|
nowcache = ""; // This case is no cache.
|
|
nowpath += "_$folder$";
|
|
if(is_special_name_folder_object(nowpath.c_str())){
|
|
// "_$folder$" type.
|
|
(*pType) = DIRTYPE_FOLDER;
|
|
result = 0; // result is OK.
|
|
}else if(-ENOTEMPTY == directory_empty(newpath.c_str())){
|
|
// "no dir object" type.
|
|
(*pType) = DIRTYPE_NOOBJ;
|
|
nowpath = ""; // now path.
|
|
result = 0; // result is OK.
|
|
}else{
|
|
// Error: Unknown type.
|
|
(*pType) = DIRTYPE_UNKNOWN;
|
|
newpath = "";
|
|
nowpath = "";
|
|
}
|
|
}
|
|
}
|
|
return result;
|
|
}
|
|
|
|
static int remove_old_type_dir(const std::string& path, dirtype type)
|
|
{
|
|
if(IS_RMTYPEDIR(type)){
|
|
S3fsCurl s3fscurl;
|
|
int result = s3fscurl.DeleteRequest(path.c_str());
|
|
if(0 != result && -ENOENT != result){
|
|
return result;
|
|
}
|
|
// succeed removing or not found the directory
|
|
}else{
|
|
// nothing to do
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
//
|
|
// Get object attributes with stat cache.
|
|
// This function is base for s3fs_getattr().
|
|
//
|
|
// [NOTICE]
|
|
// Checking order is changed following list because of reducing the number of the requests.
|
|
// 1) "dir"
|
|
// 2) "dir/"
|
|
// 3) "dir_$folder$"
|
|
//
|
|
// Special two case of the mount point directory:
|
|
// [Case 1] the mount point is the root of the bucket:
|
|
// 1) "/"
|
|
//
|
|
// [Case 2] the mount point is a directory path(ex. foo) below the bucket:
|
|
// 1) "foo"
|
|
// 2) "foo/"
|
|
// 3) "foo_$folder$"
|
|
//
|
|
static int get_object_attribute(const char* path, struct stat* pstbuf, headers_t* pmeta, bool overcheck, bool* pisforce, bool add_no_truncate_cache)
|
|
{
|
|
int result = -1;
|
|
struct stat tmpstbuf;
|
|
struct stat* pstat = pstbuf ? pstbuf : &tmpstbuf;
|
|
headers_t tmpHead;
|
|
headers_t* pheader = pmeta ? pmeta : &tmpHead;
|
|
std::string strpath;
|
|
S3fsCurl s3fscurl;
|
|
bool forcedir = false;
|
|
bool is_mountpoint = false; // path is the mount point
|
|
bool is_bucket_mountpoint = false; // path is the mount point which is the bucket root
|
|
std::string::size_type Pos;
|
|
|
|
S3FS_PRN_DBG("[path=%s]", path);
|
|
|
|
if(!path || '\0' == path[0]){
|
|
return -ENOENT;
|
|
}
|
|
|
|
memset(pstat, 0, sizeof(struct stat));
|
|
|
|
// check mount point
|
|
if(0 == strcmp(path, "/") || 0 == strcmp(path, ".")){
|
|
is_mountpoint = true;
|
|
if(mount_prefix.empty()){
|
|
is_bucket_mountpoint = true;
|
|
}
|
|
// default stat for mount point if the directory stat file is not existed.
|
|
pstat->st_mode = mp_mode;
|
|
pstat->st_uid = is_s3fs_uid ? s3fs_uid : mp_uid;
|
|
pstat->st_gid = is_s3fs_gid ? s3fs_gid : mp_gid;
|
|
}
|
|
|
|
// Check cache.
|
|
pisforce = (NULL != pisforce ? pisforce : &forcedir);
|
|
(*pisforce) = false;
|
|
strpath = path;
|
|
if(support_compat_dir && overcheck && std::string::npos != (Pos = strpath.find("_$folder$", 0))){
|
|
strpath.erase(Pos);
|
|
strpath += "/";
|
|
}
|
|
// [NOTE]
|
|
// For mount points("/"), the Stat cache key name is "/".
|
|
//
|
|
if(StatCache::getStatCacheData()->GetStat(strpath, pstat, pheader, overcheck, pisforce)){
|
|
if(is_mountpoint){
|
|
// if mount point, we need to set this.
|
|
pstat->st_nlink = 1; // see fuse faq
|
|
}
|
|
return 0;
|
|
}
|
|
if(StatCache::getStatCacheData()->IsNoObjectCache(strpath)){
|
|
// there is the path in the cache for no object, it is no object.
|
|
return -ENOENT;
|
|
}
|
|
|
|
// set query(head request) path
|
|
if(is_bucket_mountpoint){
|
|
// [NOTE]
|
|
// This is a special process for mount point
|
|
// The path is "/" for mount points.
|
|
// If the bucket mounted at a mount point, we try to find "/" object under
|
|
// the bucket for mount point's stat.
|
|
// In this case, we will send the request "HEAD // HTTP /1.1" to S3 server.
|
|
//
|
|
// If the directory under the bucket is mounted, it will be sent
|
|
// "HEAD /<directories ...>/ HTTP/1.1", so we do not need to change path at
|
|
// here.
|
|
//
|
|
strpath = "//"; // strpath is "//"
|
|
}else{
|
|
strpath = path;
|
|
}
|
|
result = s3fscurl.HeadRequest(strpath.c_str(), (*pheader));
|
|
s3fscurl.DestroyCurlHandle();
|
|
|
|
// if not found target path object, do over checking
|
|
if(-EPERM == result){
|
|
// [NOTE]
|
|
// In case of a permission error, it exists in directory
|
|
// file list but inaccessible. So there is a problem that
|
|
// it will send a HEAD request every time, because it is
|
|
// not registered in the Stats cache.
|
|
// Therefore, even if the file has a permission error, it
|
|
// should be registered in the Stats cache. However, if
|
|
// the response without modifying is registered in the
|
|
// cache, the file permission will be 0644(umask dependent)
|
|
// because the meta header does not exist.
|
|
// Thus, set the mode of 0000 here in the meta header so
|
|
// that s3fs can print a permission error when the file
|
|
// is actually accessed.
|
|
// It is better not to set meta header other than mode,
|
|
// so do not do it.
|
|
//
|
|
(*pheader)["x-amz-meta-mode"] = str(0);
|
|
|
|
}else if(0 != result){
|
|
if(overcheck && !is_bucket_mountpoint){
|
|
// when support_compat_dir is disabled, strpath maybe have "_$folder$".
|
|
if('/' != *strpath.rbegin() && std::string::npos == strpath.find("_$folder$", 0)){
|
|
// now path is "object", do check "object/" for over checking
|
|
strpath += "/";
|
|
result = s3fscurl.HeadRequest(strpath.c_str(), (*pheader));
|
|
s3fscurl.DestroyCurlHandle();
|
|
}
|
|
if(support_compat_dir && 0 != result){
|
|
// now path is "object/", do check "object_$folder$" for over checking
|
|
strpath.erase(strpath.length() - 1);
|
|
strpath += "_$folder$";
|
|
result = s3fscurl.HeadRequest(strpath.c_str(), (*pheader));
|
|
s3fscurl.DestroyCurlHandle();
|
|
|
|
if(0 != result){
|
|
// cut "_$folder$" for over checking "no dir object" after here
|
|
if(std::string::npos != (Pos = strpath.find("_$folder$", 0))){
|
|
strpath.erase(Pos);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
if(0 != result && std::string::npos == strpath.find("_$folder$", 0)){
|
|
// now path is "object" or "object/", do check "no dir object" which is not object but has only children.
|
|
//
|
|
// [NOTE]
|
|
// If the path is mount point and there is no Stat information file for it, we need this process.
|
|
//
|
|
if('/' == *strpath.rbegin()){
|
|
strpath.erase(strpath.length() - 1);
|
|
}
|
|
if(-ENOTEMPTY == directory_empty(strpath.c_str())){
|
|
// found "no dir object".
|
|
strpath += "/";
|
|
*pisforce = true;
|
|
result = 0;
|
|
}
|
|
}
|
|
}else{
|
|
if('/' != *strpath.rbegin() && std::string::npos == strpath.find("_$folder$", 0) && is_need_check_obj_detail(*pheader)){
|
|
// check a case of that "object" does not have attribute and "object" is possible to be directory.
|
|
if(-ENOTEMPTY == directory_empty(strpath.c_str())){
|
|
// found "no dir object".
|
|
strpath += "/";
|
|
*pisforce = true;
|
|
result = 0;
|
|
}
|
|
}
|
|
}
|
|
|
|
// set headers for mount point from default stat
|
|
if(is_mountpoint){
|
|
if(0 != result){
|
|
has_mp_stat = false;
|
|
|
|
// [NOTE]
|
|
// If mount point and no stat information file, create header
|
|
// information from the default stat.
|
|
//
|
|
(*pheader)["Content-Type"] = S3fsCurl::LookupMimeType(strpath);
|
|
(*pheader)["x-amz-meta-uid"] = str(pstat->st_uid);
|
|
(*pheader)["x-amz-meta-gid"] = str(pstat->st_gid);
|
|
(*pheader)["x-amz-meta-mode"] = str(pstat->st_mode);
|
|
(*pheader)["x-amz-meta-atime"] = str(pstat->st_atime);
|
|
(*pheader)["x-amz-meta-ctime"] = str(pstat->st_ctime);
|
|
(*pheader)["x-amz-meta-mtime"] = str(pstat->st_mtime);
|
|
|
|
result = 0;
|
|
}else{
|
|
has_mp_stat = true;
|
|
}
|
|
}
|
|
|
|
// [NOTE]
|
|
// If the file is listed but not allowed access, put it in
|
|
// the positive cache instead of the negative cache.
|
|
//
|
|
// When mount points, the following error does not occur.
|
|
//
|
|
if(0 != result && -EPERM != result){
|
|
// finally, "path" object did not find. Add no object cache.
|
|
strpath = path; // reset original
|
|
StatCache::getStatCacheData()->AddNoObjectCache(strpath);
|
|
return result;
|
|
}
|
|
|
|
// set cache key
|
|
if(is_bucket_mountpoint){
|
|
strpath = "/";
|
|
}else if(std::string::npos != (Pos = strpath.find("_$folder$", 0))){
|
|
// if path has "_$folder$", need to cut it.
|
|
strpath.erase(Pos);
|
|
strpath += "/";
|
|
}
|
|
|
|
// Set into cache
|
|
//
|
|
// [NOTE]
|
|
// When add_no_truncate_cache is true, the stats is always cached.
|
|
// This cached stats is only removed by DelStat().
|
|
// This is necessary for the case to access the attribute of opened file.
|
|
// (ex. getxattr() is called while writing to the opened file.)
|
|
//
|
|
if(add_no_truncate_cache || 0 != StatCache::getStatCacheData()->GetCacheSize()){
|
|
// add into stat cache
|
|
if(!StatCache::getStatCacheData()->AddStat(strpath, (*pheader), forcedir, add_no_truncate_cache)){
|
|
S3FS_PRN_ERR("failed adding stat cache [path=%s]", strpath.c_str());
|
|
return -ENOENT;
|
|
}
|
|
if(!StatCache::getStatCacheData()->GetStat(strpath, pstat, pheader, overcheck, pisforce)){
|
|
// There is not in cache.(why?) -> retry to convert.
|
|
if(!convert_header_to_stat(strpath.c_str(), (*pheader), pstat, forcedir)){
|
|
S3FS_PRN_ERR("failed convert headers to stat[path=%s]", strpath.c_str());
|
|
return -ENOENT;
|
|
}
|
|
}
|
|
}else{
|
|
// cache size is Zero -> only convert.
|
|
if(!convert_header_to_stat(strpath.c_str(), (*pheader), pstat, forcedir)){
|
|
S3FS_PRN_ERR("failed convert headers to stat[path=%s]", strpath.c_str());
|
|
return -ENOENT;
|
|
}
|
|
}
|
|
|
|
if(is_mountpoint){
|
|
// if mount point, we need to set this.
|
|
pstat->st_nlink = 1; // see fuse faq
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
//
|
|
// Check the object uid and gid for write/read/execute.
|
|
// The param "mask" is as same as access() function.
|
|
// If there is not a target file, this function returns -ENOENT.
|
|
// If the target file can be accessed, the result always is 0.
|
|
//
|
|
// path: the target object path
|
|
// mask: bit field(F_OK, R_OK, W_OK, X_OK) like access().
|
|
// stat: NULL or the pointer of struct stat.
|
|
//
|
|
static int check_object_access(const char* path, int mask, struct stat* pstbuf)
|
|
{
|
|
int result;
|
|
struct stat st;
|
|
struct stat* pst = (pstbuf ? pstbuf : &st);
|
|
struct fuse_context* pcxt;
|
|
|
|
S3FS_PRN_DBG("[path=%s]", path);
|
|
|
|
if(NULL == (pcxt = fuse_get_context())){
|
|
return -EIO;
|
|
}
|
|
S3FS_PRN_DBG("[pid=%u,uid=%u,gid=%u]", (unsigned int)(pcxt->pid), (unsigned int)(pcxt->uid), (unsigned int)(pcxt->gid));
|
|
|
|
if(0 != (result = get_object_attribute(path, pst))){
|
|
// If there is not the target file(object), result is -ENOENT.
|
|
return result;
|
|
}
|
|
if(0 == pcxt->uid){
|
|
// root is allowed all accessing.
|
|
return 0;
|
|
}
|
|
if(is_s3fs_uid && s3fs_uid == pcxt->uid){
|
|
// "uid" user is allowed all accessing.
|
|
return 0;
|
|
}
|
|
if(F_OK == mask){
|
|
// if there is a file, always return allowed.
|
|
return 0;
|
|
}
|
|
|
|
// for "uid", "gid" option
|
|
uid_t obj_uid = (is_s3fs_uid ? s3fs_uid : pst->st_uid);
|
|
gid_t obj_gid = (is_s3fs_gid ? s3fs_gid : pst->st_gid);
|
|
|
|
// compare file mode and uid/gid + mask.
|
|
mode_t mode;
|
|
mode_t base_mask = S_IRWXO;
|
|
if(is_s3fs_umask){
|
|
// If umask is set, all object attributes set ~umask.
|
|
mode = ((S_IRWXU | S_IRWXG | S_IRWXO) & ~s3fs_umask);
|
|
}else{
|
|
mode = pst->st_mode;
|
|
}
|
|
if(pcxt->uid == obj_uid){
|
|
base_mask |= S_IRWXU;
|
|
}
|
|
if(pcxt->gid == obj_gid){
|
|
base_mask |= S_IRWXG;
|
|
}
|
|
if(1 == is_uid_include_group(pcxt->uid, obj_gid)){
|
|
base_mask |= S_IRWXG;
|
|
}
|
|
mode &= base_mask;
|
|
|
|
if(X_OK == (mask & X_OK)){
|
|
if(0 == (mode & (S_IXUSR | S_IXGRP | S_IXOTH))){
|
|
return -EACCES;
|
|
}
|
|
}
|
|
if(W_OK == (mask & W_OK)){
|
|
if(0 == (mode & (S_IWUSR | S_IWGRP | S_IWOTH))){
|
|
return -EACCES;
|
|
}
|
|
}
|
|
if(R_OK == (mask & R_OK)){
|
|
if(0 == (mode & (S_IRUSR | S_IRGRP | S_IROTH))){
|
|
return -EACCES;
|
|
}
|
|
}
|
|
if(0 == mode){
|
|
return -EACCES;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int check_object_owner(const char* path, struct stat* pstbuf)
|
|
{
|
|
int result;
|
|
struct stat st;
|
|
struct stat* pst = (pstbuf ? pstbuf : &st);
|
|
struct fuse_context* pcxt;
|
|
|
|
S3FS_PRN_DBG("[path=%s]", path);
|
|
|
|
if(NULL == (pcxt = fuse_get_context())){
|
|
return -EIO;
|
|
}
|
|
if(0 != (result = get_object_attribute(path, pst))){
|
|
// If there is not the target file(object), result is -ENOENT.
|
|
return result;
|
|
}
|
|
// check owner
|
|
if(0 == pcxt->uid){
|
|
// root is allowed all accessing.
|
|
return 0;
|
|
}
|
|
if(is_s3fs_uid && s3fs_uid == pcxt->uid){
|
|
// "uid" user is allowed all accessing.
|
|
return 0;
|
|
}
|
|
if(pcxt->uid == pst->st_uid){
|
|
return 0;
|
|
}
|
|
return -EPERM;
|
|
}
|
|
|
|
//
|
|
// Check accessing the parent directories of the object by uid and gid.
|
|
//
|
|
static int check_parent_object_access(const char* path, int mask)
|
|
{
|
|
std::string parent;
|
|
int result;
|
|
|
|
S3FS_PRN_DBG("[path=%s]", path);
|
|
|
|
if(0 == strcmp(path, "/") || 0 == strcmp(path, ".")){
|
|
// path is mount point.
|
|
return 0;
|
|
}
|
|
if(X_OK == (mask & X_OK)){
|
|
for(parent = mydirname(path); !parent.empty(); parent = mydirname(parent)){
|
|
if(parent == "."){
|
|
parent = "/";
|
|
}
|
|
if(0 != (result = check_object_access(parent.c_str(), X_OK, NULL))){
|
|
return result;
|
|
}
|
|
if(parent == "/" || parent == "."){
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
mask = (mask & ~X_OK);
|
|
if(0 != mask){
|
|
parent = mydirname(path);
|
|
if(parent == "."){
|
|
parent = "/";
|
|
}
|
|
if(0 != (result = check_object_access(parent.c_str(), mask, NULL))){
|
|
return result;
|
|
}
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
//
|
|
// ssevalue is MD5 for SSE-C type, or KMS id for SSE-KMS
|
|
//
|
|
bool get_object_sse_type(const char* path, sse_type_t& ssetype, std::string& ssevalue)
|
|
{
|
|
if(!path){
|
|
return false;
|
|
}
|
|
|
|
headers_t meta;
|
|
if(0 != get_object_attribute(path, NULL, &meta)){
|
|
S3FS_PRN_ERR("Failed to get object(%s) headers", path);
|
|
return false;
|
|
}
|
|
|
|
ssetype = sse_type_t::SSE_DISABLE;
|
|
ssevalue.erase();
|
|
for(headers_t::iterator iter = meta.begin(); iter != meta.end(); ++iter){
|
|
std::string key = (*iter).first;
|
|
if(0 == strcasecmp(key.c_str(), "x-amz-server-side-encryption") && 0 == strcasecmp((*iter).second.c_str(), "AES256")){
|
|
ssetype = sse_type_t::SSE_S3;
|
|
}else if(0 == strcasecmp(key.c_str(), "x-amz-server-side-encryption-aws-kms-key-id")){
|
|
ssetype = sse_type_t::SSE_KMS;
|
|
ssevalue = (*iter).second;
|
|
}else if(0 == strcasecmp(key.c_str(), "x-amz-server-side-encryption-customer-key-md5")){
|
|
ssetype = sse_type_t::SSE_C;
|
|
ssevalue = (*iter).second;
|
|
}
|
|
}
|
|
return true;
|
|
}
|
|
|
|
static int get_local_fent(AutoFdEntity& autoent, FdEntity **entity, const char* path, int flags, bool is_load)
|
|
{
|
|
int result;
|
|
struct stat stobj;
|
|
FdEntity* ent;
|
|
headers_t meta;
|
|
|
|
S3FS_PRN_INFO2("[path=%s]", path);
|
|
|
|
if(0 != (result = get_object_attribute(path, &stobj, &meta))){
|
|
return result;
|
|
}
|
|
|
|
// open
|
|
struct timespec st_mctime;
|
|
if(!S_ISREG(stobj.st_mode) && !S_ISLNK(stobj.st_mode)){
|
|
st_mctime = S3FS_OMIT_TS;
|
|
}else{
|
|
set_stat_to_timespec(stobj, ST_TYPE_MTIME, st_mctime);
|
|
}
|
|
bool force_tmpfile = S_ISREG(stobj.st_mode) ? false : true;
|
|
|
|
if(NULL == (ent = autoent.Open(path, &meta, stobj.st_size, st_mctime, flags, force_tmpfile, true, false, AutoLock::NONE))){
|
|
S3FS_PRN_ERR("Could not open file. errno(%d)", errno);
|
|
return -EIO;
|
|
}
|
|
// load
|
|
if(is_load && !ent->LoadAll(autoent.GetPseudoFd(), &meta)){
|
|
S3FS_PRN_ERR("Could not load file. errno(%d)", errno);
|
|
autoent.Close();
|
|
return -EIO;
|
|
}
|
|
*entity = ent;
|
|
return 0;
|
|
}
|
|
|
|
//
|
|
// create or update s3 meta
|
|
// @return fuse return code
|
|
//
|
|
int put_headers(const char* path, headers_t& meta, bool is_copy, bool use_st_size)
|
|
{
|
|
int result;
|
|
S3fsCurl s3fscurl(true);
|
|
off_t size;
|
|
std::string strpath;
|
|
|
|
S3FS_PRN_INFO2("[path=%s]", path);
|
|
|
|
if(0 == strcmp(path, "/") && mount_prefix.empty()){
|
|
strpath = "//"; // for the mount point that is bucket root, change "/" to "//".
|
|
}else{
|
|
strpath = path;
|
|
}
|
|
|
|
// files larger than 5GB must be modified via the multipart interface
|
|
// call use_st_size as false when the file does not exist(ex. rename object)
|
|
if(use_st_size && '/' != *strpath.rbegin()){ // directory object("dir/") is always 0(Content-Length = 0)
|
|
struct stat buf;
|
|
if(0 != (result = get_object_attribute(path, &buf))){
|
|
return result;
|
|
}
|
|
size = buf.st_size;
|
|
}else{
|
|
size = get_size(meta);
|
|
}
|
|
|
|
if(!nocopyapi && !nomultipart && size >= multipart_threshold){
|
|
if(0 != (result = s3fscurl.MultipartHeadRequest(strpath.c_str(), size, meta, is_copy))){
|
|
return result;
|
|
}
|
|
}else{
|
|
if(0 != (result = s3fscurl.PutHeadRequest(strpath.c_str(), meta, is_copy))){
|
|
return result;
|
|
}
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int s3fs_getattr(const char* _path, struct stat* stbuf)
|
|
{
|
|
WTF8_ENCODE(path)
|
|
int result;
|
|
|
|
S3FS_PRN_INFO("[path=%s]", path);
|
|
|
|
// check parent directory attribute.
|
|
if(0 != (result = check_parent_object_access(path, X_OK))){
|
|
return result;
|
|
}
|
|
if(0 != (result = check_object_access(path, F_OK, stbuf))){
|
|
return result;
|
|
}
|
|
// If has already opened fd, the st_size should be instead.
|
|
// (See: Issue 241)
|
|
if(stbuf){
|
|
AutoFdEntity autoent;
|
|
FdEntity* ent;
|
|
if(NULL != (ent = autoent.OpenExistFdEntity(path))){
|
|
struct stat tmpstbuf;
|
|
if(ent->GetStats(tmpstbuf)){
|
|
stbuf->st_size = tmpstbuf.st_size;
|
|
}
|
|
}
|
|
stbuf->st_blksize = 4096;
|
|
stbuf->st_blocks = get_blocks(stbuf->st_size);
|
|
|
|
S3FS_PRN_DBG("[path=%s] uid=%u, gid=%u, mode=%04o", path, (unsigned int)(stbuf->st_uid), (unsigned int)(stbuf->st_gid), stbuf->st_mode);
|
|
}
|
|
S3FS_MALLOCTRIM(0);
|
|
|
|
return result;
|
|
}
|
|
|
|
static int s3fs_readlink(const char* _path, char* buf, size_t size)
|
|
{
|
|
if(!_path || !buf || 0 == size){
|
|
return 0;
|
|
}
|
|
WTF8_ENCODE(path)
|
|
std::string strValue;
|
|
|
|
// check symbolic link cache
|
|
if(!StatCache::getStatCacheData()->GetSymlink(std::string(path), strValue)){
|
|
// not found in cache, then open the path
|
|
{ // scope for AutoFdEntity
|
|
AutoFdEntity autoent;
|
|
FdEntity* ent;
|
|
int result;
|
|
if(0 != (result = get_local_fent(autoent, &ent, path, O_RDONLY))){
|
|
S3FS_PRN_ERR("could not get fent(file=%s)", path);
|
|
return result;
|
|
}
|
|
// Get size
|
|
off_t readsize;
|
|
if(!ent->GetSize(readsize)){
|
|
S3FS_PRN_ERR("could not get file size(file=%s)", path);
|
|
return -EIO;
|
|
}
|
|
if(static_cast<off_t>(size) <= readsize){
|
|
readsize = size - 1;
|
|
}
|
|
// Read
|
|
ssize_t ressize;
|
|
if(0 > (ressize = ent->Read(autoent.GetPseudoFd(), buf, 0, readsize))){
|
|
S3FS_PRN_ERR("could not read file(file=%s, ressize=%zd)", path, ressize);
|
|
return static_cast<int>(ressize);
|
|
}
|
|
buf[ressize] = '\0';
|
|
}
|
|
|
|
// check buf if it has space words.
|
|
strValue = trim(std::string(buf));
|
|
|
|
// decode wtf8. This will always be shorter
|
|
if(use_wtf8){
|
|
strValue = s3fs_wtf8_decode(strValue);
|
|
}
|
|
|
|
// add symbolic link cache
|
|
if(!StatCache::getStatCacheData()->AddSymlink(std::string(path), strValue)){
|
|
S3FS_PRN_ERR("failed to add symbolic link cache for %s", path);
|
|
}
|
|
}
|
|
// copy result
|
|
strncpy(buf, strValue.c_str(), size - 1);
|
|
buf[size - 1] = '\0';
|
|
|
|
S3FS_MALLOCTRIM(0);
|
|
|
|
return 0;
|
|
}
|
|
|
|
// common function for creation of a plain object
|
|
static int create_file_object(const char* path, mode_t mode, uid_t uid, gid_t gid)
|
|
{
|
|
S3FS_PRN_INFO2("[path=%s][mode=%04o]", path, mode);
|
|
|
|
std::string strnow = s3fs_str_realtime();
|
|
headers_t meta;
|
|
meta["Content-Type"] = S3fsCurl::LookupMimeType(std::string(path));
|
|
meta["x-amz-meta-uid"] = str(uid);
|
|
meta["x-amz-meta-gid"] = str(gid);
|
|
meta["x-amz-meta-mode"] = str(mode);
|
|
meta["x-amz-meta-atime"] = strnow;
|
|
meta["x-amz-meta-ctime"] = strnow;
|
|
meta["x-amz-meta-mtime"] = strnow;
|
|
|
|
S3fsCurl s3fscurl(true);
|
|
return s3fscurl.PutRequest(path, meta, -1); // fd=-1 means for creating zero byte object.
|
|
}
|
|
|
|
static int s3fs_mknod(const char *_path, mode_t mode, dev_t rdev)
|
|
{
|
|
WTF8_ENCODE(path)
|
|
int result;
|
|
struct fuse_context* pcxt;
|
|
|
|
S3FS_PRN_INFO("[path=%s][mode=%04o][dev=%llu]", path, mode, (unsigned long long)rdev);
|
|
|
|
if(NULL == (pcxt = fuse_get_context())){
|
|
return -EIO;
|
|
}
|
|
|
|
if(0 != (result = create_file_object(path, mode, pcxt->uid, pcxt->gid))){
|
|
S3FS_PRN_ERR("could not create object for special file(result=%d)", result);
|
|
return result;
|
|
}
|
|
StatCache::getStatCacheData()->DelStat(path);
|
|
|
|
// update parent directory timestamp
|
|
int update_result;
|
|
if(0 != (update_result = update_mctime_parent_directory(path))){
|
|
S3FS_PRN_ERR("succeed to mknod the file(%s), but could not update timestamp of its parent directory(result=%d).", path, update_result);
|
|
}
|
|
|
|
S3FS_MALLOCTRIM(0);
|
|
|
|
return result;
|
|
}
|
|
|
|
static int s3fs_create(const char* _path, mode_t mode, struct fuse_file_info* fi)
|
|
{
|
|
WTF8_ENCODE(path)
|
|
int result;
|
|
struct fuse_context* pcxt;
|
|
|
|
S3FS_PRN_INFO("[path=%s][mode=%04o][flags=0x%x]", path, mode, fi->flags);
|
|
|
|
if(NULL == (pcxt = fuse_get_context())){
|
|
return -EIO;
|
|
}
|
|
|
|
// check parent directory attribute.
|
|
if(0 != (result = check_parent_object_access(path, X_OK))){
|
|
return result;
|
|
}
|
|
result = check_object_access(path, W_OK, NULL);
|
|
if(-ENOENT == result){
|
|
if(0 != (result = check_parent_object_access(path, W_OK))){
|
|
return result;
|
|
}
|
|
}else if(0 != result){
|
|
return result;
|
|
}
|
|
|
|
std::string strnow = s3fs_str_realtime();
|
|
headers_t meta;
|
|
meta["Content-Length"] = "0";
|
|
meta["x-amz-meta-uid"] = str(pcxt->uid);
|
|
meta["x-amz-meta-gid"] = str(pcxt->gid);
|
|
meta["x-amz-meta-mode"] = str(mode);
|
|
meta["x-amz-meta-atime"] = strnow;
|
|
meta["x-amz-meta-mtime"] = strnow;
|
|
meta["x-amz-meta-ctime"] = strnow;
|
|
|
|
std::string xattrvalue;
|
|
if(build_inherited_xattr_value(path, xattrvalue)){
|
|
S3FS_PRN_DBG("Set xattrs = %s", urlDecode(xattrvalue).c_str());
|
|
meta["x-amz-meta-xattr"] = xattrvalue;
|
|
}
|
|
|
|
// [NOTE] set no_truncate flag
|
|
// At this point, the file has not been created(uploaded) and
|
|
// the data is only present in the Stats cache.
|
|
// The Stats cache should not be deleted automatically by
|
|
// timeout. If this stats is deleted, s3fs will try to get it
|
|
// from the server with a Head request and will get an
|
|
// unexpected error because the result object does not exist.
|
|
//
|
|
if(!StatCache::getStatCacheData()->AddStat(path, meta, false, true)){
|
|
return -EIO;
|
|
}
|
|
|
|
AutoFdEntity autoent;
|
|
FdEntity* ent;
|
|
if(NULL == (ent = autoent.Open(path, &meta, 0, S3FS_OMIT_TS, fi->flags, false, true, false, AutoLock::NONE))){
|
|
StatCache::getStatCacheData()->DelStat(path);
|
|
return -EIO;
|
|
}
|
|
ent->MarkDirtyNewFile();
|
|
fi->fh = autoent.Detach(); // KEEP fdentity open;
|
|
|
|
S3FS_MALLOCTRIM(0);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int create_directory_object(const char* path, mode_t mode, const struct timespec& ts_atime, const struct timespec& ts_mtime, const struct timespec& ts_ctime, uid_t uid, gid_t gid, const char* pxattrvalue)
|
|
{
|
|
S3FS_PRN_INFO1("[path=%s][mode=%04o][atime=%s][mtime=%s][ctime=%s][uid=%u][gid=%u]", path, mode, str(ts_atime).c_str(), str(ts_mtime).c_str(), str(ts_ctime).c_str(), (unsigned int)uid, (unsigned int)gid);
|
|
|
|
if(!path || '\0' == path[0]){
|
|
return -EINVAL;
|
|
}
|
|
std::string tpath = path;
|
|
if('/' != *tpath.rbegin()){
|
|
tpath += "/";
|
|
}else if("/" == tpath && mount_prefix.empty()){
|
|
tpath = "//"; // for the mount point that is bucket root, change "/" to "//".
|
|
}
|
|
|
|
headers_t meta;
|
|
meta["x-amz-meta-uid"] = str(uid);
|
|
meta["x-amz-meta-gid"] = str(gid);
|
|
meta["x-amz-meta-mode"] = str(mode);
|
|
meta["x-amz-meta-atime"] = str(ts_atime);
|
|
meta["x-amz-meta-mtime"] = str(ts_mtime);
|
|
meta["x-amz-meta-ctime"] = str(ts_ctime);
|
|
|
|
if(pxattrvalue){
|
|
S3FS_PRN_DBG("Set xattrs = %s", urlDecode(std::string(pxattrvalue)).c_str());
|
|
meta["x-amz-meta-xattr"] = std::string(pxattrvalue);
|
|
}
|
|
|
|
S3fsCurl s3fscurl;
|
|
return s3fscurl.PutRequest(tpath.c_str(), meta, -1); // fd=-1 means for creating zero byte object.
|
|
}
|
|
|
|
static int s3fs_mkdir(const char* _path, mode_t mode)
|
|
{
|
|
WTF8_ENCODE(path)
|
|
int result;
|
|
struct fuse_context* pcxt;
|
|
|
|
S3FS_PRN_INFO("[path=%s][mode=%04o]", path, mode);
|
|
|
|
if(NULL == (pcxt = fuse_get_context())){
|
|
return -EIO;
|
|
}
|
|
|
|
// check parent directory attribute.
|
|
if(0 != (result = check_parent_object_access(path, W_OK | X_OK))){
|
|
return result;
|
|
}
|
|
if(-ENOENT != (result = check_object_access(path, F_OK, NULL))){
|
|
if(0 == result){
|
|
result = -EEXIST;
|
|
}
|
|
return result;
|
|
}
|
|
|
|
std::string xattrvalue;
|
|
const char* pxattrvalue;
|
|
if(get_parent_meta_xattr_value(path, xattrvalue)){
|
|
pxattrvalue = xattrvalue.c_str();
|
|
}else{
|
|
pxattrvalue = NULL;
|
|
}
|
|
|
|
struct timespec now;
|
|
s3fs_realtime(now);
|
|
result = create_directory_object(path, mode, now, now, now, pcxt->uid, pcxt->gid, pxattrvalue);
|
|
|
|
StatCache::getStatCacheData()->DelStat(path);
|
|
|
|
// update parent directory timestamp
|
|
int update_result;
|
|
if(0 != (update_result = update_mctime_parent_directory(path))){
|
|
S3FS_PRN_ERR("succeed to create the directory(%s), but could not update timestamp of its parent directory(result=%d).", path, update_result);
|
|
}
|
|
|
|
S3FS_MALLOCTRIM(0);
|
|
|
|
return result;
|
|
}
|
|
|
|
static int s3fs_unlink(const char* _path)
|
|
{
|
|
WTF8_ENCODE(path)
|
|
int result;
|
|
|
|
S3FS_PRN_INFO("[path=%s]", path);
|
|
|
|
if(0 != (result = check_parent_object_access(path, W_OK | X_OK))){
|
|
return result;
|
|
}
|
|
S3fsCurl s3fscurl;
|
|
result = s3fscurl.DeleteRequest(path);
|
|
StatCache::getStatCacheData()->DelStat(path);
|
|
StatCache::getStatCacheData()->DelSymlink(path);
|
|
FdManager::DeleteCacheFile(path);
|
|
|
|
// update parent directory timestamp
|
|
int update_result;
|
|
if(0 != (update_result = update_mctime_parent_directory(path))){
|
|
S3FS_PRN_ERR("succeed to remove the file(%s), but could not update timestamp of its parent directory(result=%d).", path, update_result);
|
|
}
|
|
|
|
S3FS_MALLOCTRIM(0);
|
|
|
|
return result;
|
|
}
|
|
|
|
static int directory_empty(const char* path)
|
|
{
|
|
int result;
|
|
S3ObjList head;
|
|
|
|
if((result = list_bucket(path, head, "/", true)) != 0){
|
|
S3FS_PRN_ERR("list_bucket returns error.");
|
|
return result;
|
|
}
|
|
if(!head.IsEmpty()){
|
|
return -ENOTEMPTY;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int s3fs_rmdir(const char* _path)
|
|
{
|
|
WTF8_ENCODE(path)
|
|
int result;
|
|
std::string strpath;
|
|
struct stat stbuf;
|
|
|
|
S3FS_PRN_INFO("[path=%s]", path);
|
|
|
|
if(0 != (result = check_parent_object_access(path, W_OK | X_OK))){
|
|
return result;
|
|
}
|
|
|
|
// directory must be empty
|
|
if(directory_empty(path) != 0){
|
|
return -ENOTEMPTY;
|
|
}
|
|
|
|
strpath = path;
|
|
if('/' != *strpath.rbegin()){
|
|
strpath += "/";
|
|
}
|
|
S3fsCurl s3fscurl;
|
|
result = s3fscurl.DeleteRequest(strpath.c_str());
|
|
s3fscurl.DestroyCurlHandle();
|
|
StatCache::getStatCacheData()->DelStat(strpath.c_str());
|
|
|
|
// double check for old version(before 1.63)
|
|
// The old version makes "dir" object, newer version makes "dir/".
|
|
// A case, there is only "dir", the first removing object is "dir/".
|
|
// Then "dir/" is not exists, but curl_delete returns 0.
|
|
// So need to check "dir" and should be removed it.
|
|
if('/' == *strpath.rbegin()){
|
|
strpath.erase(strpath.length() - 1);
|
|
}
|
|
if(0 == get_object_attribute(strpath.c_str(), &stbuf, NULL, false)){
|
|
if(S_ISDIR(stbuf.st_mode)){
|
|
// Found "dir" object.
|
|
result = s3fscurl.DeleteRequest(strpath.c_str());
|
|
s3fscurl.DestroyCurlHandle();
|
|
StatCache::getStatCacheData()->DelStat(strpath.c_str());
|
|
}
|
|
}
|
|
// If there is no "dir" and "dir/" object(this case is made by s3cmd/s3sync),
|
|
// the cache key is "dir/". So we get error only once(delete "dir/").
|
|
|
|
// check for "_$folder$" object.
|
|
// This processing is necessary for other S3 clients compatibility.
|
|
if(is_special_name_folder_object(strpath.c_str())){
|
|
strpath += "_$folder$";
|
|
result = s3fscurl.DeleteRequest(strpath.c_str());
|
|
}
|
|
|
|
// update parent directory timestamp
|
|
int update_result;
|
|
if(0 != (update_result = update_mctime_parent_directory(path))){
|
|
S3FS_PRN_ERR("succeed to remove the directory(%s), but could not update timestamp of its parent directory(result=%d).", path, update_result);
|
|
}
|
|
|
|
S3FS_MALLOCTRIM(0);
|
|
|
|
return result;
|
|
}
|
|
|
|
static int s3fs_symlink(const char* _from, const char* _to)
|
|
{
|
|
WTF8_ENCODE(from)
|
|
WTF8_ENCODE(to)
|
|
int result;
|
|
struct fuse_context* pcxt;
|
|
|
|
S3FS_PRN_INFO("[from=%s][to=%s]", from, to);
|
|
|
|
if(NULL == (pcxt = fuse_get_context())){
|
|
return -EIO;
|
|
}
|
|
if(0 != (result = check_parent_object_access(to, W_OK | X_OK))){
|
|
return result;
|
|
}
|
|
if(-ENOENT != (result = check_object_access(to, F_OK, NULL))){
|
|
if(0 == result){
|
|
result = -EEXIST;
|
|
}
|
|
return result;
|
|
}
|
|
|
|
std::string strnow = s3fs_str_realtime();
|
|
headers_t headers;
|
|
headers["Content-Type"] = std::string("application/octet-stream"); // Static
|
|
headers["x-amz-meta-mode"] = str(S_IFLNK | S_IRWXU | S_IRWXG | S_IRWXO);
|
|
headers["x-amz-meta-atime"] = strnow;
|
|
headers["x-amz-meta-ctime"] = strnow;
|
|
headers["x-amz-meta-mtime"] = strnow;
|
|
headers["x-amz-meta-uid"] = str(pcxt->uid);
|
|
headers["x-amz-meta-gid"] = str(pcxt->gid);
|
|
|
|
// [NOTE]
|
|
// Symbolic links do not set xattrs.
|
|
|
|
// open tmpfile
|
|
std::string strFrom;
|
|
{ // scope for AutoFdEntity
|
|
AutoFdEntity autoent;
|
|
FdEntity* ent;
|
|
if(NULL == (ent = autoent.Open(to, &headers, 0, S3FS_OMIT_TS, O_RDWR, true, true, false, AutoLock::NONE))){
|
|
S3FS_PRN_ERR("could not open tmpfile(errno=%d)", errno);
|
|
return -errno;
|
|
}
|
|
// write(without space words)
|
|
strFrom = trim(std::string(from));
|
|
ssize_t from_size = static_cast<ssize_t>(strFrom.length());
|
|
ssize_t ressize;
|
|
if(from_size != (ressize = ent->Write(autoent.GetPseudoFd(), strFrom.c_str(), 0, from_size))){
|
|
if(ressize < 0){
|
|
S3FS_PRN_ERR("could not write tmpfile(errno=%d)", static_cast<int>(ressize));
|
|
return static_cast<int>(ressize);
|
|
}else{
|
|
S3FS_PRN_ERR("could not write tmpfile %zd byte(errno=%d)", ressize, errno);
|
|
return (0 == errno ? -EIO : -errno);
|
|
}
|
|
}
|
|
// upload
|
|
if(0 != (result = ent->Flush(autoent.GetPseudoFd(), AutoLock::NONE, true))){
|
|
S3FS_PRN_WARN("could not upload tmpfile(result=%d)", result);
|
|
}
|
|
}
|
|
|
|
StatCache::getStatCacheData()->DelStat(to);
|
|
if(!StatCache::getStatCacheData()->AddSymlink(std::string(to), strFrom)){
|
|
S3FS_PRN_ERR("failed to add symbolic link cache for %s", to);
|
|
}
|
|
|
|
// update parent directory timestamp
|
|
int update_result;
|
|
if(0 != (update_result = update_mctime_parent_directory(to))){
|
|
S3FS_PRN_ERR("succeed to create symbolic link(%s), but could not update timestamp of its parent directory(result=%d).", to, update_result);
|
|
}
|
|
|
|
S3FS_MALLOCTRIM(0);
|
|
|
|
return result;
|
|
}
|
|
|
|
static int rename_object(const char* from, const char* to, bool update_ctime)
|
|
{
|
|
int result;
|
|
headers_t meta;
|
|
struct stat buf;
|
|
|
|
S3FS_PRN_INFO1("[from=%s][to=%s]", from , to);
|
|
|
|
if(0 != (result = check_parent_object_access(to, W_OK | X_OK))){
|
|
// not permit writing "to" object parent dir.
|
|
return result;
|
|
}
|
|
if(0 != (result = check_parent_object_access(from, W_OK | X_OK))){
|
|
// not permit removing "from" object parent dir.
|
|
return result;
|
|
}
|
|
if(0 != (result = get_object_attribute(from, &buf, &meta))){
|
|
return result;
|
|
}
|
|
|
|
std::string strSourcePath = (mount_prefix.empty() && 0 == strcmp("/", from)) ? "//" : from;
|
|
|
|
if(update_ctime){
|
|
meta["x-amz-meta-ctime"] = s3fs_str_realtime();
|
|
}
|
|
meta["x-amz-copy-source"] = urlEncode(service_path + S3fsCred::GetBucket() + get_realpath(strSourcePath.c_str()));
|
|
meta["Content-Type"] = S3fsCurl::LookupMimeType(std::string(to));
|
|
meta["x-amz-metadata-directive"] = "REPLACE";
|
|
|
|
std::string xattrvalue;
|
|
if(get_meta_xattr_value(from, xattrvalue)){
|
|
S3FS_PRN_DBG("Set xattrs = %s", urlDecode(xattrvalue).c_str());
|
|
meta["x-amz-meta-xattr"] = xattrvalue;
|
|
}
|
|
|
|
// [NOTE]
|
|
// If it has a cache, open it first and leave it open until rename.
|
|
// The cache is renamed after put_header, because it must be open
|
|
// at the time of renaming.
|
|
{
|
|
// update time
|
|
AutoFdEntity autoent;
|
|
FdEntity* ent;
|
|
if(NULL == (ent = autoent.OpenExistFdEntity(from))){
|
|
// no opened fd
|
|
|
|
// get mtime/ctime/atime from meta
|
|
struct timespec mtime = get_mtime(meta);
|
|
struct timespec ctime = get_ctime(meta);
|
|
struct timespec atime = get_atime(meta);
|
|
if(mtime.tv_sec < 0){
|
|
mtime.tv_sec = 0L;
|
|
mtime.tv_nsec = 0L;
|
|
}
|
|
if(ctime.tv_sec < 0){
|
|
ctime.tv_sec = 0L;
|
|
ctime.tv_nsec = 0L;
|
|
}
|
|
if(atime.tv_sec < 0){
|
|
atime.tv_sec = 0L;
|
|
atime.tv_nsec = 0L;
|
|
}
|
|
|
|
if(FdManager::IsCacheDir()){
|
|
// create cache file if be needed
|
|
//
|
|
// [NOTE]
|
|
// Do not specify "S3FS_OMIT_TS" for mctime parameter.
|
|
// This is because if the cache file does not exist, the pagelist for it
|
|
// will be recreated, but the entire file area indicated by this pagelist
|
|
// will be in the "modified" state.
|
|
// This does not affect the rename process, but the cache information in
|
|
// the "modified" state remains, making it impossible to read the file correctly.
|
|
//
|
|
ent = autoent.Open(from, &meta, buf.st_size, mtime, O_RDONLY, false, true, false, AutoLock::NONE);
|
|
}
|
|
if(ent){
|
|
ent->SetMCtime(mtime, ctime);
|
|
ent->SetAtime(atime);
|
|
}
|
|
}
|
|
|
|
// copy
|
|
if(0 != (result = put_headers(to, meta, true, /* use_st_size= */ false))){
|
|
return result;
|
|
}
|
|
|
|
// rename
|
|
FdManager::get()->Rename(from, to);
|
|
}
|
|
|
|
// Remove file
|
|
result = s3fs_unlink(from);
|
|
|
|
StatCache::getStatCacheData()->DelStat(to);
|
|
|
|
return result;
|
|
}
|
|
|
|
static int rename_object_nocopy(const char* from, const char* to, bool update_ctime)
|
|
{
|
|
int result;
|
|
|
|
S3FS_PRN_INFO1("[from=%s][to=%s]", from , to);
|
|
|
|
if(0 != (result = check_parent_object_access(to, W_OK | X_OK))){
|
|
// not permit writing "to" object parent dir.
|
|
return result;
|
|
}
|
|
if(0 != (result = check_parent_object_access(from, W_OK | X_OK))){
|
|
// not permit removing "from" object parent dir.
|
|
return result;
|
|
}
|
|
|
|
// open & load
|
|
{ // scope for AutoFdEntity
|
|
AutoFdEntity autoent;
|
|
FdEntity* ent;
|
|
if(0 != (result = get_local_fent(autoent, &ent, from, O_RDWR, true))){
|
|
S3FS_PRN_ERR("could not open and read file(%s)", from);
|
|
return result;
|
|
}
|
|
|
|
// Set header
|
|
if(!ent->SetContentType(to)){
|
|
S3FS_PRN_ERR("could not set content-type for %s", to);
|
|
return -EIO;
|
|
}
|
|
|
|
// update ctime
|
|
if(update_ctime){
|
|
struct timespec ts;
|
|
s3fs_realtime(ts);
|
|
ent->SetCtime(ts);
|
|
}
|
|
|
|
// upload
|
|
if(0 != (result = ent->RowFlush(autoent.GetPseudoFd(), to, AutoLock::NONE, true))){
|
|
S3FS_PRN_ERR("could not upload file(%s): result=%d", to, result);
|
|
return result;
|
|
}
|
|
FdManager::get()->Rename(from, to);
|
|
}
|
|
|
|
// Remove file
|
|
result = s3fs_unlink(from);
|
|
|
|
// Stats
|
|
StatCache::getStatCacheData()->DelStat(to);
|
|
|
|
return result;
|
|
}
|
|
|
|
static int rename_large_object(const char* from, const char* to)
|
|
{
|
|
int result;
|
|
struct stat buf;
|
|
headers_t meta;
|
|
|
|
S3FS_PRN_INFO1("[from=%s][to=%s]", from , to);
|
|
|
|
if(0 != (result = check_parent_object_access(to, W_OK | X_OK))){
|
|
// not permit writing "to" object parent dir.
|
|
return result;
|
|
}
|
|
if(0 != (result = check_parent_object_access(from, W_OK | X_OK))){
|
|
// not permit removing "from" object parent dir.
|
|
return result;
|
|
}
|
|
if(0 != (result = get_object_attribute(from, &buf, &meta, false))){
|
|
return result;
|
|
}
|
|
|
|
S3fsCurl s3fscurl(true);
|
|
if(0 != (result = s3fscurl.MultipartRenameRequest(from, to, meta, buf.st_size))){
|
|
return result;
|
|
}
|
|
s3fscurl.DestroyCurlHandle();
|
|
|
|
// Remove file
|
|
result = s3fs_unlink(from);
|
|
|
|
StatCache::getStatCacheData()->DelStat(to);
|
|
FdManager::DeleteCacheFile(to);
|
|
|
|
return result;
|
|
}
|
|
|
|
static int clone_directory_object(const char* from, const char* to, bool update_ctime, const char* pxattrvalue)
|
|
{
|
|
int result = -1;
|
|
struct stat stbuf;
|
|
|
|
S3FS_PRN_INFO1("[from=%s][to=%s]", from, to);
|
|
|
|
// get target's attributes
|
|
if(0 != (result = get_object_attribute(from, &stbuf))){
|
|
return result;
|
|
}
|
|
|
|
struct timespec ts_atime;
|
|
struct timespec ts_mtime;
|
|
struct timespec ts_ctime;
|
|
set_stat_to_timespec(stbuf, ST_TYPE_ATIME, ts_atime);
|
|
set_stat_to_timespec(stbuf, ST_TYPE_MTIME, ts_mtime);
|
|
if(update_ctime){
|
|
s3fs_realtime(ts_ctime);
|
|
}else{
|
|
set_stat_to_timespec(stbuf, ST_TYPE_CTIME, ts_ctime);
|
|
}
|
|
result = create_directory_object(to, stbuf.st_mode, ts_atime, ts_mtime, ts_ctime, stbuf.st_uid, stbuf.st_gid, pxattrvalue);
|
|
|
|
StatCache::getStatCacheData()->DelStat(to);
|
|
|
|
return result;
|
|
}
|
|
|
|
static int rename_directory(const char* from, const char* to)
|
|
{
|
|
S3ObjList head;
|
|
s3obj_list_t headlist;
|
|
std::string strfrom = from ? from : ""; // from is without "/".
|
|
std::string strto = to ? to : ""; // to is without "/" too.
|
|
std::string basepath = strfrom + "/";
|
|
std::string newpath; // should be from name(not used)
|
|
std::string nowcache; // now cache path(not used)
|
|
dirtype DirType;
|
|
bool normdir;
|
|
MVNODE* mn_head = NULL;
|
|
MVNODE* mn_tail = NULL;
|
|
MVNODE* mn_cur;
|
|
struct stat stbuf;
|
|
int result;
|
|
bool is_dir;
|
|
|
|
S3FS_PRN_INFO1("[from=%s][to=%s]", from, to);
|
|
|
|
//
|
|
// Initiate and Add base directory into MVNODE struct.
|
|
//
|
|
strto += "/";
|
|
if(0 == chk_dir_object_type(from, newpath, strfrom, nowcache, NULL, &DirType) && DIRTYPE_UNKNOWN != DirType){
|
|
if(DIRTYPE_NOOBJ != DirType){
|
|
normdir = false;
|
|
}else{
|
|
normdir = true;
|
|
strfrom = from; // from directory is not removed, but from directory attr is needed.
|
|
}
|
|
if(NULL == (add_mvnode(&mn_head, &mn_tail, strfrom.c_str(), strto.c_str(), true, normdir))){
|
|
return -ENOMEM;
|
|
}
|
|
}else{
|
|
// Something wrong about "from" directory.
|
|
}
|
|
|
|
//
|
|
// get a list of all the objects
|
|
//
|
|
// No delimiter is specified, the result(head) is all object keys.
|
|
// (CommonPrefixes is empty, but all object is listed in Key.)
|
|
if(0 != (result = list_bucket(basepath.c_str(), head, NULL))){
|
|
S3FS_PRN_ERR("list_bucket returns error.");
|
|
return result;
|
|
}
|
|
head.GetNameList(headlist); // get name without "/".
|
|
S3ObjList::MakeHierarchizedList(headlist, false); // add hierarchized dir.
|
|
|
|
s3obj_list_t::const_iterator liter;
|
|
for(liter = headlist.begin(); headlist.end() != liter; ++liter){
|
|
// make "from" and "to" object name.
|
|
std::string from_name = basepath + (*liter);
|
|
std::string to_name = strto + (*liter);
|
|
std::string etag = head.GetETag((*liter).c_str());
|
|
|
|
// Check subdirectory.
|
|
StatCache::getStatCacheData()->HasStat(from_name, etag.c_str()); // Check ETag
|
|
if(0 != get_object_attribute(from_name.c_str(), &stbuf, NULL)){
|
|
S3FS_PRN_WARN("failed to get %s object attribute.", from_name.c_str());
|
|
continue;
|
|
}
|
|
if(S_ISDIR(stbuf.st_mode)){
|
|
is_dir = true;
|
|
if(0 != chk_dir_object_type(from_name.c_str(), newpath, from_name, nowcache, NULL, &DirType) || DIRTYPE_UNKNOWN == DirType){
|
|
S3FS_PRN_WARN("failed to get %s%s object directory type.", basepath.c_str(), (*liter).c_str());
|
|
continue;
|
|
}
|
|
if(DIRTYPE_NOOBJ != DirType){
|
|
normdir = false;
|
|
}else{
|
|
normdir = true;
|
|
from_name = basepath + (*liter); // from directory is not removed, but from directory attr is needed.
|
|
}
|
|
}else{
|
|
is_dir = false;
|
|
normdir = false;
|
|
}
|
|
|
|
// push this one onto the stack
|
|
if(NULL == add_mvnode(&mn_head, &mn_tail, from_name.c_str(), to_name.c_str(), is_dir, normdir)){
|
|
return -ENOMEM;
|
|
}
|
|
}
|
|
|
|
//
|
|
// rename
|
|
//
|
|
// rename directory objects.
|
|
for(mn_cur = mn_head; mn_cur; mn_cur = mn_cur->next){
|
|
if(mn_cur->is_dir && mn_cur->old_path && '\0' != mn_cur->old_path[0]){
|
|
std::string xattrvalue;
|
|
const char* pxattrvalue;
|
|
if(get_meta_xattr_value(mn_cur->old_path, xattrvalue)){
|
|
pxattrvalue = xattrvalue.c_str();
|
|
}else{
|
|
pxattrvalue = NULL;
|
|
}
|
|
|
|
// [NOTE]
|
|
// The ctime is updated only for the top (from) directory.
|
|
// Other than that, it will not be updated.
|
|
//
|
|
if(0 != (result = clone_directory_object(mn_cur->old_path, mn_cur->new_path, (strfrom == mn_cur->old_path), pxattrvalue))){
|
|
S3FS_PRN_ERR("clone_directory_object returned an error(%d)", result);
|
|
free_mvnodes(mn_head);
|
|
return result;
|
|
}
|
|
}
|
|
}
|
|
|
|
// iterate over the list - copy the files with rename_object
|
|
// does a safe copy - copies first and then deletes old
|
|
for(mn_cur = mn_head; mn_cur; mn_cur = mn_cur->next){
|
|
if(!mn_cur->is_dir){
|
|
if(!nocopyapi && !norenameapi){
|
|
result = rename_object(mn_cur->old_path, mn_cur->new_path, false); // keep ctime
|
|
}else{
|
|
result = rename_object_nocopy(mn_cur->old_path, mn_cur->new_path, false); // keep ctime
|
|
}
|
|
if(0 != result){
|
|
S3FS_PRN_ERR("rename_object returned an error(%d)", result);
|
|
free_mvnodes(mn_head);
|
|
return result;
|
|
}
|
|
}
|
|
}
|
|
|
|
// Iterate over old the directories, bottoms up and remove
|
|
for(mn_cur = mn_tail; mn_cur; mn_cur = mn_cur->prev){
|
|
if(mn_cur->is_dir && mn_cur->old_path && '\0' != mn_cur->old_path[0]){
|
|
if(!(mn_cur->is_normdir)){
|
|
if(0 != (result = s3fs_rmdir(mn_cur->old_path))){
|
|
S3FS_PRN_ERR("s3fs_rmdir returned an error(%d)", result);
|
|
free_mvnodes(mn_head);
|
|
return result;
|
|
}
|
|
}else{
|
|
// cache clear.
|
|
StatCache::getStatCacheData()->DelStat(mn_cur->old_path);
|
|
}
|
|
}
|
|
}
|
|
free_mvnodes(mn_head);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int s3fs_rename(const char* _from, const char* _to)
|
|
{
|
|
WTF8_ENCODE(from)
|
|
WTF8_ENCODE(to)
|
|
struct stat buf;
|
|
int result;
|
|
|
|
S3FS_PRN_INFO("[from=%s][to=%s]", from, to);
|
|
|
|
if(0 != (result = check_parent_object_access(to, W_OK | X_OK))){
|
|
// not permit writing "to" object parent dir.
|
|
return result;
|
|
}
|
|
if(0 != (result = check_parent_object_access(from, W_OK | X_OK))){
|
|
// not permit removing "from" object parent dir.
|
|
return result;
|
|
}
|
|
if(0 != (result = get_object_attribute(from, &buf, NULL))){
|
|
return result;
|
|
}
|
|
if(0 != (result = directory_empty(to))){
|
|
return result;
|
|
}
|
|
|
|
// flush pending writes if file is open
|
|
{ // scope for AutoFdEntity
|
|
AutoFdEntity autoent;
|
|
FdEntity* ent;
|
|
if(NULL != (ent = autoent.OpenExistFdEntity(from, O_RDWR))){
|
|
if(0 != (result = ent->Flush(autoent.GetPseudoFd(), AutoLock::NONE, true))){
|
|
S3FS_PRN_ERR("could not upload file(%s): result=%d", to, result);
|
|
return result;
|
|
}
|
|
StatCache::getStatCacheData()->DelStat(from);
|
|
}
|
|
}
|
|
|
|
// files larger than 5GB must be modified via the multipart interface
|
|
if(S_ISDIR(buf.st_mode)){
|
|
result = rename_directory(from, to);
|
|
}else if(!nomultipart && buf.st_size >= singlepart_copy_limit){
|
|
result = rename_large_object(from, to);
|
|
}else{
|
|
if(!nocopyapi && !norenameapi){
|
|
result = rename_object(from, to, true); // update ctime
|
|
}else{
|
|
result = rename_object_nocopy(from, to, true); // update ctime
|
|
}
|
|
}
|
|
|
|
// update parent directory timestamp
|
|
//
|
|
// [NOTE]
|
|
// already updated timestamp for original path in above functions.
|
|
//
|
|
int update_result;
|
|
if(0 != (update_result = update_mctime_parent_directory(to))){
|
|
S3FS_PRN_ERR("succeed to create the file/directory(%s), but could not update timestamp of its parent directory(result=%d).", to, update_result);
|
|
}
|
|
|
|
S3FS_MALLOCTRIM(0);
|
|
|
|
return result;
|
|
}
|
|
|
|
static int s3fs_link(const char* _from, const char* _to)
|
|
{
|
|
WTF8_ENCODE(from)
|
|
WTF8_ENCODE(to)
|
|
S3FS_PRN_INFO("[from=%s][to=%s]", from, to);
|
|
return -ENOTSUP;
|
|
}
|
|
|
|
static int s3fs_chmod(const char* _path, mode_t mode)
|
|
{
|
|
WTF8_ENCODE(path)
|
|
int result;
|
|
std::string strpath;
|
|
std::string newpath;
|
|
std::string nowcache;
|
|
headers_t meta;
|
|
struct stat stbuf;
|
|
dirtype nDirType = DIRTYPE_UNKNOWN;
|
|
|
|
S3FS_PRN_INFO("[path=%s][mode=%04o]", path, mode);
|
|
|
|
if(0 != (result = check_parent_object_access(path, X_OK))){
|
|
return result;
|
|
}
|
|
if(0 != (result = check_object_owner(path, &stbuf))){
|
|
return result;
|
|
}
|
|
|
|
if(S_ISDIR(stbuf.st_mode)){
|
|
result = chk_dir_object_type(path, newpath, strpath, nowcache, &meta, &nDirType);
|
|
}else{
|
|
strpath = path;
|
|
nowcache = strpath;
|
|
result = get_object_attribute(strpath.c_str(), NULL, &meta);
|
|
}
|
|
if(0 != result){
|
|
return result;
|
|
}
|
|
|
|
if(S_ISDIR(stbuf.st_mode) && (IS_REPLACEDIR(nDirType) || IS_CREATE_MP_STAT(path))){
|
|
std::string xattrvalue;
|
|
const char* pxattrvalue;
|
|
if(get_meta_xattr_value(path, xattrvalue)){
|
|
pxattrvalue = xattrvalue.c_str();
|
|
}else{
|
|
pxattrvalue = NULL;
|
|
}
|
|
if(IS_REPLACEDIR(nDirType)){
|
|
// Should rebuild directory object(except new type)
|
|
// Need to remove old dir("dir" etc) and make new dir("dir/")
|
|
|
|
// At first, remove directory old object
|
|
if(0 != (result = remove_old_type_dir(strpath, nDirType))){
|
|
return result;
|
|
}
|
|
}
|
|
StatCache::getStatCacheData()->DelStat(nowcache);
|
|
|
|
// Make new directory object("dir/")
|
|
struct timespec ts_atime;
|
|
struct timespec ts_mtime;
|
|
struct timespec ts_ctime;
|
|
set_stat_to_timespec(stbuf, ST_TYPE_ATIME, ts_atime);
|
|
set_stat_to_timespec(stbuf, ST_TYPE_MTIME, ts_mtime);
|
|
s3fs_realtime(ts_ctime);
|
|
|
|
if(0 != (result = create_directory_object(newpath.c_str(), mode, ts_atime, ts_mtime, ts_ctime, stbuf.st_uid, stbuf.st_gid, pxattrvalue))){
|
|
return result;
|
|
}
|
|
}else{
|
|
// normal object or directory object of newer version
|
|
std::string strSourcePath = (mount_prefix.empty() && "/" == strpath) ? "//" : strpath;
|
|
headers_t updatemeta;
|
|
updatemeta["x-amz-meta-ctime"] = s3fs_str_realtime();
|
|
updatemeta["x-amz-meta-mode"] = str(mode);
|
|
updatemeta["x-amz-copy-source"] = urlEncode(service_path + S3fsCred::GetBucket() + get_realpath(strSourcePath.c_str()));
|
|
updatemeta["x-amz-metadata-directive"] = "REPLACE";
|
|
|
|
// check opened file handle.
|
|
//
|
|
// If the file starts uploading by multipart when the disk capacity is insufficient,
|
|
// we need to put these header after finishing upload.
|
|
// Or if the file is only open, we must update to FdEntity's internal meta.
|
|
//
|
|
AutoFdEntity autoent;
|
|
FdEntity* ent;
|
|
bool need_put_header = true;
|
|
if(NULL != (ent = autoent.OpenExistFdEntity(path))){
|
|
if(ent->MergeOrgMeta(updatemeta)){
|
|
// meta is changed, but now uploading.
|
|
// then the meta is pending and accumulated to be put after the upload is complete.
|
|
S3FS_PRN_INFO("meta pending until upload is complete");
|
|
need_put_header = false;
|
|
|
|
// If there is data in the Stats cache, update the Stats cache.
|
|
StatCache::getStatCacheData()->UpdateMetaStats(strpath, updatemeta);
|
|
}
|
|
}
|
|
if(need_put_header){
|
|
// not found opened file.
|
|
merge_headers(meta, updatemeta, true);
|
|
|
|
// upload meta directly.
|
|
if(0 != (result = put_headers(strpath.c_str(), meta, true))){
|
|
return result;
|
|
}
|
|
StatCache::getStatCacheData()->DelStat(nowcache);
|
|
}
|
|
}
|
|
S3FS_MALLOCTRIM(0);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int s3fs_chmod_nocopy(const char* _path, mode_t mode)
|
|
{
|
|
WTF8_ENCODE(path)
|
|
int result;
|
|
std::string strpath;
|
|
std::string newpath;
|
|
std::string nowcache;
|
|
struct stat stbuf;
|
|
dirtype nDirType = DIRTYPE_UNKNOWN;
|
|
|
|
S3FS_PRN_INFO1("[path=%s][mode=%04o]", path, mode);
|
|
|
|
if(0 != (result = check_parent_object_access(path, X_OK))){
|
|
return result;
|
|
}
|
|
if(0 != (result = check_object_owner(path, &stbuf))){
|
|
return result;
|
|
}
|
|
|
|
// Get attributes
|
|
if(S_ISDIR(stbuf.st_mode)){
|
|
result = chk_dir_object_type(path, newpath, strpath, nowcache, NULL, &nDirType);
|
|
}else{
|
|
strpath = path;
|
|
nowcache = strpath;
|
|
result = get_object_attribute(strpath.c_str(), NULL, NULL);
|
|
}
|
|
if(0 != result){
|
|
return result;
|
|
}
|
|
|
|
if(S_ISDIR(stbuf.st_mode)){
|
|
std::string xattrvalue;
|
|
const char* pxattrvalue;
|
|
if(get_meta_xattr_value(path, xattrvalue)){
|
|
pxattrvalue = xattrvalue.c_str();
|
|
}else{
|
|
pxattrvalue = NULL;
|
|
}
|
|
|
|
if(IS_REPLACEDIR(nDirType)){
|
|
// Should rebuild all directory object
|
|
// Need to remove old dir("dir" etc) and make new dir("dir/")
|
|
|
|
// At first, remove directory old object
|
|
if(0 != (result = remove_old_type_dir(strpath, nDirType))){
|
|
return result;
|
|
}
|
|
}
|
|
StatCache::getStatCacheData()->DelStat(nowcache);
|
|
|
|
// Make new directory object("dir/")
|
|
struct timespec ts_atime;
|
|
struct timespec ts_mtime;
|
|
struct timespec ts_ctime;
|
|
set_stat_to_timespec(stbuf, ST_TYPE_ATIME, ts_atime);
|
|
set_stat_to_timespec(stbuf, ST_TYPE_MTIME, ts_mtime);
|
|
s3fs_realtime(ts_ctime);
|
|
|
|
if(0 != (result = create_directory_object(newpath.c_str(), mode, ts_atime, ts_mtime, ts_ctime, stbuf.st_uid, stbuf.st_gid, pxattrvalue))){
|
|
return result;
|
|
}
|
|
}else{
|
|
// normal object or directory object of newer version
|
|
|
|
// open & load
|
|
AutoFdEntity autoent;
|
|
FdEntity* ent;
|
|
if(0 != (result = get_local_fent(autoent, &ent, strpath.c_str(), O_RDWR, true))){
|
|
S3FS_PRN_ERR("could not open and read file(%s)", strpath.c_str());
|
|
return result;
|
|
}
|
|
|
|
struct timespec ts;
|
|
s3fs_realtime(ts);
|
|
ent->SetCtime(ts);
|
|
|
|
// Change file mode
|
|
ent->SetMode(mode);
|
|
|
|
// upload
|
|
if(0 != (result = ent->Flush(autoent.GetPseudoFd(), AutoLock::NONE, true))){
|
|
S3FS_PRN_ERR("could not upload file(%s): result=%d", strpath.c_str(), result);
|
|
return result;
|
|
}
|
|
StatCache::getStatCacheData()->DelStat(nowcache);
|
|
}
|
|
S3FS_MALLOCTRIM(0);
|
|
|
|
return result;
|
|
}
|
|
|
|
static int s3fs_chown(const char* _path, uid_t uid, gid_t gid)
|
|
{
|
|
WTF8_ENCODE(path)
|
|
int result;
|
|
std::string strpath;
|
|
std::string newpath;
|
|
std::string nowcache;
|
|
headers_t meta;
|
|
struct stat stbuf;
|
|
dirtype nDirType = DIRTYPE_UNKNOWN;
|
|
|
|
S3FS_PRN_INFO("[path=%s][uid=%u][gid=%u]", path, (unsigned int)uid, (unsigned int)gid);
|
|
|
|
if(0 != (result = check_parent_object_access(path, X_OK))){
|
|
return result;
|
|
}
|
|
if(0 != (result = check_object_owner(path, &stbuf))){
|
|
return result;
|
|
}
|
|
|
|
if((uid_t)(-1) == uid){
|
|
uid = stbuf.st_uid;
|
|
}
|
|
if((gid_t)(-1) == gid){
|
|
gid = stbuf.st_gid;
|
|
}
|
|
if(S_ISDIR(stbuf.st_mode)){
|
|
result = chk_dir_object_type(path, newpath, strpath, nowcache, &meta, &nDirType);
|
|
}else{
|
|
strpath = path;
|
|
nowcache = strpath;
|
|
result = get_object_attribute(strpath.c_str(), NULL, &meta);
|
|
}
|
|
if(0 != result){
|
|
return result;
|
|
}
|
|
|
|
if(S_ISDIR(stbuf.st_mode) && (IS_REPLACEDIR(nDirType) || IS_CREATE_MP_STAT(path))){
|
|
std::string xattrvalue;
|
|
const char* pxattrvalue;
|
|
if(get_meta_xattr_value(path, xattrvalue)){
|
|
pxattrvalue = xattrvalue.c_str();
|
|
}else{
|
|
pxattrvalue = NULL;
|
|
}
|
|
|
|
if(IS_REPLACEDIR(nDirType)){
|
|
// Should rebuild directory object(except new type)
|
|
// Need to remove old dir("dir" etc) and make new dir("dir/")
|
|
|
|
// At first, remove directory old object
|
|
if(0 != (result = remove_old_type_dir(strpath, nDirType))){
|
|
return result;
|
|
}
|
|
}
|
|
StatCache::getStatCacheData()->DelStat(nowcache);
|
|
|
|
// Make new directory object("dir/")
|
|
struct timespec ts_atime;
|
|
struct timespec ts_mtime;
|
|
struct timespec ts_ctime;
|
|
set_stat_to_timespec(stbuf, ST_TYPE_ATIME, ts_atime);
|
|
set_stat_to_timespec(stbuf, ST_TYPE_MTIME, ts_mtime);
|
|
s3fs_realtime(ts_ctime);
|
|
|
|
if(0 != (result = create_directory_object(newpath.c_str(), stbuf.st_mode, ts_atime, ts_mtime, ts_ctime, uid, gid, pxattrvalue))){
|
|
return result;
|
|
}
|
|
}else{
|
|
std::string strSourcePath = (mount_prefix.empty() && "/" == strpath) ? "//" : strpath;
|
|
headers_t updatemeta;
|
|
updatemeta["x-amz-meta-ctime"] = s3fs_str_realtime();
|
|
updatemeta["x-amz-meta-uid"] = str(uid);
|
|
updatemeta["x-amz-meta-gid"] = str(gid);
|
|
updatemeta["x-amz-copy-source"] = urlEncode(service_path + S3fsCred::GetBucket() + get_realpath(strSourcePath.c_str()));
|
|
updatemeta["x-amz-metadata-directive"] = "REPLACE";
|
|
|
|
// check opened file handle.
|
|
//
|
|
// If the file starts uploading by multipart when the disk capacity is insufficient,
|
|
// we need to put these header after finishing upload.
|
|
// Or if the file is only open, we must update to FdEntity's internal meta.
|
|
//
|
|
AutoFdEntity autoent;
|
|
FdEntity* ent;
|
|
bool need_put_header = true;
|
|
if(NULL != (ent = autoent.OpenExistFdEntity(path))){
|
|
if(ent->MergeOrgMeta(updatemeta)){
|
|
// meta is changed, but now uploading.
|
|
// then the meta is pending and accumulated to be put after the upload is complete.
|
|
S3FS_PRN_INFO("meta pending until upload is complete");
|
|
need_put_header = false;
|
|
|
|
// If there is data in the Stats cache, update the Stats cache.
|
|
StatCache::getStatCacheData()->UpdateMetaStats(strpath, updatemeta);
|
|
}
|
|
}
|
|
if(need_put_header){
|
|
// not found opened file.
|
|
merge_headers(meta, updatemeta, true);
|
|
|
|
// upload meta directly.
|
|
if(0 != (result = put_headers(strpath.c_str(), meta, true))){
|
|
return result;
|
|
}
|
|
StatCache::getStatCacheData()->DelStat(nowcache);
|
|
}
|
|
}
|
|
S3FS_MALLOCTRIM(0);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int s3fs_chown_nocopy(const char* _path, uid_t uid, gid_t gid)
|
|
{
|
|
WTF8_ENCODE(path)
|
|
int result;
|
|
std::string strpath;
|
|
std::string newpath;
|
|
std::string nowcache;
|
|
struct stat stbuf;
|
|
dirtype nDirType = DIRTYPE_UNKNOWN;
|
|
|
|
S3FS_PRN_INFO1("[path=%s][uid=%u][gid=%u]", path, (unsigned int)uid, (unsigned int)gid);
|
|
|
|
if(0 != (result = check_parent_object_access(path, X_OK))){
|
|
return result;
|
|
}
|
|
if(0 != (result = check_object_owner(path, &stbuf))){
|
|
return result;
|
|
}
|
|
|
|
if((uid_t)(-1) == uid){
|
|
uid = stbuf.st_uid;
|
|
}
|
|
if((gid_t)(-1) == gid){
|
|
gid = stbuf.st_gid;
|
|
}
|
|
|
|
// Get attributes
|
|
if(S_ISDIR(stbuf.st_mode)){
|
|
result = chk_dir_object_type(path, newpath, strpath, nowcache, NULL, &nDirType);
|
|
}else{
|
|
strpath = path;
|
|
nowcache = strpath;
|
|
result = get_object_attribute(strpath.c_str(), NULL, NULL);
|
|
}
|
|
if(0 != result){
|
|
return result;
|
|
}
|
|
|
|
if(S_ISDIR(stbuf.st_mode)){
|
|
std::string xattrvalue;
|
|
const char* pxattrvalue;
|
|
if(get_meta_xattr_value(path, xattrvalue)){
|
|
pxattrvalue = xattrvalue.c_str();
|
|
}else{
|
|
pxattrvalue = NULL;
|
|
}
|
|
|
|
if(IS_REPLACEDIR(nDirType)){
|
|
// Should rebuild all directory object
|
|
// Need to remove old dir("dir" etc) and make new dir("dir/")
|
|
|
|
// At first, remove directory old object
|
|
if(0 != (result = remove_old_type_dir(strpath, nDirType))){
|
|
return result;
|
|
}
|
|
}
|
|
StatCache::getStatCacheData()->DelStat(nowcache);
|
|
|
|
// Make new directory object("dir/")
|
|
struct timespec ts_atime;
|
|
struct timespec ts_mtime;
|
|
struct timespec ts_ctime;
|
|
set_stat_to_timespec(stbuf, ST_TYPE_ATIME, ts_atime);
|
|
set_stat_to_timespec(stbuf, ST_TYPE_MTIME, ts_mtime);
|
|
s3fs_realtime(ts_ctime);
|
|
|
|
if(0 != (result = create_directory_object(newpath.c_str(), stbuf.st_mode, ts_atime, ts_mtime, ts_ctime, uid, gid, pxattrvalue))){
|
|
return result;
|
|
}
|
|
}else{
|
|
// normal object or directory object of newer version
|
|
|
|
// open & load
|
|
AutoFdEntity autoent;
|
|
FdEntity* ent;
|
|
if(0 != (result = get_local_fent(autoent, &ent, strpath.c_str(), O_RDWR, true))){
|
|
S3FS_PRN_ERR("could not open and read file(%s)", strpath.c_str());
|
|
return result;
|
|
}
|
|
|
|
struct timespec ts;
|
|
s3fs_realtime(ts);
|
|
ent->SetCtime(ts);
|
|
|
|
// Change owner
|
|
ent->SetUId(uid);
|
|
ent->SetGId(gid);
|
|
|
|
// upload
|
|
if(0 != (result = ent->Flush(autoent.GetPseudoFd(), AutoLock::NONE, true))){
|
|
S3FS_PRN_ERR("could not upload file(%s): result=%d", strpath.c_str(), result);
|
|
return result;
|
|
}
|
|
StatCache::getStatCacheData()->DelStat(nowcache);
|
|
}
|
|
S3FS_MALLOCTRIM(0);
|
|
|
|
return result;
|
|
}
|
|
|
|
static timespec handle_utimens_special_values(timespec ts, timespec now, timespec orig)
|
|
{
|
|
if(ts.tv_nsec == UTIME_NOW){
|
|
return now;
|
|
}else if(ts.tv_nsec == UTIME_OMIT){
|
|
return orig;
|
|
}else{
|
|
return ts;
|
|
}
|
|
}
|
|
|
|
static int update_mctime_parent_directory(const char* _path)
|
|
{
|
|
if(!update_parent_dir_stat){
|
|
// Disable updating parent directory stat.
|
|
S3FS_PRN_DBG("Updating parent directory stats is disabled");
|
|
return 0;
|
|
}
|
|
|
|
WTF8_ENCODE(path)
|
|
int result;
|
|
std::string parentpath; // parent directory path
|
|
std::string nowpath; // now directory object path("dir" or "dir/" or "xxx_$folder$", etc)
|
|
std::string newpath; // directory path for the current version("dir/")
|
|
std::string nowcache;
|
|
headers_t meta;
|
|
struct stat stbuf;
|
|
struct timespec mctime;
|
|
struct timespec atime;
|
|
dirtype nDirType = DIRTYPE_UNKNOWN;
|
|
|
|
S3FS_PRN_INFO2("[path=%s]", path);
|
|
|
|
// get parent directory path
|
|
parentpath = mydirname(path);
|
|
|
|
// check & get directory type
|
|
if(0 != (result = chk_dir_object_type(parentpath.c_str(), newpath, nowpath, nowcache, &meta, &nDirType))){
|
|
return result;
|
|
}
|
|
|
|
// get directory stat
|
|
//
|
|
// [NOTE]
|
|
// It is assumed that this function is called after the operation on
|
|
// the file is completed, so there is no need to check the permissions
|
|
// on the parent directory.
|
|
//
|
|
if(0 != (result = get_object_attribute(parentpath.c_str(), &stbuf))){
|
|
// If there is not the target file(object), result is -ENOENT.
|
|
return result;
|
|
}
|
|
if(!S_ISDIR(stbuf.st_mode)){
|
|
S3FS_PRN_ERR("path(%s) is not parent directory.", parentpath.c_str());
|
|
return -EIO;
|
|
}
|
|
|
|
// make atime/mtime/ctime for updating
|
|
s3fs_realtime(mctime);
|
|
set_stat_to_timespec(stbuf, ST_TYPE_ATIME, atime);
|
|
|
|
if(0 == atime.tv_sec && 0 == atime.tv_nsec){
|
|
atime = mctime;
|
|
}
|
|
|
|
if(nocopyapi || IS_REPLACEDIR(nDirType) || IS_CREATE_MP_STAT(parentpath.c_str())){
|
|
// Should rebuild directory object(except new type)
|
|
// Need to remove old dir("dir" etc) and make new dir("dir/")
|
|
std::string xattrvalue;
|
|
const char* pxattrvalue;
|
|
if(get_meta_xattr_value(path, xattrvalue)){
|
|
pxattrvalue = xattrvalue.c_str();
|
|
}else{
|
|
pxattrvalue = NULL;
|
|
}
|
|
|
|
// At first, remove directory old object
|
|
if(!nowpath.empty()){
|
|
if(0 != (result = remove_old_type_dir(nowpath, nDirType))){
|
|
return result;
|
|
}
|
|
}
|
|
if(!nowcache.empty()){
|
|
StatCache::getStatCacheData()->DelStat(nowcache);
|
|
}
|
|
|
|
// Make new directory object("dir/")
|
|
if(0 != (result = create_directory_object(newpath.c_str(), stbuf.st_mode, atime, mctime, mctime, stbuf.st_uid, stbuf.st_gid, pxattrvalue))){
|
|
return result;
|
|
}
|
|
}else{
|
|
std::string strSourcePath = (mount_prefix.empty() && "/" == nowpath) ? "//" : nowpath;
|
|
headers_t updatemeta;
|
|
updatemeta["x-amz-meta-mtime"] = str(mctime);
|
|
updatemeta["x-amz-meta-ctime"] = str(mctime);
|
|
updatemeta["x-amz-meta-atime"] = str(atime);
|
|
updatemeta["x-amz-copy-source"] = urlEncode(service_path + S3fsCred::GetBucket() + get_realpath(strSourcePath.c_str()));
|
|
updatemeta["x-amz-metadata-directive"] = "REPLACE";
|
|
|
|
merge_headers(meta, updatemeta, true);
|
|
|
|
// upload meta for parent directory.
|
|
if(0 != (result = put_headers(nowpath.c_str(), meta, true))){
|
|
return result;
|
|
}
|
|
StatCache::getStatCacheData()->DelStat(nowcache);
|
|
}
|
|
S3FS_MALLOCTRIM(0);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int s3fs_utimens(const char* _path, const struct timespec ts[2])
|
|
{
|
|
WTF8_ENCODE(path)
|
|
int result;
|
|
std::string strpath;
|
|
std::string newpath;
|
|
std::string nowcache;
|
|
headers_t meta;
|
|
struct stat stbuf;
|
|
dirtype nDirType = DIRTYPE_UNKNOWN;
|
|
|
|
S3FS_PRN_INFO("[path=%s][mtime=%s][ctime/atime=%s]", path, str(ts[1]).c_str(), str(ts[0]).c_str());
|
|
|
|
if(0 != (result = check_parent_object_access(path, X_OK))){
|
|
return result;
|
|
}
|
|
if(0 != (result = check_object_access(path, W_OK, &stbuf))){
|
|
if(0 != check_object_owner(path, &stbuf)){
|
|
return result;
|
|
}
|
|
}
|
|
|
|
struct timespec now;
|
|
struct timespec ts_atime;
|
|
struct timespec ts_ctime;
|
|
struct timespec ts_mtime;
|
|
|
|
s3fs_realtime(now);
|
|
set_stat_to_timespec(stbuf, ST_TYPE_ATIME, ts_atime);
|
|
set_stat_to_timespec(stbuf, ST_TYPE_CTIME, ts_ctime);
|
|
set_stat_to_timespec(stbuf, ST_TYPE_MTIME, ts_mtime);
|
|
|
|
struct timespec atime = handle_utimens_special_values(ts[0], now, ts_atime);
|
|
struct timespec ctime = handle_utimens_special_values(ts[0], now, ts_ctime);
|
|
struct timespec mtime = handle_utimens_special_values(ts[1], now, ts_mtime);
|
|
|
|
if(S_ISDIR(stbuf.st_mode)){
|
|
result = chk_dir_object_type(path, newpath, strpath, nowcache, &meta, &nDirType);
|
|
}else{
|
|
strpath = path;
|
|
nowcache = strpath;
|
|
result = get_object_attribute(strpath.c_str(), NULL, &meta);
|
|
}
|
|
if(0 != result){
|
|
return result;
|
|
}
|
|
|
|
if(S_ISDIR(stbuf.st_mode) && (IS_REPLACEDIR(nDirType) || IS_CREATE_MP_STAT(path))){
|
|
std::string xattrvalue;
|
|
const char* pxattrvalue;
|
|
if(get_meta_xattr_value(path, xattrvalue)){
|
|
pxattrvalue = xattrvalue.c_str();
|
|
}else{
|
|
pxattrvalue = NULL;
|
|
}
|
|
|
|
if(IS_REPLACEDIR(nDirType)){
|
|
// Should rebuild directory object(except new type)
|
|
// Need to remove old dir("dir" etc) and make new dir("dir/")
|
|
|
|
// At first, remove directory old object
|
|
if(0 != (result = remove_old_type_dir(strpath, nDirType))){
|
|
return result;
|
|
}
|
|
}
|
|
StatCache::getStatCacheData()->DelStat(nowcache);
|
|
|
|
// Make new directory object("dir/")
|
|
if(0 != (result = create_directory_object(newpath.c_str(), stbuf.st_mode, atime, mtime, ctime, stbuf.st_uid, stbuf.st_gid, pxattrvalue))){
|
|
return result;
|
|
}
|
|
}else{
|
|
std::string strSourcePath = (mount_prefix.empty() && "/" == strpath) ? "//" : strpath;
|
|
headers_t updatemeta;
|
|
updatemeta["x-amz-meta-mtime"] = str(mtime);
|
|
updatemeta["x-amz-meta-ctime"] = str(ctime);
|
|
updatemeta["x-amz-meta-atime"] = str(atime);
|
|
updatemeta["x-amz-copy-source"] = urlEncode(service_path + S3fsCred::GetBucket() + get_realpath(strSourcePath.c_str()));
|
|
updatemeta["x-amz-metadata-directive"] = "REPLACE";
|
|
|
|
// check opened file handle.
|
|
//
|
|
// If the file starts uploading by multipart when the disk capacity is insufficient,
|
|
// we need to put these header after finishing upload.
|
|
// Or if the file is only open, we must update to FdEntity's internal meta.
|
|
//
|
|
AutoFdEntity autoent;
|
|
FdEntity* ent;
|
|
bool need_put_header = true;
|
|
bool keep_mtime = false;
|
|
if(NULL != (ent = autoent.OpenExistFdEntity(path))){
|
|
if(ent->MergeOrgMeta(updatemeta)){
|
|
// meta is changed, but now uploading.
|
|
// then the meta is pending and accumulated to be put after the upload is complete.
|
|
S3FS_PRN_INFO("meta pending until upload is complete");
|
|
need_put_header = false;
|
|
ent->SetHoldingMtime(mtime);
|
|
|
|
// If there is data in the Stats cache, update the Stats cache.
|
|
StatCache::getStatCacheData()->UpdateMetaStats(strpath, updatemeta);
|
|
|
|
}else{
|
|
S3FS_PRN_INFO("meta is not pending, but need to keep current mtime.");
|
|
|
|
// [NOTE]
|
|
// Depending on the order in which write/flush and utimens are called,
|
|
// the mtime updated here may be overwritten at the time of flush.
|
|
// To avoid that, set a special flag.
|
|
//
|
|
keep_mtime = true;
|
|
}
|
|
}
|
|
if(need_put_header){
|
|
// not found opened file.
|
|
merge_headers(meta, updatemeta, true);
|
|
|
|
// upload meta directly.
|
|
if(0 != (result = put_headers(strpath.c_str(), meta, true))){
|
|
return result;
|
|
}
|
|
StatCache::getStatCacheData()->DelStat(nowcache);
|
|
|
|
if(keep_mtime){
|
|
ent->SetHoldingMtime(mtime);
|
|
}
|
|
}
|
|
}
|
|
S3FS_MALLOCTRIM(0);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int s3fs_utimens_nocopy(const char* _path, const struct timespec ts[2])
|
|
{
|
|
WTF8_ENCODE(path)
|
|
int result;
|
|
std::string strpath;
|
|
std::string newpath;
|
|
std::string nowcache;
|
|
struct stat stbuf;
|
|
dirtype nDirType = DIRTYPE_UNKNOWN;
|
|
|
|
S3FS_PRN_INFO1("[path=%s][mtime=%s][atime/ctime=%s]", path, str(ts[1]).c_str(), str(ts[0]).c_str());
|
|
|
|
if(0 != (result = check_parent_object_access(path, X_OK))){
|
|
return result;
|
|
}
|
|
if(0 != (result = check_object_access(path, W_OK, &stbuf))){
|
|
if(0 != check_object_owner(path, &stbuf)){
|
|
return result;
|
|
}
|
|
}
|
|
|
|
struct timespec now;
|
|
struct timespec ts_atime;
|
|
struct timespec ts_ctime;
|
|
struct timespec ts_mtime;
|
|
|
|
s3fs_realtime(now);
|
|
set_stat_to_timespec(stbuf, ST_TYPE_ATIME, ts_atime);
|
|
set_stat_to_timespec(stbuf, ST_TYPE_CTIME, ts_ctime);
|
|
set_stat_to_timespec(stbuf, ST_TYPE_MTIME, ts_mtime);
|
|
|
|
struct timespec atime = handle_utimens_special_values(ts[0], now, ts_atime);
|
|
struct timespec ctime = handle_utimens_special_values(ts[0], now, ts_ctime);
|
|
struct timespec mtime = handle_utimens_special_values(ts[1], now, ts_mtime);
|
|
|
|
// Get attributes
|
|
if(S_ISDIR(stbuf.st_mode)){
|
|
result = chk_dir_object_type(path, newpath, strpath, nowcache, NULL, &nDirType);
|
|
}else{
|
|
strpath = path;
|
|
nowcache = strpath;
|
|
result = get_object_attribute(strpath.c_str(), NULL, NULL);
|
|
}
|
|
if(0 != result){
|
|
return result;
|
|
}
|
|
|
|
if(S_ISDIR(stbuf.st_mode)){
|
|
std::string xattrvalue;
|
|
const char* pxattrvalue;
|
|
if(get_meta_xattr_value(path, xattrvalue)){
|
|
pxattrvalue = xattrvalue.c_str();
|
|
}else{
|
|
pxattrvalue = NULL;
|
|
}
|
|
|
|
if(IS_REPLACEDIR(nDirType)){
|
|
// Should rebuild all directory object
|
|
// Need to remove old dir("dir" etc) and make new dir("dir/")
|
|
|
|
// At first, remove directory old object
|
|
if(0 != (result = remove_old_type_dir(strpath, nDirType))){
|
|
return result;
|
|
}
|
|
}
|
|
StatCache::getStatCacheData()->DelStat(nowcache);
|
|
|
|
// Make new directory object("dir/")
|
|
if(0 != (result = create_directory_object(newpath.c_str(), stbuf.st_mode, atime, mtime, ctime, stbuf.st_uid, stbuf.st_gid, pxattrvalue))){
|
|
return result;
|
|
}
|
|
}else{
|
|
// normal object or directory object of newer version
|
|
|
|
// open & load
|
|
AutoFdEntity autoent;
|
|
FdEntity* ent;
|
|
if(0 != (result = get_local_fent(autoent, &ent, strpath.c_str(), O_RDWR, true))){
|
|
S3FS_PRN_ERR("could not open and read file(%s)", strpath.c_str());
|
|
return result;
|
|
}
|
|
|
|
// set mtime/ctime
|
|
if(0 != (result = ent->SetMCtime(mtime, ctime))){
|
|
S3FS_PRN_ERR("could not set mtime and ctime to file(%s): result=%d", strpath.c_str(), result);
|
|
return result;
|
|
}
|
|
|
|
// set atime
|
|
if(0 != (result = ent->SetAtime(atime))){
|
|
S3FS_PRN_ERR("could not set atime to file(%s): result=%d", strpath.c_str(), result);
|
|
return result;
|
|
}
|
|
|
|
// upload
|
|
if(0 != (result = ent->Flush(autoent.GetPseudoFd(), AutoLock::NONE, true))){
|
|
S3FS_PRN_ERR("could not upload file(%s): result=%d", strpath.c_str(), result);
|
|
return result;
|
|
}
|
|
StatCache::getStatCacheData()->DelStat(nowcache);
|
|
}
|
|
S3FS_MALLOCTRIM(0);
|
|
|
|
return result;
|
|
}
|
|
|
|
static int s3fs_truncate(const char* _path, off_t size)
|
|
{
|
|
WTF8_ENCODE(path)
|
|
int result;
|
|
headers_t meta;
|
|
AutoFdEntity autoent;
|
|
FdEntity* ent = NULL;
|
|
|
|
S3FS_PRN_INFO("[path=%s][size=%lld]", path, static_cast<long long>(size));
|
|
|
|
if(size < 0){
|
|
size = 0;
|
|
}
|
|
|
|
if(0 != (result = check_parent_object_access(path, X_OK))){
|
|
return result;
|
|
}
|
|
if(0 != (result = check_object_access(path, W_OK, NULL))){
|
|
return result;
|
|
}
|
|
|
|
// Get file information
|
|
if(0 == (result = get_object_attribute(path, NULL, &meta))){
|
|
// File exists
|
|
|
|
// [NOTE]
|
|
// If the file exists, the file has already been opened by FUSE before
|
|
// truncate is called. Then the call below will change the file size.
|
|
// (When an already open file is changed the file size, FUSE will not
|
|
// reopen it.)
|
|
// The Flush is called before this file is closed, so there is no need
|
|
// to do it here.
|
|
//
|
|
// [NOTICE]
|
|
// FdManager::Open() ignores changes that reduce the file size for the
|
|
// file you are editing. However, if user opens only onece, edit it,
|
|
// and then shrink the file, it should be done.
|
|
// When this function is called, the file is already open by FUSE or
|
|
// some other operation. Therefore, if the number of open files is 1,
|
|
// no edits other than that fd will be made, and the files will be
|
|
// shrunk using ignore_modify flag even while editing.
|
|
// See the comments when merging this code for FUSE2 limitations.
|
|
// (In FUSE3, it will be possible to handle it reliably using fuse_file_info.)
|
|
//
|
|
bool ignore_modify;
|
|
if(1 < FdManager::GetOpenFdCount(path)){
|
|
ignore_modify = false;
|
|
}else{
|
|
ignore_modify = true;
|
|
}
|
|
|
|
if(NULL == (ent = autoent.Open(path, &meta, size, S3FS_OMIT_TS, O_RDWR, false, true, ignore_modify, AutoLock::NONE))){
|
|
S3FS_PRN_ERR("could not open file(%s): errno=%d", path, errno);
|
|
return -EIO;
|
|
}
|
|
ent->UpdateCtime();
|
|
|
|
#if defined(__APPLE__)
|
|
// [NOTE]
|
|
// Only for macos, this truncate calls to "size=0" do not reflect size.
|
|
// The cause is unknown now, but it can be avoided by flushing the file.
|
|
//
|
|
if(0 == size){
|
|
if(0 != (result = ent->Flush(autoent.GetPseudoFd(), AutoLock::NONE, true))){
|
|
S3FS_PRN_ERR("could not upload file(%s): result=%d", path, result);
|
|
return result;
|
|
}
|
|
StatCache::getStatCacheData()->DelStat(path);
|
|
}
|
|
#endif
|
|
|
|
}else{
|
|
// Not found -> Make tmpfile(with size)
|
|
struct fuse_context* pcxt;
|
|
if(NULL == (pcxt = fuse_get_context())){
|
|
return -EIO;
|
|
}
|
|
|
|
std::string strnow = s3fs_str_realtime();
|
|
meta["Content-Type"] = std::string("application/octet-stream"); // Static
|
|
meta["x-amz-meta-mode"] = str(S_IFLNK | S_IRWXU | S_IRWXG | S_IRWXO);
|
|
meta["x-amz-meta-ctime"] = strnow;
|
|
meta["x-amz-meta-mtime"] = strnow;
|
|
meta["x-amz-meta-uid"] = str(pcxt->uid);
|
|
meta["x-amz-meta-gid"] = str(pcxt->gid);
|
|
|
|
if(NULL == (ent = autoent.Open(path, &meta, size, S3FS_OMIT_TS, O_RDWR, true, true, false, AutoLock::NONE))){
|
|
S3FS_PRN_ERR("could not open file(%s): errno=%d", path, errno);
|
|
return -EIO;
|
|
}
|
|
if(0 != (result = ent->Flush(autoent.GetPseudoFd(), AutoLock::NONE, true))){
|
|
S3FS_PRN_ERR("could not upload file(%s): result=%d", path, result);
|
|
return result;
|
|
}
|
|
StatCache::getStatCacheData()->DelStat(path);
|
|
}
|
|
S3FS_MALLOCTRIM(0);
|
|
|
|
return result;
|
|
}
|
|
|
|
static int s3fs_open(const char* _path, struct fuse_file_info* fi)
|
|
{
|
|
WTF8_ENCODE(path)
|
|
int result;
|
|
struct stat st;
|
|
bool needs_flush = false;
|
|
|
|
S3FS_PRN_INFO("[path=%s][flags=0x%x]", path, fi->flags);
|
|
|
|
if ((fi->flags & O_ACCMODE) == O_RDONLY && fi->flags & O_TRUNC) {
|
|
return -EACCES;
|
|
}
|
|
|
|
// [NOTE]
|
|
// Delete the Stats cache only if the file is not open.
|
|
// If the file is open, the stats cache will not be deleted as
|
|
// there are cases where the object does not exist on the server
|
|
// and only the Stats cache exists.
|
|
//
|
|
if(StatCache::getStatCacheData()->HasStat(path)){
|
|
if(!FdManager::HasOpenEntityFd(path)){
|
|
StatCache::getStatCacheData()->DelStat(path);
|
|
}
|
|
}
|
|
|
|
int mask = (O_RDONLY != (fi->flags & O_ACCMODE) ? W_OK : R_OK);
|
|
if(0 != (result = check_parent_object_access(path, X_OK))){
|
|
return result;
|
|
}
|
|
|
|
result = check_object_access(path, mask, &st);
|
|
if(-ENOENT == result){
|
|
if(0 != (result = check_parent_object_access(path, W_OK))){
|
|
return result;
|
|
}
|
|
}else if(0 != result){
|
|
return result;
|
|
}
|
|
|
|
if((unsigned int)fi->flags & O_TRUNC){
|
|
if(0 != st.st_size){
|
|
st.st_size = 0;
|
|
needs_flush = true;
|
|
}
|
|
}
|
|
if(!S_ISREG(st.st_mode) || S_ISLNK(st.st_mode)){
|
|
st.st_mtime = -1;
|
|
}
|
|
|
|
AutoFdEntity autoent;
|
|
FdEntity* ent;
|
|
headers_t meta;
|
|
if(0 != (result = get_object_attribute(path, NULL, &meta, true, NULL, true))){ // no truncate cache
|
|
return result;
|
|
}
|
|
|
|
struct timespec st_mctime;
|
|
set_stat_to_timespec(st, ST_TYPE_MTIME, st_mctime);
|
|
|
|
if(NULL == (ent = autoent.Open(path, &meta, st.st_size, st_mctime, fi->flags, false, true, false, AutoLock::NONE))){
|
|
StatCache::getStatCacheData()->DelStat(path);
|
|
return -EIO;
|
|
}
|
|
|
|
if (needs_flush){
|
|
struct timespec ts;
|
|
s3fs_realtime(ts);
|
|
ent->SetMCtime(ts, ts);
|
|
|
|
if(0 != (result = ent->RowFlush(autoent.GetPseudoFd(), path, AutoLock::NONE, true))){
|
|
S3FS_PRN_ERR("could not upload file(%s): result=%d", path, result);
|
|
StatCache::getStatCacheData()->DelStat(path);
|
|
return result;
|
|
}
|
|
}
|
|
fi->fh = autoent.Detach(); // KEEP fdentity open;
|
|
|
|
S3FS_MALLOCTRIM(0);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int s3fs_read(const char* _path, char* buf, size_t size, off_t offset, struct fuse_file_info* fi)
|
|
{
|
|
WTF8_ENCODE(path)
|
|
ssize_t res;
|
|
|
|
S3FS_PRN_DBG("[path=%s][size=%zu][offset=%lld][pseudo_fd=%llu]", path, size, static_cast<long long>(offset), (unsigned long long)(fi->fh));
|
|
|
|
AutoFdEntity autoent;
|
|
FdEntity* ent;
|
|
if(NULL == (ent = autoent.GetExistFdEntity(path, static_cast<int>(fi->fh)))){
|
|
S3FS_PRN_ERR("could not find opened pseudo_fd(=%llu) for path(%s)", (unsigned long long)(fi->fh), path);
|
|
return -EIO;
|
|
}
|
|
|
|
// check real file size
|
|
off_t realsize = 0;
|
|
if(!ent->GetSize(realsize) || 0 == realsize){
|
|
S3FS_PRN_DBG("file size is 0, so break to read.");
|
|
return 0;
|
|
}
|
|
|
|
if(0 > (res = ent->Read(static_cast<int>(fi->fh), buf, offset, size, false))){
|
|
S3FS_PRN_WARN("failed to read file(%s). result=%zd", path, res);
|
|
}
|
|
|
|
return static_cast<int>(res);
|
|
}
|
|
|
|
static int s3fs_write(const char* _path, const char* buf, size_t size, off_t offset, struct fuse_file_info* fi)
|
|
{
|
|
WTF8_ENCODE(path)
|
|
ssize_t res;
|
|
|
|
S3FS_PRN_DBG("[path=%s][size=%zu][offset=%lld][pseudo_fd=%llu]", path, size, static_cast<long long int>(offset), (unsigned long long)(fi->fh));
|
|
|
|
AutoFdEntity autoent;
|
|
FdEntity* ent;
|
|
if(NULL == (ent = autoent.GetExistFdEntity(path, static_cast<int>(fi->fh)))){
|
|
S3FS_PRN_ERR("could not find opened pseudo_fd(%llu) for path(%s)", (unsigned long long)(fi->fh), path);
|
|
return -EIO;
|
|
}
|
|
|
|
if(0 > (res = ent->Write(static_cast<int>(fi->fh), buf, offset, size))){
|
|
S3FS_PRN_WARN("failed to write file(%s). result=%zd", path, res);
|
|
}
|
|
|
|
if(max_dirty_data != -1 && ent->BytesModified() >= max_dirty_data){
|
|
int flushres;
|
|
if(0 != (flushres = ent->RowFlush(static_cast<int>(fi->fh), path, AutoLock::NONE, true))){
|
|
S3FS_PRN_ERR("could not upload file(%s): result=%d", path, flushres);
|
|
StatCache::getStatCacheData()->DelStat(path);
|
|
return flushres;
|
|
}
|
|
// Punch a hole in the file to recover disk space.
|
|
if(!ent->PunchHole()){
|
|
S3FS_PRN_WARN("could not punching HOLEs to a cache file, but continue.");
|
|
}
|
|
}
|
|
|
|
return static_cast<int>(res);
|
|
}
|
|
|
|
static int s3fs_statfs(const char* _path, struct statvfs* stbuf)
|
|
{
|
|
// WTF8_ENCODE(path)
|
|
stbuf->f_bsize = 16 * 1024 * 1024;
|
|
stbuf->f_namemax = NAME_MAX;
|
|
#if defined(__MSYS__)
|
|
// WinFsp resolves the free space from f_bfree * f_frsize, and the total space from f_blocks * f_frsize (in bytes).
|
|
stbuf->f_frsize = stbuf->f_bsize;
|
|
stbuf->f_blocks = INT32_MAX;
|
|
stbuf->f_bfree = INT32_MAX;
|
|
#elif defined(__APPLE__)
|
|
stbuf->f_blocks = UINT32_MAX;
|
|
stbuf->f_bfree = UINT32_MAX;
|
|
stbuf->f_files = UINT32_MAX;
|
|
stbuf->f_ffree = UINT32_MAX;
|
|
#else
|
|
stbuf->f_blocks = static_cast<fsblkcnt_t>(~0) / stbuf->f_bsize;
|
|
stbuf->f_bfree = stbuf->f_blocks;
|
|
#endif
|
|
stbuf->f_bavail = stbuf->f_blocks;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int s3fs_flush(const char* _path, struct fuse_file_info* fi)
|
|
{
|
|
WTF8_ENCODE(path)
|
|
int result;
|
|
|
|
S3FS_PRN_INFO("[path=%s][pseudo_fd=%llu]", path, (unsigned long long)(fi->fh));
|
|
|
|
int mask = (O_RDONLY != (fi->flags & O_ACCMODE) ? W_OK : R_OK);
|
|
if(0 != (result = check_parent_object_access(path, X_OK))){
|
|
return result;
|
|
}
|
|
result = check_object_access(path, mask, NULL);
|
|
if(-ENOENT == result){
|
|
if(0 != (result = check_parent_object_access(path, W_OK))){
|
|
return result;
|
|
}
|
|
}else if(0 != result){
|
|
return result;
|
|
}
|
|
|
|
AutoFdEntity autoent;
|
|
FdEntity* ent;
|
|
if(NULL != (ent = autoent.GetExistFdEntity(path, static_cast<int>(fi->fh)))){
|
|
bool is_new_file = ent->IsDirtyNewFile();
|
|
|
|
ent->UpdateMtime(true); // clear the flag not to update mtime.
|
|
ent->UpdateCtime();
|
|
result = ent->Flush(static_cast<int>(fi->fh), AutoLock::NONE, false);
|
|
StatCache::getStatCacheData()->DelStat(path);
|
|
|
|
if(is_new_file){
|
|
// update parent directory timestamp
|
|
int update_result;
|
|
if(0 != (update_result = update_mctime_parent_directory(path))){
|
|
S3FS_PRN_ERR("succeed to create the file(%s), but could not update timestamp of its parent directory(result=%d).", path, update_result);
|
|
}
|
|
}
|
|
}
|
|
S3FS_MALLOCTRIM(0);
|
|
|
|
return result;
|
|
}
|
|
|
|
// [NOTICE]
|
|
// Assumption is a valid fd.
|
|
//
|
|
static int s3fs_fsync(const char* _path, int datasync, struct fuse_file_info* fi)
|
|
{
|
|
WTF8_ENCODE(path)
|
|
int result = 0;
|
|
|
|
S3FS_PRN_INFO("[path=%s][pseudo_fd=%llu]", path, (unsigned long long)(fi->fh));
|
|
|
|
AutoFdEntity autoent;
|
|
FdEntity* ent;
|
|
if(NULL != (ent = autoent.GetExistFdEntity(path, static_cast<int>(fi->fh)))){
|
|
bool is_new_file = ent->IsDirtyNewFile();
|
|
|
|
if(0 == datasync){
|
|
ent->UpdateMtime();
|
|
ent->UpdateCtime();
|
|
}
|
|
result = ent->Flush(static_cast<int>(fi->fh), AutoLock::NONE, false);
|
|
|
|
if(is_new_file){
|
|
// update parent directory timestamp
|
|
int update_result;
|
|
if(0 != (update_result = update_mctime_parent_directory(path))){
|
|
S3FS_PRN_ERR("succeed to create the file(%s), but could not update timestamp of its parent directory(result=%d).", path, update_result);
|
|
}
|
|
}
|
|
}
|
|
S3FS_MALLOCTRIM(0);
|
|
|
|
// Issue 320: Delete stat cache entry because st_size may have changed.
|
|
StatCache::getStatCacheData()->DelStat(path);
|
|
|
|
return result;
|
|
}
|
|
|
|
static int s3fs_release(const char* _path, struct fuse_file_info* fi)
|
|
{
|
|
WTF8_ENCODE(path)
|
|
S3FS_PRN_INFO("[path=%s][pseudo_fd=%llu]", path, (unsigned long long)(fi->fh));
|
|
|
|
// [NOTE]
|
|
// All opened file's stats is cached with no truncate flag.
|
|
// Thus we unset it here.
|
|
StatCache::getStatCacheData()->ChangeNoTruncateFlag(std::string(path), false);
|
|
|
|
// [NOTICE]
|
|
// At first, we remove stats cache.
|
|
// Because fuse does not wait for response from "release" function. :-(
|
|
// And fuse runs next command before this function returns.
|
|
// Thus we call deleting stats function ASAP.
|
|
//
|
|
if((fi->flags & O_RDWR) || (fi->flags & O_WRONLY)){
|
|
StatCache::getStatCacheData()->DelStat(path);
|
|
}
|
|
|
|
{ // scope for AutoFdEntity
|
|
AutoFdEntity autoent;
|
|
|
|
// [NOTE]
|
|
// The pseudo fd stored in fi->fh is attached to AutoFdEntry so that it can be
|
|
// destroyed here.
|
|
//
|
|
FdEntity* ent;
|
|
if(NULL == (ent = autoent.Attach(path, static_cast<int>(fi->fh)))){
|
|
S3FS_PRN_ERR("could not find pseudo_fd(%llu) for path(%s)", (unsigned long long)(fi->fh), path);
|
|
return -EIO;
|
|
}
|
|
|
|
bool is_new_file = ent->IsDirtyNewFile();
|
|
|
|
// TODO: correct locks held?
|
|
int result = ent->UploadPending(static_cast<int>(fi->fh), AutoLock::NONE);
|
|
if(0 != result){
|
|
S3FS_PRN_ERR("could not upload pending data(meta, etc) for pseudo_fd(%llu) / path(%s)", (unsigned long long)(fi->fh), path);
|
|
return result;
|
|
}
|
|
|
|
if(is_new_file){
|
|
// update parent directory timestamp
|
|
int update_result;
|
|
if(0 != (update_result = update_mctime_parent_directory(path))){
|
|
S3FS_PRN_ERR("succeed to create the file(%s), but could not update timestamp of its parent directory(result=%d).", path, update_result);
|
|
}
|
|
}
|
|
}
|
|
|
|
// check - for debug
|
|
if(S3fsLog::IsS3fsLogDbg()){
|
|
if(FdManager::HasOpenEntityFd(path)){
|
|
S3FS_PRN_WARN("file(%s) is still opened(another pseudo fd is opend).", path);
|
|
}
|
|
}
|
|
S3FS_MALLOCTRIM(0);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int s3fs_opendir(const char* _path, struct fuse_file_info* fi)
|
|
{
|
|
WTF8_ENCODE(path)
|
|
int result;
|
|
int mask = (O_RDONLY != (fi->flags & O_ACCMODE) ? W_OK : R_OK);
|
|
|
|
S3FS_PRN_INFO("[path=%s][flags=0x%x]", path, fi->flags);
|
|
|
|
if(0 == (result = check_object_access(path, mask, NULL))){
|
|
result = check_parent_object_access(path, X_OK);
|
|
}
|
|
S3FS_MALLOCTRIM(0);
|
|
|
|
return result;
|
|
}
|
|
|
|
struct multi_head_callback_param
|
|
{
|
|
void* buf;
|
|
fuse_fill_dir_t filler;
|
|
};
|
|
|
|
static bool multi_head_callback(S3fsCurl* s3fscurl, void* param)
|
|
{
|
|
if(!s3fscurl){
|
|
return false;
|
|
}
|
|
|
|
// Add stat cache
|
|
std::string saved_path = s3fscurl->GetSpecialSavedPath();
|
|
if(!StatCache::getStatCacheData()->AddStat(saved_path, *(s3fscurl->GetResponseHeaders()))){
|
|
S3FS_PRN_ERR("failed adding stat cache [path=%s]", saved_path.c_str());
|
|
return false;
|
|
}
|
|
|
|
// Get stats from stats cache(for converting from meta), and fill
|
|
std::string bpath = mybasename(saved_path);
|
|
if(use_wtf8){
|
|
bpath = s3fs_wtf8_decode(bpath);
|
|
}
|
|
if(param){
|
|
struct multi_head_callback_param* pcbparam = reinterpret_cast<struct multi_head_callback_param*>(param);
|
|
struct stat st;
|
|
if(StatCache::getStatCacheData()->GetStat(saved_path, &st)){
|
|
pcbparam->filler(pcbparam->buf, bpath.c_str(), &st, 0);
|
|
}else{
|
|
S3FS_PRN_INFO2("Could not find %s file in stat cache.", saved_path.c_str());
|
|
pcbparam->filler(pcbparam->buf, bpath.c_str(), 0, 0);
|
|
}
|
|
}else{
|
|
S3FS_PRN_WARN("param(multi_head_callback_param*) is NULL, then can not call filler.");
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
struct multi_head_notfound_callback_param
|
|
{
|
|
pthread_mutex_t list_lock;
|
|
s3obj_list_t notfound_list;
|
|
};
|
|
|
|
static bool multi_head_notfound_callback(S3fsCurl* s3fscurl, void* param)
|
|
{
|
|
if(!s3fscurl){
|
|
return false;
|
|
}
|
|
S3FS_PRN_INFO("HEAD returned NotFound(404) for %s object, it maybe only the path exists and the object does not exist.", s3fscurl->GetPath().c_str());
|
|
|
|
if(!param){
|
|
S3FS_PRN_WARN("param(multi_head_notfound_callback_param*) is NULL, then can not call filler.");
|
|
return false;
|
|
}
|
|
|
|
// set path to not found list
|
|
struct multi_head_notfound_callback_param* pcbparam = reinterpret_cast<struct multi_head_notfound_callback_param*>(param);
|
|
|
|
AutoLock auto_lock(&(pcbparam->list_lock));
|
|
pcbparam->notfound_list.push_back(s3fscurl->GetBasePath());
|
|
|
|
return true;
|
|
}
|
|
|
|
static S3fsCurl* multi_head_retry_callback(S3fsCurl* s3fscurl)
|
|
{
|
|
if(!s3fscurl){
|
|
return NULL;
|
|
}
|
|
size_t ssec_key_pos= s3fscurl->GetLastPreHeadSeecKeyPos();
|
|
int retry_count = s3fscurl->GetMultipartRetryCount();
|
|
|
|
// retry next sse key.
|
|
// if end of sse key, set retry master count is up.
|
|
ssec_key_pos = (ssec_key_pos == static_cast<size_t>(-1) ? 0 : ssec_key_pos + 1);
|
|
if(0 == S3fsCurl::GetSseKeyCount() || S3fsCurl::GetSseKeyCount() <= ssec_key_pos){
|
|
if(s3fscurl->IsOverMultipartRetryCount()){
|
|
S3FS_PRN_ERR("Over retry count(%d) limit(%s).", s3fscurl->GetMultipartRetryCount(), s3fscurl->GetSpecialSavedPath().c_str());
|
|
return NULL;
|
|
}
|
|
ssec_key_pos = -1;
|
|
retry_count++;
|
|
}
|
|
|
|
S3fsCurl* newcurl = new S3fsCurl(s3fscurl->IsUseAhbe());
|
|
std::string path = s3fscurl->GetPath();
|
|
std::string base_path = s3fscurl->GetBasePath();
|
|
std::string saved_path = s3fscurl->GetSpecialSavedPath();
|
|
|
|
if(!newcurl->PreHeadRequest(path, base_path, saved_path, ssec_key_pos)){
|
|
S3FS_PRN_ERR("Could not duplicate curl object(%s).", saved_path.c_str());
|
|
delete newcurl;
|
|
return NULL;
|
|
}
|
|
newcurl->SetMultipartRetryCount(retry_count);
|
|
|
|
return newcurl;
|
|
}
|
|
|
|
static int readdir_multi_head(const char* path, const S3ObjList& head, void* buf, fuse_fill_dir_t filler)
|
|
{
|
|
S3fsMultiCurl curlmulti(S3fsCurl::GetMaxMultiRequest());
|
|
s3obj_list_t headlist;
|
|
int result = 0;
|
|
|
|
S3FS_PRN_INFO1("[path=%s][list=%zu]", path, headlist.size());
|
|
|
|
// Make base path list.
|
|
head.GetNameList(headlist, true, false); // get name with "/".
|
|
|
|
// Initialize S3fsMultiCurl
|
|
curlmulti.SetSuccessCallback(multi_head_callback);
|
|
curlmulti.SetRetryCallback(multi_head_retry_callback);
|
|
|
|
// Success Callback function parameter
|
|
struct multi_head_callback_param success_param;
|
|
success_param.buf = buf;
|
|
success_param.filler = filler;
|
|
curlmulti.SetSuccessCallbackParam(reinterpret_cast<void*>(&success_param));
|
|
|
|
// Not found Callback function parameter
|
|
struct multi_head_notfound_callback_param notfound_param;
|
|
if(support_compat_dir){
|
|
pthread_mutexattr_t attr;
|
|
pthread_mutexattr_init(&attr);
|
|
#if S3FS_PTHREAD_ERRORCHECK
|
|
pthread_mutexattr_settype(&attr, PTHREAD_MUTEX_ERRORCHECK);
|
|
#endif
|
|
|
|
if(0 != (result = pthread_mutex_init(&(notfound_param.list_lock), &attr))){
|
|
S3FS_PRN_CRIT("failed to init notfound_param.list_lock: %d", result);
|
|
abort();
|
|
}
|
|
curlmulti.SetNotFoundCallback(multi_head_notfound_callback);
|
|
curlmulti.SetNotFoundCallbackParam(reinterpret_cast<void*>(¬found_param));
|
|
}
|
|
|
|
// Make single head request(with max).
|
|
for(s3obj_list_t::iterator iter = headlist.begin(); headlist.end() != iter; iter = headlist.erase(iter)){
|
|
std::string disppath = path + (*iter);
|
|
std::string etag = head.GetETag((*iter).c_str());
|
|
struct stat st;
|
|
|
|
// [NOTE]
|
|
// If there is a cache hit, file stat is filled by filler at here.
|
|
//
|
|
if(StatCache::getStatCacheData()->HasStat(disppath, &st, etag.c_str())){
|
|
std::string bpath = mybasename(disppath);
|
|
if(use_wtf8){
|
|
bpath = s3fs_wtf8_decode(bpath);
|
|
}
|
|
filler(buf, bpath.c_str(), &st, 0);
|
|
continue;
|
|
}
|
|
|
|
// First check for directory, start checking "not SSE-C".
|
|
// If checking failed, retry to check with "SSE-C" by retry callback func when SSE-C mode.
|
|
S3fsCurl* s3fscurl = new S3fsCurl();
|
|
if(!s3fscurl->PreHeadRequest(disppath, (*iter), disppath)){ // target path = cache key path.(ex "dir/")
|
|
S3FS_PRN_WARN("Could not make curl object for head request(%s).", disppath.c_str());
|
|
delete s3fscurl;
|
|
continue;
|
|
}
|
|
|
|
if(!curlmulti.SetS3fsCurlObject(s3fscurl)){
|
|
S3FS_PRN_WARN("Could not make curl object into multi curl(%s).", disppath.c_str());
|
|
delete s3fscurl;
|
|
continue;
|
|
}
|
|
}
|
|
|
|
// Multi request
|
|
if(0 != (result = curlmulti.Request())){
|
|
// If result is -EIO, it is something error occurred.
|
|
// This case includes that the object is encrypting(SSE) and s3fs does not have keys.
|
|
// So s3fs set result to 0 in order to continue the process.
|
|
if(-EIO == result){
|
|
S3FS_PRN_WARN("error occurred in multi request(errno=%d), but continue...", result);
|
|
result = 0;
|
|
}else{
|
|
S3FS_PRN_ERR("error occurred in multi request(errno=%d).", result);
|
|
return result;
|
|
}
|
|
}
|
|
|
|
// [NOTE]
|
|
// Objects that could not be found by HEAD request may exist only
|
|
// as a path, so search for objects under that path.(a case of no dir object)
|
|
//
|
|
if(support_compat_dir && !notfound_param.notfound_list.empty()){ // [NOTE] not need to lock to access this here.
|
|
// dummy header
|
|
mode_t dirmask = umask(0); // macos does not have getumask()
|
|
umask(dirmask);
|
|
|
|
headers_t dummy_header;
|
|
dummy_header["Content-Type"] = std::string("application/x-directory"); // directory
|
|
dummy_header["x-amz-meta-uid"] = str(is_s3fs_uid ? s3fs_uid : geteuid());
|
|
dummy_header["x-amz-meta-gid"] = str(is_s3fs_gid ? s3fs_gid : getegid());
|
|
dummy_header["x-amz-meta-mode"] = str(S_IFDIR | (~dirmask & (S_IRWXU | S_IRWXG | S_IRWXO)));
|
|
dummy_header["x-amz-meta-atime"] = str(0);
|
|
dummy_header["x-amz-meta-ctime"] = str(0);
|
|
dummy_header["x-amz-meta-mtime"] = str(0);
|
|
|
|
for(s3obj_list_t::iterator reiter = notfound_param.notfound_list.begin(); reiter != notfound_param.notfound_list.end(); ++reiter){
|
|
int dir_result;
|
|
if(0 == (dir_result = directory_empty(reiter->c_str()))){
|
|
// Found objects under the path, so the path is directory.
|
|
//
|
|
std::string dirpath = path + (*reiter);
|
|
|
|
// Add stat cache
|
|
if(StatCache::getStatCacheData()->AddStat(dirpath, dummy_header, true)){ // set forcedir=true
|
|
// Get stats from stats cache(for converting from meta), and fill
|
|
std::string base_path = mybasename(dirpath);
|
|
if(use_wtf8){
|
|
base_path = s3fs_wtf8_decode(base_path);
|
|
}
|
|
|
|
struct stat st;
|
|
if(StatCache::getStatCacheData()->GetStat(dirpath, &st)){
|
|
filler(buf, base_path.c_str(), &st, 0);
|
|
}else{
|
|
S3FS_PRN_INFO2("Could not find %s directory(no dir object) in stat cache.", dirpath.c_str());
|
|
filler(buf, base_path.c_str(), 0, 0);
|
|
}
|
|
}else{
|
|
S3FS_PRN_ERR("failed adding stat cache [path=%s], but dontinue...", dirpath.c_str());
|
|
}
|
|
}else{
|
|
S3FS_PRN_WARN("%s object does not have any object under it(errno=%d),", reiter->c_str(), dir_result);
|
|
}
|
|
}
|
|
}
|
|
|
|
return result;
|
|
}
|
|
|
|
static int s3fs_readdir(const char* _path, void* buf, fuse_fill_dir_t filler, off_t offset, struct fuse_file_info* fi)
|
|
{
|
|
WTF8_ENCODE(path)
|
|
S3ObjList head;
|
|
int result;
|
|
|
|
S3FS_PRN_INFO("[path=%s]", path);
|
|
|
|
if(0 != (result = check_object_access(path, R_OK, NULL))){
|
|
return result;
|
|
}
|
|
|
|
// get a list of all the objects
|
|
if((result = list_bucket(path, head, "/")) != 0){
|
|
S3FS_PRN_ERR("list_bucket returns error(%d).", result);
|
|
return result;
|
|
}
|
|
|
|
// force to add "." and ".." name.
|
|
filler(buf, ".", 0, 0);
|
|
filler(buf, "..", 0, 0);
|
|
if(head.IsEmpty()){
|
|
return 0;
|
|
}
|
|
|
|
// Send multi head request for stats caching.
|
|
std::string strpath = path;
|
|
if(strcmp(path, "/") != 0){
|
|
strpath += "/";
|
|
}
|
|
if(0 != (result = readdir_multi_head(strpath.c_str(), head, buf, filler))){
|
|
S3FS_PRN_ERR("readdir_multi_head returns error(%d).", result);
|
|
}
|
|
S3FS_MALLOCTRIM(0);
|
|
|
|
return result;
|
|
}
|
|
|
|
static int list_bucket(const char* path, S3ObjList& head, const char* delimiter, bool check_content_only)
|
|
{
|
|
std::string s3_realpath;
|
|
std::string query_delimiter;
|
|
std::string query_prefix;
|
|
std::string query_maxkey;
|
|
std::string next_continuation_token;
|
|
std::string next_marker;
|
|
bool truncated = true;
|
|
S3fsCurl s3fscurl;
|
|
xmlDocPtr doc;
|
|
|
|
S3FS_PRN_INFO1("[path=%s]", path);
|
|
|
|
if(delimiter && 0 < strlen(delimiter)){
|
|
query_delimiter += "delimiter=";
|
|
query_delimiter += delimiter;
|
|
query_delimiter += "&";
|
|
}
|
|
|
|
query_prefix += "&prefix=";
|
|
s3_realpath = get_realpath(path);
|
|
if(s3_realpath.empty() || '/' != *s3_realpath.rbegin()){
|
|
// last word must be "/"
|
|
query_prefix += urlEncode(s3_realpath.substr(1) + "/");
|
|
}else{
|
|
query_prefix += urlEncode(s3_realpath.substr(1));
|
|
}
|
|
if (check_content_only){
|
|
// Just need to know if there are child objects in dir
|
|
// For dir with children, expect "dir/" and "dir/child"
|
|
query_maxkey += "max-keys=2";
|
|
}else{
|
|
query_maxkey += "max-keys=" + str(max_keys_list_object);
|
|
}
|
|
|
|
while(truncated){
|
|
// append parameters to query in alphabetical order
|
|
std::string each_query;
|
|
if(!next_continuation_token.empty()){
|
|
each_query += "continuation-token=" + urlEncode(next_continuation_token) + "&";
|
|
next_continuation_token = "";
|
|
}
|
|
each_query += query_delimiter;
|
|
if(S3fsCurl::IsListObjectsV2()){
|
|
each_query += "list-type=2&";
|
|
}
|
|
if(!next_marker.empty()){
|
|
each_query += "marker=" + urlEncode(next_marker) + "&";
|
|
next_marker = "";
|
|
}
|
|
each_query += query_maxkey;
|
|
each_query += query_prefix;
|
|
|
|
// request
|
|
int result;
|
|
if(0 != (result = s3fscurl.ListBucketRequest(path, each_query.c_str()))){
|
|
S3FS_PRN_ERR("ListBucketRequest returns with error.");
|
|
return result;
|
|
}
|
|
const BodyData* body = s3fscurl.GetBodyData();
|
|
|
|
// xmlDocPtr
|
|
if(NULL == (doc = xmlReadMemory(body->str(), static_cast<int>(body->size()), "", NULL, 0))){
|
|
S3FS_PRN_ERR("xmlReadMemory returns with error.");
|
|
return -EIO;
|
|
}
|
|
if(0 != append_objects_from_xml(path, doc, head)){
|
|
S3FS_PRN_ERR("append_objects_from_xml returns with error.");
|
|
xmlFreeDoc(doc);
|
|
return -EIO;
|
|
}
|
|
if(true == (truncated = is_truncated(doc))){
|
|
xmlChar* tmpch;
|
|
if(NULL != (tmpch = get_next_continuation_token(doc))){
|
|
next_continuation_token = reinterpret_cast<char*>(tmpch);
|
|
xmlFree(tmpch);
|
|
}else if(NULL != (tmpch = get_next_marker(doc))){
|
|
next_marker = reinterpret_cast<char*>(tmpch);
|
|
xmlFree(tmpch);
|
|
}
|
|
|
|
if(next_continuation_token.empty() && next_marker.empty()){
|
|
// If did not specify "delimiter", s3 did not return "NextMarker".
|
|
// On this case, can use last name for next marker.
|
|
//
|
|
std::string lastname;
|
|
if(!head.GetLastName(lastname)){
|
|
S3FS_PRN_WARN("Could not find next marker, thus break loop.");
|
|
truncated = false;
|
|
}else{
|
|
next_marker = s3_realpath.substr(1);
|
|
if(s3_realpath.empty() || '/' != *s3_realpath.rbegin()){
|
|
next_marker += "/";
|
|
}
|
|
next_marker += lastname;
|
|
}
|
|
}
|
|
}
|
|
S3FS_XMLFREEDOC(doc);
|
|
|
|
// reset(initialize) curl object
|
|
s3fscurl.DestroyCurlHandle();
|
|
|
|
if(check_content_only){
|
|
break;
|
|
}
|
|
}
|
|
S3FS_MALLOCTRIM(0);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int remote_mountpath_exists(const char* path)
|
|
{
|
|
struct stat stbuf;
|
|
int result;
|
|
|
|
S3FS_PRN_INFO1("[path=%s]", path);
|
|
|
|
// getattr will prefix the path with the remote mountpoint
|
|
if(0 != (result = get_object_attribute("/", &stbuf, NULL))){
|
|
return result;
|
|
}
|
|
if(!S_ISDIR(stbuf.st_mode)){
|
|
return -ENOTDIR;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static bool get_meta_xattr_value(const char* path, std::string& rawvalue)
|
|
{
|
|
if(!path || '\0' == path[0]){
|
|
S3FS_PRN_ERR("path is empty.");
|
|
return false;
|
|
}
|
|
S3FS_PRN_DBG("[path=%s]", path);
|
|
|
|
rawvalue.erase();
|
|
|
|
headers_t meta;
|
|
if(0 != get_object_attribute(path, NULL, &meta)){
|
|
S3FS_PRN_ERR("Failed to get object(%s) headers", path);
|
|
return false;
|
|
}
|
|
|
|
headers_t::const_iterator iter;
|
|
if(meta.end() == (iter = meta.find("x-amz-meta-xattr"))){
|
|
return false;
|
|
}
|
|
rawvalue = iter->second;
|
|
return true;
|
|
}
|
|
|
|
static bool get_parent_meta_xattr_value(const char* path, std::string& rawvalue)
|
|
{
|
|
if(0 == strcmp(path, "/") || 0 == strcmp(path, ".")){
|
|
// path is mount point, thus does not have parent.
|
|
return false;
|
|
}
|
|
|
|
std::string parent = mydirname(path);
|
|
if(parent.empty()){
|
|
S3FS_PRN_ERR("Could not get parent path for %s.", path);
|
|
return false;
|
|
}
|
|
return get_meta_xattr_value(parent.c_str(), rawvalue);
|
|
}
|
|
|
|
static bool get_xattr_posix_key_value(const char* path, std::string& xattrvalue, bool default_key)
|
|
{
|
|
xattrvalue.erase();
|
|
|
|
std::string rawvalue;
|
|
if(!get_meta_xattr_value(path, rawvalue)){
|
|
return false;
|
|
}
|
|
|
|
xattrs_t xattrs;
|
|
if(0 == parse_xattrs(rawvalue, xattrs)){
|
|
return false;
|
|
}
|
|
|
|
std::string targetkey;
|
|
if(default_key){
|
|
targetkey = "system.posix_acl_default";
|
|
}else{
|
|
targetkey = "system.posix_acl_access";
|
|
}
|
|
|
|
xattrs_t::iterator iter;
|
|
if(xattrs.end() == (iter = xattrs.find(targetkey)) || !(iter->second)){
|
|
free_xattrs(xattrs);
|
|
return false;
|
|
}
|
|
|
|
// convert value by base64
|
|
char* base64val = s3fs_base64((iter->second)->pvalue, (iter->second)->length);
|
|
if(!base64val){
|
|
free_xattrs(xattrs);
|
|
return false;
|
|
}
|
|
free_xattrs(xattrs);
|
|
|
|
xattrvalue = base64val;
|
|
delete[] base64val;
|
|
|
|
return true;
|
|
}
|
|
|
|
// [NOTE]
|
|
// Converts and returns the POSIX ACL default(system.posix_acl_default) value of
|
|
// the parent directory as a POSIX ACL(system.posix_acl_access) value.
|
|
// Returns false if the parent directory has no POSIX ACL defaults.
|
|
//
|
|
static bool build_inherited_xattr_value(const char* path, std::string& xattrvalue)
|
|
{
|
|
S3FS_PRN_DBG("[path=%s]", path);
|
|
|
|
xattrvalue.erase();
|
|
|
|
if(0 == strcmp(path, "/") || 0 == strcmp(path, ".")){
|
|
// path is mount point, thus does not have parent.
|
|
return false;
|
|
}
|
|
|
|
std::string parent = mydirname(path);
|
|
if(parent.empty()){
|
|
S3FS_PRN_ERR("Could not get parent path for %s.", path);
|
|
return false;
|
|
}
|
|
|
|
// get parent's "system.posix_acl_default" value(base64'd).
|
|
std::string parent_default_value;
|
|
if(!get_xattr_posix_key_value(parent.c_str(), parent_default_value, true)){
|
|
return false;
|
|
}
|
|
|
|
// build "system.posix_acl_access" from parent's default value
|
|
std::string raw_xattr_value;
|
|
raw_xattr_value = "{\"system.posix_acl_access\":\"";
|
|
raw_xattr_value += parent_default_value;
|
|
raw_xattr_value += "\"}";
|
|
|
|
xattrvalue = urlEncode(raw_xattr_value);
|
|
return true;
|
|
}
|
|
|
|
static void free_xattrs(xattrs_t& xattrs)
|
|
{
|
|
for(xattrs_t::iterator iter = xattrs.begin(); iter != xattrs.end(); ++iter){
|
|
delete iter->second;
|
|
}
|
|
xattrs.clear();
|
|
}
|
|
|
|
static bool parse_xattr_keyval(const std::string& xattrpair, std::string& key, PXATTRVAL& pval)
|
|
{
|
|
// parse key and value
|
|
size_t pos;
|
|
std::string tmpval;
|
|
if(std::string::npos == (pos = xattrpair.find_first_of(':'))){
|
|
S3FS_PRN_ERR("one of xattr pair(%s) is wrong format.", xattrpair.c_str());
|
|
return false;
|
|
}
|
|
key = xattrpair.substr(0, pos);
|
|
tmpval = xattrpair.substr(pos + 1);
|
|
|
|
if(!takeout_str_dquart(key) || !takeout_str_dquart(tmpval)){
|
|
S3FS_PRN_ERR("one of xattr pair(%s) is wrong format.", xattrpair.c_str());
|
|
return false;
|
|
}
|
|
|
|
pval = new XATTRVAL;
|
|
pval->length = 0;
|
|
pval->pvalue = s3fs_decode64(tmpval.c_str(), tmpval.size(), &pval->length);
|
|
|
|
return true;
|
|
}
|
|
|
|
static size_t parse_xattrs(const std::string& strxattrs, xattrs_t& xattrs)
|
|
{
|
|
xattrs.clear();
|
|
|
|
// decode
|
|
std::string jsonxattrs = urlDecode(strxattrs);
|
|
|
|
// get from "{" to "}"
|
|
std::string restxattrs;
|
|
{
|
|
size_t startpos;
|
|
size_t endpos = std::string::npos;
|
|
if(std::string::npos != (startpos = jsonxattrs.find_first_of('{'))){
|
|
endpos = jsonxattrs.find_last_of('}');
|
|
}
|
|
if(startpos == std::string::npos || endpos == std::string::npos || endpos <= startpos){
|
|
S3FS_PRN_WARN("xattr header(%s) is not json format.", jsonxattrs.c_str());
|
|
return 0;
|
|
}
|
|
restxattrs = jsonxattrs.substr(startpos + 1, endpos - (startpos + 1));
|
|
}
|
|
|
|
// parse each key:val
|
|
for(size_t pair_nextpos = restxattrs.find_first_of(','); !restxattrs.empty(); restxattrs = (pair_nextpos != std::string::npos ? restxattrs.substr(pair_nextpos + 1) : std::string("")), pair_nextpos = restxattrs.find_first_of(',')){
|
|
std::string pair = pair_nextpos != std::string::npos ? restxattrs.substr(0, pair_nextpos) : restxattrs;
|
|
std::string key;
|
|
PXATTRVAL pval = NULL;
|
|
if(!parse_xattr_keyval(pair, key, pval)){
|
|
// something format error, so skip this.
|
|
continue;
|
|
}
|
|
xattrs[key] = pval;
|
|
}
|
|
return xattrs.size();
|
|
}
|
|
|
|
static std::string raw_build_xattrs(const xattrs_t& xattrs)
|
|
{
|
|
std::string strxattrs;
|
|
bool is_set = false;
|
|
for(xattrs_t::const_iterator iter = xattrs.begin(); iter != xattrs.end(); ++iter){
|
|
if(is_set){
|
|
strxattrs += ',';
|
|
}else{
|
|
is_set = true;
|
|
strxattrs = "{";
|
|
}
|
|
strxattrs += '\"';
|
|
strxattrs += iter->first;
|
|
strxattrs += "\":\"";
|
|
|
|
if(iter->second){
|
|
char* base64val = s3fs_base64((iter->second)->pvalue, (iter->second)->length);
|
|
if(base64val){
|
|
strxattrs += base64val;
|
|
delete[] base64val;
|
|
}
|
|
}
|
|
strxattrs += '\"';
|
|
}
|
|
if(is_set){
|
|
strxattrs += "}";
|
|
}
|
|
return strxattrs;
|
|
}
|
|
|
|
static std::string build_xattrs(const xattrs_t& xattrs)
|
|
{
|
|
std::string strxattrs = raw_build_xattrs(xattrs);
|
|
if(strxattrs.empty()){
|
|
strxattrs = "{}";
|
|
}
|
|
strxattrs = urlEncode(strxattrs);
|
|
|
|
return strxattrs;
|
|
}
|
|
|
|
static int set_xattrs_to_header(headers_t& meta, const char* name, const char* value, size_t size, int flags)
|
|
{
|
|
std::string strxattrs;
|
|
xattrs_t xattrs;
|
|
|
|
headers_t::iterator iter;
|
|
if(meta.end() == (iter = meta.find("x-amz-meta-xattr"))){
|
|
#if defined(XATTR_REPLACE)
|
|
if(XATTR_REPLACE == (flags & XATTR_REPLACE)){
|
|
// there is no xattr header but flags is replace, so failure.
|
|
return -ENOATTR;
|
|
}
|
|
#endif
|
|
}else{
|
|
#if defined(XATTR_CREATE)
|
|
if(XATTR_CREATE == (flags & XATTR_CREATE)){
|
|
// found xattr header but flags is only creating, so failure.
|
|
return -EEXIST;
|
|
}
|
|
#endif
|
|
strxattrs = iter->second;
|
|
}
|
|
|
|
// get map as xattrs_t
|
|
parse_xattrs(strxattrs, xattrs);
|
|
|
|
// add name(do not care overwrite and empty name/value)
|
|
xattrs_t::iterator xiter;
|
|
if(xattrs.end() != (xiter = xattrs.find(std::string(name)))){
|
|
// found same head. free value.
|
|
delete xiter->second;
|
|
}
|
|
|
|
PXATTRVAL pval = new XATTRVAL;
|
|
pval->length = size;
|
|
if(0 < size){
|
|
pval->pvalue = new unsigned char[size];
|
|
memcpy(pval->pvalue, value, size);
|
|
}else{
|
|
pval->pvalue = NULL;
|
|
}
|
|
xattrs[std::string(name)] = pval;
|
|
|
|
// build new strxattrs(not encoded) and set it to headers_t
|
|
meta["x-amz-meta-xattr"] = build_xattrs(xattrs);
|
|
|
|
S3FS_PRN_DBG("Set xattrs(after adding %s key) = %s", name, raw_build_xattrs(xattrs).c_str());
|
|
|
|
free_xattrs(xattrs);
|
|
|
|
return 0;
|
|
}
|
|
|
|
#if defined(__APPLE__)
|
|
static int s3fs_setxattr(const char* path, const char* name, const char* value, size_t size, int flags, uint32_t position)
|
|
#else
|
|
static int s3fs_setxattr(const char* path, const char* name, const char* value, size_t size, int flags)
|
|
#endif
|
|
{
|
|
S3FS_PRN_INFO("[path=%s][name=%s][value=%p][size=%zu][flags=0x%x]", path, name, value, size, flags);
|
|
|
|
if((value && 0 == size) || (!value && 0 < size)){
|
|
S3FS_PRN_ERR("Wrong parameter: value(%p), size(%zu)", value, size);
|
|
return 0;
|
|
}
|
|
|
|
#if defined(__APPLE__)
|
|
if (position != 0) {
|
|
// No resource fork support
|
|
return -EINVAL;
|
|
}
|
|
#endif
|
|
|
|
int result;
|
|
std::string strpath;
|
|
std::string newpath;
|
|
std::string nowcache;
|
|
headers_t meta;
|
|
struct stat stbuf;
|
|
dirtype nDirType = DIRTYPE_UNKNOWN;
|
|
|
|
if(0 != (result = check_parent_object_access(path, X_OK))){
|
|
return result;
|
|
}
|
|
if(0 != (result = check_object_owner(path, &stbuf))){
|
|
return result;
|
|
}
|
|
|
|
if(S_ISDIR(stbuf.st_mode)){
|
|
result = chk_dir_object_type(path, newpath, strpath, nowcache, &meta, &nDirType);
|
|
}else{
|
|
strpath = path;
|
|
nowcache = strpath;
|
|
result = get_object_attribute(strpath.c_str(), NULL, &meta);
|
|
}
|
|
if(0 != result){
|
|
return result;
|
|
}
|
|
|
|
if(S_ISDIR(stbuf.st_mode) && (IS_REPLACEDIR(nDirType) || IS_CREATE_MP_STAT(path))){
|
|
if(IS_REPLACEDIR(nDirType)){
|
|
// Should rebuild directory object(except new type)
|
|
// Need to remove old dir("dir" etc) and make new dir("dir/")
|
|
|
|
// At first, remove directory old object
|
|
if(0 != (result = remove_old_type_dir(strpath, nDirType))){
|
|
return result;
|
|
}
|
|
}
|
|
StatCache::getStatCacheData()->DelStat(nowcache);
|
|
|
|
// Make new directory object("dir/")
|
|
struct timespec ts_atime;
|
|
struct timespec ts_mtime;
|
|
struct timespec ts_ctime;
|
|
set_stat_to_timespec(stbuf, ST_TYPE_ATIME, ts_atime);
|
|
set_stat_to_timespec(stbuf, ST_TYPE_MTIME, ts_mtime);
|
|
set_stat_to_timespec(stbuf, ST_TYPE_CTIME, ts_ctime);
|
|
|
|
if(0 != (result = create_directory_object(newpath.c_str(), stbuf.st_mode, ts_atime, ts_mtime, ts_ctime, stbuf.st_uid, stbuf.st_gid, NULL))){
|
|
return result;
|
|
}
|
|
|
|
// need to set xattr header for directory.
|
|
strpath = newpath;
|
|
nowcache = strpath;
|
|
}
|
|
|
|
// set xattr all object
|
|
std::string strSourcePath = (mount_prefix.empty() && "/" == strpath) ? "//" : strpath;
|
|
headers_t updatemeta;
|
|
updatemeta["x-amz-meta-ctime"] = s3fs_str_realtime();
|
|
updatemeta["x-amz-copy-source"] = urlEncode(service_path + S3fsCred::GetBucket() + get_realpath(strSourcePath.c_str()));
|
|
updatemeta["x-amz-metadata-directive"] = "REPLACE";
|
|
|
|
// check opened file handle.
|
|
//
|
|
// If the file starts uploading by multipart when the disk capacity is insufficient,
|
|
// we need to put these header after finishing upload.
|
|
// Or if the file is only open, we must update to FdEntity's internal meta.
|
|
//
|
|
AutoFdEntity autoent;
|
|
FdEntity* ent;
|
|
bool need_put_header = true;
|
|
if(NULL != (ent = autoent.OpenExistFdEntity(path))){
|
|
// get xattr and make new xattr
|
|
std::string strxattr;
|
|
if(ent->GetXattr(strxattr)){
|
|
updatemeta["x-amz-meta-xattr"] = strxattr;
|
|
}else{
|
|
// [NOTE]
|
|
// Set an empty xattr.
|
|
// This requires the key to be present in order to add xattr.
|
|
ent->SetXattr(strxattr);
|
|
}
|
|
if(0 != (result = set_xattrs_to_header(updatemeta, name, value, size, flags))){
|
|
return result;
|
|
}
|
|
|
|
if(ent->MergeOrgMeta(updatemeta)){
|
|
// meta is changed, but now uploading.
|
|
// then the meta is pending and accumulated to be put after the upload is complete.
|
|
S3FS_PRN_INFO("meta pending until upload is complete");
|
|
need_put_header = false;
|
|
|
|
// If there is data in the Stats cache, update the Stats cache.
|
|
StatCache::getStatCacheData()->UpdateMetaStats(strpath, updatemeta);
|
|
}
|
|
}
|
|
if(need_put_header){
|
|
// not found opened file.
|
|
if(0 != (result = set_xattrs_to_header(meta, name, value, size, flags))){
|
|
return result;
|
|
}
|
|
merge_headers(meta, updatemeta, true);
|
|
|
|
// upload meta directly.
|
|
if(0 != (result = put_headers(strpath.c_str(), meta, true))){
|
|
return result;
|
|
}
|
|
StatCache::getStatCacheData()->DelStat(nowcache);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
#if defined(__APPLE__)
|
|
static int s3fs_getxattr(const char* path, const char* name, char* value, size_t size, uint32_t position)
|
|
#else
|
|
static int s3fs_getxattr(const char* path, const char* name, char* value, size_t size)
|
|
#endif
|
|
{
|
|
S3FS_PRN_INFO("[path=%s][name=%s][value=%p][size=%zu]", path, name, value, size);
|
|
|
|
if(!path || !name){
|
|
return -EIO;
|
|
}
|
|
|
|
#if defined(__APPLE__)
|
|
if (position != 0) {
|
|
// No resource fork support
|
|
return -EINVAL;
|
|
}
|
|
#endif
|
|
|
|
int result;
|
|
headers_t meta;
|
|
xattrs_t xattrs;
|
|
|
|
// check parent directory attribute.
|
|
if(0 != (result = check_parent_object_access(path, X_OK))){
|
|
return result;
|
|
}
|
|
|
|
// get headers
|
|
if(0 != (result = get_object_attribute(path, NULL, &meta))){
|
|
return result;
|
|
}
|
|
|
|
// get xattrs
|
|
headers_t::iterator hiter = meta.find("x-amz-meta-xattr");
|
|
if(meta.end() == hiter){
|
|
// object does not have xattrs
|
|
return -ENOATTR;
|
|
}
|
|
std::string strxattrs = hiter->second;
|
|
|
|
parse_xattrs(strxattrs, xattrs);
|
|
|
|
S3FS_PRN_DBG("Get xattrs = %s", raw_build_xattrs(xattrs).c_str());
|
|
|
|
// search name
|
|
std::string strname = name;
|
|
xattrs_t::iterator xiter = xattrs.find(strname);
|
|
if(xattrs.end() == xiter){
|
|
// not found name in xattrs
|
|
free_xattrs(xattrs);
|
|
return -ENOATTR;
|
|
}
|
|
|
|
// decode
|
|
size_t length = 0;
|
|
unsigned char* pvalue = NULL;
|
|
if(NULL != xiter->second){
|
|
length = xiter->second->length;
|
|
pvalue = xiter->second->pvalue;
|
|
}
|
|
|
|
if(0 < size){
|
|
if(static_cast<size_t>(size) < length){
|
|
// over buffer size
|
|
free_xattrs(xattrs);
|
|
return -ERANGE;
|
|
}
|
|
if(pvalue){
|
|
memcpy(value, pvalue, length);
|
|
}
|
|
}
|
|
free_xattrs(xattrs);
|
|
|
|
return static_cast<int>(length);
|
|
}
|
|
|
|
static int s3fs_listxattr(const char* path, char* list, size_t size)
|
|
{
|
|
S3FS_PRN_INFO("[path=%s][list=%p][size=%zu]", path, list, size);
|
|
|
|
if(!path){
|
|
return -EIO;
|
|
}
|
|
|
|
int result;
|
|
headers_t meta;
|
|
xattrs_t xattrs;
|
|
|
|
// check parent directory attribute.
|
|
if(0 != (result = check_parent_object_access(path, X_OK))){
|
|
return result;
|
|
}
|
|
|
|
// get headers
|
|
if(0 != (result = get_object_attribute(path, NULL, &meta))){
|
|
return result;
|
|
}
|
|
|
|
// get xattrs
|
|
headers_t::iterator iter;
|
|
if(meta.end() == (iter = meta.find("x-amz-meta-xattr"))){
|
|
// object does not have xattrs
|
|
return 0;
|
|
}
|
|
std::string strxattrs = iter->second;
|
|
|
|
parse_xattrs(strxattrs, xattrs);
|
|
|
|
S3FS_PRN_DBG("Get xattrs = %s", raw_build_xattrs(xattrs).c_str());
|
|
|
|
// calculate total name length
|
|
size_t total = 0;
|
|
for(xattrs_t::const_iterator xiter = xattrs.begin(); xiter != xattrs.end(); ++xiter){
|
|
if(!xiter->first.empty()){
|
|
total += xiter->first.length() + 1;
|
|
}
|
|
}
|
|
|
|
if(0 == total){
|
|
free_xattrs(xattrs);
|
|
return 0;
|
|
}
|
|
|
|
// check parameters
|
|
if(0 == size){
|
|
free_xattrs(xattrs);
|
|
return static_cast<int>(total);
|
|
}
|
|
if(!list || size < total){
|
|
free_xattrs(xattrs);
|
|
return -ERANGE;
|
|
}
|
|
|
|
// copy to list
|
|
char* setpos = list;
|
|
for(xattrs_t::const_iterator xiter = xattrs.begin(); xiter != xattrs.end(); ++xiter){
|
|
if(!xiter->first.empty()){
|
|
strcpy(setpos, xiter->first.c_str());
|
|
setpos = &setpos[strlen(setpos) + 1];
|
|
}
|
|
}
|
|
free_xattrs(xattrs);
|
|
|
|
return static_cast<int>(total);
|
|
}
|
|
|
|
static int s3fs_removexattr(const char* path, const char* name)
|
|
{
|
|
S3FS_PRN_INFO("[path=%s][name=%s]", path, name);
|
|
|
|
if(!path || !name){
|
|
return -EIO;
|
|
}
|
|
|
|
int result;
|
|
std::string strpath;
|
|
std::string newpath;
|
|
std::string nowcache;
|
|
headers_t meta;
|
|
xattrs_t xattrs;
|
|
struct stat stbuf;
|
|
dirtype nDirType = DIRTYPE_UNKNOWN;
|
|
|
|
if(0 == strcmp(path, "/")){
|
|
S3FS_PRN_ERR("Could not change mode for mount point.");
|
|
return -EIO;
|
|
}
|
|
if(0 != (result = check_parent_object_access(path, X_OK))){
|
|
return result;
|
|
}
|
|
if(0 != (result = check_object_owner(path, &stbuf))){
|
|
return result;
|
|
}
|
|
|
|
if(S_ISDIR(stbuf.st_mode)){
|
|
result = chk_dir_object_type(path, newpath, strpath, nowcache, &meta, &nDirType);
|
|
}else{
|
|
strpath = path;
|
|
nowcache = strpath;
|
|
result = get_object_attribute(strpath.c_str(), NULL, &meta);
|
|
}
|
|
if(0 != result){
|
|
return result;
|
|
}
|
|
|
|
// get xattrs
|
|
headers_t::iterator hiter = meta.find("x-amz-meta-xattr");
|
|
if(meta.end() == hiter){
|
|
// object does not have xattrs
|
|
return -ENOATTR;
|
|
}
|
|
std::string strxattrs = hiter->second;
|
|
|
|
parse_xattrs(strxattrs, xattrs);
|
|
|
|
// check name xattrs
|
|
std::string strname = name;
|
|
xattrs_t::iterator xiter = xattrs.find(strname);
|
|
if(xattrs.end() == xiter){
|
|
free_xattrs(xattrs);
|
|
return -ENOATTR;
|
|
}
|
|
|
|
// make new header_t after deleting name xattr
|
|
delete xiter->second;
|
|
xattrs.erase(xiter);
|
|
|
|
S3FS_PRN_DBG("Reset xattrs(after delete %s key) = %s", name, raw_build_xattrs(xattrs).c_str());
|
|
|
|
if(S_ISDIR(stbuf.st_mode) && IS_REPLACEDIR(nDirType)){
|
|
// Should rebuild directory object(except new type)
|
|
// Need to remove old dir("dir" etc) and make new dir("dir/")
|
|
|
|
// At first, remove directory old object
|
|
if(0 != (result = remove_old_type_dir(strpath, nDirType))){
|
|
return result;
|
|
}
|
|
StatCache::getStatCacheData()->DelStat(nowcache);
|
|
|
|
// Make new directory object("dir/")
|
|
struct timespec ts_atime;
|
|
struct timespec ts_mtime;
|
|
struct timespec ts_ctime;
|
|
set_stat_to_timespec(stbuf, ST_TYPE_ATIME, ts_atime);
|
|
set_stat_to_timespec(stbuf, ST_TYPE_MTIME, ts_mtime);
|
|
set_stat_to_timespec(stbuf, ST_TYPE_CTIME, ts_ctime);
|
|
|
|
if(0 != (result = create_directory_object(newpath.c_str(), stbuf.st_mode, ts_atime, ts_mtime, ts_ctime, stbuf.st_uid, stbuf.st_gid, NULL))){
|
|
free_xattrs(xattrs);
|
|
return result;
|
|
}
|
|
|
|
// need to set xattr header for directory.
|
|
strpath = newpath;
|
|
nowcache = strpath;
|
|
}
|
|
|
|
// set xattr all object
|
|
std::string strSourcePath = (mount_prefix.empty() && "/" == strpath) ? "//" : strpath;
|
|
headers_t updatemeta;
|
|
updatemeta["x-amz-copy-source"] = urlEncode(service_path + S3fsCred::GetBucket() + get_realpath(strSourcePath.c_str()));
|
|
updatemeta["x-amz-metadata-directive"] = "REPLACE";
|
|
if(!xattrs.empty()){
|
|
updatemeta["x-amz-meta-xattr"] = build_xattrs(xattrs);
|
|
}else{
|
|
updatemeta["x-amz-meta-xattr"] = std::string(""); // This is a special case. If empty, this header will eventually be removed.
|
|
}
|
|
free_xattrs(xattrs);
|
|
|
|
// check opened file handle.
|
|
//
|
|
// If the file starts uploading by multipart when the disk capacity is insufficient,
|
|
// we need to put these header after finishing upload.
|
|
// Or if the file is only open, we must update to FdEntity's internal meta.
|
|
//
|
|
AutoFdEntity autoent;
|
|
FdEntity* ent;
|
|
bool need_put_header = true;
|
|
if(NULL != (ent = autoent.OpenExistFdEntity(path))){
|
|
if(ent->MergeOrgMeta(updatemeta)){
|
|
// meta is changed, but now uploading.
|
|
// then the meta is pending and accumulated to be put after the upload is complete.
|
|
S3FS_PRN_INFO("meta pending until upload is complete");
|
|
need_put_header = false;
|
|
|
|
// If there is data in the Stats cache, update the Stats cache.
|
|
StatCache::getStatCacheData()->UpdateMetaStats(strpath, updatemeta);
|
|
}
|
|
}
|
|
if(need_put_header){
|
|
// not found opened file.
|
|
if(updatemeta["x-amz-meta-xattr"].empty()){
|
|
updatemeta.erase("x-amz-meta-xattr");
|
|
}
|
|
|
|
merge_headers(meta, updatemeta, true);
|
|
|
|
// upload meta directly.
|
|
if(0 != (result = put_headers(strpath.c_str(), meta, true))){
|
|
return result;
|
|
}
|
|
StatCache::getStatCacheData()->DelStat(nowcache);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
// s3fs_init calls this function to exit cleanly from the fuse event loop.
|
|
//
|
|
// There's no way to pass an exit status to the high-level event loop API, so
|
|
// this function stores the exit value in a global for main()
|
|
static void s3fs_exit_fuseloop(int exit_status)
|
|
{
|
|
S3FS_PRN_ERR("Exiting FUSE event loop due to errors\n");
|
|
s3fs_init_deferred_exit_status = exit_status;
|
|
struct fuse_context *ctx = fuse_get_context();
|
|
if (NULL != ctx) {
|
|
fuse_exit(ctx->fuse);
|
|
}
|
|
}
|
|
|
|
static void* s3fs_init(struct fuse_conn_info* conn)
|
|
{
|
|
S3FS_PRN_INIT_INFO("init v%s(commit:%s) with %s, credential-library(%s)", VERSION, COMMIT_HASH_VAL, s3fs_crypt_lib_name(), ps3fscred->GetCredFuncVersion(false));
|
|
|
|
// cache(remove cache dirs at first)
|
|
if(is_remove_cache && (!CacheFileStat::DeleteCacheFileStatDirectory() || !FdManager::DeleteCacheDirectory())){
|
|
S3FS_PRN_DBG("Could not initialize cache directory.");
|
|
}
|
|
|
|
// check loading IAM role name
|
|
if(!ps3fscred->LoadIAMRoleFromMetaData()){
|
|
S3FS_PRN_CRIT("could not load IAM role name from meta data.");
|
|
s3fs_exit_fuseloop(EXIT_FAILURE);
|
|
return NULL;
|
|
}
|
|
|
|
// Check Bucket
|
|
{
|
|
int result;
|
|
if(EXIT_SUCCESS != (result = s3fs_check_service())){
|
|
s3fs_exit_fuseloop(result);
|
|
return NULL;
|
|
}
|
|
}
|
|
|
|
// Investigate system capabilities
|
|
#ifndef __APPLE__
|
|
if((unsigned int)conn->capable & FUSE_CAP_ATOMIC_O_TRUNC){
|
|
conn->want |= FUSE_CAP_ATOMIC_O_TRUNC;
|
|
}
|
|
#endif
|
|
|
|
if((unsigned int)conn->capable & FUSE_CAP_BIG_WRITES){
|
|
conn->want |= FUSE_CAP_BIG_WRITES;
|
|
}
|
|
|
|
// Signal object
|
|
if(!S3fsSignals::Initialize()){
|
|
S3FS_PRN_ERR("Failed to initialize signal object, but continue...");
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static void s3fs_destroy(void*)
|
|
{
|
|
S3FS_PRN_INFO("destroy");
|
|
|
|
// Signal object
|
|
if(!S3fsSignals::Destroy()){
|
|
S3FS_PRN_WARN("Failed to clean up signal object.");
|
|
}
|
|
|
|
// cache(remove at last)
|
|
if(is_remove_cache && (!CacheFileStat::DeleteCacheFileStatDirectory() || !FdManager::DeleteCacheDirectory())){
|
|
S3FS_PRN_WARN("Could not remove cache directory.");
|
|
}
|
|
}
|
|
|
|
static int s3fs_access(const char* path, int mask)
|
|
{
|
|
S3FS_PRN_INFO("[path=%s][mask=%s%s%s%s]", path,
|
|
((mask & R_OK) == R_OK) ? "R_OK " : "",
|
|
((mask & W_OK) == W_OK) ? "W_OK " : "",
|
|
((mask & X_OK) == X_OK) ? "X_OK " : "",
|
|
(mask == F_OK) ? "F_OK" : "");
|
|
|
|
int result = check_object_access(path, mask, NULL);
|
|
S3FS_MALLOCTRIM(0);
|
|
return result;
|
|
}
|
|
|
|
//
|
|
// If calling with wrong region, s3fs gets following error body as 400 error code.
|
|
// "<Error>
|
|
// <Code>AuthorizationHeaderMalformed</Code>
|
|
// <Message>The authorization header is malformed; the region 'us-east-1' is wrong; expecting 'ap-northeast-1'</Message>
|
|
// <Region>ap-northeast-1</Region>
|
|
// <RequestId>...</RequestId>
|
|
// <HostId>...</HostId>
|
|
// </Error>"
|
|
//
|
|
// So this is cheap code but s3fs should get correct region automatically.
|
|
//
|
|
static bool check_region_error(const char* pbody, size_t len, std::string& expectregion)
|
|
{
|
|
if(!pbody){
|
|
return false;
|
|
}
|
|
|
|
std::string code;
|
|
if(!simple_parse_xml(pbody, len, "Code", code) || code != "AuthorizationHeaderMalformed"){
|
|
return false;
|
|
}
|
|
|
|
if(!simple_parse_xml(pbody, len, "Region", expectregion)){
|
|
return false;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
static bool check_endpoint_error(const char* pbody, size_t len, std::string& expectendpoint)
|
|
{
|
|
if(!pbody){
|
|
return false;
|
|
}
|
|
|
|
std::string code;
|
|
if(!simple_parse_xml(pbody, len, "Code", code) || code != "PermanentRedirect"){
|
|
return false;
|
|
}
|
|
|
|
if(!simple_parse_xml(pbody, len, "Endpoint", expectendpoint)){
|
|
return false;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
// [NOTE]
|
|
// This checks whether access to the bucket when s3fs is started.
|
|
//
|
|
// The following patterns for mount point are supported by s3fs:
|
|
// (1) Mount the bucket top
|
|
// (2) Mount the directory(folder) under the bucket. In this case, there are
|
|
// the following cases:
|
|
// (2A) Directories created by clients other than s3fs
|
|
// (2B) Directory created by s3fs
|
|
//
|
|
// At first in this functoin, if user has access to the bucket, the checking
|
|
// access to the bucket succeeds and this function returns success. However,
|
|
// if user does not have access to the bucket and has permissions to the
|
|
// directory, this first check will fail.
|
|
// But if user specifies the directory for mount point, this function retries
|
|
// to check with the path containing the directory. And it will be success.
|
|
//
|
|
// In the case of (2A), the check will succeed if the bucket allows to access,
|
|
// but will fail if permissions are granted only to the directory, as it is not
|
|
// a directory recognized by s3fs. This combination is not supported by s3fs,
|
|
// so make sure user create the directory before starting s3fs.
|
|
// In case (2B), if user does not have access to bucket, the first check(to
|
|
// bucket) fails, but the retry check(with path) succeeds.
|
|
//
|
|
static int s3fs_check_service()
|
|
{
|
|
S3FS_PRN_INFO("check services.");
|
|
|
|
// At first time for access S3, we check IAM role if it sets.
|
|
if(!ps3fscred->CheckIAMCredentialUpdate()){
|
|
S3FS_PRN_CRIT("Failed to initialize IAM credential.");
|
|
return EXIT_FAILURE;
|
|
}
|
|
|
|
S3fsCurl s3fscurl;
|
|
int res;
|
|
if(0 > (res = s3fscurl.CheckBucket("/"))){
|
|
// get response code
|
|
long responseCode = s3fscurl.GetLastResponseCode();
|
|
bool changed_endpoint = false;
|
|
|
|
// check wrong endpoint, and automatically switch endpoint
|
|
if(300 <= responseCode && responseCode < 500){
|
|
|
|
// check region error(for putting message or retrying)
|
|
const BodyData* body = s3fscurl.GetBodyData();
|
|
std::string expectregion;
|
|
std::string expectendpoint;
|
|
if(check_region_error(body->str(), body->size(), expectregion)){
|
|
// [NOTE]
|
|
// If endpoint is not specified(using us-east-1 region) and
|
|
// an error is encountered accessing a different region, we
|
|
// will retry the check on the expected region.
|
|
// see) https://docs.aws.amazon.com/AmazonS3/latest/dev/UsingBucket.html#access-bucket-intro
|
|
//
|
|
if(is_specified_endpoint){
|
|
const char* tmp_expect_ep = expectregion.c_str();
|
|
S3FS_PRN_CRIT("The bucket region is not '%s', it is correctly '%s'. You should specify 'endpoint=%s' option.", endpoint.c_str(), tmp_expect_ep, tmp_expect_ep);
|
|
|
|
}else{
|
|
// current endpoint is wrong, so try to connect to expected region.
|
|
S3FS_PRN_CRIT("Failed to connect region '%s'(default), so retry to connect region '%s'.", endpoint.c_str(), expectregion.c_str());
|
|
endpoint = expectregion;
|
|
changed_endpoint = true;
|
|
|
|
if(S3fsCurl::GetSignatureType() == V4_ONLY ||
|
|
S3fsCurl::GetSignatureType() == V2_OR_V4){
|
|
if(s3host == "http://s3.amazonaws.com"){
|
|
s3host = "http://s3-" + endpoint + ".amazonaws.com";
|
|
}else if(s3host == "https://s3.amazonaws.com"){
|
|
s3host = "https://s3-" + endpoint + ".amazonaws.com";
|
|
}
|
|
}
|
|
}
|
|
}else if(check_endpoint_error(body->str(), body->size(), expectendpoint)){
|
|
S3FS_PRN_ERR("S3 service returned PermanentRedirect with endpoint: %s", expectendpoint.c_str());
|
|
return EXIT_FAILURE;
|
|
}
|
|
}
|
|
|
|
// retry to check with new endpoint
|
|
if(changed_endpoint){
|
|
s3fscurl.DestroyCurlHandle();
|
|
res = s3fscurl.CheckBucket("/");
|
|
responseCode = s3fscurl.GetLastResponseCode();
|
|
}
|
|
|
|
// retry to check with mount prefix
|
|
if(300 <= responseCode && responseCode < 500 && !mount_prefix.empty()){
|
|
s3fscurl.DestroyCurlHandle();
|
|
res = s3fscurl.CheckBucket(get_realpath("/").c_str());
|
|
responseCode = s3fscurl.GetLastResponseCode();
|
|
}
|
|
|
|
// try signature v2
|
|
if(0 > res && (responseCode == 400 || responseCode == 403) && S3fsCurl::GetSignatureType() == V2_OR_V4){
|
|
// switch sigv2
|
|
S3FS_PRN_CRIT("Failed to connect by sigv4, so retry to connect by signature version 2.");
|
|
S3fsCurl::SetSignatureType(V2_ONLY);
|
|
|
|
// retry to check with sigv2
|
|
s3fscurl.DestroyCurlHandle();
|
|
res = s3fscurl.CheckBucket("/");
|
|
responseCode = s3fscurl.GetLastResponseCode();
|
|
|
|
// retry to check with mount prefix
|
|
if(300 <= responseCode && responseCode < 500 && !mount_prefix.empty()){
|
|
s3fscurl.DestroyCurlHandle();
|
|
res = s3fscurl.CheckBucket(get_realpath("/").c_str());
|
|
responseCode = s3fscurl.GetLastResponseCode();
|
|
}
|
|
}
|
|
|
|
// check errors(after retrying)
|
|
if(0 > res && responseCode != 200 && responseCode != 301){
|
|
if(responseCode == 400){
|
|
S3FS_PRN_CRIT("Bad Request(host=%s) - result of checking service.", s3host.c_str());
|
|
|
|
}else if(responseCode == 403){
|
|
S3FS_PRN_CRIT("invalid credentials(host=%s) - result of checking service.", s3host.c_str());
|
|
|
|
}else if(responseCode == 404){
|
|
S3FS_PRN_CRIT("bucket or key not found(host=%s) - result of checking service.", s3host.c_str());
|
|
|
|
}else{
|
|
// another error
|
|
S3FS_PRN_CRIT("unable to connect(host=%s) - result of checking service.", s3host.c_str());
|
|
}
|
|
return EXIT_FAILURE;
|
|
}
|
|
}
|
|
s3fscurl.DestroyCurlHandle();
|
|
|
|
// make sure remote mountpath exists and is a directory
|
|
if(!mount_prefix.empty()){
|
|
if(remote_mountpath_exists(mount_prefix.c_str()) != 0){
|
|
S3FS_PRN_CRIT("remote mountpath %s not found.", mount_prefix.c_str());
|
|
return EXIT_FAILURE;
|
|
}
|
|
}
|
|
S3FS_MALLOCTRIM(0);
|
|
|
|
return EXIT_SUCCESS;
|
|
}
|
|
|
|
//
|
|
// Check & Set attributes for mount point.
|
|
//
|
|
static bool set_mountpoint_attribute(struct stat& mpst)
|
|
{
|
|
mp_uid = geteuid();
|
|
mp_gid = getegid();
|
|
mp_mode = S_IFDIR | (allow_other ? (is_mp_umask ? (~mp_umask & (S_IRWXU | S_IRWXG | S_IRWXO)) : (S_IRWXU | S_IRWXG | S_IRWXO)) : S_IRWXU);
|
|
|
|
// In MSYS2 environment with WinFsp, it is not supported to change mode of mount point.
|
|
// Doing that forcely will occurs permission problem, so disabling it.
|
|
#ifdef __MSYS__
|
|
return true;
|
|
#else
|
|
S3FS_PRN_INFO2("PROC(uid=%u, gid=%u) - MountPoint(uid=%u, gid=%u, mode=%04o)",
|
|
(unsigned int)mp_uid, (unsigned int)mp_gid, (unsigned int)(mpst.st_uid), (unsigned int)(mpst.st_gid), mpst.st_mode);
|
|
|
|
// check owner
|
|
if(0 == mp_uid || mpst.st_uid == mp_uid){
|
|
return true;
|
|
}
|
|
// check group permission
|
|
if(mpst.st_gid == mp_gid || 1 == is_uid_include_group(mp_uid, mpst.st_gid)){
|
|
if(S_IRWXG == (mpst.st_mode & S_IRWXG)){
|
|
return true;
|
|
}
|
|
}
|
|
// check other permission
|
|
if(S_IRWXO == (mpst.st_mode & S_IRWXO)){
|
|
return true;
|
|
}
|
|
return false;
|
|
#endif
|
|
}
|
|
|
|
//
|
|
// Set bucket and mount_prefix based on passed bucket name.
|
|
//
|
|
static int set_bucket(const char* arg)
|
|
{
|
|
// TODO: Mutates input. Consider some other tokenization.
|
|
char *bucket_name = const_cast<char*>(arg);
|
|
if(strstr(arg, ":")){
|
|
if(strstr(arg, "://")){
|
|
S3FS_PRN_EXIT("bucket name and path(\"%s\") is wrong, it must be \"bucket[:/path]\".", arg);
|
|
return -1;
|
|
}
|
|
if(!S3fsCred::SetBucket(strtok(bucket_name, ":"))){
|
|
S3FS_PRN_EXIT("bucket name and path(\"%s\") is wrong, it must be \"bucket[:/path]\".", arg);
|
|
return -1;
|
|
}
|
|
char* pmount_prefix = strtok(NULL, "");
|
|
if(pmount_prefix){
|
|
if(0 == strlen(pmount_prefix) || '/' != pmount_prefix[0]){
|
|
S3FS_PRN_EXIT("path(%s) must be prefix \"/\".", pmount_prefix);
|
|
return -1;
|
|
}
|
|
mount_prefix = pmount_prefix;
|
|
// Trim the last consecutive '/'
|
|
mount_prefix = trim_right(mount_prefix, "/");
|
|
}
|
|
}else{
|
|
if(!S3fsCred::SetBucket(arg)){
|
|
S3FS_PRN_EXIT("bucket name and path(\"%s\") is wrong, it must be \"bucket[:/path]\".", arg);
|
|
return -1;
|
|
}
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
// This is repeatedly called by the fuse option parser
|
|
// if the key is equal to FUSE_OPT_KEY_OPT, it's an option passed in prefixed by
|
|
// '-' or '--' e.g.: -f -d -ousecache=/tmp
|
|
//
|
|
// if the key is equal to FUSE_OPT_KEY_NONOPT, it's either the bucket name
|
|
// or the mountpoint. The bucket name will always come before the mountpoint
|
|
static int my_fuse_opt_proc(void* data, const char* arg, int key, struct fuse_args* outargs)
|
|
{
|
|
int ret;
|
|
if(key == FUSE_OPT_KEY_NONOPT){
|
|
// the first NONOPT option is the bucket name
|
|
if(S3fsCred::GetBucket().empty()){
|
|
if ((ret = set_bucket(arg))){
|
|
return ret;
|
|
}
|
|
return 0;
|
|
}else if (!strcmp(arg, "s3fs")) {
|
|
return 0;
|
|
}
|
|
|
|
// the second NONOPT option is the mountpoint(not utility mode)
|
|
if(mountpoint.empty() && NO_UTILITY_MODE == utility_mode){
|
|
// save the mountpoint and do some basic error checking
|
|
mountpoint = arg;
|
|
struct stat stbuf;
|
|
|
|
// In MSYS2 environment with WinFsp, it is not needed to create the mount point before mounting.
|
|
// Also it causes a conflict with WinFsp's validation, so disabling it.
|
|
#ifdef __MSYS__
|
|
memset(&stbuf, 0, sizeof stbuf);
|
|
set_mountpoint_attribute(stbuf);
|
|
#else
|
|
if(stat(arg, &stbuf) == -1){
|
|
S3FS_PRN_EXIT("unable to access MOUNTPOINT %s: %s", mountpoint.c_str(), strerror(errno));
|
|
return -1;
|
|
}
|
|
if(!(S_ISDIR(stbuf.st_mode))){
|
|
S3FS_PRN_EXIT("MOUNTPOINT: %s is not a directory.", mountpoint.c_str());
|
|
return -1;
|
|
}
|
|
if(!set_mountpoint_attribute(stbuf)){
|
|
S3FS_PRN_EXIT("MOUNTPOINT: %s permission denied.", mountpoint.c_str());
|
|
return -1;
|
|
}
|
|
|
|
if(!nonempty){
|
|
struct dirent *ent;
|
|
DIR *dp = opendir(mountpoint.c_str());
|
|
if(dp == NULL){
|
|
S3FS_PRN_EXIT("failed to open MOUNTPOINT: %s: %s", mountpoint.c_str(), strerror(errno));
|
|
return -1;
|
|
}
|
|
while((ent = readdir(dp)) != NULL){
|
|
if(strcmp(ent->d_name, ".") != 0 && strcmp(ent->d_name, "..") != 0){
|
|
closedir(dp);
|
|
S3FS_PRN_EXIT("MOUNTPOINT directory %s is not empty. if you are sure this is safe, can use the 'nonempty' mount option.", mountpoint.c_str());
|
|
return -1;
|
|
}
|
|
}
|
|
closedir(dp);
|
|
}
|
|
#endif
|
|
return 1;
|
|
}
|
|
|
|
// Unknown option
|
|
if(NO_UTILITY_MODE == utility_mode){
|
|
S3FS_PRN_EXIT("specified unknown third option(%s).", arg);
|
|
}else{
|
|
S3FS_PRN_EXIT("specified unknown second option(%s). you don't need to specify second option(mountpoint) for utility mode(-u).", arg);
|
|
}
|
|
return -1;
|
|
|
|
}else if(key == FUSE_OPT_KEY_OPT){
|
|
if(is_prefix(arg, "uid=")){
|
|
s3fs_uid = get_uid(strchr(arg, '=') + sizeof(char));
|
|
if(0 != geteuid() && 0 == s3fs_uid){
|
|
S3FS_PRN_EXIT("root user can only specify uid=0.");
|
|
return -1;
|
|
}
|
|
is_s3fs_uid = true;
|
|
return 1; // continue for fuse option
|
|
}
|
|
if(is_prefix(arg, "gid=")){
|
|
s3fs_gid = get_gid(strchr(arg, '=') + sizeof(char));
|
|
if(0 != getegid() && 0 == s3fs_gid){
|
|
S3FS_PRN_EXIT("root user can only specify gid=0.");
|
|
return -1;
|
|
}
|
|
is_s3fs_gid = true;
|
|
return 1; // continue for fuse option
|
|
}
|
|
if(is_prefix(arg, "umask=")){
|
|
off_t s3fs_umask_tmp = cvt_strtoofft(strchr(arg, '=') + sizeof(char), /*base=*/ 8);
|
|
s3fs_umask = s3fs_umask_tmp & (S_IRWXU | S_IRWXG | S_IRWXO);
|
|
is_s3fs_umask = true;
|
|
return 1; // continue for fuse option
|
|
}
|
|
if(0 == strcmp(arg, "allow_other")){
|
|
allow_other = true;
|
|
return 1; // continue for fuse option
|
|
}
|
|
if(is_prefix(arg, "mp_umask=")){
|
|
off_t mp_umask_tmp = cvt_strtoofft(strchr(arg, '=') + sizeof(char), /*base=*/ 8);
|
|
mp_umask = mp_umask_tmp & (S_IRWXU | S_IRWXG | S_IRWXO);
|
|
is_mp_umask = true;
|
|
return 0;
|
|
}
|
|
if(is_prefix(arg, "default_acl=")){
|
|
const char* acl_string = strchr(arg, '=') + sizeof(char);
|
|
acl_t acl = acl_t::from_str(acl_string);
|
|
if(acl == acl_t::UNKNOWN){
|
|
S3FS_PRN_EXIT("unknown value for default_acl: %s", acl_string);
|
|
return -1;
|
|
}
|
|
S3fsCurl::SetDefaultAcl(acl);
|
|
return 0;
|
|
}
|
|
if(is_prefix(arg, "retries=")){
|
|
off_t retries = cvt_strtoofft(strchr(arg, '=') + sizeof(char), /*base=*/ 10);
|
|
if(retries == 0){
|
|
S3FS_PRN_EXIT("retries must be greater than zero");
|
|
return -1;
|
|
}
|
|
S3fsCurl::SetRetries(static_cast<int>(retries));
|
|
return 0;
|
|
}
|
|
if(is_prefix(arg, "tmpdir=")){
|
|
FdManager::SetTmpDir(strchr(arg, '=') + sizeof(char));
|
|
return 0;
|
|
}
|
|
if(is_prefix(arg, "use_cache=")){
|
|
FdManager::SetCacheDir(strchr(arg, '=') + sizeof(char));
|
|
return 0;
|
|
}
|
|
if(0 == strcmp(arg, "check_cache_dir_exist")){
|
|
FdManager::SetCheckCacheDirExist(true);
|
|
return 0;
|
|
}
|
|
if(0 == strcmp(arg, "del_cache")){
|
|
is_remove_cache = true;
|
|
return 0;
|
|
}
|
|
if(is_prefix(arg, "multireq_max=")){
|
|
int maxreq = static_cast<int>(cvt_strtoofft(strchr(arg, '=') + sizeof(char), /*base=*/ 10));
|
|
S3fsCurl::SetMaxMultiRequest(maxreq);
|
|
return 0;
|
|
}
|
|
if(0 == strcmp(arg, "nonempty")){
|
|
nonempty = true;
|
|
return 1; // need to continue for fuse.
|
|
}
|
|
if(0 == strcmp(arg, "nomultipart")){
|
|
nomultipart = true;
|
|
return 0;
|
|
}
|
|
// old format for storage_class
|
|
if(0 == strcmp(arg, "use_rrs") || is_prefix(arg, "use_rrs=")){
|
|
off_t rrs = 1;
|
|
// for an old format.
|
|
if(is_prefix(arg, "use_rrs=")){
|
|
rrs = cvt_strtoofft(strchr(arg, '=') + sizeof(char), /*base=*/ 10);
|
|
}
|
|
if(0 == rrs){
|
|
S3fsCurl::SetStorageClass("STANDARD");
|
|
}else if(1 == rrs){
|
|
S3fsCurl::SetStorageClass("REDUCED_REDUNDANCY");
|
|
}else{
|
|
S3FS_PRN_EXIT("poorly formed argument to option: use_rrs");
|
|
return -1;
|
|
}
|
|
return 0;
|
|
}
|
|
if(is_prefix(arg, "storage_class=")){
|
|
const char *storage_class = strchr(arg, '=') + sizeof(char);
|
|
S3fsCurl::SetStorageClass(storage_class);
|
|
return 0;
|
|
}
|
|
//
|
|
// [NOTE]
|
|
// use_sse Set Server Side Encrypting type to SSE-S3
|
|
// use_sse=1
|
|
// use_sse=file Set Server Side Encrypting type to Custom key(SSE-C) and load custom keys
|
|
// use_sse=custom(c):file
|
|
// use_sse=custom(c) Set Server Side Encrypting type to Custom key(SSE-C)
|
|
// use_sse=kmsid(k):kms-key-id Set Server Side Encrypting type to AWS Key Management key id(SSE-KMS) and load KMS id
|
|
// use_sse=kmsid(k) Set Server Side Encrypting type to AWS Key Management key id(SSE-KMS)
|
|
//
|
|
// load_sse_c=file Load Server Side Encrypting custom keys
|
|
//
|
|
// AWSSSECKEYS Loading Environment for Server Side Encrypting custom keys
|
|
// AWSSSEKMSID Loading Environment for Server Side Encrypting Key id
|
|
//
|
|
if(is_prefix(arg, "use_sse")){
|
|
if(0 == strcmp(arg, "use_sse") || 0 == strcmp(arg, "use_sse=1")){ // use_sse=1 is old type parameter
|
|
// sse type is SSE_S3
|
|
if(!S3fsCurl::IsSseDisable() && !S3fsCurl::IsSseS3Type()){
|
|
S3FS_PRN_EXIT("already set SSE another type, so conflict use_sse option or environment.");
|
|
return -1;
|
|
}
|
|
S3fsCurl::SetSseType(sse_type_t::SSE_S3);
|
|
|
|
}else if(0 == strcmp(arg, "use_sse=kmsid") || 0 == strcmp(arg, "use_sse=k")){
|
|
// sse type is SSE_KMS with out kmsid(expecting id is loaded by environment)
|
|
if(!S3fsCurl::IsSseDisable() && !S3fsCurl::IsSseKmsType()){
|
|
S3FS_PRN_EXIT("already set SSE another type, so conflict use_sse option or environment.");
|
|
return -1;
|
|
}
|
|
if(!S3fsCurl::IsSetSseKmsId()){
|
|
S3FS_PRN_EXIT("use_sse=kms but not loaded kms id by environment.");
|
|
return -1;
|
|
}
|
|
S3fsCurl::SetSseType(sse_type_t::SSE_KMS);
|
|
|
|
}else if(is_prefix(arg, "use_sse=kmsid:") || is_prefix(arg, "use_sse=k:")){
|
|
// sse type is SSE_KMS with kmsid
|
|
if(!S3fsCurl::IsSseDisable() && !S3fsCurl::IsSseKmsType()){
|
|
S3FS_PRN_EXIT("already set SSE another type, so conflict use_sse option or environment.");
|
|
return -1;
|
|
}
|
|
const char* kmsid;
|
|
if(is_prefix(arg, "use_sse=kmsid:")){
|
|
kmsid = &arg[strlen("use_sse=kmsid:")];
|
|
}else{
|
|
kmsid = &arg[strlen("use_sse=k:")];
|
|
}
|
|
if(!S3fsCurl::SetSseKmsid(kmsid)){
|
|
S3FS_PRN_EXIT("failed to load use_sse kms id.");
|
|
return -1;
|
|
}
|
|
S3fsCurl::SetSseType(sse_type_t::SSE_KMS);
|
|
|
|
}else if(0 == strcmp(arg, "use_sse=custom") || 0 == strcmp(arg, "use_sse=c")){
|
|
// sse type is SSE_C with out custom keys(expecting keys are loaded by environment or load_sse_c option)
|
|
if(!S3fsCurl::IsSseDisable() && !S3fsCurl::IsSseCType()){
|
|
S3FS_PRN_EXIT("already set SSE another type, so conflict use_sse option or environment.");
|
|
return -1;
|
|
}
|
|
// [NOTE]
|
|
// do not check ckeys exists here.
|
|
//
|
|
S3fsCurl::SetSseType(sse_type_t::SSE_C);
|
|
|
|
}else if(is_prefix(arg, "use_sse=custom:") || is_prefix(arg, "use_sse=c:")){
|
|
// sse type is SSE_C with custom keys
|
|
if(!S3fsCurl::IsSseDisable() && !S3fsCurl::IsSseCType()){
|
|
S3FS_PRN_EXIT("already set SSE another type, so conflict use_sse option or environment.");
|
|
return -1;
|
|
}
|
|
const char* ssecfile;
|
|
if(is_prefix(arg, "use_sse=custom:")){
|
|
ssecfile = &arg[strlen("use_sse=custom:")];
|
|
}else{
|
|
ssecfile = &arg[strlen("use_sse=c:")];
|
|
}
|
|
if(!S3fsCurl::SetSseCKeys(ssecfile)){
|
|
S3FS_PRN_EXIT("failed to load use_sse custom key file(%s).", ssecfile);
|
|
return -1;
|
|
}
|
|
S3fsCurl::SetSseType(sse_type_t::SSE_C);
|
|
|
|
}else if(0 == strcmp(arg, "use_sse=")){ // this type is old style(parameter is custom key file path)
|
|
// SSE_C with custom keys.
|
|
const char* ssecfile = &arg[strlen("use_sse=")];
|
|
if(!S3fsCurl::SetSseCKeys(ssecfile)){
|
|
S3FS_PRN_EXIT("failed to load use_sse custom key file(%s).", ssecfile);
|
|
return -1;
|
|
}
|
|
S3fsCurl::SetSseType(sse_type_t::SSE_C);
|
|
|
|
}else{
|
|
// never come here.
|
|
S3FS_PRN_EXIT("something wrong use_sse option.");
|
|
return -1;
|
|
}
|
|
return 0;
|
|
}
|
|
// [NOTE]
|
|
// Do only load SSE custom keys, care for set without set sse type.
|
|
if(is_prefix(arg, "load_sse_c=")){
|
|
const char* ssecfile = &arg[strlen("load_sse_c=")];
|
|
if(!S3fsCurl::SetSseCKeys(ssecfile)){
|
|
S3FS_PRN_EXIT("failed to load use_sse custom key file(%s).", ssecfile);
|
|
return -1;
|
|
}
|
|
return 0;
|
|
}
|
|
if(is_prefix(arg, "ssl_verify_hostname=")){
|
|
long sslvh = static_cast<long>(cvt_strtoofft(strchr(arg, '=') + sizeof(char), /*base=*/ 10));
|
|
if(-1 == S3fsCurl::SetSslVerifyHostname(sslvh)){
|
|
S3FS_PRN_EXIT("poorly formed argument to option: ssl_verify_hostname.");
|
|
return -1;
|
|
}
|
|
return 0;
|
|
}
|
|
//
|
|
// Detect options for credential
|
|
//
|
|
if(0 >= (ret = ps3fscred->DetectParam(arg))){
|
|
if(0 > ret){
|
|
return -1;
|
|
}
|
|
return 0;
|
|
}
|
|
if(is_prefix(arg, "public_bucket=")){
|
|
off_t pubbucket = cvt_strtoofft(strchr(arg, '=') + sizeof(char), /*base=*/ 10);
|
|
if(1 == pubbucket){
|
|
S3fsCurl::SetPublicBucket(true);
|
|
// [NOTE]
|
|
// if bucket is public(without credential), s3 do not allow copy api.
|
|
// so s3fs sets nocopyapi mode.
|
|
//
|
|
nocopyapi = true;
|
|
}else if(0 == pubbucket){
|
|
S3fsCurl::SetPublicBucket(false);
|
|
}else{
|
|
S3FS_PRN_EXIT("poorly formed argument to option: public_bucket.");
|
|
return -1;
|
|
}
|
|
return 0;
|
|
}
|
|
if(is_prefix(arg, "bucket=")){
|
|
std::string bname = strchr(arg, '=') + sizeof(char);
|
|
if ((ret = set_bucket(bname.c_str()))){
|
|
return ret;
|
|
}
|
|
return 0;
|
|
}
|
|
if(0 == strcmp(arg, "no_check_certificate")){
|
|
S3fsCurl::SetCheckCertificate(false);
|
|
return 0;
|
|
}
|
|
if(is_prefix(arg, "connect_timeout=")){
|
|
long contimeout = static_cast<long>(cvt_strtoofft(strchr(arg, '=') + sizeof(char), /*base=*/ 10));
|
|
S3fsCurl::SetConnectTimeout(contimeout);
|
|
return 0;
|
|
}
|
|
if(is_prefix(arg, "readwrite_timeout=")){
|
|
time_t rwtimeout = static_cast<time_t>(cvt_strtoofft(strchr(arg, '=') + sizeof(char), /*base=*/ 10));
|
|
S3fsCurl::SetReadwriteTimeout(rwtimeout);
|
|
return 0;
|
|
}
|
|
if(is_prefix(arg, "list_object_max_keys=")){
|
|
int max_keys = static_cast<int>(cvt_strtoofft(strchr(arg, '=') + sizeof(char), /*base=*/ 10));
|
|
if(max_keys < 1000){
|
|
S3FS_PRN_EXIT("argument should be over 1000: list_object_max_keys");
|
|
return -1;
|
|
}
|
|
max_keys_list_object = max_keys;
|
|
return 0;
|
|
}
|
|
if(is_prefix(arg, "max_stat_cache_size=")){
|
|
unsigned long cache_size = static_cast<unsigned long>(cvt_strtoofft(strchr(arg, '=') + sizeof(char), 10));
|
|
StatCache::getStatCacheData()->SetCacheSize(cache_size);
|
|
return 0;
|
|
}
|
|
if(is_prefix(arg, "stat_cache_expire=")){
|
|
time_t expr_time = static_cast<time_t>(cvt_strtoofft(strchr(arg, '=') + sizeof(char), 10));
|
|
StatCache::getStatCacheData()->SetExpireTime(expr_time);
|
|
return 0;
|
|
}
|
|
// [NOTE]
|
|
// This option is for compatibility old version.
|
|
if(is_prefix(arg, "stat_cache_interval_expire=")){
|
|
time_t expr_time = static_cast<time_t>(cvt_strtoofft(strchr(arg, '=') + sizeof(char), /*base=*/ 10));
|
|
StatCache::getStatCacheData()->SetExpireTime(expr_time, true);
|
|
return 0;
|
|
}
|
|
if(0 == strcmp(arg, "enable_noobj_cache")){
|
|
S3FS_PRN_WARN("enable_noobj_cache is enabled by default and a future version will remove this option.");
|
|
StatCache::getStatCacheData()->EnableCacheNoObject();
|
|
return 0;
|
|
}
|
|
if(0 == strcmp(arg, "disable_noobj_cache")){
|
|
StatCache::getStatCacheData()->DisableCacheNoObject();
|
|
return 0;
|
|
}
|
|
if(0 == strcmp(arg, "nodnscache")){
|
|
S3fsCurl::SetDnsCache(false);
|
|
return 0;
|
|
}
|
|
if(0 == strcmp(arg, "nosscache")){
|
|
S3fsCurl::SetSslSessionCache(false);
|
|
return 0;
|
|
}
|
|
if(is_prefix(arg, "parallel_count=") || is_prefix(arg, "parallel_upload=")){
|
|
int maxpara = static_cast<int>(cvt_strtoofft(strchr(arg, '=') + sizeof(char), /*base=*/ 10));
|
|
if(0 >= maxpara){
|
|
S3FS_PRN_EXIT("argument should be over 1: parallel_count");
|
|
return -1;
|
|
}
|
|
S3fsCurl::SetMaxParallelCount(maxpara);
|
|
return 0;
|
|
}
|
|
if(is_prefix(arg, "max_thread_count=")){
|
|
int max_thcount = static_cast<int>(cvt_strtoofft(strchr(arg, '=') + sizeof(char), /*base=*/ 10));
|
|
if(0 >= max_thcount){
|
|
S3FS_PRN_EXIT("argument should be over 1: max_thread_count");
|
|
return -1;
|
|
}
|
|
max_thread_count = max_thcount;
|
|
S3FS_PRN_WARN("The max_thread_count option is not a formal option. Please note that it will change in the future.");
|
|
return 0;
|
|
}
|
|
if(is_prefix(arg, "fd_page_size=")){
|
|
S3FS_PRN_ERR("option fd_page_size is no longer supported, so skip this option.");
|
|
return 0;
|
|
}
|
|
if(is_prefix(arg, "multipart_size=")){
|
|
off_t size = static_cast<off_t>(cvt_strtoofft(strchr(arg, '=') + sizeof(char), /*base=*/ 10));
|
|
if(!S3fsCurl::SetMultipartSize(size)){
|
|
S3FS_PRN_EXIT("multipart_size option must be at least 5 MB.");
|
|
return -1;
|
|
}
|
|
return 0;
|
|
}
|
|
if(is_prefix(arg, "multipart_copy_size=")){
|
|
off_t size = static_cast<off_t>(cvt_strtoofft(strchr(arg, '=') + sizeof(char), /*base=*/ 10));
|
|
if(!S3fsCurl::SetMultipartCopySize(size)){
|
|
S3FS_PRN_EXIT("multipart_copy_size option must be at least 5 MB.");
|
|
return -1;
|
|
}
|
|
return 0;
|
|
}
|
|
if(is_prefix(arg, "max_dirty_data=")){
|
|
off_t size = static_cast<off_t>(cvt_strtoofft(strchr(arg, '=') + sizeof(char), /*base=*/ 10));
|
|
if(size >= 50){
|
|
size *= 1024 * 1024;
|
|
}else if(size != -1){
|
|
S3FS_PRN_EXIT("max_dirty_data option must be at least 50 MB.");
|
|
return -1;
|
|
}
|
|
max_dirty_data = size;
|
|
return 0;
|
|
}
|
|
if(is_prefix(arg, "ensure_diskfree=")){
|
|
off_t dfsize = cvt_strtoofft(strchr(arg, '=') + sizeof(char), /*base=*/ 10) * 1024 * 1024;
|
|
if(dfsize < S3fsCurl::GetMultipartSize()){
|
|
S3FS_PRN_WARN("specified size to ensure disk free space is smaller than multipart size, so set multipart size to it.");
|
|
dfsize = S3fsCurl::GetMultipartSize();
|
|
}
|
|
FdManager::SetEnsureFreeDiskSpace(dfsize);
|
|
return 0;
|
|
}
|
|
if(is_prefix(arg, "fake_diskfree=")){
|
|
S3FS_PRN_WARN("The fake_diskfree option was specified. Use this option for testing or debugging.");
|
|
|
|
// [NOTE] This value is used for initializing to FdManager after parsing all options.
|
|
fake_diskfree_size = cvt_strtoofft(strchr(arg, '=') + sizeof(char), /*base=*/ 10) * 1024 * 1024;
|
|
return 0;
|
|
}
|
|
if(is_prefix(arg, "multipart_threshold=")){
|
|
multipart_threshold = static_cast<int64_t>(cvt_strtoofft(strchr(arg, '=') + sizeof(char), /*base=*/ 10)) * 1024 * 1024;
|
|
if(multipart_threshold <= MIN_MULTIPART_SIZE){
|
|
S3FS_PRN_EXIT("multipart_threshold must be at least %lld, was: %lld", static_cast<long long>(MIN_MULTIPART_SIZE), static_cast<long long>(multipart_threshold));
|
|
return -1;
|
|
}
|
|
return 0;
|
|
}
|
|
if(is_prefix(arg, "singlepart_copy_limit=")){
|
|
singlepart_copy_limit = static_cast<int64_t>(cvt_strtoofft(strchr(arg, '=') + sizeof(char), /*base=*/ 10)) * 1024 * 1024;
|
|
return 0;
|
|
}
|
|
if(is_prefix(arg, "ahbe_conf=")){
|
|
std::string ahbe_conf = strchr(arg, '=') + sizeof(char);
|
|
if(!AdditionalHeader::get()->Load(ahbe_conf.c_str())){
|
|
S3FS_PRN_EXIT("failed to load ahbe_conf file(%s).", ahbe_conf.c_str());
|
|
return -1;
|
|
}
|
|
AdditionalHeader::get()->Dump();
|
|
return 0;
|
|
}
|
|
if(0 == strcmp(arg, "noxmlns")){
|
|
noxmlns = true;
|
|
return 0;
|
|
}
|
|
if(0 == strcmp(arg, "nomixupload")){
|
|
FdEntity::SetNoMixMultipart();
|
|
return 0;
|
|
}
|
|
if(0 == strcmp(arg, "nocopyapi")){
|
|
nocopyapi = true;
|
|
return 0;
|
|
}
|
|
if(0 == strcmp(arg, "streamupload")){
|
|
FdEntity::SetStreamUpload(true);
|
|
S3FS_PRN_WARN("The streamupload option is not a formal option. Please note that it will change in the future.");
|
|
return 0;
|
|
}
|
|
if(0 == strcmp(arg, "norenameapi")){
|
|
norenameapi = true;
|
|
return 0;
|
|
}
|
|
if(0 == strcmp(arg, "complement_stat")){
|
|
complement_stat = true;
|
|
return 0;
|
|
}
|
|
if(0 == strcmp(arg, "notsup_compat_dir")){
|
|
S3FS_PRN_WARN("notsup_compat_dir is enabled by default and a future version will remove this option.");
|
|
|
|
support_compat_dir = false;
|
|
return 0;
|
|
}
|
|
if(0 == strcmp(arg, "compat_dir")){
|
|
support_compat_dir = true;
|
|
return 0;
|
|
}
|
|
if(0 == strcmp(arg, "enable_content_md5")){
|
|
S3fsCurl::SetContentMd5(true);
|
|
return 0;
|
|
}
|
|
if(0 == strcmp(arg, "enable_unsigned_payload")){
|
|
S3fsCurl::SetUnsignedPayload(true);
|
|
return 0;
|
|
}
|
|
if(0 == strcmp(arg, "update_parent_dir_stat")){
|
|
update_parent_dir_stat = true;
|
|
return 0;
|
|
}
|
|
if(is_prefix(arg, "host=")){
|
|
s3host = strchr(arg, '=') + sizeof(char);
|
|
return 0;
|
|
}
|
|
if(is_prefix(arg, "servicepath=")){
|
|
service_path = strchr(arg, '=') + sizeof(char);
|
|
return 0;
|
|
}
|
|
if(is_prefix(arg, "url=")){
|
|
s3host = strchr(arg, '=') + sizeof(char);
|
|
// strip the trailing '/', if any, off the end of the host
|
|
// std::string
|
|
size_t found, length;
|
|
found = s3host.find_last_of('/');
|
|
length = s3host.length();
|
|
while(found == (length - 1) && length > 0){
|
|
s3host.erase(found);
|
|
found = s3host.find_last_of('/');
|
|
length = s3host.length();
|
|
}
|
|
// Check url for http / https protocol std::string
|
|
if(!is_prefix(s3host.c_str(), "https://") && !is_prefix(s3host.c_str(), "http://")){
|
|
S3FS_PRN_EXIT("option url has invalid format, missing http / https protocol");
|
|
return -1;
|
|
}
|
|
return 0;
|
|
}
|
|
if(0 == strcmp(arg, "sigv2")){
|
|
S3fsCurl::SetSignatureType(V2_ONLY);
|
|
return 0;
|
|
}
|
|
if(0 == strcmp(arg, "sigv4")){
|
|
S3fsCurl::SetSignatureType(V4_ONLY);
|
|
return 0;
|
|
}
|
|
if(is_prefix(arg, "endpoint=")){
|
|
endpoint = strchr(arg, '=') + sizeof(char);
|
|
is_specified_endpoint = true;
|
|
return 0;
|
|
}
|
|
if(0 == strcmp(arg, "use_path_request_style")){
|
|
pathrequeststyle = true;
|
|
return 0;
|
|
}
|
|
if(0 == strcmp(arg, "noua")){
|
|
S3fsCurl::SetUserAgentFlag(false);
|
|
return 0;
|
|
}
|
|
if(0 == strcmp(arg, "listobjectsv2")){
|
|
S3fsCurl::SetListObjectsV2(true);
|
|
return 0;
|
|
}
|
|
if(0 == strcmp(arg, "use_xattr")){
|
|
is_use_xattr = true;
|
|
return 0;
|
|
}else if(is_prefix(arg, "use_xattr=")){
|
|
const char* strflag = strchr(arg, '=') + sizeof(char);
|
|
if(0 == strcmp(strflag, "1")){
|
|
is_use_xattr = true;
|
|
}else if(0 == strcmp(strflag, "0")){
|
|
is_use_xattr = false;
|
|
}else{
|
|
S3FS_PRN_EXIT("option use_xattr has unknown parameter(%s).", strflag);
|
|
return -1;
|
|
}
|
|
return 0;
|
|
}
|
|
if(is_prefix(arg, "cipher_suites=")){
|
|
cipher_suites = strchr(arg, '=') + sizeof(char);
|
|
return 0;
|
|
}
|
|
if(is_prefix(arg, "instance_name=")){
|
|
instance_name = strchr(arg, '=') + sizeof(char);
|
|
instance_name = "[" + instance_name + "]";
|
|
return 0;
|
|
}
|
|
if(is_prefix(arg, "mime=")){
|
|
mimetype_file = strchr(arg, '=') + sizeof(char);
|
|
return 0;
|
|
}
|
|
if(is_prefix(arg, "proxy=")){
|
|
const char* url = &arg[strlen("proxy=")];
|
|
if(!S3fsCurl::SetProxy(url)){
|
|
S3FS_PRN_EXIT("failed to set proxy(%s).", url);
|
|
return -1;
|
|
}
|
|
return 0;
|
|
}
|
|
if(is_prefix(arg, "proxy_cred_file=")){
|
|
const char* file = &arg[strlen("proxy_cred_file=")];
|
|
if(!S3fsCurl::SetProxyUserPwd(file)){
|
|
S3FS_PRN_EXIT("failed to set proxy user and passphrase from file(%s).", file);
|
|
return -1;
|
|
}
|
|
return 0;
|
|
}
|
|
//
|
|
// log file option
|
|
//
|
|
if(is_prefix(arg, "logfile=")){
|
|
const char* strlogfile = strchr(arg, '=') + sizeof(char);
|
|
if(!S3fsLog::SetLogfile(strlogfile)){
|
|
S3FS_PRN_EXIT("The file(%s) specified by logfile option could not be opened.", strlogfile);
|
|
return -1;
|
|
}
|
|
return 0;
|
|
}
|
|
//
|
|
// debug level option
|
|
//
|
|
if(is_prefix(arg, "dbglevel=")){
|
|
const char* strlevel = strchr(arg, '=') + sizeof(char);
|
|
if(0 == strcasecmp(strlevel, "silent") || 0 == strcasecmp(strlevel, "critical") || 0 == strcasecmp(strlevel, "crit")){
|
|
S3fsLog::SetLogLevel(S3fsLog::LEVEL_CRIT);
|
|
}else if(0 == strcasecmp(strlevel, "error") || 0 == strcasecmp(strlevel, "err")){
|
|
S3fsLog::SetLogLevel(S3fsLog::LEVEL_ERR);
|
|
}else if(0 == strcasecmp(strlevel, "wan") || 0 == strcasecmp(strlevel, "warn") || 0 == strcasecmp(strlevel, "warning")){
|
|
S3fsLog::SetLogLevel(S3fsLog::LEVEL_WARN);
|
|
}else if(0 == strcasecmp(strlevel, "inf") || 0 == strcasecmp(strlevel, "info") || 0 == strcasecmp(strlevel, "information")){
|
|
S3fsLog::SetLogLevel(S3fsLog::LEVEL_INFO);
|
|
}else if(0 == strcasecmp(strlevel, "dbg") || 0 == strcasecmp(strlevel, "debug")){
|
|
S3fsLog::SetLogLevel(S3fsLog::LEVEL_DBG);
|
|
}else{
|
|
S3FS_PRN_EXIT("option dbglevel has unknown parameter(%s).", strlevel);
|
|
return -1;
|
|
}
|
|
return 0;
|
|
}
|
|
//
|
|
// debug option
|
|
//
|
|
// S3fsLog level is LEVEL_INFO, after second -d is passed to fuse.
|
|
//
|
|
if(0 == strcmp(arg, "-d") || 0 == strcmp(arg, "--debug")){
|
|
if(!S3fsLog::IsS3fsLogInfo() && !S3fsLog::IsS3fsLogDbg()){
|
|
S3fsLog::SetLogLevel(S3fsLog::LEVEL_INFO);
|
|
return 0;
|
|
}
|
|
if(0 == strcmp(arg, "--debug")){
|
|
// fuse doesn't understand "--debug", but it understands -d.
|
|
// but we can't pass -d back to fuse.
|
|
return 0;
|
|
}
|
|
}
|
|
// "f2" is not used no more.
|
|
// (set S3fsLog::LEVEL_DBG)
|
|
if(0 == strcmp(arg, "f2")){
|
|
S3fsLog::SetLogLevel(S3fsLog::LEVEL_DBG);
|
|
return 0;
|
|
}
|
|
if(0 == strcmp(arg, "curldbg")){
|
|
S3fsCurl::SetVerbose(true);
|
|
return 0;
|
|
}else if(is_prefix(arg, "curldbg=")){
|
|
const char* strlevel = strchr(arg, '=') + sizeof(char);
|
|
if(0 == strcasecmp(strlevel, "normal")){
|
|
S3fsCurl::SetVerbose(true);
|
|
}else if(0 == strcasecmp(strlevel, "body")){
|
|
S3fsCurl::SetVerbose(true);
|
|
S3fsCurl::SetDumpBody(true);
|
|
}else{
|
|
S3FS_PRN_EXIT("option curldbg has unknown parameter(%s).", strlevel);
|
|
return -1;
|
|
}
|
|
return 0;
|
|
}
|
|
//
|
|
// no time stamp in debug message
|
|
//
|
|
if(0 == strcmp(arg, "no_time_stamp_msg")){
|
|
S3fsLog::SetTimeStamp(false);
|
|
return 0;
|
|
}
|
|
//
|
|
// Check cache file, using SIGUSR1
|
|
//
|
|
if(0 == strcmp(arg, "set_check_cache_sigusr1")){
|
|
if(!S3fsSignals::SetUsr1Handler(NULL)){
|
|
S3FS_PRN_EXIT("could not set sigusr1 for checking cache.");
|
|
return -1;
|
|
}
|
|
return 0;
|
|
}else if(is_prefix(arg, "set_check_cache_sigusr1=")){
|
|
const char* strfilepath = strchr(arg, '=') + sizeof(char);
|
|
if(!S3fsSignals::SetUsr1Handler(strfilepath)){
|
|
S3FS_PRN_EXIT("could not set sigusr1 for checking cache and output file(%s).", strfilepath);
|
|
return -1;
|
|
}
|
|
return 0;
|
|
}
|
|
if(is_prefix(arg, "accessKeyId=")){
|
|
S3FS_PRN_EXIT("option accessKeyId is no longer supported.");
|
|
return -1;
|
|
}
|
|
if(is_prefix(arg, "secretAccessKey=")){
|
|
S3FS_PRN_EXIT("option secretAccessKey is no longer supported.");
|
|
return -1;
|
|
}
|
|
if(0 == strcmp(arg, "use_wtf8")){
|
|
use_wtf8 = true;
|
|
return 0;
|
|
}
|
|
if(0 == strcmp(arg, "requester_pays")){
|
|
S3fsCurl::SetRequesterPays(true);
|
|
return 0;
|
|
}
|
|
// [NOTE]
|
|
// following option will be discarding, because these are not for fuse.
|
|
// (Referenced sshfs.c)
|
|
//
|
|
if(0 == strcmp(arg, "auto") ||
|
|
0 == strcmp(arg, "noauto") ||
|
|
0 == strcmp(arg, "user") ||
|
|
0 == strcmp(arg, "nouser") ||
|
|
0 == strcmp(arg, "users") ||
|
|
0 == strcmp(arg, "_netdev"))
|
|
{
|
|
return 0;
|
|
}
|
|
}
|
|
return 1;
|
|
}
|
|
|
|
int main(int argc, char* argv[])
|
|
{
|
|
int ch;
|
|
int fuse_res;
|
|
int option_index = 0;
|
|
struct fuse_operations s3fs_oper;
|
|
time_t incomp_abort_time = (24 * 60 * 60);
|
|
S3fsLog singletonLog;
|
|
|
|
static const struct option long_opts[] = {
|
|
{"help", no_argument, NULL, 'h'},
|
|
{"version", no_argument, 0, 0},
|
|
{"debug", no_argument, NULL, 'd'},
|
|
{"incomplete-mpu-list", no_argument, NULL, 'u'},
|
|
{"incomplete-mpu-abort", optional_argument, NULL, 'a'}, // 'a' is only identifier and is not option.
|
|
{NULL, 0, NULL, 0}
|
|
};
|
|
|
|
// init xml2
|
|
xmlInitParser();
|
|
LIBXML_TEST_VERSION
|
|
|
|
init_sysconf_vars();
|
|
|
|
// get program name - emulate basename
|
|
program_name = argv[0];
|
|
size_t found = program_name.find_last_of('/');
|
|
if(found != std::string::npos){
|
|
program_name.replace(0, found+1, "");
|
|
}
|
|
|
|
// set credential object
|
|
//
|
|
ps3fscred = new S3fsCred();
|
|
if(!S3fsCurl::InitCredentialObject(ps3fscred)){
|
|
S3FS_PRN_EXIT("Failed to setup credential object to s3fs curl.");
|
|
delete ps3fscred;
|
|
exit(EXIT_FAILURE);
|
|
}
|
|
|
|
while((ch = getopt_long(argc, argv, "dho:fsu", long_opts, &option_index)) != -1){
|
|
switch(ch){
|
|
case 0:
|
|
if(strcmp(long_opts[option_index].name, "version") == 0){
|
|
show_version();
|
|
delete ps3fscred;
|
|
exit(EXIT_SUCCESS);
|
|
}
|
|
break;
|
|
case 'h':
|
|
show_help();
|
|
delete ps3fscred;
|
|
exit(EXIT_SUCCESS);
|
|
case 'o':
|
|
break;
|
|
case 'd':
|
|
break;
|
|
case 'f':
|
|
foreground = true;
|
|
break;
|
|
case 's':
|
|
break;
|
|
case 'u': // --incomplete-mpu-list
|
|
if(NO_UTILITY_MODE != utility_mode){
|
|
S3FS_PRN_EXIT("already utility mode option is specified.");
|
|
delete ps3fscred;
|
|
exit(EXIT_FAILURE);
|
|
}
|
|
utility_mode = INCOMP_TYPE_LIST;
|
|
break;
|
|
case 'a': // --incomplete-mpu-abort
|
|
if(NO_UTILITY_MODE != utility_mode){
|
|
S3FS_PRN_EXIT("already utility mode option is specified.");
|
|
delete ps3fscred;
|
|
exit(EXIT_FAILURE);
|
|
}
|
|
utility_mode = INCOMP_TYPE_ABORT;
|
|
|
|
// check expire argument
|
|
if(NULL != optarg && 0 == strcasecmp(optarg, "all")){ // all is 0s
|
|
incomp_abort_time = 0;
|
|
}else if(NULL != optarg){
|
|
if(!convert_unixtime_from_option_arg(optarg, incomp_abort_time)){
|
|
S3FS_PRN_EXIT("--incomplete-mpu-abort option argument is wrong.");
|
|
delete ps3fscred;
|
|
exit(EXIT_FAILURE);
|
|
}
|
|
}
|
|
// if optarg is null, incomp_abort_time is 24H(default)
|
|
break;
|
|
default:
|
|
delete ps3fscred;
|
|
exit(EXIT_FAILURE);
|
|
}
|
|
}
|
|
// print launch message
|
|
print_launch_message(argc, argv);
|
|
|
|
// Load SSE environment
|
|
if(!S3fsCurl::LoadEnvSse()){
|
|
S3FS_PRN_EXIT("something wrong about SSE environment.");
|
|
delete ps3fscred;
|
|
exit(EXIT_FAILURE);
|
|
}
|
|
|
|
// ssl init
|
|
if(!s3fs_init_global_ssl()){
|
|
S3FS_PRN_EXIT("could not initialize for ssl libraries.");
|
|
delete ps3fscred;
|
|
exit(EXIT_FAILURE);
|
|
}
|
|
|
|
// mutex for xml
|
|
if(!init_parser_xml_lock()){
|
|
S3FS_PRN_EXIT("could not initialize mutex for xml parser.");
|
|
s3fs_destroy_global_ssl();
|
|
delete ps3fscred;
|
|
exit(EXIT_FAILURE);
|
|
}
|
|
|
|
// mutex for basename/dirname
|
|
if(!init_basename_lock()){
|
|
S3FS_PRN_EXIT("could not initialize mutex for basename/dirname.");
|
|
s3fs_destroy_global_ssl();
|
|
destroy_parser_xml_lock();
|
|
exit(EXIT_FAILURE);
|
|
}
|
|
|
|
// init curl (without mime types)
|
|
//
|
|
// [NOTE]
|
|
// The curl initialization here does not load mime types.
|
|
// The mime types file parameter are dynamic values according
|
|
// to the user's environment, and are analyzed by the my_fuse_opt_proc
|
|
// function.
|
|
// The my_fuse_opt_proc function is executed after this curl
|
|
// initialization. Because the curl method is used in the
|
|
// my_fuse_opt_proc function, then it must be called here to
|
|
// initialize. Fortunately, the processing using mime types
|
|
// is only PUT/POST processing, and it is not used until the
|
|
// call of my_fuse_opt_proc function is completed. Therefore,
|
|
// the mime type is loaded just after calling the my_fuse_opt_proc
|
|
// function.
|
|
//
|
|
if(!S3fsCurl::InitS3fsCurl()){
|
|
S3FS_PRN_EXIT("Could not initiate curl library.");
|
|
s3fs_destroy_global_ssl();
|
|
destroy_parser_xml_lock();
|
|
destroy_basename_lock();
|
|
delete ps3fscred;
|
|
exit(EXIT_FAILURE);
|
|
}
|
|
|
|
// clear this structure
|
|
memset(&s3fs_oper, 0, sizeof(s3fs_oper));
|
|
|
|
// This is the fuse-style parser for the arguments
|
|
// after which the bucket name and mountpoint names
|
|
// should have been set
|
|
struct fuse_args custom_args = FUSE_ARGS_INIT(argc, argv);
|
|
if(0 != fuse_opt_parse(&custom_args, NULL, NULL, my_fuse_opt_proc)){
|
|
S3fsCurl::DestroyS3fsCurl();
|
|
s3fs_destroy_global_ssl();
|
|
destroy_parser_xml_lock();
|
|
destroy_basename_lock();
|
|
delete ps3fscred;
|
|
exit(EXIT_FAILURE);
|
|
}
|
|
|
|
// init mime types for curl
|
|
if(!S3fsCurl::InitMimeType(mimetype_file)){
|
|
S3FS_PRN_WARN("Missing MIME types prevents setting Content-Type on uploaded objects.");
|
|
}
|
|
|
|
// [NOTE]
|
|
// exclusive option check here.
|
|
//
|
|
if(strcasecmp(S3fsCurl::GetStorageClass().c_str(), "REDUCED_REDUNDANCY") == 0 && !S3fsCurl::IsSseDisable()){
|
|
S3FS_PRN_EXIT("use_sse option could not be specified with storage class reduced_redundancy.");
|
|
S3fsCurl::DestroyS3fsCurl();
|
|
s3fs_destroy_global_ssl();
|
|
destroy_parser_xml_lock();
|
|
destroy_basename_lock();
|
|
delete ps3fscred;
|
|
exit(EXIT_FAILURE);
|
|
}
|
|
if(!S3fsCurl::FinalCheckSse()){
|
|
S3FS_PRN_EXIT("something wrong about SSE options.");
|
|
S3fsCurl::DestroyS3fsCurl();
|
|
s3fs_destroy_global_ssl();
|
|
destroy_parser_xml_lock();
|
|
destroy_basename_lock();
|
|
delete ps3fscred;
|
|
exit(EXIT_FAILURE);
|
|
}
|
|
|
|
if(S3fsCurl::GetSignatureType() == V2_ONLY && S3fsCurl::GetUnsignedPayload()){
|
|
S3FS_PRN_WARN("Ignoring enable_unsigned_payload with sigv2");
|
|
}
|
|
|
|
if(!FdEntity::GetNoMixMultipart() && max_dirty_data != -1){
|
|
S3FS_PRN_WARN("Setting max_dirty_data to -1 when nomixupload is enabled");
|
|
max_dirty_data = -1;
|
|
}
|
|
|
|
//
|
|
// Check the combination of parameters for credential
|
|
//
|
|
if(!ps3fscred->CheckAllParams()){
|
|
S3fsCurl::DestroyS3fsCurl();
|
|
s3fs_destroy_global_ssl();
|
|
destroy_parser_xml_lock();
|
|
destroy_basename_lock();
|
|
delete ps3fscred;
|
|
exit(EXIT_FAILURE);
|
|
}
|
|
|
|
// The second plain argument is the mountpoint
|
|
// if the option was given, we all ready checked for a
|
|
// readable, non-empty directory, this checks determines
|
|
// if the mountpoint option was ever supplied
|
|
if(NO_UTILITY_MODE == utility_mode){
|
|
if(mountpoint.empty()){
|
|
S3FS_PRN_EXIT("missing MOUNTPOINT argument.");
|
|
show_usage();
|
|
S3fsCurl::DestroyS3fsCurl();
|
|
s3fs_destroy_global_ssl();
|
|
destroy_parser_xml_lock();
|
|
destroy_basename_lock();
|
|
delete ps3fscred;
|
|
exit(EXIT_FAILURE);
|
|
}
|
|
}
|
|
|
|
// check tmp dir permission
|
|
if(!FdManager::CheckTmpDirExist()){
|
|
S3FS_PRN_EXIT("temporary directory doesn't exists.");
|
|
S3fsCurl::DestroyS3fsCurl();
|
|
s3fs_destroy_global_ssl();
|
|
destroy_parser_xml_lock();
|
|
destroy_basename_lock();
|
|
delete ps3fscred;
|
|
exit(EXIT_FAILURE);
|
|
}
|
|
|
|
// check cache dir permission
|
|
if(!FdManager::CheckCacheDirExist() || !FdManager::CheckCacheTopDir() || !CacheFileStat::CheckCacheFileStatTopDir()){
|
|
S3FS_PRN_EXIT("could not allow cache directory permission, check permission of cache directories.");
|
|
S3fsCurl::DestroyS3fsCurl();
|
|
s3fs_destroy_global_ssl();
|
|
destroy_parser_xml_lock();
|
|
destroy_basename_lock();
|
|
delete ps3fscred;
|
|
exit(EXIT_FAILURE);
|
|
}
|
|
|
|
// set fake free disk space
|
|
if(-1 != fake_diskfree_size){
|
|
FdManager::InitFakeUsedDiskSize(fake_diskfree_size);
|
|
}
|
|
|
|
// set user agent
|
|
S3fsCurl::InitUserAgent();
|
|
|
|
// There's room for more command line error checking
|
|
|
|
// Check to see if the bucket name contains periods and https (SSL) is
|
|
// being used. This is a known limitation:
|
|
// https://docs.amazonwebservices.com/AmazonS3/latest/dev/
|
|
// The Developers Guide suggests that either use HTTP of for us to write
|
|
// our own certificate verification logic.
|
|
// For now, this will be unsupported unless we get a request for it to
|
|
// be supported. In that case, we have a couple of options:
|
|
// - implement a command line option that bypasses the verify host
|
|
// but doesn't bypass verifying the certificate
|
|
// - write our own host verification (this might be complex)
|
|
// See issue #128strncasecmp
|
|
/*
|
|
if(1 == S3fsCurl::GetSslVerifyHostname()){
|
|
found = S3fsCred::GetBucket().find_first_of('.');
|
|
if(found != std::string::npos){
|
|
found = s3host.find("https:");
|
|
if(found != std::string::npos){
|
|
S3FS_PRN_EXIT("Using https and a bucket name with periods is unsupported.");
|
|
S3fsCurl::DestroyS3fsCurl();
|
|
s3fs_destroy_global_ssl();
|
|
destroy_parser_xml_lock();
|
|
delete ps3fscred;
|
|
exit(EXIT_FAILURE);
|
|
}
|
|
}
|
|
}
|
|
*/
|
|
|
|
if(NO_UTILITY_MODE != utility_mode){
|
|
int exitcode = s3fs_utility_processing(incomp_abort_time);
|
|
|
|
S3fsCurl::DestroyS3fsCurl();
|
|
s3fs_destroy_global_ssl();
|
|
destroy_parser_xml_lock();
|
|
destroy_basename_lock();
|
|
delete ps3fscred;
|
|
exit(exitcode);
|
|
}
|
|
|
|
// Check multipart / copy api for mix multipart uploading
|
|
if(nomultipart || nocopyapi || norenameapi){
|
|
FdEntity::SetNoMixMultipart();
|
|
max_dirty_data = -1;
|
|
}
|
|
|
|
if(!ThreadPoolMan::Initialize(max_thread_count)){
|
|
S3FS_PRN_EXIT("Could not create thread pool(%d)", max_thread_count);
|
|
S3fsCurl::DestroyS3fsCurl();
|
|
s3fs_destroy_global_ssl();
|
|
destroy_parser_xml_lock();
|
|
destroy_basename_lock();
|
|
exit(EXIT_FAILURE);
|
|
}
|
|
|
|
// check free disk space
|
|
if(!FdManager::IsSafeDiskSpace(NULL, S3fsCurl::GetMultipartSize() * S3fsCurl::GetMaxParallelCount())){
|
|
S3FS_PRN_EXIT("There is no enough disk space for used as cache(or temporary) directory by s3fs.");
|
|
S3fsCurl::DestroyS3fsCurl();
|
|
s3fs_destroy_global_ssl();
|
|
destroy_parser_xml_lock();
|
|
destroy_basename_lock();
|
|
delete ps3fscred;
|
|
exit(EXIT_FAILURE);
|
|
}
|
|
|
|
s3fs_oper.getattr = s3fs_getattr;
|
|
s3fs_oper.readlink = s3fs_readlink;
|
|
s3fs_oper.mknod = s3fs_mknod;
|
|
s3fs_oper.mkdir = s3fs_mkdir;
|
|
s3fs_oper.unlink = s3fs_unlink;
|
|
s3fs_oper.rmdir = s3fs_rmdir;
|
|
s3fs_oper.symlink = s3fs_symlink;
|
|
s3fs_oper.rename = s3fs_rename;
|
|
s3fs_oper.link = s3fs_link;
|
|
if(!nocopyapi){
|
|
s3fs_oper.chmod = s3fs_chmod;
|
|
s3fs_oper.chown = s3fs_chown;
|
|
s3fs_oper.utimens = s3fs_utimens;
|
|
}else{
|
|
s3fs_oper.chmod = s3fs_chmod_nocopy;
|
|
s3fs_oper.chown = s3fs_chown_nocopy;
|
|
s3fs_oper.utimens = s3fs_utimens_nocopy;
|
|
}
|
|
s3fs_oper.truncate = s3fs_truncate;
|
|
s3fs_oper.open = s3fs_open;
|
|
s3fs_oper.read = s3fs_read;
|
|
s3fs_oper.write = s3fs_write;
|
|
s3fs_oper.statfs = s3fs_statfs;
|
|
s3fs_oper.flush = s3fs_flush;
|
|
s3fs_oper.fsync = s3fs_fsync;
|
|
s3fs_oper.release = s3fs_release;
|
|
s3fs_oper.opendir = s3fs_opendir;
|
|
s3fs_oper.readdir = s3fs_readdir;
|
|
s3fs_oper.init = s3fs_init;
|
|
s3fs_oper.destroy = s3fs_destroy;
|
|
s3fs_oper.access = s3fs_access;
|
|
s3fs_oper.create = s3fs_create;
|
|
// extended attributes
|
|
if(is_use_xattr){
|
|
s3fs_oper.setxattr = s3fs_setxattr;
|
|
s3fs_oper.getxattr = s3fs_getxattr;
|
|
s3fs_oper.listxattr = s3fs_listxattr;
|
|
s3fs_oper.removexattr = s3fs_removexattr;
|
|
}
|
|
s3fs_oper.flag_utime_omit_ok = true;
|
|
|
|
// now passing things off to fuse, fuse will finish evaluating the command line args
|
|
fuse_res = fuse_main(custom_args.argc, custom_args.argv, &s3fs_oper, NULL);
|
|
if(fuse_res == 0){
|
|
fuse_res = s3fs_init_deferred_exit_status;
|
|
}
|
|
fuse_opt_free_args(&custom_args);
|
|
|
|
// Destroy thread pool
|
|
ThreadPoolMan::Destroy();
|
|
|
|
// Destroy curl
|
|
if(!S3fsCurl::DestroyS3fsCurl()){
|
|
S3FS_PRN_WARN("Could not release curl library.");
|
|
}
|
|
s3fs_destroy_global_ssl();
|
|
destroy_parser_xml_lock();
|
|
destroy_basename_lock();
|
|
delete ps3fscred;
|
|
|
|
// cleanup xml2
|
|
xmlCleanupParser();
|
|
S3FS_MALLOCTRIM(0);
|
|
|
|
exit(fuse_res);
|
|
}
|
|
|
|
/*
|
|
* Local variables:
|
|
* tab-width: 4
|
|
* c-basic-offset: 4
|
|
* End:
|
|
* vim600: expandtab sw=4 ts=4 fdm=marker
|
|
* vim<600: expandtab sw=4 ts=4
|
|
*/
|