mirror of
https://github.com/s3fs-fuse/s3fs-fuse.git
synced 2024-11-15 08:54:07 +00:00
2837 lines
102 KiB
Bash
Executable File
2837 lines
102 KiB
Bash
Executable File
#!/bin/bash
|
|
#
|
|
# s3fs - FUSE-based file system backed by Amazon S3
|
|
#
|
|
# Copyright 2007-2008 Randy Rizun <rrizun@gmail.com>
|
|
#
|
|
# This program is free software; you can redistribute it and/or
|
|
# modify it under the terms of the GNU General Public License
|
|
# as published by the Free Software Foundation; either version 2
|
|
# of the License, or (at your option) any later version.
|
|
#
|
|
# This program is distributed in the hope that it will be useful,
|
|
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
# GNU General Public License for more details.
|
|
#
|
|
# You should have received a copy of the GNU General Public License
|
|
# along with this program; if not, write to the Free Software
|
|
# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.
|
|
#
|
|
|
|
set -o errexit
|
|
set -o pipefail
|
|
|
|
source test-utils.sh
|
|
|
|
function test_create_empty_file {
|
|
describe "Testing creating an empty file ..."
|
|
|
|
local OBJECT_NAME; OBJECT_NAME=$(basename "${PWD}")/"${TEST_TEXT_FILE}"
|
|
|
|
touch "${TEST_TEXT_FILE}"
|
|
|
|
check_file_size "${TEST_TEXT_FILE}" 0
|
|
|
|
aws_cli s3api head-object --bucket "${TEST_BUCKET_1}" --key "${OBJECT_NAME}"
|
|
|
|
rm_test_file
|
|
}
|
|
|
|
function test_append_file {
|
|
describe "Testing append to file ..."
|
|
local TEST_INPUT="echo ${TEST_TEXT} to ${TEST_TEXT_FILE}"
|
|
|
|
# Write a small test file
|
|
for x in $(seq 1 "${TEST_TEXT_FILE_LENGTH}"); do
|
|
echo "${TEST_INPUT}"
|
|
done > "${TEST_TEXT_FILE}"
|
|
|
|
check_file_size "${TEST_TEXT_FILE}" $((TEST_TEXT_FILE_LENGTH * $((${#TEST_INPUT} + 1)) ))
|
|
|
|
rm_test_file
|
|
}
|
|
|
|
function test_truncate_file {
|
|
describe "Testing truncate file ..."
|
|
# Write a small test file
|
|
echo "${TEST_TEXT}" > "${TEST_TEXT_FILE}"
|
|
|
|
# Truncate file to 0 length. This should trigger open(path, O_RDWR | O_TRUNC...)
|
|
: > "${TEST_TEXT_FILE}"
|
|
|
|
check_file_size "${TEST_TEXT_FILE}" 0
|
|
|
|
rm_test_file
|
|
}
|
|
|
|
function test_truncate_upload {
|
|
describe "Testing truncate file for uploading ..."
|
|
|
|
# This file size uses multipart, mix upload when uploading.
|
|
# We will test these cases.
|
|
|
|
rm_test_file "${BIG_FILE}"
|
|
|
|
"${TRUNCATE_BIN}" "${BIG_FILE}" -s "${BIG_FILE_LENGTH}"
|
|
|
|
rm_test_file "${BIG_FILE}"
|
|
}
|
|
|
|
function test_truncate_empty_file {
|
|
describe "Testing truncate empty file ..."
|
|
# Write an empty test file
|
|
touch "${TEST_TEXT_FILE}"
|
|
|
|
# Truncate the file to 1024 length
|
|
local t_size=1024
|
|
"${TRUNCATE_BIN}" "${TEST_TEXT_FILE}" -s "${t_size}"
|
|
|
|
check_file_size "${TEST_TEXT_FILE}" "${t_size}"
|
|
|
|
rm_test_file
|
|
}
|
|
|
|
function test_truncate_shrink_file {
|
|
describe "Testing truncate shrinking large binary file ..."
|
|
|
|
local BIG_TRUNCATE_TEST_FILE="big-truncate-test.bin"
|
|
local t_size=$((1024 * 1024 * 32 + 64))
|
|
|
|
dd if=/dev/urandom of="${TEMP_DIR}/${BIG_TRUNCATE_TEST_FILE}" bs=1024 count=$((1024 * 64))
|
|
cp "${TEMP_DIR}/${BIG_TRUNCATE_TEST_FILE}" "${BIG_TRUNCATE_TEST_FILE}"
|
|
|
|
"${TRUNCATE_BIN}" "${TEMP_DIR}/${BIG_TRUNCATE_TEST_FILE}" -s "${t_size}"
|
|
"${TRUNCATE_BIN}" "${BIG_TRUNCATE_TEST_FILE}" -s "${t_size}"
|
|
|
|
if ! cmp "${TEMP_DIR}/${BIG_TRUNCATE_TEST_FILE}" "${BIG_TRUNCATE_TEST_FILE}"; then
|
|
return 1
|
|
fi
|
|
|
|
rm -f "${TEMP_DIR}/${BIG_TRUNCATE_TEST_FILE}"
|
|
rm_test_file "${BIG_TRUNCATE_TEST_FILE}"
|
|
}
|
|
|
|
function test_truncate_shrink_read_file {
|
|
describe "Testing truncate(shrink) and read file ..."
|
|
|
|
# Initiate file size is 1024, and shrink file size is 512
|
|
local init_size=1024
|
|
local shrink_size=512
|
|
|
|
# create file
|
|
dd if=/dev/urandom of="${TEST_TEXT_FILE}" bs="${init_size}" count=1
|
|
|
|
# truncate(shrink) file and read it before flushing
|
|
../../truncate_read_file "${TEST_TEXT_FILE}" "${shrink_size}"
|
|
|
|
# check file size
|
|
check_file_size "${TEST_TEXT_FILE}" "${shrink_size}"
|
|
|
|
# Truncate the file to 1024 length
|
|
local t_size=1024
|
|
|
|
rm_test_file
|
|
}
|
|
|
|
function test_mv_file {
|
|
describe "Testing mv file function ..."
|
|
# if the rename file exists, delete it
|
|
if [ -e "${ALT_TEST_TEXT_FILE}" ]
|
|
then
|
|
rm "${ALT_TEST_TEXT_FILE}"
|
|
fi
|
|
|
|
if [ -e "${ALT_TEST_TEXT_FILE}" ]
|
|
then
|
|
echo "Could not delete file ${ALT_TEST_TEXT_FILE}, it still exists"
|
|
return 1
|
|
fi
|
|
|
|
# create the test file again
|
|
mk_test_file
|
|
|
|
# save file length
|
|
local ALT_TEXT_LENGTH; ALT_TEXT_LENGTH=$(wc -c "${TEST_TEXT_FILE}" | awk '{print $1}')
|
|
|
|
#rename the test file
|
|
mv "${TEST_TEXT_FILE}" "${ALT_TEST_TEXT_FILE}"
|
|
if [ ! -e "${ALT_TEST_TEXT_FILE}" ]
|
|
then
|
|
echo "Could not move file"
|
|
return 1
|
|
fi
|
|
|
|
#check the renamed file content-type
|
|
if [ -f "/etc/mime.types" ]
|
|
then
|
|
check_content_type "$1/${ALT_TEST_TEXT_FILE}" "text/plain"
|
|
fi
|
|
|
|
# Check the contents of the alt file
|
|
local ALT_FILE_LENGTH; ALT_FILE_LENGTH=$(wc -c "${ALT_TEST_TEXT_FILE}" | awk '{print $1}')
|
|
if [ "$ALT_FILE_LENGTH" -ne "$ALT_TEXT_LENGTH" ]
|
|
then
|
|
echo "moved file length is not as expected expected: $ALT_TEXT_LENGTH got: $ALT_FILE_LENGTH"
|
|
return 1
|
|
fi
|
|
|
|
# clean up
|
|
rm_test_file "${ALT_TEST_TEXT_FILE}"
|
|
}
|
|
|
|
function test_mv_to_exist_file {
|
|
describe "Testing mv file to exist file function ..."
|
|
|
|
local BIG_MV_FILE_BLOCK_SIZE=$((BIG_FILE_BLOCK_SIZE + 1))
|
|
|
|
../../junk_data $((BIG_FILE_BLOCK_SIZE * BIG_FILE_COUNT)) > "${BIG_FILE}"
|
|
../../junk_data $((BIG_MV_FILE_BLOCK_SIZE * BIG_FILE_COUNT)) > "${BIG_FILE}-mv"
|
|
|
|
mv "${BIG_FILE}" "${BIG_FILE}-mv"
|
|
|
|
rm_test_file "${BIG_FILE}-mv"
|
|
}
|
|
|
|
function test_mv_empty_directory {
|
|
describe "Testing mv directory function ..."
|
|
if [ -e "${TEST_DIR}" ]; then
|
|
echo "Unexpected, this file/directory exists: ${TEST_DIR}"
|
|
return 1
|
|
fi
|
|
|
|
mk_test_dir
|
|
|
|
mv "${TEST_DIR}" "${TEST_DIR}_rename"
|
|
if [ ! -d "${TEST_DIR}_rename" ]; then
|
|
echo "Directory ${TEST_DIR} was not renamed"
|
|
return 1
|
|
fi
|
|
|
|
rmdir "${TEST_DIR}_rename"
|
|
if [ -e "${TEST_DIR}_rename" ]; then
|
|
echo "Could not remove the test directory, it still exists: ${TEST_DIR}_rename"
|
|
return 1
|
|
fi
|
|
}
|
|
|
|
function test_mv_nonempty_directory {
|
|
describe "Testing mv directory function ..."
|
|
if [ -e "${TEST_DIR}" ]; then
|
|
echo "Unexpected, this file/directory exists: ${TEST_DIR}"
|
|
return 1
|
|
fi
|
|
|
|
mk_test_dir
|
|
|
|
touch "${TEST_DIR}"/file
|
|
|
|
mv "${TEST_DIR}" "${TEST_DIR}_rename"
|
|
if [ ! -d "${TEST_DIR}_rename" ]; then
|
|
echo "Directory ${TEST_DIR} was not renamed"
|
|
return 1
|
|
fi
|
|
|
|
rm -r "${TEST_DIR}_rename"
|
|
if [ -e "${TEST_DIR}_rename" ]; then
|
|
echo "Could not remove the test directory, it still exists: ${TEST_DIR}_rename"
|
|
return 1
|
|
fi
|
|
}
|
|
|
|
function test_redirects {
|
|
describe "Testing redirects ..."
|
|
|
|
mk_test_file "ABCDEF"
|
|
|
|
local CONTENT; CONTENT=$(cat "${TEST_TEXT_FILE}")
|
|
|
|
if [ "${CONTENT}" != "ABCDEF" ]; then
|
|
echo "CONTENT read is unexpected, got ${CONTENT}, expected ABCDEF"
|
|
return 1
|
|
fi
|
|
|
|
echo "XYZ" > "${TEST_TEXT_FILE}"
|
|
|
|
CONTENT=$(cat "${TEST_TEXT_FILE}")
|
|
|
|
if [ "${CONTENT}" != "XYZ" ]; then
|
|
echo "CONTENT read is unexpected, got ${CONTENT}, expected XYZ"
|
|
return 1
|
|
fi
|
|
|
|
echo "123456" >> "${TEST_TEXT_FILE}"
|
|
|
|
local LINE1; LINE1=$("${SED_BIN}" -n '1,1p' "${TEST_TEXT_FILE}")
|
|
local LINE2; LINE2=$("${SED_BIN}" -n '2,2p' "${TEST_TEXT_FILE}")
|
|
|
|
if [ "${LINE1}" != "XYZ" ]; then
|
|
echo "LINE1 was not as expected, got ${LINE1}, expected XYZ"
|
|
return 1
|
|
fi
|
|
|
|
if [ "${LINE2}" != "123456" ]; then
|
|
echo "LINE2 was not as expected, got ${LINE2}, expected 123456"
|
|
return 1
|
|
fi
|
|
|
|
# clean up
|
|
rm_test_file
|
|
}
|
|
|
|
function test_mkdir_rmdir {
|
|
describe "Testing creation/removal of a directory ..."
|
|
|
|
if [ -e "${TEST_DIR}" ]; then
|
|
echo "Unexpected, this file/directory exists: ${TEST_DIR}"
|
|
return 1
|
|
fi
|
|
|
|
mk_test_dir
|
|
rm_test_dir
|
|
}
|
|
|
|
function test_chmod {
|
|
describe "Testing chmod file function ..."
|
|
|
|
# create the test file again
|
|
mk_test_file
|
|
|
|
local ORIGINAL_PERMISSIONS; ORIGINAL_PERMISSIONS=$(get_permissions "${TEST_TEXT_FILE}")
|
|
|
|
chmod 777 "${TEST_TEXT_FILE}";
|
|
|
|
# if they're the same, we have a problem.
|
|
local CHANGED_PERMISSIONS; CHANGED_PERMISSIONS=$(get_permissions "${TEST_TEXT_FILE}")
|
|
if [ "${CHANGED_PERMISSIONS}" = "${ORIGINAL_PERMISSIONS}" ]
|
|
then
|
|
echo "Could not modify ${TEST_TEXT_FILE} permissions"
|
|
return 1
|
|
fi
|
|
|
|
# clean up
|
|
rm_test_file
|
|
}
|
|
|
|
function test_chown {
|
|
describe "Testing chown file function ..."
|
|
|
|
# create the test file again
|
|
mk_test_file
|
|
|
|
local ORIGINAL_PERMISSIONS
|
|
ORIGINAL_PERMISSIONS=$(get_user_and_group "${TEST_TEXT_FILE}")
|
|
|
|
# [NOTE]
|
|
# Prevents test interruptions due to permission errors, etc.
|
|
# If the chown command fails, an error will occur with the
|
|
# following judgment statement. So skip the chown command error.
|
|
# '|| true' was added due to a problem with Travis CI and MacOS
|
|
# and ensure_diskfree option.
|
|
#
|
|
chown 1000:1000 "${TEST_TEXT_FILE}" || true
|
|
|
|
# if they're the same, we have a problem.
|
|
local CHANGED_PERMISSIONS
|
|
CHANGED_PERMISSIONS=$(get_user_and_group "${TEST_TEXT_FILE}")
|
|
if [ "${CHANGED_PERMISSIONS}" = "${ORIGINAL_PERMISSIONS}" ]
|
|
then
|
|
if [ "${ORIGINAL_PERMISSIONS}" = "1000:1000" ]
|
|
then
|
|
echo "Could not be strict check because original file permission 1000:1000"
|
|
else
|
|
echo "Could not modify ${TEST_TEXT_FILE} ownership($ORIGINAL_PERMISSIONS to 1000:1000)"
|
|
return 1
|
|
fi
|
|
fi
|
|
|
|
# clean up
|
|
rm_test_file
|
|
}
|
|
|
|
function test_list {
|
|
describe "Testing list ..."
|
|
mk_test_file
|
|
mk_test_dir
|
|
|
|
local file_list=(*)
|
|
local file_cnt=${#file_list[@]}
|
|
if [ "${file_cnt}" -ne 2 ]; then
|
|
echo "Expected 2 file but got ${file_cnt}"
|
|
echo "Files: " "${file_list[@]}"
|
|
return 1
|
|
fi
|
|
|
|
rm_test_file
|
|
rm_test_dir
|
|
}
|
|
|
|
function test_remove_nonempty_directory {
|
|
describe "Testing removing a non-empty directory ..."
|
|
mk_test_dir
|
|
touch "${TEST_DIR}/file"
|
|
(
|
|
set +o pipefail
|
|
rmdir "${TEST_DIR}" 2>&1 | grep -q "Directory not empty"
|
|
)
|
|
rm "${TEST_DIR}/file"
|
|
rm_test_dir
|
|
}
|
|
|
|
function test_external_directory_creation {
|
|
describe "Test external directory creation ..."
|
|
local OBJECT_NAME; OBJECT_NAME=$(basename "${PWD}")/directory/"${TEST_TEXT_FILE}"
|
|
echo "data" | aws_cli s3 cp - "s3://${TEST_BUCKET_1}/${OBJECT_NAME}"
|
|
# shellcheck disable=SC2010
|
|
ls | grep -q directory
|
|
stat directory >/dev/null 2>&1
|
|
get_permissions directory | grep -q 750$
|
|
ls directory
|
|
cmp <(echo "data") directory/"${TEST_TEXT_FILE}"
|
|
rm -f directory/"${TEST_TEXT_FILE}"
|
|
}
|
|
|
|
function test_external_modification {
|
|
describe "Test external modification to an object ..."
|
|
echo "old" > "${TEST_TEXT_FILE}"
|
|
|
|
# [NOTE]
|
|
# If the stat and file cache directory are enabled, an error will
|
|
# occur if the unixtime(sec) value does not change.
|
|
# If mtime(ctime/atime) when updating from the external program
|
|
# (awscli) is the same unixtime value as immediately before, the
|
|
# cache will be read out.
|
|
# Therefore, we need to wait over 1 second here.
|
|
#
|
|
sleep 1
|
|
|
|
local OBJECT_NAME; OBJECT_NAME=$(basename "${PWD}")/"${TEST_TEXT_FILE}"
|
|
echo "new new" | aws_cli s3 cp - "s3://${TEST_BUCKET_1}/${OBJECT_NAME}"
|
|
|
|
cmp "${TEST_TEXT_FILE}" <(echo "new new")
|
|
rm -f "${TEST_TEXT_FILE}"
|
|
}
|
|
|
|
function test_external_creation {
|
|
describe "Test external creation of an object ..."
|
|
local OBJECT_NAME; OBJECT_NAME=$(basename "${PWD}")/"${TEST_TEXT_FILE}"
|
|
|
|
# [NOTE]
|
|
# If noobj_cache is enabled, register that cache here.
|
|
#
|
|
[ ! -e "${TEST_TEXT_FILE}" ]
|
|
|
|
# [NOTE]
|
|
# If noobj_cache is enabled, we cannot be sure that it is registered in that cache.
|
|
# That's because an error will occur if the upload by aws cli takes more than 1 second.
|
|
#
|
|
echo "data" | aws_cli s3 cp - "s3://${TEST_BUCKET_1}/${OBJECT_NAME}"
|
|
|
|
wait_ostype 1
|
|
|
|
[ -e "${TEST_TEXT_FILE}" ]
|
|
|
|
rm -f "${TEST_TEXT_FILE}"
|
|
}
|
|
|
|
function test_read_external_object() {
|
|
describe "create objects via aws CLI and read via s3fs ..."
|
|
local OBJECT_NAME; OBJECT_NAME=$(basename "${PWD}")/"${TEST_TEXT_FILE}"
|
|
echo "test" | aws_cli s3 cp - "s3://${TEST_BUCKET_1}/${OBJECT_NAME}"
|
|
cmp "${TEST_TEXT_FILE}" <(echo "test")
|
|
rm -f "${TEST_TEXT_FILE}"
|
|
}
|
|
|
|
function test_read_external_dir_object() {
|
|
describe "create directory objects via aws CLI and read via s3fs ..."
|
|
local SUB_DIR_NAME; SUB_DIR_NAME="subdir"
|
|
local SUB_DIR_TEST_FILE; SUB_DIR_TEST_FILE="${SUB_DIR_NAME}/${TEST_TEXT_FILE}"
|
|
local OBJECT_NAME; OBJECT_NAME=$(basename "${PWD}")/"${SUB_DIR_TEST_FILE}"
|
|
|
|
echo "test" | aws_cli s3 cp - "s3://${TEST_BUCKET_1}/${OBJECT_NAME}"
|
|
|
|
if stat "${SUB_DIR_NAME}" | grep -q '1969-12-31[[:space:]]23:59:59[.]000000000'; then
|
|
echo "sub directory a/c/m time is underflow(-1)."
|
|
return 1
|
|
fi
|
|
rm -rf "${SUB_DIR_NAME}"
|
|
}
|
|
|
|
function test_update_metadata_external_small_object() {
|
|
describe "update meta to small file after created file by aws cli"
|
|
|
|
# [NOTE]
|
|
# Use the only filename in the test to avoid being affected by noobjcache.
|
|
#
|
|
local TEST_FILE_EXT; TEST_FILE_EXT=$(make_random_string)
|
|
local TEST_CHMOD_FILE="${TEST_TEXT_FILE}_chmod.${TEST_FILE_EXT}"
|
|
local TEST_CHOWN_FILE="${TEST_TEXT_FILE}_chown.${TEST_FILE_EXT}"
|
|
local TEST_UTIMENS_FILE="${TEST_TEXT_FILE}_utimens.${TEST_FILE_EXT}"
|
|
local TEST_SETXATTR_FILE="${TEST_TEXT_FILE}_xattr.${TEST_FILE_EXT}"
|
|
local TEST_RMXATTR_FILE="${TEST_TEXT_FILE}_xattr.${TEST_FILE_EXT}"
|
|
|
|
local TEST_INPUT="TEST_STRING_IN_SMALL_FILE"
|
|
|
|
#
|
|
# chmod
|
|
#
|
|
local OBJECT_NAME; OBJECT_NAME=$(basename "${PWD}")/"${TEST_CHMOD_FILE}"
|
|
echo "${TEST_INPUT}" | aws_cli s3 cp - "s3://${TEST_BUCKET_1}/${OBJECT_NAME}"
|
|
chmod +x "${TEST_CHMOD_FILE}"
|
|
cmp "${TEST_CHMOD_FILE}" <(echo "${TEST_INPUT}")
|
|
|
|
#
|
|
# chown
|
|
#
|
|
OBJECT_NAME=$(basename "${PWD}")/"${TEST_CHOWN_FILE}"
|
|
echo "${TEST_INPUT}" | aws_cli s3 cp - "s3://${TEST_BUCKET_1}/${OBJECT_NAME}"
|
|
chown "${UID}" "${TEST_CHOWN_FILE}"
|
|
cmp "${TEST_CHOWN_FILE}" <(echo "${TEST_INPUT}")
|
|
|
|
#
|
|
# utimens
|
|
#
|
|
OBJECT_NAME=$(basename "${PWD}")/"${TEST_UTIMENS_FILE}"
|
|
echo "${TEST_INPUT}" | aws_cli s3 cp - "s3://${TEST_BUCKET_1}/${OBJECT_NAME}"
|
|
touch "${TEST_UTIMENS_FILE}"
|
|
cmp "${TEST_UTIMENS_FILE}" <(echo "${TEST_INPUT}")
|
|
|
|
#
|
|
# set xattr
|
|
#
|
|
OBJECT_NAME=$(basename "${PWD}")/"${TEST_SETXATTR_FILE}"
|
|
echo "${TEST_INPUT}" | aws_cli s3 cp - "s3://${TEST_BUCKET_1}/${OBJECT_NAME}"
|
|
set_xattr key value "${TEST_SETXATTR_FILE}"
|
|
cmp "${TEST_SETXATTR_FILE}" <(echo "${TEST_INPUT}")
|
|
|
|
#
|
|
# remove xattr
|
|
#
|
|
# "%7B%22key%22%3A%22dmFsdWU%3D%22%7D" = {"key":"value"}
|
|
#
|
|
OBJECT_NAME=$(basename "${PWD}")/"${TEST_RMXATTR_FILE}"
|
|
echo "${TEST_INPUT}" | aws_cli s3 cp - "s3://${TEST_BUCKET_1}/${OBJECT_NAME}" --metadata xattr=%7B%22key%22%3A%22dmFsdWU%3D%22%7D
|
|
del_xattr key "${TEST_RMXATTR_FILE}"
|
|
cmp "${TEST_RMXATTR_FILE}" <(echo "${TEST_INPUT}")
|
|
|
|
rm -f "${TEST_CHMOD_FILE}"
|
|
rm -f "${TEST_CHOWN_FILE}"
|
|
rm -f "${TEST_UTIMENS_FILE}"
|
|
rm -f "${TEST_SETXATTR_FILE}"
|
|
rm -f "${TEST_RMXATTR_FILE}"
|
|
}
|
|
|
|
function test_update_metadata_external_large_object() {
|
|
describe "update meta to large file after created file by aws cli"
|
|
|
|
# [NOTE]
|
|
# Use the only filename in the test to avoid being affected by noobjcache.
|
|
#
|
|
local TEST_FILE_EXT; TEST_FILE_EXT=$(make_random_string)
|
|
local TEST_CHMOD_FILE="${TEST_TEXT_FILE}_chmod.${TEST_FILE_EXT}"
|
|
local TEST_CHOWN_FILE="${TEST_TEXT_FILE}_chown.${TEST_FILE_EXT}"
|
|
local TEST_UTIMENS_FILE="${TEST_TEXT_FILE}_utimens.${TEST_FILE_EXT}"
|
|
local TEST_SETXATTR_FILE="${TEST_TEXT_FILE}_xattr.${TEST_FILE_EXT}"
|
|
local TEST_RMXATTR_FILE="${TEST_TEXT_FILE}_xattr.${TEST_FILE_EXT}"
|
|
|
|
../../junk_data $((BIG_FILE_BLOCK_SIZE * BIG_FILE_COUNT)) > "${TEMP_DIR}/${BIG_FILE}"
|
|
|
|
#
|
|
# chmod
|
|
#
|
|
local OBJECT_NAME; OBJECT_NAME=$(basename "${PWD}")/"${TEST_CHMOD_FILE}"
|
|
aws_cli s3 cp "${TEMP_DIR}/${BIG_FILE}" "s3://${TEST_BUCKET_1}/${OBJECT_NAME}" --no-progress
|
|
chmod +x "${TEST_CHMOD_FILE}"
|
|
cmp "${TEST_CHMOD_FILE}" "${TEMP_DIR}/${BIG_FILE}"
|
|
|
|
#
|
|
# chown
|
|
#
|
|
OBJECT_NAME=$(basename "${PWD}")/"${TEST_CHOWN_FILE}"
|
|
aws_cli s3 cp "${TEMP_DIR}/${BIG_FILE}" "s3://${TEST_BUCKET_1}/${OBJECT_NAME}" --no-progress
|
|
chown "${UID}" "${TEST_CHOWN_FILE}"
|
|
cmp "${TEST_CHOWN_FILE}" "${TEMP_DIR}/${BIG_FILE}"
|
|
|
|
#
|
|
# utimens
|
|
#
|
|
OBJECT_NAME=$(basename "${PWD}")/"${TEST_UTIMENS_FILE}"
|
|
aws_cli s3 cp "${TEMP_DIR}/${BIG_FILE}" "s3://${TEST_BUCKET_1}/${OBJECT_NAME}" --no-progress
|
|
touch "${TEST_UTIMENS_FILE}"
|
|
cmp "${TEST_UTIMENS_FILE}" "${TEMP_DIR}/${BIG_FILE}"
|
|
|
|
#
|
|
# set xattr
|
|
#
|
|
OBJECT_NAME=$(basename "${PWD}")/"${TEST_SETXATTR_FILE}"
|
|
aws_cli s3 cp "${TEMP_DIR}/${BIG_FILE}" "s3://${TEST_BUCKET_1}/${OBJECT_NAME}" --no-progress
|
|
set_xattr key value "${TEST_SETXATTR_FILE}"
|
|
cmp "${TEST_SETXATTR_FILE}" "${TEMP_DIR}/${BIG_FILE}"
|
|
|
|
#
|
|
# remove xattr
|
|
#
|
|
# "%7B%22key%22%3A%22dmFsdWU%3D%22%7D" = {"key":"value"}
|
|
#
|
|
OBJECT_NAME=$(basename "${PWD}")/"${TEST_RMXATTR_FILE}"
|
|
aws_cli s3 cp "${TEMP_DIR}/${BIG_FILE}" "s3://${TEST_BUCKET_1}/${OBJECT_NAME}" --no-progress --metadata xattr=%7B%22key%22%3A%22dmFsdWU%3D%22%7D
|
|
del_xattr key "${TEST_RMXATTR_FILE}"
|
|
cmp "${TEST_RMXATTR_FILE}" "${TEMP_DIR}/${BIG_FILE}"
|
|
|
|
rm -f "${TEMP_DIR}/${BIG_FILE}"
|
|
rm -f "${TEST_CHMOD_FILE}"
|
|
rm -f "${TEST_CHOWN_FILE}"
|
|
rm -f "${TEST_UTIMENS_FILE}"
|
|
rm -f "${TEST_SETXATTR_FILE}"
|
|
rm -f "${TEST_RMXATTR_FILE}"
|
|
}
|
|
|
|
function test_rename_before_close {
|
|
describe "Testing rename before close ..."
|
|
|
|
# shellcheck disable=SC2094
|
|
(
|
|
echo foo
|
|
mv "${TEST_TEXT_FILE}" "${TEST_TEXT_FILE}.new"
|
|
) > "${TEST_TEXT_FILE}"
|
|
|
|
if ! cmp <(echo "foo") "${TEST_TEXT_FILE}.new"; then
|
|
echo "rename before close failed"
|
|
return 1
|
|
fi
|
|
|
|
rm_test_file "${TEST_TEXT_FILE}.new"
|
|
rm -f "${TEST_TEXT_FILE}"
|
|
}
|
|
|
|
function test_multipart_upload {
|
|
describe "Testing multi-part upload ..."
|
|
|
|
../../junk_data $((BIG_FILE_BLOCK_SIZE * BIG_FILE_COUNT)) > "${TEMP_DIR}/${BIG_FILE}"
|
|
dd if="${TEMP_DIR}/${BIG_FILE}" of="${BIG_FILE}" bs="${BIG_FILE_BLOCK_SIZE}" count="${BIG_FILE_COUNT}"
|
|
|
|
# Verify contents of file
|
|
echo "Comparing test file"
|
|
if ! cmp "${TEMP_DIR}/${BIG_FILE}" "${BIG_FILE}"
|
|
then
|
|
return 1
|
|
fi
|
|
|
|
rm -f "${TEMP_DIR}/${BIG_FILE}"
|
|
rm_test_file "${BIG_FILE}"
|
|
}
|
|
|
|
function test_multipart_copy {
|
|
describe "Testing multi-part copy ..."
|
|
|
|
../../junk_data $((BIG_FILE_BLOCK_SIZE * BIG_FILE_COUNT)) > "${TEMP_DIR}/${BIG_FILE}"
|
|
dd if="${TEMP_DIR}/${BIG_FILE}" of="${BIG_FILE}" bs="${BIG_FILE_BLOCK_SIZE}" count="${BIG_FILE_COUNT}"
|
|
mv "${BIG_FILE}" "${BIG_FILE}-copy"
|
|
|
|
# Verify contents of file
|
|
echo "Comparing test file"
|
|
if ! cmp "${TEMP_DIR}/${BIG_FILE}" "${BIG_FILE}-copy"
|
|
then
|
|
return 1
|
|
fi
|
|
|
|
#check the renamed file content-type
|
|
check_content_type "$1/${BIG_FILE}-copy" "application/octet-stream"
|
|
|
|
rm -f "${TEMP_DIR}/${BIG_FILE}"
|
|
rm_test_file "${BIG_FILE}-copy"
|
|
}
|
|
|
|
function test_multipart_mix {
|
|
describe "Testing multi-part mix ..."
|
|
|
|
../../junk_data $((BIG_FILE_BLOCK_SIZE * BIG_FILE_COUNT)) > "${TEMP_DIR}/${BIG_FILE}"
|
|
dd if="${TEMP_DIR}/${BIG_FILE}" of="${BIG_FILE}" bs="${BIG_FILE_BLOCK_SIZE}" count="${BIG_FILE_COUNT}"
|
|
|
|
# (1) Edit the middle of an existing file
|
|
# modify directly(seek 7.5MB offset)
|
|
# In the case of nomultipart and nocopyapi,
|
|
# it makes no sense, but copying files is because it leaves no cache.
|
|
#
|
|
cp "${TEMP_DIR}/${BIG_FILE}" "${TEMP_DIR}/${BIG_FILE}-mix"
|
|
cp "${BIG_FILE}" "${BIG_FILE}-mix"
|
|
|
|
local MODIFY_START_BLOCK=$((15*1024*1024/2/4))
|
|
echo -n "0123456789ABCDEF" | dd of="${BIG_FILE}-mix" bs=4 count=4 seek="${MODIFY_START_BLOCK}" conv=notrunc
|
|
echo -n "0123456789ABCDEF" | dd of="${TEMP_DIR}/${BIG_FILE}-mix" bs=4 count=4 seek="${MODIFY_START_BLOCK}" conv=notrunc
|
|
|
|
# Verify contents of file
|
|
echo "Comparing test file (1)"
|
|
if ! cmp "${TEMP_DIR}/${BIG_FILE}-mix" "${BIG_FILE}-mix"
|
|
then
|
|
return 1
|
|
fi
|
|
|
|
# (2) Write to an area larger than the size of the existing file
|
|
# modify directly(over file end offset)
|
|
#
|
|
cp "${TEMP_DIR}/${BIG_FILE}" "${TEMP_DIR}/${BIG_FILE}-mix"
|
|
cp "${BIG_FILE}" "${BIG_FILE}-mix"
|
|
|
|
local OVER_FILE_BLOCK_POS=$((26*1024*1024/4))
|
|
echo -n "0123456789ABCDEF" | dd of="${BIG_FILE}-mix" bs=4 count=4 seek="${OVER_FILE_BLOCK_POS}" conv=notrunc
|
|
echo -n "0123456789ABCDEF" | dd of="${TEMP_DIR}/${BIG_FILE}-mix" bs=4 count=4 seek="${OVER_FILE_BLOCK_POS}" conv=notrunc
|
|
|
|
# Verify contents of file
|
|
echo "Comparing test file (2)"
|
|
if ! cmp "${TEMP_DIR}/${BIG_FILE}-mix" "${BIG_FILE}-mix"
|
|
then
|
|
return 1
|
|
fi
|
|
|
|
# (3) Writing from the 0th byte
|
|
#
|
|
cp "${TEMP_DIR}/${BIG_FILE}" "${TEMP_DIR}/${BIG_FILE}-mix"
|
|
cp "${BIG_FILE}" "${BIG_FILE}-mix"
|
|
|
|
echo -n "0123456789ABCDEF" | dd of="${BIG_FILE}-mix" bs=4 count=4 seek=0 conv=notrunc
|
|
echo -n "0123456789ABCDEF" | dd of="${TEMP_DIR}/${BIG_FILE}-mix" bs=4 count=4 seek=0 conv=notrunc
|
|
|
|
# Verify contents of file
|
|
echo "Comparing test file (3)"
|
|
if ! cmp "${TEMP_DIR}/${BIG_FILE}-mix" "${BIG_FILE}-mix"
|
|
then
|
|
return 1
|
|
fi
|
|
|
|
# (4) Write to the area within 5MB from the top
|
|
# modify directly(seek 1MB offset)
|
|
#
|
|
cp "${TEMP_DIR}/${BIG_FILE}" "${TEMP_DIR}/${BIG_FILE}-mix"
|
|
cp "${BIG_FILE}" "${BIG_FILE}-mix"
|
|
|
|
local MODIFY_START_BLOCK=$((1*1024*1024))
|
|
echo -n "0123456789ABCDEF" | dd of="${BIG_FILE}-mix" bs=4 count=4 seek="${MODIFY_START_BLOCK}" conv=notrunc
|
|
echo -n "0123456789ABCDEF" | dd of="${TEMP_DIR}/${BIG_FILE}-mix" bs=4 count=4 seek="${MODIFY_START_BLOCK}" conv=notrunc
|
|
|
|
# Verify contents of file
|
|
echo "Comparing test file (4)"
|
|
if ! cmp "${TEMP_DIR}/${BIG_FILE}-mix" "${BIG_FILE}-mix"
|
|
then
|
|
return 1
|
|
fi
|
|
|
|
# [NOTE]
|
|
# For macos, in order to free up disk space for statvfs (or df command),
|
|
# it is necessary to zero out the file size, delete it, and sync it.
|
|
# In the case of macos, even if you delete a file, there seems to be a
|
|
# delay in the free space being reflected.
|
|
# Testing the ensure_diskfree option shows that if this is not done, free
|
|
# disk space will be exhausted.
|
|
#
|
|
if [ "$(uname)" = "Darwin" ]; then
|
|
cat /dev/null > "${TEMP_DIR}/${BIG_FILE}"
|
|
cat /dev/null > "${TEMP_DIR}/${BIG_FILE}-mix"
|
|
cat /dev/null > "${BIG_FILE}"
|
|
cat /dev/null > "${BIG_FILE}-mix"
|
|
fi
|
|
|
|
rm -f "${TEMP_DIR}/${BIG_FILE}"
|
|
rm -f "${TEMP_DIR}/${BIG_FILE}-mix"
|
|
rm_test_file "${BIG_FILE}"
|
|
rm_test_file "${BIG_FILE}-mix"
|
|
|
|
if [ "$(uname)" = "Darwin" ]; then
|
|
sync
|
|
fi
|
|
}
|
|
|
|
function test_utimens_during_multipart {
|
|
describe "Testing utimens calling during multipart copy ..."
|
|
|
|
../../junk_data $((BIG_FILE_BLOCK_SIZE * BIG_FILE_COUNT)) > "${TEMP_DIR}/${BIG_FILE}"
|
|
|
|
cp "${TEMP_DIR}/${BIG_FILE}" "${BIG_FILE}"
|
|
|
|
# The second copy of the "-p" option calls utimens during multipart upload.
|
|
cp -p "${TEMP_DIR}/${BIG_FILE}" "${BIG_FILE}"
|
|
|
|
rm -f "${TEMP_DIR}/${BIG_FILE}"
|
|
rm_test_file "${BIG_FILE}"
|
|
}
|
|
|
|
function test_special_characters {
|
|
describe "Testing special characters ..."
|
|
|
|
(
|
|
set +o pipefail
|
|
# shellcheck disable=SC2010
|
|
ls 'special' 2>&1 | grep -q 'No such file or directory'
|
|
# shellcheck disable=SC2010
|
|
ls 'special?' 2>&1 | grep -q 'No such file or directory'
|
|
# shellcheck disable=SC2010
|
|
ls 'special*' 2>&1 | grep -q 'No such file or directory'
|
|
# shellcheck disable=SC2010
|
|
ls 'special~' 2>&1 | grep -q 'No such file or directory'
|
|
# shellcheck disable=SC2010
|
|
ls 'specialμ' 2>&1 | grep -q 'No such file or directory'
|
|
)
|
|
|
|
mkdir "TOYOTA TRUCK 8.2.2"
|
|
rm -rf "TOYOTA TRUCK 8.2.2"
|
|
}
|
|
|
|
function test_hardlink {
|
|
describe "Testing hardlinks ..."
|
|
|
|
rm -f "${TEST_TEXT_FILE}"
|
|
rm -f "${ALT_TEST_TEXT_FILE}"
|
|
echo foo > "${TEST_TEXT_FILE}"
|
|
|
|
(
|
|
# [NOTE]
|
|
# macos-fuse-t returns 'Input/output error'
|
|
#
|
|
set +o pipefail
|
|
ln "${TEST_TEXT_FILE}" "${ALT_TEST_TEXT_FILE}" 2>&1 | grep -q -e 'Operation not supported' -e 'Not supported' -e 'Input/output error'
|
|
)
|
|
|
|
rm_test_file
|
|
rm_test_file "${ALT_TEST_TEXT_FILE}"
|
|
}
|
|
|
|
function test_mknod {
|
|
describe "Testing mknod system call function ..."
|
|
|
|
local MKNOD_TEST_FILE_BASENAME="mknod_testfile"
|
|
|
|
rm -f "${MKNOD_TEST_FILE_BASENAME}*"
|
|
|
|
../../mknod_test "${MKNOD_TEST_FILE_BASENAME}"
|
|
}
|
|
|
|
function test_symlink {
|
|
describe "Testing symlinks ..."
|
|
|
|
rm -f "${TEST_TEXT_FILE}"
|
|
rm -f "${ALT_TEST_TEXT_FILE}"
|
|
echo foo > "${TEST_TEXT_FILE}"
|
|
|
|
ln -s "${TEST_TEXT_FILE}" "${ALT_TEST_TEXT_FILE}"
|
|
cmp "${TEST_TEXT_FILE}" "${ALT_TEST_TEXT_FILE}"
|
|
|
|
rm -f "${TEST_TEXT_FILE}"
|
|
|
|
[ -L "${ALT_TEST_TEXT_FILE}" ]
|
|
[ ! -f "${ALT_TEST_TEXT_FILE}" ]
|
|
|
|
rm -f "${ALT_TEST_TEXT_FILE}"
|
|
}
|
|
|
|
function test_extended_attributes {
|
|
describe "Testing extended attributes ..."
|
|
|
|
rm -f "${TEST_TEXT_FILE}"
|
|
touch "${TEST_TEXT_FILE}"
|
|
|
|
# set value
|
|
set_xattr key1 value0 "${TEST_TEXT_FILE}"
|
|
get_xattr key1 "${TEST_TEXT_FILE}" | grep -q '^value0$'
|
|
|
|
# over write value
|
|
set_xattr key1 value1 "${TEST_TEXT_FILE}"
|
|
get_xattr key1 "${TEST_TEXT_FILE}" | grep -q '^value1$'
|
|
|
|
# [NOTE]
|
|
# macOS still caches extended attributes even when told not to.
|
|
# Thus we need to wait one second here.
|
|
wait_ostype 1 "Darwin"
|
|
|
|
# append value
|
|
set_xattr key2 value2 "${TEST_TEXT_FILE}"
|
|
get_xattr key1 "${TEST_TEXT_FILE}" | grep -q '^value1$'
|
|
get_xattr key2 "${TEST_TEXT_FILE}" | grep -q '^value2$'
|
|
|
|
# remove value
|
|
del_xattr key1 "${TEST_TEXT_FILE}"
|
|
get_xattr key1 "${TEST_TEXT_FILE}" && return 1
|
|
get_xattr key2 "${TEST_TEXT_FILE}" | grep -q '^value2$'
|
|
|
|
rm_test_file
|
|
}
|
|
|
|
function test_mtime_file {
|
|
describe "Testing mtime preservation function ..."
|
|
|
|
# if the rename file exists, delete it
|
|
if [ -e "${ALT_TEST_TEXT_FILE}" ] || [ -L "${ALT_TEST_TEXT_FILE}" ]
|
|
then
|
|
rm "${ALT_TEST_TEXT_FILE}"
|
|
fi
|
|
|
|
if [ -e "${ALT_TEST_TEXT_FILE}" ]
|
|
then
|
|
echo "Could not delete file ${ALT_TEST_TEXT_FILE}, it still exists"
|
|
return 1
|
|
fi
|
|
|
|
# create the test file again
|
|
mk_test_file
|
|
|
|
#copy the test file with preserve mode
|
|
cp -p "${TEST_TEXT_FILE}" "${ALT_TEST_TEXT_FILE}"
|
|
|
|
local testmtime; testmtime=$(get_mtime "${TEST_TEXT_FILE}")
|
|
local testctime; testctime=$(get_ctime "${TEST_TEXT_FILE}")
|
|
local testatime; testatime=$(get_atime "${TEST_TEXT_FILE}")
|
|
local altmtime; altmtime=$(get_mtime "${ALT_TEST_TEXT_FILE}")
|
|
local altctime; altctime=$(get_ctime "${ALT_TEST_TEXT_FILE}")
|
|
local altatime; altatime=$(get_atime "${ALT_TEST_TEXT_FILE}")
|
|
|
|
if [ "${testmtime}" != "${altmtime}" ] || [ "${testctime}" = "${altctime}" ] || [ "${testatime}" != "${altatime}" ]; then
|
|
echo "cp(-p) expected times: mtime( ${testmtime} == ${altmtime} ), ctime( ${testctime} != ${altctime} ), atime( ${testatime} == ${altatime} )"
|
|
fi
|
|
|
|
rm_test_file
|
|
rm_test_file "${ALT_TEST_TEXT_FILE}"
|
|
}
|
|
|
|
# [NOTE]
|
|
# If it mounted with relatime or noatime options , the "touch -a"
|
|
# command may not update the atime.
|
|
# In ubuntu:xenial, atime was updated even if relatime was granted.
|
|
# However, it was not updated in bionic/focal.
|
|
# We can probably update atime by explicitly specifying the strictatime
|
|
# option and running the "touch -a" command. However, the strictatime
|
|
# option cannot be set.
|
|
# Therefore, if the relatime option is set, the test with the "touch -a"
|
|
# command is bypassed.
|
|
# We do not know why atime is not updated may or not be affected by
|
|
# these options.(can't say for sure)
|
|
# However, if atime has not been updated, the s3fs_utimens entry point
|
|
# will not be called from FUSE library. We added this bypass because
|
|
# the test became unstable.
|
|
#
|
|
function test_update_time_chmod() {
|
|
describe "Testing update time function chmod..."
|
|
|
|
local t0=1000000000 # 9 September 2001
|
|
local OBJECT_NAME; OBJECT_NAME=$(basename "${PWD}")/"${TEST_TEXT_FILE}"
|
|
echo data | aws_cli s3 cp --metadata="atime=${t0},ctime=${t0},mtime=${t0}" - "s3://${TEST_BUCKET_1}/${OBJECT_NAME}"
|
|
local base_atime; base_atime=$(get_atime "${TEST_TEXT_FILE}")
|
|
local base_ctime; base_ctime=$(get_ctime "${TEST_TEXT_FILE}")
|
|
local base_mtime; base_mtime=$(get_mtime "${TEST_TEXT_FILE}")
|
|
|
|
#
|
|
# chmod -> update only ctime
|
|
#
|
|
chmod +x "${TEST_TEXT_FILE}"
|
|
local atime; atime=$(get_atime "${TEST_TEXT_FILE}")
|
|
local ctime; ctime=$(get_ctime "${TEST_TEXT_FILE}")
|
|
local mtime; mtime=$(get_mtime "${TEST_TEXT_FILE}")
|
|
if [ "${base_atime}" != "${atime}" ] || [ "${base_ctime}" = "${ctime}" ] || [ "${base_mtime}" != "${mtime}" ]; then
|
|
echo "chmod expected updated ctime: $base_ctime != $ctime and same mtime: $base_mtime == $mtime, atime: $base_atime == $atime"
|
|
return 1
|
|
fi
|
|
rm_test_file
|
|
}
|
|
|
|
function test_update_time_chown() {
|
|
describe "Testing update time function chown..."
|
|
|
|
#
|
|
# chown -> update only ctime
|
|
#
|
|
local t0=1000000000 # 9 September 2001
|
|
local OBJECT_NAME; OBJECT_NAME=$(basename "${PWD}")/"${TEST_TEXT_FILE}"
|
|
echo data | aws_cli s3 cp --metadata="atime=${t0},ctime=${t0},mtime=${t0}" - "s3://${TEST_BUCKET_1}/${OBJECT_NAME}"
|
|
local base_atime; base_atime=$(get_atime "${TEST_TEXT_FILE}")
|
|
local base_ctime; base_ctime=$(get_ctime "${TEST_TEXT_FILE}")
|
|
local base_mtime; base_mtime=$(get_mtime "${TEST_TEXT_FILE}")
|
|
|
|
# [NOTE]
|
|
# In this test, chown is called with the same UID.
|
|
#
|
|
chown "${UID}" "${TEST_TEXT_FILE}"
|
|
|
|
local atime; atime=$(get_atime "${TEST_TEXT_FILE}")
|
|
local ctime; ctime=$(get_ctime "${TEST_TEXT_FILE}")
|
|
local mtime; mtime=$(get_mtime "${TEST_TEXT_FILE}")
|
|
|
|
if ! uname | grep -q Darwin; then
|
|
if [ "${base_atime}" != "${atime}" ] || [ "${base_ctime}" = "${ctime}" ] || [ "${base_mtime}" != "${mtime}" ]; then
|
|
echo "chown expected updated ctime: $base_ctime != $ctime and same mtime: $base_mtime == $mtime, atime: $base_atime == $atime"
|
|
return 1
|
|
fi
|
|
else
|
|
# [FIXME] macos fuse-t
|
|
# macos fuse-t doesn't update stat if UID doesn't change.
|
|
# There is a way to specify "uid=1000" with aws cli and use sudo when chown is executed, but the
|
|
# test is not finished.
|
|
# For now, we are just leaving the chown call with the same UID as the parameter.
|
|
# This test will be fixed in the future.
|
|
if [ "${base_atime}" = "${atime}" ] || [ "${base_ctime}" = "${ctime}" ] || [ "${base_mtime}" = "${mtime}" ]; then
|
|
if [ "${base_atime}" = "${atime}" ] && [ "${base_ctime}" = "${ctime}" ] && [ "${base_mtime}" = "${mtime}" ]; then
|
|
echo "[FIXME] Doing a temporary test bypass : same ctime $base_ctime = $ctime and same mtime: $base_mtime = $mtime and same atime: $base_atime = $atime"
|
|
else
|
|
echo "chown expected updated ctime: $base_ctime != $ctime and same mtime: $base_mtime != $mtime, atime: $base_atime != $atime"
|
|
return 1
|
|
fi
|
|
fi
|
|
fi
|
|
rm_test_file
|
|
}
|
|
|
|
function test_update_time_xattr() {
|
|
describe "Testing update time function set_xattr..."
|
|
|
|
local t0=1000000000 # 9 September 2001
|
|
local OBJECT_NAME; OBJECT_NAME=$(basename "${PWD}")/"${TEST_TEXT_FILE}"
|
|
echo data | aws_cli s3 cp --metadata="atime=${t0},ctime=${t0},mtime=${t0}" - "s3://${TEST_BUCKET_1}/${OBJECT_NAME}"
|
|
local base_atime; base_atime=$(get_atime "${TEST_TEXT_FILE}")
|
|
local base_ctime; base_ctime=$(get_ctime "${TEST_TEXT_FILE}")
|
|
local base_mtime; base_mtime=$(get_mtime "${TEST_TEXT_FILE}")
|
|
|
|
#
|
|
# set_xattr -> update only ctime
|
|
#
|
|
set_xattr key value "${TEST_TEXT_FILE}"
|
|
local atime; atime=$(get_atime "${TEST_TEXT_FILE}")
|
|
local ctime; ctime=$(get_ctime "${TEST_TEXT_FILE}")
|
|
local mtime; mtime=$(get_mtime "${TEST_TEXT_FILE}")
|
|
if [ "${base_atime}" != "${atime}" ] || [ "${base_ctime}" = "${ctime}" ] || [ "${base_mtime}" != "${mtime}" ]; then
|
|
echo "set_xattr expected updated ctime: $base_ctime != $ctime and same mtime: $base_mtime == $mtime, atime: $base_atime == $atime"
|
|
return 1
|
|
fi
|
|
rm_test_file
|
|
}
|
|
|
|
function test_update_time_touch() {
|
|
describe "Testing update time function touch..."
|
|
|
|
local t0=1000000000 # 9 September 2001
|
|
local OBJECT_NAME; OBJECT_NAME=$(basename "${PWD}")/"${TEST_TEXT_FILE}"
|
|
echo data | aws_cli s3 cp --metadata="atime=${t0},ctime=${t0},mtime=${t0}" - "s3://${TEST_BUCKET_1}/${OBJECT_NAME}"
|
|
local base_atime; base_atime=$(get_atime "${TEST_TEXT_FILE}")
|
|
local base_ctime; base_ctime=$(get_ctime "${TEST_TEXT_FILE}")
|
|
local base_mtime; base_mtime=$(get_mtime "${TEST_TEXT_FILE}")
|
|
|
|
#
|
|
# touch -> update ctime/atime/mtime
|
|
#
|
|
touch "${TEST_TEXT_FILE}"
|
|
local atime; atime=$(get_atime "${TEST_TEXT_FILE}")
|
|
local ctime; ctime=$(get_ctime "${TEST_TEXT_FILE}")
|
|
local mtime; mtime=$(get_mtime "${TEST_TEXT_FILE}")
|
|
if [ "${base_atime}" = "${atime}" ] || [ "${base_ctime}" = "${ctime}" ] || [ "${base_mtime}" = "${mtime}" ]; then
|
|
echo "touch expected updated ctime: $base_ctime != $ctime, mtime: $base_mtime != $mtime, atime: $base_atime != $atime"
|
|
return 1
|
|
fi
|
|
rm_test_file
|
|
}
|
|
|
|
function test_update_time_touch_a() {
|
|
describe "Testing update time function touch -a..."
|
|
|
|
local t0=1000000000 # 9 September 2001
|
|
local OBJECT_NAME; OBJECT_NAME=$(basename "${PWD}")/"${TEST_TEXT_FILE}"
|
|
echo data | aws_cli s3 cp --metadata="atime=${t0},ctime=${t0},mtime=${t0}" - "s3://${TEST_BUCKET_1}/${OBJECT_NAME}"
|
|
local base_atime; base_atime=$(get_atime "${TEST_TEXT_FILE}")
|
|
local base_ctime; base_ctime=$(get_ctime "${TEST_TEXT_FILE}")
|
|
local base_mtime; base_mtime=$(get_mtime "${TEST_TEXT_FILE}")
|
|
|
|
#
|
|
# "touch -a" -> update ctime/atime, not update mtime
|
|
#
|
|
touch -a "${TEST_TEXT_FILE}"
|
|
|
|
local atime; atime=$(get_atime "${TEST_TEXT_FILE}")
|
|
local ctime; ctime=$(get_ctime "${TEST_TEXT_FILE}")
|
|
local mtime; mtime=$(get_mtime "${TEST_TEXT_FILE}")
|
|
|
|
if ! uname | grep -q Darwin; then
|
|
if [ "${base_atime}" = "${atime}" ] || [ "${base_ctime}" = "${ctime}" ] || [ "${base_mtime}" != "${mtime}" ]; then
|
|
echo "touch with -a option expected updated ctime: $base_ctime != $ctime, atime: $base_atime != $atime and same mtime: $base_mtime == $mtime"
|
|
return 1
|
|
fi
|
|
else
|
|
# [macos] fuse-t
|
|
# atime/ctime/mtime are all updated.
|
|
#
|
|
if [ "${base_atime}" = "${atime}" ] || [ "${base_ctime}" = "${ctime}" ] || [ "${base_mtime}" = "${mtime}" ]; then
|
|
echo "touch with -a option expected updated ctime: $base_ctime != $ctime, atime: $base_atime != $atime and same mtime: $base_mtime != $mtime"
|
|
return 1
|
|
fi
|
|
fi
|
|
rm_test_file
|
|
}
|
|
|
|
function test_update_time_append() {
|
|
describe "Testing update time function append..."
|
|
|
|
local t0=1000000000 # 9 September 2001
|
|
local OBJECT_NAME; OBJECT_NAME=$(basename "${PWD}")/"${TEST_TEXT_FILE}"
|
|
echo data | aws_cli s3 cp --metadata="atime=${t0},ctime=${t0},mtime=${t0}" - "s3://${TEST_BUCKET_1}/${OBJECT_NAME}"
|
|
local base_atime; base_atime=$(get_atime "${TEST_TEXT_FILE}")
|
|
local base_ctime; base_ctime=$(get_ctime "${TEST_TEXT_FILE}")
|
|
local base_mtime; base_mtime=$(get_mtime "${TEST_TEXT_FILE}")
|
|
|
|
#
|
|
# append -> update ctime/mtime, not update atime
|
|
#
|
|
echo foo >> "${TEST_TEXT_FILE}"
|
|
local atime; atime=$(get_atime "${TEST_TEXT_FILE}")
|
|
local ctime; ctime=$(get_ctime "${TEST_TEXT_FILE}")
|
|
local mtime; mtime=$(get_mtime "${TEST_TEXT_FILE}")
|
|
if [ "${base_atime}" != "${atime}" ] || [ "${base_ctime}" = "${ctime}" ] || [ "${base_mtime}" = "${mtime}" ]; then
|
|
echo "append expected updated ctime: $base_ctime != $ctime, mtime: $base_mtime != $mtime and same atime: $base_atime == $atime"
|
|
return 1
|
|
fi
|
|
rm_test_file
|
|
}
|
|
|
|
function test_update_time_cp_p() {
|
|
describe "Testing update time function cp -p..."
|
|
|
|
local t0=1000000000 # 9 September 2001
|
|
local OBJECT_NAME; OBJECT_NAME=$(basename "${PWD}")/"${TEST_TEXT_FILE}"
|
|
echo data | aws_cli s3 cp --metadata="atime=${t0},ctime=${t0},mtime=${t0}" - "s3://${TEST_BUCKET_1}/${OBJECT_NAME}"
|
|
local base_atime; base_atime=$(get_atime "${TEST_TEXT_FILE}")
|
|
local base_ctime; base_ctime=$(get_ctime "${TEST_TEXT_FILE}")
|
|
local base_mtime; base_mtime=$(get_mtime "${TEST_TEXT_FILE}")
|
|
|
|
#
|
|
# cp -p -> update ctime, not update atime/mtime
|
|
#
|
|
local TIME_TEST_TEXT_FILE=test-s3fs-time.txt
|
|
cp -p "${TEST_TEXT_FILE}" "${TIME_TEST_TEXT_FILE}"
|
|
local atime; atime=$(get_atime "${TIME_TEST_TEXT_FILE}")
|
|
local ctime; ctime=$(get_ctime "${TIME_TEST_TEXT_FILE}")
|
|
local mtime; mtime=$(get_mtime "${TIME_TEST_TEXT_FILE}")
|
|
if [ "${base_atime}" != "${atime}" ] || [ "${base_ctime}" = "${ctime}" ] || [ "${base_mtime}" != "${mtime}" ]; then
|
|
echo "cp with -p option expected updated ctime: $base_ctime != $ctime and same mtime: $base_mtime == $mtime, atime: $base_atime == $atime"
|
|
return 1
|
|
fi
|
|
rm_test_file
|
|
rm_test_file "${TIME_TEST_TEXT_FILE}"
|
|
}
|
|
|
|
function test_update_time_mv() {
|
|
describe "Testing update time function mv..."
|
|
|
|
local t0=1000000000 # 9 September 2001
|
|
local OBJECT_NAME; OBJECT_NAME=$(basename "${PWD}")/"${TEST_TEXT_FILE}"
|
|
echo data | aws_cli s3 cp --metadata="atime=${t0},ctime=${t0},mtime=${t0}" - "s3://${TEST_BUCKET_1}/${OBJECT_NAME}"
|
|
local base_atime; base_atime=$(get_atime "${TEST_TEXT_FILE}")
|
|
local base_ctime; base_ctime=$(get_ctime "${TEST_TEXT_FILE}")
|
|
local base_mtime; base_mtime=$(get_mtime "${TEST_TEXT_FILE}")
|
|
|
|
#
|
|
# mv -> update ctime, not update atime/mtime
|
|
#
|
|
local TIME2_TEST_TEXT_FILE=test-s3fs-time2.txt
|
|
mv "${TEST_TEXT_FILE}" "${TIME2_TEST_TEXT_FILE}"
|
|
local atime; atime=$(get_atime "${TIME2_TEST_TEXT_FILE}")
|
|
local ctime; ctime=$(get_ctime "${TIME2_TEST_TEXT_FILE}")
|
|
local mtime; mtime=$(get_mtime "${TIME2_TEST_TEXT_FILE}")
|
|
if [ "${base_atime}" != "${atime}" ] || [ "${base_ctime}" = "${ctime}" ] || [ "${base_mtime}" != "${mtime}" ]; then
|
|
echo "mv expected updated ctime: $base_ctime != $ctime and same mtime: $base_mtime == $mtime, atime: $base_atime == $atime"
|
|
return 1
|
|
fi
|
|
|
|
rm_test_file "${TIME_TEST_TEXT_FILE}"
|
|
rm_test_file "${TIME2_TEST_TEXT_FILE}"
|
|
}
|
|
|
|
# [NOTE]
|
|
# See the description of test_update_time () for notes about the
|
|
# "touch -a" command and atime.
|
|
#
|
|
function test_update_directory_time_chmod() {
|
|
describe "Testing update time for directory mv..."
|
|
|
|
#
|
|
# create the directory and sub-directory and a file in directory
|
|
#
|
|
local t0=1000000000 # 9 September 2001
|
|
local DIRECTORY_NAME; DIRECTORY_NAME=$(basename "${PWD}")/"${TEST_DIR}"
|
|
aws_cli s3api put-object --content-type="application/x-directory" --metadata="atime=${t0},ctime=${t0},mtime=${t0}" --bucket "${TEST_BUCKET_1}" --key "$DIRECTORY_NAME/"
|
|
|
|
local base_atime; base_atime=$(get_atime "${TEST_DIR}")
|
|
local base_ctime; base_ctime=$(get_ctime "${TEST_DIR}")
|
|
local base_mtime; base_mtime=$(get_mtime "${TEST_DIR}")
|
|
|
|
#
|
|
# chmod -> update only ctime
|
|
#
|
|
chmod 0777 "${TEST_DIR}"
|
|
local atime; atime=$(get_atime "${TEST_DIR}")
|
|
local ctime; ctime=$(get_ctime "${TEST_DIR}")
|
|
local mtime; mtime=$(get_mtime "${TEST_DIR}")
|
|
if [ "${base_atime}" != "${atime}" ] || [ "${base_ctime}" = "${ctime}" ] || [ "${base_mtime}" != "${mtime}" ]; then
|
|
echo "chmod expected updated ctime: $base_ctime != $ctime and same mtime: $base_mtime == $mtime, atime: $base_atime == $atime"
|
|
return 1
|
|
fi
|
|
|
|
rm -rf "${TEST_DIR}"
|
|
}
|
|
|
|
function test_update_directory_time_chown {
|
|
describe "Testing update time for directory chown..."
|
|
|
|
local t0=1000000000 # 9 September 2001
|
|
local DIRECTORY_NAME; DIRECTORY_NAME=$(basename "${PWD}")/"${TEST_DIR}"
|
|
aws_cli s3api put-object --content-type="application/x-directory" --metadata="atime=${t0},ctime=${t0},mtime=${t0}" --bucket "${TEST_BUCKET_1}" --key "$DIRECTORY_NAME/"
|
|
|
|
local base_atime; base_atime=$(get_atime "${TEST_DIR}")
|
|
local base_ctime; base_ctime=$(get_ctime "${TEST_DIR}")
|
|
local base_mtime; base_mtime=$(get_mtime "${TEST_DIR}")
|
|
#
|
|
# chown -> update only ctime
|
|
#
|
|
chown $UID "${TEST_DIR}"
|
|
local atime; atime=$(get_atime "${TEST_DIR}")
|
|
local ctime; ctime=$(get_ctime "${TEST_DIR}")
|
|
local mtime; mtime=$(get_mtime "${TEST_DIR}")
|
|
|
|
if ! uname | grep -q Darwin; then
|
|
if [ "${base_atime}" != "${atime}" ] || [ "${base_ctime}" = "${ctime}" ] || [ "${base_mtime}" != "${mtime}" ]; then
|
|
echo "chown expected updated ctime: $base_ctime != $ctime and same mtime: $base_mtime == $mtime, atime: $base_atime == $atime"
|
|
return 1
|
|
fi
|
|
else
|
|
# [macos] fuse-t
|
|
# atime/ctime/mtime are not updated.
|
|
#
|
|
if [ "${base_atime}" != "${atime}" ] || [ "${base_ctime}" != "${ctime}" ] || [ "${base_mtime}" != "${mtime}" ]; then
|
|
echo "touch with -a option expected updated ctime: $base_ctime == $ctime, atime: $base_atime == $atime and same mtime: $base_mtime == $mtime"
|
|
return 1
|
|
fi
|
|
fi
|
|
|
|
rm -rf "${TEST_DIR}"
|
|
}
|
|
|
|
function test_update_directory_time_set_xattr {
|
|
describe "Testing update time for directory set_xattr..."
|
|
|
|
local t0=1000000000 # 9 September 2001
|
|
local DIRECTORY_NAME; DIRECTORY_NAME=$(basename "${PWD}")/"${TEST_DIR}"
|
|
aws_cli s3api put-object --content-type="application/x-directory" --metadata="atime=${t0},ctime=${t0},mtime=${t0}" --bucket "${TEST_BUCKET_1}" --key "$DIRECTORY_NAME/"
|
|
|
|
local base_atime; base_atime=$(get_atime "${TEST_DIR}")
|
|
local base_ctime; base_ctime=$(get_ctime "${TEST_DIR}")
|
|
local base_mtime; base_mtime=$(get_mtime "${TEST_DIR}")
|
|
#
|
|
# set_xattr -> update only ctime
|
|
#
|
|
set_xattr key value "${TEST_DIR}"
|
|
local atime; atime=$(get_atime "${TEST_DIR}")
|
|
local ctime; ctime=$(get_ctime "${TEST_DIR}")
|
|
local mtime; mtime=$(get_mtime "${TEST_DIR}")
|
|
|
|
if ! uname | grep -q Darwin; then
|
|
if [ "${base_atime}" != "${atime}" ] || [ "${base_ctime}" = "${ctime}" ] || [ "${base_mtime}" != "${mtime}" ]; then
|
|
echo "set_xattr expected updated ctime: $base_ctime != $ctime and same mtime: $base_mtime == $mtime, atime: $base_atime == $atime"
|
|
return 1
|
|
fi
|
|
else
|
|
# [macos] fuse-t
|
|
# atime/mtime are not updated.
|
|
#
|
|
if [ "${base_atime}" != "${atime}" ] || [ "${base_ctime}" = "${ctime}" ] || [ "${base_mtime}" != "${mtime}" ]; then
|
|
echo "set_xattr expected updated ctime: $base_ctime != $ctime and same mtime: $base_mtime == $mtime, atime: $base_atime == $atime"
|
|
return 1
|
|
fi
|
|
fi
|
|
|
|
rm -rf "${TEST_DIR}"
|
|
}
|
|
|
|
function test_update_directory_time_touch {
|
|
describe "Testing update time for directory touch..."
|
|
|
|
local t0=1000000000 # 9 September 2001
|
|
local DIRECTORY_NAME; DIRECTORY_NAME=$(basename "${PWD}")/"${TEST_DIR}"
|
|
aws_cli s3api put-object --content-type="application/x-directory" --metadata="atime=${t0},ctime=${t0},mtime=${t0}" --bucket "${TEST_BUCKET_1}" --key "$DIRECTORY_NAME/"
|
|
|
|
local base_atime; base_atime=$(get_atime "${TEST_DIR}")
|
|
local base_ctime; base_ctime=$(get_ctime "${TEST_DIR}")
|
|
local base_mtime; base_mtime=$(get_mtime "${TEST_DIR}")
|
|
#
|
|
# touch -> update ctime/atime/mtime
|
|
#
|
|
touch "${TEST_DIR}"
|
|
local atime; atime=$(get_atime "${TEST_DIR}")
|
|
local ctime; ctime=$(get_ctime "${TEST_DIR}")
|
|
local mtime; mtime=$(get_mtime "${TEST_DIR}")
|
|
if [ "${base_atime}" = "${atime}" ] || [ "${base_ctime}" = "${ctime}" ] || [ "${base_mtime}" = "${mtime}" ]; then
|
|
echo "touch expected updated ctime: $base_ctime != $ctime, mtime: $base_mtime != $mtime, atime: $base_atime != $atime"
|
|
return 1
|
|
fi
|
|
|
|
rm -rf "${TEST_DIR}"
|
|
}
|
|
|
|
function test_update_directory_time_touch_a {
|
|
describe "Testing update time for directory touch -a..."
|
|
|
|
local t0=1000000000 # 9 September 2001
|
|
local DIRECTORY_NAME; DIRECTORY_NAME=$(basename "${PWD}")/"${TEST_DIR}"
|
|
aws_cli s3api put-object --content-type="application/x-directory" --metadata="atime=${t0},ctime=${t0},mtime=${t0}" --bucket "${TEST_BUCKET_1}" --key "$DIRECTORY_NAME/"
|
|
|
|
local base_atime; base_atime=$(get_atime "${TEST_DIR}")
|
|
local base_ctime; base_ctime=$(get_ctime "${TEST_DIR}")
|
|
local base_mtime; base_mtime=$(get_mtime "${TEST_DIR}")
|
|
#
|
|
# "touch -a" -> update ctime/atime, not update mtime
|
|
#
|
|
touch -a "${TEST_DIR}"
|
|
local atime; atime=$(get_atime "${TEST_DIR}")
|
|
local ctime; ctime=$(get_ctime "${TEST_DIR}")
|
|
local mtime; mtime=$(get_mtime "${TEST_DIR}")
|
|
|
|
if ! uname | grep -q Darwin; then
|
|
if [ "${base_atime}" = "${atime}" ] || [ "${base_ctime}" = "${ctime}" ] || [ "${base_mtime}" != "${mtime}" ]; then
|
|
echo "touch with -a option expected updated ctime: $base_ctime != $ctime, atime: $base_atime != $atime and same mtime: $base_mtime == $mtime"
|
|
return 1
|
|
fi
|
|
else
|
|
# [macos] fuse-t
|
|
# atime/ctime/mtime are all updated.
|
|
#
|
|
if [ "${base_atime}" = "${atime}" ] || [ "${base_ctime}" = "${ctime}" ] || [ "${base_mtime}" = "${mtime}" ]; then
|
|
echo "touch with -a option expected updated ctime: $base_ctime != $ctime, atime: $base_atime != $atime and same mtime: $base_mtime != $mtime"
|
|
return 1
|
|
fi
|
|
fi
|
|
|
|
rm -rf "${TEST_DIR}"
|
|
}
|
|
|
|
function test_update_directory_time_subdir() {
|
|
describe "Testing update time for directory subdirectory..."
|
|
|
|
local TIME_TEST_SUBDIR="${TEST_DIR}/testsubdir"
|
|
local TIME_TEST_FILE_INDIR="${TEST_DIR}/testfile"
|
|
mk_test_dir
|
|
mkdir "${TIME_TEST_SUBDIR}"
|
|
touch "${TIME_TEST_FILE_INDIR}"
|
|
# TODO: remove sleep after improving AWS CLI speed
|
|
sleep 1
|
|
|
|
local base_atime; base_atime=$(get_atime "${TEST_DIR}")
|
|
local base_ctime; base_ctime=$(get_ctime "${TEST_DIR}")
|
|
local base_mtime; base_mtime=$(get_mtime "${TEST_DIR}")
|
|
local subdir_atime; subdir_atime=$(get_atime "${TIME_TEST_SUBDIR}")
|
|
local subdir_ctime; subdir_ctime=$(get_ctime "${TIME_TEST_SUBDIR}")
|
|
local subdir_mtime; subdir_mtime=$(get_mtime "${TIME_TEST_SUBDIR}")
|
|
local subfile_atime; subfile_atime=$(get_atime "${TIME_TEST_FILE_INDIR}")
|
|
local subfile_ctime; subfile_ctime=$(get_ctime "${TIME_TEST_FILE_INDIR}")
|
|
local subfile_mtime; subfile_mtime=$(get_mtime "${TIME_TEST_FILE_INDIR}")
|
|
#
|
|
# mv -> update ctime, not update atime/mtime for target directory
|
|
# not update any for sub-directory and a file
|
|
#
|
|
local TIME_TEST_DIR=timetestdir
|
|
local TIME2_TEST_SUBDIR="${TIME_TEST_DIR}/testsubdir"
|
|
local TIME2_TEST_FILE_INDIR="${TIME_TEST_DIR}/testfile"
|
|
mv "${TEST_DIR}" "${TIME_TEST_DIR}"
|
|
local atime; atime=$(get_atime "${TIME_TEST_DIR}")
|
|
local ctime; ctime=$(get_ctime "${TIME_TEST_DIR}")
|
|
local mtime; mtime=$(get_mtime "${TIME_TEST_DIR}")
|
|
if [ "${base_atime}" != "${atime}" ] || [ "${base_ctime}" = "${ctime}" ] || [ "${base_mtime}" != "${mtime}" ]; then
|
|
echo "mv expected updated ctime: $base_ctime != $ctime and same mtime: $base_mtime == $mtime, atime: $base_atime == $atime"
|
|
return 1
|
|
fi
|
|
atime=$(get_atime "${TIME2_TEST_SUBDIR}")
|
|
ctime=$(get_ctime "${TIME2_TEST_SUBDIR}")
|
|
mtime=$(get_mtime "${TIME2_TEST_SUBDIR}")
|
|
if [ "${subdir_atime}" != "${atime}" ] || [ "${subdir_ctime}" != "${ctime}" ] || [ "${subdir_mtime}" != "${mtime}" ]; then
|
|
echo "mv for sub-directory expected same ctime: $subdir_ctime == $ctime, mtime: $subdir_mtime == $mtime, atime: $subdir_atime == $atime"
|
|
return 1
|
|
fi
|
|
atime=$(get_atime "${TIME2_TEST_FILE_INDIR}")
|
|
ctime=$(get_ctime "${TIME2_TEST_FILE_INDIR}")
|
|
mtime=$(get_mtime "${TIME2_TEST_FILE_INDIR}")
|
|
if [ "${subfile_atime}" != "${atime}" ] || [ "${subfile_ctime}" != "${ctime}" ] || [ "${subfile_mtime}" != "${mtime}" ]; then
|
|
echo "mv for a file in directory expected same ctime: $subfile_ctime == $ctime, mtime: $subfile_mtime == $mtime, atime: $subfile_atime == $atime"
|
|
return 1
|
|
fi
|
|
|
|
rm -rf "${TIME_TEST_SUBDIR}"
|
|
rm -rf "${TIME_TEST_DIR}"
|
|
rm -rf "${TEST_DIR}"
|
|
}
|
|
|
|
# [NOTE]
|
|
# This test changes the file mode while creating/editing a new file,
|
|
# and finally closes it.
|
|
# Test with the sed command as it occurs when in place mode of the sed
|
|
# command. (If trying it with a standard C function(and shell script),
|
|
# it will be not the same result of sed, so sed is used.)
|
|
#
|
|
function test_update_chmod_opened_file() {
|
|
describe "Testing create, modify the file by sed in place mode"
|
|
|
|
# test file
|
|
local BEFORE_STRING_DATA; BEFORE_STRING_DATA="sed in place test : BEFORE DATA"
|
|
local AFTER_STRING_DATA; AFTER_STRING_DATA="sed in place test : AFTER DATA"
|
|
echo "${BEFORE_STRING_DATA}" > "${TEST_TEXT_FILE}"
|
|
|
|
# sed in place
|
|
sed -i -e 's/BEFORE DATA/AFTER DATA/g' "${TEST_TEXT_FILE}"
|
|
|
|
# compare result
|
|
local RESULT_STRING; RESULT_STRING=$(cat "${TEST_TEXT_FILE}")
|
|
|
|
if [ -z "${RESULT_STRING}" ] || [ "${RESULT_STRING}" != "${AFTER_STRING_DATA}" ]; then
|
|
echo "the file conversion by sed in place command failed."
|
|
return 1
|
|
fi
|
|
|
|
# clean up
|
|
rm_test_file "${ALT_TEST_TEXT_FILE}"
|
|
}
|
|
|
|
function test_update_parent_directory_time_sub() {
|
|
if [ $# -ne 1 ]; then
|
|
echo "Internal error: parameter is wrong."
|
|
return 1
|
|
fi
|
|
|
|
# [NOTE]
|
|
# Skip test for mknod/mkfifo command.
|
|
# If run them, ctime/mtime of the parent directory will be updated.
|
|
#
|
|
local TEST_PARENTDIR_PARENT="${1}"
|
|
local TEST_PARENTDIR_FILE="${TEST_PARENTDIR_PARENT}/testfile"
|
|
local TEST_PARENTDIR_SYMFILE_BASE="testfile2"
|
|
local TEST_PARENTDIR_FILE_MV="${TEST_PARENTDIR_PARENT}/${TEST_PARENTDIR_SYMFILE_BASE}"
|
|
local TEST_PARENTDIR_SYMFILE="${TEST_PARENTDIR_PARENT}/symfile"
|
|
local TEST_PARENTDIR_SYMFILE_MV="${TEST_PARENTDIR_PARENT}/symfile2"
|
|
local TEST_PARENTDIR_DIR="${TEST_PARENTDIR_PARENT}/testdir"
|
|
local TEST_PARENTDIR_DIR_MV="${TEST_PARENTDIR_PARENT}/testdir2"
|
|
|
|
#
|
|
# Create file -> Darwin: Not update any
|
|
# -> Others: Update parent directory's mtime/ctime
|
|
#
|
|
local base_atime; base_atime=$(get_atime "${TEST_PARENTDIR_PARENT}")
|
|
local base_ctime; base_ctime=$(get_ctime "${TEST_PARENTDIR_PARENT}")
|
|
local base_mtime; base_mtime=$(get_mtime "${TEST_PARENTDIR_PARENT}")
|
|
|
|
touch "${TEST_PARENTDIR_FILE}"
|
|
|
|
local after_atime; after_atime=$(get_atime "${TEST_PARENTDIR_PARENT}")
|
|
local after_ctime; after_ctime=$(get_ctime "${TEST_PARENTDIR_PARENT}")
|
|
local after_mtime; after_mtime=$(get_mtime "${TEST_PARENTDIR_PARENT}")
|
|
|
|
if [ "${base_atime}" != "${after_atime}" ] || [ "${base_ctime}" = "${after_ctime}" ] || [ "${base_mtime}" = "${after_mtime}" ]; then
|
|
echo "creating file expected updating ctime/mtime: atime( ${base_atime} == ${after_atime} ), ctime( ${base_ctime} != ${after_ctime} ), mtime( ${base_mtime} != ${after_mtime} )"
|
|
return 1
|
|
fi
|
|
|
|
#
|
|
# Update file -> Not update parent directory's atime/mtime/ctime
|
|
#
|
|
base_atime="${after_atime}"
|
|
base_ctime="${after_ctime}"
|
|
base_mtime="${after_mtime}"
|
|
|
|
touch "${TEST_PARENTDIR_FILE}"
|
|
|
|
local after_atime; after_atime=$(get_atime "${TEST_PARENTDIR_PARENT}")
|
|
local after_ctime; after_ctime=$(get_ctime "${TEST_PARENTDIR_PARENT}")
|
|
local after_mtime; after_mtime=$(get_mtime "${TEST_PARENTDIR_PARENT}")
|
|
|
|
if [ "${base_atime}" != "${after_atime}" ] || [ "${base_ctime}" != "${after_ctime}" ] || [ "${base_mtime}" != "${after_mtime}" ]; then
|
|
echo "updating file expected updating ctime/mtime: atime( ${base_atime} == ${after_atime} ), ctime( ${base_ctime} == ${after_ctime} ), mtime( ${base_mtime} == ${after_mtime} )"
|
|
return 1
|
|
fi
|
|
|
|
#
|
|
# Rename file -> Update parent directory's mtime/ctime
|
|
#
|
|
base_atime="${after_atime}"
|
|
base_ctime="${after_ctime}"
|
|
base_mtime="${after_mtime}"
|
|
|
|
mv "${TEST_PARENTDIR_FILE}" "${TEST_PARENTDIR_FILE_MV}"
|
|
|
|
local after_atime; after_atime=$(get_atime "${TEST_PARENTDIR_PARENT}")
|
|
local after_ctime; after_ctime=$(get_ctime "${TEST_PARENTDIR_PARENT}")
|
|
local after_mtime; after_mtime=$(get_mtime "${TEST_PARENTDIR_PARENT}")
|
|
|
|
if [ "${base_atime}" != "${after_atime}" ] || [ "${base_ctime}" = "${after_ctime}" ] || [ "${base_mtime}" = "${after_mtime}" ]; then
|
|
echo "renaming file expected updating ctime/mtime: atime( ${base_atime} == ${after_atime} ), ctime( ${base_ctime} != ${after_ctime} ), mtime( ${base_mtime} != ${after_mtime} )"
|
|
return 1
|
|
fi
|
|
|
|
#
|
|
# Create symbolic link -> Update parent directory's mtime/ctime
|
|
#
|
|
base_atime="${after_atime}"
|
|
base_ctime="${after_ctime}"
|
|
base_mtime="${after_mtime}"
|
|
|
|
ln -s "${TEST_PARENTDIR_SYMFILE_BASE}" "${TEST_PARENTDIR_SYMFILE}"
|
|
|
|
local after_atime; after_atime=$(get_atime "${TEST_PARENTDIR_PARENT}")
|
|
local after_ctime; after_ctime=$(get_ctime "${TEST_PARENTDIR_PARENT}")
|
|
local after_mtime; after_mtime=$(get_mtime "${TEST_PARENTDIR_PARENT}")
|
|
|
|
if [ "${base_atime}" != "${after_atime}" ] || [ "${base_ctime}" = "${after_ctime}" ] || [ "${base_mtime}" = "${after_mtime}" ]; then
|
|
echo "creating symbolic file expected updating ctime/mtime: atime( ${base_atime} == ${after_atime} ), ctime( ${base_ctime} != ${after_ctime} ), mtime( ${base_mtime} != ${after_mtime} )"
|
|
return 1
|
|
fi
|
|
|
|
#
|
|
# Update symbolic file -> Not update parent directory's atime/mtime/ctime
|
|
#
|
|
base_atime="${after_atime}"
|
|
base_ctime="${after_ctime}"
|
|
base_mtime="${after_mtime}"
|
|
|
|
touch "${TEST_PARENTDIR_SYMFILE}"
|
|
|
|
local after_atime; after_atime=$(get_atime "${TEST_PARENTDIR_PARENT}")
|
|
local after_ctime; after_ctime=$(get_ctime "${TEST_PARENTDIR_PARENT}")
|
|
local after_mtime; after_mtime=$(get_mtime "${TEST_PARENTDIR_PARENT}")
|
|
|
|
if [ "${base_atime}" != "${after_atime}" ] || [ "${base_ctime}" != "${after_ctime}" ] || [ "${base_mtime}" != "${after_mtime}" ]; then
|
|
echo "updating symbolic file expected updating ctime/mtime: atime( ${base_atime} == ${after_atime} ), ctime( ${base_ctime} == ${after_ctime} ), mtime( ${base_mtime} == ${after_mtime} )"
|
|
return 1
|
|
fi
|
|
|
|
#
|
|
# Rename symbolic link -> Update parent directory's mtime/ctime
|
|
#
|
|
base_atime="${after_atime}"
|
|
base_ctime="${after_ctime}"
|
|
base_mtime="${after_mtime}"
|
|
|
|
mv "${TEST_PARENTDIR_SYMFILE}" "${TEST_PARENTDIR_SYMFILE_MV}"
|
|
|
|
local after_atime; after_atime=$(get_atime "${TEST_PARENTDIR_PARENT}")
|
|
local after_ctime; after_ctime=$(get_ctime "${TEST_PARENTDIR_PARENT}")
|
|
local after_mtime; after_mtime=$(get_mtime "${TEST_PARENTDIR_PARENT}")
|
|
|
|
if [ "${base_atime}" != "${after_atime}" ] || [ "${base_ctime}" = "${after_ctime}" ] || [ "${base_mtime}" = "${after_mtime}" ]; then
|
|
echo "renaming symbolic file expected updating ctime/mtime: atime( ${base_atime} == ${after_atime} ), ctime( ${base_ctime} != ${after_ctime} ), mtime( ${base_mtime} != ${after_mtime} )"
|
|
return 1
|
|
fi
|
|
|
|
#
|
|
# Delete symbolic link -> Update parent directory's mtime/ctime
|
|
#
|
|
base_atime="${after_atime}"
|
|
base_ctime="${after_ctime}"
|
|
base_mtime="${after_mtime}"
|
|
|
|
rm "${TEST_PARENTDIR_SYMFILE_MV}"
|
|
|
|
local after_atime; after_atime=$(get_atime "${TEST_PARENTDIR_PARENT}")
|
|
local after_ctime; after_ctime=$(get_ctime "${TEST_PARENTDIR_PARENT}")
|
|
local after_mtime; after_mtime=$(get_mtime "${TEST_PARENTDIR_PARENT}")
|
|
|
|
if [ "${base_atime}" != "${after_atime}" ] || [ "${base_ctime}" = "${after_ctime}" ] || [ "${base_mtime}" = "${after_mtime}" ]; then
|
|
echo "deleting symbolic file expected updating ctime/mtime: atime( ${base_atime} == ${after_atime} ), ctime( ${base_ctime} != ${after_ctime} ), mtime( ${base_mtime} != ${after_mtime} )"
|
|
return 1
|
|
fi
|
|
|
|
#
|
|
# Delete file -> Update parent directory's mtime/ctime
|
|
#
|
|
base_atime="${after_atime}"
|
|
base_ctime="${after_ctime}"
|
|
base_mtime="${after_mtime}"
|
|
|
|
rm "${TEST_PARENTDIR_FILE_MV}"
|
|
|
|
local after_atime; after_atime=$(get_atime "${TEST_PARENTDIR_PARENT}")
|
|
local after_ctime; after_ctime=$(get_ctime "${TEST_PARENTDIR_PARENT}")
|
|
local after_mtime; after_mtime=$(get_mtime "${TEST_PARENTDIR_PARENT}")
|
|
|
|
if [ "${base_atime}" != "${after_atime}" ] || [ "${base_ctime}" = "${after_ctime}" ] || [ "${base_mtime}" = "${after_mtime}" ]; then
|
|
echo "deleting file expected updating ctime/mtime: atime( ${base_atime} == ${after_atime} ), ctime( ${base_ctime} != ${after_ctime} ), mtime( ${base_mtime} != ${after_mtime} )"
|
|
return 1
|
|
fi
|
|
|
|
#
|
|
# Create directory -> Update parent directory's mtime/ctime
|
|
#
|
|
local base_atime; base_atime=$(get_atime "${TEST_PARENTDIR_PARENT}")
|
|
local base_ctime; base_ctime=$(get_ctime "${TEST_PARENTDIR_PARENT}")
|
|
local base_mtime; base_mtime=$(get_mtime "${TEST_PARENTDIR_PARENT}")
|
|
|
|
mkdir "${TEST_PARENTDIR_DIR}"
|
|
|
|
local after_atime; after_atime=$(get_atime "${TEST_PARENTDIR_PARENT}")
|
|
local after_ctime; after_ctime=$(get_ctime "${TEST_PARENTDIR_PARENT}")
|
|
local after_mtime; after_mtime=$(get_mtime "${TEST_PARENTDIR_PARENT}")
|
|
|
|
if [ "${base_atime}" != "${after_atime}" ] || [ "${base_ctime}" = "${after_ctime}" ] || [ "${base_mtime}" = "${after_mtime}" ]; then
|
|
echo "creating directory expected updating ctime/mtime: atime( ${base_atime} == ${after_atime} ), ctime( ${base_ctime} != ${after_ctime} ), mtime( ${base_mtime} != ${after_mtime} )"
|
|
return 1
|
|
fi
|
|
|
|
#
|
|
# Update directory -> Not update parent directory's atime/mtime/ctime
|
|
#
|
|
base_atime="${after_atime}"
|
|
base_ctime="${after_ctime}"
|
|
base_mtime="${after_mtime}"
|
|
|
|
touch "${TEST_PARENTDIR_DIR}"
|
|
|
|
local after_atime; after_atime=$(get_atime "${TEST_PARENTDIR_PARENT}")
|
|
local after_ctime; after_ctime=$(get_ctime "${TEST_PARENTDIR_PARENT}")
|
|
local after_mtime; after_mtime=$(get_mtime "${TEST_PARENTDIR_PARENT}")
|
|
|
|
if [ "${base_atime}" != "${after_atime}" ] || [ "${base_ctime}" != "${after_ctime}" ] || [ "${base_mtime}" != "${after_mtime}" ]; then
|
|
echo "updating directory expected updating ctime/mtime: atime( ${base_atime} == ${after_atime} ), ctime( ${base_ctime} == ${after_ctime} ), mtime( ${base_mtime} == ${after_mtime} )"
|
|
return 1
|
|
fi
|
|
|
|
#
|
|
# Rename directory -> Update parent directory's mtime/ctime
|
|
#
|
|
base_atime="${after_atime}"
|
|
base_ctime="${after_ctime}"
|
|
base_mtime="${after_mtime}"
|
|
|
|
mv "${TEST_PARENTDIR_DIR}" "${TEST_PARENTDIR_DIR_MV}"
|
|
|
|
local after_atime; after_atime=$(get_atime "${TEST_PARENTDIR_PARENT}")
|
|
local after_ctime; after_ctime=$(get_ctime "${TEST_PARENTDIR_PARENT}")
|
|
local after_mtime; after_mtime=$(get_mtime "${TEST_PARENTDIR_PARENT}")
|
|
|
|
if [ "${base_atime}" != "${after_atime}" ] || [ "${base_ctime}" = "${after_ctime}" ] || [ "${base_mtime}" = "${after_mtime}" ]; then
|
|
echo "renaming directory expected updating ctime/mtime: atime( ${base_atime} == ${after_atime} ), ctime( ${base_ctime} != ${after_ctime} ), mtime( ${base_mtime} != ${after_mtime} )"
|
|
return 1
|
|
fi
|
|
|
|
#
|
|
# Delete directory -> Update parent directory's mtime/ctime
|
|
#
|
|
base_atime="${after_atime}"
|
|
base_ctime="${after_ctime}"
|
|
base_mtime="${after_mtime}"
|
|
|
|
rm -r "${TEST_PARENTDIR_DIR_MV}"
|
|
|
|
local after_atime; after_atime=$(get_atime "${TEST_PARENTDIR_PARENT}")
|
|
local after_ctime; after_ctime=$(get_ctime "${TEST_PARENTDIR_PARENT}")
|
|
local after_mtime; after_mtime=$(get_mtime "${TEST_PARENTDIR_PARENT}")
|
|
|
|
if [ "${base_atime}" != "${after_atime}" ] || [ "${base_ctime}" = "${after_ctime}" ] || [ "${base_mtime}" = "${after_mtime}" ]; then
|
|
echo "deleting directory expected updating ctime/mtime: atime( ${base_atime} == ${after_atime} ), ctime( ${base_ctime} != ${after_ctime} ), mtime( ${base_mtime} != ${after_mtime} )"
|
|
return 1
|
|
fi
|
|
|
|
return 0
|
|
}
|
|
|
|
function test_update_parent_directory_time() {
|
|
describe "Testing update time of parent directory..."
|
|
|
|
#
|
|
# Test sub directory
|
|
#
|
|
mk_test_dir
|
|
if ! test_update_parent_directory_time_sub "${TEST_DIR}"; then
|
|
echo "failed test about updating time of parent directory: ${TEST_DIR}"
|
|
return 1
|
|
fi
|
|
rm -rf "${TEST_DIR}"
|
|
|
|
#
|
|
# Test bucket top directory
|
|
#
|
|
# [NOTE]
|
|
# The current directory for test execution is "<mount point>/testrun-xxxx".
|
|
# This test checks in the directory at the top of the bucket.
|
|
#
|
|
if ! test_update_parent_directory_time_sub ".."; then
|
|
echo "failed test about updating time of parent directory: ${TEST_DIR}"
|
|
return 1
|
|
fi
|
|
|
|
return 0
|
|
}
|
|
|
|
function test_rm_rf_dir {
|
|
describe "Test that rm -rf will remove directory with contents ..."
|
|
# Create a dir with some files and directories
|
|
mkdir dir1
|
|
mkdir dir1/dir2
|
|
touch dir1/file1
|
|
touch dir1/dir2/file2
|
|
|
|
# Remove the dir with recursive rm
|
|
rm -rf dir1
|
|
|
|
if [ -e dir1 ]; then
|
|
echo "rm -rf did not remove $PWD/dir1"
|
|
return 1
|
|
fi
|
|
}
|
|
|
|
function test_posix_acl {
|
|
describe "Testing posix acl function ..."
|
|
|
|
#------------------------------------------------------
|
|
# Directory
|
|
#------------------------------------------------------
|
|
local POSIX_ACL_TEST_DIR1="posix_acl_dir1"
|
|
local POSIX_ACL_TEST_DIR2="posix_acl_dir2"
|
|
rm -rf "${POSIX_ACL_TEST_DIR1}"
|
|
rm -rf "${POSIX_ACL_TEST_DIR2}"
|
|
|
|
mkdir "${POSIX_ACL_TEST_DIR1}"
|
|
|
|
# [NOTE]
|
|
# Let the target username be "root".
|
|
# On some systems the USER environment variable may be
|
|
# empty, so using ${USER} will not give an accurate test.
|
|
#
|
|
TARGET_ACL_USER="root"
|
|
|
|
#
|
|
# Set posix acl(not default)
|
|
#
|
|
setfacl -m "u:${TARGET_ACL_USER}:rwx" "${POSIX_ACL_TEST_DIR1}"
|
|
if ! getfacl "${POSIX_ACL_TEST_DIR1}" | grep -q "^user:${TARGET_ACL_USER}:rwx"; then
|
|
echo "Could not set posix acl(not default) to ${POSIX_ACL_TEST_DIR1} directory"
|
|
return 1
|
|
fi
|
|
|
|
#
|
|
# Set posix acl(default)
|
|
#
|
|
setfacl -d -m "u:${TARGET_ACL_USER}:rwx" "${POSIX_ACL_TEST_DIR1}"
|
|
if ! getfacl "${POSIX_ACL_TEST_DIR1}" | grep -q "^default:user:${TARGET_ACL_USER}:rwx"; then
|
|
echo "Could not set posix acl(default) to ${POSIX_ACL_TEST_DIR1} directory"
|
|
return 1
|
|
fi
|
|
|
|
#
|
|
# Rename
|
|
#
|
|
mv "${POSIX_ACL_TEST_DIR1}" "${POSIX_ACL_TEST_DIR2}"
|
|
if ! getfacl "${POSIX_ACL_TEST_DIR2}" | grep -q "^user:${TARGET_ACL_USER}:rwx"; then
|
|
echo "Could not move with posix acl(not default) to ${POSIX_ACL_TEST_DIR2} directory"
|
|
return 1
|
|
fi
|
|
if ! getfacl "${POSIX_ACL_TEST_DIR2}" | grep -q "^default:user:${TARGET_ACL_USER}:rwx"; then
|
|
echo "Could not move with posix acl(default) to ${POSIX_ACL_TEST_DIR2} directory"
|
|
return 1
|
|
fi
|
|
|
|
#
|
|
# Copy with permission
|
|
#
|
|
cp -rp "${POSIX_ACL_TEST_DIR2}" "${POSIX_ACL_TEST_DIR1}"
|
|
if ! getfacl "${POSIX_ACL_TEST_DIR1}" | grep -q "^user:${TARGET_ACL_USER}:rwx"; then
|
|
echo "Could not copy with posix acl(not default) to ${POSIX_ACL_TEST_DIR1} directory"
|
|
return 1
|
|
fi
|
|
if ! getfacl "${POSIX_ACL_TEST_DIR1}" | grep -q "^default:user:${TARGET_ACL_USER}:rwx"; then
|
|
echo "Could not copy with posix acl(default) to ${POSIX_ACL_TEST_DIR1} directory"
|
|
return 1
|
|
fi
|
|
|
|
#
|
|
# Overwrite posix acl(not default)
|
|
#
|
|
setfacl -m "u:${TARGET_ACL_USER}:r-x" "${POSIX_ACL_TEST_DIR2}"
|
|
if ! getfacl "${POSIX_ACL_TEST_DIR2}" | grep -q "^user:${TARGET_ACL_USER}:r-x"; then
|
|
echo "Could not set posix acl(not default) to ${POSIX_ACL_TEST_DIR2} directory"
|
|
return 1
|
|
fi
|
|
|
|
#
|
|
# Overwrite posix acl(default)
|
|
#
|
|
setfacl -d -m "u:${TARGET_ACL_USER}:r-x" "${POSIX_ACL_TEST_DIR2}"
|
|
if ! getfacl "${POSIX_ACL_TEST_DIR2}" | grep -q "^default:user:${TARGET_ACL_USER}:r-x"; then
|
|
echo "Could not set posix acl(default) to ${POSIX_ACL_TEST_DIR2} directory"
|
|
return 1
|
|
fi
|
|
|
|
#
|
|
# Remove posix acl(default)
|
|
#
|
|
setfacl -k "${POSIX_ACL_TEST_DIR2}"
|
|
if getfacl "${POSIX_ACL_TEST_DIR2}" | grep -q "^default:user:${TARGET_ACL_USER}"; then
|
|
echo "Could not remove posix acl(default) to ${POSIX_ACL_TEST_DIR2} directory"
|
|
return 1
|
|
fi
|
|
|
|
#
|
|
# Remove posix acl(all)
|
|
#
|
|
setfacl -b "${POSIX_ACL_TEST_DIR2}"
|
|
if getfacl "${POSIX_ACL_TEST_DIR2}" | grep -q "^user:${TARGET_ACL_USER}"; then
|
|
echo "Could not remove posix acl(all) to ${POSIX_ACL_TEST_DIR2} directory"
|
|
return 1
|
|
fi
|
|
|
|
#
|
|
# Copy without permission
|
|
#
|
|
rm -rf "${POSIX_ACL_TEST_DIR2}"
|
|
cp -r "${POSIX_ACL_TEST_DIR1}" "${POSIX_ACL_TEST_DIR2}"
|
|
if getfacl "${POSIX_ACL_TEST_DIR2}" | grep -q "^default:user:${TARGET_ACL_USER}"; then
|
|
echo "Could not copy without posix acl(default) to ${POSIX_ACL_TEST_DIR2} directory"
|
|
return 1
|
|
fi
|
|
if getfacl "${POSIX_ACL_TEST_DIR2}" | grep -q "^user:${TARGET_ACL_USER}"; then
|
|
echo "Could not copy without posix acl(all) to ${POSIX_ACL_TEST_DIR2} directory"
|
|
return 1
|
|
fi
|
|
|
|
#------------------------------------------------------
|
|
# File
|
|
#------------------------------------------------------
|
|
local POSIX_ACL_TEST_FILE1="posix_acl_dir1/posix_acl_file1"
|
|
local POSIX_ACL_TEST_FILE2="posix_acl_dir1/posix_acl_file2"
|
|
local POSIX_ACL_TEST_FILE3="posix_acl_dir2/posix_acl_file3"
|
|
local POSIX_ACL_TEST_FILE4="posix_acl_dir2/posix_acl_file4"
|
|
mkdir "${POSIX_ACL_TEST_DIR2}"
|
|
touch "${POSIX_ACL_TEST_FILE1}"
|
|
|
|
#
|
|
# Check default inherited posix acl
|
|
#
|
|
if ! getfacl "${POSIX_ACL_TEST_FILE1}" | grep -q "^user:${TARGET_ACL_USER}:rwx"; then
|
|
echo "Could not set posix acl(inherited default) to ${POSIX_ACL_TEST_FILE1} file"
|
|
return 1
|
|
fi
|
|
|
|
#
|
|
# Overwrite posix acl
|
|
#
|
|
setfacl -m "u:${TARGET_ACL_USER}:r-x" "${POSIX_ACL_TEST_FILE1}"
|
|
if ! getfacl "${POSIX_ACL_TEST_FILE1}" | grep -q "^user:${TARGET_ACL_USER}:r-x"; then
|
|
echo "Could not overwrite posix acl to ${POSIX_ACL_TEST_FILE1} file"
|
|
return 1
|
|
fi
|
|
|
|
#
|
|
# Rename
|
|
#
|
|
mv "${POSIX_ACL_TEST_FILE1}" "${POSIX_ACL_TEST_FILE2}"
|
|
if ! getfacl "${POSIX_ACL_TEST_FILE2}" | grep -q "^user:${TARGET_ACL_USER}:r-x"; then
|
|
echo "Could not move with posix acl to ${POSIX_ACL_TEST_FILE2} file"
|
|
return 1
|
|
fi
|
|
|
|
#
|
|
# Copy with permission
|
|
#
|
|
cp -p "${POSIX_ACL_TEST_FILE2}" "${POSIX_ACL_TEST_FILE1}"
|
|
if ! getfacl "${POSIX_ACL_TEST_FILE1}" | grep -q "^user:${TARGET_ACL_USER}:r-x"; then
|
|
echo "Could not copy with posix acl to ${POSIX_ACL_TEST_FILE1} file"
|
|
return 1
|
|
fi
|
|
|
|
#
|
|
# Remove posix acl
|
|
#
|
|
setfacl -b "${POSIX_ACL_TEST_FILE2}"
|
|
if getfacl "${POSIX_ACL_TEST_FILE2}" | grep -q "^default:user:${TARGET_ACL_USER}"; then
|
|
echo "Could not remove posix acl to ${POSIX_ACL_TEST_FILE2} file"
|
|
return 1
|
|
fi
|
|
|
|
#
|
|
# Copy without permission(set parent directory default acl)
|
|
#
|
|
rm -f "${POSIX_ACL_TEST_FILE2}"
|
|
cp "${POSIX_ACL_TEST_FILE1}" "${POSIX_ACL_TEST_FILE2}"
|
|
if ! getfacl "${POSIX_ACL_TEST_FILE2}" | grep -q "^user:${TARGET_ACL_USER}:rwx"; then
|
|
echo "Could not copy without posix acl(inherited parent) to ${POSIX_ACL_TEST_FILE2} file"
|
|
return 1
|
|
fi
|
|
|
|
#
|
|
# Copy with permission(to no-acl directory)
|
|
#
|
|
cp -p "${POSIX_ACL_TEST_FILE1}" "${POSIX_ACL_TEST_FILE3}"
|
|
if ! getfacl "${POSIX_ACL_TEST_FILE3}" | grep -q "^user:${TARGET_ACL_USER}:r-x"; then
|
|
echo "Could not copy with posix acl to ${POSIX_ACL_TEST_FILE3} file in no-acl directory"
|
|
return 1
|
|
fi
|
|
|
|
#
|
|
# Copy without permission(to no-acl directory)
|
|
#
|
|
cp "${POSIX_ACL_TEST_FILE1}" "${POSIX_ACL_TEST_FILE4}"
|
|
if getfacl "${POSIX_ACL_TEST_FILE4}" | grep -q "^user:${TARGET_ACL_USER}"; then
|
|
echo "Could not copy without posix acl to ${POSIX_ACL_TEST_FILE4} file in no-acl directory"
|
|
return 1
|
|
fi
|
|
|
|
rm -rf "${POSIX_ACL_TEST_DIR1}"
|
|
rm -rf "${POSIX_ACL_TEST_DIR2}"
|
|
}
|
|
|
|
function test_copy_file {
|
|
describe "Test simple copy ..."
|
|
|
|
dd if=/dev/urandom of=/tmp/simple_file bs=1024 count=1
|
|
cp /tmp/simple_file copied_simple_file
|
|
cmp /tmp/simple_file copied_simple_file
|
|
|
|
rm_test_file /tmp/simple_file
|
|
rm_test_file copied_simple_file
|
|
}
|
|
|
|
function test_write_after_seek_ahead {
|
|
describe "Test writes succeed after a seek ahead ..."
|
|
dd if=/dev/zero of=testfile seek=1 count=1 bs=1024
|
|
rm_test_file testfile
|
|
}
|
|
|
|
function test_overwrite_existing_file_range {
|
|
describe "Test overwrite range succeeds ..."
|
|
dd if=<(seq 1000) of="${TEST_TEXT_FILE}"
|
|
dd if=/dev/zero of="${TEST_TEXT_FILE}" seek=1 count=1 bs=1024 conv=notrunc
|
|
|
|
# [NOTE]
|
|
# In ALPINE, comparisons using redirects often fail.
|
|
# Therefore, create a real file for comparison and use it.
|
|
#
|
|
dd if=<(seq 1000) of=/tmp/cmp_base_file
|
|
dd if=/dev/zero of=/tmp/cmp_base_file seek=1 count=1 bs=1024 conv=notrunc
|
|
|
|
cmp "${TEST_TEXT_FILE}" /tmp/cmp_base_file
|
|
|
|
rm_test_file /tmp/cmp_base_file
|
|
rm_test_file
|
|
}
|
|
|
|
function test_concurrent_directory_updates {
|
|
describe "Test concurrent updates to a directory ..."
|
|
for i in $(seq 5); do
|
|
echo foo > "${i}"
|
|
done
|
|
for _ in $(seq 10); do
|
|
for i in $(seq 5); do
|
|
local file
|
|
# shellcheck disable=SC2012,SC2046
|
|
file=$(ls $(seq 5) | "${SED_BIN}" -n "$((RANDOM % 5 + 1))p")
|
|
cat "${file}" >/dev/null || true
|
|
rm -f "${file}"
|
|
echo "foo" > "${file}" || true
|
|
done &
|
|
done
|
|
wait
|
|
# shellcheck disable=SC2046
|
|
rm -f $(seq 5)
|
|
}
|
|
|
|
function test_concurrent_reads {
|
|
describe "Test concurrent reads from a file ..."
|
|
../../junk_data $((BIG_FILE_BLOCK_SIZE * BIG_FILE_COUNT)) > "${TEST_TEXT_FILE}"
|
|
for _ in $(seq 10); do
|
|
dd if="${TEST_TEXT_FILE}" of=/dev/null seek=$((RANDOM % BIG_FILE_LENGTH)) count=16 bs=1024 &
|
|
done
|
|
wait
|
|
rm_test_file
|
|
}
|
|
|
|
function test_concurrent_writes {
|
|
describe "Test concurrent writes to a file ..."
|
|
|
|
../../junk_data $((BIG_FILE_BLOCK_SIZE * BIG_FILE_COUNT)) > "${TEST_TEXT_FILE}"
|
|
|
|
NUM_PROCS=10
|
|
PIDS=()
|
|
for _ in $(seq "${NUM_PROCS}"); do
|
|
dd if=/dev/zero of="${TEST_TEXT_FILE}" seek=$((RANDOM % BIG_FILE_LENGTH)) count=16 bs=1024 conv=notrunc &
|
|
PIDS+=($!)
|
|
done
|
|
|
|
GRC=0
|
|
for PID in "${PIDS[@]}"; do
|
|
wait "${PID}"
|
|
RC=$?
|
|
[ $RC -ne 0 ] && GRC="${RC}"
|
|
done
|
|
rm_test_file
|
|
|
|
if [ "${GRC}" -ne 0 ]; then
|
|
echo "unexpected return code: $GRC"
|
|
return 1
|
|
fi
|
|
}
|
|
|
|
function test_open_second_fd {
|
|
describe "read from an open fd ..."
|
|
rm_test_file second_fd_file
|
|
|
|
local RESULT
|
|
# shellcheck disable=SC2094
|
|
RESULT=$( (echo foo ; wc -c < second_fd_file >&2) 2>& 1>second_fd_file)
|
|
if [ "${RESULT}" -ne 4 ]; then
|
|
echo "size mismatch, expected: 4, was: ${RESULT}"
|
|
return 1
|
|
fi
|
|
rm_test_file second_fd_file
|
|
}
|
|
|
|
function test_write_multiple_offsets {
|
|
describe "test writing to multiple offsets ..."
|
|
../../write_multiblock -f "${TEST_TEXT_FILE}" -p "1024:1" -p "$((16 * 1024 * 1024)):1" -p "$((18 * 1024 * 1024)):1"
|
|
rm_test_file "${TEST_TEXT_FILE}"
|
|
}
|
|
|
|
function test_write_multiple_offsets_backwards {
|
|
describe "test writing to multiple offsets ..."
|
|
../../write_multiblock -f "${TEST_TEXT_FILE}" -p "$((20 * 1024 * 1024 + 1)):1" -p "$((10 * 1024 * 1024)):1"
|
|
rm_test_file "${TEST_TEXT_FILE}"
|
|
}
|
|
|
|
function test_clean_up_cache() {
|
|
describe "Test clean up cache ..."
|
|
|
|
local dir="many_files"
|
|
local count=25
|
|
mkdir -p "${dir}"
|
|
|
|
for x in $(seq "${count}"); do
|
|
../../junk_data 10485760 > "${dir}"/file-"${x}"
|
|
done
|
|
|
|
local file_list=("${dir}"/*);
|
|
local file_cnt="${#file_list[@]}"
|
|
if [ "${file_cnt}" != "${count}" ]; then
|
|
echo "Expected $count files but got ${file_cnt}"
|
|
echo "Files: " "${file_list[@]}"
|
|
rm -rf "${dir}"
|
|
return 1
|
|
fi
|
|
local CACHE_DISK_AVAIL_SIZE; CACHE_DISK_AVAIL_SIZE=$(get_disk_avail_size "${CACHE_DIR}")
|
|
if [ "${CACHE_DISK_AVAIL_SIZE}" -lt "${ENSURE_DISKFREE_SIZE}" ];then
|
|
echo "Cache disk avail size:${CACHE_DISK_AVAIL_SIZE} less than ensure_diskfree size:${ENSURE_DISKFREE_SIZE}"
|
|
rm -rf "${dir}"
|
|
return 1
|
|
fi
|
|
rm -rf "${dir}"
|
|
}
|
|
|
|
function test_content_type() {
|
|
describe "Test Content-Type detection ..."
|
|
|
|
local DIR_NAME; DIR_NAME=$(basename "${PWD}")
|
|
|
|
touch "test.txt"
|
|
check_content_type "${DIR_NAME}/test.txt" "text/plain"
|
|
|
|
touch "test.jpg"
|
|
check_content_type "${DIR_NAME}/test.jpg" "image/jpeg"
|
|
|
|
touch "test.bin"
|
|
check_content_type "${DIR_NAME}/test.bin" "application/octet-stream"
|
|
|
|
mkdir "test.dir"
|
|
check_content_type "${DIR_NAME}/test.dir/" "application/x-directory"
|
|
|
|
rm -f test.txt
|
|
rm -f test.jpg
|
|
rm -f test.bin
|
|
rm -rf test.dir
|
|
}
|
|
|
|
# create more files than -o max_stat_cache_size
|
|
function test_truncate_cache() {
|
|
describe "Test make cache files over max cache file size ..."
|
|
|
|
for dir in $(seq 2); do
|
|
mkdir "${dir}"
|
|
for file in $(seq 75); do
|
|
touch "${dir}/${file}"
|
|
done
|
|
ls "${dir}"
|
|
done
|
|
|
|
# shellcheck disable=SC2046
|
|
rm -rf $(seq 2)
|
|
}
|
|
|
|
function test_cache_file_stat() {
|
|
describe "Test cache file stat ..."
|
|
|
|
../../junk_data $((BIG_FILE_BLOCK_SIZE * BIG_FILE_COUNT)) > "${BIG_FILE}"
|
|
|
|
#
|
|
# The first argument of the script is "testrun-<random>" the directory name.
|
|
#
|
|
local CACHE_TESTRUN_DIR=$1
|
|
|
|
#
|
|
# get cache file inode number
|
|
#
|
|
local CACHE_FILE_INODE
|
|
CACHE_FILE_INODE=$(get_inode "${CACHE_DIR}/${TEST_BUCKET_1}/${CACHE_TESTRUN_DIR}/${BIG_FILE}")
|
|
if [ -z "${CACHE_FILE_INODE}" ]; then
|
|
echo "Not found cache file or failed to get inode: ${CACHE_DIR}/${TEST_BUCKET_1}/${CACHE_TESTRUN_DIR}/${BIG_FILE}"
|
|
return 1;
|
|
fi
|
|
|
|
#
|
|
# get lines from cache stat file
|
|
#
|
|
local CACHE_FILE_STAT_LINE_1; CACHE_FILE_STAT_LINE_1=$("${SED_BIN}" -n 1p "${CACHE_DIR}/.${TEST_BUCKET_1}.stat/${CACHE_TESTRUN_DIR}/${BIG_FILE}")
|
|
local CACHE_FILE_STAT_LINE_2; CACHE_FILE_STAT_LINE_2=$("${SED_BIN}" -n 2p "${CACHE_DIR}/.${TEST_BUCKET_1}.stat/${CACHE_TESTRUN_DIR}/${BIG_FILE}")
|
|
if [ -z "${CACHE_FILE_STAT_LINE_1}" ] || [ -z "${CACHE_FILE_STAT_LINE_2}" ]; then
|
|
echo "could not get first or second line from cache file stat: ${CACHE_DIR}/.${TEST_BUCKET_1}.stat/${CACHE_TESTRUN_DIR}/${BIG_FILE}"
|
|
return 1;
|
|
fi
|
|
|
|
#
|
|
# compare
|
|
#
|
|
if [ "${CACHE_FILE_STAT_LINE_1}" != "${CACHE_FILE_INODE}:${BIG_FILE_LENGTH}" ]; then
|
|
echo "first line(cache file stat) is different: \"${CACHE_FILE_STAT_LINE_1}\" != \"${CACHE_FILE_INODE}:${BIG_FILE_LENGTH}\""
|
|
return 1;
|
|
fi
|
|
if [ "${CACHE_FILE_STAT_LINE_2}" != "0:${BIG_FILE_LENGTH}:1:0" ]; then
|
|
echo "last line(cache file stat) is different: \"${CACHE_FILE_STAT_LINE_2}\" != \"0:${BIG_FILE_LENGTH}:1:0\""
|
|
return 1;
|
|
fi
|
|
|
|
#
|
|
# remove cache files directly
|
|
#
|
|
rm -f "${CACHE_DIR}/${TEST_BUCKET_1}/${CACHE_TESTRUN_DIR}/${BIG_FILE}"
|
|
rm -f "${CACHE_DIR}/.${TEST_BUCKET_1}.stat/${CACHE_TESTRUN_DIR}/${BIG_FILE}"
|
|
|
|
#
|
|
# write a byte into the middle(not the boundary) of the file
|
|
#
|
|
local CHECK_UPLOAD_OFFSET=$((10 * 1024 * 1024 + 17))
|
|
dd if=/dev/urandom of="${BIG_FILE}" bs=1 count=1 seek="${CHECK_UPLOAD_OFFSET}" conv=notrunc
|
|
|
|
#
|
|
# get cache file inode number
|
|
#
|
|
CACHE_FILE_INODE=$(get_inode "${CACHE_DIR}/${TEST_BUCKET_1}/${CACHE_TESTRUN_DIR}/${BIG_FILE}")
|
|
if [ -z "${CACHE_FILE_INODE}" ]; then
|
|
echo "Not found cache file or failed to get inode: ${CACHE_DIR}/${TEST_BUCKET_1}/${CACHE_TESTRUN_DIR}/${BIG_FILE}"
|
|
return 1;
|
|
fi
|
|
|
|
#
|
|
# get lines from cache stat file
|
|
#
|
|
CACHE_FILE_STAT_LINE_1=$("${SED_BIN}" -n 1p "${CACHE_DIR}/.${TEST_BUCKET_1}.stat/${CACHE_TESTRUN_DIR}/${BIG_FILE}")
|
|
local CACHE_FILE_STAT_LINE_E; CACHE_FILE_STAT_LINE_E=$(tail -1 "${CACHE_DIR}/.${TEST_BUCKET_1}.stat/${CACHE_TESTRUN_DIR}/${BIG_FILE}" 2>/dev/null)
|
|
if [ -z "${CACHE_FILE_STAT_LINE_1}" ] || [ -z "${CACHE_FILE_STAT_LINE_E}" ]; then
|
|
echo "could not get first or end line from cache file stat: ${CACHE_DIR}/.${TEST_BUCKET_1}.stat/${CACHE_TESTRUN_DIR}/${BIG_FILE}"
|
|
return 1;
|
|
fi
|
|
|
|
#
|
|
# check first and cache file length from last line
|
|
#
|
|
# we should check all stat lines, but there are cases where the value
|
|
# differs depending on the processing system etc., then the cache file
|
|
# size is calculated and compared.
|
|
#
|
|
local CACHE_LAST_OFFSET; CACHE_LAST_OFFSET=$(echo "${CACHE_FILE_STAT_LINE_E}" | cut -d ":" -f1)
|
|
local CACHE_LAST_SIZE; CACHE_LAST_SIZE=$(echo "${CACHE_FILE_STAT_LINE_E}" | cut -d ":" -f2)
|
|
local CACHE_TOTAL_SIZE=$((CACHE_LAST_OFFSET + CACHE_LAST_SIZE))
|
|
|
|
if [ "${CACHE_FILE_STAT_LINE_1}" != "${CACHE_FILE_INODE}:${BIG_FILE_LENGTH}" ]; then
|
|
echo "first line(cache file stat) is different: \"${CACHE_FILE_STAT_LINE_1}\" != \"${CACHE_FILE_INODE}:${BIG_FILE_LENGTH}\""
|
|
return 1;
|
|
fi
|
|
if [ "${BIG_FILE_LENGTH}" -ne "${CACHE_TOTAL_SIZE}" ]; then
|
|
echo "the file size indicated by the cache stat file is different: \"${BIG_FILE_LENGTH}\" != \"${CACHE_TOTAL_SIZE}\""
|
|
return 1;
|
|
fi
|
|
|
|
rm_test_file "${BIG_FILE}"
|
|
}
|
|
|
|
function test_zero_cache_file_stat() {
|
|
describe "Test zero byte cache file stat ..."
|
|
|
|
rm_test_file "${TEST_TEXT_FILE}"
|
|
|
|
#
|
|
# create empty file
|
|
#
|
|
touch "${TEST_TEXT_FILE}"
|
|
|
|
#
|
|
# The first argument of the script is "testrun-<random>" the directory name.
|
|
#
|
|
local CACHE_TESTRUN_DIR=$1
|
|
|
|
# [NOTE]
|
|
# The stat file is a one-line text file, expecting for "<inode>:0"(ex. "4543937: 0").
|
|
#
|
|
if ! head -1 "${CACHE_DIR}/.${TEST_BUCKET_1}.stat/${CACHE_TESTRUN_DIR}/${TEST_TEXT_FILE}" 2>/dev/null | grep -q ':0$' 2>/dev/null; then
|
|
echo "The cache file stat after creating an empty file is incorrect : ${CACHE_DIR}/.${TEST_BUCKET_1}.stat/${CACHE_TESTRUN_DIR}/${TEST_TEXT_FILE}"
|
|
return 1;
|
|
fi
|
|
rm_test_file "${TEST_TEXT_FILE}"
|
|
}
|
|
|
|
function test_upload_sparsefile {
|
|
describe "Testing upload sparse file ..."
|
|
|
|
rm_test_file "${BIG_FILE}"
|
|
rm -f "${TEMP_DIR}/${BIG_FILE}"
|
|
|
|
#
|
|
# Make all HOLE file
|
|
#
|
|
"${TRUNCATE_BIN}" "${BIG_FILE}" -s "${BIG_FILE_LENGTH}"
|
|
|
|
#
|
|
# Write some bytes to ABOUT middle in the file
|
|
# (Dare to remove the block breaks)
|
|
#
|
|
local WRITE_POS=$((BIG_FILE_LENGTH / 2 - 128))
|
|
echo -n "0123456789ABCDEF" | dd of="${TEMP_DIR}/${BIG_FILE}" bs=1 count=16 seek="${WRITE_POS}" conv=notrunc
|
|
|
|
#
|
|
# copy(upload) the file
|
|
#
|
|
cp "${TEMP_DIR}/${BIG_FILE}" "${BIG_FILE}"
|
|
|
|
#
|
|
# check
|
|
#
|
|
cmp "${TEMP_DIR}/${BIG_FILE}" "${BIG_FILE}"
|
|
|
|
rm_test_file "${BIG_FILE}"
|
|
rm -f "${TEMP_DIR}/${BIG_FILE}"
|
|
}
|
|
|
|
function test_mix_upload_entities() {
|
|
describe "Testing upload sparse files ..."
|
|
|
|
#
|
|
# Make test file
|
|
#
|
|
../../junk_data $((BIG_FILE_BLOCK_SIZE * BIG_FILE_COUNT)) > "${BIG_FILE}"
|
|
|
|
#
|
|
# If the cache option is enabled, delete the cache of uploaded files.
|
|
#
|
|
if [ -f "${CACHE_DIR}/${TEST_BUCKET_1}/${BIG_FILE}" ]; then
|
|
rm -f "${CACHE_DIR}/${TEST_BUCKET_1}/${BIG_FILE}"
|
|
fi
|
|
if [ -f "${CACHE_DIR}/.${TEST_BUCKET_1}.stat/${BIG_FILE}" ]; then
|
|
rm -f "${CACHE_DIR}/.${TEST_BUCKET_1}.stat/${BIG_FILE}"
|
|
fi
|
|
|
|
#
|
|
# Do a partial write to the file.
|
|
#
|
|
echo -n "0123456789ABCDEF" | dd of="${BIG_FILE}" bs=1 count=16 seek=0 conv=notrunc
|
|
echo -n "0123456789ABCDEF" | dd of="${BIG_FILE}" bs=1 count=16 seek=8192 conv=notrunc
|
|
echo -n "0123456789ABCDEF" | dd of="${BIG_FILE}" bs=1 count=16 seek=1073152 conv=notrunc
|
|
echo -n "0123456789ABCDEF" | dd of="${BIG_FILE}" bs=1 count=16 seek=26214400 conv=notrunc
|
|
echo -n "0123456789ABCDEF" | dd of="${BIG_FILE}" bs=1 count=16 seek=26222592 conv=notrunc
|
|
|
|
rm_test_file "${BIG_FILE}"
|
|
}
|
|
|
|
#
|
|
# [NOTE]
|
|
# This test runs last because it uses up disk space and may not recover.
|
|
# This may be a problem, especially on MacOS. (See the comment near the definition
|
|
# line for the ENSURE_DISKFREE_SIZE variable)
|
|
#
|
|
function test_ensurespace_move_file() {
|
|
describe "Testing upload(mv) file when disk space is not enough ..."
|
|
|
|
#
|
|
# Make test file which is not under mountpoint
|
|
#
|
|
mkdir -p "${CACHE_DIR}/.s3fs_test_tmpdir"
|
|
../../junk_data $((BIG_FILE_BLOCK_SIZE * BIG_FILE_COUNT)) > "${CACHE_DIR}/.s3fs_test_tmpdir/${BIG_FILE}"
|
|
|
|
#
|
|
# Backup file stat
|
|
#
|
|
local ORIGINAL_PERMISSIONS
|
|
ORIGINAL_PERMISSIONS=$(get_user_and_group "${CACHE_DIR}/.s3fs_test_tmpdir/${BIG_FILE}")
|
|
|
|
#
|
|
# Fill the disk size
|
|
#
|
|
local NOW_CACHE_DISK_AVAIL_SIZE; NOW_CACHE_DISK_AVAIL_SIZE=$(get_disk_avail_size "${CACHE_DIR}")
|
|
local TMP_FILE_NO=0
|
|
while true; do
|
|
local ALLOWED_USING_SIZE=$((NOW_CACHE_DISK_AVAIL_SIZE - ENSURE_DISKFREE_SIZE))
|
|
if [ "${ALLOWED_USING_SIZE}" -gt "${BIG_FILE_LENGTH}" ]; then
|
|
cp -p "${CACHE_DIR}/.s3fs_test_tmpdir/${BIG_FILE}" "${CACHE_DIR}/.s3fs_test_tmpdir/${BIG_FILE}_${TMP_FILE_NO}"
|
|
local TMP_FILE_NO=$((TMP_FILE_NO + 1))
|
|
else
|
|
break;
|
|
fi
|
|
done
|
|
|
|
#
|
|
# move file
|
|
#
|
|
mv "${CACHE_DIR}/.s3fs_test_tmpdir/${BIG_FILE}" "${BIG_FILE}"
|
|
|
|
#
|
|
# file stat
|
|
#
|
|
local MOVED_PERMISSIONS
|
|
MOVED_PERMISSIONS=$(get_user_and_group "${BIG_FILE}")
|
|
local MOVED_FILE_LENGTH
|
|
MOVED_FILE_LENGTH=$(get_size "${BIG_FILE}")
|
|
|
|
#
|
|
# check
|
|
#
|
|
if [ "${MOVED_PERMISSIONS}" != "${ORIGINAL_PERMISSIONS}" ]; then
|
|
echo "Failed to move file with permission"
|
|
return 1
|
|
fi
|
|
if [ "${MOVED_FILE_LENGTH}" -ne "${BIG_FILE_LENGTH}" ]; then
|
|
echo "Failed to move file with file length: ${MOVED_FILE_LENGTH} ${BIG_FILE_LENGTH}"
|
|
return 1
|
|
fi
|
|
|
|
rm_test_file "${BIG_FILE}"
|
|
rm -rf "${CACHE_DIR}/.s3fs_test_tmpdir"
|
|
}
|
|
|
|
function test_not_existed_dir_obj() {
|
|
describe "Test not existed directory object..."
|
|
|
|
local DIR_NAME; DIR_NAME=$(basename "${PWD}")
|
|
|
|
#
|
|
# Create files under not existed directory by aws command
|
|
#
|
|
local OBJECT_NAME_1; OBJECT_NAME_1="${DIR_NAME}/not_existed_dir_single/${TEST_TEXT_FILE}"
|
|
local OBJECT_NAME_2; OBJECT_NAME_2="${DIR_NAME}/not_existed_dir_parent/not_existed_dir_child/${TEST_TEXT_FILE}"
|
|
echo data1 | aws_cli s3 cp - "s3://${TEST_BUCKET_1}/${OBJECT_NAME_1}"
|
|
echo data2 | aws_cli s3 cp - "s3://${TEST_BUCKET_1}/${OBJECT_NAME_2}"
|
|
|
|
# Top directory
|
|
# shellcheck disable=SC2010
|
|
if ! ls -1 | grep -q '^not_existed_dir_single$'; then
|
|
echo "Expect to find \"not_existed_dir_single\" directory, but it is not found"
|
|
return 1
|
|
fi
|
|
# shellcheck disable=SC2010
|
|
if ! ls -1 | grep -q '^not_existed_dir_parent$'; then
|
|
echo "Expect to find \"not_existed_dir_parent\" directory, but it is not found"
|
|
return 1
|
|
fi
|
|
|
|
# Single nest directory
|
|
if ! stat not_existed_dir_single; then
|
|
echo "Expect to find \"not_existed_dir_single\" directory, but it is not found"
|
|
return 1
|
|
fi
|
|
# shellcheck disable=SC2010
|
|
if ! ls -1 not_existed_dir_single | grep -q "^${TEST_TEXT_FILE}\$"; then
|
|
echo "Expect to find \"not_existed_dir_single/${TEST_TEXT_FILE}\" file, but it is not found"
|
|
return 1
|
|
fi
|
|
# shellcheck disable=SC2010
|
|
if ! ls -1 "not_existed_dir_single/${TEST_TEXT_FILE}" | grep -q "^not_existed_dir_single/${TEST_TEXT_FILE}\$"; then
|
|
echo "Expect to find \"not_existed_dir_single/${TEST_TEXT_FILE}\" file, but it is not found"
|
|
return 1
|
|
fi
|
|
|
|
# Double nest directory
|
|
if ! stat not_existed_dir_parent; then
|
|
echo "Expect to find \"not_existed_dir_parent\" directory, but it is not found"
|
|
return 1
|
|
fi
|
|
# shellcheck disable=SC2010
|
|
if ! ls -1 not_existed_dir_parent | grep -q '^not_existed_dir_child'; then
|
|
echo "Expect to find \"not_existed_dir_parent/not_existed_dir_child\" directory, but it is not found"
|
|
return 1
|
|
fi
|
|
if ! stat not_existed_dir_parent/not_existed_dir_child; then
|
|
echo "Expect to find \"not_existed_dir_parent/not_existed_dir_child\" directory, but it is not found"
|
|
return 1
|
|
fi
|
|
# shellcheck disable=SC2010
|
|
if ! ls -1 not_existed_dir_parent/not_existed_dir_child | grep -q "^${TEST_TEXT_FILE}\$"; then
|
|
echo "Expect to find \"not_existed_dir_parent/not_existed_dir_child/${TEST_TEXT_FILE}\" directory, but it is not found"
|
|
return 1
|
|
fi
|
|
# shellcheck disable=SC2010
|
|
if ! ls -1 "not_existed_dir_parent/not_existed_dir_child/${TEST_TEXT_FILE}" | grep -q "^not_existed_dir_parent/not_existed_dir_child/${TEST_TEXT_FILE}\$"; then
|
|
echo "Expect to find \"not_existed_dir_parent/not_existed_dir_child/${TEST_TEXT_FILE}\" directory, but it is not found"
|
|
return 1
|
|
fi
|
|
|
|
rm -rf not_existed_dir_single
|
|
rm -rf not_existed_dir_parent
|
|
}
|
|
|
|
function test_ut_ossfs {
|
|
describe "Testing ossfs python ut..."
|
|
|
|
# shellcheck disable=SC2153
|
|
export TEST_BUCKET_MOUNT_POINT="${TEST_BUCKET_MOUNT_POINT_1}"
|
|
../../ut_test.py
|
|
}
|
|
|
|
function test_cr_filename {
|
|
describe "Testing filename with CR code ..."
|
|
|
|
# The following tests create a file, test it, and delete it.
|
|
# So this test just calls the following program.
|
|
#
|
|
../../cr_filename "${TEST_TEXT_FILE}"
|
|
}
|
|
|
|
#
|
|
# This test opens a file and writes multiple sets of data.
|
|
# The file is opened only once and multiple blocks of data are written
|
|
# to the file descriptor with a gap.
|
|
#
|
|
# That is, the data sets are written discontinuously.
|
|
# The data to be written uses multiple data that is less than or larger
|
|
# than the part size of the multi-part upload.
|
|
# The gap should be at least the part size of the multi-part upload.
|
|
# Write as shown below:
|
|
# <SOF>....<write data>....<write data>....<write data><EOF>
|
|
#
|
|
# There are two types of tests: new files and existing files.
|
|
# For existing files, the file size must be larger than where this test
|
|
# writes last position.
|
|
# <SOF>....<write data>....<write data>....<write data>...<EOF>
|
|
#
|
|
function test_write_data_with_skip() {
|
|
describe "Testing write data block with skipping block..."
|
|
|
|
#
|
|
# The first argument of the script is "testrun-<random>" the directory name.
|
|
#
|
|
local CACHE_TESTRUN_DIR=$1
|
|
|
|
local _SKIPWRITE_FILE="test_skipwrite"
|
|
local _TMP_SKIPWRITE_FILE="/tmp/${_SKIPWRITE_FILE}"
|
|
|
|
#------------------------------------------------------
|
|
# (1) test new file
|
|
#------------------------------------------------------
|
|
#
|
|
# Clean files
|
|
#
|
|
rm_test_file "${_SKIPWRITE_FILE}"
|
|
rm_test_file "${_TMP_SKIPWRITE_FILE}"
|
|
|
|
#
|
|
# Create new file in bucket and temporary directory(/tmp)
|
|
#
|
|
# Writing to the file is as follows:
|
|
# |<-- skip(12MB) --><-- write(1MB) --><-- skip(22MB) --><-- write(20MB) --><-- skip(23MB) --><-- write(1MB) -->| (79MB)
|
|
#
|
|
# As a result, areas that are not written to the file are mixed.
|
|
# The part that is not written has a HOLE that is truncate and filled
|
|
# with 0x00.
|
|
# Assuming that multipart upload is performed on a part-by-part basis,
|
|
# it will be as follows:
|
|
# part 1) 0x0.. 0x9FFFFF : <not write area(0x00)>
|
|
# part 2) 0xA00000..0x13FFFFF : 0xA00000..0xBFFFFF <not write area(0x00)>
|
|
# 0xC00000..0xCFFFFF <write area>
|
|
# 0xD00000..0x13FFFFF <not write area(0x00)>
|
|
# part 3) 0x1400000..0x1DFFFFF : <not write area(0x00)>
|
|
# part 4) 0x1E00000..0x27FFFFF : 0x1E00000..0x22FFFFF <not write area(0x00)>
|
|
# 0x2300000..0x27FFFFF <write area>
|
|
# part 5) 0x2800000..0x31FFFFF : <write area>
|
|
# part 6) 0x3200000..0x3BFFFFF : 0x3200000..0x36FFFFF <write area>
|
|
# 0x3700000..0x3BFFFFF <not write area(0x00)>
|
|
# part 7) 0x3C00000..0x45FFFFF : <not write area(0x00)>
|
|
# part 8) 0x4600000..0x4BFFFFF : 0x4600000..0x4AFFFFF <not write area(0x00)>
|
|
# 0x4B00000..0x4BFFFFF <write area>
|
|
#
|
|
../../write_multiblock -f "${_SKIPWRITE_FILE}" -f "${_TMP_SKIPWRITE_FILE}" -p 12582912:65536 -p 36700160:20971520 -p 78643200:65536
|
|
|
|
#
|
|
# delete cache file if using cache
|
|
#
|
|
if s3fs_args | grep -q use_cache; then
|
|
rm -f "${CACHE_DIR}/${TEST_BUCKET_1}/${CACHE_TESTRUN_DIR}/${_SKIPWRITE_FILE}"
|
|
rm -f "${CACHE_DIR}/.${TEST_BUCKET_1}.stat/${CACHE_TESTRUN_DIR}/${_SKIPWRITE_FILE}"
|
|
fi
|
|
|
|
#
|
|
# Compare
|
|
#
|
|
cmp "${_SKIPWRITE_FILE}" "${_TMP_SKIPWRITE_FILE}"
|
|
|
|
#------------------------------------------------------
|
|
# (2) test existed file
|
|
#------------------------------------------------------
|
|
# [NOTE]
|
|
# This test uses the file used in the previous test as an existing file.
|
|
#
|
|
if s3fs_args | grep -q use_cache; then
|
|
rm -f "${CACHE_DIR}/${TEST_BUCKET_1}/${CACHE_TESTRUN_DIR}/${_SKIPWRITE_FILE}"
|
|
rm -f "${CACHE_DIR}/.${TEST_BUCKET_1}.stat/${CACHE_TESTRUN_DIR}/${_SKIPWRITE_FILE}"
|
|
fi
|
|
|
|
#
|
|
# Over write data to existed file in bucket and temporary directory(/tmp)
|
|
#
|
|
# Writing to the file is as follows:
|
|
# |<----------------------------------------------- existed file ----------------------------------------------------------->| (79MB)
|
|
# |<-- skip(12MB) --><-- write(1MB) --><-- skip(22MB) --><-- write(20MB) --><-- skip(22MB) --><-- write(1MB) --><-- 1MB -->| (79MB)
|
|
#
|
|
# As a result, areas that are not written to the file are mixed.
|
|
# The part that is not written has a HOLE that is truncate and filled
|
|
# with 0x00.
|
|
# Assuming that multipart upload is performed on a part-by-part basis,
|
|
# it will be as follows:
|
|
# part 1) 0x0.. 0x9FFFFF : <not write area(0x00)>
|
|
# part 2) 0xA00000..0x13FFFFF : 0xA00000..0xBFFFFF <not write area(0x00)>
|
|
# 0xC00000..0xCFFFFF <write area>
|
|
# 0xD00000..0x13FFFFF <not write area(0x00)>
|
|
# part 3) 0x1400000..0x1DFFFFF : <not write area(0x00)>
|
|
# part 4) 0x1E00000..0x27FFFFF : 0x1E00000..0x22FFFFF <not write area(0x00)>
|
|
# 0x2300000..0x27FFFFF <write area>
|
|
# part 5) 0x2800000..0x31FFFFF : <write area>
|
|
# part 6) 0x3200000..0x3BFFFFF : 0x3200000..0x36FFFFF <write area>
|
|
# 0x3700000..0x3BFFFFF <not write area(0x00)>
|
|
# part 7) 0x3C00000..0x45FFFFF : <not write area(0x00)>
|
|
# part 8) 0x4600000..0x4BFFFFF : 0x4600000..0x49FFFFF <not write area(0x00)>
|
|
# part 8) 0x4600000..0x4BFFFFF : 0x4A00000..0x4AFFFFF <write area>
|
|
# 0x4B00000..0x4BFFFFF <not write area(0x00)>
|
|
#
|
|
../../write_multiblock -f "${_SKIPWRITE_FILE}" -f "${_TMP_SKIPWRITE_FILE}" -p 12582912:65536 -p 36700160:20971520 -p 77594624:65536
|
|
|
|
#
|
|
# delete cache file if using cache
|
|
#
|
|
if s3fs_args | grep -q use_cache; then
|
|
rm -f "${CACHE_DIR}/${TEST_BUCKET_1}/${CACHE_TESTRUN_DIR}/${_SKIPWRITE_FILE}"
|
|
rm -f "${CACHE_DIR}/.${TEST_BUCKET_1}.stat/${CACHE_TESTRUN_DIR}/${_SKIPWRITE_FILE}"
|
|
fi
|
|
|
|
#
|
|
# Compare
|
|
#
|
|
cmp "${_SKIPWRITE_FILE}" "${_TMP_SKIPWRITE_FILE}"
|
|
|
|
#
|
|
# Clean files
|
|
#
|
|
rm_test_file "${_SKIPWRITE_FILE}"
|
|
rm_test_file "${_TMP_SKIPWRITE_FILE}"
|
|
}
|
|
|
|
function test_not_boundary_writes {
|
|
describe "Test non-boundary write ..."
|
|
|
|
# [MEMO]
|
|
# Files used in this test, multipart related sizes, etc.
|
|
#
|
|
# Test file size: 25MB(25 * 1024 * 1024)
|
|
# Multipart size: 10MB
|
|
# Multipart minimum upload size: 5MB
|
|
#
|
|
# The multipart upload part that should be executed here is as follows:
|
|
# Part number 1: 0 - 10,485,759 (size = 10MB)
|
|
# Part number 2: 10,485,760 - 20,971,519 (size = 10MB)
|
|
# Part number 3: 20,971,520 - 26,214,399 (size = 5MB)
|
|
#
|
|
local BOUNDARY_TEST_FILE_SIZE; BOUNDARY_TEST_FILE_SIZE=$((BIG_FILE_BLOCK_SIZE * BIG_FILE_COUNT))
|
|
|
|
../../junk_data "${BOUNDARY_TEST_FILE_SIZE}" > "${TEST_TEXT_FILE}"
|
|
|
|
#
|
|
# Write in First boundary
|
|
#
|
|
# Write 0 - 3,145,727(3MB) : less than the multipart minimum size from the beginning
|
|
dd if=/dev/zero of="${TEST_TEXT_FILE}" seek=0 count=3072 bs=1024 conv=notrunc
|
|
|
|
# Write 0 - 7,340,031(7MB) : multipart exceeding the minimum size from the beginning
|
|
dd if=/dev/zero of="${TEST_TEXT_FILE}" seek=0 count=7168 bs=1024 conv=notrunc
|
|
|
|
# Write 0 - 12,582,911(12MB) : beyond the multipart size boundary from the beginning
|
|
dd if=/dev/zero of="${TEST_TEXT_FILE}" seek=0 count=12288 bs=1024 conv=notrunc
|
|
|
|
#
|
|
# Write in First and second boundary
|
|
#
|
|
# Write 3,145,728 - 4,194,303(1MB) : less than the minimum multipart size from the middle of the first multipart area
|
|
dd if=/dev/zero of="${TEST_TEXT_FILE}" seek=3072 count=1024 bs=1024 conv=notrunc
|
|
|
|
# Write 3,145,728 - 9,437,183(6MB) : exceeding the minimum multipart size from the middle of the first multipart area
|
|
dd if=/dev/zero of="${TEST_TEXT_FILE}" seek=3072 count=6144 bs=1024 conv=notrunc
|
|
|
|
# Write 3,145,728 - 12,582,911(9MB) : beyond the multipart boundary from the middle of the first multipart area
|
|
dd if=/dev/zero of="${TEST_TEXT_FILE}" seek=3072 count=9216 bs=1024 conv=notrunc
|
|
|
|
#
|
|
# Write in Second boundary
|
|
#
|
|
# Write 12,582,912 - 14,680,063(2MB) : below the minimum multipart size from the middle of the multipart area
|
|
dd if=/dev/zero of="${TEST_TEXT_FILE}" seek=12288 count=2048 bs=1024 conv=notrunc
|
|
|
|
# Write 12,582,912 - 18,874,367(6MB) : data exceeding the minimum multipart size from the middle of the multipart area
|
|
dd if=/dev/zero of="${TEST_TEXT_FILE}" seek=12288 count=6144 bs=1024 conv=notrunc
|
|
|
|
# Write 12,582,912 - 23,068,671(10MB) : beyond the multipart boundary from the middle of the multipart area
|
|
dd if=/dev/zero of="${TEST_TEXT_FILE}" seek=12288 count=10240 bs=1024 conv=notrunc
|
|
|
|
# Write 12,582,912 - 26,214,399(13MB) : beyond the multipart boundary(last) from the middle of the multipart area
|
|
dd if=/dev/zero of="${TEST_TEXT_FILE}" seek=12288 count=13312 bs=1024 conv=notrunc
|
|
|
|
#
|
|
# Write in Last boundary
|
|
#
|
|
# Write 23,068,672 - 24,117,247(1MB) : below the minimum multipart size from the middle of the final multipart area
|
|
dd if=/dev/zero of="${TEST_TEXT_FILE}" seek=22528 count=1024 bs=1024 conv=notrunc
|
|
|
|
# Write 23,068,672 - 26,214,399(3MB) : beyond the multipart boundary(last) from the middle of the final multipart area
|
|
dd if=/dev/zero of="${TEST_TEXT_FILE}" seek=22528 count=3072 bs=1024 conv=notrunc
|
|
|
|
rm_test_file
|
|
}
|
|
|
|
function test_chmod_mountpoint {
|
|
describe "Testing chmod to mount point..."
|
|
|
|
local MOUNTPOINT_DIR; MOUNTPOINT_DIR=$(cd ..; pwd)
|
|
|
|
local ORIGINAL_PERMISSIONS; ORIGINAL_PERMISSIONS=$(get_permissions "${MOUNTPOINT_DIR}")
|
|
|
|
chmod 0777 "${MOUNTPOINT_DIR}";
|
|
|
|
# if they're the same, we have a problem.
|
|
local CHANGED_PERMISSIONS; CHANGED_PERMISSIONS=$(get_permissions "${MOUNTPOINT_DIR}")
|
|
if [ "${CHANGED_PERMISSIONS}" = "${ORIGINAL_PERMISSIONS}" ]
|
|
then
|
|
echo "Could not modify mount point(${MOUNTPOINT_DIR}) permissions"
|
|
return 1
|
|
fi
|
|
}
|
|
|
|
function test_chown_mountpoint {
|
|
describe "Testing chown mount point..."
|
|
|
|
local MOUNTPOINT_DIR; MOUNTPOINT_DIR=$(cd ..; pwd)
|
|
|
|
local ORIGINAL_PERMISSIONS
|
|
ORIGINAL_PERMISSIONS=$(get_user_and_group "${MOUNTPOINT_DIR}")
|
|
|
|
# [NOTE]
|
|
# Prevents test interruptions due to permission errors, etc.
|
|
# If the chown command fails, an error will occur with the
|
|
# following judgment statement. So skip the chown command error.
|
|
# '|| true' was added due to a problem with MacOS and ensure_diskfree option.
|
|
#
|
|
chown 1000:1000 "${MOUNTPOINT_DIR}" || true
|
|
|
|
# if they're the same, we have a problem.
|
|
local CHANGED_PERMISSIONS
|
|
CHANGED_PERMISSIONS=$(get_user_and_group "${MOUNTPOINT_DIR}")
|
|
if [ "${CHANGED_PERMISSIONS}" = "${ORIGINAL_PERMISSIONS}" ]
|
|
then
|
|
if [ "${ORIGINAL_PERMISSIONS}" = "1000:1000" ]
|
|
then
|
|
echo "Could not be strict check because original file permission 1000:1000"
|
|
else
|
|
echo "Could not modify mount point(${MOUNTPOINT_DIR}) ownership($ORIGINAL_PERMISSIONS to 1000:1000)"
|
|
return 1
|
|
fi
|
|
fi
|
|
}
|
|
|
|
function test_time_mountpoint {
|
|
describe "Testing atime/ctime/mtime to mount point..."
|
|
|
|
local MOUNTPOINT_DIR; MOUNTPOINT_DIR=$(cd ..; pwd)
|
|
|
|
local base_atime; base_atime=$(get_atime "${MOUNTPOINT_DIR}")
|
|
local base_ctime; base_ctime=$(get_ctime "${MOUNTPOINT_DIR}")
|
|
local base_mtime; base_mtime=$(get_mtime "${MOUNTPOINT_DIR}")
|
|
|
|
touch "${MOUNTPOINT_DIR}"
|
|
|
|
local atime; atime=$(get_atime "${MOUNTPOINT_DIR}")
|
|
local ctime; ctime=$(get_ctime "${MOUNTPOINT_DIR}")
|
|
local mtime; mtime=$(get_mtime "${MOUNTPOINT_DIR}")
|
|
|
|
if [ "${base_atime}" = "${atime}" ] || [ "${base_ctime}" = "${ctime}" ] || [ "${base_mtime}" = "${mtime}" ]; then
|
|
echo "chmod expected updated ctime: $base_ctime != $ctime, mtime: $base_mtime == $mtime, atime: $base_atime == $atime"
|
|
return 1
|
|
fi
|
|
}
|
|
|
|
function test_file_names_longer_than_posix() {
|
|
local DIR_NAME; DIR_NAME=$(basename "${PWD}")
|
|
a256="aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa"
|
|
#a256="aaaa"
|
|
|
|
if ! touch "${a256}"; then
|
|
echo "could not create long file name"
|
|
return 1
|
|
fi
|
|
rm -f "${a256}"
|
|
|
|
echo data | aws_cli s3 cp - "s3://${TEST_BUCKET_1}/${DIR_NAME}/${a256}"
|
|
files=(*)
|
|
if [ "${#files[@]}" = 0 ]; then
|
|
echo "failed to list long file name"
|
|
return 1
|
|
fi
|
|
rm -f "${a256}"
|
|
}
|
|
|
|
function test_statvfs() {
|
|
describe "Testing the free/available size on mount point(statvfs)..."
|
|
|
|
# [NOTE]
|
|
# The df command result format is different between Linux and macos,
|
|
# but the order of Total/Used/Available size is the same.
|
|
#
|
|
local MOUNTPOINT_DIR; MOUNTPOINT_DIR=$(cd ..; pwd)
|
|
local DF_RESULT; DF_RESULT=$(df "${MOUNTPOINT_DIR}" 2>/dev/null | tail -n +2)
|
|
local TOTAL_SIZE; TOTAL_SIZE=$(echo "${DF_RESULT}" | awk '{print $2}')
|
|
local USED_SIZE; USED_SIZE=$(echo "${DF_RESULT}" | awk '{print $3}')
|
|
local AVAIL_SIZE; AVAIL_SIZE=$(echo "${DF_RESULT}" | awk '{print $4}')
|
|
|
|
# [NOTE]
|
|
# In the disk information (statvfs) provided by s3fs, Total size and
|
|
# Available size are always the same and not 0, and used size is always 0.
|
|
#
|
|
if [ -z "${TOTAL_SIZE}" ] || [ -z "${AVAIL_SIZE}" ] || [ -z "${USED_SIZE}" ] || [ "${TOTAL_SIZE}" = "0" ] || [ "${AVAIL_SIZE}" = "0" ] || [ "${TOTAL_SIZE}" != "${AVAIL_SIZE}" ] || [ "${USED_SIZE}" != "0" ]; then
|
|
echo "The result of df <mount point> command is wrong: Total=${TOTAL_SIZE}, Used=${USED_SIZE}, Available=${AVAIL_SIZE}"
|
|
return 1
|
|
fi
|
|
}
|
|
|
|
function add_all_tests {
|
|
if s3fs_args | grep -q use_cache; then
|
|
add_tests test_cache_file_stat
|
|
add_tests test_zero_cache_file_stat
|
|
else
|
|
add_tests test_file_names_longer_than_posix
|
|
fi
|
|
if ! s3fs_args | grep -q ensure_diskfree && ! uname | grep -q Darwin; then
|
|
add_tests test_clean_up_cache
|
|
fi
|
|
add_tests test_create_empty_file
|
|
add_tests test_append_file
|
|
add_tests test_truncate_file
|
|
add_tests test_truncate_upload
|
|
add_tests test_truncate_empty_file
|
|
add_tests test_truncate_shrink_file
|
|
add_tests test_truncate_shrink_read_file
|
|
add_tests test_mv_file
|
|
add_tests test_mv_to_exist_file
|
|
add_tests test_mv_empty_directory
|
|
add_tests test_mv_nonempty_directory
|
|
add_tests test_redirects
|
|
add_tests test_mkdir_rmdir
|
|
add_tests test_chmod
|
|
add_tests test_chown
|
|
add_tests test_list
|
|
add_tests test_remove_nonempty_directory
|
|
add_tests test_external_directory_creation
|
|
add_tests test_external_modification
|
|
add_tests test_external_creation
|
|
add_tests test_read_external_object
|
|
add_tests test_read_external_dir_object
|
|
add_tests test_update_metadata_external_small_object
|
|
add_tests test_update_metadata_external_large_object
|
|
add_tests test_rename_before_close
|
|
add_tests test_multipart_upload
|
|
add_tests test_multipart_copy
|
|
add_tests test_multipart_mix
|
|
add_tests test_utimens_during_multipart
|
|
add_tests test_special_characters
|
|
add_tests test_hardlink
|
|
add_tests test_symlink
|
|
if ! uname | grep -q Darwin; then
|
|
add_tests test_mknod
|
|
fi
|
|
add_tests test_extended_attributes
|
|
add_tests test_mtime_file
|
|
|
|
add_tests test_update_time_chmod
|
|
add_tests test_update_time_chown
|
|
add_tests test_update_time_xattr
|
|
add_tests test_update_time_touch
|
|
if ! mount -t fuse.s3fs | grep "$TEST_BUCKET_MOUNT_POINT_1 " | grep -q -e noatime -e relatime ; then
|
|
add_tests test_update_time_touch_a
|
|
fi
|
|
add_tests test_update_time_append
|
|
add_tests test_update_time_cp_p
|
|
add_tests test_update_time_mv
|
|
|
|
add_tests test_update_directory_time_chmod
|
|
add_tests test_update_directory_time_chown
|
|
add_tests test_update_directory_time_set_xattr
|
|
add_tests test_update_directory_time_touch
|
|
if ! mount -t fuse.s3fs | grep "$TEST_BUCKET_MOUNT_POINT_1 " | grep -q -e noatime -e relatime ; then
|
|
add_tests test_update_directory_time_touch_a
|
|
fi
|
|
add_tests test_update_directory_time_subdir
|
|
add_tests test_update_chmod_opened_file
|
|
if s3fs_args | grep -q update_parent_dir_stat; then
|
|
add_tests test_update_parent_directory_time
|
|
fi
|
|
if ! s3fs_args | grep -q use_xattr; then
|
|
add_tests test_posix_acl
|
|
fi
|
|
|
|
add_tests test_rm_rf_dir
|
|
add_tests test_copy_file
|
|
add_tests test_write_after_seek_ahead
|
|
add_tests test_overwrite_existing_file_range
|
|
add_tests test_concurrent_directory_updates
|
|
add_tests test_concurrent_reads
|
|
add_tests test_concurrent_writes
|
|
add_tests test_open_second_fd
|
|
add_tests test_write_multiple_offsets
|
|
add_tests test_write_multiple_offsets_backwards
|
|
add_tests test_content_type
|
|
add_tests test_truncate_cache
|
|
add_tests test_upload_sparsefile
|
|
add_tests test_mix_upload_entities
|
|
# TODO: investigate why only Alpine cannot see the implicit directory objects.
|
|
if ! test -f /etc/os-release || ! grep -q -i -e 'ID=alpine' -e 'ID="alpine"' /etc/os-release; then
|
|
add_tests test_not_existed_dir_obj
|
|
fi
|
|
add_tests test_ut_ossfs
|
|
add_tests test_cr_filename
|
|
if ! s3fs_args | grep -q ensure_diskfree && ! uname | grep -q Darwin; then
|
|
add_tests test_ensurespace_move_file
|
|
fi
|
|
add_tests test_write_data_with_skip
|
|
add_tests test_not_boundary_writes
|
|
|
|
# [NOTE]
|
|
# The test on CI will fail depending on the permissions, so skip these(chmod/chown).
|
|
#
|
|
# add_tests test_chmod_mountpoint
|
|
# add_tests test_chown_mountpoint
|
|
add_tests test_time_mountpoint
|
|
add_tests test_statvfs
|
|
}
|
|
|
|
init_suite
|
|
add_all_tests
|
|
run_suite
|
|
|
|
#
|
|
# Local variables:
|
|
# tab-width: 4
|
|
# c-basic-offset: 4
|
|
# End:
|
|
# vim600: expandtab sw=4 ts=4 fdm=marker
|
|
# vim<600: expandtab sw=4 ts=4
|
|
#
|