#!/bin/bash # # s3fs - FUSE-based file system backed by Amazon S3 # # Copyright 2007-2008 Randy Rizun # # This program is free software; you can redistribute it and/or # modify it under the terms of the GNU General Public License # as published by the Free Software Foundation; either version 2 # of the License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program; if not, write to the Free Software # Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. # set -o errexit set -o pipefail source test-utils.sh function test_create_empty_file { describe "Testing creating an empty file ..." local OBJECT_NAME; OBJECT_NAME=$(basename "${PWD}")/"${TEST_TEXT_FILE}" touch "${TEST_TEXT_FILE}" check_file_size "${TEST_TEXT_FILE}" 0 aws_cli s3api head-object --bucket "${TEST_BUCKET_1}" --key "${OBJECT_NAME}" rm_test_file } function test_append_file { describe "Testing append to file ..." local TEST_INPUT="echo ${TEST_TEXT} to ${TEST_TEXT_FILE}" # Write a small test file for x in $(seq 1 "${TEST_TEXT_FILE_LENGTH}"); do echo "${TEST_INPUT}" done > "${TEST_TEXT_FILE}" check_file_size "${TEST_TEXT_FILE}" $((TEST_TEXT_FILE_LENGTH * $((${#TEST_INPUT} + 1)) )) rm_test_file } function test_truncate_file { describe "Testing truncate file ..." # Write a small test file echo "${TEST_TEXT}" > "${TEST_TEXT_FILE}" # Truncate file to 0 length. This should trigger open(path, O_RDWR | O_TRUNC...) : > "${TEST_TEXT_FILE}" check_file_size "${TEST_TEXT_FILE}" 0 rm_test_file } function test_truncate_upload { describe "Testing truncate file for uploading ..." # This file size uses multipart, mix upload when uploading. # We will test these cases. rm_test_file "${BIG_FILE}" "${TRUNCATE_BIN}" "${BIG_FILE}" -s "${BIG_FILE_LENGTH}" rm_test_file "${BIG_FILE}" } function test_truncate_empty_file { describe "Testing truncate empty file ..." # Write an empty test file touch "${TEST_TEXT_FILE}" # Truncate the file to 1024 length local t_size=1024 "${TRUNCATE_BIN}" "${TEST_TEXT_FILE}" -s "${t_size}" check_file_size "${TEST_TEXT_FILE}" "${t_size}" rm_test_file } function test_truncate_shrink_file { describe "Testing truncate shrinking large binary file ..." local BIG_TRUNCATE_TEST_FILE="big-truncate-test.bin" local t_size=$((1024 * 1024 * 32 + 64)) dd if=/dev/urandom of="${TEMP_DIR}/${BIG_TRUNCATE_TEST_FILE}" bs=1024 count=$((1024 * 64)) cp "${TEMP_DIR}/${BIG_TRUNCATE_TEST_FILE}" "${BIG_TRUNCATE_TEST_FILE}" "${TRUNCATE_BIN}" "${TEMP_DIR}/${BIG_TRUNCATE_TEST_FILE}" -s "${t_size}" "${TRUNCATE_BIN}" "${BIG_TRUNCATE_TEST_FILE}" -s "${t_size}" if ! cmp "${TEMP_DIR}/${BIG_TRUNCATE_TEST_FILE}" "${BIG_TRUNCATE_TEST_FILE}"; then return 1 fi rm -f "${TEMP_DIR}/${BIG_TRUNCATE_TEST_FILE}" rm_test_file "${BIG_TRUNCATE_TEST_FILE}" } function test_truncate_shrink_read_file { describe "Testing truncate(shrink) and read file ..." # Initiate file size is 1024, and shrink file size is 512 local init_size=1024 local shrink_size=512 # create file dd if=/dev/urandom of="${TEST_TEXT_FILE}" bs="${init_size}" count=1 # truncate(shrink) file and read it before flusing ../../truncate_read_file "${TEST_TEXT_FILE}" "${shrink_size}" # check file size check_file_size "${TEST_TEXT_FILE}" "${shrink_size}" # Truncate the file to 1024 length local t_size=1024 rm_test_file } function test_mv_file { describe "Testing mv file function ..." # if the rename file exists, delete it if [ -e "${ALT_TEST_TEXT_FILE}" ] then rm "${ALT_TEST_TEXT_FILE}" fi if [ -e "${ALT_TEST_TEXT_FILE}" ] then echo "Could not delete file ${ALT_TEST_TEXT_FILE}, it still exists" return 1 fi # create the test file again mk_test_file # save file length local ALT_TEXT_LENGTH; ALT_TEXT_LENGTH=$(wc -c "${TEST_TEXT_FILE}" | awk '{print $1}') #rename the test file mv "${TEST_TEXT_FILE}" "${ALT_TEST_TEXT_FILE}" if [ ! -e "${ALT_TEST_TEXT_FILE}" ] then echo "Could not move file" return 1 fi #check the renamed file content-type if [ -f "/etc/mime.types" ] then check_content_type "$1/${ALT_TEST_TEXT_FILE}" "text/plain" fi # Check the contents of the alt file local ALT_FILE_LENGTH; ALT_FILE_LENGTH=$(wc -c "${ALT_TEST_TEXT_FILE}" | awk '{print $1}') if [ "$ALT_FILE_LENGTH" -ne "$ALT_TEXT_LENGTH" ] then echo "moved file length is not as expected expected: $ALT_TEXT_LENGTH got: $ALT_FILE_LENGTH" return 1 fi # clean up rm_test_file "${ALT_TEST_TEXT_FILE}" } function test_mv_to_exist_file { describe "Testing mv file to exist file function ..." local BIG_MV_FILE_BLOCK_SIZE=$((BIG_FILE_BLOCK_SIZE + 1)) ../../junk_data $((BIG_FILE_BLOCK_SIZE * BIG_FILE_COUNT)) > "${BIG_FILE}" ../../junk_data $((BIG_MV_FILE_BLOCK_SIZE * BIG_FILE_COUNT)) > "${BIG_FILE}-mv" mv "${BIG_FILE}" "${BIG_FILE}-mv" rm_test_file "${BIG_FILE}-mv" } function test_mv_empty_directory { describe "Testing mv directory function ..." if [ -e "${TEST_DIR}" ]; then echo "Unexpected, this file/directory exists: ${TEST_DIR}" return 1 fi mk_test_dir mv "${TEST_DIR}" "${TEST_DIR}_rename" if [ ! -d "${TEST_DIR}_rename" ]; then echo "Directory ${TEST_DIR} was not renamed" return 1 fi rmdir "${TEST_DIR}_rename" if [ -e "${TEST_DIR}_rename" ]; then echo "Could not remove the test directory, it still exists: ${TEST_DIR}_rename" return 1 fi } function test_mv_nonempty_directory { describe "Testing mv directory function ..." if [ -e "${TEST_DIR}" ]; then echo "Unexpected, this file/directory exists: ${TEST_DIR}" return 1 fi mk_test_dir touch "${TEST_DIR}"/file mv "${TEST_DIR}" "${TEST_DIR}_rename" if [ ! -d "${TEST_DIR}_rename" ]; then echo "Directory ${TEST_DIR} was not renamed" return 1 fi rm -r "${TEST_DIR}_rename" if [ -e "${TEST_DIR}_rename" ]; then echo "Could not remove the test directory, it still exists: ${TEST_DIR}_rename" return 1 fi } function test_redirects { describe "Testing redirects ..." mk_test_file "ABCDEF" local CONTENT; CONTENT=$(cat "${TEST_TEXT_FILE}") if [ "${CONTENT}" != "ABCDEF" ]; then echo "CONTENT read is unexpected, got ${CONTENT}, expected ABCDEF" return 1 fi echo "XYZ" > "${TEST_TEXT_FILE}" CONTENT=$(cat "${TEST_TEXT_FILE}") if [ "${CONTENT}" != "XYZ" ]; then echo "CONTENT read is unexpected, got ${CONTENT}, expected XYZ" return 1 fi echo "123456" >> "${TEST_TEXT_FILE}" local LINE1; LINE1=$("${SED_BIN}" -n '1,1p' "${TEST_TEXT_FILE}") local LINE2; LINE2=$("${SED_BIN}" -n '2,2p' "${TEST_TEXT_FILE}") if [ "${LINE1}" != "XYZ" ]; then echo "LINE1 was not as expected, got ${LINE1}, expected XYZ" return 1 fi if [ "${LINE2}" != "123456" ]; then echo "LINE2 was not as expected, got ${LINE2}, expected 123456" return 1 fi # clean up rm_test_file } function test_mkdir_rmdir { describe "Testing creation/removal of a directory ..." if [ -e "${TEST_DIR}" ]; then echo "Unexpected, this file/directory exists: ${TEST_DIR}" return 1 fi mk_test_dir rm_test_dir } function test_chmod { describe "Testing chmod file function ..." # create the test file again mk_test_file local ORIGINAL_PERMISSIONS; ORIGINAL_PERMISSIONS=$(get_permissions "${TEST_TEXT_FILE}") chmod 777 "${TEST_TEXT_FILE}"; # if they're the same, we have a problem. local CHANGED_PERMISSIONS; CHANGED_PERMISSIONS=$(get_permissions "${TEST_TEXT_FILE}") if [ "${CHANGED_PERMISSIONS}" = "${ORIGINAL_PERMISSIONS}" ] then echo "Could not modify ${TEST_TEXT_FILE} permissions" return 1 fi # clean up rm_test_file } function test_chown { describe "Testing chown file function ..." # create the test file again mk_test_file local ORIGINAL_PERMISSIONS ORIGINAL_PERMISSIONS=$(get_user_and_group "${TEST_TEXT_FILE}") # [NOTE] # Prevents test interruptions due to permission errors, etc. # If the chown command fails, an error will occur with the # following judgment statement. So skip the chown command error. # '|| true' was added due to a problem with Travis CI and MacOS # and ensure_diskfree option. # chown 1000:1000 "${TEST_TEXT_FILE}" || true # if they're the same, we have a problem. local CHANGED_PERMISSIONS CHANGED_PERMISSIONS=$(get_user_and_group "${TEST_TEXT_FILE}") if [ "${CHANGED_PERMISSIONS}" = "${ORIGINAL_PERMISSIONS}" ] then if [ "${ORIGINAL_PERMISSIONS}" = "1000:1000" ] then echo "Could not be strict check because original file permission 1000:1000" else echo "Could not modify ${TEST_TEXT_FILE} ownership($ORIGINAL_PERMISSIONS to 1000:1000)" return 1 fi fi # clean up rm_test_file } function test_list { describe "Testing list ..." mk_test_file mk_test_dir local file_list=(*) local file_cnt=${#file_list[@]} if [ "${file_cnt}" -ne 2 ]; then echo "Expected 2 file but got ${file_cnt}" echo "Files: " "${file_list[@]}" return 1 fi rm_test_file rm_test_dir } function test_remove_nonempty_directory { describe "Testing removing a non-empty directory ..." mk_test_dir touch "${TEST_DIR}/file" ( set +o pipefail rmdir "${TEST_DIR}" 2>&1 | grep -q "Directory not empty" ) rm "${TEST_DIR}/file" rm_test_dir } function test_external_directory_creation { describe "Test external directory creation ..." local OBJECT_NAME; OBJECT_NAME=$(basename "${PWD}")/directory/"${TEST_TEXT_FILE}" echo "data" | aws_cli s3 cp - "s3://${TEST_BUCKET_1}/${OBJECT_NAME}" # shellcheck disable=SC2010 ls | grep -q directory stat directory >/dev/null 2>&1 get_permissions directory | grep -q 750$ ls directory cmp <(echo "data") directory/"${TEST_TEXT_FILE}" rm -f directory/"${TEST_TEXT_FILE}" } function test_external_modification { describe "Test external modification to an object ..." echo "old" > "${TEST_TEXT_FILE}" # [NOTE] # If the stat and file cache directory are enabled, an error will # occur if the unixtime(sec) value does not change. # If mtime(ctime/atime) when updating from the external program # (awscli) is the same unixtime value as immediately before, the # cache will be read out. # Therefore, we need to wait over 1 second here. # sleep 1 local OBJECT_NAME; OBJECT_NAME=$(basename "${PWD}")/"${TEST_TEXT_FILE}" echo "new new" | aws_cli s3 cp - "s3://${TEST_BUCKET_1}/${OBJECT_NAME}" cmp "${TEST_TEXT_FILE}" <(echo "new new") rm -f "${TEST_TEXT_FILE}" } function test_external_creation { describe "Test external creation of an object ..." local OBJECT_NAME; OBJECT_NAME=$(basename "${PWD}")/"${TEST_TEXT_FILE}" # [NOTE] # If noobj_cache is enabled, register that cache here. # [ ! -e "${TEST_TEXT_FILE}" ] # [NOTE] # If noobj_cache is enabled, we cannot be sure that it is registered in that cache. # That's because an error will occur if the upload by aws cli takes more than 1 second. # echo "data" | aws_cli s3 cp - "s3://${TEST_BUCKET_1}/${OBJECT_NAME}" sleep 1 [ -e "${TEST_TEXT_FILE}" ] rm -f "${TEST_TEXT_FILE}" } function test_read_external_object() { describe "create objects via aws CLI and read via s3fs ..." local OBJECT_NAME; OBJECT_NAME=$(basename "${PWD}")/"${TEST_TEXT_FILE}" echo "test" | aws_cli s3 cp - "s3://${TEST_BUCKET_1}/${OBJECT_NAME}" cmp "${TEST_TEXT_FILE}" <(echo "test") rm -f "${TEST_TEXT_FILE}" } function test_read_external_dir_object() { describe "create directory objects via aws CLI and read via s3fs ..." local SUB_DIR_NAME; SUB_DIR_NAME="subdir" local SUB_DIR_TEST_FILE; SUB_DIR_TEST_FILE="${SUB_DIR_NAME}/${TEST_TEXT_FILE}" local OBJECT_NAME; OBJECT_NAME=$(basename "${PWD}")/"${SUB_DIR_TEST_FILE}" echo "test" | aws_cli s3 cp - "s3://${TEST_BUCKET_1}/${OBJECT_NAME}" if stat "${SUB_DIR_NAME}" | grep -q '1969-12-31[[:space:]]23:59:59[.]000000000'; then echo "sub directory a/c/m time is underflow(-1)." return 1 fi rm -rf "${SUB_DIR_NAME}" } function test_update_metadata_external_small_object() { describe "update meta to small file after created file by aws cli" # [NOTE] # Use the only filename in the test to avoid being affected by noobjcache. # local TEST_FILE_EXT; TEST_FILE_EXT=$(make_random_string) local TEST_CHMOD_FILE="${TEST_TEXT_FILE}_chmod.${TEST_FILE_EXT}" local TEST_CHOWN_FILE="${TEST_TEXT_FILE}_chown.${TEST_FILE_EXT}" local TEST_UTIMENS_FILE="${TEST_TEXT_FILE}_utimens.${TEST_FILE_EXT}" local TEST_SETXATTR_FILE="${TEST_TEXT_FILE}_xattr.${TEST_FILE_EXT}" local TEST_RMXATTR_FILE="${TEST_TEXT_FILE}_xattr.${TEST_FILE_EXT}" local TEST_INPUT="TEST_STRING_IN_SMALL_FILE" # # chmod # local OBJECT_NAME; OBJECT_NAME=$(basename "${PWD}")/"${TEST_CHMOD_FILE}" echo "${TEST_INPUT}" | aws_cli s3 cp - "s3://${TEST_BUCKET_1}/${OBJECT_NAME}" chmod +x "${TEST_CHMOD_FILE}" cmp "${TEST_CHMOD_FILE}" <(echo "${TEST_INPUT}") # # chown # OBJECT_NAME=$(basename "${PWD}")/"${TEST_CHOWN_FILE}" echo "${TEST_INPUT}" | aws_cli s3 cp - "s3://${TEST_BUCKET_1}/${OBJECT_NAME}" chown "${UID}" "${TEST_CHOWN_FILE}" cmp "${TEST_CHOWN_FILE}" <(echo "${TEST_INPUT}") # # utimens # OBJECT_NAME=$(basename "${PWD}")/"${TEST_UTIMENS_FILE}" echo "${TEST_INPUT}" | aws_cli s3 cp - "s3://${TEST_BUCKET_1}/${OBJECT_NAME}" touch "${TEST_UTIMENS_FILE}" cmp "${TEST_UTIMENS_FILE}" <(echo "${TEST_INPUT}") # # set xattr # OBJECT_NAME=$(basename "${PWD}")/"${TEST_SETXATTR_FILE}" echo "${TEST_INPUT}" | aws_cli s3 cp - "s3://${TEST_BUCKET_1}/${OBJECT_NAME}" set_xattr key value "${TEST_SETXATTR_FILE}" cmp "${TEST_SETXATTR_FILE}" <(echo "${TEST_INPUT}") # # remove xattr # # "%7B%22key%22%3A%22dmFsdWU%3D%22%7D" = {"key":"value"} # OBJECT_NAME=$(basename "${PWD}")/"${TEST_RMXATTR_FILE}" echo "${TEST_INPUT}" | aws_cli s3 cp - "s3://${TEST_BUCKET_1}/${OBJECT_NAME}" --metadata xattr=%7B%22key%22%3A%22dmFsdWU%3D%22%7D del_xattr key "${TEST_RMXATTR_FILE}" cmp "${TEST_RMXATTR_FILE}" <(echo "${TEST_INPUT}") rm -f "${TEST_CHMOD_FILE}" rm -f "${TEST_CHOWN_FILE}" rm -f "${TEST_UTIMENS_FILE}" rm -f "${TEST_SETXATTR_FILE}" rm -f "${TEST_RMXATTR_FILE}" } function test_update_metadata_external_large_object() { describe "update meta to large file after created file by aws cli" # [NOTE] # Use the only filename in the test to avoid being affected by noobjcache. # local TEST_FILE_EXT; TEST_FILE_EXT=$(make_random_string) local TEST_CHMOD_FILE="${TEST_TEXT_FILE}_chmod.${TEST_FILE_EXT}" local TEST_CHOWN_FILE="${TEST_TEXT_FILE}_chown.${TEST_FILE_EXT}" local TEST_UTIMENS_FILE="${TEST_TEXT_FILE}_utimens.${TEST_FILE_EXT}" local TEST_SETXATTR_FILE="${TEST_TEXT_FILE}_xattr.${TEST_FILE_EXT}" local TEST_RMXATTR_FILE="${TEST_TEXT_FILE}_xattr.${TEST_FILE_EXT}" ../../junk_data $((BIG_FILE_BLOCK_SIZE * BIG_FILE_COUNT)) > "${TEMP_DIR}/${BIG_FILE}" # # chmod # local OBJECT_NAME; OBJECT_NAME=$(basename "${PWD}")/"${TEST_CHMOD_FILE}" aws_cli s3 cp "${TEMP_DIR}/${BIG_FILE}" "s3://${TEST_BUCKET_1}/${OBJECT_NAME}" --no-progress chmod +x "${TEST_CHMOD_FILE}" cmp "${TEST_CHMOD_FILE}" "${TEMP_DIR}/${BIG_FILE}" # # chown # OBJECT_NAME=$(basename "${PWD}")/"${TEST_CHOWN_FILE}" aws_cli s3 cp "${TEMP_DIR}/${BIG_FILE}" "s3://${TEST_BUCKET_1}/${OBJECT_NAME}" --no-progress chown "${UID}" "${TEST_CHOWN_FILE}" cmp "${TEST_CHOWN_FILE}" "${TEMP_DIR}/${BIG_FILE}" # # utimens # OBJECT_NAME=$(basename "${PWD}")/"${TEST_UTIMENS_FILE}" aws_cli s3 cp "${TEMP_DIR}/${BIG_FILE}" "s3://${TEST_BUCKET_1}/${OBJECT_NAME}" --no-progress touch "${TEST_UTIMENS_FILE}" cmp "${TEST_UTIMENS_FILE}" "${TEMP_DIR}/${BIG_FILE}" # # set xattr # OBJECT_NAME=$(basename "${PWD}")/"${TEST_SETXATTR_FILE}" aws_cli s3 cp "${TEMP_DIR}/${BIG_FILE}" "s3://${TEST_BUCKET_1}/${OBJECT_NAME}" --no-progress set_xattr key value "${TEST_SETXATTR_FILE}" cmp "${TEST_SETXATTR_FILE}" "${TEMP_DIR}/${BIG_FILE}" # # remove xattr # # "%7B%22key%22%3A%22dmFsdWU%3D%22%7D" = {"key":"value"} # OBJECT_NAME=$(basename "${PWD}")/"${TEST_RMXATTR_FILE}" aws_cli s3 cp "${TEMP_DIR}/${BIG_FILE}" "s3://${TEST_BUCKET_1}/${OBJECT_NAME}" --no-progress --metadata xattr=%7B%22key%22%3A%22dmFsdWU%3D%22%7D del_xattr key "${TEST_RMXATTR_FILE}" cmp "${TEST_RMXATTR_FILE}" "${TEMP_DIR}/${BIG_FILE}" rm -f "${TEMP_DIR}/${BIG_FILE}" rm -f "${TEST_CHMOD_FILE}" rm -f "${TEST_CHOWN_FILE}" rm -f "${TEST_UTIMENS_FILE}" rm -f "${TEST_SETXATTR_FILE}" rm -f "${TEST_RMXATTR_FILE}" } function test_rename_before_close { describe "Testing rename before close ..." # shellcheck disable=SC2094 ( echo foo mv "${TEST_TEXT_FILE}" "${TEST_TEXT_FILE}.new" ) > "${TEST_TEXT_FILE}" if ! cmp <(echo "foo") "${TEST_TEXT_FILE}.new"; then echo "rename before close failed" return 1 fi rm_test_file "${TEST_TEXT_FILE}.new" rm -f "${TEST_TEXT_FILE}" } function test_multipart_upload { describe "Testing multi-part upload ..." ../../junk_data $((BIG_FILE_BLOCK_SIZE * BIG_FILE_COUNT)) > "${TEMP_DIR}/${BIG_FILE}" dd if="${TEMP_DIR}/${BIG_FILE}" of="${BIG_FILE}" bs="${BIG_FILE_BLOCK_SIZE}" count="${BIG_FILE_COUNT}" # Verify contents of file echo "Comparing test file" if ! cmp "${TEMP_DIR}/${BIG_FILE}" "${BIG_FILE}" then return 1 fi rm -f "${TEMP_DIR}/${BIG_FILE}" rm_test_file "${BIG_FILE}" } function test_multipart_copy { describe "Testing multi-part copy ..." ../../junk_data $((BIG_FILE_BLOCK_SIZE * BIG_FILE_COUNT)) > "${TEMP_DIR}/${BIG_FILE}" dd if="${TEMP_DIR}/${BIG_FILE}" of="${BIG_FILE}" bs="${BIG_FILE_BLOCK_SIZE}" count="${BIG_FILE_COUNT}" mv "${BIG_FILE}" "${BIG_FILE}-copy" # Verify contents of file echo "Comparing test file" if ! cmp "${TEMP_DIR}/${BIG_FILE}" "${BIG_FILE}-copy" then return 1 fi #check the renamed file content-type check_content_type "$1/${BIG_FILE}-copy" "application/octet-stream" rm -f "${TEMP_DIR}/${BIG_FILE}" rm_test_file "${BIG_FILE}-copy" } function test_multipart_mix { describe "Testing multi-part mix ..." ../../junk_data $((BIG_FILE_BLOCK_SIZE * BIG_FILE_COUNT)) > "${TEMP_DIR}/${BIG_FILE}" dd if="${TEMP_DIR}/${BIG_FILE}" of="${BIG_FILE}" bs="${BIG_FILE_BLOCK_SIZE}" count="${BIG_FILE_COUNT}" # (1) Edit the middle of an existing file # modify directly(seek 7.5MB offset) # In the case of nomultipart and nocopyapi, # it makes no sense, but copying files is because it leaves no cache. # cp "${TEMP_DIR}/${BIG_FILE}" "${TEMP_DIR}/${BIG_FILE}-mix" cp "${BIG_FILE}" "${BIG_FILE}-mix" local MODIFY_START_BLOCK=$((15*1024*1024/2/4)) echo -n "0123456789ABCDEF" | dd of="${BIG_FILE}-mix" bs=4 count=4 seek="${MODIFY_START_BLOCK}" conv=notrunc echo -n "0123456789ABCDEF" | dd of="${TEMP_DIR}/${BIG_FILE}-mix" bs=4 count=4 seek="${MODIFY_START_BLOCK}" conv=notrunc # Verify contents of file echo "Comparing test file (1)" if ! cmp "${TEMP_DIR}/${BIG_FILE}-mix" "${BIG_FILE}-mix" then return 1 fi # (2) Write to an area larger than the size of the existing file # modify directly(over file end offset) # cp "${TEMP_DIR}/${BIG_FILE}" "${TEMP_DIR}/${BIG_FILE}-mix" cp "${BIG_FILE}" "${BIG_FILE}-mix" local OVER_FILE_BLOCK_POS=$((26*1024*1024/4)) echo -n "0123456789ABCDEF" | dd of="${BIG_FILE}-mix" bs=4 count=4 seek="${OVER_FILE_BLOCK_POS}" conv=notrunc echo -n "0123456789ABCDEF" | dd of="${TEMP_DIR}/${BIG_FILE}-mix" bs=4 count=4 seek="${OVER_FILE_BLOCK_POS}" conv=notrunc # Verify contents of file echo "Comparing test file (2)" if ! cmp "${TEMP_DIR}/${BIG_FILE}-mix" "${BIG_FILE}-mix" then return 1 fi # (3) Writing from the 0th byte # cp "${TEMP_DIR}/${BIG_FILE}" "${TEMP_DIR}/${BIG_FILE}-mix" cp "${BIG_FILE}" "${BIG_FILE}-mix" echo -n "0123456789ABCDEF" | dd of="${BIG_FILE}-mix" bs=4 count=4 seek=0 conv=notrunc echo -n "0123456789ABCDEF" | dd of="${TEMP_DIR}/${BIG_FILE}-mix" bs=4 count=4 seek=0 conv=notrunc # Verify contents of file echo "Comparing test file (3)" if ! cmp "${TEMP_DIR}/${BIG_FILE}-mix" "${BIG_FILE}-mix" then return 1 fi # (4) Write to the area within 5MB from the top # modify directly(seek 1MB offset) # cp "${TEMP_DIR}/${BIG_FILE}" "${TEMP_DIR}/${BIG_FILE}-mix" cp "${BIG_FILE}" "${BIG_FILE}-mix" local MODIFY_START_BLOCK=$((1*1024*1024)) echo -n "0123456789ABCDEF" | dd of="${BIG_FILE}-mix" bs=4 count=4 seek="${MODIFY_START_BLOCK}" conv=notrunc echo -n "0123456789ABCDEF" | dd of="${TEMP_DIR}/${BIG_FILE}-mix" bs=4 count=4 seek="${MODIFY_START_BLOCK}" conv=notrunc # Verify contents of file echo "Comparing test file (4)" if ! cmp "${TEMP_DIR}/${BIG_FILE}-mix" "${BIG_FILE}-mix" then return 1 fi # [NOTE] # For macos, in order to free up disk space for statvfs (or df command), # it is necessary to zero out the file size, delete it, and sync it. # In the case of macos, even if you delete a file, there seems to be a # delay in the free space being reflected. # Testing the ensure_diskfree option shows that if this is not done, free # disk space will be exhausted. # if [ "$(uname)" = "Darwin" ]; then cat /dev/null > "${TEMP_DIR}/${BIG_FILE}" cat /dev/null > "${TEMP_DIR}/${BIG_FILE}-mix" cat /dev/null > "${BIG_FILE}" cat /dev/null > "${BIG_FILE}-mix" fi rm -f "${TEMP_DIR}/${BIG_FILE}" rm -f "${TEMP_DIR}/${BIG_FILE}-mix" rm_test_file "${BIG_FILE}" rm_test_file "${BIG_FILE}-mix" if [ "$(uname)" = "Darwin" ]; then sync fi } function test_utimens_during_multipart { describe "Testing utimens calling during multipart copy ..." ../../junk_data $((BIG_FILE_BLOCK_SIZE * BIG_FILE_COUNT)) > "${TEMP_DIR}/${BIG_FILE}" cp "${TEMP_DIR}/${BIG_FILE}" "${BIG_FILE}" # The second copy of the "-p" option calls utimens during multipart upload. cp -p "${TEMP_DIR}/${BIG_FILE}" "${BIG_FILE}" rm -f "${TEMP_DIR}/${BIG_FILE}" rm_test_file "${BIG_FILE}" } function test_special_characters { describe "Testing special characters ..." ( set +o pipefail # shellcheck disable=SC2010 ls 'special' 2>&1 | grep -q 'No such file or directory' # shellcheck disable=SC2010 ls 'special?' 2>&1 | grep -q 'No such file or directory' # shellcheck disable=SC2010 ls 'special*' 2>&1 | grep -q 'No such file or directory' # shellcheck disable=SC2010 ls 'special~' 2>&1 | grep -q 'No such file or directory' # shellcheck disable=SC2010 ls 'specialĪ¼' 2>&1 | grep -q 'No such file or directory' ) mkdir "TOYOTA TRUCK 8.2.2" rm -rf "TOYOTA TRUCK 8.2.2" } function test_hardlink { describe "Testing hardlinks ..." rm -f "${TEST_TEXT_FILE}" rm -f "${ALT_TEST_TEXT_FILE}" echo foo > "${TEST_TEXT_FILE}" ( set +o pipefail ln "${TEST_TEXT_FILE}" "${ALT_TEST_TEXT_FILE}" 2>&1 | grep -q -e 'Operation not supported' -e 'Not supported' ) rm_test_file rm_test_file "${ALT_TEST_TEXT_FILE}" } function test_mknod { describe "Testing mknod system call function ..." local MKNOD_TEST_FILE_BASENAME="mknod_testfile" rm -f "${MKNOD_TEST_FILE_BASENAME}*" ../../mknod_test "${MKNOD_TEST_FILE_BASENAME}" } function test_symlink { describe "Testing symlinks ..." rm -f "${TEST_TEXT_FILE}" rm -f "${ALT_TEST_TEXT_FILE}" echo foo > "${TEST_TEXT_FILE}" ln -s "${TEST_TEXT_FILE}" "${ALT_TEST_TEXT_FILE}" cmp "${TEST_TEXT_FILE}" "${ALT_TEST_TEXT_FILE}" rm -f "${TEST_TEXT_FILE}" [ -L "${ALT_TEST_TEXT_FILE}" ] [ ! -f "${ALT_TEST_TEXT_FILE}" ] rm -f "${ALT_TEST_TEXT_FILE}" } function test_extended_attributes { describe "Testing extended attributes ..." rm -f "${TEST_TEXT_FILE}" touch "${TEST_TEXT_FILE}" # set value set_xattr key1 value0 "${TEST_TEXT_FILE}" get_xattr key1 "${TEST_TEXT_FILE}" | grep -q '^value0$' # over write value set_xattr key1 value1 "${TEST_TEXT_FILE}" get_xattr key1 "${TEST_TEXT_FILE}" | grep -q '^value1$' # append value set_xattr key2 value2 "${TEST_TEXT_FILE}" get_xattr key1 "${TEST_TEXT_FILE}" | grep -q '^value1$' get_xattr key2 "${TEST_TEXT_FILE}" | grep -q '^value2$' # remove value del_xattr key1 "${TEST_TEXT_FILE}" get_xattr key1 "${TEST_TEXT_FILE}" && return 1 get_xattr key2 "${TEST_TEXT_FILE}" | grep -q '^value2$' rm_test_file } function test_mtime_file { describe "Testing mtime preservation function ..." # if the rename file exists, delete it if [ -e "${ALT_TEST_TEXT_FILE}" ] || [ -L "${ALT_TEST_TEXT_FILE}" ] then rm "${ALT_TEST_TEXT_FILE}" fi if [ -e "${ALT_TEST_TEXT_FILE}" ] then echo "Could not delete file ${ALT_TEST_TEXT_FILE}, it still exists" return 1 fi # create the test file again mk_test_file #copy the test file with preserve mode cp -p "${TEST_TEXT_FILE}" "${ALT_TEST_TEXT_FILE}" local testmtime; testmtime=$(get_mtime "${TEST_TEXT_FILE}") local testctime; testctime=$(get_ctime "${TEST_TEXT_FILE}") local testatime; testatime=$(get_atime "${TEST_TEXT_FILE}") local altmtime; altmtime=$(get_mtime "${ALT_TEST_TEXT_FILE}") local altctime; altctime=$(get_ctime "${ALT_TEST_TEXT_FILE}") local altatime; altatime=$(get_atime "${ALT_TEST_TEXT_FILE}") if [ "${testmtime}" != "${altmtime}" ] || [ "${testctime}" = "${altctime}" ] || [ "${testatime}" != "${altatime}" ]; then # [NOTE]{FIXME] # On macos10, the mtime of the file copied by "cp -p" is # truncated to usec from nsec, and it cannot be solved. # This is because the timespec.tv_sec value of the mtime # of the original file is truncated in usec units at calling # s3fs_utimens. # (ex. "1658768609.505917125" vs "1658768609.505917000") # Now this workaround is not found, so for macos compare # mtime with only usec. # if ! uname | grep -q Darwin; then echo "cp(-p) expected times: mtime( ${testmtime} == ${altmtime} ), ctime( ${testctime} != ${altctime} ), atime( ${testatime} == ${altatime} )" return 1 else testmtime=$(echo "${testmtime}" | cut -c 1-17) altmtime=$(echo "${altmtime}" | cut -c 1-17) if [ "${testmtime}" != "${altmtime}" ] || [ "${testctime}" = "${altctime}" ] || [ "${testatime}" != "${altatime}" ]; then echo "cp(-p) expected times: mtime( ${testmtime} == ${altmtime} ), ctime( ${testctime} != ${altctime} ), atime( ${testatime} == ${altatime} )" return 1 fi fi fi rm_test_file rm_test_file "${ALT_TEST_TEXT_FILE}" } # [NOTE] # If it mounted with relatime or noatime options , the "touch -a" # command may not update the atime. # In ubuntu:xenial, atime was updated even if relatime was granted. # However, it was not updated in bionic/focal. # We can probably update atime by explicitly specifying the strictatime # option and running the "touch -a" command. However, the strictatime # option cannot be set. # Therefore, if the relatime option is set, the test with the "touch -a" # command is bypassed. # We do not know why atime is not updated may or not be affected by # these options.(can't say for sure) # However, if atime has not been updated, the s3fs_utimens entry point # will not be called from FUSE library. We added this bypass because # the test became unstable. # function test_update_time_chmod() { describe "Testing update time function chmod..." local t0=1000000000 # 9 September 2001 local OBJECT_NAME; OBJECT_NAME=$(basename "${PWD}")/"${TEST_TEXT_FILE}" echo data | aws_cli s3 cp --metadata="atime=${t0},ctime=${t0},mtime=${t0}" - "s3://${TEST_BUCKET_1}/${OBJECT_NAME}" local base_atime; base_atime=$(get_atime "${TEST_TEXT_FILE}") local base_ctime; base_ctime=$(get_ctime "${TEST_TEXT_FILE}") local base_mtime; base_mtime=$(get_mtime "${TEST_TEXT_FILE}") # # chmod -> update only ctime # chmod +x "${TEST_TEXT_FILE}" local atime; atime=$(get_atime "${TEST_TEXT_FILE}") local ctime; ctime=$(get_ctime "${TEST_TEXT_FILE}") local mtime; mtime=$(get_mtime "${TEST_TEXT_FILE}") if [ "${base_atime}" != "${atime}" ] || [ "${base_ctime}" = "${ctime}" ] || [ "${base_mtime}" != "${mtime}" ]; then echo "chmod expected updated ctime: $base_ctime != $ctime and same mtime: $base_mtime == $mtime, atime: $base_atime == $atime" return 1 fi rm_test_file } function test_update_time_chown() { describe "Testing update time function chown..." # # chown -> update only ctime # local t0=1000000000 # 9 September 2001 local OBJECT_NAME; OBJECT_NAME=$(basename "${PWD}")/"${TEST_TEXT_FILE}" echo data | aws_cli s3 cp --metadata="atime=${t0},ctime=${t0},mtime=${t0}" - "s3://${TEST_BUCKET_1}/${OBJECT_NAME}" local base_atime; base_atime=$(get_atime "${TEST_TEXT_FILE}") local base_ctime; base_ctime=$(get_ctime "${TEST_TEXT_FILE}") local base_mtime; base_mtime=$(get_mtime "${TEST_TEXT_FILE}") chown $UID "${TEST_TEXT_FILE}" local atime; atime=$(get_atime "${TEST_TEXT_FILE}") local ctime; ctime=$(get_ctime "${TEST_TEXT_FILE}") local mtime; mtime=$(get_mtime "${TEST_TEXT_FILE}") if [ "${base_atime}" != "${atime}" ] || [ "${base_ctime}" = "${ctime}" ] || [ "${base_mtime}" != "${mtime}" ]; then echo "chown expected updated ctime: $base_ctime != $ctime and same mtime: $base_mtime == $mtime, atime: $base_atime == $atime" return 1 fi rm_test_file } function test_update_time_xattr() { describe "Testing update time function set_xattr..." local t0=1000000000 # 9 September 2001 local OBJECT_NAME; OBJECT_NAME=$(basename "${PWD}")/"${TEST_TEXT_FILE}" echo data | aws_cli s3 cp --metadata="atime=${t0},ctime=${t0},mtime=${t0}" - "s3://${TEST_BUCKET_1}/${OBJECT_NAME}" local base_atime; base_atime=$(get_atime "${TEST_TEXT_FILE}") local base_ctime; base_ctime=$(get_ctime "${TEST_TEXT_FILE}") local base_mtime; base_mtime=$(get_mtime "${TEST_TEXT_FILE}") # # set_xattr -> update only ctime # set_xattr key value "${TEST_TEXT_FILE}" local atime; atime=$(get_atime "${TEST_TEXT_FILE}") local ctime; ctime=$(get_ctime "${TEST_TEXT_FILE}") local mtime; mtime=$(get_mtime "${TEST_TEXT_FILE}") if [ "${base_atime}" != "${atime}" ] || [ "${base_ctime}" = "${ctime}" ] || [ "${base_mtime}" != "${mtime}" ]; then echo "set_xattr expected updated ctime: $base_ctime != $ctime and same mtime: $base_mtime == $mtime, atime: $base_atime == $atime" return 1 fi rm_test_file } function test_update_time_touch() { describe "Testing update time function touch..." local t0=1000000000 # 9 September 2001 local OBJECT_NAME; OBJECT_NAME=$(basename "${PWD}")/"${TEST_TEXT_FILE}" echo data | aws_cli s3 cp --metadata="atime=${t0},ctime=${t0},mtime=${t0}" - "s3://${TEST_BUCKET_1}/${OBJECT_NAME}" sleep 1 local base_atime; base_atime=$(get_atime "${TEST_TEXT_FILE}") local base_ctime; base_ctime=$(get_ctime "${TEST_TEXT_FILE}") local base_mtime; base_mtime=$(get_mtime "${TEST_TEXT_FILE}") # # touch -> update ctime/atime/mtime # touch "${TEST_TEXT_FILE}" local atime; atime=$(get_atime "${TEST_TEXT_FILE}") local ctime; ctime=$(get_ctime "${TEST_TEXT_FILE}") local mtime; mtime=$(get_mtime "${TEST_TEXT_FILE}") if [ "${base_atime}" = "${atime}" ] || [ "${base_ctime}" = "${ctime}" ] || [ "${base_mtime}" = "${mtime}" ]; then echo "touch expected updated ctime: $base_ctime != $ctime, mtime: $base_mtime != $mtime, atime: $base_atime != $atime" return 1 fi rm_test_file } function test_update_time_touch_a() { describe "Testing update time function touch -a..." local t0=1000000000 # 9 September 2001 local OBJECT_NAME; OBJECT_NAME=$(basename "${PWD}")/"${TEST_TEXT_FILE}" echo data | aws_cli s3 cp --metadata="atime=${t0},ctime=${t0},mtime=${t0}" - "s3://${TEST_BUCKET_1}/${OBJECT_NAME}" local base_atime; base_atime=$(get_atime "${TEST_TEXT_FILE}") local base_ctime; base_ctime=$(get_ctime "${TEST_TEXT_FILE}") local base_mtime; base_mtime=$(get_mtime "${TEST_TEXT_FILE}") # # "touch -a" -> update ctime/atime, not update mtime # touch -a "${TEST_TEXT_FILE}" local atime; atime=$(get_atime "${TEST_TEXT_FILE}") local ctime; ctime=$(get_ctime "${TEST_TEXT_FILE}") local mtime; mtime=$(get_mtime "${TEST_TEXT_FILE}") if [ "${base_atime}" = "${atime}" ] || [ "${base_ctime}" = "${ctime}" ] || [ "${base_mtime}" != "${mtime}" ]; then echo "touch with -a option expected updated ctime: $base_ctime != $ctime, atime: $base_atime != $atime and same mtime: $base_mtime == $mtime" return 1 fi rm_test_file } function test_update_time_append() { describe "Testing update time function append..." local t0=1000000000 # 9 September 2001 local OBJECT_NAME; OBJECT_NAME=$(basename "${PWD}")/"${TEST_TEXT_FILE}" echo data | aws_cli s3 cp --metadata="atime=${t0},ctime=${t0},mtime=${t0}" - "s3://${TEST_BUCKET_1}/${OBJECT_NAME}" local base_atime; base_atime=$(get_atime "${TEST_TEXT_FILE}") local base_ctime; base_ctime=$(get_ctime "${TEST_TEXT_FILE}") local base_mtime; base_mtime=$(get_mtime "${TEST_TEXT_FILE}") # # append -> update ctime/mtime, not update atime # echo foo >> "${TEST_TEXT_FILE}" local atime; atime=$(get_atime "${TEST_TEXT_FILE}") local ctime; ctime=$(get_ctime "${TEST_TEXT_FILE}") local mtime; mtime=$(get_mtime "${TEST_TEXT_FILE}") if [ "${base_atime}" != "${atime}" ] || [ "${base_ctime}" = "${ctime}" ] || [ "${base_mtime}" = "${mtime}" ]; then echo "append expected updated ctime: $base_ctime != $ctime, mtime: $base_mtime != $mtime and same atime: $base_atime == $atime" return 1 fi rm_test_file } function test_update_time_cp_p() { describe "Testing update time function cp -p..." local t0=1000000000 # 9 September 2001 local OBJECT_NAME; OBJECT_NAME=$(basename "${PWD}")/"${TEST_TEXT_FILE}" echo data | aws_cli s3 cp --metadata="atime=${t0},ctime=${t0},mtime=${t0}" - "s3://${TEST_BUCKET_1}/${OBJECT_NAME}" local base_atime; base_atime=$(get_atime "${TEST_TEXT_FILE}") local base_ctime; base_ctime=$(get_ctime "${TEST_TEXT_FILE}") local base_mtime; base_mtime=$(get_mtime "${TEST_TEXT_FILE}") # # cp -p -> update ctime, not update atime/mtime # local TIME_TEST_TEXT_FILE=test-s3fs-time.txt cp -p "${TEST_TEXT_FILE}" "${TIME_TEST_TEXT_FILE}" local atime; atime=$(get_atime "${TIME_TEST_TEXT_FILE}") local ctime; ctime=$(get_ctime "${TIME_TEST_TEXT_FILE}") local mtime; mtime=$(get_mtime "${TIME_TEST_TEXT_FILE}") if [ "${base_atime}" != "${atime}" ] || [ "${base_ctime}" = "${ctime}" ] || [ "${base_mtime}" != "${mtime}" ]; then echo "cp with -p option expected updated ctime: $base_ctime != $ctime and same mtime: $base_mtime == $mtime, atime: $base_atime == $atime" return 1 fi } function test_update_time_mv() { describe "Testing update time function mv..." local t0=1000000000 # 9 September 2001 local OBJECT_NAME; OBJECT_NAME=$(basename "${PWD}")/"${TEST_TEXT_FILE}" echo data | aws_cli s3 cp --metadata="atime=${t0},ctime=${t0},mtime=${t0}" - "s3://${TEST_BUCKET_1}/${OBJECT_NAME}" local base_atime; base_atime=$(get_atime "${TEST_TEXT_FILE}") local base_ctime; base_ctime=$(get_ctime "${TEST_TEXT_FILE}") local base_mtime; base_mtime=$(get_mtime "${TEST_TEXT_FILE}") # # mv -> update ctime, not update atime/mtime # local TIME2_TEST_TEXT_FILE=test-s3fs-time2.txt mv "${TEST_TEXT_FILE}" "${TIME2_TEST_TEXT_FILE}" local atime; atime=$(get_atime "${TIME2_TEST_TEXT_FILE}") local ctime; ctime=$(get_ctime "${TIME2_TEST_TEXT_FILE}") local mtime; mtime=$(get_mtime "${TIME2_TEST_TEXT_FILE}") if [ "${base_atime}" != "${atime}" ] || [ "${base_ctime}" = "${ctime}" ] || [ "${base_mtime}" != "${mtime}" ]; then echo "mv expected updated ctime: $base_ctime != $ctime and same mtime: $base_mtime == $mtime, atime: $base_atime == $atime" return 1 fi rm_test_file "${TIME_TEST_TEXT_FILE}" rm_test_file "${TIME2_TEST_TEXT_FILE}" } # [NOTE] # See the description of test_update_time () for notes about the # "touch -a" command and atime. # function test_update_directory_time_chmod() { describe "Testing update time for directory mv..." # # create the directory and sub-directory and a file in directory # local t0=1000000000 # 9 September 2001 local DIRECTORY_NAME; DIRECTORY_NAME=$(basename "${PWD}")/"${TEST_DIR}" aws_cli s3api put-object --content-type="application/x-directory" --metadata="atime=${t0},ctime=${t0},mtime=${t0}" --bucket "${TEST_BUCKET_1}" --key "$DIRECTORY_NAME/" local base_atime; base_atime=$(get_atime "${TEST_DIR}") local base_ctime; base_ctime=$(get_ctime "${TEST_DIR}") local base_mtime; base_mtime=$(get_mtime "${TEST_DIR}") # # chmod -> update only ctime # chmod 0777 "${TEST_DIR}" local atime; atime=$(get_atime "${TEST_DIR}") local ctime; ctime=$(get_ctime "${TEST_DIR}") local mtime; mtime=$(get_mtime "${TEST_DIR}") if [ "${base_atime}" != "${atime}" ] || [ "${base_ctime}" = "${ctime}" ] || [ "${base_mtime}" != "${mtime}" ]; then echo "chmod expected updated ctime: $base_ctime != $ctime and same mtime: $base_mtime == $mtime, atime: $base_atime == $atime" return 1 fi rm -rf "${TEST_DIR}" } function test_update_directory_time_chown { describe "Testing update time for directory chown..." local t0=1000000000 # 9 September 2001 local DIRECTORY_NAME; DIRECTORY_NAME=$(basename "${PWD}")/"${TEST_DIR}" aws_cli s3api put-object --content-type="application/x-directory" --metadata="atime=${t0},ctime=${t0},mtime=${t0}" --bucket "${TEST_BUCKET_1}" --key "$DIRECTORY_NAME/" local base_atime; base_atime=$(get_atime "${TEST_DIR}") local base_ctime; base_ctime=$(get_ctime "${TEST_DIR}") local base_mtime; base_mtime=$(get_mtime "${TEST_DIR}") # # chown -> update only ctime # chown $UID "${TEST_DIR}" local atime; atime=$(get_atime "${TEST_DIR}") local ctime; ctime=$(get_ctime "${TEST_DIR}") local mtime; mtime=$(get_mtime "${TEST_DIR}") if [ "${base_atime}" != "${atime}" ] || [ "${base_ctime}" = "${ctime}" ] || [ "${base_mtime}" != "${mtime}" ]; then echo "chown expected updated ctime: $base_ctime != $ctime and same mtime: $base_mtime == $mtime, atime: $base_atime == $atime" return 1 fi rm -rf "${TEST_DIR}" } function test_update_directory_time_set_xattr { describe "Testing update time for directory set_xattr..." local t0=1000000000 # 9 September 2001 local DIRECTORY_NAME; DIRECTORY_NAME=$(basename "${PWD}")/"${TEST_DIR}" aws_cli s3api put-object --content-type="application/x-directory" --metadata="atime=${t0},ctime=${t0},mtime=${t0}" --bucket "${TEST_BUCKET_1}" --key "$DIRECTORY_NAME/" local base_atime; base_atime=$(get_atime "${TEST_DIR}") local base_ctime; base_ctime=$(get_ctime "${TEST_DIR}") local base_mtime; base_mtime=$(get_mtime "${TEST_DIR}") # # set_xattr -> update only ctime # set_xattr key value "${TEST_DIR}" local atime; atime=$(get_atime "${TEST_DIR}") local ctime; ctime=$(get_ctime "${TEST_DIR}") local mtime; mtime=$(get_mtime "${TEST_DIR}") if [ "${base_atime}" != "${atime}" ] || [ "${base_ctime}" = "${ctime}" ] || [ "${base_mtime}" != "${mtime}" ]; then echo "set_xattr expected updated ctime: $base_ctime != $ctime and same mtime: $base_mtime == $mtime, atime: $base_atime == $atime" return 1 fi rm -rf "${TEST_DIR}" } function test_update_directory_time_touch { describe "Testing update time for directory touch..." local t0=1000000000 # 9 September 2001 local DIRECTORY_NAME; DIRECTORY_NAME=$(basename "${PWD}")/"${TEST_DIR}" aws_cli s3api put-object --content-type="application/x-directory" --metadata="atime=${t0},ctime=${t0},mtime=${t0}" --bucket "${TEST_BUCKET_1}" --key "$DIRECTORY_NAME/" local base_atime; base_atime=$(get_atime "${TEST_DIR}") local base_ctime; base_ctime=$(get_ctime "${TEST_DIR}") local base_mtime; base_mtime=$(get_mtime "${TEST_DIR}") # # touch -> update ctime/atime/mtime # touch "${TEST_DIR}" local atime; atime=$(get_atime "${TEST_DIR}") local ctime; ctime=$(get_ctime "${TEST_DIR}") local mtime; mtime=$(get_mtime "${TEST_DIR}") if [ "${base_atime}" = "${atime}" ] || [ "${base_ctime}" = "${ctime}" ] || [ "${base_mtime}" = "${mtime}" ]; then echo "touch expected updated ctime: $base_ctime != $ctime, mtime: $base_mtime != $mtime, atime: $base_atime != $atime" return 1 fi rm -rf "${TEST_DIR}" } function test_update_directory_time_touch_a { describe "Testing update time for directory touch -a..." local t0=1000000000 # 9 September 2001 local DIRECTORY_NAME; DIRECTORY_NAME=$(basename "${PWD}")/"${TEST_DIR}" aws_cli s3api put-object --content-type="application/x-directory" --metadata="atime=${t0},ctime=${t0},mtime=${t0}" --bucket "${TEST_BUCKET_1}" --key "$DIRECTORY_NAME/" local base_atime; base_atime=$(get_atime "${TEST_DIR}") local base_ctime; base_ctime=$(get_ctime "${TEST_DIR}") local base_mtime; base_mtime=$(get_mtime "${TEST_DIR}") # # "touch -a" -> update ctime/atime, not update mtime # touch -a "${TEST_DIR}" local atime; atime=$(get_atime "${TEST_DIR}") local ctime; ctime=$(get_ctime "${TEST_DIR}") local mtime; mtime=$(get_mtime "${TEST_DIR}") if [ "${base_atime}" = "${atime}" ] || [ "${base_ctime}" = "${ctime}" ] || [ "${base_mtime}" != "${mtime}" ]; then echo "touch with -a option expected updated ctime: $base_ctime != $ctime, atime: $base_atime != $atime and same mtime: $base_mtime == $mtime" return 1 fi rm -rf "${TEST_DIR}" } function test_update_directory_time_subdir() { describe "Testing update time for directory subdirectory..." local TIME_TEST_SUBDIR="${TEST_DIR}/testsubdir" local TIME_TEST_FILE_INDIR="${TEST_DIR}/testfile" mk_test_dir mkdir "${TIME_TEST_SUBDIR}" touch "${TIME_TEST_FILE_INDIR}" # TODO: remove sleep after improving AWS CLI speed sleep 1 local base_atime; base_atime=$(get_atime "${TEST_DIR}") local base_ctime; base_ctime=$(get_ctime "${TEST_DIR}") local base_mtime; base_mtime=$(get_mtime "${TEST_DIR}") local subdir_atime; subdir_atime=$(get_atime "${TIME_TEST_SUBDIR}") local subdir_ctime; subdir_ctime=$(get_ctime "${TIME_TEST_SUBDIR}") local subdir_mtime; subdir_mtime=$(get_mtime "${TIME_TEST_SUBDIR}") local subfile_atime; subfile_atime=$(get_atime "${TIME_TEST_FILE_INDIR}") local subfile_ctime; subfile_ctime=$(get_ctime "${TIME_TEST_FILE_INDIR}") local subfile_mtime; subfile_mtime=$(get_mtime "${TIME_TEST_FILE_INDIR}") # # mv -> update ctime, not update atime/mtime for target directory # not update any for sub-directory and a file # local TIME_TEST_DIR=timetestdir local TIME2_TEST_SUBDIR="${TIME_TEST_DIR}/testsubdir" local TIME2_TEST_FILE_INDIR="${TIME_TEST_DIR}/testfile" mv "${TEST_DIR}" "${TIME_TEST_DIR}" local atime; atime=$(get_atime "${TIME_TEST_DIR}") local ctime; ctime=$(get_ctime "${TIME_TEST_DIR}") local mtime; mtime=$(get_mtime "${TIME_TEST_DIR}") if [ "${base_atime}" != "${atime}" ] || [ "${base_ctime}" = "${ctime}" ] || [ "${base_mtime}" != "${mtime}" ]; then echo "mv expected updated ctime: $base_ctime != $ctime and same mtime: $base_mtime == $mtime, atime: $base_atime == $atime" return 1 fi atime=$(get_atime "${TIME2_TEST_SUBDIR}") ctime=$(get_ctime "${TIME2_TEST_SUBDIR}") mtime=$(get_mtime "${TIME2_TEST_SUBDIR}") if [ "${subdir_atime}" != "${atime}" ] || [ "${subdir_ctime}" != "${ctime}" ] || [ "${subdir_mtime}" != "${mtime}" ]; then echo "mv for sub-directory expected same ctime: $subdir_ctime == $ctime, mtime: $subdir_mtime == $mtime, atime: $subdir_atime == $atime" return 1 fi atime=$(get_atime "${TIME2_TEST_FILE_INDIR}") ctime=$(get_ctime "${TIME2_TEST_FILE_INDIR}") mtime=$(get_mtime "${TIME2_TEST_FILE_INDIR}") if [ "${subfile_atime}" != "${atime}" ] || [ "${subfile_ctime}" != "${ctime}" ] || [ "${subfile_mtime}" != "${mtime}" ]; then echo "mv for a file in directory expected same ctime: $subfile_ctime == $ctime, mtime: $subfile_mtime == $mtime, atime: $subfile_atime == $atime" return 1 fi rm -rf "${TIME_TEST_SUBDIR}" rm -rf "${TIME_TEST_DIR}" rm -rf "${TEST_DIR}" } # [NOTE] # This test changes the file mode while creating/editing a new file, # and finally closes it. # Test with the sed command as it occurs when in place mode of the sed # command. (If trying it with a standard C function(and shell script), # it will be not the same result of sed, so sed is used.) # function test_update_chmod_opened_file() { describe "Testing create, modify the file by sed in place mode" # test file local BEFORE_STRING_DATA; BEFORE_STRING_DATA="sed in place test : BEFORE DATA" local AFTER_STRING_DATA; AFTER_STRING_DATA="sed in place test : AFTER DATA" echo "${BEFORE_STRING_DATA}" > "${TEST_TEXT_FILE}" # sed in place sed -i -e 's/BEFORE DATA/AFTER DATA/g' "${TEST_TEXT_FILE}" # compare result local RESULT_STRING; RESULT_STRING=$(cat "${TEST_TEXT_FILE}") if [ -z "${RESULT_STRING}" ] || [ "${RESULT_STRING}" != "${AFTER_STRING_DATA}" ]; then echo "the file conversion by sed in place command failed." return 1 fi # clean up rm_test_file "${ALT_TEST_TEXT_FILE}" } function test_update_parent_directory_time_sub() { if [ $# -ne 1 ]; then echo "Internal error: parameter is wrong." return 1 fi # [NOTE] # Skip test for mknod/mkfifo command. # If run them, ctime/mtime of the parent directory will be updated. # local TEST_PARENTDIR_PARENT="${1}" local TEST_PARENTDIR_FILE="${TEST_PARENTDIR_PARENT}/testfile" local TEST_PARENTDIR_SYMFILE_BASE="testfile2" local TEST_PARENTDIR_FILE_MV="${TEST_PARENTDIR_PARENT}/${TEST_PARENTDIR_SYMFILE_BASE}" local TEST_PARENTDIR_SYMFILE="${TEST_PARENTDIR_PARENT}/symfile" local TEST_PARENTDIR_SYMFILE_MV="${TEST_PARENTDIR_PARENT}/symfile2" local TEST_PARENTDIR_DIR="${TEST_PARENTDIR_PARENT}/testdir" local TEST_PARENTDIR_DIR_MV="${TEST_PARENTDIR_PARENT}/testdir2" # # Create file -> Update parent directory's mtime/ctime # local base_atime; base_atime=$(get_atime "${TEST_PARENTDIR_PARENT}") local base_ctime; base_ctime=$(get_ctime "${TEST_PARENTDIR_PARENT}") local base_mtime; base_mtime=$(get_mtime "${TEST_PARENTDIR_PARENT}") touch "${TEST_PARENTDIR_FILE}" local after_atime; after_atime=$(get_atime "${TEST_PARENTDIR_PARENT}") local after_ctime; after_ctime=$(get_ctime "${TEST_PARENTDIR_PARENT}") local after_mtime; after_mtime=$(get_mtime "${TEST_PARENTDIR_PARENT}") if [ "${base_atime}" != "${after_atime}" ] || [ "${base_ctime}" = "${after_ctime}" ] || [ "${base_mtime}" = "${after_mtime}" ]; then echo "creating file expected updating ctime/mtime: atime( ${base_atime} == ${after_atime} ), ctime( ${base_ctime} != ${after_ctime} ), mtime( ${base_mtime} != ${after_mtime} )" return 1 fi # # Update file -> Not update parent directory's atime/mtime/ctime # base_atime="${after_atime}" base_ctime="${after_ctime}" base_mtime="${after_mtime}" touch "${TEST_PARENTDIR_FILE}" local after_atime; after_atime=$(get_atime "${TEST_PARENTDIR_PARENT}") local after_ctime; after_ctime=$(get_ctime "${TEST_PARENTDIR_PARENT}") local after_mtime; after_mtime=$(get_mtime "${TEST_PARENTDIR_PARENT}") if [ "${base_atime}" != "${after_atime}" ] || [ "${base_ctime}" != "${after_ctime}" ] || [ "${base_mtime}" != "${after_mtime}" ]; then echo "updating file expected updating ctime/mtime: atime( ${base_atime} == ${after_atime} ), ctime( ${base_ctime} == ${after_ctime} ), mtime( ${base_mtime} == ${after_mtime} )" return 1 fi # # Rename file -> Update parent directory's mtime/ctime # base_atime="${after_atime}" base_ctime="${after_ctime}" base_mtime="${after_mtime}" mv "${TEST_PARENTDIR_FILE}" "${TEST_PARENTDIR_FILE_MV}" local after_atime; after_atime=$(get_atime "${TEST_PARENTDIR_PARENT}") local after_ctime; after_ctime=$(get_ctime "${TEST_PARENTDIR_PARENT}") local after_mtime; after_mtime=$(get_mtime "${TEST_PARENTDIR_PARENT}") if [ "${base_atime}" != "${after_atime}" ] || [ "${base_ctime}" = "${after_ctime}" ] || [ "${base_mtime}" = "${after_mtime}" ]; then echo "renaming file expected updating ctime/mtime: atime( ${base_atime} == ${after_atime} ), ctime( ${base_ctime} != ${after_ctime} ), mtime( ${base_mtime} != ${after_mtime} )" return 1 fi # # Create symbolic link -> Update parent directory's mtime/ctime # base_atime="${after_atime}" base_ctime="${after_ctime}" base_mtime="${after_mtime}" ln -s "${TEST_PARENTDIR_SYMFILE_BASE}" "${TEST_PARENTDIR_SYMFILE}" local after_atime; after_atime=$(get_atime "${TEST_PARENTDIR_PARENT}") local after_ctime; after_ctime=$(get_ctime "${TEST_PARENTDIR_PARENT}") local after_mtime; after_mtime=$(get_mtime "${TEST_PARENTDIR_PARENT}") if [ "${base_atime}" != "${after_atime}" ] || [ "${base_ctime}" = "${after_ctime}" ] || [ "${base_mtime}" = "${after_mtime}" ]; then echo "creating symbolic file expected updating ctime/mtime: atime( ${base_atime} == ${after_atime} ), ctime( ${base_ctime} != ${after_ctime} ), mtime( ${base_mtime} != ${after_mtime} )" return 1 fi # # Update symbolic file -> Not update parent directory's atime/mtime/ctime # base_atime="${after_atime}" base_ctime="${after_ctime}" base_mtime="${after_mtime}" touch "${TEST_PARENTDIR_SYMFILE}" local after_atime; after_atime=$(get_atime "${TEST_PARENTDIR_PARENT}") local after_ctime; after_ctime=$(get_ctime "${TEST_PARENTDIR_PARENT}") local after_mtime; after_mtime=$(get_mtime "${TEST_PARENTDIR_PARENT}") if [ "${base_atime}" != "${after_atime}" ] || [ "${base_ctime}" != "${after_ctime}" ] || [ "${base_mtime}" != "${after_mtime}" ]; then echo "updating symbolic file expected updating ctime/mtime: atime( ${base_atime} == ${after_atime} ), ctime( ${base_ctime} == ${after_ctime} ), mtime( ${base_mtime} == ${after_mtime} )" return 1 fi # # Rename symbolic link -> Update parent directory's mtime/ctime # base_atime="${after_atime}" base_ctime="${after_ctime}" base_mtime="${after_mtime}" mv "${TEST_PARENTDIR_SYMFILE}" "${TEST_PARENTDIR_SYMFILE_MV}" local after_atime; after_atime=$(get_atime "${TEST_PARENTDIR_PARENT}") local after_ctime; after_ctime=$(get_ctime "${TEST_PARENTDIR_PARENT}") local after_mtime; after_mtime=$(get_mtime "${TEST_PARENTDIR_PARENT}") if [ "${base_atime}" != "${after_atime}" ] || [ "${base_ctime}" = "${after_ctime}" ] || [ "${base_mtime}" = "${after_mtime}" ]; then echo "renaming symbolic file expected updating ctime/mtime: atime( ${base_atime} == ${after_atime} ), ctime( ${base_ctime} != ${after_ctime} ), mtime( ${base_mtime} != ${after_mtime} )" return 1 fi # # Delete symbolic link -> Update parent directory's mtime/ctime # base_atime="${after_atime}" base_ctime="${after_ctime}" base_mtime="${after_mtime}" rm "${TEST_PARENTDIR_SYMFILE_MV}" local after_atime; after_atime=$(get_atime "${TEST_PARENTDIR_PARENT}") local after_ctime; after_ctime=$(get_ctime "${TEST_PARENTDIR_PARENT}") local after_mtime; after_mtime=$(get_mtime "${TEST_PARENTDIR_PARENT}") if [ "${base_atime}" != "${after_atime}" ] || [ "${base_ctime}" = "${after_ctime}" ] || [ "${base_mtime}" = "${after_mtime}" ]; then echo "deleting symbolic file expected updating ctime/mtime: atime( ${base_atime} == ${after_atime} ), ctime( ${base_ctime} != ${after_ctime} ), mtime( ${base_mtime} != ${after_mtime} )" return 1 fi # # Delete file -> Update parent directory's mtime/ctime # base_atime="${after_atime}" base_ctime="${after_ctime}" base_mtime="${after_mtime}" rm "${TEST_PARENTDIR_FILE_MV}" local after_atime; after_atime=$(get_atime "${TEST_PARENTDIR_PARENT}") local after_ctime; after_ctime=$(get_ctime "${TEST_PARENTDIR_PARENT}") local after_mtime; after_mtime=$(get_mtime "${TEST_PARENTDIR_PARENT}") if [ "${base_atime}" != "${after_atime}" ] || [ "${base_ctime}" = "${after_ctime}" ] || [ "${base_mtime}" = "${after_mtime}" ]; then echo "deleting file expected updating ctime/mtime: atime( ${base_atime} == ${after_atime} ), ctime( ${base_ctime} != ${after_ctime} ), mtime( ${base_mtime} != ${after_mtime} )" return 1 fi # # Create directory -> Update parent directory's mtime/ctime # local base_atime; base_atime=$(get_atime "${TEST_PARENTDIR_PARENT}") local base_ctime; base_ctime=$(get_ctime "${TEST_PARENTDIR_PARENT}") local base_mtime; base_mtime=$(get_mtime "${TEST_PARENTDIR_PARENT}") mkdir "${TEST_PARENTDIR_DIR}" local after_atime; after_atime=$(get_atime "${TEST_PARENTDIR_PARENT}") local after_ctime; after_ctime=$(get_ctime "${TEST_PARENTDIR_PARENT}") local after_mtime; after_mtime=$(get_mtime "${TEST_PARENTDIR_PARENT}") if [ "${base_atime}" != "${after_atime}" ] || [ "${base_ctime}" = "${after_ctime}" ] || [ "${base_mtime}" = "${after_mtime}" ]; then echo "creating directory expected updating ctime/mtime: atime( ${base_atime} == ${after_atime} ), ctime( ${base_ctime} != ${after_ctime} ), mtime( ${base_mtime} != ${after_mtime} )" return 1 fi # # Update directory -> Not update parent directory's atime/mtime/ctime # base_atime="${after_atime}" base_ctime="${after_ctime}" base_mtime="${after_mtime}" touch "${TEST_PARENTDIR_DIR}" local after_atime; after_atime=$(get_atime "${TEST_PARENTDIR_PARENT}") local after_ctime; after_ctime=$(get_ctime "${TEST_PARENTDIR_PARENT}") local after_mtime; after_mtime=$(get_mtime "${TEST_PARENTDIR_PARENT}") if [ "${base_atime}" != "${after_atime}" ] || [ "${base_ctime}" != "${after_ctime}" ] || [ "${base_mtime}" != "${after_mtime}" ]; then echo "updating directory expected updating ctime/mtime: atime( ${base_atime} == ${after_atime} ), ctime( ${base_ctime} == ${after_ctime} ), mtime( ${base_mtime} == ${after_mtime} )" return 1 fi # # Rename directory -> Update parent directory's mtime/ctime # base_atime="${after_atime}" base_ctime="${after_ctime}" base_mtime="${after_mtime}" mv "${TEST_PARENTDIR_DIR}" "${TEST_PARENTDIR_DIR_MV}" local after_atime; after_atime=$(get_atime "${TEST_PARENTDIR_PARENT}") local after_ctime; after_ctime=$(get_ctime "${TEST_PARENTDIR_PARENT}") local after_mtime; after_mtime=$(get_mtime "${TEST_PARENTDIR_PARENT}") if [ "${base_atime}" != "${after_atime}" ] || [ "${base_ctime}" = "${after_ctime}" ] || [ "${base_mtime}" = "${after_mtime}" ]; then echo "renaming directory expected updating ctime/mtime: atime( ${base_atime} == ${after_atime} ), ctime( ${base_ctime} != ${after_ctime} ), mtime( ${base_mtime} != ${after_mtime} )" return 1 fi # # Delete directory -> Update parent directory's mtime/ctime # base_atime="${after_atime}" base_ctime="${after_ctime}" base_mtime="${after_mtime}" rm -r "${TEST_PARENTDIR_DIR_MV}" local after_atime; after_atime=$(get_atime "${TEST_PARENTDIR_PARENT}") local after_ctime; after_ctime=$(get_ctime "${TEST_PARENTDIR_PARENT}") local after_mtime; after_mtime=$(get_mtime "${TEST_PARENTDIR_PARENT}") if [ "${base_atime}" != "${after_atime}" ] || [ "${base_ctime}" = "${after_ctime}" ] || [ "${base_mtime}" = "${after_mtime}" ]; then echo "deleting directory expected updating ctime/mtime: atime( ${base_atime} == ${after_atime} ), ctime( ${base_ctime} != ${after_ctime} ), mtime( ${base_mtime} != ${after_mtime} )" return 1 fi return 0 } function test_update_parent_directory_time() { describe "Testing update time of parent directory..." # # Test sub directory # mk_test_dir if ! test_update_parent_directory_time_sub "${TEST_DIR}"; then echo "failed test about updating time of parent directory: ${TEST_DIR}" return 1 fi rm -rf "${TEST_DIR}" # # Test bucket top directory # # [NOTE] # The current directory for test execution is "/testrun-xxxx". # This test checks in the directory at the top of the bucket. # if ! test_update_parent_directory_time_sub ".."; then echo "failed test about updating time of parent directory: ${TEST_DIR}" return 1 fi return 0 } function test_rm_rf_dir { describe "Test that rm -rf will remove directory with contents ..." # Create a dir with some files and directories mkdir dir1 mkdir dir1/dir2 touch dir1/file1 touch dir1/dir2/file2 # Remove the dir with recursive rm rm -rf dir1 if [ -e dir1 ]; then echo "rm -rf did not remove $PWD/dir1" return 1 fi } function test_posix_acl { describe "Testing posix acl function ..." #------------------------------------------------------ # Directory #------------------------------------------------------ local POSIX_ACL_TEST_DIR1="posix_acl_dir1" local POSIX_ACL_TEST_DIR2="posix_acl_dir2" rm -rf "${POSIX_ACL_TEST_DIR1}" rm -rf "${POSIX_ACL_TEST_DIR2}" mkdir "${POSIX_ACL_TEST_DIR1}" # [NOTE] # Let the target username be "root". # On some systems the USER environment variable may be # empty, so using ${USER} will not give an accurate test. # TARGET_ACL_USER="root" # # Set posix acl(not default) # setfacl -m "u:${TARGET_ACL_USER}:rwx" "${POSIX_ACL_TEST_DIR1}" if ! getfacl "${POSIX_ACL_TEST_DIR1}" | grep -q "^user:${TARGET_ACL_USER}:rwx"; then echo "Could not set posix acl(not default) to ${POSIX_ACL_TEST_DIR1} directory" return 1 fi # # Set posix acl(default) # setfacl -d -m "u:${TARGET_ACL_USER}:rwx" "${POSIX_ACL_TEST_DIR1}" if ! getfacl "${POSIX_ACL_TEST_DIR1}" | grep -q "^default:user:${TARGET_ACL_USER}:rwx"; then echo "Could not set posix acl(default) to ${POSIX_ACL_TEST_DIR1} directory" return 1 fi # # Rename # mv "${POSIX_ACL_TEST_DIR1}" "${POSIX_ACL_TEST_DIR2}" if ! getfacl "${POSIX_ACL_TEST_DIR2}" | grep -q "^user:${TARGET_ACL_USER}:rwx"; then echo "Could not move with posix acl(not default) to ${POSIX_ACL_TEST_DIR2} directory" return 1 fi if ! getfacl "${POSIX_ACL_TEST_DIR2}" | grep -q "^default:user:${TARGET_ACL_USER}:rwx"; then echo "Could not move with posix acl(default) to ${POSIX_ACL_TEST_DIR2} directory" return 1 fi # # Copy with permission # cp -rp "${POSIX_ACL_TEST_DIR2}" "${POSIX_ACL_TEST_DIR1}" if ! getfacl "${POSIX_ACL_TEST_DIR1}" | grep -q "^user:${TARGET_ACL_USER}:rwx"; then echo "Could not copy with posix acl(not default) to ${POSIX_ACL_TEST_DIR1} directory" return 1 fi if ! getfacl "${POSIX_ACL_TEST_DIR1}" | grep -q "^default:user:${TARGET_ACL_USER}:rwx"; then echo "Could not copy with posix acl(default) to ${POSIX_ACL_TEST_DIR1} directory" return 1 fi # # Overwrite posix acl(not default) # setfacl -m "u:${TARGET_ACL_USER}:r-x" "${POSIX_ACL_TEST_DIR2}" if ! getfacl "${POSIX_ACL_TEST_DIR2}" | grep -q "^user:${TARGET_ACL_USER}:r-x"; then echo "Could not set posix acl(not default) to ${POSIX_ACL_TEST_DIR2} directory" return 1 fi # # Overwrite posix acl(default) # setfacl -d -m "u:${TARGET_ACL_USER}:r-x" "${POSIX_ACL_TEST_DIR2}" if ! getfacl "${POSIX_ACL_TEST_DIR2}" | grep -q "^default:user:${TARGET_ACL_USER}:r-x"; then echo "Could not set posix acl(default) to ${POSIX_ACL_TEST_DIR2} directory" return 1 fi # # Remove posix acl(default) # setfacl -k "${POSIX_ACL_TEST_DIR2}" if getfacl "${POSIX_ACL_TEST_DIR2}" | grep -q "^default:user:${TARGET_ACL_USER}"; then echo "Could not remove posix acl(default) to ${POSIX_ACL_TEST_DIR2} directory" return 1 fi # # Remove posix acl(all) # setfacl -b "${POSIX_ACL_TEST_DIR2}" if getfacl "${POSIX_ACL_TEST_DIR2}" | grep -q "^user:${TARGET_ACL_USER}"; then echo "Could not remove posix acl(all) to ${POSIX_ACL_TEST_DIR2} directory" return 1 fi # # Copy without permission # rm -rf "${POSIX_ACL_TEST_DIR2}" cp -r "${POSIX_ACL_TEST_DIR1}" "${POSIX_ACL_TEST_DIR2}" if getfacl "${POSIX_ACL_TEST_DIR2}" | grep -q "^default:user:${TARGET_ACL_USER}"; then echo "Could not copy without posix acl(default) to ${POSIX_ACL_TEST_DIR2} directory" return 1 fi if getfacl "${POSIX_ACL_TEST_DIR2}" | grep -q "^user:${TARGET_ACL_USER}"; then echo "Could not copy without posix acl(all) to ${POSIX_ACL_TEST_DIR2} directory" return 1 fi #------------------------------------------------------ # File #------------------------------------------------------ local POSIX_ACL_TEST_FILE1="posix_acl_dir1/posix_acl_file1" local POSIX_ACL_TEST_FILE2="posix_acl_dir1/posix_acl_file2" local POSIX_ACL_TEST_FILE3="posix_acl_dir2/posix_acl_file3" local POSIX_ACL_TEST_FILE4="posix_acl_dir2/posix_acl_file4" mkdir "${POSIX_ACL_TEST_DIR2}" touch "${POSIX_ACL_TEST_FILE1}" # # Check default inherited posix acl # if ! getfacl "${POSIX_ACL_TEST_FILE1}" | grep -q "^user:${TARGET_ACL_USER}:rwx"; then echo "Could not set posix acl(inherited default) to ${POSIX_ACL_TEST_FILE1} file" return 1 fi # # Overwrite posix acl # setfacl -m "u:${TARGET_ACL_USER}:r-x" "${POSIX_ACL_TEST_FILE1}" if ! getfacl "${POSIX_ACL_TEST_FILE1}" | grep -q "^user:${TARGET_ACL_USER}:r-x"; then echo "Could not overwrite posix acl to ${POSIX_ACL_TEST_FILE1} file" return 1 fi # # Rename # mv "${POSIX_ACL_TEST_FILE1}" "${POSIX_ACL_TEST_FILE2}" if ! getfacl "${POSIX_ACL_TEST_FILE2}" | grep -q "^user:${TARGET_ACL_USER}:r-x"; then echo "Could not move with posix acl to ${POSIX_ACL_TEST_FILE2} file" return 1 fi # # Copy with permission # cp -p "${POSIX_ACL_TEST_FILE2}" "${POSIX_ACL_TEST_FILE1}" if ! getfacl "${POSIX_ACL_TEST_FILE1}" | grep -q "^user:${TARGET_ACL_USER}:r-x"; then echo "Could not copy with posix acl to ${POSIX_ACL_TEST_FILE1} file" return 1 fi # # Remove posix acl # setfacl -b "${POSIX_ACL_TEST_FILE2}" if getfacl "${POSIX_ACL_TEST_FILE2}" | grep -q "^default:user:${TARGET_ACL_USER}"; then echo "Could not remove posix acl to ${POSIX_ACL_TEST_FILE2} file" return 1 fi # # Copy without permission(set parent directory default acl) # rm -f "${POSIX_ACL_TEST_FILE2}" cp "${POSIX_ACL_TEST_FILE1}" "${POSIX_ACL_TEST_FILE2}" if ! getfacl "${POSIX_ACL_TEST_FILE2}" | grep -q "^user:${TARGET_ACL_USER}:rwx"; then echo "Could not copy without posix acl(inherited parent) to ${POSIX_ACL_TEST_FILE2} file" return 1 fi # # Copy with permission(to no-acl directory) # cp -p "${POSIX_ACL_TEST_FILE1}" "${POSIX_ACL_TEST_FILE3}" if ! getfacl "${POSIX_ACL_TEST_FILE3}" | grep -q "^user:${TARGET_ACL_USER}:r-x"; then echo "Could not copy with posix acl to ${POSIX_ACL_TEST_FILE3} file in no-acl directory" return 1 fi # # Copy without permission(to no-acl directory) # cp "${POSIX_ACL_TEST_FILE1}" "${POSIX_ACL_TEST_FILE4}" if getfacl "${POSIX_ACL_TEST_FILE4}" | grep -q "^user:${TARGET_ACL_USER}"; then echo "Could not copy without posix acl to ${POSIX_ACL_TEST_FILE4} file in no-acl directory" return 1 fi rm -rf "${POSIX_ACL_TEST_DIR1}" rm -rf "${POSIX_ACL_TEST_DIR2}" } function test_copy_file { describe "Test simple copy ..." dd if=/dev/urandom of=/tmp/simple_file bs=1024 count=1 cp /tmp/simple_file copied_simple_file cmp /tmp/simple_file copied_simple_file rm_test_file /tmp/simple_file rm_test_file copied_simple_file } function test_write_after_seek_ahead { describe "Test writes succeed after a seek ahead ..." dd if=/dev/zero of=testfile seek=1 count=1 bs=1024 rm_test_file testfile } function test_overwrite_existing_file_range { describe "Test overwrite range succeeds ..." dd if=<(seq 1000) of="${TEST_TEXT_FILE}" dd if=/dev/zero of="${TEST_TEXT_FILE}" seek=1 count=1 bs=1024 conv=notrunc # [NOTE] # In ALPINE, comparisons using redirects often fail. # Therefore, create a real file for comparison and use it. # dd if=<(seq 1000) of=/tmp/cmp_base_file dd if=/dev/zero of=/tmp/cmp_base_file seek=1 count=1 bs=1024 conv=notrunc cmp "${TEST_TEXT_FILE}" /tmp/cmp_base_file rm_test_file /tmp/cmp_base_file rm_test_file } function test_concurrent_directory_updates { describe "Test concurrent updates to a directory ..." for i in $(seq 5); do echo foo > "${i}" done for _ in $(seq 10); do for i in $(seq 5); do local file # shellcheck disable=SC2012,SC2046 file=$(ls $(seq 5) | "${SED_BIN}" -n "$((RANDOM % 5 + 1))p") cat "${file}" >/dev/null || true rm -f "${file}" echo "foo" > "${file}" || true done & done wait # shellcheck disable=SC2046 rm -f $(seq 5) } function test_concurrent_reads { describe "Test concurrent reads from a file ..." ../../junk_data $((BIG_FILE_BLOCK_SIZE * BIG_FILE_COUNT)) > "${TEST_TEXT_FILE}" for _ in $(seq 10); do dd if="${TEST_TEXT_FILE}" of=/dev/null seek=$((RANDOM % BIG_FILE_LENGTH)) count=16 bs=1024 & done wait rm_test_file } function test_concurrent_writes { describe "Test concurrent writes to a file ..." ../../junk_data $((BIG_FILE_BLOCK_SIZE * BIG_FILE_COUNT)) > "${TEST_TEXT_FILE}" NUM_PROCS=10 PIDS=() for _ in $(seq "${NUM_PROCS}"); do dd if=/dev/zero of="${TEST_TEXT_FILE}" seek=$((RANDOM % BIG_FILE_LENGTH)) count=16 bs=1024 conv=notrunc & PIDS+=($!) done GRC=0 for PID in "${PIDS[@]}"; do wait "${PID}" RC=$? [ $RC -ne 0 ] && GRC="${RC}" done rm_test_file if [ "${GRC}" -ne 0 ]; then echo "unexpected return code: $GRC" return 1 fi } function test_open_second_fd { describe "read from an open fd ..." rm_test_file second_fd_file local RESULT # shellcheck disable=SC2094 RESULT=$( (echo foo ; wc -c < second_fd_file >&2) 2>& 1>second_fd_file) if [ "${RESULT}" -ne 4 ]; then echo "size mismatch, expected: 4, was: ${RESULT}" return 1 fi rm_test_file second_fd_file } function test_write_multiple_offsets { describe "test writing to multiple offsets ..." ../../write_multiblock -f "${TEST_TEXT_FILE}" -p "1024:1" -p "$((16 * 1024 * 1024)):1" -p "$((18 * 1024 * 1024)):1" rm_test_file "${TEST_TEXT_FILE}" } function test_write_multiple_offsets_backwards { describe "test writing to multiple offsets ..." ../../write_multiblock -f "${TEST_TEXT_FILE}" -p "$((20 * 1024 * 1024 + 1)):1" -p "$((10 * 1024 * 1024)):1" rm_test_file "${TEST_TEXT_FILE}" } function test_clean_up_cache() { describe "Test clean up cache ..." local dir="many_files" local count=25 mkdir -p "${dir}" for x in $(seq "${count}"); do ../../junk_data 10485760 > "${dir}"/file-"${x}" done local file_list=("${dir}"/*); local file_cnt="${#file_list[@]}" if [ "${file_cnt}" != "${count}" ]; then echo "Expected $count files but got ${file_cnt}" echo "Files: " "${file_list[@]}" rm -rf "${dir}" return 1 fi local CACHE_DISK_AVAIL_SIZE; CACHE_DISK_AVAIL_SIZE=$(get_disk_avail_size "${CACHE_DIR}") if [ "${CACHE_DISK_AVAIL_SIZE}" -lt "${ENSURE_DISKFREE_SIZE}" ];then echo "Cache disk avail size:${CACHE_DISK_AVAIL_SIZE} less than ensure_diskfree size:${ENSURE_DISKFREE_SIZE}" rm -rf "${dir}" return 1 fi rm -rf "${dir}" } function test_content_type() { describe "Test Content-Type detection ..." local DIR_NAME; DIR_NAME=$(basename "${PWD}") touch "test.txt" check_content_type "${DIR_NAME}/test.txt" "text/plain" touch "test.jpg" check_content_type "${DIR_NAME}/test.jpg" "image/jpeg" touch "test.bin" check_content_type "${DIR_NAME}/test.bin" "application/octet-stream" mkdir "test.dir" check_content_type "${DIR_NAME}/test.dir/" "application/x-directory" rm -f test.txt rm -f test.jpg rm -f test.bin rm -rf test.dir } # create more files than -o max_stat_cache_size function test_truncate_cache() { describe "Test make cache files over max cache file size ..." for dir in $(seq 2); do mkdir "${dir}" for file in $(seq 75); do touch "${dir}/${file}" done ls "${dir}" done # shellcheck disable=SC2046 rm -rf $(seq 2) } function test_cache_file_stat() { describe "Test cache file stat ..." ../../junk_data $((BIG_FILE_BLOCK_SIZE * BIG_FILE_COUNT)) > "${BIG_FILE}" # # The first argument of the script is "testrun-" the directory name. # local CACHE_TESTRUN_DIR=$1 # # get cache file inode number # local CACHE_FILE_INODE CACHE_FILE_INODE=$(get_inode "${CACHE_DIR}/${TEST_BUCKET_1}/${CACHE_TESTRUN_DIR}/${BIG_FILE}") if [ -z "${CACHE_FILE_INODE}" ]; then echo "Not found cache file or failed to get inode: ${CACHE_DIR}/${TEST_BUCKET_1}/${CACHE_TESTRUN_DIR}/${BIG_FILE}" return 1; fi # # get lines from cache stat file # local CACHE_FILE_STAT_LINE_1; CACHE_FILE_STAT_LINE_1=$("${SED_BIN}" -n 1p "${CACHE_DIR}/.${TEST_BUCKET_1}.stat/${CACHE_TESTRUN_DIR}/${BIG_FILE}") local CACHE_FILE_STAT_LINE_2; CACHE_FILE_STAT_LINE_2=$("${SED_BIN}" -n 2p "${CACHE_DIR}/.${TEST_BUCKET_1}.stat/${CACHE_TESTRUN_DIR}/${BIG_FILE}") if [ -z "${CACHE_FILE_STAT_LINE_1}" ] || [ -z "${CACHE_FILE_STAT_LINE_2}" ]; then echo "could not get first or second line from cache file stat: ${CACHE_DIR}/.${TEST_BUCKET_1}.stat/${CACHE_TESTRUN_DIR}/${BIG_FILE}" return 1; fi # # compare # if [ "${CACHE_FILE_STAT_LINE_1}" != "${CACHE_FILE_INODE}:${BIG_FILE_LENGTH}" ]; then echo "first line(cache file stat) is different: \"${CACHE_FILE_STAT_LINE_1}\" != \"${CACHE_FILE_INODE}:${BIG_FILE_LENGTH}\"" return 1; fi if [ "${CACHE_FILE_STAT_LINE_2}" != "0:${BIG_FILE_LENGTH}:1:0" ]; then echo "last line(cache file stat) is different: \"${CACHE_FILE_STAT_LINE_2}\" != \"0:${BIG_FILE_LENGTH}:1:0\"" return 1; fi # # remove cache files directly # rm -f "${CACHE_DIR}/${TEST_BUCKET_1}/${CACHE_TESTRUN_DIR}/${BIG_FILE}" rm -f "${CACHE_DIR}/.${TEST_BUCKET_1}.stat/${CACHE_TESTRUN_DIR}/${BIG_FILE}" # # write a byte into the middle(not the boundary) of the file # local CHECK_UPLOAD_OFFSET=$((10 * 1024 * 1024 + 17)) dd if=/dev/urandom of="${BIG_FILE}" bs=1 count=1 seek="${CHECK_UPLOAD_OFFSET}" conv=notrunc # # get cache file inode number # CACHE_FILE_INODE=$(get_inode "${CACHE_DIR}/${TEST_BUCKET_1}/${CACHE_TESTRUN_DIR}/${BIG_FILE}") if [ -z "${CACHE_FILE_INODE}" ]; then echo "Not found cache file or failed to get inode: ${CACHE_DIR}/${TEST_BUCKET_1}/${CACHE_TESTRUN_DIR}/${BIG_FILE}" return 1; fi # # get lines from cache stat file # CACHE_FILE_STAT_LINE_1=$("${SED_BIN}" -n 1p "${CACHE_DIR}/.${TEST_BUCKET_1}.stat/${CACHE_TESTRUN_DIR}/${BIG_FILE}") local CACHE_FILE_STAT_LINE_E; CACHE_FILE_STAT_LINE_E=$(tail -1 "${CACHE_DIR}/.${TEST_BUCKET_1}.stat/${CACHE_TESTRUN_DIR}/${BIG_FILE}" 2>/dev/null) if [ -z "${CACHE_FILE_STAT_LINE_1}" ] || [ -z "${CACHE_FILE_STAT_LINE_E}" ]; then echo "could not get first or end line from cache file stat: ${CACHE_DIR}/.${TEST_BUCKET_1}.stat/${CACHE_TESTRUN_DIR}/${BIG_FILE}" return 1; fi # # check first and cache file length from last line # # we should check all stat lines, but there are cases where the value # differs depending on the processing system etc., then the cache file # size is calculated and compared. # local CACHE_LAST_OFFSET; CACHE_LAST_OFFSET=$(echo "${CACHE_FILE_STAT_LINE_E}" | cut -d ":" -f1) local CACHE_LAST_SIZE; CACHE_LAST_SIZE=$(echo "${CACHE_FILE_STAT_LINE_E}" | cut -d ":" -f2) local CACHE_TOTAL_SIZE=$((CACHE_LAST_OFFSET + CACHE_LAST_SIZE)) if [ "${CACHE_FILE_STAT_LINE_1}" != "${CACHE_FILE_INODE}:${BIG_FILE_LENGTH}" ]; then echo "first line(cache file stat) is different: \"${CACHE_FILE_STAT_LINE_1}\" != \"${CACHE_FILE_INODE}:${BIG_FILE_LENGTH}\"" return 1; fi if [ "${BIG_FILE_LENGTH}" -ne "${CACHE_TOTAL_SIZE}" ]; then echo "the file size indicated by the cache stat file is different: \"${BIG_FILE_LENGTH}\" != \"${CACHE_TOTAL_SIZE}\"" return 1; fi rm_test_file "${BIG_FILE}" } function test_zero_cache_file_stat() { describe "Test zero byte cache file stat ..." rm_test_file "${TEST_TEXT_FILE}" # # create empty file # touch "${TEST_TEXT_FILE}" # # The first argument of the script is "testrun-" the directory name. # local CACHE_TESTRUN_DIR=$1 # [NOTE] # The stat file is a one-line text file, expecting for ":0"(ex. "4543937: 0"). # if ! head -1 "${CACHE_DIR}/.${TEST_BUCKET_1}.stat/${CACHE_TESTRUN_DIR}/${TEST_TEXT_FILE}" 2>/dev/null | grep -q ':0$' 2>/dev/null; then echo "The cache file stat after creating an empty file is incorrect : ${CACHE_DIR}/.${TEST_BUCKET_1}.stat/${CACHE_TESTRUN_DIR}/${TEST_TEXT_FILE}" return 1; fi rm_test_file "${TEST_TEXT_FILE}" } function test_upload_sparsefile { describe "Testing upload sparse file ..." rm_test_file "${BIG_FILE}" rm -f "${TEMP_DIR}/${BIG_FILE}" # # Make all HOLE file # "${TRUNCATE_BIN}" "${BIG_FILE}" -s "${BIG_FILE_LENGTH}" # # Write some bytes to ABOUT middle in the file # (Dare to remove the block breaks) # local WRITE_POS=$((BIG_FILE_LENGTH / 2 - 128)) echo -n "0123456789ABCDEF" | dd of="${TEMP_DIR}/${BIG_FILE}" bs=1 count=16 seek="${WRITE_POS}" conv=notrunc # # copy(upload) the file # cp "${TEMP_DIR}/${BIG_FILE}" "${BIG_FILE}" # # check # cmp "${TEMP_DIR}/${BIG_FILE}" "${BIG_FILE}" rm_test_file "${BIG_FILE}" rm -f "${TEMP_DIR}/${BIG_FILE}" } function test_mix_upload_entities() { describe "Testing upload sparse files ..." # # Make test file # ../../junk_data $((BIG_FILE_BLOCK_SIZE * BIG_FILE_COUNT)) > "${BIG_FILE}" # # If the cache option is enabled, delete the cache of uploaded files. # if [ -f "${CACHE_DIR}/${TEST_BUCKET_1}/${BIG_FILE}" ]; then rm -f "${CACHE_DIR}/${TEST_BUCKET_1}/${BIG_FILE}" fi if [ -f "${CACHE_DIR}/.${TEST_BUCKET_1}.stat/${BIG_FILE}" ]; then rm -f "${CACHE_DIR}/.${TEST_BUCKET_1}.stat/${BIG_FILE}" fi # # Do a partial write to the file. # echo -n "0123456789ABCDEF" | dd of="${BIG_FILE}" bs=1 count=16 seek=0 conv=notrunc echo -n "0123456789ABCDEF" | dd of="${BIG_FILE}" bs=1 count=16 seek=8192 conv=notrunc echo -n "0123456789ABCDEF" | dd of="${BIG_FILE}" bs=1 count=16 seek=1073152 conv=notrunc echo -n "0123456789ABCDEF" | dd of="${BIG_FILE}" bs=1 count=16 seek=26214400 conv=notrunc echo -n "0123456789ABCDEF" | dd of="${BIG_FILE}" bs=1 count=16 seek=26222592 conv=notrunc rm_test_file "${BIG_FILE}" } # # [NOTE] # This test runs last because it uses up disk space and may not recover. # This may be a problem, especially on MacOS. (See the comment near the definition # line for the ENSURE_DISKFREE_SIZE variable) # function test_ensurespace_move_file() { describe "Testing upload(mv) file when disk space is not enough ..." # # Make test file which is not under mountpoint # mkdir -p "${CACHE_DIR}/.s3fs_test_tmpdir" ../../junk_data $((BIG_FILE_BLOCK_SIZE * BIG_FILE_COUNT)) > "${CACHE_DIR}/.s3fs_test_tmpdir/${BIG_FILE}" # # Backup file stat # local ORIGINAL_PERMISSIONS ORIGINAL_PERMISSIONS=$(get_user_and_group "${CACHE_DIR}/.s3fs_test_tmpdir/${BIG_FILE}") # # Fill the disk size # local NOW_CACHE_DISK_AVAIL_SIZE; NOW_CACHE_DISK_AVAIL_SIZE=$(get_disk_avail_size "${CACHE_DIR}") local TMP_FILE_NO=0 while true; do local ALLOWED_USING_SIZE=$((NOW_CACHE_DISK_AVAIL_SIZE - ENSURE_DISKFREE_SIZE)) if [ "${ALLOWED_USING_SIZE}" -gt "${BIG_FILE_LENGTH}" ]; then cp -p "${CACHE_DIR}/.s3fs_test_tmpdir/${BIG_FILE}" "${CACHE_DIR}/.s3fs_test_tmpdir/${BIG_FILE}_${TMP_FILE_NO}" local TMP_FILE_NO=$((TMP_FILE_NO + 1)) else break; fi done # # move file # mv "${CACHE_DIR}/.s3fs_test_tmpdir/${BIG_FILE}" "${BIG_FILE}" # # file stat # local MOVED_PERMISSIONS MOVED_PERMISSIONS=$(get_user_and_group "${BIG_FILE}") local MOVED_FILE_LENGTH MOVED_FILE_LENGTH=$(get_size "${BIG_FILE}") # # check # if [ "${MOVED_PERMISSIONS}" != "${ORIGINAL_PERMISSIONS}" ]; then echo "Failed to move file with permission" return 1 fi if [ "${MOVED_FILE_LENGTH}" -ne "${BIG_FILE_LENGTH}" ]; then echo "Failed to move file with file length: ${MOVED_FILE_LENGTH} ${BIG_FILE_LENGTH}" return 1 fi rm_test_file "${BIG_FILE}" rm -rf "${CACHE_DIR}/.s3fs_test_tmpdir" } function test_not_existed_dir_obj() { describe "Test not existed directory object..." local DIR_NAME; DIR_NAME=$(basename "${PWD}") # # Create files under not existed directory by aws command # local OBJECT_NAME_1; OBJECT_NAME_1="${DIR_NAME}/not_existed_dir_single/${TEST_TEXT_FILE}" local OBJECT_NAME_2; OBJECT_NAME_2="${DIR_NAME}/not_existed_dir_parent/not_existed_dir_child/${TEST_TEXT_FILE}" echo data1 | aws_cli s3 cp - "s3://${TEST_BUCKET_1}/${OBJECT_NAME_1}" echo data2 | aws_cli s3 cp - "s3://${TEST_BUCKET_1}/${OBJECT_NAME_2}" # Top directory # # [NOTE] # Immediately after uploading with the aws command, there is a case where # the directory path cannot be found (404 error). # In this case, it may be listed correctly after a few retries. # This error is a response problem on the S3 server side, not a problem # with the s3fs process, so we try again. # _RETRY_COUNT=0 # shellcheck disable=SC2010 while ! ls | grep -q 'not_existed_dir_single'; do if [ "${_RETRY_COUNT}" -ge 3 ]; then echo "Expect to find \"not_existed_dir_single\" directory by ls command, but it is not found (after ${_RETRY_COUNT} retries)" return 1 fi sleep 1 done _RETRY_COUNT=0 # shellcheck disable=SC2010 while ! ls | grep -q 'not_existed_dir_parent'; do if [ "${_RETRY_COUNT}" -ge 3 ]; then echo "Expect to find \"not_existed_dir_parent\" directory by ls command, but it is not found (after ${_RETRY_COUNT} retries)" return 1 fi sleep 1 done # Single nest directory # shellcheck disable=SC2086 if ! "${STAT_BIN[@]}" not_existed_dir_single; then echo "Expect to find \"not_existed_dir_single\" directory, but it is not found" return 1; fi # shellcheck disable=SC2010 if ! ls not_existed_dir_single | grep -q "^${TEST_TEXT_FILE}\$"; then echo "Expect to find \"not_existed_dir_single/${TEST_TEXT_FILE}\" file, but it is not found" return 1; fi # shellcheck disable=SC2010 if ! ls "not_existed_dir_single/${TEST_TEXT_FILE}" | grep -q "^not_existed_dir_single/${TEST_TEXT_FILE}\$"; then echo "Expect to find \"not_existed_dir_single/${TEST_TEXT_FILE}\" file, but it is not found" return 1; fi # Double nest directory # shellcheck disable=SC2086 if ! "${STAT_BIN[@]}" not_existed_dir_parent; then echo "Expect to find \"not_existed_dir_parent\" directory, but it is not found" return 1; fi # shellcheck disable=SC2010 if ! ls not_existed_dir_parent | grep -q '^not_existed_dir_child'; then echo "Expect to find \"not_existed_dir_parent/not_existed_dir_child\" directory, but it is not found" return 1; fi # shellcheck disable=SC2086 if ! "${STAT_BIN[@]}" not_existed_dir_parent/not_existed_dir_child; then echo "Expect to find \"not_existed_dir_parent/not_existed_dir_child\" directory, but it is not found" return 1; fi # shellcheck disable=SC2010 if ! ls not_existed_dir_parent/not_existed_dir_child | grep -q "^${TEST_TEXT_FILE}\$"; then echo "Expect to find \"not_existed_dir_parent/not_existed_dir_child/${TEST_TEXT_FILE}\" directory, but it is not found" return 1; fi # shellcheck disable=SC2010 if ! ls "not_existed_dir_parent/not_existed_dir_child/${TEST_TEXT_FILE}" | grep -q "^not_existed_dir_parent/not_existed_dir_child/${TEST_TEXT_FILE}\$"; then echo "Expect to find \"not_existed_dir_parent/not_existed_dir_child/${TEST_TEXT_FILE}\" directory, but it is not found" return 1; fi rm -rf not_existed_dir_single rm -rf not_existed_dir_parent } function test_ut_ossfs { describe "Testing ossfs python ut..." # shellcheck disable=SC2153 export TEST_BUCKET_MOUNT_POINT="${TEST_BUCKET_MOUNT_POINT_1}" ../../ut_test.py } function test_cr_filename { describe "Testing filename with CR code ..." # The following tests create a file, test it, and delete it. # So this test just calls the following program. # ../../cr_filename "${TEST_TEXT_FILE}" } # # This test opens a file and writes multiple sets of data. # The file is opened only once and multiple blocks of data are written # to the file descriptor with a gap. # # That is, the data sets are written discontinuously. # The data to be written uses multiple data that is less than or larger # than the part size of the multi-part upload. # The gap should be at least the part size of the multi-part upload. # Write as shown below: # ............ # # There are two types of tests: new files and existing files. # For existing files, the file size must be larger than where this test # writes last position. # ............... # function test_write_data_with_skip() { describe "Testing write data block with skipping block..." # # The first argument of the script is "testrun-" the directory name. # local CACHE_TESTRUN_DIR=$1 local _SKIPWRITE_FILE="test_skipwrite" local _TMP_SKIPWRITE_FILE="/tmp/${_SKIPWRITE_FILE}" #------------------------------------------------------ # (1) test new file #------------------------------------------------------ # # Clean files # rm_test_file "${_SKIPWRITE_FILE}" rm_test_file "${_TMP_SKIPWRITE_FILE}" # # Create new file in bucket and temporary directory(/tmp) # # Writing to the file is as follows: # |<-- skip(12MB) --><-- write(1MB) --><-- skip(22MB) --><-- write(20MB) --><-- skip(23MB) --><-- write(1MB) -->| (79MB) # # As a result, areas that are not written to the file are mixed. # The part that is not written has a HOLE that is truncate and filled # with 0x00. # Assuming that multipart upload is performed on a part-by-part basis, # it will be as follows: # part 1) 0x0.. 0x9FFFFF : # part 2) 0xA00000..0x13FFFFF : 0xA00000..0xBFFFFF # 0xC00000..0xCFFFFF # 0xD00000..0x13FFFFF # part 3) 0x1400000..0x1DFFFFF : # part 4) 0x1E00000..0x27FFFFF : 0x1E00000..0x22FFFFF # 0x2300000..0x27FFFFF # part 5) 0x2800000..0x31FFFFF : # part 6) 0x3200000..0x3BFFFFF : 0x3200000..0x36FFFFF # 0x3700000..0x3BFFFFF # part 7) 0x3C00000..0x45FFFFF : # part 8) 0x4600000..0x4BFFFFF : 0x4600000..0x4AFFFFF # 0x4B00000..0x4BFFFFF # ../../write_multiblock -f "${_SKIPWRITE_FILE}" -f "${_TMP_SKIPWRITE_FILE}" -p 12582912:65536 -p 36700160:20971520 -p 78643200:65536 # # delete cache file if using cache # if s3fs_args | grep -q use_cache; then rm -f "${CACHE_DIR}/${TEST_BUCKET_1}/${CACHE_TESTRUN_DIR}/${_SKIPWRITE_FILE}" rm -f "${CACHE_DIR}/.${TEST_BUCKET_1}.stat/${CACHE_TESTRUN_DIR}/${_SKIPWRITE_FILE}" fi # # Compare # cmp "${_SKIPWRITE_FILE}" "${_TMP_SKIPWRITE_FILE}" #------------------------------------------------------ # (2) test existed file #------------------------------------------------------ # [NOTE] # This test uses the file used in the previous test as an existing file. # if s3fs_args | grep -q use_cache; then rm -f "${CACHE_DIR}/${TEST_BUCKET_1}/${CACHE_TESTRUN_DIR}/${_SKIPWRITE_FILE}" rm -f "${CACHE_DIR}/.${TEST_BUCKET_1}.stat/${CACHE_TESTRUN_DIR}/${_SKIPWRITE_FILE}" fi # # Over write data to existed file in bucket and temporary directory(/tmp) # # Writing to the file is as follows: # |<----------------------------------------------- existed file ----------------------------------------------------------->| (79MB) # |<-- skip(12MB) --><-- write(1MB) --><-- skip(22MB) --><-- write(20MB) --><-- skip(22MB) --><-- write(1MB) --><-- 1MB -->| (79MB) # # As a result, areas that are not written to the file are mixed. # The part that is not written has a HOLE that is truncate and filled # with 0x00. # Assuming that multipart upload is performed on a part-by-part basis, # it will be as follows: # part 1) 0x0.. 0x9FFFFF : # part 2) 0xA00000..0x13FFFFF : 0xA00000..0xBFFFFF # 0xC00000..0xCFFFFF # 0xD00000..0x13FFFFF # part 3) 0x1400000..0x1DFFFFF : # part 4) 0x1E00000..0x27FFFFF : 0x1E00000..0x22FFFFF # 0x2300000..0x27FFFFF # part 5) 0x2800000..0x31FFFFF : # part 6) 0x3200000..0x3BFFFFF : 0x3200000..0x36FFFFF # 0x3700000..0x3BFFFFF # part 7) 0x3C00000..0x45FFFFF : # part 8) 0x4600000..0x4BFFFFF : 0x4600000..0x49FFFFF # part 8) 0x4600000..0x4BFFFFF : 0x4A00000..0x4AFFFFF # 0x4B00000..0x4BFFFFF # ../../write_multiblock -f "${_SKIPWRITE_FILE}" -f "${_TMP_SKIPWRITE_FILE}" -p 12582912:65536 -p 36700160:20971520 -p 77594624:65536 # # delete cache file if using cache # if s3fs_args | grep -q use_cache; then rm -f "${CACHE_DIR}/${TEST_BUCKET_1}/${CACHE_TESTRUN_DIR}/${_SKIPWRITE_FILE}" rm -f "${CACHE_DIR}/.${TEST_BUCKET_1}.stat/${CACHE_TESTRUN_DIR}/${_SKIPWRITE_FILE}" fi # # Compare # cmp "${_SKIPWRITE_FILE}" "${_TMP_SKIPWRITE_FILE}" # # Clean files # rm_test_file "${_SKIPWRITE_FILE}" rm_test_file "${_TMP_SKIPWRITE_FILE}" } function test_not_boundary_writes { describe "Test non-boundary write ..." # [MEMO] # Files used in this test, multipart related sizes, etc. # # Test file size: 25MB(25 * 1024 * 1024) # Multipart size: 10MB # Multipart minimum upload size: 5MB # # The multipart upload part that should be executed here is as follows: # Part number 1: 0 - 10,485,759 (size = 10MB) # Part number 2: 10,485,760 - 20,971,519 (size = 10MB) # Part number 3: 20,971,520 - 26,214,399 (size = 5MB) # local BOUNDAY_TEST_FILE_SIZE; BOUNDAY_TEST_FILE_SIZE=$((BIG_FILE_BLOCK_SIZE * BIG_FILE_COUNT)) ../../junk_data "${BOUNDAY_TEST_FILE_SIZE}" > "${TEST_TEXT_FILE}" # # Write in First boundary # # Write 0 - 3,145,727(3MB) : less than the multipart minimum size from the beginning dd if=/dev/zero of="${TEST_TEXT_FILE}" seek=0 count=3072 bs=1024 conv=notrunc # Write 0 - 7,340,031(7MB) : multipart exceeding the minimum size from the beginning dd if=/dev/zero of="${TEST_TEXT_FILE}" seek=0 count=7168 bs=1024 conv=notrunc # Write 0 - 12,582,911(12MB) : beyond the multipart size boundary from the beginning dd if=/dev/zero of="${TEST_TEXT_FILE}" seek=0 count=12288 bs=1024 conv=notrunc # # Write in First and second boundary # # Write 3,145,728 - 4,194,303(1MB) : less than the minimum multipart size from the middle of the first multipart area dd if=/dev/zero of="${TEST_TEXT_FILE}" seek=3072 count=1024 bs=1024 conv=notrunc # Write 3,145,728 - 9,437,183(6MB) : exceeding the minimum multipart size from the middle of the first multipart area dd if=/dev/zero of="${TEST_TEXT_FILE}" seek=3072 count=6144 bs=1024 conv=notrunc # Write 3,145,728 - 12,582,911(9MB) : beyond the multipart boundary from the middle of the first multipart area dd if=/dev/zero of="${TEST_TEXT_FILE}" seek=3072 count=9216 bs=1024 conv=notrunc # # Write in Second boundary # # Write 12,582,912 - 14,680,063(2MB) : below the minimum multipart size from the middle of the multipart area dd if=/dev/zero of="${TEST_TEXT_FILE}" seek=12288 count=2048 bs=1024 conv=notrunc # Write 12,582,912 - 18,874,367(6MB) : data exceeding the minimum multipart size from the middle of the multipart area dd if=/dev/zero of="${TEST_TEXT_FILE}" seek=12288 count=6144 bs=1024 conv=notrunc # Write 12,582,912 - 23,068,671(10MB) : beyond the multipart boundary from the middle of the multipart area dd if=/dev/zero of="${TEST_TEXT_FILE}" seek=12288 count=10240 bs=1024 conv=notrunc # Write 12,582,912 - 26,214,399(13MB) : beyond the multipart boundary(last) from the middle of the multipart area dd if=/dev/zero of="${TEST_TEXT_FILE}" seek=12288 count=13312 bs=1024 conv=notrunc # # Write in Last boundary # # Write 23,068,672 - 24,117,247(1MB) : below the minimum multipart size from the middle of the final multipart area dd if=/dev/zero of="${TEST_TEXT_FILE}" seek=22528 count=1024 bs=1024 conv=notrunc # Write 23,068,672 - 26,214,399(3MB) : beyond the multipart boundary(last) from the middle of the final multipart area dd if=/dev/zero of="${TEST_TEXT_FILE}" seek=22528 count=3072 bs=1024 conv=notrunc rm_test_file } function test_chmod_mountpoint { describe "Testing chmod to mount point..." local MOUNTPOINT_DIR; MOUNTPOINT_DIR=$(cd ..; pwd) local ORIGINAL_PERMISSIONS; ORIGINAL_PERMISSIONS=$(get_permissions "${MOUNTPOINT_DIR}") chmod 0777 "${MOUNTPOINT_DIR}"; # if they're the same, we have a problem. local CHANGED_PERMISSIONS; CHANGED_PERMISSIONS=$(get_permissions "${MOUNTPOINT_DIR}") if [ "${CHANGED_PERMISSIONS}" = "${ORIGINAL_PERMISSIONS}" ] then echo "Could not modify mount point(${MOUNTPOINT_DIR}) permissions" return 1 fi } function test_chown_mountpoint { describe "Testing chown mount point..." local MOUNTPOINT_DIR; MOUNTPOINT_DIR=$(cd ..; pwd) local ORIGINAL_PERMISSIONS ORIGINAL_PERMISSIONS=$(get_user_and_group "${MOUNTPOINT_DIR}") # [NOTE] # Prevents test interruptions due to permission errors, etc. # If the chown command fails, an error will occur with the # following judgment statement. So skip the chown command error. # '|| true' was added due to a problem with MacOS and ensure_diskfree option. # chown 1000:1000 "${MOUNTPOINT_DIR}" || true # if they're the same, we have a problem. local CHANGED_PERMISSIONS CHANGED_PERMISSIONS=$(get_user_and_group "${MOUNTPOINT_DIR}") if [ "${CHANGED_PERMISSIONS}" = "${ORIGINAL_PERMISSIONS}" ] then if [ "${ORIGINAL_PERMISSIONS}" = "1000:1000" ] then echo "Could not be strict check because original file permission 1000:1000" else echo "Could not modify mount point(${MOUNTPOINT_DIR}) ownership($ORIGINAL_PERMISSIONS to 1000:1000)" return 1 fi fi } function test_time_mountpoint { describe "Testing atime/ctime/mtime to mount point..." local MOUNTPOINT_DIR; MOUNTPOINT_DIR=$(cd ..; pwd) local base_atime; base_atime=$(get_atime "${MOUNTPOINT_DIR}") local base_ctime; base_ctime=$(get_ctime "${MOUNTPOINT_DIR}") local base_mtime; base_mtime=$(get_mtime "${MOUNTPOINT_DIR}") touch "${MOUNTPOINT_DIR}" local atime; atime=$(get_atime "${MOUNTPOINT_DIR}") local ctime; ctime=$(get_ctime "${MOUNTPOINT_DIR}") local mtime; mtime=$(get_mtime "${MOUNTPOINT_DIR}") if [ "${base_atime}" = "${atime}" ] || [ "${base_ctime}" = "${ctime}" ] || [ "${base_mtime}" = "${mtime}" ]; then echo "chmod expected updated ctime: $base_ctime != $ctime, mtime: $base_mtime == $mtime, atime: $base_atime == $atime" return 1 fi } function test_file_names_longer_than_posix() { local DIR_NAME; DIR_NAME=$(basename "${PWD}") a256="aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa" #a256="aaaa" if ! touch "${a256}"; then echo "could not create long file name" return 1 fi rm -f "${a256}" echo data | aws_cli s3 cp - "s3://${TEST_BUCKET_1}/${DIR_NAME}/${a256}" files=(*) if [ "${#files[@]}" = 0 ]; then echo "failed to list long file name" return 1 fi rm -f "${a256}" } function test_statvfs() { describe "Testing the free/available size on mount point(statvfs)..." # [NOTE] # The df command result format is different between Linux and macos, # but the order of Total/Used/Available size is the same. # local MOUNTPOINT_DIR; MOUNTPOINT_DIR=$(cd ..; pwd) local DF_RESULT; DF_RESULT=$(df "${MOUNTPOINT_DIR}" 2>/dev/null | tail -n +2) local TOTAL_SIZE; TOTAL_SIZE=$(echo "${DF_RESULT}" | awk '{print $2}') local USED_SIZE; USED_SIZE=$(echo "${DF_RESULT}" | awk '{print $3}') local AVAIL_SIZE; AVAIL_SIZE=$(echo "${DF_RESULT}" | awk '{print $4}') # [NOTE] # In the disk information (statvfs) provided by s3fs, Total size and # Available size are always the same and not 0, and used size is always 0. # if [ -z "${TOTAL_SIZE}" ] || [ -z "${AVAIL_SIZE}" ] || [ -z "${USED_SIZE}" ] || [ "${TOTAL_SIZE}" = "0" ] || [ "${AVAIL_SIZE}" = "0" ] || [ "${TOTAL_SIZE}" != "${AVAIL_SIZE}" ] || [ "${USED_SIZE}" != "0" ]; then echo "The result of df command is wrong: Total=${TOTAL_SIZE}, Used=${USED_SIZE}, Available=${AVAIL_SIZE}" return 1 fi } function add_all_tests { if s3fs_args | grep -q use_cache; then add_tests test_cache_file_stat add_tests test_zero_cache_file_stat else add_tests test_file_names_longer_than_posix fi if ! s3fs_args | grep -q ensure_diskfree && ! uname | grep -q Darwin; then add_tests test_clean_up_cache fi add_tests test_create_empty_file add_tests test_append_file add_tests test_truncate_file add_tests test_truncate_upload add_tests test_truncate_empty_file add_tests test_truncate_shrink_file add_tests test_truncate_shrink_read_file add_tests test_mv_file add_tests test_mv_to_exist_file add_tests test_mv_empty_directory add_tests test_mv_nonempty_directory add_tests test_redirects add_tests test_mkdir_rmdir add_tests test_chmod add_tests test_chown add_tests test_list add_tests test_remove_nonempty_directory add_tests test_external_directory_creation add_tests test_external_modification add_tests test_external_creation add_tests test_read_external_object add_tests test_read_external_dir_object add_tests test_update_metadata_external_small_object add_tests test_update_metadata_external_large_object add_tests test_rename_before_close add_tests test_multipart_upload add_tests test_multipart_copy add_tests test_multipart_mix add_tests test_utimens_during_multipart add_tests test_special_characters add_tests test_hardlink add_tests test_symlink if ! uname | grep -q Darwin; then add_tests test_mknod fi add_tests test_extended_attributes add_tests test_mtime_file add_tests test_update_time_chmod add_tests test_update_time_chown add_tests test_update_time_xattr add_tests test_update_time_touch if ! mount -t fuse.s3fs | grep "$TEST_BUCKET_MOUNT_POINT_1 " | grep -q -e noatime -e relatime ; then add_tests test_update_time_touch_a fi add_tests test_update_time_append add_tests test_update_time_cp_p add_tests test_update_time_mv add_tests test_update_directory_time_chmod add_tests test_update_directory_time_chown add_tests test_update_directory_time_set_xattr add_tests test_update_directory_time_touch if ! mount -t fuse.s3fs | grep "$TEST_BUCKET_MOUNT_POINT_1 " | grep -q -e noatime -e relatime ; then add_tests test_update_directory_time_touch_a fi add_tests test_update_directory_time_subdir add_tests test_update_chmod_opened_file if s3fs_args | grep -q update_parent_dir_stat; then add_tests test_update_parent_directory_time fi if ! s3fs_args | grep -q use_xattr; then add_tests test_posix_acl fi add_tests test_rm_rf_dir add_tests test_copy_file add_tests test_write_after_seek_ahead add_tests test_overwrite_existing_file_range add_tests test_concurrent_directory_updates add_tests test_concurrent_reads add_tests test_concurrent_writes add_tests test_open_second_fd add_tests test_write_multiple_offsets add_tests test_write_multiple_offsets_backwards add_tests test_content_type add_tests test_truncate_cache add_tests test_upload_sparsefile add_tests test_mix_upload_entities # TODO: investigate why only Alpine cannot see the implicit directory objects. if ! test -f /etc/os-release || ! grep -q -i -e 'ID=alpine' -e 'ID="alpine"' /etc/os-release; then add_tests test_not_existed_dir_obj fi add_tests test_ut_ossfs add_tests test_cr_filename if ! s3fs_args | grep -q ensure_diskfree && ! uname | grep -q Darwin; then add_tests test_ensurespace_move_file fi add_tests test_write_data_with_skip add_tests test_not_boundary_writes # [NOTE] # The test on CI will fail depending on the permissions, so skip these(chmod/chown). # # add_tests test_chmod_mountpoint # add_tests test_chown_mountpoint add_tests test_time_mountpoint add_tests test_statvfs } init_suite add_all_tests run_suite # # Local variables: # tab-width: 4 # c-basic-offset: 4 # End: # vim600: expandtab sw=4 ts=4 fdm=marker # vim<600: expandtab sw=4 ts=4 #