Skip to content

Commit

Permalink
Fix typos (s3fs-fuse#2473)
Browse files Browse the repository at this point in the history
  • Loading branch information
szepeviktor authored Jun 23, 2024
1 parent 254d717 commit fa807a5
Show file tree
Hide file tree
Showing 19 changed files with 36 additions and 36 deletions.
2 changes: 1 addition & 1 deletion .github/workflows/ci.yml
Original file line number Diff line number Diff line change
Expand Up @@ -141,7 +141,7 @@ jobs:
# This product(package) is a workaround for osxfuse which required an OS reboot(macos 11 and later).
# see. https://github.com/macos-fuse-t/fuse-t
# About osxfuse
# This job doesn't work with Github Actions using macOS 11+ because "load_osxfuse" returns
# This job doesn't work with GitHub Actions using macOS 11+ because "load_osxfuse" returns
# "exit code = 1".(requires OS reboot)
#
macos12:
Expand Down
2 changes: 1 addition & 1 deletion COMPILATION.md
Original file line number Diff line number Diff line change
Expand Up @@ -42,7 +42,7 @@ Keep in mind using the pre-built packages when available.
./configure
```
Depending on the TLS library (OpenSSL/GnuTLS/NSS), add `--with-openssl`, `--with-gnutls` or `--with-nss` when executing `configure`. (If omitted, it is equivalent to `--with-openssl`.)
3. Bulding:
3. Building:
```sh
make
```
Expand Down
2 changes: 1 addition & 1 deletion ChangeLog
Original file line number Diff line number Diff line change
Expand Up @@ -535,7 +535,7 @@ issue #3 - Fixed local timezone was incorrectly being applied to IAM and Last-Mo
issue #4 - Fix compilation error on MacOSX with missing const

Version 1.74 -- Nov 24, 2013
This version is initial version on Github, same as on GoogleCodes(s3fs).
This version is initial version on GitHub, same as on GoogleCodes(s3fs).
https://github.com/s3fs-fuse/s3fs-fuse/releases/tag/v1.74
see more detail on googlecodes: https://code.google.com/p/s3fs/downloads/detail?name=s3fs-1.74.tar.gz

Expand Down
2 changes: 1 addition & 1 deletion doc/man/s3fs.1.in
Original file line number Diff line number Diff line change
Expand Up @@ -298,7 +298,7 @@ Note that this option is still experimental and may change in the future.
.TP
\fB\-o\fR max_thread_count (default is "5")
Specifies the number of threads waiting for stream uploads.
Note that this option and Streamm Upload are still experimental and subject to change in the future.
Note that this option and Stream Upload are still experimental and subject to change in the future.
This option will be merged with "parallel_count" in the future.
.TP
\fB\-o\fR enable_content_md5 (default is disable)
Expand Down
2 changes: 1 addition & 1 deletion src/curl.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -3782,7 +3782,7 @@ int S3fsCurl::CheckBucket(const char* check_path, bool compat_dir, bool force_no
query_string += '&';
}
query_string += "prefix=";
query_string += &check_path[1]; // skip first '/' charactor
query_string += &check_path[1]; // skip first '/' character
}
}
if(!query_string.empty()){
Expand Down
2 changes: 1 addition & 1 deletion src/curl_multi.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -252,7 +252,7 @@ int S3fsMultiCurl::MultiRead()
}
}else{
S3FS_PRN_ERR("failed a request(Unknown response code: %s)", s3fscurl->url.c_str());
// Reuse partical file
// Reuse particular file
switch(curlCode){
case CURLE_OPERATION_TIMEDOUT:
isRetry = true;
Expand Down
2 changes: 1 addition & 1 deletion src/fdcache_entity.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -1868,7 +1868,7 @@ int FdEntity::RowFlushStreamMultipart(PseudoFdInfo* pseudo_obj, const char* tpat
// [NOTE]
// If there is a part where has already been uploading, that part
// is re-updated after finishing uploading, so the part of the last
// uploded must be canceled.
// uploaded must be canceled.
// (These are cancel_uploaded_list, cancellation processing means
// re-uploading the same area.)
//
Expand Down
2 changes: 1 addition & 1 deletion src/fdcache_entity.h
Original file line number Diff line number Diff line change
Expand Up @@ -134,7 +134,7 @@ class FdEntity
return GetOpenCountHasLock();
}
int GetOpenCountHasLock() const REQUIRES(FdEntity::fdent_lock);
// TODO: should thsi require a lock?
// TODO: should this require a lock?
const std::string& GetPath() const { return path; }
bool RenamePath(const std::string& newpath, std::string& fentmapkey);
int GetPhysicalFd() const { return physical_fd; }
Expand Down
16 changes: 8 additions & 8 deletions src/fdcache_fdinfo.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -129,7 +129,7 @@ PseudoFdInfo::PseudoFdInfo(int fd, int open_flags) : pseudo_fd(-1), physical_fd(

PseudoFdInfo::~PseudoFdInfo()
{
Clear(); // call before destrying the mutex
Clear(); // call before destroying the mutex

if(is_lock_init){
int result;
Expand Down Expand Up @@ -686,11 +686,11 @@ bool PseudoFdInfo::CancelAllThreads()
bool PseudoFdInfo::ExtractUploadPartsFromUntreatedArea(const off_t& untreated_start, const off_t& untreated_size, mp_part_list_t& to_upload_list, filepart_list_t& cancel_upload_list, off_t max_mp_size)
{
if(untreated_start < 0 || untreated_size <= 0){
S3FS_PRN_ERR("Paramters are wrong(untreated_start=%lld, untreated_size=%lld).", static_cast<long long int>(untreated_start), static_cast<long long int>(untreated_size));
S3FS_PRN_ERR("Parameters are wrong(untreated_start=%lld, untreated_size=%lld).", static_cast<long long int>(untreated_start), static_cast<long long int>(untreated_size));
return false;
}

// Initiliaze lists
// Initialize lists
to_upload_list.clear();
cancel_upload_list.clear();

Expand Down Expand Up @@ -775,7 +775,7 @@ bool PseudoFdInfo::ExtractUploadPartsFromAllArea(UntreatedParts& untreated_list,
{
AutoLock auto_lock(&upload_list_lock);

// Initiliaze lists
// Initialize lists
to_upload_list.clear();
to_copy_list.clear();
to_download_list.clear();
Expand Down Expand Up @@ -843,13 +843,13 @@ bool PseudoFdInfo::ExtractUploadPartsFromAllArea(UntreatedParts& untreated_list,
// The untreated area exceeds the end of the current area
//

// Ajust untreated area
// Adjust untreated area
tmp_untreated_size = (cur_start + cur_size) - tmp_untreated_start;

// Add ajusted untreated area to cur_untreated_list
// Add adjusted untreated area to cur_untreated_list
cur_untreated_list.emplace_back(tmp_untreated_start, tmp_untreated_size);

// Remove this ajusted untreated area from the area pointed
// Remove this adjusted untreated area from the area pointed
// to by dup_untreated_iter.
dup_untreated_iter->size = (dup_untreated_iter->start + dup_untreated_iter->size) - (cur_start + cur_size);
dup_untreated_iter->start = tmp_untreated_start + tmp_untreated_size;
Expand Down Expand Up @@ -995,7 +995,7 @@ bool PseudoFdInfo::ExtractUploadPartsFromAllArea(UntreatedParts& untreated_list,
((tmp_cur_start + tmp_cur_size) - tmp_cur_untreated_iter->start) >= MIN_MULTIPART_SIZE )
{
//
// Unify to this area to previouse copy area.
// Unify to this area to previous copy area.
//
copy_riter->size += tmp_cur_untreated_iter->start - tmp_cur_start;
S3FS_PRN_DBG("Resize to copy: start=%lld, size=%lld", static_cast<long long int>(copy_riter->start), static_cast<long long int>(copy_riter->size));
Expand Down
6 changes: 3 additions & 3 deletions src/fdcache_page.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -270,7 +270,7 @@ bool PageList::CheckZeroAreaInFile(int fd, off_t start, size_t bytes)
// checkpage: This is one state of the cache file, it is loaded from the stats file.
// sparse_list: This is a list of the results of directly checking the cache file status(HOLE/DATA).
// In the HOLE area, the "loaded" flag of fdpage is false. The DATA area has it set to true.
// fd: opened file discriptor to target cache file.
// fd: opened file descriptor to target cache file.
//
bool PageList::CheckAreaInSparseFile(const struct fdpage& checkpage, const fdpage_list_t& sparse_list, int fd, fdpage_list_t& err_area_list, fdpage_list_t& warn_area_list)
{
Expand Down Expand Up @@ -353,7 +353,7 @@ void PageList::FreeList(fdpage_list_t& list)
list.clear();
}

PageList::PageList(off_t size, bool is_loaded, bool is_modified, bool shrinked) : is_shrink(shrinked)
PageList::PageList(off_t size, bool is_loaded, bool is_modified, bool shrunk) : is_shrink(shrunk)
{
Init(size, is_loaded, is_modified);
}
Expand Down Expand Up @@ -975,7 +975,7 @@ void PageList::Dump() const
{
int cnt = 0;

S3FS_PRN_DBG("pages (shrinked=%s) = {", (is_shrink ? "yes" : "no"));
S3FS_PRN_DBG("pages (shrunk=%s) = {", (is_shrink ? "yes" : "no"));
for(fdpage_list_t::const_iterator iter = pages.begin(); iter != pages.end(); ++iter, ++cnt){
S3FS_PRN_DBG(" [%08d] -> {%014lld - %014lld : %s / %s}", cnt, static_cast<long long int>(iter->offset), static_cast<long long int>(iter->bytes), iter->loaded ? "loaded" : "unloaded", iter->modified ? "modified" : "not modified");
}
Expand Down
4 changes: 2 additions & 2 deletions src/fdcache_page.h
Original file line number Diff line number Diff line change
Expand Up @@ -77,7 +77,7 @@ class PageList

private:
fdpage_list_t pages;
bool is_shrink; // [NOTE] true if it has been shrinked even once
bool is_shrink; // [NOTE] true if it has been shrunk even once

public:
enum class page_status : int8_t {
Expand All @@ -98,7 +98,7 @@ class PageList
public:
static void FreeList(fdpage_list_t& list);

explicit PageList(off_t size = 0, bool is_loaded = false, bool is_modified = false, bool shrinked = false);
explicit PageList(off_t size = 0, bool is_loaded = false, bool is_modified = false, bool shrunk = false);
PageList(const PageList&) = delete;
PageList(PageList&&) = delete;
PageList& operator=(const PageList&) = delete;
Expand Down
6 changes: 3 additions & 3 deletions src/fdcache_untreated.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -64,7 +64,7 @@ bool UntreatedParts::empty()
bool UntreatedParts::AddPart(off_t start, off_t size)
{
if(start < 0 || size <= 0){
S3FS_PRN_ERR("Paramter are wrong(start=%lld, size=%lld).", static_cast<long long int>(start), static_cast<long long int>(size));
S3FS_PRN_ERR("Parameter are wrong(start=%lld, size=%lld).", static_cast<long long int>(start), static_cast<long long int>(size));
return false;
}
AutoLock auto_lock(&untreated_list_lock);
Expand Down Expand Up @@ -102,7 +102,7 @@ bool UntreatedParts::AddPart(off_t start, off_t size)
bool UntreatedParts::RowGetPart(off_t& start, off_t& size, off_t max_size, off_t min_size, bool lastpart) const
{
if(max_size <= 0 || min_size < 0 || max_size < min_size){
S3FS_PRN_ERR("Paramter are wrong(max_size=%lld, min_size=%lld).", static_cast<long long int>(max_size), static_cast<long long int>(min_size));
S3FS_PRN_ERR("Parameter are wrong(max_size=%lld, min_size=%lld).", static_cast<long long int>(max_size), static_cast<long long int>(min_size));
return false;
}
AutoLock auto_lock(&untreated_list_lock);
Expand Down Expand Up @@ -137,7 +137,7 @@ bool UntreatedParts::RowGetPart(off_t& start, off_t& size, off_t max_size, off_t
bool UntreatedParts::ClearParts(off_t start, off_t size)
{
if(start < 0 || size < 0){
S3FS_PRN_ERR("Paramter are wrong(start=%lld, size=%lld).", static_cast<long long int>(start), static_cast<long long int>(size));
S3FS_PRN_ERR("Parameter are wrong(start=%lld, size=%lld).", static_cast<long long int>(start), static_cast<long long int>(size));
return false;
}
AutoLock auto_lock(&untreated_list_lock);
Expand Down
4 changes: 2 additions & 2 deletions src/s3fs.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -2743,7 +2743,7 @@ static int s3fs_truncate(const char* _path, off_t size)
//
// [NOTICE]
// FdManager::Open() ignores changes that reduce the file size for the
// file you are editing. However, if user opens only onece, edit it,
// file you are editing. However, if user opens only once, edit it,
// and then shrink the file, it should be done.
// When this function is called, the file is already open by FUSE or
// some other operation. Therefore, if the number of open files is 1,
Expand Down Expand Up @@ -3151,7 +3151,7 @@ static int s3fs_release(const char* _path, struct fuse_file_info* fi)
// check - for debug
if(S3fsLog::IsS3fsLogDbg()){
if(FdManager::HasOpenEntityFd(path)){
S3FS_PRN_DBG("file(%s) is still opened(another pseudo fd is opend).", path);
S3FS_PRN_DBG("file(%s) is still opened(another pseudo fd is opened).", path);
}
}
S3FS_MALLOCTRIM(0);
Expand Down
6 changes: 3 additions & 3 deletions src/s3fs_cred.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -48,7 +48,7 @@ static constexpr char DEFAULT_AWS_PROFILE_NAME[] = "default";
//
// detail=false ex. "Custom AWS Credential Library - v1.0.0"
// detail=true ex. "Custom AWS Credential Library - v1.0.0
// s3fs-fuse credential I/F library for S3 compatible strage X.
// s3fs-fuse credential I/F library for S3 compatible storage X.
// Copyright(C) 2022 Foo"
//
const char* VersionS3fsCredential(bool detail)
Expand Down Expand Up @@ -1119,7 +1119,7 @@ bool S3fsCred::CheckIAMCredentialUpdate(std::string* access_key_id, std::string*
AutoLock auto_lock(&token_lock);

if(IsIBMIAMAuth() || IsSetExtCredLib() || is_ecs || IsSetIAMRole()){
if(AWSAccessTokenExpire < (time(nullptr) + S3fsCred::IAM_EXPIRE_MERGIN)){
if(AWSAccessTokenExpire < (time(nullptr) + S3fsCred::IAM_EXPIRE_MERGING)){
S3FS_PRN_INFO("IAM Access Token refreshing...");

// update
Expand Down Expand Up @@ -1579,7 +1579,7 @@ bool S3fsCred::CheckAllParams()
// Load and Initialize external credential library
if(IsSetExtCredLib() || IsSetExtCredLibOpts()){
if(!IsSetExtCredLib()){
S3FS_PRN_EXIT("The \"credlib_opts\"(%s) is specifyed but \"credlib\" option is not specified.", credlib_opts.c_str());
S3FS_PRN_EXIT("The \"credlib_opts\"(%s) is specified but \"credlib\" option is not specified.", credlib_opts.c_str());
return false;
}

Expand Down
2 changes: 1 addition & 1 deletion src/s3fs_cred.h
Original file line number Diff line number Diff line change
Expand Up @@ -49,7 +49,7 @@ class S3fsCred
static constexpr char AWS_ACCESSKEYID[] = "AWSAccessKeyId";
static constexpr char AWS_SECRETKEY[] = "AWSSecretKey";

static constexpr int IAM_EXPIRE_MERGIN = 20 * 60; // update timing
static constexpr int IAM_EXPIRE_MERGING = 20 * 60; // update timing
static constexpr char ECS_IAM_ENV_VAR[] = "AWS_CONTAINER_CREDENTIALS_RELATIVE_URI";
static constexpr char IAMCRED_ACCESSKEYID[] = "AccessKeyId";
static constexpr char IAMCRED_SECRETACCESSKEY[] = "SecretAccessKey";
Expand Down
2 changes: 1 addition & 1 deletion src/s3fs_help.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -359,7 +359,7 @@ static constexpr char help_string[] =
"\n"
" max_thread_count (default is \"5\")\n"
" - Specifies the number of threads waiting for stream uploads.\n"
" Note that this option and Streamm Upload are still experimental\n"
" Note that this option and Stream Upload are still experimental\n"
" and subject to change in the future.\n"
" This option will be merged with \"parallel_count\" in the future.\n"
"\n"
Expand Down
2 changes: 1 addition & 1 deletion src/threadpoolman.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -100,7 +100,7 @@ void* ThreadPoolMan::Worker(void* arg)

void* retval = param.pfunc(param.args);
if(nullptr != retval){
S3FS_PRN_WARN("The instruction function returned with somthign error code(%ld).", reinterpret_cast<long>(retval));
S3FS_PRN_WARN("The instruction function returned with something error code(%ld).", reinterpret_cast<long>(retval));
}
if(param.psem){
param.psem->post();
Expand Down
6 changes: 3 additions & 3 deletions test/integration-test-main.sh
Original file line number Diff line number Diff line change
Expand Up @@ -122,7 +122,7 @@ function test_truncate_shrink_read_file {
# create file
dd if=/dev/urandom of="${TEST_TEXT_FILE}" bs="${init_size}" count=1

# truncate(shrink) file and read it before flusing
# truncate(shrink) file and read it before flushing
../../truncate_read_file "${TEST_TEXT_FILE}" "${shrink_size}"

# check file size
Expand Down Expand Up @@ -2551,9 +2551,9 @@ function test_not_boundary_writes {
# Part number 2: 10,485,760 - 20,971,519 (size = 10MB)
# Part number 3: 20,971,520 - 26,214,399 (size = 5MB)
#
local BOUNDAY_TEST_FILE_SIZE; BOUNDAY_TEST_FILE_SIZE=$((BIG_FILE_BLOCK_SIZE * BIG_FILE_COUNT))
local BOUNDARY_TEST_FILE_SIZE; BOUNDARY_TEST_FILE_SIZE=$((BIG_FILE_BLOCK_SIZE * BIG_FILE_COUNT))

../../junk_data "${BOUNDAY_TEST_FILE_SIZE}" > "${TEST_TEXT_FILE}"
../../junk_data "${BOUNDARY_TEST_FILE_SIZE}" > "${TEST_TEXT_FILE}"

#
# Write in First boundary
Expand Down
2 changes: 1 addition & 1 deletion test/sample_delcache.sh
Original file line number Diff line number Diff line change
Expand Up @@ -20,7 +20,7 @@
#

#
# This is unsupport sample deleting cache files script.
# This is unsupported sample deleting cache files script.
# So s3fs's local cache files(stats and objects) grow up,
# you need to delete these.
# This script deletes these files with total size limit
Expand Down

0 comments on commit fa807a5

Please sign in to comment.