| File: | home/bhubbard/working/src/ceph/src/rocksdb/util/sst_file_manager_impl.cc |
| Warning: | line 276, column 24 The left operand of '<' is a garbage value |
[?] Use j/k keys for keyboard navigation
| 1 | // Copyright (c) 2011-present, Facebook, Inc. All rights reserved. | |||
| 2 | // This source code is licensed under both the GPLv2 (found in the | |||
| 3 | // COPYING file in the root directory) and Apache 2.0 License | |||
| 4 | // (found in the LICENSE.Apache file in the root directory). | |||
| 5 | ||||
| 6 | #include "util/sst_file_manager_impl.h" | |||
| 7 | ||||
| 8 | #ifndef __STDC_FORMAT_MACROS | |||
| 9 | #define __STDC_FORMAT_MACROS | |||
| 10 | #endif | |||
| 11 | ||||
| 12 | #include <inttypes.h> | |||
| 13 | #include <vector> | |||
| 14 | ||||
| 15 | #include "db/db_impl.h" | |||
| 16 | #include "port/port.h" | |||
| 17 | #include "rocksdb/env.h" | |||
| 18 | #include "rocksdb/sst_file_manager.h" | |||
| 19 | #include "util/mutexlock.h" | |||
| 20 | #include "util/sync_point.h" | |||
| 21 | ||||
| 22 | namespace rocksdb { | |||
| 23 | ||||
| 24 | #ifndef ROCKSDB_LITE | |||
| 25 | SstFileManagerImpl::SstFileManagerImpl(Env* env, std::shared_ptr<Logger> logger, | |||
| 26 | int64_t rate_bytes_per_sec, | |||
| 27 | double max_trash_db_ratio, | |||
| 28 | uint64_t bytes_max_delete_chunk) | |||
| 29 | : env_(env), | |||
| 30 | logger_(logger), | |||
| 31 | total_files_size_(0), | |||
| 32 | in_progress_files_size_(0), | |||
| 33 | compaction_buffer_size_(0), | |||
| 34 | cur_compactions_reserved_size_(0), | |||
| 35 | max_allowed_space_(0), | |||
| 36 | delete_scheduler_(env, rate_bytes_per_sec, logger.get(), this, | |||
| 37 | max_trash_db_ratio, bytes_max_delete_chunk), | |||
| 38 | cv_(&mu_), | |||
| 39 | closing_(false), | |||
| 40 | bg_thread_(nullptr), | |||
| 41 | reserved_disk_buffer_(0), | |||
| 42 | free_space_trigger_(0), | |||
| 43 | cur_instance_(nullptr) { | |||
| 44 | } | |||
| 45 | ||||
| 46 | SstFileManagerImpl::~SstFileManagerImpl() { | |||
| 47 | Close(); | |||
| 48 | } | |||
| 49 | ||||
| 50 | void SstFileManagerImpl::Close() { | |||
| 51 | { | |||
| 52 | MutexLock l(&mu_); | |||
| 53 | if (closing_) { | |||
| 54 | return; | |||
| 55 | } | |||
| 56 | closing_ = true; | |||
| 57 | cv_.SignalAll(); | |||
| 58 | } | |||
| 59 | if (bg_thread_) { | |||
| 60 | bg_thread_->join(); | |||
| 61 | } | |||
| 62 | } | |||
| 63 | ||||
| 64 | Status SstFileManagerImpl::OnAddFile(const std::string& file_path, | |||
| 65 | bool compaction) { | |||
| 66 | uint64_t file_size; | |||
| 67 | Status s = env_->GetFileSize(file_path, &file_size); | |||
| 68 | if (s.ok()) { | |||
| 69 | MutexLock l(&mu_); | |||
| 70 | OnAddFileImpl(file_path, file_size, compaction); | |||
| 71 | } | |||
| 72 | TEST_SYNC_POINT("SstFileManagerImpl::OnAddFile"); | |||
| 73 | return s; | |||
| 74 | } | |||
| 75 | ||||
| 76 | Status SstFileManagerImpl::OnDeleteFile(const std::string& file_path) { | |||
| 77 | { | |||
| 78 | MutexLock l(&mu_); | |||
| 79 | OnDeleteFileImpl(file_path); | |||
| 80 | } | |||
| 81 | TEST_SYNC_POINT("SstFileManagerImpl::OnDeleteFile"); | |||
| 82 | return Status::OK(); | |||
| 83 | } | |||
| 84 | ||||
| 85 | void SstFileManagerImpl::OnCompactionCompletion(Compaction* c) { | |||
| 86 | MutexLock l(&mu_); | |||
| 87 | uint64_t size_added_by_compaction = 0; | |||
| 88 | for (size_t i = 0; i < c->num_input_levels(); i++) { | |||
| 89 | for (size_t j = 0; j < c->num_input_files(i); j++) { | |||
| 90 | FileMetaData* filemeta = c->input(i, j); | |||
| 91 | size_added_by_compaction += filemeta->fd.GetFileSize(); | |||
| 92 | } | |||
| 93 | } | |||
| 94 | cur_compactions_reserved_size_ -= size_added_by_compaction; | |||
| 95 | ||||
| 96 | auto new_files = c->edit()->GetNewFiles(); | |||
| 97 | for (auto& new_file : new_files) { | |||
| 98 | auto fn = TableFileName(c->immutable_cf_options()->cf_paths, | |||
| 99 | new_file.second.fd.GetNumber(), | |||
| 100 | new_file.second.fd.GetPathId()); | |||
| 101 | if (in_progress_files_.find(fn) != in_progress_files_.end()) { | |||
| 102 | auto tracked_file = tracked_files_.find(fn); | |||
| 103 | assert(tracked_file != tracked_files_.end())(static_cast<void> (0)); | |||
| 104 | in_progress_files_size_ -= tracked_file->second; | |||
| 105 | in_progress_files_.erase(fn); | |||
| 106 | } | |||
| 107 | } | |||
| 108 | } | |||
| 109 | ||||
| 110 | Status SstFileManagerImpl::OnMoveFile(const std::string& old_path, | |||
| 111 | const std::string& new_path, | |||
| 112 | uint64_t* file_size) { | |||
| 113 | { | |||
| 114 | MutexLock l(&mu_); | |||
| 115 | if (file_size != nullptr) { | |||
| 116 | *file_size = tracked_files_[old_path]; | |||
| 117 | } | |||
| 118 | OnAddFileImpl(new_path, tracked_files_[old_path], false); | |||
| 119 | OnDeleteFileImpl(old_path); | |||
| 120 | } | |||
| 121 | TEST_SYNC_POINT("SstFileManagerImpl::OnMoveFile"); | |||
| 122 | return Status::OK(); | |||
| 123 | } | |||
| 124 | ||||
| 125 | void SstFileManagerImpl::SetMaxAllowedSpaceUsage(uint64_t max_allowed_space) { | |||
| 126 | MutexLock l(&mu_); | |||
| 127 | max_allowed_space_ = max_allowed_space; | |||
| 128 | } | |||
| 129 | ||||
| 130 | void SstFileManagerImpl::SetCompactionBufferSize( | |||
| 131 | uint64_t compaction_buffer_size) { | |||
| 132 | MutexLock l(&mu_); | |||
| 133 | compaction_buffer_size_ = compaction_buffer_size; | |||
| 134 | } | |||
| 135 | ||||
| 136 | bool SstFileManagerImpl::IsMaxAllowedSpaceReached() { | |||
| 137 | MutexLock l(&mu_); | |||
| 138 | if (max_allowed_space_ <= 0) { | |||
| 139 | return false; | |||
| 140 | } | |||
| 141 | return total_files_size_ >= max_allowed_space_; | |||
| 142 | } | |||
| 143 | ||||
| 144 | bool SstFileManagerImpl::IsMaxAllowedSpaceReachedIncludingCompactions() { | |||
| 145 | MutexLock l(&mu_); | |||
| 146 | if (max_allowed_space_ <= 0) { | |||
| 147 | return false; | |||
| 148 | } | |||
| 149 | return total_files_size_ + cur_compactions_reserved_size_ >= | |||
| 150 | max_allowed_space_; | |||
| 151 | } | |||
| 152 | ||||
| 153 | bool SstFileManagerImpl::EnoughRoomForCompaction( | |||
| 154 | ColumnFamilyData* cfd, const std::vector<CompactionInputFiles>& inputs, | |||
| 155 | Status bg_error) { | |||
| 156 | MutexLock l(&mu_); | |||
| 157 | uint64_t size_added_by_compaction = 0; | |||
| 158 | // First check if we even have the space to do the compaction | |||
| 159 | for (size_t i = 0; i < inputs.size(); i++) { | |||
| 160 | for (size_t j = 0; j < inputs[i].size(); j++) { | |||
| 161 | FileMetaData* filemeta = inputs[i][j]; | |||
| 162 | size_added_by_compaction += filemeta->fd.GetFileSize(); | |||
| 163 | } | |||
| 164 | } | |||
| 165 | ||||
| 166 | // Update cur_compactions_reserved_size_ so concurrent compaction | |||
| 167 | // don't max out space | |||
| 168 | size_t needed_headroom = | |||
| 169 | cur_compactions_reserved_size_ + size_added_by_compaction + | |||
| 170 | compaction_buffer_size_; | |||
| 171 | if (max_allowed_space_ != 0 && | |||
| 172 | (needed_headroom + total_files_size_ > max_allowed_space_)) { | |||
| 173 | return false; | |||
| 174 | } | |||
| 175 | ||||
| 176 | // Implement more aggressive checks only if this DB instance has already | |||
| 177 | // seen a NoSpace() error. This is tin order to contain a single potentially | |||
| 178 | // misbehaving DB instance and prevent it from slowing down compactions of | |||
| 179 | // other DB instances | |||
| 180 | if (CheckFreeSpace() && bg_error == Status::NoSpace()) { | |||
| 181 | auto fn = | |||
| 182 | TableFileName(cfd->ioptions()->cf_paths, inputs[0][0]->fd.GetNumber(), | |||
| 183 | inputs[0][0]->fd.GetPathId()); | |||
| 184 | uint64_t free_space = 0; | |||
| 185 | env_->GetFreeSpace(fn, &free_space); | |||
| 186 | // needed_headroom is based on current size reserved by compactions, | |||
| 187 | // minus any files created by running compactions as they would count | |||
| 188 | // against the reserved size. If user didn't specify any compaction | |||
| 189 | // buffer, add reserved_disk_buffer_ that's calculated by default so the | |||
| 190 | // compaction doesn't end up leaving nothing for logs and flush SSTs | |||
| 191 | if (compaction_buffer_size_ == 0) { | |||
| 192 | needed_headroom += reserved_disk_buffer_; | |||
| 193 | } | |||
| 194 | needed_headroom -= in_progress_files_size_; | |||
| 195 | if (free_space < needed_headroom + size_added_by_compaction) { | |||
| 196 | // We hit the condition of not enough disk space | |||
| 197 | ROCKS_LOG_ERROR(logger_,rocksdb::Log(InfoLogLevel::ERROR_LEVEL, logger_, ("[%s:" "201" "] " "free space [%" "l" "u" " bytes] is less than " "needed headroom [%" "zu" " bytes]\n"), RocksLogShorterFileName("/home/bhubbard/working/src/ceph/src/rocksdb/util/sst_file_manager_impl.cc" ), free_space, needed_headroom) | |||
| 198 | "free space [%" PRIu64rocksdb::Log(InfoLogLevel::ERROR_LEVEL, logger_, ("[%s:" "201" "] " "free space [%" "l" "u" " bytes] is less than " "needed headroom [%" "zu" " bytes]\n"), RocksLogShorterFileName("/home/bhubbard/working/src/ceph/src/rocksdb/util/sst_file_manager_impl.cc" ), free_space, needed_headroom) | |||
| 199 | " bytes] is less than "rocksdb::Log(InfoLogLevel::ERROR_LEVEL, logger_, ("[%s:" "201" "] " "free space [%" "l" "u" " bytes] is less than " "needed headroom [%" "zu" " bytes]\n"), RocksLogShorterFileName("/home/bhubbard/working/src/ceph/src/rocksdb/util/sst_file_manager_impl.cc" ), free_space, needed_headroom) | |||
| 200 | "needed headroom [%" ROCKSDB_PRIszt " bytes]\n",rocksdb::Log(InfoLogLevel::ERROR_LEVEL, logger_, ("[%s:" "201" "] " "free space [%" "l" "u" " bytes] is less than " "needed headroom [%" "zu" " bytes]\n"), RocksLogShorterFileName("/home/bhubbard/working/src/ceph/src/rocksdb/util/sst_file_manager_impl.cc" ), free_space, needed_headroom) | |||
| 201 | free_space, needed_headroom)rocksdb::Log(InfoLogLevel::ERROR_LEVEL, logger_, ("[%s:" "201" "] " "free space [%" "l" "u" " bytes] is less than " "needed headroom [%" "zu" " bytes]\n"), RocksLogShorterFileName("/home/bhubbard/working/src/ceph/src/rocksdb/util/sst_file_manager_impl.cc" ), free_space, needed_headroom); | |||
| 202 | return false; | |||
| 203 | } | |||
| 204 | } | |||
| 205 | ||||
| 206 | cur_compactions_reserved_size_ += size_added_by_compaction; | |||
| 207 | // Take a snapshot of cur_compactions_reserved_size_ for when we encounter | |||
| 208 | // a NoSpace error. | |||
| 209 | free_space_trigger_ = cur_compactions_reserved_size_; | |||
| 210 | return true; | |||
| 211 | } | |||
| 212 | ||||
| 213 | uint64_t SstFileManagerImpl::GetCompactionsReservedSize() { | |||
| 214 | MutexLock l(&mu_); | |||
| 215 | return cur_compactions_reserved_size_; | |||
| 216 | } | |||
| 217 | ||||
| 218 | uint64_t SstFileManagerImpl::GetTotalSize() { | |||
| 219 | MutexLock l(&mu_); | |||
| 220 | return total_files_size_; | |||
| 221 | } | |||
| 222 | ||||
| 223 | std::unordered_map<std::string, uint64_t> | |||
| 224 | SstFileManagerImpl::GetTrackedFiles() { | |||
| 225 | MutexLock l(&mu_); | |||
| 226 | return tracked_files_; | |||
| 227 | } | |||
| 228 | ||||
| 229 | int64_t SstFileManagerImpl::GetDeleteRateBytesPerSecond() { | |||
| 230 | return delete_scheduler_.GetRateBytesPerSecond(); | |||
| 231 | } | |||
| 232 | ||||
| 233 | void SstFileManagerImpl::SetDeleteRateBytesPerSecond(int64_t delete_rate) { | |||
| 234 | return delete_scheduler_.SetRateBytesPerSecond(delete_rate); | |||
| 235 | } | |||
| 236 | ||||
| 237 | double SstFileManagerImpl::GetMaxTrashDBRatio() { | |||
| 238 | return delete_scheduler_.GetMaxTrashDBRatio(); | |||
| 239 | } | |||
| 240 | ||||
| 241 | void SstFileManagerImpl::SetMaxTrashDBRatio(double r) { | |||
| 242 | return delete_scheduler_.SetMaxTrashDBRatio(r); | |||
| 243 | } | |||
| 244 | ||||
| 245 | uint64_t SstFileManagerImpl::GetTotalTrashSize() { | |||
| 246 | return delete_scheduler_.GetTotalTrashSize(); | |||
| 247 | } | |||
| 248 | ||||
| 249 | void SstFileManagerImpl::ReserveDiskBuffer(uint64_t size, | |||
| 250 | const std::string& path) { | |||
| 251 | MutexLock l(&mu_); | |||
| 252 | ||||
| 253 | reserved_disk_buffer_ += size; | |||
| 254 | if (path_.empty()) { | |||
| 255 | path_ = path; | |||
| 256 | } | |||
| 257 | } | |||
| 258 | ||||
| 259 | void SstFileManagerImpl::ClearError() { | |||
| 260 | while (true) { | |||
| ||||
| 261 | MutexLock l(&mu_); | |||
| 262 | ||||
| 263 | if (closing_) { | |||
| 264 | return; | |||
| 265 | } | |||
| 266 | ||||
| 267 | uint64_t free_space; | |||
| 268 | Status s = env_->GetFreeSpace(path_, &free_space); | |||
| 269 | if (s.ok()) { | |||
| 270 | // In case of multi-DB instances, some of them may have experienced a | |||
| 271 | // soft error and some a hard error. In the SstFileManagerImpl, a hard | |||
| 272 | // error will basically override previously reported soft errors. Once | |||
| 273 | // we clear the hard error, we don't keep track of previous errors for | |||
| 274 | // now | |||
| 275 | if (bg_err_.severity() == Status::Severity::kHardError) { | |||
| 276 | if (free_space < reserved_disk_buffer_) { | |||
| ||||
| 277 | ROCKS_LOG_ERROR(logger_,rocksdb::Log(InfoLogLevel::ERROR_LEVEL, logger_, ("[%s:" "281" "] " "free space [%" "l" "u" " bytes] is less than " "required disk buffer [%" "l" "u" " bytes]\n"), RocksLogShorterFileName("/home/bhubbard/working/src/ceph/src/rocksdb/util/sst_file_manager_impl.cc" ), free_space, reserved_disk_buffer_) | |||
| 278 | "free space [%" PRIu64rocksdb::Log(InfoLogLevel::ERROR_LEVEL, logger_, ("[%s:" "281" "] " "free space [%" "l" "u" " bytes] is less than " "required disk buffer [%" "l" "u" " bytes]\n"), RocksLogShorterFileName("/home/bhubbard/working/src/ceph/src/rocksdb/util/sst_file_manager_impl.cc" ), free_space, reserved_disk_buffer_) | |||
| 279 | " bytes] is less than "rocksdb::Log(InfoLogLevel::ERROR_LEVEL, logger_, ("[%s:" "281" "] " "free space [%" "l" "u" " bytes] is less than " "required disk buffer [%" "l" "u" " bytes]\n"), RocksLogShorterFileName("/home/bhubbard/working/src/ceph/src/rocksdb/util/sst_file_manager_impl.cc" ), free_space, reserved_disk_buffer_) | |||
| 280 | "required disk buffer [%" PRIu64 " bytes]\n",rocksdb::Log(InfoLogLevel::ERROR_LEVEL, logger_, ("[%s:" "281" "] " "free space [%" "l" "u" " bytes] is less than " "required disk buffer [%" "l" "u" " bytes]\n"), RocksLogShorterFileName("/home/bhubbard/working/src/ceph/src/rocksdb/util/sst_file_manager_impl.cc" ), free_space, reserved_disk_buffer_) | |||
| 281 | free_space, reserved_disk_buffer_)rocksdb::Log(InfoLogLevel::ERROR_LEVEL, logger_, ("[%s:" "281" "] " "free space [%" "l" "u" " bytes] is less than " "required disk buffer [%" "l" "u" " bytes]\n"), RocksLogShorterFileName("/home/bhubbard/working/src/ceph/src/rocksdb/util/sst_file_manager_impl.cc" ), free_space, reserved_disk_buffer_); | |||
| 282 | ROCKS_LOG_ERROR(logger_, "Cannot clear hard error\n")rocksdb::Log(InfoLogLevel::ERROR_LEVEL, logger_, ("[%s:" "282" "] " "Cannot clear hard error\n"), RocksLogShorterFileName("/home/bhubbard/working/src/ceph/src/rocksdb/util/sst_file_manager_impl.cc" )); | |||
| 283 | s = Status::NoSpace(); | |||
| 284 | } | |||
| 285 | } else if (bg_err_.severity() == Status::Severity::kSoftError) { | |||
| 286 | if (free_space < free_space_trigger_) { | |||
| 287 | ROCKS_LOG_WARN(logger_,rocksdb::Log(InfoLogLevel::WARN_LEVEL, logger_, ("[%s:" "292" "] " "free space [%" "l" "u" " bytes] is less than " "free space for compaction trigger [%" "l" "u" " bytes]\n"), RocksLogShorterFileName("/home/bhubbard/working/src/ceph/src/rocksdb/util/sst_file_manager_impl.cc" ), free_space, free_space_trigger_) | |||
| 288 | "free space [%" PRIu64rocksdb::Log(InfoLogLevel::WARN_LEVEL, logger_, ("[%s:" "292" "] " "free space [%" "l" "u" " bytes] is less than " "free space for compaction trigger [%" "l" "u" " bytes]\n"), RocksLogShorterFileName("/home/bhubbard/working/src/ceph/src/rocksdb/util/sst_file_manager_impl.cc" ), free_space, free_space_trigger_) | |||
| 289 | " bytes] is less than "rocksdb::Log(InfoLogLevel::WARN_LEVEL, logger_, ("[%s:" "292" "] " "free space [%" "l" "u" " bytes] is less than " "free space for compaction trigger [%" "l" "u" " bytes]\n"), RocksLogShorterFileName("/home/bhubbard/working/src/ceph/src/rocksdb/util/sst_file_manager_impl.cc" ), free_space, free_space_trigger_) | |||
| 290 | "free space for compaction trigger [%" PRIu64rocksdb::Log(InfoLogLevel::WARN_LEVEL, logger_, ("[%s:" "292" "] " "free space [%" "l" "u" " bytes] is less than " "free space for compaction trigger [%" "l" "u" " bytes]\n"), RocksLogShorterFileName("/home/bhubbard/working/src/ceph/src/rocksdb/util/sst_file_manager_impl.cc" ), free_space, free_space_trigger_) | |||
| 291 | " bytes]\n",rocksdb::Log(InfoLogLevel::WARN_LEVEL, logger_, ("[%s:" "292" "] " "free space [%" "l" "u" " bytes] is less than " "free space for compaction trigger [%" "l" "u" " bytes]\n"), RocksLogShorterFileName("/home/bhubbard/working/src/ceph/src/rocksdb/util/sst_file_manager_impl.cc" ), free_space, free_space_trigger_) | |||
| 292 | free_space, free_space_trigger_)rocksdb::Log(InfoLogLevel::WARN_LEVEL, logger_, ("[%s:" "292" "] " "free space [%" "l" "u" " bytes] is less than " "free space for compaction trigger [%" "l" "u" " bytes]\n"), RocksLogShorterFileName("/home/bhubbard/working/src/ceph/src/rocksdb/util/sst_file_manager_impl.cc" ), free_space, free_space_trigger_); | |||
| 293 | ROCKS_LOG_WARN(logger_, "Cannot clear soft error\n")rocksdb::Log(InfoLogLevel::WARN_LEVEL, logger_, ("[%s:" "293" "] " "Cannot clear soft error\n"), RocksLogShorterFileName("/home/bhubbard/working/src/ceph/src/rocksdb/util/sst_file_manager_impl.cc" )); | |||
| 294 | s = Status::NoSpace(); | |||
| 295 | } | |||
| 296 | } | |||
| 297 | } | |||
| 298 | ||||
| 299 | // Someone could have called CancelErrorRecovery() and the list could have | |||
| 300 | // become empty, so check again here | |||
| 301 | if (s.ok() && !error_handler_list_.empty()) { | |||
| 302 | auto error_handler = error_handler_list_.front(); | |||
| 303 | // Since we will release the mutex, set cur_instance_ to signal to the | |||
| 304 | // shutdown thread, if it calls // CancelErrorRecovery() the meantime, | |||
| 305 | // to indicate that this DB instance is busy. The DB instance is | |||
| 306 | // guaranteed to not be deleted before RecoverFromBGError() returns, | |||
| 307 | // since the ErrorHandler::recovery_in_prog_ flag would be true | |||
| 308 | cur_instance_ = error_handler; | |||
| 309 | mu_.Unlock(); | |||
| 310 | s = error_handler->RecoverFromBGError(); | |||
| 311 | mu_.Lock(); | |||
| 312 | // The DB instance might have been deleted while we were | |||
| 313 | // waiting for the mutex, so check cur_instance_ to make sure its | |||
| 314 | // still non-null | |||
| 315 | if (cur_instance_) { | |||
| 316 | // Check for error again, since the instance may have recovered but | |||
| 317 | // immediately got another error. If that's the case, and the new | |||
| 318 | // error is also a NoSpace() non-fatal error, leave the instance in | |||
| 319 | // the list | |||
| 320 | Status err = cur_instance_->GetBGError(); | |||
| 321 | if (s.ok() && err == Status::NoSpace() && | |||
| 322 | err.severity() < Status::Severity::kFatalError) { | |||
| 323 | s = err; | |||
| 324 | } | |||
| 325 | cur_instance_ = nullptr; | |||
| 326 | } | |||
| 327 | ||||
| 328 | if (s.ok() || s.IsShutdownInProgress() || | |||
| 329 | (!s.ok() && s.severity() >= Status::Severity::kFatalError)) { | |||
| 330 | // If shutdown is in progress, abandon this handler instance | |||
| 331 | // and continue with the others | |||
| 332 | error_handler_list_.pop_front(); | |||
| 333 | } | |||
| 334 | } | |||
| 335 | ||||
| 336 | if (!error_handler_list_.empty()) { | |||
| 337 | // If there are more instances to be recovered, reschedule after 5 | |||
| 338 | // seconds | |||
| 339 | int64_t wait_until = env_->NowMicros() + 5000000; | |||
| 340 | cv_.TimedWait(wait_until); | |||
| 341 | } | |||
| 342 | ||||
| 343 | // Check again for error_handler_list_ empty, as a DB instance shutdown | |||
| 344 | // could have removed it from the queue while we were in timed wait | |||
| 345 | if (error_handler_list_.empty()) { | |||
| 346 | ROCKS_LOG_INFO(logger_, "Clearing error\n")rocksdb::Log(InfoLogLevel::INFO_LEVEL, logger_, ("[%s:" "346" "] " "Clearing error\n"), RocksLogShorterFileName("/home/bhubbard/working/src/ceph/src/rocksdb/util/sst_file_manager_impl.cc" )); | |||
| 347 | bg_err_ = Status::OK(); | |||
| 348 | return; | |||
| 349 | } | |||
| 350 | } | |||
| 351 | } | |||
| 352 | ||||
| 353 | void SstFileManagerImpl::StartErrorRecovery(ErrorHandler* handler, | |||
| 354 | Status bg_error) { | |||
| 355 | MutexLock l(&mu_); | |||
| 356 | if (bg_error.severity() == Status::Severity::kSoftError) { | |||
| 357 | if (bg_err_.ok()) { | |||
| 358 | // Setting bg_err_ basically means we're in degraded mode | |||
| 359 | // Assume that all pending compactions will fail similarly. The trigger | |||
| 360 | // for clearing this condition is set to current compaction reserved | |||
| 361 | // size, so we stop checking disk space available in | |||
| 362 | // EnoughRoomForCompaction once this much free space is available | |||
| 363 | bg_err_ = bg_error; | |||
| 364 | } | |||
| 365 | } else if (bg_error.severity() == Status::Severity::kHardError) { | |||
| 366 | bg_err_ = bg_error; | |||
| 367 | } else { | |||
| 368 | assert(false)(static_cast<void> (0)); | |||
| 369 | } | |||
| 370 | ||||
| 371 | // If this is the first instance of this error, kick of a thread to poll | |||
| 372 | // and recover from this condition | |||
| 373 | if (error_handler_list_.empty()) { | |||
| 374 | error_handler_list_.push_back(handler); | |||
| 375 | // Release lock before calling join. Its ok to do so because | |||
| 376 | // error_handler_list_ is now non-empty, so no other invocation of this | |||
| 377 | // function will execute this piece of code | |||
| 378 | mu_.Unlock(); | |||
| 379 | if (bg_thread_) { | |||
| 380 | bg_thread_->join(); | |||
| 381 | } | |||
| 382 | // Start a new thread. The previous one would have exited. | |||
| 383 | bg_thread_.reset(new port::Thread(&SstFileManagerImpl::ClearError, this)); | |||
| 384 | mu_.Lock(); | |||
| 385 | } else { | |||
| 386 | // Check if this DB instance is already in the list | |||
| 387 | for (auto iter = error_handler_list_.begin(); | |||
| 388 | iter != error_handler_list_.end(); ++iter) { | |||
| 389 | if ((*iter) == handler) { | |||
| 390 | return; | |||
| 391 | } | |||
| 392 | } | |||
| 393 | error_handler_list_.push_back(handler); | |||
| 394 | } | |||
| 395 | } | |||
| 396 | ||||
| 397 | bool SstFileManagerImpl::CancelErrorRecovery(ErrorHandler* handler) { | |||
| 398 | MutexLock l(&mu_); | |||
| 399 | ||||
| 400 | if (cur_instance_ == handler) { | |||
| 401 | // This instance is currently busy attempting to recover | |||
| 402 | // Nullify it so the recovery thread doesn't attempt to access it again | |||
| 403 | cur_instance_ = nullptr; | |||
| 404 | return false; | |||
| 405 | } | |||
| 406 | ||||
| 407 | for (auto iter = error_handler_list_.begin(); | |||
| 408 | iter != error_handler_list_.end(); ++iter) { | |||
| 409 | if ((*iter) == handler) { | |||
| 410 | error_handler_list_.erase(iter); | |||
| 411 | return true; | |||
| 412 | } | |||
| 413 | } | |||
| 414 | return false; | |||
| 415 | } | |||
| 416 | ||||
| 417 | Status SstFileManagerImpl::ScheduleFileDeletion( | |||
| 418 | const std::string& file_path, const std::string& path_to_sync, | |||
| 419 | const bool force_bg) { | |||
| 420 | TEST_SYNC_POINT("SstFileManagerImpl::ScheduleFileDeletion"); | |||
| 421 | return delete_scheduler_.DeleteFile(file_path, path_to_sync, | |||
| 422 | force_bg); | |||
| 423 | } | |||
| 424 | ||||
| 425 | void SstFileManagerImpl::WaitForEmptyTrash() { | |||
| 426 | delete_scheduler_.WaitForEmptyTrash(); | |||
| 427 | } | |||
| 428 | ||||
| 429 | void SstFileManagerImpl::OnAddFileImpl(const std::string& file_path, | |||
| 430 | uint64_t file_size, bool compaction) { | |||
| 431 | auto tracked_file = tracked_files_.find(file_path); | |||
| 432 | if (tracked_file != tracked_files_.end()) { | |||
| 433 | // File was added before, we will just update the size | |||
| 434 | assert(!compaction)(static_cast<void> (0)); | |||
| 435 | total_files_size_ -= tracked_file->second; | |||
| 436 | total_files_size_ += file_size; | |||
| 437 | cur_compactions_reserved_size_ -= file_size; | |||
| 438 | } else { | |||
| 439 | total_files_size_ += file_size; | |||
| 440 | if (compaction) { | |||
| 441 | // Keep track of the size of files created by in-progress compactions. | |||
| 442 | // When calculating whether there's enough headroom for new compactions, | |||
| 443 | // this will be subtracted from cur_compactions_reserved_size_. | |||
| 444 | // Otherwise, compactions will be double counted. | |||
| 445 | in_progress_files_size_ += file_size; | |||
| 446 | in_progress_files_.insert(file_path); | |||
| 447 | } | |||
| 448 | } | |||
| 449 | tracked_files_[file_path] = file_size; | |||
| 450 | } | |||
| 451 | ||||
| 452 | void SstFileManagerImpl::OnDeleteFileImpl(const std::string& file_path) { | |||
| 453 | auto tracked_file = tracked_files_.find(file_path); | |||
| 454 | if (tracked_file == tracked_files_.end()) { | |||
| 455 | // File is not tracked | |||
| 456 | assert(in_progress_files_.find(file_path) == in_progress_files_.end())(static_cast<void> (0)); | |||
| 457 | return; | |||
| 458 | } | |||
| 459 | ||||
| 460 | total_files_size_ -= tracked_file->second; | |||
| 461 | // Check if it belonged to an in-progress compaction | |||
| 462 | if (in_progress_files_.find(file_path) != in_progress_files_.end()) { | |||
| 463 | in_progress_files_size_ -= tracked_file->second; | |||
| 464 | in_progress_files_.erase(file_path); | |||
| 465 | } | |||
| 466 | tracked_files_.erase(tracked_file); | |||
| 467 | } | |||
| 468 | ||||
| 469 | SstFileManager* NewSstFileManager(Env* env, std::shared_ptr<Logger> info_log, | |||
| 470 | std::string trash_dir, | |||
| 471 | int64_t rate_bytes_per_sec, | |||
| 472 | bool delete_existing_trash, Status* status, | |||
| 473 | double max_trash_db_ratio, | |||
| 474 | uint64_t bytes_max_delete_chunk) { | |||
| 475 | SstFileManagerImpl* res = | |||
| 476 | new SstFileManagerImpl(env, info_log, rate_bytes_per_sec, | |||
| 477 | max_trash_db_ratio, bytes_max_delete_chunk); | |||
| 478 | ||||
| 479 | // trash_dir is deprecated and not needed anymore, but if user passed it | |||
| 480 | // we will still remove files in it. | |||
| 481 | Status s; | |||
| 482 | if (delete_existing_trash && trash_dir != "") { | |||
| 483 | std::vector<std::string> files_in_trash; | |||
| 484 | s = env->GetChildren(trash_dir, &files_in_trash); | |||
| 485 | if (s.ok()) { | |||
| 486 | for (const std::string& trash_file : files_in_trash) { | |||
| 487 | if (trash_file == "." || trash_file == "..") { | |||
| 488 | continue; | |||
| 489 | } | |||
| 490 | ||||
| 491 | std::string path_in_trash = trash_dir + "/" + trash_file; | |||
| 492 | res->OnAddFile(path_in_trash); | |||
| 493 | Status file_delete = | |||
| 494 | res->ScheduleFileDeletion(path_in_trash, trash_dir); | |||
| 495 | if (s.ok() && !file_delete.ok()) { | |||
| 496 | s = file_delete; | |||
| 497 | } | |||
| 498 | } | |||
| 499 | } | |||
| 500 | } | |||
| 501 | ||||
| 502 | if (status) { | |||
| 503 | *status = s; | |||
| 504 | } | |||
| 505 | ||||
| 506 | return res; | |||
| 507 | } | |||
| 508 | ||||
| 509 | #else | |||
| 510 | ||||
| 511 | SstFileManager* NewSstFileManager(Env* /*env*/, | |||
| 512 | std::shared_ptr<Logger> /*info_log*/, | |||
| 513 | std::string /*trash_dir*/, | |||
| 514 | int64_t /*rate_bytes_per_sec*/, | |||
| 515 | bool /*delete_existing_trash*/, | |||
| 516 | Status* status, double /*max_trash_db_ratio*/, | |||
| 517 | uint64_t /*bytes_max_delete_chunk*/) { | |||
| 518 | if (status) { | |||
| 519 | *status = | |||
| 520 | Status::NotSupported("SstFileManager is not supported in ROCKSDB_LITE"); | |||
| 521 | } | |||
| 522 | return nullptr; | |||
| 523 | } | |||
| 524 | ||||
| 525 | #endif // ROCKSDB_LITE | |||
| 526 | ||||
| 527 | } // namespace rocksdb |
| 1 | // Copyright (c) 2011-present, Facebook, Inc. All rights reserved. |
| 2 | // This source code is licensed under both the GPLv2 (found in the |
| 3 | // COPYING file in the root directory) and Apache 2.0 License |
| 4 | // (found in the LICENSE.Apache file in the root directory). |
| 5 | // Copyright (c) 2011 The LevelDB Authors. All rights reserved. |
| 6 | // Use of this source code is governed by a BSD-style license that can be |
| 7 | // found in the LICENSE file. See the AUTHORS file for names of contributors. |
| 8 | // |
| 9 | // An Env is an interface used by the rocksdb implementation to access |
| 10 | // operating system functionality like the filesystem etc. Callers |
| 11 | // may wish to provide a custom Env object when opening a database to |
| 12 | // get fine gain control; e.g., to rate limit file system operations. |
| 13 | // |
| 14 | // All Env implementations are safe for concurrent access from |
| 15 | // multiple threads without any external synchronization. |
| 16 | |
| 17 | #pragma once |
| 18 | |
| 19 | #include <stdint.h> |
| 20 | #include <cstdarg> |
| 21 | #include <functional> |
| 22 | #include <limits> |
| 23 | #include <memory> |
| 24 | #include <string> |
| 25 | #include <vector> |
| 26 | #include "rocksdb/status.h" |
| 27 | #include "rocksdb/thread_status.h" |
| 28 | |
| 29 | #ifdef _WIN32 |
| 30 | // Windows API macro interference |
| 31 | #undef DeleteFile |
| 32 | #undef GetCurrentTime |
| 33 | #endif |
| 34 | |
| 35 | #if defined(__GNUC__4) || defined(__clang__1) |
| 36 | #define ROCKSDB_PRINTF_FORMAT_ATTR(format_param, dots_param)__attribute__((__format__(__printf__, format_param, dots_param ))) \ |
| 37 | __attribute__((__format__(__printf__, format_param, dots_param))) |
| 38 | #else |
| 39 | #define ROCKSDB_PRINTF_FORMAT_ATTR(format_param, dots_param)__attribute__((__format__(__printf__, format_param, dots_param ))) |
| 40 | #endif |
| 41 | |
| 42 | namespace rocksdb { |
| 43 | |
| 44 | class FileLock; |
| 45 | class Logger; |
| 46 | class RandomAccessFile; |
| 47 | class SequentialFile; |
| 48 | class Slice; |
| 49 | class WritableFile; |
| 50 | class RandomRWFile; |
| 51 | class MemoryMappedFileBuffer; |
| 52 | class Directory; |
| 53 | struct DBOptions; |
| 54 | struct ImmutableDBOptions; |
| 55 | struct MutableDBOptions; |
| 56 | class RateLimiter; |
| 57 | class ThreadStatusUpdater; |
| 58 | struct ThreadStatus; |
| 59 | |
| 60 | const size_t kDefaultPageSize = 4 * 1024; |
| 61 | |
| 62 | // Options while opening a file to read/write |
| 63 | struct EnvOptions { |
| 64 | // Construct with default Options |
| 65 | EnvOptions(); |
| 66 | |
| 67 | // Construct from Options |
| 68 | explicit EnvOptions(const DBOptions& options); |
| 69 | |
| 70 | // If true, then use mmap to read data |
| 71 | bool use_mmap_reads = false; |
| 72 | |
| 73 | // If true, then use mmap to write data |
| 74 | bool use_mmap_writes = true; |
| 75 | |
| 76 | // If true, then use O_DIRECT for reading data |
| 77 | bool use_direct_reads = false; |
| 78 | |
| 79 | // If true, then use O_DIRECT for writing data |
| 80 | bool use_direct_writes = false; |
| 81 | |
| 82 | // If false, fallocate() calls are bypassed |
| 83 | bool allow_fallocate = true; |
| 84 | |
| 85 | // If true, set the FD_CLOEXEC on open fd. |
| 86 | bool set_fd_cloexec = true; |
| 87 | |
| 88 | // Allows OS to incrementally sync files to disk while they are being |
| 89 | // written, in the background. Issue one request for every bytes_per_sync |
| 90 | // written. 0 turns it off. |
| 91 | // Default: 0 |
| 92 | uint64_t bytes_per_sync = 0; |
| 93 | |
| 94 | // If true, we will preallocate the file with FALLOC_FL_KEEP_SIZE flag, which |
| 95 | // means that file size won't change as part of preallocation. |
| 96 | // If false, preallocation will also change the file size. This option will |
| 97 | // improve the performance in workloads where you sync the data on every |
| 98 | // write. By default, we set it to true for MANIFEST writes and false for |
| 99 | // WAL writes |
| 100 | bool fallocate_with_keep_size = true; |
| 101 | |
| 102 | // See DBOptions doc |
| 103 | size_t compaction_readahead_size; |
| 104 | |
| 105 | // See DBOptions doc |
| 106 | size_t random_access_max_buffer_size; |
| 107 | |
| 108 | // See DBOptions doc |
| 109 | size_t writable_file_max_buffer_size = 1024 * 1024; |
| 110 | |
| 111 | // If not nullptr, write rate limiting is enabled for flush and compaction |
| 112 | RateLimiter* rate_limiter = nullptr; |
| 113 | }; |
| 114 | |
| 115 | class Env { |
| 116 | public: |
| 117 | struct FileAttributes { |
| 118 | // File name |
| 119 | std::string name; |
| 120 | |
| 121 | // Size of file in bytes |
| 122 | uint64_t size_bytes; |
| 123 | }; |
| 124 | |
| 125 | Env() : thread_status_updater_(nullptr) {} |
| 126 | |
| 127 | virtual ~Env(); |
| 128 | |
| 129 | // Return a default environment suitable for the current operating |
| 130 | // system. Sophisticated users may wish to provide their own Env |
| 131 | // implementation instead of relying on this default environment. |
| 132 | // |
| 133 | // The result of Default() belongs to rocksdb and must never be deleted. |
| 134 | static Env* Default(); |
| 135 | |
| 136 | // Create a brand new sequentially-readable file with the specified name. |
| 137 | // On success, stores a pointer to the new file in *result and returns OK. |
| 138 | // On failure stores nullptr in *result and returns non-OK. If the file does |
| 139 | // not exist, returns a non-OK status. |
| 140 | // |
| 141 | // The returned file will only be accessed by one thread at a time. |
| 142 | virtual Status NewSequentialFile(const std::string& fname, |
| 143 | std::unique_ptr<SequentialFile>* result, |
| 144 | const EnvOptions& options) = 0; |
| 145 | |
| 146 | // Create a brand new random access read-only file with the |
| 147 | // specified name. On success, stores a pointer to the new file in |
| 148 | // *result and returns OK. On failure stores nullptr in *result and |
| 149 | // returns non-OK. If the file does not exist, returns a non-OK |
| 150 | // status. |
| 151 | // |
| 152 | // The returned file may be concurrently accessed by multiple threads. |
| 153 | virtual Status NewRandomAccessFile(const std::string& fname, |
| 154 | std::unique_ptr<RandomAccessFile>* result, |
| 155 | const EnvOptions& options) = 0; |
| 156 | // These values match Linux definition |
| 157 | // https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git/tree/include/uapi/linux/fcntl.h#n56 |
| 158 | enum WriteLifeTimeHint { |
| 159 | WLTH_NOT_SET = 0, // No hint information set |
| 160 | WLTH_NONE, // No hints about write life time |
| 161 | WLTH_SHORT, // Data written has a short life time |
| 162 | WLTH_MEDIUM, // Data written has a medium life time |
| 163 | WLTH_LONG, // Data written has a long life time |
| 164 | WLTH_EXTREME, // Data written has an extremely long life time |
| 165 | }; |
| 166 | |
| 167 | // Create an object that writes to a new file with the specified |
| 168 | // name. Deletes any existing file with the same name and creates a |
| 169 | // new file. On success, stores a pointer to the new file in |
| 170 | // *result and returns OK. On failure stores nullptr in *result and |
| 171 | // returns non-OK. |
| 172 | // |
| 173 | // The returned file will only be accessed by one thread at a time. |
| 174 | virtual Status NewWritableFile(const std::string& fname, |
| 175 | std::unique_ptr<WritableFile>* result, |
| 176 | const EnvOptions& options) = 0; |
| 177 | |
| 178 | // Create an object that writes to a new file with the specified |
| 179 | // name. Deletes any existing file with the same name and creates a |
| 180 | // new file. On success, stores a pointer to the new file in |
| 181 | // *result and returns OK. On failure stores nullptr in *result and |
| 182 | // returns non-OK. |
| 183 | // |
| 184 | // The returned file will only be accessed by one thread at a time. |
| 185 | virtual Status ReopenWritableFile(const std::string& /*fname*/, |
| 186 | std::unique_ptr<WritableFile>* /*result*/, |
| 187 | const EnvOptions& /*options*/) { |
| 188 | return Status::NotSupported(); |
| 189 | } |
| 190 | |
| 191 | // Reuse an existing file by renaming it and opening it as writable. |
| 192 | virtual Status ReuseWritableFile(const std::string& fname, |
| 193 | const std::string& old_fname, |
| 194 | std::unique_ptr<WritableFile>* result, |
| 195 | const EnvOptions& options); |
| 196 | |
| 197 | // Open `fname` for random read and write, if file doesn't exist the file |
| 198 | // will be created. On success, stores a pointer to the new file in |
| 199 | // *result and returns OK. On failure returns non-OK. |
| 200 | // |
| 201 | // The returned file will only be accessed by one thread at a time. |
| 202 | virtual Status NewRandomRWFile(const std::string& /*fname*/, |
| 203 | std::unique_ptr<RandomRWFile>* /*result*/, |
| 204 | const EnvOptions& /*options*/) { |
| 205 | return Status::NotSupported("RandomRWFile is not implemented in this Env"); |
| 206 | } |
| 207 | |
| 208 | // Opens `fname` as a memory-mapped file for read and write (in-place updates |
| 209 | // only, i.e., no appends). On success, stores a raw buffer covering the whole |
| 210 | // file in `*result`. The file must exist prior to this call. |
| 211 | virtual Status NewMemoryMappedFileBuffer( |
| 212 | const std::string& /*fname*/, |
| 213 | std::unique_ptr<MemoryMappedFileBuffer>* /*result*/) { |
| 214 | return Status::NotSupported( |
| 215 | "MemoryMappedFileBuffer is not implemented in this Env"); |
| 216 | } |
| 217 | |
| 218 | // Create an object that represents a directory. Will fail if directory |
| 219 | // doesn't exist. If the directory exists, it will open the directory |
| 220 | // and create a new Directory object. |
| 221 | // |
| 222 | // On success, stores a pointer to the new Directory in |
| 223 | // *result and returns OK. On failure stores nullptr in *result and |
| 224 | // returns non-OK. |
| 225 | virtual Status NewDirectory(const std::string& name, |
| 226 | std::unique_ptr<Directory>* result) = 0; |
| 227 | |
| 228 | // Returns OK if the named file exists. |
| 229 | // NotFound if the named file does not exist, |
| 230 | // the calling process does not have permission to determine |
| 231 | // whether this file exists, or if the path is invalid. |
| 232 | // IOError if an IO Error was encountered |
| 233 | virtual Status FileExists(const std::string& fname) = 0; |
| 234 | |
| 235 | // Store in *result the names of the children of the specified directory. |
| 236 | // The names are relative to "dir". |
| 237 | // Original contents of *results are dropped. |
| 238 | // Returns OK if "dir" exists and "*result" contains its children. |
| 239 | // NotFound if "dir" does not exist, the calling process does not have |
| 240 | // permission to access "dir", or if "dir" is invalid. |
| 241 | // IOError if an IO Error was encountered |
| 242 | virtual Status GetChildren(const std::string& dir, |
| 243 | std::vector<std::string>* result) = 0; |
| 244 | |
| 245 | // Store in *result the attributes of the children of the specified directory. |
| 246 | // In case the implementation lists the directory prior to iterating the files |
| 247 | // and files are concurrently deleted, the deleted files will be omitted from |
| 248 | // result. |
| 249 | // The name attributes are relative to "dir". |
| 250 | // Original contents of *results are dropped. |
| 251 | // Returns OK if "dir" exists and "*result" contains its children. |
| 252 | // NotFound if "dir" does not exist, the calling process does not have |
| 253 | // permission to access "dir", or if "dir" is invalid. |
| 254 | // IOError if an IO Error was encountered |
| 255 | virtual Status GetChildrenFileAttributes(const std::string& dir, |
| 256 | std::vector<FileAttributes>* result); |
| 257 | |
| 258 | // Delete the named file. |
| 259 | virtual Status DeleteFile(const std::string& fname) = 0; |
| 260 | |
| 261 | // Truncate the named file to the specified size. |
| 262 | virtual Status Truncate(const std::string& /*fname*/, size_t /*size*/) { |
| 263 | return Status::NotSupported("Truncate is not supported for this Env"); |
| 264 | } |
| 265 | |
| 266 | // Create the specified directory. Returns error if directory exists. |
| 267 | virtual Status CreateDir(const std::string& dirname) = 0; |
| 268 | |
| 269 | // Creates directory if missing. Return Ok if it exists, or successful in |
| 270 | // Creating. |
| 271 | virtual Status CreateDirIfMissing(const std::string& dirname) = 0; |
| 272 | |
| 273 | // Delete the specified directory. |
| 274 | virtual Status DeleteDir(const std::string& dirname) = 0; |
| 275 | |
| 276 | // Store the size of fname in *file_size. |
| 277 | virtual Status GetFileSize(const std::string& fname, uint64_t* file_size) = 0; |
| 278 | |
| 279 | // Store the last modification time of fname in *file_mtime. |
| 280 | virtual Status GetFileModificationTime(const std::string& fname, |
| 281 | uint64_t* file_mtime) = 0; |
| 282 | // Rename file src to target. |
| 283 | virtual Status RenameFile(const std::string& src, |
| 284 | const std::string& target) = 0; |
| 285 | |
| 286 | // Hard Link file src to target. |
| 287 | virtual Status LinkFile(const std::string& /*src*/, |
| 288 | const std::string& /*target*/) { |
| 289 | return Status::NotSupported("LinkFile is not supported for this Env"); |
| 290 | } |
| 291 | |
| 292 | virtual Status NumFileLinks(const std::string& /*fname*/, |
| 293 | uint64_t* /*count*/) { |
| 294 | return Status::NotSupported( |
| 295 | "Getting number of file links is not supported for this Env"); |
| 296 | } |
| 297 | |
| 298 | virtual Status AreFilesSame(const std::string& /*first*/, |
| 299 | const std::string& /*second*/, bool* /*res*/) { |
| 300 | return Status::NotSupported("AreFilesSame is not supported for this Env"); |
| 301 | } |
| 302 | |
| 303 | // Lock the specified file. Used to prevent concurrent access to |
| 304 | // the same db by multiple processes. On failure, stores nullptr in |
| 305 | // *lock and returns non-OK. |
| 306 | // |
| 307 | // On success, stores a pointer to the object that represents the |
| 308 | // acquired lock in *lock and returns OK. The caller should call |
| 309 | // UnlockFile(*lock) to release the lock. If the process exits, |
| 310 | // the lock will be automatically released. |
| 311 | // |
| 312 | // If somebody else already holds the lock, finishes immediately |
| 313 | // with a failure. I.e., this call does not wait for existing locks |
| 314 | // to go away. |
| 315 | // |
| 316 | // May create the named file if it does not already exist. |
| 317 | virtual Status LockFile(const std::string& fname, FileLock** lock) = 0; |
| 318 | |
| 319 | // Release the lock acquired by a previous successful call to LockFile. |
| 320 | // REQUIRES: lock was returned by a successful LockFile() call |
| 321 | // REQUIRES: lock has not already been unlocked. |
| 322 | virtual Status UnlockFile(FileLock* lock) = 0; |
| 323 | |
| 324 | // Priority for scheduling job in thread pool |
| 325 | enum Priority { BOTTOM, LOW, HIGH, USER, TOTAL }; |
| 326 | |
| 327 | static std::string PriorityToString(Priority priority); |
| 328 | |
| 329 | // Priority for requesting bytes in rate limiter scheduler |
| 330 | enum IOPriority { IO_LOW = 0, IO_HIGH = 1, IO_TOTAL = 2 }; |
| 331 | |
| 332 | // Arrange to run "(*function)(arg)" once in a background thread, in |
| 333 | // the thread pool specified by pri. By default, jobs go to the 'LOW' |
| 334 | // priority thread pool. |
| 335 | |
| 336 | // "function" may run in an unspecified thread. Multiple functions |
| 337 | // added to the same Env may run concurrently in different threads. |
| 338 | // I.e., the caller may not assume that background work items are |
| 339 | // serialized. |
| 340 | // When the UnSchedule function is called, the unschedFunction |
| 341 | // registered at the time of Schedule is invoked with arg as a parameter. |
| 342 | virtual void Schedule(void (*function)(void* arg), void* arg, |
| 343 | Priority pri = LOW, void* tag = nullptr, |
| 344 | void (*unschedFunction)(void* arg) = nullptr) = 0; |
| 345 | |
| 346 | // Arrange to remove jobs for given arg from the queue_ if they are not |
| 347 | // already scheduled. Caller is expected to have exclusive lock on arg. |
| 348 | virtual int UnSchedule(void* /*arg*/, Priority /*pri*/) { return 0; } |
| 349 | |
| 350 | // Start a new thread, invoking "function(arg)" within the new thread. |
| 351 | // When "function(arg)" returns, the thread will be destroyed. |
| 352 | virtual void StartThread(void (*function)(void* arg), void* arg) = 0; |
| 353 | |
| 354 | // Wait for all threads started by StartThread to terminate. |
| 355 | virtual void WaitForJoin() {} |
| 356 | |
| 357 | // Get thread pool queue length for specific thread pool. |
| 358 | virtual unsigned int GetThreadPoolQueueLen(Priority /*pri*/ = LOW) const { |
| 359 | return 0; |
| 360 | } |
| 361 | |
| 362 | // *path is set to a temporary directory that can be used for testing. It may |
| 363 | // or many not have just been created. The directory may or may not differ |
| 364 | // between runs of the same process, but subsequent calls will return the |
| 365 | // same directory. |
| 366 | virtual Status GetTestDirectory(std::string* path) = 0; |
| 367 | |
| 368 | // Create and return a log file for storing informational messages. |
| 369 | virtual Status NewLogger(const std::string& fname, |
| 370 | std::shared_ptr<Logger>* result) = 0; |
| 371 | |
| 372 | // Returns the number of micro-seconds since some fixed point in time. |
| 373 | // It is often used as system time such as in GenericRateLimiter |
| 374 | // and other places so a port needs to return system time in order to work. |
| 375 | virtual uint64_t NowMicros() = 0; |
| 376 | |
| 377 | // Returns the number of nano-seconds since some fixed point in time. Only |
| 378 | // useful for computing deltas of time in one run. |
| 379 | // Default implementation simply relies on NowMicros. |
| 380 | // In platform-specific implementations, NowNanos() should return time points |
| 381 | // that are MONOTONIC. |
| 382 | virtual uint64_t NowNanos() { return NowMicros() * 1000; } |
| 383 | |
| 384 | // 0 indicates not supported. |
| 385 | virtual uint64_t NowCPUNanos() { return 0; } |
| 386 | |
| 387 | // Sleep/delay the thread for the prescribed number of micro-seconds. |
| 388 | virtual void SleepForMicroseconds(int micros) = 0; |
| 389 | |
| 390 | // Get the current host name. |
| 391 | virtual Status GetHostName(char* name, uint64_t len) = 0; |
| 392 | |
| 393 | // Get the number of seconds since the Epoch, 1970-01-01 00:00:00 (UTC). |
| 394 | // Only overwrites *unix_time on success. |
| 395 | virtual Status GetCurrentTime(int64_t* unix_time) = 0; |
| 396 | |
| 397 | // Get full directory name for this db. |
| 398 | virtual Status GetAbsolutePath(const std::string& db_path, |
| 399 | std::string* output_path) = 0; |
| 400 | |
| 401 | // The number of background worker threads of a specific thread pool |
| 402 | // for this environment. 'LOW' is the default pool. |
| 403 | // default number: 1 |
| 404 | virtual void SetBackgroundThreads(int number, Priority pri = LOW) = 0; |
| 405 | virtual int GetBackgroundThreads(Priority pri = LOW) = 0; |
| 406 | |
| 407 | virtual Status SetAllowNonOwnerAccess(bool /*allow_non_owner_access*/) { |
| 408 | return Status::NotSupported("Not supported."); |
| 409 | } |
| 410 | |
| 411 | // Enlarge number of background worker threads of a specific thread pool |
| 412 | // for this environment if it is smaller than specified. 'LOW' is the default |
| 413 | // pool. |
| 414 | virtual void IncBackgroundThreadsIfNeeded(int number, Priority pri) = 0; |
| 415 | |
| 416 | // Lower IO priority for threads from the specified pool. |
| 417 | virtual void LowerThreadPoolIOPriority(Priority /*pool*/ = LOW) {} |
| 418 | |
| 419 | // Lower CPU priority for threads from the specified pool. |
| 420 | virtual void LowerThreadPoolCPUPriority(Priority /*pool*/ = LOW) {} |
| 421 | |
| 422 | // Converts seconds-since-Jan-01-1970 to a printable string |
| 423 | virtual std::string TimeToString(uint64_t time) = 0; |
| 424 | |
| 425 | // Generates a unique id that can be used to identify a db |
| 426 | virtual std::string GenerateUniqueId(); |
| 427 | |
| 428 | // OptimizeForLogWrite will create a new EnvOptions object that is a copy of |
| 429 | // the EnvOptions in the parameters, but is optimized for reading log files. |
| 430 | virtual EnvOptions OptimizeForLogRead(const EnvOptions& env_options) const; |
| 431 | |
| 432 | // OptimizeForManifestRead will create a new EnvOptions object that is a copy |
| 433 | // of the EnvOptions in the parameters, but is optimized for reading manifest |
| 434 | // files. |
| 435 | virtual EnvOptions OptimizeForManifestRead( |
| 436 | const EnvOptions& env_options) const; |
| 437 | |
| 438 | // OptimizeForLogWrite will create a new EnvOptions object that is a copy of |
| 439 | // the EnvOptions in the parameters, but is optimized for writing log files. |
| 440 | // Default implementation returns the copy of the same object. |
| 441 | virtual EnvOptions OptimizeForLogWrite(const EnvOptions& env_options, |
| 442 | const DBOptions& db_options) const; |
| 443 | // OptimizeForManifestWrite will create a new EnvOptions object that is a copy |
| 444 | // of the EnvOptions in the parameters, but is optimized for writing manifest |
| 445 | // files. Default implementation returns the copy of the same object. |
| 446 | virtual EnvOptions OptimizeForManifestWrite( |
| 447 | const EnvOptions& env_options) const; |
| 448 | |
| 449 | // OptimizeForCompactionTableWrite will create a new EnvOptions object that is |
| 450 | // a copy of the EnvOptions in the parameters, but is optimized for writing |
| 451 | // table files. |
| 452 | virtual EnvOptions OptimizeForCompactionTableWrite( |
| 453 | const EnvOptions& env_options, |
| 454 | const ImmutableDBOptions& immutable_ops) const; |
| 455 | |
| 456 | // OptimizeForCompactionTableWrite will create a new EnvOptions object that |
| 457 | // is a copy of the EnvOptions in the parameters, but is optimized for reading |
| 458 | // table files. |
| 459 | virtual EnvOptions OptimizeForCompactionTableRead( |
| 460 | const EnvOptions& env_options, |
| 461 | const ImmutableDBOptions& db_options) const; |
| 462 | |
| 463 | // Returns the status of all threads that belong to the current Env. |
| 464 | virtual Status GetThreadList(std::vector<ThreadStatus>* /*thread_list*/) { |
| 465 | return Status::NotSupported("Not supported."); |
| 466 | } |
| 467 | |
| 468 | // Returns the pointer to ThreadStatusUpdater. This function will be |
| 469 | // used in RocksDB internally to update thread status and supports |
| 470 | // GetThreadList(). |
| 471 | virtual ThreadStatusUpdater* GetThreadStatusUpdater() const { |
| 472 | return thread_status_updater_; |
| 473 | } |
| 474 | |
| 475 | // Returns the ID of the current thread. |
| 476 | virtual uint64_t GetThreadID() const; |
| 477 | |
| 478 | // This seems to clash with a macro on Windows, so #undef it here |
| 479 | #undef GetFreeSpace |
| 480 | |
| 481 | // Get the amount of free disk space |
| 482 | virtual Status GetFreeSpace(const std::string& /*path*/, |
| 483 | uint64_t* /*diskfree*/) { |
| 484 | return Status::NotSupported(); |
| 485 | } |
| 486 | |
| 487 | // If you're adding methods here, remember to add them to EnvWrapper too. |
| 488 | |
| 489 | protected: |
| 490 | // The pointer to an internal structure that will update the |
| 491 | // status of each thread. |
| 492 | ThreadStatusUpdater* thread_status_updater_; |
| 493 | |
| 494 | private: |
| 495 | // No copying allowed |
| 496 | Env(const Env&); |
| 497 | void operator=(const Env&); |
| 498 | }; |
| 499 | |
| 500 | // The factory function to construct a ThreadStatusUpdater. Any Env |
| 501 | // that supports GetThreadList() feature should call this function in its |
| 502 | // constructor to initialize thread_status_updater_. |
| 503 | ThreadStatusUpdater* CreateThreadStatusUpdater(); |
| 504 | |
| 505 | // A file abstraction for reading sequentially through a file |
| 506 | class SequentialFile { |
| 507 | public: |
| 508 | SequentialFile() {} |
| 509 | virtual ~SequentialFile(); |
| 510 | |
| 511 | // Read up to "n" bytes from the file. "scratch[0..n-1]" may be |
| 512 | // written by this routine. Sets "*result" to the data that was |
| 513 | // read (including if fewer than "n" bytes were successfully read). |
| 514 | // May set "*result" to point at data in "scratch[0..n-1]", so |
| 515 | // "scratch[0..n-1]" must be live when "*result" is used. |
| 516 | // If an error was encountered, returns a non-OK status. |
| 517 | // |
| 518 | // REQUIRES: External synchronization |
| 519 | virtual Status Read(size_t n, Slice* result, char* scratch) = 0; |
| 520 | |
| 521 | // Skip "n" bytes from the file. This is guaranteed to be no |
| 522 | // slower that reading the same data, but may be faster. |
| 523 | // |
| 524 | // If end of file is reached, skipping will stop at the end of the |
| 525 | // file, and Skip will return OK. |
| 526 | // |
| 527 | // REQUIRES: External synchronization |
| 528 | virtual Status Skip(uint64_t n) = 0; |
| 529 | |
| 530 | // Indicates the upper layers if the current SequentialFile implementation |
| 531 | // uses direct IO. |
| 532 | virtual bool use_direct_io() const { return false; } |
| 533 | |
| 534 | // Use the returned alignment value to allocate |
| 535 | // aligned buffer for Direct I/O |
| 536 | virtual size_t GetRequiredBufferAlignment() const { return kDefaultPageSize; } |
| 537 | |
| 538 | // Remove any kind of caching of data from the offset to offset+length |
| 539 | // of this file. If the length is 0, then it refers to the end of file. |
| 540 | // If the system is not caching the file contents, then this is a noop. |
| 541 | virtual Status InvalidateCache(size_t /*offset*/, size_t /*length*/) { |
| 542 | return Status::NotSupported("InvalidateCache not supported."); |
| 543 | } |
| 544 | |
| 545 | // Positioned Read for direct I/O |
| 546 | // If Direct I/O enabled, offset, n, and scratch should be properly aligned |
| 547 | virtual Status PositionedRead(uint64_t /*offset*/, size_t /*n*/, |
| 548 | Slice* /*result*/, char* /*scratch*/) { |
| 549 | return Status::NotSupported(); |
| 550 | } |
| 551 | |
| 552 | // If you're adding methods here, remember to add them to |
| 553 | // SequentialFileWrapper too. |
| 554 | }; |
| 555 | |
| 556 | // A file abstraction for randomly reading the contents of a file. |
| 557 | class RandomAccessFile { |
| 558 | public: |
| 559 | RandomAccessFile() {} |
| 560 | virtual ~RandomAccessFile(); |
| 561 | |
| 562 | // Read up to "n" bytes from the file starting at "offset". |
| 563 | // "scratch[0..n-1]" may be written by this routine. Sets "*result" |
| 564 | // to the data that was read (including if fewer than "n" bytes were |
| 565 | // successfully read). May set "*result" to point at data in |
| 566 | // "scratch[0..n-1]", so "scratch[0..n-1]" must be live when |
| 567 | // "*result" is used. If an error was encountered, returns a non-OK |
| 568 | // status. |
| 569 | // |
| 570 | // Safe for concurrent use by multiple threads. |
| 571 | // If Direct I/O enabled, offset, n, and scratch should be aligned properly. |
| 572 | virtual Status Read(uint64_t offset, size_t n, Slice* result, |
| 573 | char* scratch) const = 0; |
| 574 | |
| 575 | // Readahead the file starting from offset by n bytes for caching. |
| 576 | virtual Status Prefetch(uint64_t /*offset*/, size_t /*n*/) { |
| 577 | return Status::OK(); |
| 578 | } |
| 579 | |
| 580 | // Tries to get an unique ID for this file that will be the same each time |
| 581 | // the file is opened (and will stay the same while the file is open). |
| 582 | // Furthermore, it tries to make this ID at most "max_size" bytes. If such an |
| 583 | // ID can be created this function returns the length of the ID and places it |
| 584 | // in "id"; otherwise, this function returns 0, in which case "id" |
| 585 | // may not have been modified. |
| 586 | // |
| 587 | // This function guarantees, for IDs from a given environment, two unique ids |
| 588 | // cannot be made equal to each other by adding arbitrary bytes to one of |
| 589 | // them. That is, no unique ID is the prefix of another. |
| 590 | // |
| 591 | // This function guarantees that the returned ID will not be interpretable as |
| 592 | // a single varint. |
| 593 | // |
| 594 | // Note: these IDs are only valid for the duration of the process. |
| 595 | virtual size_t GetUniqueId(char* /*id*/, size_t /*max_size*/) const { |
| 596 | return 0; // Default implementation to prevent issues with backwards |
| 597 | // compatibility. |
| 598 | }; |
| 599 | |
| 600 | enum AccessPattern { NORMAL, RANDOM, SEQUENTIAL, WILLNEED, DONTNEED }; |
| 601 | |
| 602 | virtual void Hint(AccessPattern /*pattern*/) {} |
| 603 | |
| 604 | // Indicates the upper layers if the current RandomAccessFile implementation |
| 605 | // uses direct IO. |
| 606 | virtual bool use_direct_io() const { return false; } |
| 607 | |
| 608 | // Use the returned alignment value to allocate |
| 609 | // aligned buffer for Direct I/O |
| 610 | virtual size_t GetRequiredBufferAlignment() const { return kDefaultPageSize; } |
| 611 | |
| 612 | // Remove any kind of caching of data from the offset to offset+length |
| 613 | // of this file. If the length is 0, then it refers to the end of file. |
| 614 | // If the system is not caching the file contents, then this is a noop. |
| 615 | virtual Status InvalidateCache(size_t /*offset*/, size_t /*length*/) { |
| 616 | return Status::NotSupported("InvalidateCache not supported."); |
| 617 | } |
| 618 | |
| 619 | // If you're adding methods here, remember to add them to |
| 620 | // RandomAccessFileWrapper too. |
| 621 | }; |
| 622 | |
| 623 | // A file abstraction for sequential writing. The implementation |
| 624 | // must provide buffering since callers may append small fragments |
| 625 | // at a time to the file. |
| 626 | class WritableFile { |
| 627 | public: |
| 628 | WritableFile() |
| 629 | : last_preallocated_block_(0), |
| 630 | preallocation_block_size_(0), |
| 631 | io_priority_(Env::IO_TOTAL), |
| 632 | write_hint_(Env::WLTH_NOT_SET) {} |
| 633 | virtual ~WritableFile(); |
| 634 | |
| 635 | // Append data to the end of the file |
| 636 | // Note: A WriteabelFile object must support either Append or |
| 637 | // PositionedAppend, so the users cannot mix the two. |
| 638 | virtual Status Append(const Slice& data) = 0; |
| 639 | |
| 640 | // PositionedAppend data to the specified offset. The new EOF after append |
| 641 | // must be larger than the previous EOF. This is to be used when writes are |
| 642 | // not backed by OS buffers and hence has to always start from the start of |
| 643 | // the sector. The implementation thus needs to also rewrite the last |
| 644 | // partial sector. |
| 645 | // Note: PositionAppend does not guarantee moving the file offset after the |
| 646 | // write. A WritableFile object must support either Append or |
| 647 | // PositionedAppend, so the users cannot mix the two. |
| 648 | // |
| 649 | // PositionedAppend() can only happen on the page/sector boundaries. For that |
| 650 | // reason, if the last write was an incomplete sector we still need to rewind |
| 651 | // back to the nearest sector/page and rewrite the portion of it with whatever |
| 652 | // we need to add. We need to keep where we stop writing. |
| 653 | // |
| 654 | // PositionedAppend() can only write whole sectors. For that reason we have to |
| 655 | // pad with zeros for the last write and trim the file when closing according |
| 656 | // to the position we keep in the previous step. |
| 657 | // |
| 658 | // PositionedAppend() requires aligned buffer to be passed in. The alignment |
| 659 | // required is queried via GetRequiredBufferAlignment() |
| 660 | virtual Status PositionedAppend(const Slice& /* data */, |
| 661 | uint64_t /* offset */) { |
| 662 | return Status::NotSupported(); |
| 663 | } |
| 664 | |
| 665 | // Truncate is necessary to trim the file to the correct size |
| 666 | // before closing. It is not always possible to keep track of the file |
| 667 | // size due to whole pages writes. The behavior is undefined if called |
| 668 | // with other writes to follow. |
| 669 | virtual Status Truncate(uint64_t /*size*/) { return Status::OK(); } |
| 670 | virtual Status Close() = 0; |
| 671 | virtual Status Flush() = 0; |
| 672 | virtual Status Sync() = 0; // sync data |
| 673 | |
| 674 | /* |
| 675 | * Sync data and/or metadata as well. |
| 676 | * By default, sync only data. |
| 677 | * Override this method for environments where we need to sync |
| 678 | * metadata as well. |
| 679 | */ |
| 680 | virtual Status Fsync() { return Sync(); } |
| 681 | |
| 682 | // true if Sync() and Fsync() are safe to call concurrently with Append() |
| 683 | // and Flush(). |
| 684 | virtual bool IsSyncThreadSafe() const { return false; } |
| 685 | |
| 686 | // Indicates the upper layers if the current WritableFile implementation |
| 687 | // uses direct IO. |
| 688 | virtual bool use_direct_io() const { return false; } |
| 689 | |
| 690 | // Use the returned alignment value to allocate |
| 691 | // aligned buffer for Direct I/O |
| 692 | virtual size_t GetRequiredBufferAlignment() const { return kDefaultPageSize; } |
| 693 | /* |
| 694 | * Change the priority in rate limiter if rate limiting is enabled. |
| 695 | * If rate limiting is not enabled, this call has no effect. |
| 696 | */ |
| 697 | virtual void SetIOPriority(Env::IOPriority pri) { io_priority_ = pri; } |
| 698 | |
| 699 | virtual Env::IOPriority GetIOPriority() { return io_priority_; } |
| 700 | |
| 701 | virtual void SetWriteLifeTimeHint(Env::WriteLifeTimeHint hint) { |
| 702 | write_hint_ = hint; |
| 703 | } |
| 704 | |
| 705 | virtual Env::WriteLifeTimeHint GetWriteLifeTimeHint() { return write_hint_; } |
| 706 | /* |
| 707 | * Get the size of valid data in the file. |
| 708 | */ |
| 709 | virtual uint64_t GetFileSize() { return 0; } |
| 710 | |
| 711 | /* |
| 712 | * Get and set the default pre-allocation block size for writes to |
| 713 | * this file. If non-zero, then Allocate will be used to extend the |
| 714 | * underlying storage of a file (generally via fallocate) if the Env |
| 715 | * instance supports it. |
| 716 | */ |
| 717 | virtual void SetPreallocationBlockSize(size_t size) { |
| 718 | preallocation_block_size_ = size; |
| 719 | } |
| 720 | |
| 721 | virtual void GetPreallocationStatus(size_t* block_size, |
| 722 | size_t* last_allocated_block) { |
| 723 | *last_allocated_block = last_preallocated_block_; |
| 724 | *block_size = preallocation_block_size_; |
| 725 | } |
| 726 | |
| 727 | // For documentation, refer to RandomAccessFile::GetUniqueId() |
| 728 | virtual size_t GetUniqueId(char* /*id*/, size_t /*max_size*/) const { |
| 729 | return 0; // Default implementation to prevent issues with backwards |
| 730 | } |
| 731 | |
| 732 | // Remove any kind of caching of data from the offset to offset+length |
| 733 | // of this file. If the length is 0, then it refers to the end of file. |
| 734 | // If the system is not caching the file contents, then this is a noop. |
| 735 | // This call has no effect on dirty pages in the cache. |
| 736 | virtual Status InvalidateCache(size_t /*offset*/, size_t /*length*/) { |
| 737 | return Status::NotSupported("InvalidateCache not supported."); |
| 738 | } |
| 739 | |
| 740 | // Sync a file range with disk. |
| 741 | // offset is the starting byte of the file range to be synchronized. |
| 742 | // nbytes specifies the length of the range to be synchronized. |
| 743 | // This asks the OS to initiate flushing the cached data to disk, |
| 744 | // without waiting for completion. |
| 745 | // Default implementation does nothing. |
| 746 | virtual Status RangeSync(uint64_t /*offset*/, uint64_t /*nbytes*/) { |
| 747 | return Status::OK(); |
| 748 | } |
| 749 | |
| 750 | // PrepareWrite performs any necessary preparation for a write |
| 751 | // before the write actually occurs. This allows for pre-allocation |
| 752 | // of space on devices where it can result in less file |
| 753 | // fragmentation and/or less waste from over-zealous filesystem |
| 754 | // pre-allocation. |
| 755 | virtual void PrepareWrite(size_t offset, size_t len) { |
| 756 | if (preallocation_block_size_ == 0) { |
| 757 | return; |
| 758 | } |
| 759 | // If this write would cross one or more preallocation blocks, |
| 760 | // determine what the last preallocation block necessary to |
| 761 | // cover this write would be and Allocate to that point. |
| 762 | const auto block_size = preallocation_block_size_; |
| 763 | size_t new_last_preallocated_block = |
| 764 | (offset + len + block_size - 1) / block_size; |
| 765 | if (new_last_preallocated_block > last_preallocated_block_) { |
| 766 | size_t num_spanned_blocks = |
| 767 | new_last_preallocated_block - last_preallocated_block_; |
| 768 | Allocate(block_size * last_preallocated_block_, |
| 769 | block_size * num_spanned_blocks); |
| 770 | last_preallocated_block_ = new_last_preallocated_block; |
| 771 | } |
| 772 | } |
| 773 | |
| 774 | // Pre-allocates space for a file. |
| 775 | virtual Status Allocate(uint64_t /*offset*/, uint64_t /*len*/) { |
| 776 | return Status::OK(); |
| 777 | } |
| 778 | |
| 779 | // If you're adding methods here, remember to add them to |
| 780 | // WritableFileWrapper too. |
| 781 | |
| 782 | protected: |
| 783 | size_t preallocation_block_size() { return preallocation_block_size_; } |
| 784 | |
| 785 | private: |
| 786 | size_t last_preallocated_block_; |
| 787 | size_t preallocation_block_size_; |
| 788 | // No copying allowed |
| 789 | WritableFile(const WritableFile&); |
| 790 | void operator=(const WritableFile&); |
| 791 | |
| 792 | protected: |
| 793 | Env::IOPriority io_priority_; |
| 794 | Env::WriteLifeTimeHint write_hint_; |
| 795 | }; |
| 796 | |
| 797 | // A file abstraction for random reading and writing. |
| 798 | class RandomRWFile { |
| 799 | public: |
| 800 | RandomRWFile() {} |
| 801 | virtual ~RandomRWFile() {} |
| 802 | |
| 803 | // Indicates if the class makes use of direct I/O |
| 804 | // If false you must pass aligned buffer to Write() |
| 805 | virtual bool use_direct_io() const { return false; } |
| 806 | |
| 807 | // Use the returned alignment value to allocate |
| 808 | // aligned buffer for Direct I/O |
| 809 | virtual size_t GetRequiredBufferAlignment() const { return kDefaultPageSize; } |
| 810 | |
| 811 | // Write bytes in `data` at offset `offset`, Returns Status::OK() on success. |
| 812 | // Pass aligned buffer when use_direct_io() returns true. |
| 813 | virtual Status Write(uint64_t offset, const Slice& data) = 0; |
| 814 | |
| 815 | // Read up to `n` bytes starting from offset `offset` and store them in |
| 816 | // result, provided `scratch` size should be at least `n`. |
| 817 | // Returns Status::OK() on success. |
| 818 | virtual Status Read(uint64_t offset, size_t n, Slice* result, |
| 819 | char* scratch) const = 0; |
| 820 | |
| 821 | virtual Status Flush() = 0; |
| 822 | |
| 823 | virtual Status Sync() = 0; |
| 824 | |
| 825 | virtual Status Fsync() { return Sync(); } |
| 826 | |
| 827 | virtual Status Close() = 0; |
| 828 | |
| 829 | // If you're adding methods here, remember to add them to |
| 830 | // RandomRWFileWrapper too. |
| 831 | |
| 832 | // No copying allowed |
| 833 | RandomRWFile(const RandomRWFile&) = delete; |
| 834 | RandomRWFile& operator=(const RandomRWFile&) = delete; |
| 835 | }; |
| 836 | |
| 837 | // MemoryMappedFileBuffer object represents a memory-mapped file's raw buffer. |
| 838 | // Subclasses should release the mapping upon destruction. |
| 839 | class MemoryMappedFileBuffer { |
| 840 | public: |
| 841 | MemoryMappedFileBuffer(void* _base, size_t _length) |
| 842 | : base_(_base), length_(_length) {} |
| 843 | |
| 844 | virtual ~MemoryMappedFileBuffer() = 0; |
| 845 | |
| 846 | // We do not want to unmap this twice. We can make this class |
| 847 | // movable if desired, however, since |
| 848 | MemoryMappedFileBuffer(const MemoryMappedFileBuffer&) = delete; |
| 849 | MemoryMappedFileBuffer& operator=(const MemoryMappedFileBuffer&) = delete; |
| 850 | |
| 851 | void* GetBase() const { return base_; } |
| 852 | size_t GetLen() const { return length_; } |
| 853 | |
| 854 | protected: |
| 855 | void* base_; |
| 856 | const size_t length_; |
| 857 | }; |
| 858 | |
| 859 | // Directory object represents collection of files and implements |
| 860 | // filesystem operations that can be executed on directories. |
| 861 | class Directory { |
| 862 | public: |
| 863 | virtual ~Directory() {} |
| 864 | // Fsync directory. Can be called concurrently from multiple threads. |
| 865 | virtual Status Fsync() = 0; |
| 866 | |
| 867 | virtual size_t GetUniqueId(char* /*id*/, size_t /*max_size*/) const { |
| 868 | return 0; |
| 869 | } |
| 870 | |
| 871 | // If you're adding methods here, remember to add them to |
| 872 | // DirectoryWrapper too. |
| 873 | }; |
| 874 | |
| 875 | enum InfoLogLevel : unsigned char { |
| 876 | DEBUG_LEVEL = 0, |
| 877 | INFO_LEVEL, |
| 878 | WARN_LEVEL, |
| 879 | ERROR_LEVEL, |
| 880 | FATAL_LEVEL, |
| 881 | HEADER_LEVEL, |
| 882 | NUM_INFO_LOG_LEVELS, |
| 883 | }; |
| 884 | |
| 885 | // An interface for writing log messages. |
| 886 | class Logger { |
| 887 | public: |
| 888 | size_t kDoNotSupportGetLogFileSize = (std::numeric_limits<size_t>::max)(); |
| 889 | |
| 890 | explicit Logger(const InfoLogLevel log_level = InfoLogLevel::INFO_LEVEL) |
| 891 | : closed_(false), log_level_(log_level) {} |
| 892 | virtual ~Logger(); |
| 893 | |
| 894 | // Close the log file. Must be called before destructor. If the return |
| 895 | // status is NotSupported(), it means the implementation does cleanup in |
| 896 | // the destructor |
| 897 | virtual Status Close(); |
| 898 | |
| 899 | // Write a header to the log file with the specified format |
| 900 | // It is recommended that you log all header information at the start of the |
| 901 | // application. But it is not enforced. |
| 902 | virtual void LogHeader(const char* format, va_list ap) { |
| 903 | // Default implementation does a simple INFO level log write. |
| 904 | // Please override as per the logger class requirement. |
| 905 | Logv(format, ap); |
| 906 | } |
| 907 | |
| 908 | // Write an entry to the log file with the specified format. |
| 909 | virtual void Logv(const char* format, va_list ap) = 0; |
| 910 | |
| 911 | // Write an entry to the log file with the specified log level |
| 912 | // and format. Any log with level under the internal log level |
| 913 | // of *this (see @SetInfoLogLevel and @GetInfoLogLevel) will not be |
| 914 | // printed. |
| 915 | virtual void Logv(const InfoLogLevel log_level, const char* format, |
| 916 | va_list ap); |
| 917 | |
| 918 | virtual size_t GetLogFileSize() const { return kDoNotSupportGetLogFileSize; } |
| 919 | // Flush to the OS buffers |
| 920 | virtual void Flush() {} |
| 921 | virtual InfoLogLevel GetInfoLogLevel() const { return log_level_; } |
| 922 | virtual void SetInfoLogLevel(const InfoLogLevel log_level) { |
| 923 | log_level_ = log_level; |
| 924 | } |
| 925 | |
| 926 | // If you're adding methods here, remember to add them to LoggerWrapper too. |
| 927 | |
| 928 | protected: |
| 929 | virtual Status CloseImpl(); |
| 930 | bool closed_; |
| 931 | |
| 932 | private: |
| 933 | // No copying allowed |
| 934 | Logger(const Logger&); |
| 935 | void operator=(const Logger&); |
| 936 | InfoLogLevel log_level_; |
| 937 | }; |
| 938 | |
| 939 | // Identifies a locked file. |
| 940 | class FileLock { |
| 941 | public: |
| 942 | FileLock() {} |
| 943 | virtual ~FileLock(); |
| 944 | |
| 945 | private: |
| 946 | // No copying allowed |
| 947 | FileLock(const FileLock&); |
| 948 | void operator=(const FileLock&); |
| 949 | }; |
| 950 | |
| 951 | extern void LogFlush(const std::shared_ptr<Logger>& info_log); |
| 952 | |
| 953 | extern void Log(const InfoLogLevel log_level, |
| 954 | const std::shared_ptr<Logger>& info_log, const char* format, |
| 955 | ...) ROCKSDB_PRINTF_FORMAT_ATTR(3, 4)__attribute__((__format__(__printf__, 3, 4))); |
| 956 | |
| 957 | // a set of log functions with different log levels. |
| 958 | extern void Header(const std::shared_ptr<Logger>& info_log, const char* format, |
| 959 | ...) ROCKSDB_PRINTF_FORMAT_ATTR(2, 3)__attribute__((__format__(__printf__, 2, 3))); |
| 960 | extern void Debug(const std::shared_ptr<Logger>& info_log, const char* format, |
| 961 | ...) ROCKSDB_PRINTF_FORMAT_ATTR(2, 3)__attribute__((__format__(__printf__, 2, 3))); |
| 962 | extern void Info(const std::shared_ptr<Logger>& info_log, const char* format, |
| 963 | ...) ROCKSDB_PRINTF_FORMAT_ATTR(2, 3)__attribute__((__format__(__printf__, 2, 3))); |
| 964 | extern void Warn(const std::shared_ptr<Logger>& info_log, const char* format, |
| 965 | ...) ROCKSDB_PRINTF_FORMAT_ATTR(2, 3)__attribute__((__format__(__printf__, 2, 3))); |
| 966 | extern void Error(const std::shared_ptr<Logger>& info_log, const char* format, |
| 967 | ...) ROCKSDB_PRINTF_FORMAT_ATTR(2, 3)__attribute__((__format__(__printf__, 2, 3))); |
| 968 | extern void Fatal(const std::shared_ptr<Logger>& info_log, const char* format, |
| 969 | ...) ROCKSDB_PRINTF_FORMAT_ATTR(2, 3)__attribute__((__format__(__printf__, 2, 3))); |
| 970 | |
| 971 | // Log the specified data to *info_log if info_log is non-nullptr. |
| 972 | // The default info log level is InfoLogLevel::INFO_LEVEL. |
| 973 | extern void Log(const std::shared_ptr<Logger>& info_log, const char* format, |
| 974 | ...) ROCKSDB_PRINTF_FORMAT_ATTR(2, 3)__attribute__((__format__(__printf__, 2, 3))); |
| 975 | |
| 976 | extern void LogFlush(Logger* info_log); |
| 977 | |
| 978 | extern void Log(const InfoLogLevel log_level, Logger* info_log, |
| 979 | const char* format, ...) ROCKSDB_PRINTF_FORMAT_ATTR(3, 4)__attribute__((__format__(__printf__, 3, 4))); |
| 980 | |
| 981 | // The default info log level is InfoLogLevel::INFO_LEVEL. |
| 982 | extern void Log(Logger* info_log, const char* format, ...) |
| 983 | ROCKSDB_PRINTF_FORMAT_ATTR(2, 3)__attribute__((__format__(__printf__, 2, 3))); |
| 984 | |
| 985 | // a set of log functions with different log levels. |
| 986 | extern void Header(Logger* info_log, const char* format, ...) |
| 987 | ROCKSDB_PRINTF_FORMAT_ATTR(2, 3)__attribute__((__format__(__printf__, 2, 3))); |
| 988 | extern void Debug(Logger* info_log, const char* format, ...) |
| 989 | ROCKSDB_PRINTF_FORMAT_ATTR(2, 3)__attribute__((__format__(__printf__, 2, 3))); |
| 990 | extern void Info(Logger* info_log, const char* format, ...) |
| 991 | ROCKSDB_PRINTF_FORMAT_ATTR(2, 3)__attribute__((__format__(__printf__, 2, 3))); |
| 992 | extern void Warn(Logger* info_log, const char* format, ...) |
| 993 | ROCKSDB_PRINTF_FORMAT_ATTR(2, 3)__attribute__((__format__(__printf__, 2, 3))); |
| 994 | extern void Error(Logger* info_log, const char* format, ...) |
| 995 | ROCKSDB_PRINTF_FORMAT_ATTR(2, 3)__attribute__((__format__(__printf__, 2, 3))); |
| 996 | extern void Fatal(Logger* info_log, const char* format, ...) |
| 997 | ROCKSDB_PRINTF_FORMAT_ATTR(2, 3)__attribute__((__format__(__printf__, 2, 3))); |
| 998 | |
| 999 | // A utility routine: write "data" to the named file. |
| 1000 | extern Status WriteStringToFile(Env* env, const Slice& data, |
| 1001 | const std::string& fname, |
| 1002 | bool should_sync = false); |
| 1003 | |
| 1004 | // A utility routine: read contents of named file into *data |
| 1005 | extern Status ReadFileToString(Env* env, const std::string& fname, |
| 1006 | std::string* data); |
| 1007 | |
| 1008 | // Below are helpers for wrapping most of the classes in this file. |
| 1009 | // They forward all calls to another instance of the class. |
| 1010 | // Useful when wrapping the default implementations. |
| 1011 | // Typical usage is to inherit your wrapper from *Wrapper, e.g.: |
| 1012 | // |
| 1013 | // class MySequentialFileWrapper : public rocksdb::SequentialFileWrapper { |
| 1014 | // public: |
| 1015 | // MySequentialFileWrapper(rocksdb::SequentialFile* target): |
| 1016 | // rocksdb::SequentialFileWrapper(target) {} |
| 1017 | // Status Read(size_t n, Slice* result, char* scratch) override { |
| 1018 | // cout << "Doing a read of size " << n << "!" << endl; |
| 1019 | // return rocksdb::SequentialFileWrapper::Read(n, result, scratch); |
| 1020 | // } |
| 1021 | // // All other methods are forwarded to target_ automatically. |
| 1022 | // }; |
| 1023 | // |
| 1024 | // This is often more convenient than inheriting the class directly because |
| 1025 | // (a) Don't have to override and forward all methods - the Wrapper will |
| 1026 | // forward everything you're not explicitly overriding. |
| 1027 | // (b) Don't need to update the wrapper when more methods are added to the |
| 1028 | // rocksdb class. Unless you actually want to override the behavior. |
| 1029 | // (And unless rocksdb people forgot to update the *Wrapper class.) |
| 1030 | |
| 1031 | // An implementation of Env that forwards all calls to another Env. |
| 1032 | // May be useful to clients who wish to override just part of the |
| 1033 | // functionality of another Env. |
| 1034 | class EnvWrapper : public Env { |
| 1035 | public: |
| 1036 | // Initialize an EnvWrapper that delegates all calls to *t |
| 1037 | explicit EnvWrapper(Env* t) : target_(t) {} |
| 1038 | ~EnvWrapper() override; |
| 1039 | |
| 1040 | // Return the target to which this Env forwards all calls |
| 1041 | Env* target() const { return target_; } |
| 1042 | |
| 1043 | // The following text is boilerplate that forwards all methods to target() |
| 1044 | Status NewSequentialFile(const std::string& f, |
| 1045 | std::unique_ptr<SequentialFile>* r, |
| 1046 | const EnvOptions& options) override { |
| 1047 | return target_->NewSequentialFile(f, r, options); |
| 1048 | } |
| 1049 | Status NewRandomAccessFile(const std::string& f, |
| 1050 | std::unique_ptr<RandomAccessFile>* r, |
| 1051 | const EnvOptions& options) override { |
| 1052 | return target_->NewRandomAccessFile(f, r, options); |
| 1053 | } |
| 1054 | Status NewWritableFile(const std::string& f, std::unique_ptr<WritableFile>* r, |
| 1055 | const EnvOptions& options) override { |
| 1056 | return target_->NewWritableFile(f, r, options); |
| 1057 | } |
| 1058 | Status ReopenWritableFile(const std::string& fname, |
| 1059 | std::unique_ptr<WritableFile>* result, |
| 1060 | const EnvOptions& options) override { |
| 1061 | return target_->ReopenWritableFile(fname, result, options); |
| 1062 | } |
| 1063 | Status ReuseWritableFile(const std::string& fname, |
| 1064 | const std::string& old_fname, |
| 1065 | std::unique_ptr<WritableFile>* r, |
| 1066 | const EnvOptions& options) override { |
| 1067 | return target_->ReuseWritableFile(fname, old_fname, r, options); |
| 1068 | } |
| 1069 | Status NewRandomRWFile(const std::string& fname, |
| 1070 | std::unique_ptr<RandomRWFile>* result, |
| 1071 | const EnvOptions& options) override { |
| 1072 | return target_->NewRandomRWFile(fname, result, options); |
| 1073 | } |
| 1074 | Status NewMemoryMappedFileBuffer( |
| 1075 | const std::string& fname, |
| 1076 | std::unique_ptr<MemoryMappedFileBuffer>* result) override { |
| 1077 | return target_->NewMemoryMappedFileBuffer(fname, result); |
| 1078 | } |
| 1079 | Status NewDirectory(const std::string& name, |
| 1080 | std::unique_ptr<Directory>* result) override { |
| 1081 | return target_->NewDirectory(name, result); |
| 1082 | } |
| 1083 | Status FileExists(const std::string& f) override { |
| 1084 | return target_->FileExists(f); |
| 1085 | } |
| 1086 | Status GetChildren(const std::string& dir, |
| 1087 | std::vector<std::string>* r) override { |
| 1088 | return target_->GetChildren(dir, r); |
| 1089 | } |
| 1090 | Status GetChildrenFileAttributes( |
| 1091 | const std::string& dir, std::vector<FileAttributes>* result) override { |
| 1092 | return target_->GetChildrenFileAttributes(dir, result); |
| 1093 | } |
| 1094 | Status DeleteFile(const std::string& f) override { |
| 1095 | return target_->DeleteFile(f); |
| 1096 | } |
| 1097 | Status Truncate(const std::string& fname, size_t size) override { |
| 1098 | return target_->Truncate(fname, size); |
| 1099 | } |
| 1100 | Status CreateDir(const std::string& d) override { |
| 1101 | return target_->CreateDir(d); |
| 1102 | } |
| 1103 | Status CreateDirIfMissing(const std::string& d) override { |
| 1104 | return target_->CreateDirIfMissing(d); |
| 1105 | } |
| 1106 | Status DeleteDir(const std::string& d) override { |
| 1107 | return target_->DeleteDir(d); |
| 1108 | } |
| 1109 | Status GetFileSize(const std::string& f, uint64_t* s) override { |
| 1110 | return target_->GetFileSize(f, s); |
| 1111 | } |
| 1112 | |
| 1113 | Status GetFileModificationTime(const std::string& fname, |
| 1114 | uint64_t* file_mtime) override { |
| 1115 | return target_->GetFileModificationTime(fname, file_mtime); |
| 1116 | } |
| 1117 | |
| 1118 | Status RenameFile(const std::string& s, const std::string& t) override { |
| 1119 | return target_->RenameFile(s, t); |
| 1120 | } |
| 1121 | |
| 1122 | Status LinkFile(const std::string& s, const std::string& t) override { |
| 1123 | return target_->LinkFile(s, t); |
| 1124 | } |
| 1125 | |
| 1126 | Status NumFileLinks(const std::string& fname, uint64_t* count) override { |
| 1127 | return target_->NumFileLinks(fname, count); |
| 1128 | } |
| 1129 | |
| 1130 | Status AreFilesSame(const std::string& first, const std::string& second, |
| 1131 | bool* res) override { |
| 1132 | return target_->AreFilesSame(first, second, res); |
| 1133 | } |
| 1134 | |
| 1135 | Status LockFile(const std::string& f, FileLock** l) override { |
| 1136 | return target_->LockFile(f, l); |
| 1137 | } |
| 1138 | |
| 1139 | Status UnlockFile(FileLock* l) override { return target_->UnlockFile(l); } |
| 1140 | |
| 1141 | void Schedule(void (*f)(void* arg), void* a, Priority pri, |
| 1142 | void* tag = nullptr, void (*u)(void* arg) = nullptr) override { |
| 1143 | return target_->Schedule(f, a, pri, tag, u); |
| 1144 | } |
| 1145 | |
| 1146 | int UnSchedule(void* tag, Priority pri) override { |
| 1147 | return target_->UnSchedule(tag, pri); |
| 1148 | } |
| 1149 | |
| 1150 | void StartThread(void (*f)(void*), void* a) override { |
| 1151 | return target_->StartThread(f, a); |
| 1152 | } |
| 1153 | void WaitForJoin() override { return target_->WaitForJoin(); } |
| 1154 | unsigned int GetThreadPoolQueueLen(Priority pri = LOW) const override { |
| 1155 | return target_->GetThreadPoolQueueLen(pri); |
| 1156 | } |
| 1157 | Status GetTestDirectory(std::string* path) override { |
| 1158 | return target_->GetTestDirectory(path); |
| 1159 | } |
| 1160 | Status NewLogger(const std::string& fname, |
| 1161 | std::shared_ptr<Logger>* result) override { |
| 1162 | return target_->NewLogger(fname, result); |
| 1163 | } |
| 1164 | uint64_t NowMicros() override { return target_->NowMicros(); } |
| 1165 | uint64_t NowNanos() override { return target_->NowNanos(); } |
| 1166 | uint64_t NowCPUNanos() override { return target_->NowCPUNanos(); } |
| 1167 | |
| 1168 | void SleepForMicroseconds(int micros) override { |
| 1169 | target_->SleepForMicroseconds(micros); |
| 1170 | } |
| 1171 | Status GetHostName(char* name, uint64_t len) override { |
| 1172 | return target_->GetHostName(name, len); |
| 1173 | } |
| 1174 | Status GetCurrentTime(int64_t* unix_time) override { |
| 1175 | return target_->GetCurrentTime(unix_time); |
| 1176 | } |
| 1177 | Status GetAbsolutePath(const std::string& db_path, |
| 1178 | std::string* output_path) override { |
| 1179 | return target_->GetAbsolutePath(db_path, output_path); |
| 1180 | } |
| 1181 | void SetBackgroundThreads(int num, Priority pri) override { |
| 1182 | return target_->SetBackgroundThreads(num, pri); |
| 1183 | } |
| 1184 | int GetBackgroundThreads(Priority pri) override { |
| 1185 | return target_->GetBackgroundThreads(pri); |
| 1186 | } |
| 1187 | |
| 1188 | Status SetAllowNonOwnerAccess(bool allow_non_owner_access) override { |
| 1189 | return target_->SetAllowNonOwnerAccess(allow_non_owner_access); |
| 1190 | } |
| 1191 | |
| 1192 | void IncBackgroundThreadsIfNeeded(int num, Priority pri) override { |
| 1193 | return target_->IncBackgroundThreadsIfNeeded(num, pri); |
| 1194 | } |
| 1195 | |
| 1196 | void LowerThreadPoolIOPriority(Priority pool = LOW) override { |
| 1197 | target_->LowerThreadPoolIOPriority(pool); |
| 1198 | } |
| 1199 | |
| 1200 | void LowerThreadPoolCPUPriority(Priority pool = LOW) override { |
| 1201 | target_->LowerThreadPoolCPUPriority(pool); |
| 1202 | } |
| 1203 | |
| 1204 | std::string TimeToString(uint64_t time) override { |
| 1205 | return target_->TimeToString(time); |
| 1206 | } |
| 1207 | |
| 1208 | Status GetThreadList(std::vector<ThreadStatus>* thread_list) override { |
| 1209 | return target_->GetThreadList(thread_list); |
| 1210 | } |
| 1211 | |
| 1212 | ThreadStatusUpdater* GetThreadStatusUpdater() const override { |
| 1213 | return target_->GetThreadStatusUpdater(); |
| 1214 | } |
| 1215 | |
| 1216 | uint64_t GetThreadID() const override { return target_->GetThreadID(); } |
| 1217 | |
| 1218 | std::string GenerateUniqueId() override { |
| 1219 | return target_->GenerateUniqueId(); |
| 1220 | } |
| 1221 | |
| 1222 | EnvOptions OptimizeForLogRead(const EnvOptions& env_options) const override { |
| 1223 | return target_->OptimizeForLogRead(env_options); |
| 1224 | } |
| 1225 | EnvOptions OptimizeForManifestRead( |
| 1226 | const EnvOptions& env_options) const override { |
| 1227 | return target_->OptimizeForManifestRead(env_options); |
| 1228 | } |
| 1229 | EnvOptions OptimizeForLogWrite(const EnvOptions& env_options, |
| 1230 | const DBOptions& db_options) const override { |
| 1231 | return target_->OptimizeForLogWrite(env_options, db_options); |
| 1232 | } |
| 1233 | EnvOptions OptimizeForManifestWrite( |
| 1234 | const EnvOptions& env_options) const override { |
| 1235 | return target_->OptimizeForManifestWrite(env_options); |
| 1236 | } |
| 1237 | EnvOptions OptimizeForCompactionTableWrite( |
| 1238 | const EnvOptions& env_options, |
| 1239 | const ImmutableDBOptions& immutable_ops) const override { |
| 1240 | return target_->OptimizeForCompactionTableWrite(env_options, immutable_ops); |
| 1241 | } |
| 1242 | EnvOptions OptimizeForCompactionTableRead( |
| 1243 | const EnvOptions& env_options, |
| 1244 | const ImmutableDBOptions& db_options) const override { |
| 1245 | return target_->OptimizeForCompactionTableRead(env_options, db_options); |
| 1246 | } |
| 1247 | Status GetFreeSpace(const std::string& path, uint64_t* diskfree) override { |
| 1248 | return target_->GetFreeSpace(path, diskfree); |
| 1249 | } |
| 1250 | |
| 1251 | private: |
| 1252 | Env* target_; |
| 1253 | }; |
| 1254 | |
| 1255 | class SequentialFileWrapper : public SequentialFile { |
| 1256 | public: |
| 1257 | explicit SequentialFileWrapper(SequentialFile* target) : target_(target) {} |
| 1258 | |
| 1259 | Status Read(size_t n, Slice* result, char* scratch) override { |
| 1260 | return target_->Read(n, result, scratch); |
| 1261 | } |
| 1262 | Status Skip(uint64_t n) override { return target_->Skip(n); } |
| 1263 | bool use_direct_io() const override { return target_->use_direct_io(); } |
| 1264 | size_t GetRequiredBufferAlignment() const override { |
| 1265 | return target_->GetRequiredBufferAlignment(); |
| 1266 | } |
| 1267 | Status InvalidateCache(size_t offset, size_t length) override { |
| 1268 | return target_->InvalidateCache(offset, length); |
| 1269 | } |
| 1270 | Status PositionedRead(uint64_t offset, size_t n, Slice* result, |
| 1271 | char* scratch) override { |
| 1272 | return target_->PositionedRead(offset, n, result, scratch); |
| 1273 | } |
| 1274 | |
| 1275 | private: |
| 1276 | SequentialFile* target_; |
| 1277 | }; |
| 1278 | |
| 1279 | class RandomAccessFileWrapper : public RandomAccessFile { |
| 1280 | public: |
| 1281 | explicit RandomAccessFileWrapper(RandomAccessFile* target) |
| 1282 | : target_(target) {} |
| 1283 | |
| 1284 | Status Read(uint64_t offset, size_t n, Slice* result, |
| 1285 | char* scratch) const override { |
| 1286 | return target_->Read(offset, n, result, scratch); |
| 1287 | } |
| 1288 | Status Prefetch(uint64_t offset, size_t n) override { |
| 1289 | return target_->Prefetch(offset, n); |
| 1290 | } |
| 1291 | size_t GetUniqueId(char* id, size_t max_size) const override { |
| 1292 | return target_->GetUniqueId(id, max_size); |
| 1293 | }; |
| 1294 | void Hint(AccessPattern pattern) override { target_->Hint(pattern); } |
| 1295 | bool use_direct_io() const override { return target_->use_direct_io(); } |
| 1296 | size_t GetRequiredBufferAlignment() const override { |
| 1297 | return target_->GetRequiredBufferAlignment(); |
| 1298 | } |
| 1299 | Status InvalidateCache(size_t offset, size_t length) override { |
| 1300 | return target_->InvalidateCache(offset, length); |
| 1301 | } |
| 1302 | |
| 1303 | private: |
| 1304 | RandomAccessFile* target_; |
| 1305 | }; |
| 1306 | |
| 1307 | class WritableFileWrapper : public WritableFile { |
| 1308 | public: |
| 1309 | explicit WritableFileWrapper(WritableFile* t) : target_(t) {} |
| 1310 | |
| 1311 | Status Append(const Slice& data) override { return target_->Append(data); } |
| 1312 | Status PositionedAppend(const Slice& data, uint64_t offset) override { |
| 1313 | return target_->PositionedAppend(data, offset); |
| 1314 | } |
| 1315 | Status Truncate(uint64_t size) override { return target_->Truncate(size); } |
| 1316 | Status Close() override { return target_->Close(); } |
| 1317 | Status Flush() override { return target_->Flush(); } |
| 1318 | Status Sync() override { return target_->Sync(); } |
| 1319 | Status Fsync() override { return target_->Fsync(); } |
| 1320 | bool IsSyncThreadSafe() const override { return target_->IsSyncThreadSafe(); } |
| 1321 | |
| 1322 | bool use_direct_io() const override { return target_->use_direct_io(); } |
| 1323 | |
| 1324 | size_t GetRequiredBufferAlignment() const override { |
| 1325 | return target_->GetRequiredBufferAlignment(); |
| 1326 | } |
| 1327 | |
| 1328 | void SetIOPriority(Env::IOPriority pri) override { |
| 1329 | target_->SetIOPriority(pri); |
| 1330 | } |
| 1331 | |
| 1332 | Env::IOPriority GetIOPriority() override { return target_->GetIOPriority(); } |
| 1333 | |
| 1334 | void SetWriteLifeTimeHint(Env::WriteLifeTimeHint hint) override { |
| 1335 | target_->SetWriteLifeTimeHint(hint); |
| 1336 | } |
| 1337 | |
| 1338 | Env::WriteLifeTimeHint GetWriteLifeTimeHint() override { |
| 1339 | return target_->GetWriteLifeTimeHint(); |
| 1340 | } |
| 1341 | |
| 1342 | uint64_t GetFileSize() override { return target_->GetFileSize(); } |
| 1343 | |
| 1344 | void SetPreallocationBlockSize(size_t size) override { |
| 1345 | target_->SetPreallocationBlockSize(size); |
| 1346 | } |
| 1347 | |
| 1348 | void GetPreallocationStatus(size_t* block_size, |
| 1349 | size_t* last_allocated_block) override { |
| 1350 | target_->GetPreallocationStatus(block_size, last_allocated_block); |
| 1351 | } |
| 1352 | |
| 1353 | size_t GetUniqueId(char* id, size_t max_size) const override { |
| 1354 | return target_->GetUniqueId(id, max_size); |
| 1355 | } |
| 1356 | |
| 1357 | Status InvalidateCache(size_t offset, size_t length) override { |
| 1358 | return target_->InvalidateCache(offset, length); |
| 1359 | } |
| 1360 | |
| 1361 | Status RangeSync(uint64_t offset, uint64_t nbytes) override { |
| 1362 | return target_->RangeSync(offset, nbytes); |
| 1363 | } |
| 1364 | |
| 1365 | void PrepareWrite(size_t offset, size_t len) override { |
| 1366 | target_->PrepareWrite(offset, len); |
| 1367 | } |
| 1368 | |
| 1369 | Status Allocate(uint64_t offset, uint64_t len) override { |
| 1370 | return target_->Allocate(offset, len); |
| 1371 | } |
| 1372 | |
| 1373 | private: |
| 1374 | WritableFile* target_; |
| 1375 | }; |
| 1376 | |
| 1377 | class RandomRWFileWrapper : public RandomRWFile { |
| 1378 | public: |
| 1379 | explicit RandomRWFileWrapper(RandomRWFile* target) : target_(target) {} |
| 1380 | |
| 1381 | bool use_direct_io() const override { return target_->use_direct_io(); } |
| 1382 | size_t GetRequiredBufferAlignment() const override { |
| 1383 | return target_->GetRequiredBufferAlignment(); |
| 1384 | } |
| 1385 | Status Write(uint64_t offset, const Slice& data) override { |
| 1386 | return target_->Write(offset, data); |
| 1387 | } |
| 1388 | Status Read(uint64_t offset, size_t n, Slice* result, |
| 1389 | char* scratch) const override { |
| 1390 | return target_->Read(offset, n, result, scratch); |
| 1391 | } |
| 1392 | Status Flush() override { return target_->Flush(); } |
| 1393 | Status Sync() override { return target_->Sync(); } |
| 1394 | Status Fsync() override { return target_->Fsync(); } |
| 1395 | Status Close() override { return target_->Close(); } |
| 1396 | |
| 1397 | private: |
| 1398 | RandomRWFile* target_; |
| 1399 | }; |
| 1400 | |
| 1401 | class DirectoryWrapper : public Directory { |
| 1402 | public: |
| 1403 | explicit DirectoryWrapper(Directory* target) : target_(target) {} |
| 1404 | |
| 1405 | Status Fsync() override { return target_->Fsync(); } |
| 1406 | size_t GetUniqueId(char* id, size_t max_size) const override { |
| 1407 | return target_->GetUniqueId(id, max_size); |
| 1408 | } |
| 1409 | |
| 1410 | private: |
| 1411 | Directory* target_; |
| 1412 | }; |
| 1413 | |
| 1414 | class LoggerWrapper : public Logger { |
| 1415 | public: |
| 1416 | explicit LoggerWrapper(Logger* target) : target_(target) {} |
| 1417 | |
| 1418 | Status Close() override { return target_->Close(); } |
| 1419 | void LogHeader(const char* format, va_list ap) override { |
| 1420 | return target_->LogHeader(format, ap); |
| 1421 | } |
| 1422 | void Logv(const char* format, va_list ap) override { |
| 1423 | return target_->Logv(format, ap); |
| 1424 | } |
| 1425 | void Logv(const InfoLogLevel log_level, const char* format, |
| 1426 | va_list ap) override { |
| 1427 | return target_->Logv(log_level, format, ap); |
| 1428 | } |
| 1429 | size_t GetLogFileSize() const override { return target_->GetLogFileSize(); } |
| 1430 | void Flush() override { return target_->Flush(); } |
| 1431 | InfoLogLevel GetInfoLogLevel() const override { |
| 1432 | return target_->GetInfoLogLevel(); |
| 1433 | } |
| 1434 | void SetInfoLogLevel(const InfoLogLevel log_level) override { |
| 1435 | return target_->SetInfoLogLevel(log_level); |
| 1436 | } |
| 1437 | |
| 1438 | private: |
| 1439 | Logger* target_; |
| 1440 | }; |
| 1441 | |
| 1442 | // Returns a new environment that stores its data in memory and delegates |
| 1443 | // all non-file-storage tasks to base_env. The caller must delete the result |
| 1444 | // when it is no longer needed. |
| 1445 | // *base_env must remain live while the result is in use. |
| 1446 | Env* NewMemEnv(Env* base_env); |
| 1447 | |
| 1448 | // Returns a new environment that is used for HDFS environment. |
| 1449 | // This is a factory method for HdfsEnv declared in hdfs/env_hdfs.h |
| 1450 | Status NewHdfsEnv(Env** hdfs_env, const std::string& fsname); |
| 1451 | |
| 1452 | // Returns a new environment that measures function call times for filesystem |
| 1453 | // operations, reporting results to variables in PerfContext. |
| 1454 | // This is a factory method for TimedEnv defined in utilities/env_timed.cc. |
| 1455 | Env* NewTimedEnv(Env* base_env); |
| 1456 | |
| 1457 | } // namespace rocksdb |