| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571 | //  Copyright (c) 2011-present, Facebook, Inc.  All rights reserved.//  This source code is licensed under both the GPLv2 (found in the//  COPYING file in the root directory) and Apache 2.0 License//  (found in the LICENSE.Apache file in the root directory).//// Copyright (c) 2011 The LevelDB Authors. All rights reserved.// Use of this source code is governed by a BSD-style license that can be// found in the LICENSE file. See the AUTHORS file for names of contributors.#ifndef ROCKSDB_LITE#include <stdlib.h>#include <map>#include <string>#include <vector>#include "db/db_impl/db_impl.h"#include "db/db_test_util.h"#include "db/version_set.h"#include "db/write_batch_internal.h"#include "file/filename.h"#include "port/stack_trace.h"#include "rocksdb/db.h"#include "rocksdb/env.h"#include "rocksdb/transaction_log.h"#include "test_util/sync_point.h"#include "test_util/testharness.h"#include "test_util/testutil.h"#include "util/string_util.h"namespace ROCKSDB_NAMESPACE {class DeleteFileTest : public DBTestBase { public:  const int numlevels_;  const std::string wal_dir_;  DeleteFileTest()      : DBTestBase("/deletefile_test"),        numlevels_(7),        wal_dir_(dbname_ + "/wal_files") {}  void SetOptions(Options* options) {    assert(options);    options->delete_obsolete_files_period_micros = 0;  // always do full purge    options->enable_thread_tracking = true;    options->write_buffer_size = 1024 * 1024 * 1000;    options->target_file_size_base = 1024 * 1024 * 1000;    options->max_bytes_for_level_base = 1024 * 1024 * 1000;    options->WAL_ttl_seconds = 300;     // Used to test log files    options->WAL_size_limit_MB = 1024;  // Used to test log files    options->wal_dir = wal_dir_;  }  void AddKeys(int numkeys, int startkey = 0) {    WriteOptions options;    options.sync = false;    ReadOptions roptions;    for (int i = startkey; i < (numkeys + startkey) ; i++) {      std::string temp = ToString(i);      Slice key(temp);      Slice value(temp);      ASSERT_OK(db_->Put(options, key, value));    }  }  int numKeysInLevels(    std::vector<LiveFileMetaData> &metadata,    std::vector<int> *keysperlevel = nullptr) {    if (keysperlevel != nullptr) {      keysperlevel->resize(numlevels_);    }    int numKeys = 0;    for (size_t i = 0; i < metadata.size(); i++) {      int startkey = atoi(metadata[i].smallestkey.c_str());      int endkey = atoi(metadata[i].largestkey.c_str());      int numkeysinfile = (endkey - startkey + 1);      numKeys += numkeysinfile;      if (keysperlevel != nullptr) {        (*keysperlevel)[(int)metadata[i].level] += numkeysinfile;      }      fprintf(stderr, "level %d name %s smallest %s largest %s\n",              metadata[i].level, metadata[i].name.c_str(),              metadata[i].smallestkey.c_str(),              metadata[i].largestkey.c_str());    }    return numKeys;  }  void CreateTwoLevels() {    AddKeys(50000, 10000);    ASSERT_OK(dbfull()->TEST_FlushMemTable());    ASSERT_OK(dbfull()->TEST_WaitForFlushMemTable());    for (int i = 0; i < 2; ++i) {      ASSERT_OK(dbfull()->TEST_CompactRange(i, nullptr, nullptr));    }    AddKeys(50000, 10000);    ASSERT_OK(dbfull()->TEST_FlushMemTable());    ASSERT_OK(dbfull()->TEST_WaitForFlushMemTable());    ASSERT_OK(dbfull()->TEST_CompactRange(0, nullptr, nullptr));  }  void CheckFileTypeCounts(const std::string& dir, int required_log,                           int required_sst, int required_manifest) {    std::vector<std::string> filenames;    env_->GetChildren(dir, &filenames);    int log_cnt = 0, sst_cnt = 0, manifest_cnt = 0;    for (auto file : filenames) {      uint64_t number;      FileType type;      if (ParseFileName(file, &number, &type)) {        log_cnt += (type == kLogFile);        sst_cnt += (type == kTableFile);        manifest_cnt += (type == kDescriptorFile);      }    }    ASSERT_EQ(required_log, log_cnt);    ASSERT_EQ(required_sst, sst_cnt);    ASSERT_EQ(required_manifest, manifest_cnt);  }  static void DoSleep(void* arg) {    auto test = reinterpret_cast<DeleteFileTest*>(arg);    test->env_->SleepForMicroseconds(2 * 1000 * 1000);  }  // An empty job to guard all jobs are processed  static void GuardFinish(void* /*arg*/) {    TEST_SYNC_POINT("DeleteFileTest::GuardFinish");  }};TEST_F(DeleteFileTest, AddKeysAndQueryLevels) {  Options options = CurrentOptions();  SetOptions(&options);  Destroy(options);  options.create_if_missing = true;  Reopen(options);  CreateTwoLevels();  std::vector<LiveFileMetaData> metadata;  db_->GetLiveFilesMetaData(&metadata);  std::string level1file = "";  int level1keycount = 0;  std::string level2file = "";  int level2keycount = 0;  int level1index = 0;  int level2index = 1;  ASSERT_EQ((int)metadata.size(), 2);  if (metadata[0].level == 2) {    level1index = 1;    level2index = 0;  }  level1file = metadata[level1index].name;  int startkey = atoi(metadata[level1index].smallestkey.c_str());  int endkey = atoi(metadata[level1index].largestkey.c_str());  level1keycount = (endkey - startkey + 1);  level2file = metadata[level2index].name;  startkey = atoi(metadata[level2index].smallestkey.c_str());  endkey = atoi(metadata[level2index].largestkey.c_str());  level2keycount = (endkey - startkey + 1);  // COntrolled setup. Levels 1 and 2 should both have 50K files.  // This is a little fragile as it depends on the current  // compaction heuristics.  ASSERT_EQ(level1keycount, 50000);  ASSERT_EQ(level2keycount, 50000);  Status status = db_->DeleteFile("0.sst");  ASSERT_TRUE(status.IsInvalidArgument());  // intermediate level files cannot be deleted.  status = db_->DeleteFile(level1file);  ASSERT_TRUE(status.IsInvalidArgument());  // Lowest level file deletion should succeed.  ASSERT_OK(db_->DeleteFile(level2file));}TEST_F(DeleteFileTest, PurgeObsoleteFilesTest) {  Options options = CurrentOptions();  SetOptions(&options);  Destroy(options);  options.create_if_missing = true;  Reopen(options);  CreateTwoLevels();  // there should be only one (empty) log file because CreateTwoLevels()  // flushes the memtables to disk  CheckFileTypeCounts(wal_dir_, 1, 0, 0);  // 2 ssts, 1 manifest  CheckFileTypeCounts(dbname_, 0, 2, 1);  std::string first("0"), last("999999");  CompactRangeOptions compact_options;  compact_options.change_level = true;  compact_options.target_level = 2;  Slice first_slice(first), last_slice(last);  db_->CompactRange(compact_options, &first_slice, &last_slice);  // 1 sst after compaction  CheckFileTypeCounts(dbname_, 0, 1, 1);  // this time, we keep an iterator alive  Reopen(options);  Iterator *itr = nullptr;  CreateTwoLevels();  itr = db_->NewIterator(ReadOptions());  db_->CompactRange(compact_options, &first_slice, &last_slice);  // 3 sst after compaction with live iterator  CheckFileTypeCounts(dbname_, 0, 3, 1);  delete itr;  // 1 sst after iterator deletion  CheckFileTypeCounts(dbname_, 0, 1, 1);}TEST_F(DeleteFileTest, BackgroundPurgeIteratorTest) {  Options options = CurrentOptions();  SetOptions(&options);  Destroy(options);  options.create_if_missing = true;  Reopen(options);  std::string first("0"), last("999999");  CompactRangeOptions compact_options;  compact_options.change_level = true;  compact_options.target_level = 2;  Slice first_slice(first), last_slice(last);  // We keep an iterator alive  Iterator* itr = nullptr;  CreateTwoLevels();  ReadOptions read_options;  read_options.background_purge_on_iterator_cleanup = true;  itr = db_->NewIterator(read_options);  db_->CompactRange(compact_options, &first_slice, &last_slice);  // 3 sst after compaction with live iterator  CheckFileTypeCounts(dbname_, 0, 3, 1);  test::SleepingBackgroundTask sleeping_task_before;  env_->Schedule(&test::SleepingBackgroundTask::DoSleepTask,                 &sleeping_task_before, Env::Priority::HIGH);  delete itr;  test::SleepingBackgroundTask sleeping_task_after;  env_->Schedule(&test::SleepingBackgroundTask::DoSleepTask,                 &sleeping_task_after, Env::Priority::HIGH);  // Make sure no purges are executed foreground  CheckFileTypeCounts(dbname_, 0, 3, 1);  sleeping_task_before.WakeUp();  sleeping_task_before.WaitUntilDone();  // Make sure all background purges are executed  sleeping_task_after.WakeUp();  sleeping_task_after.WaitUntilDone();  // 1 sst after iterator deletion  CheckFileTypeCounts(dbname_, 0, 1, 1);}TEST_F(DeleteFileTest, BackgroundPurgeCFDropTest) {  Options options = CurrentOptions();  SetOptions(&options);  Destroy(options);  options.create_if_missing = true;  Reopen(options);  auto do_test = [&](bool bg_purge) {    ColumnFamilyOptions co;    co.max_write_buffer_size_to_maintain =        static_cast<int64_t>(co.write_buffer_size);    WriteOptions wo;    FlushOptions fo;    ColumnFamilyHandle* cfh = nullptr;    ASSERT_OK(db_->CreateColumnFamily(co, "dropme", &cfh));    ASSERT_OK(db_->Put(wo, cfh, "pika", "chu"));    ASSERT_OK(db_->Flush(fo, cfh));    // Expect 1 sst file.    CheckFileTypeCounts(dbname_, 0, 1, 1);    ASSERT_OK(db_->DropColumnFamily(cfh));    // Still 1 file, it won't be deleted while ColumnFamilyHandle is alive.    CheckFileTypeCounts(dbname_, 0, 1, 1);    delete cfh;    test::SleepingBackgroundTask sleeping_task_after;    env_->Schedule(&test::SleepingBackgroundTask::DoSleepTask,                   &sleeping_task_after, Env::Priority::HIGH);    // If background purge is enabled, the file should still be there.    CheckFileTypeCounts(dbname_, 0, bg_purge ? 1 : 0, 1);    TEST_SYNC_POINT("DeleteFileTest::BackgroundPurgeCFDropTest:1");    // Execute background purges.    sleeping_task_after.WakeUp();    sleeping_task_after.WaitUntilDone();    // The file should have been deleted.    CheckFileTypeCounts(dbname_, 0, 0, 1);  };  {    SCOPED_TRACE("avoid_unnecessary_blocking_io = false");    do_test(false);  }  SyncPoint::GetInstance()->DisableProcessing();  SyncPoint::GetInstance()->ClearAllCallBacks();  SyncPoint::GetInstance()->LoadDependency(      {{"DeleteFileTest::BackgroundPurgeCFDropTest:1",        "DBImpl::BGWorkPurge:start"}});  SyncPoint::GetInstance()->EnableProcessing();  options.avoid_unnecessary_blocking_io = true;  options.create_if_missing = false;  Reopen(options);  {    SCOPED_TRACE("avoid_unnecessary_blocking_io = true");    do_test(true);  }}// This test is to reproduce a bug that read invalid ReadOption in iterator// cleanup functionTEST_F(DeleteFileTest, BackgroundPurgeCopyOptions) {  Options options = CurrentOptions();  SetOptions(&options);  Destroy(options);  options.create_if_missing = true;  Reopen(options);  std::string first("0"), last("999999");  CompactRangeOptions compact_options;  compact_options.change_level = true;  compact_options.target_level = 2;  Slice first_slice(first), last_slice(last);  // We keep an iterator alive  Iterator* itr = nullptr;  CreateTwoLevels();  {    ReadOptions read_options;    read_options.background_purge_on_iterator_cleanup = true;    itr = db_->NewIterator(read_options);    // ReadOptions is deleted, but iterator cleanup function should not be    // affected  }  db_->CompactRange(compact_options, &first_slice, &last_slice);  // 3 sst after compaction with live iterator  CheckFileTypeCounts(dbname_, 0, 3, 1);  delete itr;  test::SleepingBackgroundTask sleeping_task_after;  env_->Schedule(&test::SleepingBackgroundTask::DoSleepTask,                 &sleeping_task_after, Env::Priority::HIGH);  // Make sure all background purges are executed  sleeping_task_after.WakeUp();  sleeping_task_after.WaitUntilDone();  // 1 sst after iterator deletion  CheckFileTypeCounts(dbname_, 0, 1, 1);}TEST_F(DeleteFileTest, BackgroundPurgeTestMultipleJobs) {  Options options = CurrentOptions();  SetOptions(&options);  Destroy(options);  options.create_if_missing = true;  Reopen(options);  std::string first("0"), last("999999");  CompactRangeOptions compact_options;  compact_options.change_level = true;  compact_options.target_level = 2;  Slice first_slice(first), last_slice(last);  // We keep an iterator alive  CreateTwoLevels();  ReadOptions read_options;  read_options.background_purge_on_iterator_cleanup = true;  Iterator* itr1 = db_->NewIterator(read_options);  CreateTwoLevels();  Iterator* itr2 = db_->NewIterator(read_options);  db_->CompactRange(compact_options, &first_slice, &last_slice);  // 5 sst files after 2 compactions with 2 live iterators  CheckFileTypeCounts(dbname_, 0, 5, 1);  ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->DisableProcessing();  // ~DBImpl should wait until all BGWorkPurge are finished  ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->LoadDependency(      {{"DBImpl::~DBImpl:WaitJob", "DBImpl::BGWorkPurge"},       {"DeleteFileTest::GuardFinish",        "DeleteFileTest::BackgroundPurgeTestMultipleJobs:DBClose"}});  ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->EnableProcessing();  delete itr1;  env_->Schedule(&DeleteFileTest::DoSleep, this, Env::Priority::HIGH);  delete itr2;  env_->Schedule(&DeleteFileTest::GuardFinish, nullptr, Env::Priority::HIGH);  Close();  TEST_SYNC_POINT("DeleteFileTest::BackgroundPurgeTestMultipleJobs:DBClose");  // 1 sst after iterator deletion  CheckFileTypeCounts(dbname_, 0, 1, 1);}TEST_F(DeleteFileTest, DeleteFileWithIterator) {  Options options = CurrentOptions();  SetOptions(&options);  Destroy(options);  options.create_if_missing = true;  Reopen(options);  CreateTwoLevels();  ReadOptions read_options;  Iterator* it = db_->NewIterator(read_options);  std::vector<LiveFileMetaData> metadata;  db_->GetLiveFilesMetaData(&metadata);  std::string level2file;  ASSERT_EQ(metadata.size(), static_cast<size_t>(2));  if (metadata[0].level == 1) {    level2file = metadata[1].name;  } else {    level2file = metadata[0].name;  }  Status status = db_->DeleteFile(level2file);  fprintf(stdout, "Deletion status %s: %s\n",          level2file.c_str(), status.ToString().c_str());  ASSERT_TRUE(status.ok());  it->SeekToFirst();  int numKeysIterated = 0;  while(it->Valid()) {    numKeysIterated++;    it->Next();  }  ASSERT_EQ(numKeysIterated, 50000);  delete it;}TEST_F(DeleteFileTest, DeleteLogFiles) {  Options options = CurrentOptions();  SetOptions(&options);  Destroy(options);  options.create_if_missing = true;  Reopen(options);  AddKeys(10, 0);  VectorLogPtr logfiles;  db_->GetSortedWalFiles(logfiles);  ASSERT_GT(logfiles.size(), 0UL);  // Take the last log file which is expected to be alive and try to delete it  // Should not succeed because live logs are not allowed to be deleted  std::unique_ptr<LogFile> alive_log = std::move(logfiles.back());  ASSERT_EQ(alive_log->Type(), kAliveLogFile);  ASSERT_OK(env_->FileExists(wal_dir_ + "/" + alive_log->PathName()));  fprintf(stdout, "Deleting alive log file %s\n",          alive_log->PathName().c_str());  ASSERT_TRUE(!db_->DeleteFile(alive_log->PathName()).ok());  ASSERT_OK(env_->FileExists(wal_dir_ + "/" + alive_log->PathName()));  logfiles.clear();  // Call Flush to bring about a new working log file and add more keys  // Call Flush again to flush out memtable and move alive log to archived log  // and try to delete the archived log file  FlushOptions fopts;  db_->Flush(fopts);  AddKeys(10, 0);  db_->Flush(fopts);  db_->GetSortedWalFiles(logfiles);  ASSERT_GT(logfiles.size(), 0UL);  std::unique_ptr<LogFile> archived_log = std::move(logfiles.front());  ASSERT_EQ(archived_log->Type(), kArchivedLogFile);  ASSERT_OK(env_->FileExists(wal_dir_ + "/" + archived_log->PathName()));  fprintf(stdout, "Deleting archived log file %s\n",          archived_log->PathName().c_str());  ASSERT_OK(db_->DeleteFile(archived_log->PathName()));  ASSERT_EQ(Status::NotFound(),            env_->FileExists(wal_dir_ + "/" + archived_log->PathName()));}TEST_F(DeleteFileTest, DeleteNonDefaultColumnFamily) {  Options options = CurrentOptions();  SetOptions(&options);  Destroy(options);  options.create_if_missing = true;  Reopen(options);  CreateAndReopenWithCF({"new_cf"}, options);  Random rnd(5);  for (int i = 0; i < 1000; ++i) {    ASSERT_OK(db_->Put(WriteOptions(), handles_[1], test::RandomKey(&rnd, 10),                       test::RandomKey(&rnd, 10)));  }  ASSERT_OK(db_->Flush(FlushOptions(), handles_[1]));  for (int i = 0; i < 1000; ++i) {    ASSERT_OK(db_->Put(WriteOptions(), handles_[1], test::RandomKey(&rnd, 10),                       test::RandomKey(&rnd, 10)));  }  ASSERT_OK(db_->Flush(FlushOptions(), handles_[1]));  std::vector<LiveFileMetaData> metadata;  db_->GetLiveFilesMetaData(&metadata);  ASSERT_EQ(2U, metadata.size());  ASSERT_EQ("new_cf", metadata[0].column_family_name);  ASSERT_EQ("new_cf", metadata[1].column_family_name);  auto old_file = metadata[0].smallest_seqno < metadata[1].smallest_seqno                      ? metadata[0].name                      : metadata[1].name;  auto new_file = metadata[0].smallest_seqno > metadata[1].smallest_seqno                      ? metadata[0].name                      : metadata[1].name;  ASSERT_TRUE(db_->DeleteFile(new_file).IsInvalidArgument());  ASSERT_OK(db_->DeleteFile(old_file));  {    std::unique_ptr<Iterator> itr(db_->NewIterator(ReadOptions(), handles_[1]));    int count = 0;    for (itr->SeekToFirst(); itr->Valid(); itr->Next()) {      ASSERT_OK(itr->status());      ++count;    }    ASSERT_EQ(count, 1000);  }  Close();  ReopenWithColumnFamilies({kDefaultColumnFamilyName, "new_cf"}, options);  {    std::unique_ptr<Iterator> itr(db_->NewIterator(ReadOptions(), handles_[1]));    int count = 0;    for (itr->SeekToFirst(); itr->Valid(); itr->Next()) {      ASSERT_OK(itr->status());      ++count;    }    ASSERT_EQ(count, 1000);  }}}  // namespace ROCKSDB_NAMESPACE#ifdef ROCKSDB_UNITTESTS_WITH_CUSTOM_OBJECTS_FROM_STATIC_LIBSextern "C" {void RegisterCustomObjects(int argc, char** argv);}#elsevoid RegisterCustomObjects(int /*argc*/, char** /*argv*/) {}#endif  // !ROCKSDB_UNITTESTS_WITH_CUSTOM_OBJECTS_FROM_STATIC_LIBSint main(int argc, char** argv) {  ROCKSDB_NAMESPACE::port::InstallStackTraceHandler();  ::testing::InitGoogleTest(&argc, argv);  RegisterCustomObjects(argc, argv);  return RUN_ALL_TESTS();}#else#include <stdio.h>int main(int /*argc*/, char** /*argv*/) {  fprintf(stderr,          "SKIPPED as DBImpl::DeleteFile is not supported in ROCKSDB_LITE\n");  return 0;}#endif  // !ROCKSDB_LITE
 |