db_universal_compaction_test.cc 77 KB

1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798991001011021031041051061071081091101111121131141151161171181191201211221231241251261271281291301311321331341351361371381391401411421431441451461471481491501511521531541551561571581591601611621631641651661671681691701711721731741751761771781791801811821831841851861871881891901911921931941951961971981992002012022032042052062072082092102112122132142152162172182192202212222232242252262272282292302312322332342352362372382392402412422432442452462472482492502512522532542552562572582592602612622632642652662672682692702712722732742752762772782792802812822832842852862872882892902912922932942952962972982993003013023033043053063073083093103113123133143153163173183193203213223233243253263273283293303313323333343353363373383393403413423433443453463473483493503513523533543553563573583593603613623633643653663673683693703713723733743753763773783793803813823833843853863873883893903913923933943953963973983994004014024034044054064074084094104114124134144154164174184194204214224234244254264274284294304314324334344354364374384394404414424434444454464474484494504514524534544554564574584594604614624634644654664674684694704714724734744754764774784794804814824834844854864874884894904914924934944954964974984995005015025035045055065075085095105115125135145155165175185195205215225235245255265275285295305315325335345355365375385395405415425435445455465475485495505515525535545555565575585595605615625635645655665675685695705715725735745755765775785795805815825835845855865875885895905915925935945955965975985996006016026036046056066076086096106116126136146156166176186196206216226236246256266276286296306316326336346356366376386396406416426436446456466476486496506516526536546556566576586596606616626636646656666676686696706716726736746756766776786796806816826836846856866876886896906916926936946956966976986997007017027037047057067077087097107117127137147157167177187197207217227237247257267277287297307317327337347357367377387397407417427437447457467477487497507517527537547557567577587597607617627637647657667677687697707717727737747757767777787797807817827837847857867877887897907917927937947957967977987998008018028038048058068078088098108118128138148158168178188198208218228238248258268278288298308318328338348358368378388398408418428438448458468478488498508518528538548558568578588598608618628638648658668678688698708718728738748758768778788798808818828838848858868878888898908918928938948958968978988999009019029039049059069079089099109119129139149159169179189199209219229239249259269279289299309319329339349359369379389399409419429439449459469479489499509519529539549559569579589599609619629639649659669679689699709719729739749759769779789799809819829839849859869879889899909919929939949959969979989991000100110021003100410051006100710081009101010111012101310141015101610171018101910201021102210231024102510261027102810291030103110321033103410351036103710381039104010411042104310441045104610471048104910501051105210531054105510561057105810591060106110621063106410651066106710681069107010711072107310741075107610771078107910801081108210831084108510861087108810891090109110921093109410951096109710981099110011011102110311041105110611071108110911101111111211131114111511161117111811191120112111221123112411251126112711281129113011311132113311341135113611371138113911401141114211431144114511461147114811491150115111521153115411551156115711581159116011611162116311641165116611671168116911701171117211731174117511761177117811791180118111821183118411851186118711881189119011911192119311941195119611971198119912001201120212031204120512061207120812091210121112121213121412151216121712181219122012211222122312241225122612271228122912301231123212331234123512361237123812391240124112421243124412451246124712481249125012511252125312541255125612571258125912601261126212631264126512661267126812691270127112721273127412751276127712781279128012811282128312841285128612871288128912901291129212931294129512961297129812991300130113021303130413051306130713081309131013111312131313141315131613171318131913201321132213231324132513261327132813291330133113321333133413351336133713381339134013411342134313441345134613471348134913501351135213531354135513561357135813591360136113621363136413651366136713681369137013711372137313741375137613771378137913801381138213831384138513861387138813891390139113921393139413951396139713981399140014011402140314041405140614071408140914101411141214131414141514161417141814191420142114221423142414251426142714281429143014311432143314341435143614371438143914401441144214431444144514461447144814491450145114521453145414551456145714581459146014611462146314641465146614671468146914701471147214731474147514761477147814791480148114821483148414851486148714881489149014911492149314941495149614971498149915001501150215031504150515061507150815091510151115121513151415151516151715181519152015211522152315241525152615271528152915301531153215331534153515361537153815391540154115421543154415451546154715481549155015511552155315541555155615571558155915601561156215631564156515661567156815691570157115721573157415751576157715781579158015811582158315841585158615871588158915901591159215931594159515961597159815991600160116021603160416051606160716081609161016111612161316141615161616171618161916201621162216231624162516261627162816291630163116321633163416351636163716381639164016411642164316441645164616471648164916501651165216531654165516561657165816591660166116621663166416651666166716681669167016711672167316741675167616771678167916801681168216831684168516861687168816891690169116921693169416951696169716981699170017011702170317041705170617071708170917101711171217131714171517161717171817191720172117221723172417251726172717281729173017311732173317341735173617371738173917401741174217431744174517461747174817491750175117521753175417551756175717581759176017611762176317641765176617671768176917701771177217731774177517761777177817791780178117821783178417851786178717881789179017911792179317941795179617971798179918001801180218031804180518061807180818091810181118121813181418151816181718181819182018211822182318241825182618271828182918301831183218331834183518361837183818391840184118421843184418451846184718481849185018511852185318541855185618571858185918601861186218631864186518661867186818691870187118721873187418751876187718781879188018811882188318841885188618871888188918901891189218931894189518961897189818991900190119021903190419051906190719081909191019111912191319141915191619171918191919201921192219231924192519261927192819291930193119321933193419351936193719381939194019411942194319441945194619471948194919501951195219531954195519561957195819591960196119621963196419651966196719681969197019711972197319741975197619771978197919801981198219831984198519861987198819891990199119921993199419951996199719981999200020012002200320042005200620072008200920102011201220132014201520162017201820192020202120222023202420252026202720282029203020312032203320342035203620372038203920402041204220432044204520462047204820492050205120522053205420552056205720582059206020612062206320642065206620672068206920702071207220732074207520762077207820792080208120822083208420852086208720882089209020912092209320942095209620972098209921002101210221032104210521062107210821092110211121122113211421152116211721182119212021212122212321242125212621272128212921302131213221332134213521362137213821392140214121422143214421452146214721482149215021512152215321542155215621572158215921602161216221632164216521662167216821692170217121722173217421752176217721782179218021812182218321842185218621872188218921902191219221932194219521962197219821992200220122022203220422052206220722082209221022112212221322142215221622172218221922202221222222232224222522262227222822292230223122322233223422352236223722382239224022412242224322442245224622472248224922502251225222532254
  1. // Copyright (c) 2011-present, Facebook, Inc. All rights reserved.
  2. // This source code is licensed under both the GPLv2 (found in the
  3. // COPYING file in the root directory) and Apache 2.0 License
  4. // (found in the LICENSE.Apache file in the root directory).
  5. //
  6. // Copyright (c) 2011 The LevelDB Authors. All rights reserved.
  7. // Use of this source code is governed by a BSD-style license that can be
  8. // found in the LICENSE file. See the AUTHORS file for names of contributors.
  9. #include "db/db_test_util.h"
  10. #include "port/stack_trace.h"
  11. #if !defined(ROCKSDB_LITE)
  12. #include "rocksdb/utilities/table_properties_collectors.h"
  13. #include "test_util/sync_point.h"
  14. namespace ROCKSDB_NAMESPACE {
  15. static std::string CompressibleString(Random* rnd, int len) {
  16. std::string r;
  17. test::CompressibleString(rnd, 0.8, len, &r);
  18. return r;
  19. }
  20. class DBTestUniversalCompactionBase
  21. : public DBTestBase,
  22. public ::testing::WithParamInterface<std::tuple<int, bool>> {
  23. public:
  24. explicit DBTestUniversalCompactionBase(
  25. const std::string& path) : DBTestBase(path) {}
  26. void SetUp() override {
  27. num_levels_ = std::get<0>(GetParam());
  28. exclusive_manual_compaction_ = std::get<1>(GetParam());
  29. }
  30. int num_levels_;
  31. bool exclusive_manual_compaction_;
  32. };
  33. class DBTestUniversalCompaction : public DBTestUniversalCompactionBase {
  34. public:
  35. DBTestUniversalCompaction() :
  36. DBTestUniversalCompactionBase("/db_universal_compaction_test") {}
  37. };
  38. class DBTestUniversalCompaction2 : public DBTestBase {
  39. public:
  40. DBTestUniversalCompaction2() : DBTestBase("/db_universal_compaction_test2") {}
  41. };
  42. namespace {
  43. void VerifyCompactionResult(
  44. const ColumnFamilyMetaData& cf_meta,
  45. const std::set<std::string>& overlapping_file_numbers) {
  46. #ifndef NDEBUG
  47. for (auto& level : cf_meta.levels) {
  48. for (auto& file : level.files) {
  49. assert(overlapping_file_numbers.find(file.name) ==
  50. overlapping_file_numbers.end());
  51. }
  52. }
  53. #endif
  54. }
  55. class KeepFilter : public CompactionFilter {
  56. public:
  57. bool Filter(int /*level*/, const Slice& /*key*/, const Slice& /*value*/,
  58. std::string* /*new_value*/,
  59. bool* /*value_changed*/) const override {
  60. return false;
  61. }
  62. const char* Name() const override { return "KeepFilter"; }
  63. };
  64. class KeepFilterFactory : public CompactionFilterFactory {
  65. public:
  66. explicit KeepFilterFactory(bool check_context = false)
  67. : check_context_(check_context) {}
  68. std::unique_ptr<CompactionFilter> CreateCompactionFilter(
  69. const CompactionFilter::Context& context) override {
  70. if (check_context_) {
  71. EXPECT_EQ(expect_full_compaction_.load(), context.is_full_compaction);
  72. EXPECT_EQ(expect_manual_compaction_.load(), context.is_manual_compaction);
  73. }
  74. return std::unique_ptr<CompactionFilter>(new KeepFilter());
  75. }
  76. const char* Name() const override { return "KeepFilterFactory"; }
  77. bool check_context_;
  78. std::atomic_bool expect_full_compaction_;
  79. std::atomic_bool expect_manual_compaction_;
  80. };
  81. class DelayFilter : public CompactionFilter {
  82. public:
  83. explicit DelayFilter(DBTestBase* d) : db_test(d) {}
  84. bool Filter(int /*level*/, const Slice& /*key*/, const Slice& /*value*/,
  85. std::string* /*new_value*/,
  86. bool* /*value_changed*/) const override {
  87. db_test->env_->addon_time_.fetch_add(1000);
  88. return true;
  89. }
  90. const char* Name() const override { return "DelayFilter"; }
  91. private:
  92. DBTestBase* db_test;
  93. };
  94. class DelayFilterFactory : public CompactionFilterFactory {
  95. public:
  96. explicit DelayFilterFactory(DBTestBase* d) : db_test(d) {}
  97. std::unique_ptr<CompactionFilter> CreateCompactionFilter(
  98. const CompactionFilter::Context& /*context*/) override {
  99. return std::unique_ptr<CompactionFilter>(new DelayFilter(db_test));
  100. }
  101. const char* Name() const override { return "DelayFilterFactory"; }
  102. private:
  103. DBTestBase* db_test;
  104. };
  105. } // namespace
  106. // Make sure we don't trigger a problem if the trigger condtion is given
  107. // to be 0, which is invalid.
  108. TEST_P(DBTestUniversalCompaction, UniversalCompactionSingleSortedRun) {
  109. Options options = CurrentOptions();
  110. options.compaction_style = kCompactionStyleUniversal;
  111. options.num_levels = num_levels_;
  112. // Config universal compaction to always compact to one single sorted run.
  113. options.level0_file_num_compaction_trigger = 0;
  114. options.compaction_options_universal.size_ratio = 10;
  115. options.compaction_options_universal.min_merge_width = 2;
  116. options.compaction_options_universal.max_size_amplification_percent = 0;
  117. options.write_buffer_size = 105 << 10; // 105KB
  118. options.arena_block_size = 4 << 10;
  119. options.target_file_size_base = 32 << 10; // 32KB
  120. // trigger compaction if there are >= 4 files
  121. KeepFilterFactory* filter = new KeepFilterFactory(true);
  122. filter->expect_manual_compaction_.store(false);
  123. options.compaction_filter_factory.reset(filter);
  124. DestroyAndReopen(options);
  125. ASSERT_EQ(1, db_->GetOptions().level0_file_num_compaction_trigger);
  126. Random rnd(301);
  127. int key_idx = 0;
  128. filter->expect_full_compaction_.store(true);
  129. for (int num = 0; num < 16; num++) {
  130. // Write 100KB file. And immediately it should be compacted to one file.
  131. GenerateNewFile(&rnd, &key_idx);
  132. dbfull()->TEST_WaitForCompact();
  133. ASSERT_EQ(NumSortedRuns(0), 1);
  134. }
  135. ASSERT_OK(Put(Key(key_idx), ""));
  136. dbfull()->TEST_WaitForCompact();
  137. ASSERT_EQ(NumSortedRuns(0), 1);
  138. }
  139. TEST_P(DBTestUniversalCompaction, OptimizeFiltersForHits) {
  140. Options options = CurrentOptions();
  141. options.compaction_style = kCompactionStyleUniversal;
  142. options.compaction_options_universal.size_ratio = 5;
  143. options.num_levels = num_levels_;
  144. options.write_buffer_size = 105 << 10; // 105KB
  145. options.arena_block_size = 4 << 10;
  146. options.target_file_size_base = 32 << 10; // 32KB
  147. // trigger compaction if there are >= 4 files
  148. options.level0_file_num_compaction_trigger = 4;
  149. BlockBasedTableOptions bbto;
  150. bbto.cache_index_and_filter_blocks = true;
  151. bbto.filter_policy.reset(NewBloomFilterPolicy(10, false));
  152. bbto.whole_key_filtering = true;
  153. options.table_factory.reset(NewBlockBasedTableFactory(bbto));
  154. options.optimize_filters_for_hits = true;
  155. options.statistics = ROCKSDB_NAMESPACE::CreateDBStatistics();
  156. options.memtable_factory.reset(new SpecialSkipListFactory(3));
  157. DestroyAndReopen(options);
  158. // block compaction from happening
  159. env_->SetBackgroundThreads(1, Env::LOW);
  160. test::SleepingBackgroundTask sleeping_task_low;
  161. env_->Schedule(&test::SleepingBackgroundTask::DoSleepTask, &sleeping_task_low,
  162. Env::Priority::LOW);
  163. for (int num = 0; num < options.level0_file_num_compaction_trigger; num++) {
  164. Put(Key(num * 10), "val");
  165. if (num) {
  166. dbfull()->TEST_WaitForFlushMemTable();
  167. }
  168. Put(Key(30 + num * 10), "val");
  169. Put(Key(60 + num * 10), "val");
  170. }
  171. Put("", "");
  172. dbfull()->TEST_WaitForFlushMemTable();
  173. // Query set of non existing keys
  174. for (int i = 5; i < 90; i += 10) {
  175. ASSERT_EQ(Get(Key(i)), "NOT_FOUND");
  176. }
  177. // Make sure bloom filter is used at least once.
  178. ASSERT_GT(TestGetTickerCount(options, BLOOM_FILTER_USEFUL), 0);
  179. auto prev_counter = TestGetTickerCount(options, BLOOM_FILTER_USEFUL);
  180. // Make sure bloom filter is used for all but the last L0 file when looking
  181. // up a non-existent key that's in the range of all L0 files.
  182. ASSERT_EQ(Get(Key(35)), "NOT_FOUND");
  183. ASSERT_EQ(prev_counter + NumTableFilesAtLevel(0) - 1,
  184. TestGetTickerCount(options, BLOOM_FILTER_USEFUL));
  185. prev_counter = TestGetTickerCount(options, BLOOM_FILTER_USEFUL);
  186. // Unblock compaction and wait it for happening.
  187. sleeping_task_low.WakeUp();
  188. dbfull()->TEST_WaitForCompact();
  189. // The same queries will not trigger bloom filter
  190. for (int i = 5; i < 90; i += 10) {
  191. ASSERT_EQ(Get(Key(i)), "NOT_FOUND");
  192. }
  193. ASSERT_EQ(prev_counter, TestGetTickerCount(options, BLOOM_FILTER_USEFUL));
  194. }
  195. // TODO(kailiu) The tests on UniversalCompaction has some issues:
  196. // 1. A lot of magic numbers ("11" or "12").
  197. // 2. Made assumption on the memtable flush conditions, which may change from
  198. // time to time.
  199. TEST_P(DBTestUniversalCompaction, UniversalCompactionTrigger) {
  200. Options options;
  201. options.compaction_style = kCompactionStyleUniversal;
  202. options.compaction_options_universal.size_ratio = 5;
  203. options.num_levels = num_levels_;
  204. options.write_buffer_size = 105 << 10; // 105KB
  205. options.arena_block_size = 4 << 10;
  206. options.target_file_size_base = 32 << 10; // 32KB
  207. // trigger compaction if there are >= 4 files
  208. options.level0_file_num_compaction_trigger = 4;
  209. KeepFilterFactory* filter = new KeepFilterFactory(true);
  210. filter->expect_manual_compaction_.store(false);
  211. options.compaction_filter_factory.reset(filter);
  212. options = CurrentOptions(options);
  213. DestroyAndReopen(options);
  214. CreateAndReopenWithCF({"pikachu"}, options);
  215. ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->SetCallBack(
  216. "DBTestWritableFile.GetPreallocationStatus", [&](void* arg) {
  217. ASSERT_TRUE(arg != nullptr);
  218. size_t preallocation_size = *(static_cast<size_t*>(arg));
  219. if (num_levels_ > 3) {
  220. ASSERT_LE(preallocation_size, options.target_file_size_base * 1.1);
  221. }
  222. });
  223. ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->EnableProcessing();
  224. Random rnd(301);
  225. int key_idx = 0;
  226. filter->expect_full_compaction_.store(true);
  227. // Stage 1:
  228. // Generate a set of files at level 0, but don't trigger level-0
  229. // compaction.
  230. for (int num = 0; num < options.level0_file_num_compaction_trigger - 1;
  231. num++) {
  232. // Write 100KB
  233. GenerateNewFile(1, &rnd, &key_idx);
  234. }
  235. // Generate one more file at level-0, which should trigger level-0
  236. // compaction.
  237. GenerateNewFile(1, &rnd, &key_idx);
  238. // Suppose each file flushed from mem table has size 1. Now we compact
  239. // (level0_file_num_compaction_trigger+1)=4 files and should have a big
  240. // file of size 4.
  241. ASSERT_EQ(NumSortedRuns(1), 1);
  242. // Stage 2:
  243. // Now we have one file at level 0, with size 4. We also have some data in
  244. // mem table. Let's continue generating new files at level 0, but don't
  245. // trigger level-0 compaction.
  246. // First, clean up memtable before inserting new data. This will generate
  247. // a level-0 file, with size around 0.4 (according to previously written
  248. // data amount).
  249. filter->expect_full_compaction_.store(false);
  250. ASSERT_OK(Flush(1));
  251. for (int num = 0; num < options.level0_file_num_compaction_trigger - 3;
  252. num++) {
  253. GenerateNewFile(1, &rnd, &key_idx);
  254. ASSERT_EQ(NumSortedRuns(1), num + 3);
  255. }
  256. // Generate one more file at level-0, which should trigger level-0
  257. // compaction.
  258. GenerateNewFile(1, &rnd, &key_idx);
  259. // Before compaction, we have 4 files at level 0, with size 4, 0.4, 1, 1.
  260. // After compaction, we should have 2 files, with size 4, 2.4.
  261. ASSERT_EQ(NumSortedRuns(1), 2);
  262. // Stage 3:
  263. // Now we have 2 files at level 0, with size 4 and 2.4. Continue
  264. // generating new files at level 0.
  265. for (int num = 0; num < options.level0_file_num_compaction_trigger - 3;
  266. num++) {
  267. GenerateNewFile(1, &rnd, &key_idx);
  268. ASSERT_EQ(NumSortedRuns(1), num + 3);
  269. }
  270. // Generate one more file at level-0, which should trigger level-0
  271. // compaction.
  272. GenerateNewFile(1, &rnd, &key_idx);
  273. // Before compaction, we have 4 files at level 0, with size 4, 2.4, 1, 1.
  274. // After compaction, we should have 3 files, with size 4, 2.4, 2.
  275. ASSERT_EQ(NumSortedRuns(1), 3);
  276. // Stage 4:
  277. // Now we have 3 files at level 0, with size 4, 2.4, 2. Let's generate a
  278. // new file of size 1.
  279. GenerateNewFile(1, &rnd, &key_idx);
  280. dbfull()->TEST_WaitForCompact();
  281. // Level-0 compaction is triggered, but no file will be picked up.
  282. ASSERT_EQ(NumSortedRuns(1), 4);
  283. // Stage 5:
  284. // Now we have 4 files at level 0, with size 4, 2.4, 2, 1. Let's generate
  285. // a new file of size 1.
  286. filter->expect_full_compaction_.store(true);
  287. GenerateNewFile(1, &rnd, &key_idx);
  288. dbfull()->TEST_WaitForCompact();
  289. // All files at level 0 will be compacted into a single one.
  290. ASSERT_EQ(NumSortedRuns(1), 1);
  291. ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->DisableProcessing();
  292. }
  293. TEST_P(DBTestUniversalCompaction, UniversalCompactionSizeAmplification) {
  294. Options options = CurrentOptions();
  295. options.compaction_style = kCompactionStyleUniversal;
  296. options.num_levels = num_levels_;
  297. options.write_buffer_size = 100 << 10; // 100KB
  298. options.target_file_size_base = 32 << 10; // 32KB
  299. options.level0_file_num_compaction_trigger = 3;
  300. DestroyAndReopen(options);
  301. CreateAndReopenWithCF({"pikachu"}, options);
  302. // Trigger compaction if size amplification exceeds 110%
  303. options.compaction_options_universal.max_size_amplification_percent = 110;
  304. options = CurrentOptions(options);
  305. ReopenWithColumnFamilies({"default", "pikachu"}, options);
  306. Random rnd(301);
  307. int key_idx = 0;
  308. // Generate two files in Level 0. Both files are approx the same size.
  309. for (int num = 0; num < options.level0_file_num_compaction_trigger - 1;
  310. num++) {
  311. // Write 110KB (11 values, each 10K)
  312. for (int i = 0; i < 11; i++) {
  313. ASSERT_OK(Put(1, Key(key_idx), RandomString(&rnd, 10000)));
  314. key_idx++;
  315. }
  316. dbfull()->TEST_WaitForFlushMemTable(handles_[1]);
  317. ASSERT_EQ(NumSortedRuns(1), num + 1);
  318. }
  319. ASSERT_EQ(NumSortedRuns(1), 2);
  320. // Flush whatever is remaining in memtable. This is typically
  321. // small, which should not trigger size ratio based compaction
  322. // but will instead trigger size amplification.
  323. ASSERT_OK(Flush(1));
  324. dbfull()->TEST_WaitForCompact();
  325. // Verify that size amplification did occur
  326. ASSERT_EQ(NumSortedRuns(1), 1);
  327. }
  328. TEST_P(DBTestUniversalCompaction, DynamicUniversalCompactionSizeAmplification) {
  329. Options options = CurrentOptions();
  330. options.compaction_style = kCompactionStyleUniversal;
  331. options.num_levels = 1;
  332. options.write_buffer_size = 100 << 10; // 100KB
  333. options.target_file_size_base = 32 << 10; // 32KB
  334. options.level0_file_num_compaction_trigger = 3;
  335. // Initial setup of compaction_options_universal will prevent universal
  336. // compaction from happening
  337. options.compaction_options_universal.size_ratio = 100;
  338. options.compaction_options_universal.min_merge_width = 100;
  339. DestroyAndReopen(options);
  340. int total_picked_compactions = 0;
  341. int total_size_amp_compactions = 0;
  342. ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->SetCallBack(
  343. "UniversalCompactionBuilder::PickCompaction:Return", [&](void* arg) {
  344. if (arg) {
  345. total_picked_compactions++;
  346. Compaction* c = static_cast<Compaction*>(arg);
  347. if (c->compaction_reason() ==
  348. CompactionReason::kUniversalSizeAmplification) {
  349. total_size_amp_compactions++;
  350. }
  351. }
  352. });
  353. ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->EnableProcessing();
  354. MutableCFOptions mutable_cf_options;
  355. CreateAndReopenWithCF({"pikachu"}, options);
  356. Random rnd(301);
  357. int key_idx = 0;
  358. // Generate two files in Level 0. Both files are approx the same size.
  359. for (int num = 0; num < options.level0_file_num_compaction_trigger - 1;
  360. num++) {
  361. // Write 110KB (11 values, each 10K)
  362. for (int i = 0; i < 11; i++) {
  363. ASSERT_OK(Put(1, Key(key_idx), RandomString(&rnd, 10000)));
  364. key_idx++;
  365. }
  366. dbfull()->TEST_WaitForFlushMemTable(handles_[1]);
  367. ASSERT_EQ(NumSortedRuns(1), num + 1);
  368. }
  369. ASSERT_EQ(NumSortedRuns(1), 2);
  370. // Flush whatever is remaining in memtable. This is typically
  371. // small, which should not trigger size ratio based compaction
  372. // but could instead trigger size amplification if it's set
  373. // to 110.
  374. ASSERT_OK(Flush(1));
  375. dbfull()->TEST_WaitForCompact();
  376. // Verify compaction did not happen
  377. ASSERT_EQ(NumSortedRuns(1), 3);
  378. // Trigger compaction if size amplification exceeds 110% without reopening DB
  379. ASSERT_EQ(dbfull()
  380. ->GetOptions(handles_[1])
  381. .compaction_options_universal.max_size_amplification_percent,
  382. 200U);
  383. ASSERT_OK(dbfull()->SetOptions(handles_[1],
  384. {{"compaction_options_universal",
  385. "{max_size_amplification_percent=110;}"}}));
  386. ASSERT_EQ(dbfull()
  387. ->GetOptions(handles_[1])
  388. .compaction_options_universal.max_size_amplification_percent,
  389. 110u);
  390. ASSERT_OK(dbfull()->TEST_GetLatestMutableCFOptions(handles_[1],
  391. &mutable_cf_options));
  392. ASSERT_EQ(110u, mutable_cf_options.compaction_options_universal
  393. .max_size_amplification_percent);
  394. dbfull()->TEST_WaitForCompact();
  395. // Verify that size amplification did happen
  396. ASSERT_EQ(NumSortedRuns(1), 1);
  397. ASSERT_EQ(total_picked_compactions, 1);
  398. ASSERT_EQ(total_size_amp_compactions, 1);
  399. }
  400. TEST_P(DBTestUniversalCompaction, DynamicUniversalCompactionReadAmplification) {
  401. Options options = CurrentOptions();
  402. options.compaction_style = kCompactionStyleUniversal;
  403. options.num_levels = 1;
  404. options.write_buffer_size = 100 << 10; // 100KB
  405. options.target_file_size_base = 32 << 10; // 32KB
  406. options.level0_file_num_compaction_trigger = 3;
  407. // Initial setup of compaction_options_universal will prevent universal
  408. // compaction from happening
  409. options.compaction_options_universal.max_size_amplification_percent = 2000;
  410. options.compaction_options_universal.size_ratio = 0;
  411. options.compaction_options_universal.min_merge_width = 100;
  412. DestroyAndReopen(options);
  413. int total_picked_compactions = 0;
  414. int total_size_ratio_compactions = 0;
  415. ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->SetCallBack(
  416. "UniversalCompactionBuilder::PickCompaction:Return", [&](void* arg) {
  417. if (arg) {
  418. total_picked_compactions++;
  419. Compaction* c = static_cast<Compaction*>(arg);
  420. if (c->compaction_reason() == CompactionReason::kUniversalSizeRatio) {
  421. total_size_ratio_compactions++;
  422. }
  423. }
  424. });
  425. ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->EnableProcessing();
  426. MutableCFOptions mutable_cf_options;
  427. CreateAndReopenWithCF({"pikachu"}, options);
  428. Random rnd(301);
  429. int key_idx = 0;
  430. // Generate three files in Level 0. All files are approx the same size.
  431. for (int num = 0; num < options.level0_file_num_compaction_trigger; num++) {
  432. // Write 110KB (11 values, each 10K)
  433. for (int i = 0; i < 11; i++) {
  434. ASSERT_OK(Put(1, Key(key_idx), RandomString(&rnd, 10000)));
  435. key_idx++;
  436. }
  437. dbfull()->TEST_WaitForFlushMemTable(handles_[1]);
  438. ASSERT_EQ(NumSortedRuns(1), num + 1);
  439. }
  440. ASSERT_EQ(NumSortedRuns(1), options.level0_file_num_compaction_trigger);
  441. // Flush whatever is remaining in memtable. This is typically small, about
  442. // 30KB.
  443. ASSERT_OK(Flush(1));
  444. dbfull()->TEST_WaitForCompact();
  445. // Verify compaction did not happen
  446. ASSERT_EQ(NumSortedRuns(1), options.level0_file_num_compaction_trigger + 1);
  447. ASSERT_EQ(total_picked_compactions, 0);
  448. ASSERT_OK(dbfull()->SetOptions(
  449. handles_[1],
  450. {{"compaction_options_universal",
  451. "{min_merge_width=2;max_merge_width=2;size_ratio=100;}"}}));
  452. ASSERT_EQ(dbfull()
  453. ->GetOptions(handles_[1])
  454. .compaction_options_universal.min_merge_width,
  455. 2u);
  456. ASSERT_EQ(dbfull()
  457. ->GetOptions(handles_[1])
  458. .compaction_options_universal.max_merge_width,
  459. 2u);
  460. ASSERT_EQ(
  461. dbfull()->GetOptions(handles_[1]).compaction_options_universal.size_ratio,
  462. 100u);
  463. ASSERT_OK(dbfull()->TEST_GetLatestMutableCFOptions(handles_[1],
  464. &mutable_cf_options));
  465. ASSERT_EQ(mutable_cf_options.compaction_options_universal.size_ratio, 100u);
  466. ASSERT_EQ(mutable_cf_options.compaction_options_universal.min_merge_width,
  467. 2u);
  468. ASSERT_EQ(mutable_cf_options.compaction_options_universal.max_merge_width,
  469. 2u);
  470. dbfull()->TEST_WaitForCompact();
  471. // Files in L0 are approx: 0.3 (30KB), 1, 1, 1.
  472. // On compaction: the files are below the size amp threshold, so we
  473. // fallthrough to checking read amp conditions. The configured size ratio is
  474. // not big enough to take 0.3 into consideration. So the next files 1 and 1
  475. // are compacted together first as they satisfy size ratio condition and
  476. // (min_merge_width, max_merge_width) condition, to give out a file size of 2.
  477. // Next, the newly generated 2 and the last file 1 are compacted together. So
  478. // at the end: #sortedRuns = 2, #picked_compactions = 2, and all the picked
  479. // ones are size ratio based compactions.
  480. ASSERT_EQ(NumSortedRuns(1), 2);
  481. // If max_merge_width had not been changed dynamically above, and if it
  482. // continued to be the default value of UINIT_MAX, total_picked_compactions
  483. // would have been 1.
  484. ASSERT_EQ(total_picked_compactions, 2);
  485. ASSERT_EQ(total_size_ratio_compactions, 2);
  486. }
  487. TEST_P(DBTestUniversalCompaction, CompactFilesOnUniversalCompaction) {
  488. const int kTestKeySize = 16;
  489. const int kTestValueSize = 984;
  490. const int kEntrySize = kTestKeySize + kTestValueSize;
  491. const int kEntriesPerBuffer = 10;
  492. ChangeCompactOptions();
  493. Options options;
  494. options.create_if_missing = true;
  495. options.compaction_style = kCompactionStyleLevel;
  496. options.num_levels = 1;
  497. options.target_file_size_base = options.write_buffer_size;
  498. options.compression = kNoCompression;
  499. options = CurrentOptions(options);
  500. options.write_buffer_size = kEntrySize * kEntriesPerBuffer;
  501. CreateAndReopenWithCF({"pikachu"}, options);
  502. ASSERT_EQ(options.compaction_style, kCompactionStyleUniversal);
  503. Random rnd(301);
  504. for (int key = 1024 * kEntriesPerBuffer; key >= 0; --key) {
  505. ASSERT_OK(Put(1, ToString(key), RandomString(&rnd, kTestValueSize)));
  506. }
  507. dbfull()->TEST_WaitForFlushMemTable(handles_[1]);
  508. dbfull()->TEST_WaitForCompact();
  509. ColumnFamilyMetaData cf_meta;
  510. dbfull()->GetColumnFamilyMetaData(handles_[1], &cf_meta);
  511. std::vector<std::string> compaction_input_file_names;
  512. for (auto file : cf_meta.levels[0].files) {
  513. if (rnd.OneIn(2)) {
  514. compaction_input_file_names.push_back(file.name);
  515. }
  516. }
  517. if (compaction_input_file_names.size() == 0) {
  518. compaction_input_file_names.push_back(
  519. cf_meta.levels[0].files[0].name);
  520. }
  521. // expect fail since universal compaction only allow L0 output
  522. ASSERT_FALSE(dbfull()
  523. ->CompactFiles(CompactionOptions(), handles_[1],
  524. compaction_input_file_names, 1)
  525. .ok());
  526. // expect ok and verify the compacted files no longer exist.
  527. ASSERT_OK(dbfull()->CompactFiles(
  528. CompactionOptions(), handles_[1],
  529. compaction_input_file_names, 0));
  530. dbfull()->GetColumnFamilyMetaData(handles_[1], &cf_meta);
  531. VerifyCompactionResult(
  532. cf_meta,
  533. std::set<std::string>(compaction_input_file_names.begin(),
  534. compaction_input_file_names.end()));
  535. compaction_input_file_names.clear();
  536. // Pick the first and the last file, expect everything is
  537. // compacted into one single file.
  538. compaction_input_file_names.push_back(
  539. cf_meta.levels[0].files[0].name);
  540. compaction_input_file_names.push_back(
  541. cf_meta.levels[0].files[
  542. cf_meta.levels[0].files.size() - 1].name);
  543. ASSERT_OK(dbfull()->CompactFiles(
  544. CompactionOptions(), handles_[1],
  545. compaction_input_file_names, 0));
  546. dbfull()->GetColumnFamilyMetaData(handles_[1], &cf_meta);
  547. ASSERT_EQ(cf_meta.levels[0].files.size(), 1U);
  548. }
  549. TEST_P(DBTestUniversalCompaction, UniversalCompactionTargetLevel) {
  550. Options options = CurrentOptions();
  551. options.compaction_style = kCompactionStyleUniversal;
  552. options.write_buffer_size = 100 << 10; // 100KB
  553. options.num_levels = 7;
  554. options.disable_auto_compactions = true;
  555. DestroyAndReopen(options);
  556. // Generate 3 overlapping files
  557. Random rnd(301);
  558. for (int i = 0; i < 210; i++) {
  559. ASSERT_OK(Put(Key(i), RandomString(&rnd, 100)));
  560. }
  561. ASSERT_OK(Flush());
  562. for (int i = 200; i < 300; i++) {
  563. ASSERT_OK(Put(Key(i), RandomString(&rnd, 100)));
  564. }
  565. ASSERT_OK(Flush());
  566. for (int i = 250; i < 260; i++) {
  567. ASSERT_OK(Put(Key(i), RandomString(&rnd, 100)));
  568. }
  569. ASSERT_OK(Flush());
  570. ASSERT_EQ("3", FilesPerLevel(0));
  571. // Compact all files into 1 file and put it in L4
  572. CompactRangeOptions compact_options;
  573. compact_options.change_level = true;
  574. compact_options.target_level = 4;
  575. compact_options.exclusive_manual_compaction = exclusive_manual_compaction_;
  576. db_->CompactRange(compact_options, nullptr, nullptr);
  577. ASSERT_EQ("0,0,0,0,1", FilesPerLevel(0));
  578. }
  579. #ifndef ROCKSDB_VALGRIND_RUN
  580. class DBTestUniversalCompactionMultiLevels
  581. : public DBTestUniversalCompactionBase {
  582. public:
  583. DBTestUniversalCompactionMultiLevels() :
  584. DBTestUniversalCompactionBase(
  585. "/db_universal_compaction_multi_levels_test") {}
  586. };
  587. TEST_P(DBTestUniversalCompactionMultiLevels, UniversalCompactionMultiLevels) {
  588. Options options = CurrentOptions();
  589. options.compaction_style = kCompactionStyleUniversal;
  590. options.num_levels = num_levels_;
  591. options.write_buffer_size = 100 << 10; // 100KB
  592. options.level0_file_num_compaction_trigger = 8;
  593. options.max_background_compactions = 3;
  594. options.target_file_size_base = 32 * 1024;
  595. CreateAndReopenWithCF({"pikachu"}, options);
  596. // Trigger compaction if size amplification exceeds 110%
  597. options.compaction_options_universal.max_size_amplification_percent = 110;
  598. options = CurrentOptions(options);
  599. ReopenWithColumnFamilies({"default", "pikachu"}, options);
  600. Random rnd(301);
  601. int num_keys = 100000;
  602. for (int i = 0; i < num_keys * 2; i++) {
  603. ASSERT_OK(Put(1, Key(i % num_keys), Key(i)));
  604. }
  605. dbfull()->TEST_WaitForCompact();
  606. for (int i = num_keys; i < num_keys * 2; i++) {
  607. ASSERT_EQ(Get(1, Key(i % num_keys)), Key(i));
  608. }
  609. }
  610. // Tests universal compaction with trivial move enabled
  611. TEST_P(DBTestUniversalCompactionMultiLevels, UniversalCompactionTrivialMove) {
  612. int32_t trivial_move = 0;
  613. int32_t non_trivial_move = 0;
  614. ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->SetCallBack(
  615. "DBImpl::BackgroundCompaction:TrivialMove",
  616. [&](void* /*arg*/) { trivial_move++; });
  617. ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->SetCallBack(
  618. "DBImpl::BackgroundCompaction:NonTrivial", [&](void* arg) {
  619. non_trivial_move++;
  620. ASSERT_TRUE(arg != nullptr);
  621. int output_level = *(static_cast<int*>(arg));
  622. ASSERT_EQ(output_level, 0);
  623. });
  624. ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->EnableProcessing();
  625. Options options = CurrentOptions();
  626. options.compaction_style = kCompactionStyleUniversal;
  627. options.compaction_options_universal.allow_trivial_move = true;
  628. options.num_levels = 3;
  629. options.write_buffer_size = 100 << 10; // 100KB
  630. options.level0_file_num_compaction_trigger = 3;
  631. options.max_background_compactions = 2;
  632. options.target_file_size_base = 32 * 1024;
  633. DestroyAndReopen(options);
  634. CreateAndReopenWithCF({"pikachu"}, options);
  635. // Trigger compaction if size amplification exceeds 110%
  636. options.compaction_options_universal.max_size_amplification_percent = 110;
  637. options = CurrentOptions(options);
  638. ReopenWithColumnFamilies({"default", "pikachu"}, options);
  639. Random rnd(301);
  640. int num_keys = 150000;
  641. for (int i = 0; i < num_keys; i++) {
  642. ASSERT_OK(Put(1, Key(i), Key(i)));
  643. }
  644. std::vector<std::string> values;
  645. ASSERT_OK(Flush(1));
  646. dbfull()->TEST_WaitForCompact();
  647. ASSERT_GT(trivial_move, 0);
  648. ASSERT_GT(non_trivial_move, 0);
  649. ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->DisableProcessing();
  650. }
  651. INSTANTIATE_TEST_CASE_P(MultiLevels, DBTestUniversalCompactionMultiLevels,
  652. ::testing::Combine(::testing::Values(3, 20),
  653. ::testing::Bool()));
  654. class DBTestUniversalCompactionParallel :
  655. public DBTestUniversalCompactionBase {
  656. public:
  657. DBTestUniversalCompactionParallel() :
  658. DBTestUniversalCompactionBase(
  659. "/db_universal_compaction_prallel_test") {}
  660. };
  661. TEST_P(DBTestUniversalCompactionParallel, UniversalCompactionParallel) {
  662. Options options = CurrentOptions();
  663. options.compaction_style = kCompactionStyleUniversal;
  664. options.num_levels = num_levels_;
  665. options.write_buffer_size = 1 << 10; // 1KB
  666. options.level0_file_num_compaction_trigger = 3;
  667. options.max_background_compactions = 3;
  668. options.max_background_flushes = 3;
  669. options.target_file_size_base = 1 * 1024;
  670. options.compaction_options_universal.max_size_amplification_percent = 110;
  671. DestroyAndReopen(options);
  672. CreateAndReopenWithCF({"pikachu"}, options);
  673. // Delay every compaction so multiple compactions will happen.
  674. std::atomic<int> num_compactions_running(0);
  675. std::atomic<bool> has_parallel(false);
  676. ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->SetCallBack(
  677. "CompactionJob::Run():Start", [&](void* /*arg*/) {
  678. if (num_compactions_running.fetch_add(1) > 0) {
  679. has_parallel.store(true);
  680. return;
  681. }
  682. for (int nwait = 0; nwait < 20000; nwait++) {
  683. if (has_parallel.load() || num_compactions_running.load() > 1) {
  684. has_parallel.store(true);
  685. break;
  686. }
  687. env_->SleepForMicroseconds(1000);
  688. }
  689. });
  690. ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->SetCallBack(
  691. "CompactionJob::Run():End",
  692. [&](void* /*arg*/) { num_compactions_running.fetch_add(-1); });
  693. ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->EnableProcessing();
  694. options = CurrentOptions(options);
  695. ReopenWithColumnFamilies({"default", "pikachu"}, options);
  696. Random rnd(301);
  697. int num_keys = 30000;
  698. for (int i = 0; i < num_keys * 2; i++) {
  699. ASSERT_OK(Put(1, Key(i % num_keys), Key(i)));
  700. }
  701. dbfull()->TEST_WaitForCompact();
  702. ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->DisableProcessing();
  703. ASSERT_EQ(num_compactions_running.load(), 0);
  704. ASSERT_TRUE(has_parallel.load());
  705. for (int i = num_keys; i < num_keys * 2; i++) {
  706. ASSERT_EQ(Get(1, Key(i % num_keys)), Key(i));
  707. }
  708. // Reopen and check.
  709. ReopenWithColumnFamilies({"default", "pikachu"}, options);
  710. for (int i = num_keys; i < num_keys * 2; i++) {
  711. ASSERT_EQ(Get(1, Key(i % num_keys)), Key(i));
  712. }
  713. }
  714. TEST_P(DBTestUniversalCompactionParallel, PickByFileNumberBug) {
  715. Options options = CurrentOptions();
  716. options.compaction_style = kCompactionStyleUniversal;
  717. options.num_levels = num_levels_;
  718. options.write_buffer_size = 1 * 1024; // 1KB
  719. options.level0_file_num_compaction_trigger = 7;
  720. options.max_background_compactions = 2;
  721. options.target_file_size_base = 1024 * 1024; // 1MB
  722. // Disable size amplifiction compaction
  723. options.compaction_options_universal.max_size_amplification_percent =
  724. UINT_MAX;
  725. DestroyAndReopen(options);
  726. ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->LoadDependency(
  727. {{"DBTestUniversalCompactionParallel::PickByFileNumberBug:0",
  728. "BackgroundCallCompaction:0"},
  729. {"UniversalCompactionBuilder::PickCompaction:Return",
  730. "DBTestUniversalCompactionParallel::PickByFileNumberBug:1"},
  731. {"DBTestUniversalCompactionParallel::PickByFileNumberBug:2",
  732. "CompactionJob::Run():Start"}});
  733. int total_picked_compactions = 0;
  734. ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->SetCallBack(
  735. "UniversalCompactionBuilder::PickCompaction:Return", [&](void* arg) {
  736. if (arg) {
  737. total_picked_compactions++;
  738. }
  739. });
  740. ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->EnableProcessing();
  741. // Write 7 files to trigger compaction
  742. int key_idx = 1;
  743. for (int i = 1; i <= 70; i++) {
  744. std::string k = Key(key_idx++);
  745. ASSERT_OK(Put(k, k));
  746. if (i % 10 == 0) {
  747. ASSERT_OK(Flush());
  748. }
  749. }
  750. // Wait for the 1st background compaction process to start
  751. TEST_SYNC_POINT("DBTestUniversalCompactionParallel::PickByFileNumberBug:0");
  752. TEST_SYNC_POINT("DBTestUniversalCompactionParallel::PickByFileNumberBug:1");
  753. ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->ClearTrace();
  754. // Write 3 files while 1st compaction is held
  755. // These 3 files have different sizes to avoid compacting based on size_ratio
  756. int num_keys = 1000;
  757. for (int i = 0; i < 3; i++) {
  758. for (int j = 1; j <= num_keys; j++) {
  759. std::string k = Key(key_idx++);
  760. ASSERT_OK(Put(k, k));
  761. }
  762. ASSERT_OK(Flush());
  763. num_keys -= 100;
  764. }
  765. // Hold the 1st compaction from finishing
  766. TEST_SYNC_POINT("DBTestUniversalCompactionParallel::PickByFileNumberBug:2");
  767. dbfull()->TEST_WaitForCompact();
  768. // There should only be one picked compaction as the score drops below one
  769. // after the first one is picked.
  770. EXPECT_EQ(total_picked_compactions, 1);
  771. EXPECT_EQ(TotalTableFiles(), 4);
  772. // Stop SyncPoint and destroy the DB and reopen it again
  773. ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->ClearTrace();
  774. ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->DisableProcessing();
  775. key_idx = 1;
  776. total_picked_compactions = 0;
  777. DestroyAndReopen(options);
  778. ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->EnableProcessing();
  779. // Write 7 files to trigger compaction
  780. for (int i = 1; i <= 70; i++) {
  781. std::string k = Key(key_idx++);
  782. ASSERT_OK(Put(k, k));
  783. if (i % 10 == 0) {
  784. ASSERT_OK(Flush());
  785. }
  786. }
  787. // Wait for the 1st background compaction process to start
  788. TEST_SYNC_POINT("DBTestUniversalCompactionParallel::PickByFileNumberBug:0");
  789. TEST_SYNC_POINT("DBTestUniversalCompactionParallel::PickByFileNumberBug:1");
  790. ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->ClearTrace();
  791. // Write 8 files while 1st compaction is held
  792. // These 8 files have different sizes to avoid compacting based on size_ratio
  793. num_keys = 1000;
  794. for (int i = 0; i < 8; i++) {
  795. for (int j = 1; j <= num_keys; j++) {
  796. std::string k = Key(key_idx++);
  797. ASSERT_OK(Put(k, k));
  798. }
  799. ASSERT_OK(Flush());
  800. num_keys -= 100;
  801. }
  802. // Wait for the 2nd background compaction process to start
  803. TEST_SYNC_POINT("DBTestUniversalCompactionParallel::PickByFileNumberBug:0");
  804. TEST_SYNC_POINT("DBTestUniversalCompactionParallel::PickByFileNumberBug:1");
  805. // Hold the 1st and 2nd compaction from finishing
  806. TEST_SYNC_POINT("DBTestUniversalCompactionParallel::PickByFileNumberBug:2");
  807. dbfull()->TEST_WaitForCompact();
  808. // This time we will trigger a compaction because of size ratio and
  809. // another compaction because of number of files that are not compacted
  810. // greater than 7
  811. EXPECT_GE(total_picked_compactions, 2);
  812. }
  813. INSTANTIATE_TEST_CASE_P(Parallel, DBTestUniversalCompactionParallel,
  814. ::testing::Combine(::testing::Values(1, 10),
  815. ::testing::Values(false)));
  816. #endif // ROCKSDB_VALGRIND_RUN
  817. TEST_P(DBTestUniversalCompaction, UniversalCompactionOptions) {
  818. Options options = CurrentOptions();
  819. options.compaction_style = kCompactionStyleUniversal;
  820. options.write_buffer_size = 105 << 10; // 105KB
  821. options.arena_block_size = 4 << 10; // 4KB
  822. options.target_file_size_base = 32 << 10; // 32KB
  823. options.level0_file_num_compaction_trigger = 4;
  824. options.num_levels = num_levels_;
  825. options.compaction_options_universal.compression_size_percent = -1;
  826. DestroyAndReopen(options);
  827. CreateAndReopenWithCF({"pikachu"}, options);
  828. Random rnd(301);
  829. int key_idx = 0;
  830. for (int num = 0; num < options.level0_file_num_compaction_trigger; num++) {
  831. // Write 100KB (100 values, each 1K)
  832. for (int i = 0; i < 100; i++) {
  833. ASSERT_OK(Put(1, Key(key_idx), RandomString(&rnd, 990)));
  834. key_idx++;
  835. }
  836. dbfull()->TEST_WaitForFlushMemTable(handles_[1]);
  837. if (num < options.level0_file_num_compaction_trigger - 1) {
  838. ASSERT_EQ(NumSortedRuns(1), num + 1);
  839. }
  840. }
  841. dbfull()->TEST_WaitForCompact();
  842. ASSERT_EQ(NumSortedRuns(1), 1);
  843. }
  844. TEST_P(DBTestUniversalCompaction, UniversalCompactionStopStyleSimilarSize) {
  845. Options options = CurrentOptions();
  846. options.compaction_style = kCompactionStyleUniversal;
  847. options.write_buffer_size = 105 << 10; // 105KB
  848. options.arena_block_size = 4 << 10; // 4KB
  849. options.target_file_size_base = 32 << 10; // 32KB
  850. // trigger compaction if there are >= 4 files
  851. options.level0_file_num_compaction_trigger = 4;
  852. options.compaction_options_universal.size_ratio = 10;
  853. options.compaction_options_universal.stop_style =
  854. kCompactionStopStyleSimilarSize;
  855. options.num_levels = num_levels_;
  856. DestroyAndReopen(options);
  857. Random rnd(301);
  858. int key_idx = 0;
  859. // Stage 1:
  860. // Generate a set of files at level 0, but don't trigger level-0
  861. // compaction.
  862. for (int num = 0; num < options.level0_file_num_compaction_trigger - 1;
  863. num++) {
  864. // Write 100KB (100 values, each 1K)
  865. for (int i = 0; i < 100; i++) {
  866. ASSERT_OK(Put(Key(key_idx), RandomString(&rnd, 990)));
  867. key_idx++;
  868. }
  869. dbfull()->TEST_WaitForFlushMemTable();
  870. ASSERT_EQ(NumSortedRuns(), num + 1);
  871. }
  872. // Generate one more file at level-0, which should trigger level-0
  873. // compaction.
  874. for (int i = 0; i < 100; i++) {
  875. ASSERT_OK(Put(Key(key_idx), RandomString(&rnd, 990)));
  876. key_idx++;
  877. }
  878. dbfull()->TEST_WaitForCompact();
  879. // Suppose each file flushed from mem table has size 1. Now we compact
  880. // (level0_file_num_compaction_trigger+1)=4 files and should have a big
  881. // file of size 4.
  882. ASSERT_EQ(NumSortedRuns(), 1);
  883. // Stage 2:
  884. // Now we have one file at level 0, with size 4. We also have some data in
  885. // mem table. Let's continue generating new files at level 0, but don't
  886. // trigger level-0 compaction.
  887. // First, clean up memtable before inserting new data. This will generate
  888. // a level-0 file, with size around 0.4 (according to previously written
  889. // data amount).
  890. dbfull()->Flush(FlushOptions());
  891. for (int num = 0; num < options.level0_file_num_compaction_trigger - 3;
  892. num++) {
  893. // Write 110KB (11 values, each 10K)
  894. for (int i = 0; i < 100; i++) {
  895. ASSERT_OK(Put(Key(key_idx), RandomString(&rnd, 990)));
  896. key_idx++;
  897. }
  898. dbfull()->TEST_WaitForFlushMemTable();
  899. ASSERT_EQ(NumSortedRuns(), num + 3);
  900. }
  901. // Generate one more file at level-0, which should trigger level-0
  902. // compaction.
  903. for (int i = 0; i < 100; i++) {
  904. ASSERT_OK(Put(Key(key_idx), RandomString(&rnd, 990)));
  905. key_idx++;
  906. }
  907. dbfull()->TEST_WaitForCompact();
  908. // Before compaction, we have 4 files at level 0, with size 4, 0.4, 1, 1.
  909. // After compaction, we should have 3 files, with size 4, 0.4, 2.
  910. ASSERT_EQ(NumSortedRuns(), 3);
  911. // Stage 3:
  912. // Now we have 3 files at level 0, with size 4, 0.4, 2. Generate one
  913. // more file at level-0, which should trigger level-0 compaction.
  914. for (int i = 0; i < 100; i++) {
  915. ASSERT_OK(Put(Key(key_idx), RandomString(&rnd, 990)));
  916. key_idx++;
  917. }
  918. dbfull()->TEST_WaitForCompact();
  919. // Level-0 compaction is triggered, but no file will be picked up.
  920. ASSERT_EQ(NumSortedRuns(), 4);
  921. }
  922. TEST_P(DBTestUniversalCompaction, UniversalCompactionCompressRatio1) {
  923. if (!Snappy_Supported()) {
  924. return;
  925. }
  926. Options options = CurrentOptions();
  927. options.compaction_style = kCompactionStyleUniversal;
  928. options.write_buffer_size = 100 << 10; // 100KB
  929. options.target_file_size_base = 32 << 10; // 32KB
  930. options.level0_file_num_compaction_trigger = 2;
  931. options.num_levels = num_levels_;
  932. options.compaction_options_universal.compression_size_percent = 70;
  933. DestroyAndReopen(options);
  934. Random rnd(301);
  935. int key_idx = 0;
  936. // The first compaction (2) is compressed.
  937. for (int num = 0; num < 2; num++) {
  938. // Write 110KB (11 values, each 10K)
  939. for (int i = 0; i < 11; i++) {
  940. ASSERT_OK(Put(Key(key_idx), CompressibleString(&rnd, 10000)));
  941. key_idx++;
  942. }
  943. dbfull()->TEST_WaitForFlushMemTable();
  944. dbfull()->TEST_WaitForCompact();
  945. }
  946. ASSERT_LT(TotalSize(), 110000U * 2 * 0.9);
  947. // The second compaction (4) is compressed
  948. for (int num = 0; num < 2; num++) {
  949. // Write 110KB (11 values, each 10K)
  950. for (int i = 0; i < 11; i++) {
  951. ASSERT_OK(Put(Key(key_idx), CompressibleString(&rnd, 10000)));
  952. key_idx++;
  953. }
  954. dbfull()->TEST_WaitForFlushMemTable();
  955. dbfull()->TEST_WaitForCompact();
  956. }
  957. ASSERT_LT(TotalSize(), 110000 * 4 * 0.9);
  958. // The third compaction (2 4) is compressed since this time it is
  959. // (1 1 3.2) and 3.2/5.2 doesn't reach ratio.
  960. for (int num = 0; num < 2; num++) {
  961. // Write 110KB (11 values, each 10K)
  962. for (int i = 0; i < 11; i++) {
  963. ASSERT_OK(Put(Key(key_idx), CompressibleString(&rnd, 10000)));
  964. key_idx++;
  965. }
  966. dbfull()->TEST_WaitForFlushMemTable();
  967. dbfull()->TEST_WaitForCompact();
  968. }
  969. ASSERT_LT(TotalSize(), 110000 * 6 * 0.9);
  970. // When we start for the compaction up to (2 4 8), the latest
  971. // compressed is not compressed.
  972. for (int num = 0; num < 8; num++) {
  973. // Write 110KB (11 values, each 10K)
  974. for (int i = 0; i < 11; i++) {
  975. ASSERT_OK(Put(Key(key_idx), CompressibleString(&rnd, 10000)));
  976. key_idx++;
  977. }
  978. dbfull()->TEST_WaitForFlushMemTable();
  979. dbfull()->TEST_WaitForCompact();
  980. }
  981. ASSERT_GT(TotalSize(), 110000 * 11 * 0.8 + 110000 * 2);
  982. }
  983. TEST_P(DBTestUniversalCompaction, UniversalCompactionCompressRatio2) {
  984. if (!Snappy_Supported()) {
  985. return;
  986. }
  987. Options options = CurrentOptions();
  988. options.compaction_style = kCompactionStyleUniversal;
  989. options.write_buffer_size = 100 << 10; // 100KB
  990. options.target_file_size_base = 32 << 10; // 32KB
  991. options.level0_file_num_compaction_trigger = 2;
  992. options.num_levels = num_levels_;
  993. options.compaction_options_universal.compression_size_percent = 95;
  994. DestroyAndReopen(options);
  995. Random rnd(301);
  996. int key_idx = 0;
  997. // When we start for the compaction up to (2 4 8), the latest
  998. // compressed is compressed given the size ratio to compress.
  999. for (int num = 0; num < 14; num++) {
  1000. // Write 120KB (12 values, each 10K)
  1001. for (int i = 0; i < 12; i++) {
  1002. ASSERT_OK(Put(Key(key_idx), CompressibleString(&rnd, 10000)));
  1003. key_idx++;
  1004. }
  1005. dbfull()->TEST_WaitForFlushMemTable();
  1006. dbfull()->TEST_WaitForCompact();
  1007. }
  1008. ASSERT_LT(TotalSize(), 120000U * 12 * 0.82 + 120000 * 2);
  1009. }
  1010. #ifndef ROCKSDB_VALGRIND_RUN
  1011. // Test that checks trivial move in universal compaction
  1012. TEST_P(DBTestUniversalCompaction, UniversalCompactionTrivialMoveTest1) {
  1013. int32_t trivial_move = 0;
  1014. int32_t non_trivial_move = 0;
  1015. ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->SetCallBack(
  1016. "DBImpl::BackgroundCompaction:TrivialMove",
  1017. [&](void* /*arg*/) { trivial_move++; });
  1018. ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->SetCallBack(
  1019. "DBImpl::BackgroundCompaction:NonTrivial", [&](void* arg) {
  1020. non_trivial_move++;
  1021. ASSERT_TRUE(arg != nullptr);
  1022. int output_level = *(static_cast<int*>(arg));
  1023. ASSERT_EQ(output_level, 0);
  1024. });
  1025. ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->EnableProcessing();
  1026. Options options = CurrentOptions();
  1027. options.compaction_style = kCompactionStyleUniversal;
  1028. options.compaction_options_universal.allow_trivial_move = true;
  1029. options.num_levels = 2;
  1030. options.write_buffer_size = 100 << 10; // 100KB
  1031. options.level0_file_num_compaction_trigger = 3;
  1032. options.max_background_compactions = 1;
  1033. options.target_file_size_base = 32 * 1024;
  1034. DestroyAndReopen(options);
  1035. CreateAndReopenWithCF({"pikachu"}, options);
  1036. // Trigger compaction if size amplification exceeds 110%
  1037. options.compaction_options_universal.max_size_amplification_percent = 110;
  1038. options = CurrentOptions(options);
  1039. ReopenWithColumnFamilies({"default", "pikachu"}, options);
  1040. Random rnd(301);
  1041. int num_keys = 250000;
  1042. for (int i = 0; i < num_keys; i++) {
  1043. ASSERT_OK(Put(1, Key(i), Key(i)));
  1044. }
  1045. std::vector<std::string> values;
  1046. ASSERT_OK(Flush(1));
  1047. dbfull()->TEST_WaitForCompact();
  1048. ASSERT_GT(trivial_move, 0);
  1049. ASSERT_GT(non_trivial_move, 0);
  1050. ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->DisableProcessing();
  1051. }
  1052. // Test that checks trivial move in universal compaction
  1053. TEST_P(DBTestUniversalCompaction, UniversalCompactionTrivialMoveTest2) {
  1054. int32_t trivial_move = 0;
  1055. ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->SetCallBack(
  1056. "DBImpl::BackgroundCompaction:TrivialMove",
  1057. [&](void* /*arg*/) { trivial_move++; });
  1058. ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->SetCallBack(
  1059. "DBImpl::BackgroundCompaction:NonTrivial", [&](void* arg) {
  1060. ASSERT_TRUE(arg != nullptr);
  1061. int output_level = *(static_cast<int*>(arg));
  1062. ASSERT_EQ(output_level, 0);
  1063. });
  1064. ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->EnableProcessing();
  1065. Options options = CurrentOptions();
  1066. options.compaction_style = kCompactionStyleUniversal;
  1067. options.compaction_options_universal.allow_trivial_move = true;
  1068. options.num_levels = 15;
  1069. options.write_buffer_size = 100 << 10; // 100KB
  1070. options.level0_file_num_compaction_trigger = 8;
  1071. options.max_background_compactions = 2;
  1072. options.target_file_size_base = 64 * 1024;
  1073. DestroyAndReopen(options);
  1074. CreateAndReopenWithCF({"pikachu"}, options);
  1075. // Trigger compaction if size amplification exceeds 110%
  1076. options.compaction_options_universal.max_size_amplification_percent = 110;
  1077. options = CurrentOptions(options);
  1078. ReopenWithColumnFamilies({"default", "pikachu"}, options);
  1079. Random rnd(301);
  1080. int num_keys = 500000;
  1081. for (int i = 0; i < num_keys; i++) {
  1082. ASSERT_OK(Put(1, Key(i), Key(i)));
  1083. }
  1084. std::vector<std::string> values;
  1085. ASSERT_OK(Flush(1));
  1086. dbfull()->TEST_WaitForCompact();
  1087. ASSERT_GT(trivial_move, 0);
  1088. ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->DisableProcessing();
  1089. }
  1090. #endif // ROCKSDB_VALGRIND_RUN
  1091. TEST_P(DBTestUniversalCompaction, UniversalCompactionFourPaths) {
  1092. Options options = CurrentOptions();
  1093. options.db_paths.emplace_back(dbname_, 300 * 1024);
  1094. options.db_paths.emplace_back(dbname_ + "_2", 300 * 1024);
  1095. options.db_paths.emplace_back(dbname_ + "_3", 500 * 1024);
  1096. options.db_paths.emplace_back(dbname_ + "_4", 1024 * 1024 * 1024);
  1097. options.memtable_factory.reset(
  1098. new SpecialSkipListFactory(KNumKeysByGenerateNewFile - 1));
  1099. options.compaction_style = kCompactionStyleUniversal;
  1100. options.compaction_options_universal.size_ratio = 5;
  1101. options.write_buffer_size = 111 << 10; // 114KB
  1102. options.arena_block_size = 4 << 10;
  1103. options.level0_file_num_compaction_trigger = 2;
  1104. options.num_levels = 1;
  1105. std::vector<std::string> filenames;
  1106. env_->GetChildren(options.db_paths[1].path, &filenames);
  1107. // Delete archival files.
  1108. for (size_t i = 0; i < filenames.size(); ++i) {
  1109. env_->DeleteFile(options.db_paths[1].path + "/" + filenames[i]);
  1110. }
  1111. env_->DeleteDir(options.db_paths[1].path);
  1112. Reopen(options);
  1113. Random rnd(301);
  1114. int key_idx = 0;
  1115. // First three 110KB files are not going to second path.
  1116. // After that, (100K, 200K)
  1117. for (int num = 0; num < 3; num++) {
  1118. GenerateNewFile(&rnd, &key_idx);
  1119. }
  1120. // Another 110KB triggers a compaction to 400K file to second path
  1121. GenerateNewFile(&rnd, &key_idx);
  1122. ASSERT_EQ(1, GetSstFileCount(options.db_paths[2].path));
  1123. // (1, 4)
  1124. GenerateNewFile(&rnd, &key_idx);
  1125. ASSERT_EQ(1, GetSstFileCount(options.db_paths[2].path));
  1126. ASSERT_EQ(1, GetSstFileCount(dbname_));
  1127. // (1,1,4) -> (2, 4)
  1128. GenerateNewFile(&rnd, &key_idx);
  1129. ASSERT_EQ(1, GetSstFileCount(options.db_paths[2].path));
  1130. ASSERT_EQ(1, GetSstFileCount(options.db_paths[1].path));
  1131. ASSERT_EQ(0, GetSstFileCount(dbname_));
  1132. // (1, 2, 4) -> (3, 4)
  1133. GenerateNewFile(&rnd, &key_idx);
  1134. ASSERT_EQ(1, GetSstFileCount(options.db_paths[2].path));
  1135. ASSERT_EQ(1, GetSstFileCount(options.db_paths[1].path));
  1136. ASSERT_EQ(0, GetSstFileCount(dbname_));
  1137. // (1, 3, 4) -> (8)
  1138. GenerateNewFile(&rnd, &key_idx);
  1139. ASSERT_EQ(1, GetSstFileCount(options.db_paths[3].path));
  1140. // (1, 8)
  1141. GenerateNewFile(&rnd, &key_idx);
  1142. ASSERT_EQ(1, GetSstFileCount(options.db_paths[3].path));
  1143. ASSERT_EQ(1, GetSstFileCount(dbname_));
  1144. // (1, 1, 8) -> (2, 8)
  1145. GenerateNewFile(&rnd, &key_idx);
  1146. ASSERT_EQ(1, GetSstFileCount(options.db_paths[3].path));
  1147. ASSERT_EQ(1, GetSstFileCount(options.db_paths[1].path));
  1148. // (1, 2, 8) -> (3, 8)
  1149. GenerateNewFile(&rnd, &key_idx);
  1150. ASSERT_EQ(1, GetSstFileCount(options.db_paths[3].path));
  1151. ASSERT_EQ(1, GetSstFileCount(options.db_paths[1].path));
  1152. ASSERT_EQ(0, GetSstFileCount(dbname_));
  1153. // (1, 3, 8) -> (4, 8)
  1154. GenerateNewFile(&rnd, &key_idx);
  1155. ASSERT_EQ(1, GetSstFileCount(options.db_paths[2].path));
  1156. ASSERT_EQ(1, GetSstFileCount(options.db_paths[3].path));
  1157. // (1, 4, 8) -> (5, 8)
  1158. GenerateNewFile(&rnd, &key_idx);
  1159. ASSERT_EQ(1, GetSstFileCount(options.db_paths[3].path));
  1160. ASSERT_EQ(1, GetSstFileCount(options.db_paths[2].path));
  1161. ASSERT_EQ(0, GetSstFileCount(dbname_));
  1162. for (int i = 0; i < key_idx; i++) {
  1163. auto v = Get(Key(i));
  1164. ASSERT_NE(v, "NOT_FOUND");
  1165. ASSERT_TRUE(v.size() == 1 || v.size() == 990);
  1166. }
  1167. Reopen(options);
  1168. for (int i = 0; i < key_idx; i++) {
  1169. auto v = Get(Key(i));
  1170. ASSERT_NE(v, "NOT_FOUND");
  1171. ASSERT_TRUE(v.size() == 1 || v.size() == 990);
  1172. }
  1173. Destroy(options);
  1174. }
  1175. TEST_P(DBTestUniversalCompaction, UniversalCompactionCFPathUse) {
  1176. Options options = CurrentOptions();
  1177. options.db_paths.emplace_back(dbname_, 300 * 1024);
  1178. options.db_paths.emplace_back(dbname_ + "_2", 300 * 1024);
  1179. options.db_paths.emplace_back(dbname_ + "_3", 500 * 1024);
  1180. options.db_paths.emplace_back(dbname_ + "_4", 1024 * 1024 * 1024);
  1181. options.memtable_factory.reset(
  1182. new SpecialSkipListFactory(KNumKeysByGenerateNewFile - 1));
  1183. options.compaction_style = kCompactionStyleUniversal;
  1184. options.compaction_options_universal.size_ratio = 10;
  1185. options.write_buffer_size = 111 << 10; // 114KB
  1186. options.arena_block_size = 4 << 10;
  1187. options.level0_file_num_compaction_trigger = 2;
  1188. options.num_levels = 1;
  1189. std::vector<Options> option_vector;
  1190. option_vector.emplace_back(options);
  1191. ColumnFamilyOptions cf_opt1(options), cf_opt2(options);
  1192. // Configure CF1 specific paths.
  1193. cf_opt1.cf_paths.emplace_back(dbname_ + "cf1", 300 * 1024);
  1194. cf_opt1.cf_paths.emplace_back(dbname_ + "cf1_2", 300 * 1024);
  1195. cf_opt1.cf_paths.emplace_back(dbname_ + "cf1_3", 500 * 1024);
  1196. cf_opt1.cf_paths.emplace_back(dbname_ + "cf1_4", 1024 * 1024 * 1024);
  1197. option_vector.emplace_back(DBOptions(options), cf_opt1);
  1198. CreateColumnFamilies({"one"},option_vector[1]);
  1199. // Configura CF2 specific paths.
  1200. cf_opt2.cf_paths.emplace_back(dbname_ + "cf2", 300 * 1024);
  1201. cf_opt2.cf_paths.emplace_back(dbname_ + "cf2_2", 300 * 1024);
  1202. cf_opt2.cf_paths.emplace_back(dbname_ + "cf2_3", 500 * 1024);
  1203. cf_opt2.cf_paths.emplace_back(dbname_ + "cf2_4", 1024 * 1024 * 1024);
  1204. option_vector.emplace_back(DBOptions(options), cf_opt2);
  1205. CreateColumnFamilies({"two"},option_vector[2]);
  1206. ReopenWithColumnFamilies({"default", "one", "two"}, option_vector);
  1207. Random rnd(301);
  1208. int key_idx = 0;
  1209. int key_idx1 = 0;
  1210. int key_idx2 = 0;
  1211. auto generate_file = [&]() {
  1212. GenerateNewFile(0, &rnd, &key_idx);
  1213. GenerateNewFile(1, &rnd, &key_idx1);
  1214. GenerateNewFile(2, &rnd, &key_idx2);
  1215. };
  1216. auto check_sstfilecount = [&](int path_id, int expected) {
  1217. ASSERT_EQ(expected, GetSstFileCount(options.db_paths[path_id].path));
  1218. ASSERT_EQ(expected, GetSstFileCount(cf_opt1.cf_paths[path_id].path));
  1219. ASSERT_EQ(expected, GetSstFileCount(cf_opt2.cf_paths[path_id].path));
  1220. };
  1221. auto check_getvalues = [&]() {
  1222. for (int i = 0; i < key_idx; i++) {
  1223. auto v = Get(0, Key(i));
  1224. ASSERT_NE(v, "NOT_FOUND");
  1225. ASSERT_TRUE(v.size() == 1 || v.size() == 990);
  1226. }
  1227. for (int i = 0; i < key_idx1; i++) {
  1228. auto v = Get(1, Key(i));
  1229. ASSERT_NE(v, "NOT_FOUND");
  1230. ASSERT_TRUE(v.size() == 1 || v.size() == 990);
  1231. }
  1232. for (int i = 0; i < key_idx2; i++) {
  1233. auto v = Get(2, Key(i));
  1234. ASSERT_NE(v, "NOT_FOUND");
  1235. ASSERT_TRUE(v.size() == 1 || v.size() == 990);
  1236. }
  1237. };
  1238. // First three 110KB files are not going to second path.
  1239. // After that, (100K, 200K)
  1240. for (int num = 0; num < 3; num++) {
  1241. generate_file();
  1242. }
  1243. // Another 110KB triggers a compaction to 400K file to second path
  1244. generate_file();
  1245. check_sstfilecount(2, 1);
  1246. // (1, 4)
  1247. generate_file();
  1248. check_sstfilecount(2, 1);
  1249. check_sstfilecount(0, 1);
  1250. // (1,1,4) -> (2, 4)
  1251. generate_file();
  1252. check_sstfilecount(2, 1);
  1253. check_sstfilecount(1, 1);
  1254. check_sstfilecount(0, 0);
  1255. // (1, 2, 4) -> (3, 4)
  1256. generate_file();
  1257. check_sstfilecount(2, 1);
  1258. check_sstfilecount(1, 1);
  1259. check_sstfilecount(0, 0);
  1260. // (1, 3, 4) -> (8)
  1261. generate_file();
  1262. check_sstfilecount(3, 1);
  1263. // (1, 8)
  1264. generate_file();
  1265. check_sstfilecount(3, 1);
  1266. check_sstfilecount(0, 1);
  1267. // (1, 1, 8) -> (2, 8)
  1268. generate_file();
  1269. check_sstfilecount(3, 1);
  1270. check_sstfilecount(1, 1);
  1271. // (1, 2, 8) -> (3, 8)
  1272. generate_file();
  1273. check_sstfilecount(3, 1);
  1274. check_sstfilecount(1, 1);
  1275. check_sstfilecount(0, 0);
  1276. // (1, 3, 8) -> (4, 8)
  1277. generate_file();
  1278. check_sstfilecount(2, 1);
  1279. check_sstfilecount(3, 1);
  1280. // (1, 4, 8) -> (5, 8)
  1281. generate_file();
  1282. check_sstfilecount(3, 1);
  1283. check_sstfilecount(2, 1);
  1284. check_sstfilecount(0, 0);
  1285. check_getvalues();
  1286. ReopenWithColumnFamilies({"default", "one", "two"}, option_vector);
  1287. check_getvalues();
  1288. Destroy(options, true);
  1289. }
  1290. TEST_P(DBTestUniversalCompaction, IncreaseUniversalCompactionNumLevels) {
  1291. std::function<void(int)> verify_func = [&](int num_keys_in_db) {
  1292. std::string keys_in_db;
  1293. Iterator* iter = dbfull()->NewIterator(ReadOptions(), handles_[1]);
  1294. for (iter->SeekToFirst(); iter->Valid(); iter->Next()) {
  1295. keys_in_db.append(iter->key().ToString());
  1296. keys_in_db.push_back(',');
  1297. }
  1298. delete iter;
  1299. std::string expected_keys;
  1300. for (int i = 0; i <= num_keys_in_db; i++) {
  1301. expected_keys.append(Key(i));
  1302. expected_keys.push_back(',');
  1303. }
  1304. ASSERT_EQ(keys_in_db, expected_keys);
  1305. };
  1306. Random rnd(301);
  1307. int max_key1 = 200;
  1308. int max_key2 = 600;
  1309. int max_key3 = 800;
  1310. const int KNumKeysPerFile = 10;
  1311. // Stage 1: open a DB with universal compaction, num_levels=1
  1312. Options options = CurrentOptions();
  1313. options.compaction_style = kCompactionStyleUniversal;
  1314. options.num_levels = 1;
  1315. options.write_buffer_size = 200 << 10; // 200KB
  1316. options.level0_file_num_compaction_trigger = 3;
  1317. options.memtable_factory.reset(new SpecialSkipListFactory(KNumKeysPerFile));
  1318. options = CurrentOptions(options);
  1319. CreateAndReopenWithCF({"pikachu"}, options);
  1320. for (int i = 0; i <= max_key1; i++) {
  1321. // each value is 10K
  1322. ASSERT_OK(Put(1, Key(i), RandomString(&rnd, 10000)));
  1323. dbfull()->TEST_WaitForFlushMemTable(handles_[1]);
  1324. dbfull()->TEST_WaitForCompact();
  1325. }
  1326. ASSERT_OK(Flush(1));
  1327. dbfull()->TEST_WaitForCompact();
  1328. // Stage 2: reopen with universal compaction, num_levels=4
  1329. options.compaction_style = kCompactionStyleUniversal;
  1330. options.num_levels = 4;
  1331. options = CurrentOptions(options);
  1332. ReopenWithColumnFamilies({"default", "pikachu"}, options);
  1333. verify_func(max_key1);
  1334. // Insert more keys
  1335. for (int i = max_key1 + 1; i <= max_key2; i++) {
  1336. // each value is 10K
  1337. ASSERT_OK(Put(1, Key(i), RandomString(&rnd, 10000)));
  1338. dbfull()->TEST_WaitForFlushMemTable(handles_[1]);
  1339. dbfull()->TEST_WaitForCompact();
  1340. }
  1341. ASSERT_OK(Flush(1));
  1342. dbfull()->TEST_WaitForCompact();
  1343. verify_func(max_key2);
  1344. // Compaction to non-L0 has happened.
  1345. ASSERT_GT(NumTableFilesAtLevel(options.num_levels - 1, 1), 0);
  1346. // Stage 3: Revert it back to one level and revert to num_levels=1.
  1347. options.num_levels = 4;
  1348. options.target_file_size_base = INT_MAX;
  1349. ReopenWithColumnFamilies({"default", "pikachu"}, options);
  1350. // Compact all to level 0
  1351. CompactRangeOptions compact_options;
  1352. compact_options.change_level = true;
  1353. compact_options.target_level = 0;
  1354. compact_options.exclusive_manual_compaction = exclusive_manual_compaction_;
  1355. dbfull()->CompactRange(compact_options, handles_[1], nullptr, nullptr);
  1356. // Need to restart it once to remove higher level records in manifest.
  1357. ReopenWithColumnFamilies({"default", "pikachu"}, options);
  1358. // Final reopen
  1359. options.compaction_style = kCompactionStyleUniversal;
  1360. options.num_levels = 1;
  1361. options = CurrentOptions(options);
  1362. ReopenWithColumnFamilies({"default", "pikachu"}, options);
  1363. // Insert more keys
  1364. for (int i = max_key2 + 1; i <= max_key3; i++) {
  1365. // each value is 10K
  1366. ASSERT_OK(Put(1, Key(i), RandomString(&rnd, 10000)));
  1367. dbfull()->TEST_WaitForFlushMemTable(handles_[1]);
  1368. dbfull()->TEST_WaitForCompact();
  1369. }
  1370. ASSERT_OK(Flush(1));
  1371. dbfull()->TEST_WaitForCompact();
  1372. verify_func(max_key3);
  1373. }
  1374. TEST_P(DBTestUniversalCompaction, UniversalCompactionSecondPathRatio) {
  1375. if (!Snappy_Supported()) {
  1376. return;
  1377. }
  1378. Options options = CurrentOptions();
  1379. options.db_paths.emplace_back(dbname_, 500 * 1024);
  1380. options.db_paths.emplace_back(dbname_ + "_2", 1024 * 1024 * 1024);
  1381. options.compaction_style = kCompactionStyleUniversal;
  1382. options.compaction_options_universal.size_ratio = 5;
  1383. options.write_buffer_size = 111 << 10; // 114KB
  1384. options.arena_block_size = 4 << 10;
  1385. options.level0_file_num_compaction_trigger = 2;
  1386. options.num_levels = 1;
  1387. options.memtable_factory.reset(
  1388. new SpecialSkipListFactory(KNumKeysByGenerateNewFile - 1));
  1389. std::vector<std::string> filenames;
  1390. env_->GetChildren(options.db_paths[1].path, &filenames);
  1391. // Delete archival files.
  1392. for (size_t i = 0; i < filenames.size(); ++i) {
  1393. env_->DeleteFile(options.db_paths[1].path + "/" + filenames[i]);
  1394. }
  1395. env_->DeleteDir(options.db_paths[1].path);
  1396. Reopen(options);
  1397. Random rnd(301);
  1398. int key_idx = 0;
  1399. // First three 110KB files are not going to second path.
  1400. // After that, (100K, 200K)
  1401. for (int num = 0; num < 3; num++) {
  1402. GenerateNewFile(&rnd, &key_idx);
  1403. }
  1404. // Another 110KB triggers a compaction to 400K file to second path
  1405. GenerateNewFile(&rnd, &key_idx);
  1406. ASSERT_EQ(1, GetSstFileCount(options.db_paths[1].path));
  1407. // (1, 4)
  1408. GenerateNewFile(&rnd, &key_idx);
  1409. ASSERT_EQ(1, GetSstFileCount(options.db_paths[1].path));
  1410. ASSERT_EQ(1, GetSstFileCount(dbname_));
  1411. // (1,1,4) -> (2, 4)
  1412. GenerateNewFile(&rnd, &key_idx);
  1413. ASSERT_EQ(1, GetSstFileCount(options.db_paths[1].path));
  1414. ASSERT_EQ(1, GetSstFileCount(dbname_));
  1415. // (1, 2, 4) -> (3, 4)
  1416. GenerateNewFile(&rnd, &key_idx);
  1417. ASSERT_EQ(2, GetSstFileCount(options.db_paths[1].path));
  1418. ASSERT_EQ(0, GetSstFileCount(dbname_));
  1419. // (1, 3, 4) -> (8)
  1420. GenerateNewFile(&rnd, &key_idx);
  1421. ASSERT_EQ(1, GetSstFileCount(options.db_paths[1].path));
  1422. ASSERT_EQ(0, GetSstFileCount(dbname_));
  1423. // (1, 8)
  1424. GenerateNewFile(&rnd, &key_idx);
  1425. ASSERT_EQ(1, GetSstFileCount(options.db_paths[1].path));
  1426. ASSERT_EQ(1, GetSstFileCount(dbname_));
  1427. // (1, 1, 8) -> (2, 8)
  1428. GenerateNewFile(&rnd, &key_idx);
  1429. ASSERT_EQ(1, GetSstFileCount(options.db_paths[1].path));
  1430. ASSERT_EQ(1, GetSstFileCount(dbname_));
  1431. // (1, 2, 8) -> (3, 8)
  1432. GenerateNewFile(&rnd, &key_idx);
  1433. ASSERT_EQ(2, GetSstFileCount(options.db_paths[1].path));
  1434. ASSERT_EQ(0, GetSstFileCount(dbname_));
  1435. // (1, 3, 8) -> (4, 8)
  1436. GenerateNewFile(&rnd, &key_idx);
  1437. ASSERT_EQ(2, GetSstFileCount(options.db_paths[1].path));
  1438. ASSERT_EQ(0, GetSstFileCount(dbname_));
  1439. // (1, 4, 8) -> (5, 8)
  1440. GenerateNewFile(&rnd, &key_idx);
  1441. ASSERT_EQ(2, GetSstFileCount(options.db_paths[1].path));
  1442. ASSERT_EQ(0, GetSstFileCount(dbname_));
  1443. for (int i = 0; i < key_idx; i++) {
  1444. auto v = Get(Key(i));
  1445. ASSERT_NE(v, "NOT_FOUND");
  1446. ASSERT_TRUE(v.size() == 1 || v.size() == 990);
  1447. }
  1448. Reopen(options);
  1449. for (int i = 0; i < key_idx; i++) {
  1450. auto v = Get(Key(i));
  1451. ASSERT_NE(v, "NOT_FOUND");
  1452. ASSERT_TRUE(v.size() == 1 || v.size() == 990);
  1453. }
  1454. Destroy(options);
  1455. }
  1456. TEST_P(DBTestUniversalCompaction, ConcurrentBottomPriLowPriCompactions) {
  1457. if (num_levels_ == 1) {
  1458. // for single-level universal, everything's bottom level so nothing should
  1459. // be executed in bottom-pri thread pool.
  1460. return;
  1461. }
  1462. const int kNumFilesTrigger = 3;
  1463. Env::Default()->SetBackgroundThreads(1, Env::Priority::BOTTOM);
  1464. Options options = CurrentOptions();
  1465. options.compaction_style = kCompactionStyleUniversal;
  1466. options.num_levels = num_levels_;
  1467. options.write_buffer_size = 100 << 10; // 100KB
  1468. options.target_file_size_base = 32 << 10; // 32KB
  1469. options.level0_file_num_compaction_trigger = kNumFilesTrigger;
  1470. // Trigger compaction if size amplification exceeds 110%
  1471. options.compaction_options_universal.max_size_amplification_percent = 110;
  1472. DestroyAndReopen(options);
  1473. ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->LoadDependency(
  1474. {// wait for the full compaction to be picked before adding files intended
  1475. // for the second one.
  1476. {"DBImpl::BackgroundCompaction:ForwardToBottomPriPool",
  1477. "DBTestUniversalCompaction:ConcurrentBottomPriLowPriCompactions:0"},
  1478. // the full (bottom-pri) compaction waits until a partial (low-pri)
  1479. // compaction has started to verify they can run in parallel.
  1480. {"DBImpl::BackgroundCompaction:NonTrivial",
  1481. "DBImpl::BGWorkBottomCompaction"}});
  1482. SyncPoint::GetInstance()->EnableProcessing();
  1483. Random rnd(301);
  1484. for (int i = 0; i < 2; ++i) {
  1485. for (int num = 0; num < kNumFilesTrigger; num++) {
  1486. int key_idx = 0;
  1487. GenerateNewFile(&rnd, &key_idx, true /* no_wait */);
  1488. // use no_wait above because that one waits for flush and compaction. We
  1489. // don't want to wait for compaction because the full compaction is
  1490. // intentionally blocked while more files are flushed.
  1491. dbfull()->TEST_WaitForFlushMemTable();
  1492. }
  1493. if (i == 0) {
  1494. TEST_SYNC_POINT(
  1495. "DBTestUniversalCompaction:ConcurrentBottomPriLowPriCompactions:0");
  1496. }
  1497. }
  1498. dbfull()->TEST_WaitForCompact();
  1499. // First compaction should output to bottom level. Second should output to L0
  1500. // since older L0 files pending compaction prevent it from being placed lower.
  1501. ASSERT_EQ(NumSortedRuns(), 2);
  1502. ASSERT_GT(NumTableFilesAtLevel(0), 0);
  1503. ASSERT_GT(NumTableFilesAtLevel(num_levels_ - 1), 0);
  1504. ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->DisableProcessing();
  1505. Env::Default()->SetBackgroundThreads(0, Env::Priority::BOTTOM);
  1506. }
  1507. TEST_P(DBTestUniversalCompaction, RecalculateScoreAfterPicking) {
  1508. // Regression test for extra compactions scheduled. Once enough compactions
  1509. // have been scheduled to bring the score below one, we should stop
  1510. // scheduling more; otherwise, other CFs/DBs may be delayed unnecessarily.
  1511. const int kNumFilesTrigger = 8;
  1512. Options options = CurrentOptions();
  1513. options.memtable_factory.reset(
  1514. new SpecialSkipListFactory(KNumKeysByGenerateNewFile - 1));
  1515. options.compaction_options_universal.max_merge_width = kNumFilesTrigger / 2;
  1516. options.compaction_options_universal.max_size_amplification_percent =
  1517. static_cast<unsigned int>(-1);
  1518. options.compaction_style = kCompactionStyleUniversal;
  1519. options.level0_file_num_compaction_trigger = kNumFilesTrigger;
  1520. options.num_levels = num_levels_;
  1521. Reopen(options);
  1522. std::atomic<int> num_compactions_attempted(0);
  1523. ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->SetCallBack(
  1524. "DBImpl::BackgroundCompaction:Start",
  1525. [&](void* /*arg*/) { ++num_compactions_attempted; });
  1526. ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->EnableProcessing();
  1527. Random rnd(301);
  1528. for (int num = 0; num < kNumFilesTrigger; num++) {
  1529. ASSERT_EQ(NumSortedRuns(), num);
  1530. int key_idx = 0;
  1531. GenerateNewFile(&rnd, &key_idx);
  1532. }
  1533. dbfull()->TEST_WaitForCompact();
  1534. // Compacting the first four files was enough to bring the score below one so
  1535. // there's no need to schedule any more compactions.
  1536. ASSERT_EQ(1, num_compactions_attempted);
  1537. ASSERT_EQ(NumSortedRuns(), 5);
  1538. }
  1539. TEST_P(DBTestUniversalCompaction, FinalSortedRunCompactFilesConflict) {
  1540. // Regression test for conflict between:
  1541. // (1) Running CompactFiles including file in the final sorted run; and
  1542. // (2) Picking universal size-amp-triggered compaction, which always includes
  1543. // the final sorted run.
  1544. if (exclusive_manual_compaction_) {
  1545. return;
  1546. }
  1547. Options opts = CurrentOptions();
  1548. opts.compaction_style = kCompactionStyleUniversal;
  1549. opts.compaction_options_universal.max_size_amplification_percent = 50;
  1550. opts.compaction_options_universal.min_merge_width = 2;
  1551. opts.compression = kNoCompression;
  1552. opts.level0_file_num_compaction_trigger = 2;
  1553. opts.max_background_compactions = 2;
  1554. opts.num_levels = num_levels_;
  1555. Reopen(opts);
  1556. // make sure compaction jobs can be parallelized
  1557. auto stop_token =
  1558. dbfull()->TEST_write_controler().GetCompactionPressureToken();
  1559. Put("key", "val");
  1560. Flush();
  1561. dbfull()->CompactRange(CompactRangeOptions(), nullptr, nullptr);
  1562. ASSERT_EQ(NumTableFilesAtLevel(num_levels_ - 1), 1);
  1563. ColumnFamilyMetaData cf_meta;
  1564. ColumnFamilyHandle* default_cfh = db_->DefaultColumnFamily();
  1565. dbfull()->GetColumnFamilyMetaData(default_cfh, &cf_meta);
  1566. ASSERT_EQ(1, cf_meta.levels[num_levels_ - 1].files.size());
  1567. std::string first_sst_filename =
  1568. cf_meta.levels[num_levels_ - 1].files[0].name;
  1569. ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->LoadDependency(
  1570. {{"CompactFilesImpl:0",
  1571. "DBTestUniversalCompaction:FinalSortedRunCompactFilesConflict:0"},
  1572. {"DBImpl::BackgroundCompaction():AfterPickCompaction",
  1573. "CompactFilesImpl:1"}});
  1574. ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->EnableProcessing();
  1575. port::Thread compact_files_thread([&]() {
  1576. ASSERT_OK(dbfull()->CompactFiles(CompactionOptions(), default_cfh,
  1577. {first_sst_filename}, num_levels_ - 1));
  1578. });
  1579. TEST_SYNC_POINT(
  1580. "DBTestUniversalCompaction:FinalSortedRunCompactFilesConflict:0");
  1581. for (int i = 0; i < 2; ++i) {
  1582. Put("key", "val");
  1583. Flush();
  1584. }
  1585. dbfull()->TEST_WaitForCompact();
  1586. compact_files_thread.join();
  1587. }
  1588. INSTANTIATE_TEST_CASE_P(NumLevels, DBTestUniversalCompaction,
  1589. ::testing::Combine(::testing::Values(1, 3, 5),
  1590. ::testing::Bool()));
  1591. class DBTestUniversalManualCompactionOutputPathId
  1592. : public DBTestUniversalCompactionBase {
  1593. public:
  1594. DBTestUniversalManualCompactionOutputPathId() :
  1595. DBTestUniversalCompactionBase(
  1596. "/db_universal_compaction_manual_pid_test") {}
  1597. };
  1598. TEST_P(DBTestUniversalManualCompactionOutputPathId,
  1599. ManualCompactionOutputPathId) {
  1600. Options options = CurrentOptions();
  1601. options.create_if_missing = true;
  1602. options.db_paths.emplace_back(dbname_, 1000000000);
  1603. options.db_paths.emplace_back(dbname_ + "_2", 1000000000);
  1604. options.compaction_style = kCompactionStyleUniversal;
  1605. options.num_levels = num_levels_;
  1606. options.target_file_size_base = 1 << 30; // Big size
  1607. options.level0_file_num_compaction_trigger = 10;
  1608. Destroy(options);
  1609. DestroyAndReopen(options);
  1610. CreateAndReopenWithCF({"pikachu"}, options);
  1611. MakeTables(3, "p", "q", 1);
  1612. dbfull()->TEST_WaitForCompact();
  1613. ASSERT_EQ(2, TotalLiveFiles(1));
  1614. ASSERT_EQ(2, GetSstFileCount(options.db_paths[0].path));
  1615. ASSERT_EQ(0, GetSstFileCount(options.db_paths[1].path));
  1616. // Full compaction to DB path 0
  1617. CompactRangeOptions compact_options;
  1618. compact_options.target_path_id = 1;
  1619. compact_options.exclusive_manual_compaction = exclusive_manual_compaction_;
  1620. db_->CompactRange(compact_options, handles_[1], nullptr, nullptr);
  1621. ASSERT_EQ(1, TotalLiveFiles(1));
  1622. ASSERT_EQ(0, GetSstFileCount(options.db_paths[0].path));
  1623. ASSERT_EQ(1, GetSstFileCount(options.db_paths[1].path));
  1624. ReopenWithColumnFamilies({kDefaultColumnFamilyName, "pikachu"}, options);
  1625. ASSERT_EQ(1, TotalLiveFiles(1));
  1626. ASSERT_EQ(0, GetSstFileCount(options.db_paths[0].path));
  1627. ASSERT_EQ(1, GetSstFileCount(options.db_paths[1].path));
  1628. MakeTables(1, "p", "q", 1);
  1629. ASSERT_EQ(2, TotalLiveFiles(1));
  1630. ASSERT_EQ(1, GetSstFileCount(options.db_paths[0].path));
  1631. ASSERT_EQ(1, GetSstFileCount(options.db_paths[1].path));
  1632. ReopenWithColumnFamilies({kDefaultColumnFamilyName, "pikachu"}, options);
  1633. ASSERT_EQ(2, TotalLiveFiles(1));
  1634. ASSERT_EQ(1, GetSstFileCount(options.db_paths[0].path));
  1635. ASSERT_EQ(1, GetSstFileCount(options.db_paths[1].path));
  1636. // Full compaction to DB path 0
  1637. compact_options.target_path_id = 0;
  1638. compact_options.exclusive_manual_compaction = exclusive_manual_compaction_;
  1639. db_->CompactRange(compact_options, handles_[1], nullptr, nullptr);
  1640. ASSERT_EQ(1, TotalLiveFiles(1));
  1641. ASSERT_EQ(1, GetSstFileCount(options.db_paths[0].path));
  1642. ASSERT_EQ(0, GetSstFileCount(options.db_paths[1].path));
  1643. // Fail when compacting to an invalid path ID
  1644. compact_options.target_path_id = 2;
  1645. compact_options.exclusive_manual_compaction = exclusive_manual_compaction_;
  1646. ASSERT_TRUE(db_->CompactRange(compact_options, handles_[1], nullptr, nullptr)
  1647. .IsInvalidArgument());
  1648. }
  1649. INSTANTIATE_TEST_CASE_P(OutputPathId,
  1650. DBTestUniversalManualCompactionOutputPathId,
  1651. ::testing::Combine(::testing::Values(1, 8),
  1652. ::testing::Bool()));
  1653. TEST_F(DBTestUniversalCompaction2, BasicL0toL1) {
  1654. const int kNumKeys = 3000;
  1655. const int kWindowSize = 100;
  1656. const int kNumDelsTrigger = 90;
  1657. Options opts = CurrentOptions();
  1658. opts.table_properties_collector_factories.emplace_back(
  1659. NewCompactOnDeletionCollectorFactory(kWindowSize, kNumDelsTrigger));
  1660. opts.compaction_style = kCompactionStyleUniversal;
  1661. opts.level0_file_num_compaction_trigger = 2;
  1662. opts.compression = kNoCompression;
  1663. opts.compaction_options_universal.size_ratio = 10;
  1664. opts.compaction_options_universal.min_merge_width = 2;
  1665. opts.compaction_options_universal.max_size_amplification_percent = 200;
  1666. Reopen(opts);
  1667. // add an L1 file to prevent tombstones from dropping due to obsolescence
  1668. // during flush
  1669. int i;
  1670. for (i = 0; i < 2000; ++i) {
  1671. Put(Key(i), "val");
  1672. }
  1673. Flush();
  1674. // MoveFilesToLevel(6);
  1675. dbfull()->CompactRange(CompactRangeOptions(), nullptr, nullptr);
  1676. for (i = 1999; i < kNumKeys; ++i) {
  1677. if (i >= kNumKeys - kWindowSize &&
  1678. i < kNumKeys - kWindowSize + kNumDelsTrigger) {
  1679. Delete(Key(i));
  1680. } else {
  1681. Put(Key(i), "val");
  1682. }
  1683. }
  1684. Flush();
  1685. dbfull()->TEST_WaitForCompact();
  1686. ASSERT_EQ(0, NumTableFilesAtLevel(0));
  1687. ASSERT_GT(NumTableFilesAtLevel(6), 0);
  1688. }
  1689. TEST_F(DBTestUniversalCompaction2, SingleLevel) {
  1690. const int kNumKeys = 3000;
  1691. const int kWindowSize = 100;
  1692. const int kNumDelsTrigger = 90;
  1693. Options opts = CurrentOptions();
  1694. opts.table_properties_collector_factories.emplace_back(
  1695. NewCompactOnDeletionCollectorFactory(kWindowSize, kNumDelsTrigger));
  1696. opts.compaction_style = kCompactionStyleUniversal;
  1697. opts.level0_file_num_compaction_trigger = 2;
  1698. opts.compression = kNoCompression;
  1699. opts.num_levels = 1;
  1700. opts.compaction_options_universal.size_ratio = 10;
  1701. opts.compaction_options_universal.min_merge_width = 2;
  1702. opts.compaction_options_universal.max_size_amplification_percent = 200;
  1703. Reopen(opts);
  1704. // add an L1 file to prevent tombstones from dropping due to obsolescence
  1705. // during flush
  1706. int i;
  1707. for (i = 0; i < 2000; ++i) {
  1708. Put(Key(i), "val");
  1709. }
  1710. Flush();
  1711. for (i = 1999; i < kNumKeys; ++i) {
  1712. if (i >= kNumKeys - kWindowSize &&
  1713. i < kNumKeys - kWindowSize + kNumDelsTrigger) {
  1714. Delete(Key(i));
  1715. } else {
  1716. Put(Key(i), "val");
  1717. }
  1718. }
  1719. Flush();
  1720. dbfull()->TEST_WaitForCompact();
  1721. ASSERT_EQ(1, NumTableFilesAtLevel(0));
  1722. }
  1723. TEST_F(DBTestUniversalCompaction2, MultipleLevels) {
  1724. const int kWindowSize = 100;
  1725. const int kNumDelsTrigger = 90;
  1726. Options opts = CurrentOptions();
  1727. opts.table_properties_collector_factories.emplace_back(
  1728. NewCompactOnDeletionCollectorFactory(kWindowSize, kNumDelsTrigger));
  1729. opts.compaction_style = kCompactionStyleUniversal;
  1730. opts.level0_file_num_compaction_trigger = 4;
  1731. opts.compression = kNoCompression;
  1732. opts.compaction_options_universal.size_ratio = 10;
  1733. opts.compaction_options_universal.min_merge_width = 2;
  1734. opts.compaction_options_universal.max_size_amplification_percent = 200;
  1735. Reopen(opts);
  1736. // add an L1 file to prevent tombstones from dropping due to obsolescence
  1737. // during flush
  1738. int i;
  1739. for (i = 0; i < 500; ++i) {
  1740. Put(Key(i), "val");
  1741. }
  1742. Flush();
  1743. for (i = 500; i < 1000; ++i) {
  1744. Put(Key(i), "val");
  1745. }
  1746. Flush();
  1747. for (i = 1000; i < 1500; ++i) {
  1748. Put(Key(i), "val");
  1749. }
  1750. Flush();
  1751. for (i = 1500; i < 2000; ++i) {
  1752. Put(Key(i), "val");
  1753. }
  1754. Flush();
  1755. dbfull()->TEST_WaitForCompact();
  1756. ASSERT_EQ(0, NumTableFilesAtLevel(0));
  1757. ASSERT_GT(NumTableFilesAtLevel(6), 0);
  1758. for (i = 1999; i < 2333; ++i) {
  1759. Put(Key(i), "val");
  1760. }
  1761. Flush();
  1762. for (i = 2333; i < 2666; ++i) {
  1763. Put(Key(i), "val");
  1764. }
  1765. Flush();
  1766. for (i = 2666; i < 2999; ++i) {
  1767. Put(Key(i), "val");
  1768. }
  1769. Flush();
  1770. dbfull()->TEST_WaitForCompact();
  1771. ASSERT_EQ(0, NumTableFilesAtLevel(0));
  1772. ASSERT_GT(NumTableFilesAtLevel(6), 0);
  1773. ASSERT_GT(NumTableFilesAtLevel(5), 0);
  1774. for (i = 1900; i < 2100; ++i) {
  1775. Delete(Key(i));
  1776. }
  1777. Flush();
  1778. dbfull()->TEST_WaitForCompact();
  1779. ASSERT_EQ(0, NumTableFilesAtLevel(0));
  1780. ASSERT_EQ(0, NumTableFilesAtLevel(1));
  1781. ASSERT_EQ(0, NumTableFilesAtLevel(2));
  1782. ASSERT_EQ(0, NumTableFilesAtLevel(3));
  1783. ASSERT_EQ(0, NumTableFilesAtLevel(4));
  1784. ASSERT_EQ(0, NumTableFilesAtLevel(5));
  1785. ASSERT_GT(NumTableFilesAtLevel(6), 0);
  1786. }
  1787. TEST_F(DBTestUniversalCompaction2, OverlappingL0) {
  1788. const int kWindowSize = 100;
  1789. const int kNumDelsTrigger = 90;
  1790. Options opts = CurrentOptions();
  1791. opts.table_properties_collector_factories.emplace_back(
  1792. NewCompactOnDeletionCollectorFactory(kWindowSize, kNumDelsTrigger));
  1793. opts.compaction_style = kCompactionStyleUniversal;
  1794. opts.level0_file_num_compaction_trigger = 5;
  1795. opts.compression = kNoCompression;
  1796. opts.compaction_options_universal.size_ratio = 10;
  1797. opts.compaction_options_universal.min_merge_width = 2;
  1798. opts.compaction_options_universal.max_size_amplification_percent = 200;
  1799. Reopen(opts);
  1800. // add an L1 file to prevent tombstones from dropping due to obsolescence
  1801. // during flush
  1802. int i;
  1803. for (i = 0; i < 2000; ++i) {
  1804. Put(Key(i), "val");
  1805. }
  1806. Flush();
  1807. for (i = 2000; i < 3000; ++i) {
  1808. Put(Key(i), "val");
  1809. }
  1810. Flush();
  1811. for (i = 3500; i < 4000; ++i) {
  1812. Put(Key(i), "val");
  1813. }
  1814. Flush();
  1815. for (i = 2900; i < 3100; ++i) {
  1816. Delete(Key(i));
  1817. }
  1818. Flush();
  1819. dbfull()->TEST_WaitForCompact();
  1820. ASSERT_EQ(2, NumTableFilesAtLevel(0));
  1821. ASSERT_GT(NumTableFilesAtLevel(6), 0);
  1822. }
  1823. TEST_F(DBTestUniversalCompaction2, IngestBehind) {
  1824. const int kNumKeys = 3000;
  1825. const int kWindowSize = 100;
  1826. const int kNumDelsTrigger = 90;
  1827. Options opts = CurrentOptions();
  1828. opts.table_properties_collector_factories.emplace_back(
  1829. NewCompactOnDeletionCollectorFactory(kWindowSize, kNumDelsTrigger));
  1830. opts.compaction_style = kCompactionStyleUniversal;
  1831. opts.level0_file_num_compaction_trigger = 2;
  1832. opts.compression = kNoCompression;
  1833. opts.allow_ingest_behind = true;
  1834. opts.compaction_options_universal.size_ratio = 10;
  1835. opts.compaction_options_universal.min_merge_width = 2;
  1836. opts.compaction_options_universal.max_size_amplification_percent = 200;
  1837. Reopen(opts);
  1838. // add an L1 file to prevent tombstones from dropping due to obsolescence
  1839. // during flush
  1840. int i;
  1841. for (i = 0; i < 2000; ++i) {
  1842. Put(Key(i), "val");
  1843. }
  1844. Flush();
  1845. // MoveFilesToLevel(6);
  1846. dbfull()->CompactRange(CompactRangeOptions(), nullptr, nullptr);
  1847. for (i = 1999; i < kNumKeys; ++i) {
  1848. if (i >= kNumKeys - kWindowSize &&
  1849. i < kNumKeys - kWindowSize + kNumDelsTrigger) {
  1850. Delete(Key(i));
  1851. } else {
  1852. Put(Key(i), "val");
  1853. }
  1854. }
  1855. Flush();
  1856. dbfull()->TEST_WaitForCompact();
  1857. ASSERT_EQ(0, NumTableFilesAtLevel(0));
  1858. ASSERT_EQ(0, NumTableFilesAtLevel(6));
  1859. ASSERT_GT(NumTableFilesAtLevel(5), 0);
  1860. }
  1861. TEST_F(DBTestUniversalCompaction2, PeriodicCompactionDefault) {
  1862. Options options;
  1863. options.compaction_style = kCompactionStyleUniversal;
  1864. KeepFilterFactory* filter = new KeepFilterFactory(true);
  1865. options.compaction_filter_factory.reset(filter);
  1866. Reopen(options);
  1867. ASSERT_EQ(30 * 24 * 60 * 60,
  1868. dbfull()->GetOptions().periodic_compaction_seconds);
  1869. KeepFilter df;
  1870. options.compaction_filter_factory.reset();
  1871. options.compaction_filter = &df;
  1872. Reopen(options);
  1873. ASSERT_EQ(30 * 24 * 60 * 60,
  1874. dbfull()->GetOptions().periodic_compaction_seconds);
  1875. options.ttl = 60 * 24 * 60 * 60;
  1876. options.compaction_filter = nullptr;
  1877. Reopen(options);
  1878. ASSERT_EQ(60 * 24 * 60 * 60,
  1879. dbfull()->GetOptions().periodic_compaction_seconds);
  1880. }
  1881. TEST_F(DBTestUniversalCompaction2, PeriodicCompaction) {
  1882. Options opts = CurrentOptions();
  1883. opts.env = env_;
  1884. opts.compaction_style = kCompactionStyleUniversal;
  1885. opts.level0_file_num_compaction_trigger = 10;
  1886. opts.max_open_files = -1;
  1887. opts.compaction_options_universal.size_ratio = 10;
  1888. opts.compaction_options_universal.min_merge_width = 2;
  1889. opts.compaction_options_universal.max_size_amplification_percent = 200;
  1890. opts.periodic_compaction_seconds = 48 * 60 * 60; // 2 days
  1891. opts.num_levels = 5;
  1892. env_->addon_time_.store(0);
  1893. Reopen(opts);
  1894. int periodic_compactions = 0;
  1895. int start_level = -1;
  1896. int output_level = -1;
  1897. ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->SetCallBack(
  1898. "UniversalCompactionPicker::PickPeriodicCompaction:Return",
  1899. [&](void* arg) {
  1900. Compaction* compaction = reinterpret_cast<Compaction*>(arg);
  1901. ASSERT_TRUE(arg != nullptr);
  1902. ASSERT_TRUE(compaction->compaction_reason() ==
  1903. CompactionReason::kPeriodicCompaction);
  1904. start_level = compaction->start_level();
  1905. output_level = compaction->output_level();
  1906. periodic_compactions++;
  1907. });
  1908. ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->EnableProcessing();
  1909. // Case 1: Oldest flushed file excceeds periodic compaction threshold.
  1910. ASSERT_OK(Put("foo", "bar"));
  1911. Flush();
  1912. ASSERT_EQ(0, periodic_compactions);
  1913. // Move clock forward so that the flushed file would qualify periodic
  1914. // compaction.
  1915. env_->addon_time_.store(48 * 60 * 60 + 100);
  1916. // Another flush would trigger compaction the oldest file.
  1917. ASSERT_OK(Put("foo", "bar2"));
  1918. Flush();
  1919. dbfull()->TEST_WaitForCompact();
  1920. ASSERT_EQ(1, periodic_compactions);
  1921. ASSERT_EQ(0, start_level);
  1922. ASSERT_EQ(4, output_level);
  1923. // Case 2: Oldest compacted file excceeds periodic compaction threshold
  1924. periodic_compactions = 0;
  1925. // A flush doesn't trigger a periodic compaction when threshold not hit
  1926. ASSERT_OK(Put("foo", "bar2"));
  1927. Flush();
  1928. dbfull()->TEST_WaitForCompact();
  1929. ASSERT_EQ(0, periodic_compactions);
  1930. // After periodic compaction threshold hits, a flush will trigger
  1931. // a compaction
  1932. ASSERT_OK(Put("foo", "bar2"));
  1933. env_->addon_time_.fetch_add(48 * 60 * 60 + 100);
  1934. Flush();
  1935. dbfull()->TEST_WaitForCompact();
  1936. ASSERT_EQ(1, periodic_compactions);
  1937. ASSERT_EQ(0, start_level);
  1938. ASSERT_EQ(4, output_level);
  1939. }
  1940. } // namespace ROCKSDB_NAMESPACE
  1941. #endif // !defined(ROCKSDB_LITE)
  1942. int main(int argc, char** argv) {
  1943. #if !defined(ROCKSDB_LITE)
  1944. ROCKSDB_NAMESPACE::port::InstallStackTraceHandler();
  1945. ::testing::InitGoogleTest(&argc, argv);
  1946. return RUN_ALL_TESTS();
  1947. #else
  1948. (void) argc;
  1949. (void) argv;
  1950. return 0;
  1951. #endif
  1952. }