compare mace remote store and rocksdb blobdb -- v1

This commit is contained in:
abbycin 2025-10-15 08:57:21 +08:00
parent 07baaae51e
commit 1bb175101c
Signed by: abby
GPG Key ID: B636E0F0307EF8EB
2 changed files with 43 additions and 17 deletions

View File

@ -1,10 +1,13 @@
#include <algorithm> #include <algorithm>
#include <atomic> #include <atomic>
#include <cstdint>
#include <fmt/format.h> #include <fmt/format.h>
#include <memory>
#include <random> #include <random>
#include <rocksdb/cache.h> #include <rocksdb/cache.h>
#include <rocksdb/db.h> #include <rocksdb/db.h>
#include <rocksdb/options.h> #include <rocksdb/options.h>
#include <rocksdb/table.h>
#include <rocksdb/utilities/optimistic_transaction_db.h> #include <rocksdb/utilities/optimistic_transaction_db.h>
#include <rocksdb/utilities/transaction.h> #include <rocksdb/utilities/transaction.h>
#include <rocksdb/utilities/transaction_db.h> #include <rocksdb/utilities/transaction_db.h>
@ -76,17 +79,30 @@ int main(int argc, char *argv[]) {
return 1; return 1;
} }
rocksdb::Options options; rocksdb::ColumnFamilyOptions cfo{};
cfo.enable_blob_files = true;
cfo.min_blob_size = 8192;
// use 1GB block cache
auto cache = rocksdb::NewLRUCache(1 << 30);
rocksdb::BlockBasedTableOptions table_options{};
table_options.block_cache = cache;
cfo.table_factory.reset(NewBlockBasedTableFactory(table_options));
// the following three options makes it not trigger GC in test
cfo.level0_file_num_compaction_trigger = 1000;
cfo.write_buffer_size = 1 << 30;
cfo.max_write_buffer_number = 5;
std::vector<rocksdb::ColumnFamilyDescriptor> cfd{};
cfd.push_back(rocksdb::ColumnFamilyDescriptor("default", cfo));
rocksdb::DBOptions options;
options.create_if_missing = true; options.create_if_missing = true;
options.allow_concurrent_memtable_write = true; options.allow_concurrent_memtable_write = true;
options.enable_pipelined_write = true; options.enable_pipelined_write = true;
// the following three options makes it not trigger GC in test
options.level0_file_num_compaction_trigger = 1000;
options.write_buffer_size = 1 << 30;
options.max_write_buffer_number = 5;
auto ropt = rocksdb::ReadOptions(); auto ropt = rocksdb::ReadOptions();
auto wopt = rocksdb::WriteOptions(); auto wopt = rocksdb::WriteOptions();
wopt.no_slowdown = true;
// wopt.disableWAL = true; // wopt.disableWAL = true;
std::vector<std::thread> wg; std::vector<std::thread> wg;
std::vector<std::vector<std::string>> keys{}; std::vector<std::vector<std::string>> keys{};
@ -94,7 +110,8 @@ int main(int argc, char *argv[]) {
rocksdb::OptimisticTransactionDB *db; rocksdb::OptimisticTransactionDB *db;
auto b = nm::Instant::now(); auto b = nm::Instant::now();
std::mutex mtx{}; std::mutex mtx{};
auto s = rocksdb::OptimisticTransactionDB::Open(options, args.path, &db); std::vector<rocksdb::ColumnFamilyHandle *> handles{};
auto s = rocksdb::OptimisticTransactionDB::Open(options, args.path, cfd, &handles, &db);
assert(s.ok()); assert(s.ok());
std::barrier barrier{static_cast<ptrdiff_t>(args.threads)}; std::barrier barrier{static_cast<ptrdiff_t>(args.threads)};
@ -117,23 +134,27 @@ int main(int argc, char *argv[]) {
keys.emplace_back(std::move(key)); keys.emplace_back(std::move(key));
} }
auto *handle = handles[0];
if (args.mode == "get") { if (args.mode == "get") {
auto *kv = db->BeginTransaction(wopt); auto *kv = db->BeginTransaction(wopt);
for (size_t tid = 0; tid < args.threads; ++tid) { for (size_t tid = 0; tid < args.threads; ++tid) {
auto *tk = &keys[tid]; auto *tk = &keys[tid];
for (auto &key: *tk) { for (auto &key: *tk) {
kv->Put(key, val); kv->Put(handle, key, val);
} }
} }
kv->Commit(); kv->Commit();
delete kv; delete kv;
delete handle;
delete db; delete db;
handles.clear();
// re-open db // re-open db
s = rocksdb::OptimisticTransactionDB::Open(options, args.path, &db); s = rocksdb::OptimisticTransactionDB::Open(options, args.path, cfd, &handles, &db);
assert(s.ok()); assert(s.ok());
} }
handle = handles[0];
for (size_t tid = 0; tid < args.threads; ++tid) { for (size_t tid = 0; tid < args.threads; ++tid) {
auto *tk = &keys[tid]; auto *tk = &keys[tid];
wg.emplace_back([&] { wg.emplace_back([&] {
@ -147,7 +168,7 @@ int main(int argc, char *argv[]) {
if (args.mode == "insert") { if (args.mode == "insert") {
for (auto &key: *tk) { for (auto &key: *tk) {
auto *kv = db->BeginTransaction(wopt); auto *kv = db->BeginTransaction(wopt);
kv->Put(key, val); kv->Put(handle, key, val);
kv->Commit(); kv->Commit();
delete kv; delete kv;
} }
@ -155,7 +176,7 @@ int main(int argc, char *argv[]) {
} else if (args.mode == "get") { } else if (args.mode == "get") {
for (auto &key: *tk) { for (auto &key: *tk) {
auto *kv = db->BeginTransaction(wopt); auto *kv = db->BeginTransaction(wopt);
kv->Get(ropt, key, &rval); kv->Get(ropt, handle, key, &rval);
kv->Commit(); kv->Commit();
delete kv; delete kv;
} }
@ -164,9 +185,9 @@ int main(int argc, char *argv[]) {
auto is_insert = dist(gen) < args.insert_ratio; auto is_insert = dist(gen) < args.insert_ratio;
auto *kv = db->BeginTransaction(wopt); auto *kv = db->BeginTransaction(wopt);
if (is_insert) { if (is_insert) {
kv->Put(key, val); kv->Put(handle, key, val);
} else { } else {
kv->Get(ropt, key, &rval); // not found kv->Get(ropt, handle, key, &rval); // not found
} }
kv->Commit(); kv->Commit();
delete kv; delete kv;
@ -184,9 +205,10 @@ int main(int argc, char *argv[]) {
return args.insert_ratio; return args.insert_ratio;
return args.mode == "insert" ? 100 : 0; return args.mode == "insert" ? 100 : 0;
}(); }();
double ops = static_cast<double>(total_op.load(std::memory_order_relaxed)) / b.elapse_sec(); uint64_t ops = total_op.load(std::memory_order_relaxed) / b.elapse_sec();
fmt::println("{},{},{},{},{},{:.2f},{}", args.mode, args.threads, args.key_size, args.value_size, ratio, ops, fmt::println("{},{},{},{},{},{},{}", args.mode, args.threads, args.key_size, args.value_size, ratio, (uint64_t) ops,
b.elapse_ms()); (uint64_t) b.elapse_ms());
delete handle;
delete db; delete db;
std::filesystem::remove_all(args.path); std::filesystem::remove_all(args.path);
} }

View File

@ -41,6 +41,9 @@ struct Args {
#[arg(long, default_value = "false")] #[arg(long, default_value = "false")]
random: bool, random: bool,
#[arg(long, default_value = "8192")]
blob_size: usize,
} }
fn main() { fn main() {
@ -76,10 +79,11 @@ fn main() {
let mut keys: Vec<Vec<Vec<u8>>> = Vec::with_capacity(args.threads); let mut keys: Vec<Vec<Vec<u8>>> = Vec::with_capacity(args.threads);
let mut opt = Options::new(path); let mut opt = Options::new(path);
opt.sync_on_write = false; opt.sync_on_write = false;
opt.over_provision = true; // large value will use lots of memeory
opt.inline_size = args.blob_size;
opt.tmp_store = args.mode != "get"; opt.tmp_store = args.mode != "get";
let mut saved = opt.clone(); let mut saved = opt.clone();
saved.tmp_store = false; saved.tmp_store = false;
// opt.cache_capacity = 3 << 30; // this is very important for large key-value store
let mut db = Mace::new(opt.validate().unwrap()).unwrap(); let mut db = Mace::new(opt.validate().unwrap()).unwrap();
db.disable_gc(); db.disable_gc();
@ -180,7 +184,7 @@ fn main() {
let test_start = start_time.lock().unwrap(); let test_start = start_time.lock().unwrap();
let duration = test_start.elapsed(); let duration = test_start.elapsed();
let total = total_ops.load(std::sync::atomic::Ordering::Relaxed); let total = total_ops.load(std::sync::atomic::Ordering::Relaxed);
let ops = total as f64 / duration.as_secs_f64(); let ops = (total as f64 / duration.as_secs_f64()) as usize;
// println!("{:<20} {}", "Test Mode:", args.mode); // println!("{:<20} {}", "Test Mode:", args.mode);
// println!("{:<20} {}", "Threads:", args.threads); // println!("{:<20} {}", "Threads:", args.threads);