Compare commits

..

No commits in common. "prefix-scan" and "master" have entirely different histories.

4 changed files with 34 additions and 107 deletions

View File

@ -4,14 +4,12 @@
#include <cstdio>
#include <fmt/base.h>
#include <fmt/format.h>
#include <iostream>
#include <memory>
#include <random>
#include <rocksdb/cache.h>
#include <rocksdb/db.h>
#include <rocksdb/env.h>
#include <rocksdb/options.h>
#include <rocksdb/slice.h>
#include <rocksdb/table.h>
#include <rocksdb/utilities/optimistic_transaction_db.h>
#include <rocksdb/utilities/transaction.h>
@ -21,23 +19,16 @@
#include <filesystem>
#include <format>
#include <string>
#include <syncstream>
#include "CLI/CLI.hpp"
#include "instant.h"
template<class T>
static void black_box(const T &t) {
asm volatile("" ::"m"(t) : "memory");
}
struct Args {
size_t threads;
size_t iterations;
size_t key_size;
size_t value_size;
size_t insert_ratio;
size_t blob_size;
bool random;
std::string mode;
std::string path;
@ -59,7 +50,6 @@ int main(int argc, char *argv[]) {
app.add_option("-t,--threads", args.threads, "Threads");
app.add_option("-k,--key-size", args.key_size, "Key Size");
app.add_option("-v,--value-size", args.value_size, "Value Size");
app.add_option("-b,--blob-size", args.value_size, "Blob Size");
app.add_option("-i,--iterations", args.iterations, "Iterations");
app.add_option("-r,--insert-ratio", args.insert_ratio, "Insert Ratio for mixed mode");
app.add_option("-p,--path", args.path, "DataBase Home");
@ -92,25 +82,9 @@ int main(int argc, char *argv[]) {
return 1;
}
auto find_upper_bound = [](std::string prefix) {
std::string upper_bound_key = prefix;
for (int i = upper_bound_key.length() - 1; i >= 0; --i) {
if ((unsigned char) upper_bound_key[i] != 0xff) {
upper_bound_key[i] = (unsigned char) upper_bound_key[i] + 1;
upper_bound_key.resize(i + 1);
break;
}
if (i == 0) {
upper_bound_key = "";
break;
}
}
return upper_bound_key;
};
rocksdb::ColumnFamilyOptions cfo{};
cfo.enable_blob_files = true;
cfo.min_blob_size = args.blob_size;
cfo.min_blob_size = 8192;
// use 1GB block cache
auto cache = rocksdb::NewLRUCache(1 << 30);
rocksdb::BlockBasedTableOptions table_options{};
@ -157,7 +131,7 @@ int main(int argc, char *argv[]) {
tmp.resize(args.key_size, 'x');
key.emplace_back(std::move(tmp));
}
if (args.mode == "get" || args.random || args.mode == "scan") {
if (args.mode == "get" || args.random) {
std::shuffle(keys.begin(), keys.end(), gen);
}
keys.emplace_back(std::move(key));
@ -176,32 +150,21 @@ int main(int argc, char *argv[]) {
}
kv->Commit();
delete kv;
if (args.mode == "get") {
delete handle;
delete db;
handles.clear();
// re-open db
s = rocksdb::OptimisticTransactionDB::Open(options, args.path, cfd, &handles, &db);
assert(s.ok());
}
delete handle;
delete db;
handles.clear();
// re-open db
s = rocksdb::OptimisticTransactionDB::Open(options, args.path, cfd, &handles, &db);
assert(s.ok());
}
handle = handles[0];
auto *snapshot = db->GetSnapshot();
for (size_t tid = 0; tid < args.threads; ++tid) {
auto *tk = &keys[tid];
wg.emplace_back([&, tid] {
std::string rval(args.value_size, '0');
auto prefix = std::format("key_{}", tid);
auto ropt = rocksdb::ReadOptions();
auto upper_bound = find_upper_bound(prefix);
auto upper_bound_slice = rocksdb::Slice(upper_bound);
if (!upper_bound.empty()) {
ropt.iterate_upper_bound = &upper_bound_slice;
}
auto *tk = &keys[tid];
ropt.prefix_same_as_start = true;
ropt.snapshot = snapshot;
size_t round = 0;
barrier.arrive_and_wait();
if (mtx.try_lock()) {
@ -211,7 +174,6 @@ int main(int argc, char *argv[]) {
if (args.mode == "insert") {
for (auto &key: *tk) {
round += 1;
auto *kv = db->BeginTransaction(wopt);
kv->Put(handle, key, val);
kv->Commit();
@ -220,7 +182,6 @@ int main(int argc, char *argv[]) {
} else if (args.mode == "get") {
for (auto &key: *tk) {
round += 1;
auto *kv = db->BeginTransaction(wopt);
kv->Get(ropt, handle, key, &rval);
kv->Commit();
@ -228,7 +189,6 @@ int main(int argc, char *argv[]) {
}
} else if (args.mode == "mixed") {
for (auto &key: *tk) {
round += 1;
auto is_insert = dist(gen) < args.insert_ratio;
auto *kv = db->BeginTransaction(wopt);
if (is_insert) {
@ -242,19 +202,12 @@ int main(int argc, char *argv[]) {
} else if (args.mode == "scan") {
auto *iter = db->NewIterator(ropt);
iter->Seek(prefix);
size_t n = 0;
while (iter->Valid()) {
round += 1;
auto k = iter->key();
auto v = iter->value();
black_box(k);
black_box(v);
iter->Next();
n += 1;
}
delete iter;
}
total_op.fetch_add(round, std::memory_order::relaxed);
total_op.fetch_add(args.iterations, std::memory_order::relaxed);
});
}
@ -269,7 +222,6 @@ int main(int argc, char *argv[]) {
uint64_t ops = total_op.load(std::memory_order_relaxed) / b.elapse_sec();
fmt::println("{},{},{},{},{},{},{}", args.mode, args.threads, args.key_size, args.value_size, ratio, (uint64_t) ops,
(uint64_t) b.elapse_ms());
db->ReleaseSnapshot(snapshot);
delete handle;
delete db;
std::filesystem::remove_all(args.path);

View File

@ -1,11 +1,5 @@
#!/usr/bin/env bash
if [ "$#" -ne 1 ]
then
printf "\033[m$0 path\033[0m\n"
exit 1
fi
pushd .
cd ..
cargo build --release 1>/dev/null 2> /dev/null
@ -14,32 +8,31 @@ function samples() {
export RUST_BACKTRACE=full
kv_sz=(16 16 100 1024 1024 1024 16 10240)
# set -x
db_root=$1
cnt=10000
for ((i = 1; i <= $(nproc); i *= 2))
do
for ((j = 0; j < ${#kv_sz[@]}; j += 2))
do
./target/release/kv_bench --path $db_root --threads $i --iterations $cnt --mode insert --key-size ${kv_sz[j]} --value-size ${kv_sz[j+1]}
./target/release/kv_bench --path /home/abby/mace_bench --threads $i --iterations $cnt --mode insert --key-size ${kv_sz[j]} --value-size ${kv_sz[j+1]}
if test $? -ne 0
then
echo "insert threads $i ksz ${kv_sz[j]} vsz ${kv_sz[j+1]} fail"
exit 1
fi
./target/release/kv_bench --path $db_root --threads $i --iterations $cnt --mode get --key-size ${kv_sz[j]} --value-size ${kv_sz[j+1]}
./target/release/kv_bench --path /home/abby/mace_bench --threads $i --iterations $cnt --mode get --key-size ${kv_sz[j]} --value-size ${kv_sz[j+1]}
if test $? -ne 0
then
echo "get threads $i ksz ${kv_sz[j]} vsz ${kv_sz[j+1]} fail"
exit 1
fi
./target/release/kv_bench --path $db_root --threads $i --iterations $cnt --mode mixed --key-size ${kv_sz[j]} --value-size ${kv_sz[j+1]} --insert-ratio 30
./target/release/kv_bench --path /home/abby/mace_bench --threads $i --iterations $cnt --mode mixed --key-size ${kv_sz[j]} --value-size ${kv_sz[j+1]} --insert-ratio 30
if test $? -ne 0
then
echo "mixed threads $i ksz ${kv_sz[j]} vsz ${kv_sz[j+1]} fail"
exit 1
fi
./target/release/kv_bench --path $db_root --threads $i --iterations $cnt --mode scan --key-size ${kv_sz[j]} --value-size ${kv_sz[j+1]} --insert-ratio 30
./target/release/kv_bench --path /home/abby/mace_bench --threads $i --iterations $cnt --mode scan --key-size ${kv_sz[j]} --value-size ${kv_sz[j+1]} --insert-ratio 30
if test $? -ne 0
then
echo "mixed threads $i ksz ${kv_sz[j]} vsz ${kv_sz[j+1]} fail"
@ -50,6 +43,6 @@ function samples() {
}
echo mode,threads,key_size,value_size,insert_ratio,ops,elasped > scripts/mace.csv
samples $1 2>> scripts/mace.csv
samples 2>> scripts/mace.csv
popd
./bin/python plot.py mace.csv

View File

@ -1,11 +1,5 @@
#!/usr/bin/env bash
if [ "$#" -ne 1 ]
then
printf "\033[m$0 path\033[0m\n"
exit 1
fi
pushd .
cd ../rocksdb
cmake --preset release 1>/dev/null 2>/dev/null
@ -14,31 +8,30 @@ cmake --build --preset release 1>/dev/null 2>/dev/null
function samples() {
kv_sz=(16 16 100 1024 1024 1024 16 10240)
# set -x
db_root=$1
cnt=10000
for ((i = 1; i <= $(nproc); i *= 2))
do
for ((j = 0; j < ${#kv_sz[@]}; j += 2))
do
./build/release/rocksdb_bench --path $db_root --threads $i --iterations $cnt --mode insert --key-size ${kv_sz[j]} --value-size ${kv_sz[j+1]}
./build/release/rocksdb_bench --path /home/abby/rocksdb_tmp --threads $i --iterations $cnt --mode insert --key-size ${kv_sz[j]} --value-size ${kv_sz[j+1]}
if test $? -ne 0
then
echo "insert threads $i ksz ${kv_sz[j]} vsz ${kv_sz[j+1]} fail"
exit 1
fi
./build/release/rocksdb_bench --path $db_root --threads $i --iterations $cnt --mode get --key-size ${kv_sz[j]} --value-size ${kv_sz[j+1]}
./build/release/rocksdb_bench --path /home/abby/rocksdb_tmp --threads $i --iterations $cnt --mode get --key-size ${kv_sz[j]} --value-size ${kv_sz[j+1]}
if test $? -ne 0
then
echo "get threads $i ksz ${kv_sz[j]} vsz ${kv_sz[j+1]} fail"
exit 1
fi
./build/release/rocksdb_bench --path $db_root --threads $i --iterations $cnt --mode mixed --key-size ${kv_sz[j]} --value-size ${kv_sz[j+1]} --insert-ratio 30
./build/release/rocksdb_bench --path /home/abby/rocksdb_tmp --threads $i --iterations $cnt --mode mixed --key-size ${kv_sz[j]} --value-size ${kv_sz[j+1]} --insert-ratio 30
if test $? -ne 0
then
echo "mixed threads $i ksz ${kv_sz[j]} vsz ${kv_sz[j+1]} fail"
exit 1
fi
./build/release/rocksdb_bench --path $db_root --threads $i --iterations $cnt --mode scan --key-size ${kv_sz[j]} --value-size ${kv_sz[j+1]} --insert-ratio 30
./build/release/rocksdb_bench --path /home/abby/rocksdb_tmp --threads $i --iterations $cnt --mode scan --key-size ${kv_sz[j]} --value-size ${kv_sz[j+1]} --insert-ratio 30
if test $? -ne 0
then
echo "mixed threads $i ksz ${kv_sz[j]} vsz ${kv_sz[j+1]} fail"
@ -49,6 +42,6 @@ function samples() {
}
echo mode,threads,key_size,value_size,insert_ratio,ops,elapsed > ../scripts/rocksdb.csv
samples $1 1>> ../scripts/rocksdb.csv
samples 1>> ../scripts/rocksdb.csv
popd
./bin/python plot.py rocksdb.csv

View File

@ -81,7 +81,7 @@ fn main() {
opt.sync_on_write = false;
opt.over_provision = true; // large value will use lots of memeory
opt.inline_size = args.blob_size;
opt.tmp_store = args.mode != "get";
opt.tmp_store = args.mode != "get" && args.mode != "scan";
let mut saved = opt.clone();
saved.tmp_store = false;
let mut db = Mace::new(opt.validate().unwrap()).unwrap();
@ -96,7 +96,7 @@ fn main() {
key.resize(args.key_size, b'x');
tk.push(key);
}
if args.random || args.mode == "get" || args.mode == "scan" {
if args.random || args.mode == "get" {
tk.shuffle(&mut rng);
}
keys.push(tk);
@ -105,17 +105,15 @@ fn main() {
if args.mode == "get" || args.mode == "scan" {
let pre_tx = db.begin().unwrap();
(0..args.threads).for_each(|tid| {
for k in &keys[tid] {
pre_tx.put(k, &*value).unwrap();
for i in 0..args.iterations {
pre_tx.put(&keys[tid][i], &*value).unwrap();
}
});
pre_tx.commit().unwrap();
if args.mode == "get" {
drop(db);
// re-open db
saved.tmp_store = true;
db = Mace::new(saved.validate().unwrap()).unwrap();
}
drop(db);
// re-open db
saved.tmp_store = true;
db = Mace::new(saved.validate().unwrap()).unwrap();
}
let barrier = Arc::new(std::sync::Barrier::new(args.threads));
@ -136,7 +134,6 @@ fn main() {
std::thread::spawn(move || {
// coreid::bind_core(tid);
let mut round = 0;
barrier.wait();
{
@ -144,10 +141,10 @@ fn main() {
*guard = Instant::now();
}
}
match mode.as_str() {
"insert" => {
for key in tk {
round += 1;
let tx = db.begin().unwrap();
tx.put(key.as_slice(), val.as_slice()).unwrap();
tx.commit().unwrap();
@ -155,16 +152,13 @@ fn main() {
}
"get" => {
for key in tk {
round += 1;
let tx = db.view().unwrap();
let x = tx.get(key).unwrap();
std::hint::black_box(x);
tx.get(key).unwrap();
}
}
"mixed" => {
for key in tk {
let is_insert = rand::random_range(0..100) < insert_ratio;
round += 1;
if is_insert {
let tx = db.begin().unwrap();
@ -172,8 +166,7 @@ fn main() {
tx.commit().unwrap();
} else {
let tx = db.view().unwrap();
let x = tx.get(key); // not found
let _ = std::hint::black_box(x);
let _ = tx.get(key); // not found
}
}
}
@ -181,17 +174,13 @@ fn main() {
let view = db.view().unwrap();
let iter = view.seek(prefix);
for x in iter {
round += 1;
let k = x.key();
let v = x.val();
std::hint::black_box(k);
std::hint::black_box(v);
std::hint::black_box(x);
}
}
_ => panic!("Invalid mode"),
}
total_ops.fetch_add(round, std::sync::atomic::Ordering::Relaxed);
total_ops.fetch_add(args.iterations, std::sync::atomic::Ordering::Relaxed);
})
})
.collect();
@ -224,7 +213,7 @@ fn main() {
};
// eprintln!("mode,threads,key_size,value_size,insert_ratio,ops");
eprintln!(
"{},{},{},{},{},{},{}",
"{},{},{},{},{},{:.2},{}",
args.mode,
args.threads,
args.key_size,