mirror of https://gitlab.com/famedly/conduit.git
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
246 lines
7.9 KiB
246 lines
7.9 KiB
use super::{super::Config, watchers::Watchers, DatabaseEngine, Tree}; |
|
use crate::{utils, Result}; |
|
use std::{future::Future, pin::Pin, sync::Arc, sync::RwLock}; |
|
|
|
pub struct Engine { |
|
rocks: rocksdb::DBWithThreadMode<rocksdb::MultiThreaded>, |
|
cache_capacity_bytes: usize, |
|
max_open_files: i32, |
|
cache: rocksdb::Cache, |
|
old_cfs: Vec<String>, |
|
} |
|
|
|
pub struct RocksDbEngineTree<'a> { |
|
db: Arc<Engine>, |
|
name: &'a str, |
|
watchers: Watchers, |
|
write_lock: RwLock<()>, |
|
} |
|
|
|
fn db_options( |
|
cache_capacity_bytes: usize, |
|
max_open_files: i32, |
|
rocksdb_cache: &rocksdb::Cache, |
|
) -> rocksdb::Options { |
|
let mut block_based_options = rocksdb::BlockBasedOptions::default(); |
|
block_based_options.set_block_cache(rocksdb_cache); |
|
|
|
// "Difference of spinning disk" |
|
// https://zhangyuchi.gitbooks.io/rocksdbbook/content/RocksDB-Tuning-Guide.html |
|
block_based_options.set_block_size(4 * 1024); |
|
block_based_options.set_cache_index_and_filter_blocks(true); |
|
|
|
let mut db_opts = rocksdb::Options::default(); |
|
db_opts.set_block_based_table_factory(&block_based_options); |
|
db_opts.set_optimize_filters_for_hits(true); |
|
db_opts.set_skip_stats_update_on_db_open(true); |
|
db_opts.set_level_compaction_dynamic_level_bytes(true); |
|
db_opts.set_target_file_size_base(256 * 1024 * 1024); |
|
//db_opts.set_compaction_readahead_size(2 * 1024 * 1024); |
|
//db_opts.set_use_direct_reads(true); |
|
//db_opts.set_use_direct_io_for_flush_and_compaction(true); |
|
db_opts.create_if_missing(true); |
|
db_opts.increase_parallelism(num_cpus::get() as i32); |
|
db_opts.set_max_open_files(max_open_files); |
|
db_opts.set_compression_type(rocksdb::DBCompressionType::Zstd); |
|
db_opts.set_compaction_style(rocksdb::DBCompactionStyle::Level); |
|
db_opts.optimize_level_style_compaction(10 * 1024 * 1024); |
|
|
|
let prefix_extractor = rocksdb::SliceTransform::create_fixed_prefix(1); |
|
db_opts.set_prefix_extractor(prefix_extractor); |
|
|
|
db_opts |
|
} |
|
|
|
impl DatabaseEngine for Arc<Engine> { |
|
fn open(config: &Config) -> Result<Self> { |
|
let cache_capacity_bytes = (config.db_cache_capacity_mb * 1024.0 * 1024.0) as usize; |
|
let rocksdb_cache = rocksdb::Cache::new_lru_cache(cache_capacity_bytes).unwrap(); |
|
|
|
let db_opts = db_options( |
|
cache_capacity_bytes, |
|
config.rocksdb_max_open_files, |
|
&rocksdb_cache, |
|
); |
|
|
|
let cfs = rocksdb::DBWithThreadMode::<rocksdb::MultiThreaded>::list_cf( |
|
&db_opts, |
|
&config.database_path, |
|
) |
|
.unwrap_or_default(); |
|
|
|
let db = rocksdb::DBWithThreadMode::<rocksdb::MultiThreaded>::open_cf_descriptors( |
|
&db_opts, |
|
&config.database_path, |
|
cfs.iter().map(|name| { |
|
rocksdb::ColumnFamilyDescriptor::new( |
|
name, |
|
db_options( |
|
cache_capacity_bytes, |
|
config.rocksdb_max_open_files, |
|
&rocksdb_cache, |
|
), |
|
) |
|
}), |
|
)?; |
|
|
|
Ok(Arc::new(Engine { |
|
rocks: db, |
|
cache_capacity_bytes, |
|
max_open_files: config.rocksdb_max_open_files, |
|
cache: rocksdb_cache, |
|
old_cfs: cfs, |
|
})) |
|
} |
|
|
|
fn open_tree(&self, name: &'static str) -> Result<Arc<dyn Tree>> { |
|
if !self.old_cfs.contains(&name.to_owned()) { |
|
// Create if it didn't exist |
|
let _ = self.rocks.create_cf( |
|
name, |
|
&db_options(self.cache_capacity_bytes, self.max_open_files, &self.cache), |
|
); |
|
} |
|
|
|
Ok(Arc::new(RocksDbEngineTree { |
|
name, |
|
db: Arc::clone(self), |
|
watchers: Watchers::default(), |
|
write_lock: RwLock::new(()), |
|
})) |
|
} |
|
|
|
fn flush(&self) -> Result<()> { |
|
// TODO? |
|
Ok(()) |
|
} |
|
|
|
fn memory_usage(&self) -> Result<String> { |
|
let stats = |
|
rocksdb::perf::get_memory_usage_stats(Some(&[&self.rocks]), Some(&[&self.cache]))?; |
|
Ok(format!( |
|
"Approximate memory usage of all the mem-tables: {:.3} MB\n\ |
|
Approximate memory usage of un-flushed mem-tables: {:.3} MB\n\ |
|
Approximate memory usage of all the table readers: {:.3} MB\n\ |
|
Approximate memory usage by cache: {:.3} MB\n\ |
|
Approximate memory usage by cache pinned: {:.3} MB\n\ |
|
", |
|
stats.mem_table_total as f64 / 1024.0 / 1024.0, |
|
stats.mem_table_unflushed as f64 / 1024.0 / 1024.0, |
|
stats.mem_table_readers_total as f64 / 1024.0 / 1024.0, |
|
stats.cache_total as f64 / 1024.0 / 1024.0, |
|
self.cache.get_pinned_usage() as f64 / 1024.0 / 1024.0, |
|
)) |
|
} |
|
} |
|
|
|
impl RocksDbEngineTree<'_> { |
|
fn cf(&self) -> Arc<rocksdb::BoundColumnFamily<'_>> { |
|
self.db.rocks.cf_handle(self.name).unwrap() |
|
} |
|
} |
|
|
|
impl Tree for RocksDbEngineTree<'_> { |
|
fn get(&self, key: &[u8]) -> Result<Option<Vec<u8>>> { |
|
Ok(self.db.rocks.get_cf(&self.cf(), key)?) |
|
} |
|
|
|
fn insert(&self, key: &[u8], value: &[u8]) -> Result<()> { |
|
let lock = self.write_lock.read().unwrap(); |
|
self.db.rocks.put_cf(&self.cf(), key, value)?; |
|
drop(lock); |
|
|
|
self.watchers.wake(key); |
|
|
|
Ok(()) |
|
} |
|
|
|
fn insert_batch<'a>(&self, iter: &mut dyn Iterator<Item = (Vec<u8>, Vec<u8>)>) -> Result<()> { |
|
for (key, value) in iter { |
|
self.db.rocks.put_cf(&self.cf(), key, value)?; |
|
} |
|
|
|
Ok(()) |
|
} |
|
|
|
fn remove(&self, key: &[u8]) -> Result<()> { |
|
Ok(self.db.rocks.delete_cf(&self.cf(), key)?) |
|
} |
|
|
|
fn iter<'a>(&'a self) -> Box<dyn Iterator<Item = (Vec<u8>, Vec<u8>)> + 'a> { |
|
Box::new( |
|
self.db |
|
.rocks |
|
.iterator_cf(&self.cf(), rocksdb::IteratorMode::Start) |
|
.map(|(k, v)| (Vec::from(k), Vec::from(v))), |
|
) |
|
} |
|
|
|
fn iter_from<'a>( |
|
&'a self, |
|
from: &[u8], |
|
backwards: bool, |
|
) -> Box<dyn Iterator<Item = (Vec<u8>, Vec<u8>)> + 'a> { |
|
Box::new( |
|
self.db |
|
.rocks |
|
.iterator_cf( |
|
&self.cf(), |
|
rocksdb::IteratorMode::From( |
|
from, |
|
if backwards { |
|
rocksdb::Direction::Reverse |
|
} else { |
|
rocksdb::Direction::Forward |
|
}, |
|
), |
|
) |
|
.map(|(k, v)| (Vec::from(k), Vec::from(v))), |
|
) |
|
} |
|
|
|
fn increment(&self, key: &[u8]) -> Result<Vec<u8>> { |
|
let lock = self.write_lock.write().unwrap(); |
|
|
|
let old = self.db.rocks.get_cf(&self.cf(), &key)?; |
|
let new = utils::increment(old.as_deref()).unwrap(); |
|
self.db.rocks.put_cf(&self.cf(), key, &new)?; |
|
|
|
drop(lock); |
|
Ok(new) |
|
} |
|
|
|
fn increment_batch<'a>(&self, iter: &mut dyn Iterator<Item = Vec<u8>>) -> Result<()> { |
|
let lock = self.write_lock.write().unwrap(); |
|
|
|
for key in iter { |
|
let old = self.db.rocks.get_cf(&self.cf(), &key)?; |
|
let new = utils::increment(old.as_deref()).unwrap(); |
|
self.db.rocks.put_cf(&self.cf(), key, new)?; |
|
} |
|
|
|
drop(lock); |
|
|
|
Ok(()) |
|
} |
|
|
|
fn scan_prefix<'a>( |
|
&'a self, |
|
prefix: Vec<u8>, |
|
) -> Box<dyn Iterator<Item = (Vec<u8>, Vec<u8>)> + 'a> { |
|
Box::new( |
|
self.db |
|
.rocks |
|
.iterator_cf( |
|
&self.cf(), |
|
rocksdb::IteratorMode::From(&prefix, rocksdb::Direction::Forward), |
|
) |
|
.map(|(k, v)| (Vec::from(k), Vec::from(v))) |
|
.take_while(move |(k, _)| k.starts_with(&prefix)), |
|
) |
|
} |
|
|
|
fn watch_prefix<'a>(&'a self, prefix: &[u8]) -> Pin<Box<dyn Future<Output = ()> + Send + 'a>> { |
|
self.watchers.watch(prefix) |
|
} |
|
}
|
|
|