38 changed files with 1887 additions and 886 deletions
@ -1,32 +1,47 @@
|
||||
use crate::{utils, ConduitResult}; |
||||
use ruma::api::client::r0::filter::{self, create_filter, get_filter}; |
||||
use crate::{database::DatabaseGuard, ConduitResult, Error, Ruma}; |
||||
use ruma::api::client::{ |
||||
error::ErrorKind, |
||||
r0::filter::{create_filter, get_filter}, |
||||
}; |
||||
|
||||
#[cfg(feature = "conduit_bin")] |
||||
use rocket::{get, post}; |
||||
|
||||
/// # `GET /_matrix/client/r0/user/{userId}/filter/{filterId}`
|
||||
///
|
||||
/// TODO: Loads a filter that was previously created.
|
||||
#[cfg_attr(feature = "conduit_bin", get("/_matrix/client/r0/user/<_>/filter/<_>"))] |
||||
#[tracing::instrument] |
||||
pub async fn get_filter_route() -> ConduitResult<get_filter::Response> { |
||||
// TODO
|
||||
Ok(get_filter::Response::new(filter::IncomingFilterDefinition { |
||||
event_fields: None, |
||||
event_format: filter::EventFormat::default(), |
||||
account_data: filter::IncomingFilter::default(), |
||||
room: filter::IncomingRoomFilter::default(), |
||||
presence: filter::IncomingFilter::default(), |
||||
}) |
||||
.into()) |
||||
/// Loads a filter that was previously created.
|
||||
///
|
||||
/// - A user can only access their own filters
|
||||
#[cfg_attr(
|
||||
feature = "conduit_bin", |
||||
get("/_matrix/client/r0/user/<_>/filter/<_>", data = "<body>") |
||||
)] |
||||
#[tracing::instrument(skip(db, body))] |
||||
pub async fn get_filter_route( |
||||
db: DatabaseGuard, |
||||
body: Ruma<get_filter::Request<'_>>, |
||||
) -> ConduitResult<get_filter::Response> { |
||||
let sender_user = body.sender_user.as_ref().expect("user is authenticated"); |
||||
let filter = match db.users.get_filter(sender_user, &body.filter_id)? { |
||||
Some(filter) => filter, |
||||
None => return Err(Error::BadRequest(ErrorKind::NotFound, "Filter not found.")), |
||||
}; |
||||
|
||||
Ok(get_filter::Response::new(filter).into()) |
||||
} |
||||
|
||||
/// # `PUT /_matrix/client/r0/user/{userId}/filter`
|
||||
///
|
||||
/// TODO: Creates a new filter to be used by other endpoints.
|
||||
#[cfg_attr(feature = "conduit_bin", post("/_matrix/client/r0/user/<_>/filter"))] |
||||
#[tracing::instrument] |
||||
pub async fn create_filter_route() -> ConduitResult<create_filter::Response> { |
||||
// TODO
|
||||
Ok(create_filter::Response::new(utils::random_string(10)).into()) |
||||
/// Creates a new filter to be used by other endpoints.
|
||||
#[cfg_attr(
|
||||
feature = "conduit_bin", |
||||
post("/_matrix/client/r0/user/<_>/filter", data = "<body>") |
||||
)] |
||||
#[tracing::instrument(skip(db, body))] |
||||
pub async fn create_filter_route( |
||||
db: DatabaseGuard, |
||||
body: Ruma<create_filter::Request<'_>>, |
||||
) -> ConduitResult<create_filter::Response> { |
||||
let sender_user = body.sender_user.as_ref().expect("user is authenticated"); |
||||
Ok(create_filter::Response::new(db.users.create_filter(sender_user, &body.filter)?).into()) |
||||
} |
||||
|
||||
@ -0,0 +1,231 @@
|
||||
use super::{super::Config, watchers::Watchers, DatabaseEngine, Tree}; |
||||
use crate::{utils, Result}; |
||||
use std::{future::Future, pin::Pin, sync::Arc, sync::RwLock}; |
||||
|
||||
pub struct Engine { |
||||
rocks: rocksdb::DBWithThreadMode<rocksdb::MultiThreaded>, |
||||
max_open_files: i32, |
||||
cache: rocksdb::Cache, |
||||
old_cfs: Vec<String>, |
||||
} |
||||
|
||||
pub struct RocksDbEngineTree<'a> { |
||||
db: Arc<Engine>, |
||||
name: &'a str, |
||||
watchers: Watchers, |
||||
write_lock: RwLock<()>, |
||||
} |
||||
|
||||
fn db_options(max_open_files: i32, rocksdb_cache: &rocksdb::Cache) -> rocksdb::Options { |
||||
let mut block_based_options = rocksdb::BlockBasedOptions::default(); |
||||
block_based_options.set_block_cache(rocksdb_cache); |
||||
|
||||
// "Difference of spinning disk"
|
||||
// https://zhangyuchi.gitbooks.io/rocksdbbook/content/RocksDB-Tuning-Guide.html
|
||||
block_based_options.set_block_size(4 * 1024); |
||||
block_based_options.set_cache_index_and_filter_blocks(true); |
||||
|
||||
let mut db_opts = rocksdb::Options::default(); |
||||
db_opts.set_block_based_table_factory(&block_based_options); |
||||
db_opts.set_optimize_filters_for_hits(true); |
||||
db_opts.set_skip_stats_update_on_db_open(true); |
||||
db_opts.set_level_compaction_dynamic_level_bytes(true); |
||||
db_opts.set_target_file_size_base(256 * 1024 * 1024); |
||||
//db_opts.set_compaction_readahead_size(2 * 1024 * 1024);
|
||||
//db_opts.set_use_direct_reads(true);
|
||||
//db_opts.set_use_direct_io_for_flush_and_compaction(true);
|
||||
db_opts.create_if_missing(true); |
||||
db_opts.increase_parallelism(num_cpus::get() as i32); |
||||
db_opts.set_max_open_files(max_open_files); |
||||
db_opts.set_compression_type(rocksdb::DBCompressionType::Zstd); |
||||
db_opts.set_compaction_style(rocksdb::DBCompactionStyle::Level); |
||||
db_opts.optimize_level_style_compaction(10 * 1024 * 1024); |
||||
|
||||
let prefix_extractor = rocksdb::SliceTransform::create_fixed_prefix(1); |
||||
db_opts.set_prefix_extractor(prefix_extractor); |
||||
|
||||
db_opts |
||||
} |
||||
|
||||
impl DatabaseEngine for Arc<Engine> { |
||||
fn open(config: &Config) -> Result<Self> { |
||||
let cache_capacity_bytes = (config.db_cache_capacity_mb * 1024.0 * 1024.0) as usize; |
||||
let rocksdb_cache = rocksdb::Cache::new_lru_cache(cache_capacity_bytes).unwrap(); |
||||
|
||||
let db_opts = db_options(config.rocksdb_max_open_files, &rocksdb_cache); |
||||
|
||||
let cfs = rocksdb::DBWithThreadMode::<rocksdb::MultiThreaded>::list_cf( |
||||
&db_opts, |
||||
&config.database_path, |
||||
) |
||||
.unwrap_or_default(); |
||||
|
||||
let db = rocksdb::DBWithThreadMode::<rocksdb::MultiThreaded>::open_cf_descriptors( |
||||
&db_opts, |
||||
&config.database_path, |
||||
cfs.iter().map(|name| { |
||||
rocksdb::ColumnFamilyDescriptor::new( |
||||
name, |
||||
db_options(config.rocksdb_max_open_files, &rocksdb_cache), |
||||
) |
||||
}), |
||||
)?; |
||||
|
||||
Ok(Arc::new(Engine { |
||||
rocks: db, |
||||
max_open_files: config.rocksdb_max_open_files, |
||||
cache: rocksdb_cache, |
||||
old_cfs: cfs, |
||||
})) |
||||
} |
||||
|
||||
fn open_tree(&self, name: &'static str) -> Result<Arc<dyn Tree>> { |
||||
if !self.old_cfs.contains(&name.to_owned()) { |
||||
// Create if it didn't exist
|
||||
let _ = self |
||||
.rocks |
||||
.create_cf(name, &db_options(self.max_open_files, &self.cache)); |
||||
} |
||||
|
||||
Ok(Arc::new(RocksDbEngineTree { |
||||
name, |
||||
db: Arc::clone(self), |
||||
watchers: Watchers::default(), |
||||
write_lock: RwLock::new(()), |
||||
})) |
||||
} |
||||
|
||||
fn flush(&self) -> Result<()> { |
||||
// TODO?
|
||||
Ok(()) |
||||
} |
||||
|
||||
fn memory_usage(&self) -> Result<String> { |
||||
let stats = |
||||
rocksdb::perf::get_memory_usage_stats(Some(&[&self.rocks]), Some(&[&self.cache]))?; |
||||
Ok(format!( |
||||
"Approximate memory usage of all the mem-tables: {:.3} MB\n\ |
||||
Approximate memory usage of un-flushed mem-tables: {:.3} MB\n\ |
||||
Approximate memory usage of all the table readers: {:.3} MB\n\ |
||||
Approximate memory usage by cache: {:.3} MB\n\ |
||||
Approximate memory usage by cache pinned: {:.3} MB\n\ |
||||
", |
||||
stats.mem_table_total as f64 / 1024.0 / 1024.0, |
||||
stats.mem_table_unflushed as f64 / 1024.0 / 1024.0, |
||||
stats.mem_table_readers_total as f64 / 1024.0 / 1024.0, |
||||
stats.cache_total as f64 / 1024.0 / 1024.0, |
||||
self.cache.get_pinned_usage() as f64 / 1024.0 / 1024.0, |
||||
)) |
||||
} |
||||
} |
||||
|
||||
impl RocksDbEngineTree<'_> { |
||||
fn cf(&self) -> Arc<rocksdb::BoundColumnFamily<'_>> { |
||||
self.db.rocks.cf_handle(self.name).unwrap() |
||||
} |
||||
} |
||||
|
||||
impl Tree for RocksDbEngineTree<'_> { |
||||
fn get(&self, key: &[u8]) -> Result<Option<Vec<u8>>> { |
||||
Ok(self.db.rocks.get_cf(&self.cf(), key)?) |
||||
} |
||||
|
||||
fn insert(&self, key: &[u8], value: &[u8]) -> Result<()> { |
||||
let lock = self.write_lock.read().unwrap(); |
||||
self.db.rocks.put_cf(&self.cf(), key, value)?; |
||||
drop(lock); |
||||
|
||||
self.watchers.wake(key); |
||||
|
||||
Ok(()) |
||||
} |
||||
|
||||
fn insert_batch<'a>(&self, iter: &mut dyn Iterator<Item = (Vec<u8>, Vec<u8>)>) -> Result<()> { |
||||
for (key, value) in iter { |
||||
self.db.rocks.put_cf(&self.cf(), key, value)?; |
||||
} |
||||
|
||||
Ok(()) |
||||
} |
||||
|
||||
fn remove(&self, key: &[u8]) -> Result<()> { |
||||
Ok(self.db.rocks.delete_cf(&self.cf(), key)?) |
||||
} |
||||
|
||||
fn iter<'a>(&'a self) -> Box<dyn Iterator<Item = (Vec<u8>, Vec<u8>)> + 'a> { |
||||
Box::new( |
||||
self.db |
||||
.rocks |
||||
.iterator_cf(&self.cf(), rocksdb::IteratorMode::Start) |
||||
.map(|(k, v)| (Vec::from(k), Vec::from(v))), |
||||
) |
||||
} |
||||
|
||||
fn iter_from<'a>( |
||||
&'a self, |
||||
from: &[u8], |
||||
backwards: bool, |
||||
) -> Box<dyn Iterator<Item = (Vec<u8>, Vec<u8>)> + 'a> { |
||||
Box::new( |
||||
self.db |
||||
.rocks |
||||
.iterator_cf( |
||||
&self.cf(), |
||||
rocksdb::IteratorMode::From( |
||||
from, |
||||
if backwards { |
||||
rocksdb::Direction::Reverse |
||||
} else { |
||||
rocksdb::Direction::Forward |
||||
}, |
||||
), |
||||
) |
||||
.map(|(k, v)| (Vec::from(k), Vec::from(v))), |
||||
) |
||||
} |
||||
|
||||
fn increment(&self, key: &[u8]) -> Result<Vec<u8>> { |
||||
let lock = self.write_lock.write().unwrap(); |
||||
|
||||
let old = self.db.rocks.get_cf(&self.cf(), &key)?; |
||||
let new = utils::increment(old.as_deref()).unwrap(); |
||||
self.db.rocks.put_cf(&self.cf(), key, &new)?; |
||||
|
||||
drop(lock); |
||||
Ok(new) |
||||
} |
||||
|
||||
fn increment_batch<'a>(&self, iter: &mut dyn Iterator<Item = Vec<u8>>) -> Result<()> { |
||||
let lock = self.write_lock.write().unwrap(); |
||||
|
||||
for key in iter { |
||||
let old = self.db.rocks.get_cf(&self.cf(), &key)?; |
||||
let new = utils::increment(old.as_deref()).unwrap(); |
||||
self.db.rocks.put_cf(&self.cf(), key, new)?; |
||||
} |
||||
|
||||
drop(lock); |
||||
|
||||
Ok(()) |
||||
} |
||||
|
||||
fn scan_prefix<'a>( |
||||
&'a self, |
||||
prefix: Vec<u8>, |
||||
) -> Box<dyn Iterator<Item = (Vec<u8>, Vec<u8>)> + 'a> { |
||||
Box::new( |
||||
self.db |
||||
.rocks |
||||
.iterator_cf( |
||||
&self.cf(), |
||||
rocksdb::IteratorMode::From(&prefix, rocksdb::Direction::Forward), |
||||
) |
||||
.map(|(k, v)| (Vec::from(k), Vec::from(v))) |
||||
.take_while(move |(k, _)| k.starts_with(&prefix)), |
||||
) |
||||
} |
||||
|
||||
fn watch_prefix<'a>(&'a self, prefix: &[u8]) -> Pin<Box<dyn Future<Output = ()> + Send + 'a>> { |
||||
self.watchers.watch(prefix) |
||||
} |
||||
} |
||||
@ -0,0 +1,54 @@
|
||||
use std::{ |
||||
collections::{hash_map, HashMap}, |
||||
future::Future, |
||||
pin::Pin, |
||||
sync::RwLock, |
||||
}; |
||||
use tokio::sync::watch; |
||||
|
||||
#[derive(Default)] |
||||
pub(super) struct Watchers { |
||||
watchers: RwLock<HashMap<Vec<u8>, (watch::Sender<()>, watch::Receiver<()>)>>, |
||||
} |
||||
|
||||
impl Watchers { |
||||
pub(super) fn watch<'a>( |
||||
&'a self, |
||||
prefix: &[u8], |
||||
) -> Pin<Box<dyn Future<Output = ()> + Send + 'a>> { |
||||
let mut rx = match self.watchers.write().unwrap().entry(prefix.to_vec()) { |
||||
hash_map::Entry::Occupied(o) => o.get().1.clone(), |
||||
hash_map::Entry::Vacant(v) => { |
||||
let (tx, rx) = tokio::sync::watch::channel(()); |
||||
v.insert((tx, rx.clone())); |
||||
rx |
||||
} |
||||
}; |
||||
|
||||
Box::pin(async move { |
||||
// Tx is never destroyed
|
||||
rx.changed().await.unwrap(); |
||||
}) |
||||
} |
||||
pub(super) fn wake(&self, key: &[u8]) { |
||||
let watchers = self.watchers.read().unwrap(); |
||||
let mut triggered = Vec::new(); |
||||
|
||||
for length in 0..=key.len() { |
||||
if watchers.contains_key(&key[..length]) { |
||||
triggered.push(&key[..length]); |
||||
} |
||||
} |
||||
|
||||
drop(watchers); |
||||
|
||||
if !triggered.is_empty() { |
||||
let mut watchers = self.watchers.write().unwrap(); |
||||
for prefix in triggered { |
||||
if let Some(tx) = watchers.remove(prefix) { |
||||
let _ = tx.0.send(()); |
||||
} |
||||
} |
||||
}; |
||||
} |
||||
} |
||||
Loading…
Reference in new issue