batch queries to maximize throughput
query-side streams for first level of callsites Signed-off-by: Jason Volk <jason@zemos.net>
This commit is contained in:
parent
1792711d09
commit
2259e2c82f
13 changed files with 191 additions and 56 deletions
|
@ -1457,6 +1457,19 @@
|
||||||
#
|
#
|
||||||
#stream_width_scale = 1.0
|
#stream_width_scale = 1.0
|
||||||
|
|
||||||
|
# Sets the initial amplification factor. This controls batch sizes of
|
||||||
|
# requests made by each pool worker, multiplying the throughput of each
|
||||||
|
# stream. This value is somewhat abstract from specific hardware
|
||||||
|
# characteristics and can be significantly larger than any thread count or
|
||||||
|
# queue size. This is because each database query may require several
|
||||||
|
# index lookups, thus many database queries in a batch may make progress
|
||||||
|
# independently while also sharing index and data blocks which may or may
|
||||||
|
# not be cached. It is worthwhile to submit huge batches to reduce
|
||||||
|
# complexity. The maximum value is 32768, though sufficient hardware is
|
||||||
|
# still advised for that.
|
||||||
|
#
|
||||||
|
#stream_amplification = 1024
|
||||||
|
|
||||||
# Number of sender task workers; determines sender parallelism. Default is
|
# Number of sender task workers; determines sender parallelism. Default is
|
||||||
# '0' which means the value is determined internally, likely matching the
|
# '0' which means the value is determined internally, likely matching the
|
||||||
# number of tokio worker-threads or number of cores, etc. Override by
|
# number of tokio worker-threads or number of cores, etc. Override by
|
||||||
|
|
|
@ -1653,6 +1653,21 @@ pub struct Config {
|
||||||
#[serde(default = "default_stream_width_scale")]
|
#[serde(default = "default_stream_width_scale")]
|
||||||
pub stream_width_scale: f32,
|
pub stream_width_scale: f32,
|
||||||
|
|
||||||
|
/// Sets the initial amplification factor. This controls batch sizes of
|
||||||
|
/// requests made by each pool worker, multiplying the throughput of each
|
||||||
|
/// stream. This value is somewhat abstract from specific hardware
|
||||||
|
/// characteristics and can be significantly larger than any thread count or
|
||||||
|
/// queue size. This is because each database query may require several
|
||||||
|
/// index lookups, thus many database queries in a batch may make progress
|
||||||
|
/// independently while also sharing index and data blocks which may or may
|
||||||
|
/// not be cached. It is worthwhile to submit huge batches to reduce
|
||||||
|
/// complexity. The maximum value is 32768, though sufficient hardware is
|
||||||
|
/// still advised for that.
|
||||||
|
///
|
||||||
|
/// default: 1024
|
||||||
|
#[serde(default = "default_stream_amplification")]
|
||||||
|
pub stream_amplification: usize,
|
||||||
|
|
||||||
/// Number of sender task workers; determines sender parallelism. Default is
|
/// Number of sender task workers; determines sender parallelism. Default is
|
||||||
/// '0' which means the value is determined internally, likely matching the
|
/// '0' which means the value is determined internally, likely matching the
|
||||||
/// number of tokio worker-threads or number of cores, etc. Override by
|
/// number of tokio worker-threads or number of cores, etc. Override by
|
||||||
|
@ -2467,3 +2482,5 @@ fn default_db_pool_queue_mult() -> usize { 4 }
|
||||||
fn default_stream_width_default() -> usize { 32 }
|
fn default_stream_width_default() -> usize { 32 }
|
||||||
|
|
||||||
fn default_stream_width_scale() -> f32 { 1.0 }
|
fn default_stream_width_scale() -> f32 { 1.0 }
|
||||||
|
|
||||||
|
fn default_stream_amplification() -> usize { 1024 }
|
||||||
|
|
|
@ -3,9 +3,15 @@ use std::sync::atomic::{AtomicUsize, Ordering};
|
||||||
/// Stream concurrency factor; this is a live value.
|
/// Stream concurrency factor; this is a live value.
|
||||||
static WIDTH: AtomicUsize = AtomicUsize::new(32);
|
static WIDTH: AtomicUsize = AtomicUsize::new(32);
|
||||||
|
|
||||||
/// Practicable limits on the stream width
|
/// Stream throughput amplifier; this is a live value.
|
||||||
|
static AMPLIFICATION: AtomicUsize = AtomicUsize::new(1024);
|
||||||
|
|
||||||
|
/// Practicable limits on the stream width.
|
||||||
pub const WIDTH_LIMIT: (usize, usize) = (1, 1024);
|
pub const WIDTH_LIMIT: (usize, usize) = (1, 1024);
|
||||||
|
|
||||||
|
/// Practicable limits on the stream amplifier.
|
||||||
|
pub const AMPLIFICATION_LIMIT: (usize, usize) = (32, 32768);
|
||||||
|
|
||||||
/// Sets the live concurrency factor. The first return value is the previous
|
/// Sets the live concurrency factor. The first return value is the previous
|
||||||
/// width which was replaced. The second return value is the value which was set
|
/// width which was replaced. The second return value is the value which was set
|
||||||
/// after any applied limits.
|
/// after any applied limits.
|
||||||
|
@ -14,6 +20,14 @@ pub fn set_width(width: usize) -> (usize, usize) {
|
||||||
(WIDTH.swap(width, Ordering::Relaxed), width)
|
(WIDTH.swap(width, Ordering::Relaxed), width)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Sets the live concurrency amplification. The first return value is the
|
||||||
|
/// previous width which was replaced. The second return value is the value
|
||||||
|
/// which was set after any applied limits.
|
||||||
|
pub fn set_amplification(width: usize) -> (usize, usize) {
|
||||||
|
let width = width.clamp(AMPLIFICATION_LIMIT.0, AMPLIFICATION_LIMIT.1);
|
||||||
|
(AMPLIFICATION.swap(width, Ordering::Relaxed), width)
|
||||||
|
}
|
||||||
|
|
||||||
/// Used by stream operations where the concurrency factor hasn't been manually
|
/// Used by stream operations where the concurrency factor hasn't been manually
|
||||||
/// supplied by the caller (most uses). Instead we provide a default value which
|
/// supplied by the caller (most uses). Instead we provide a default value which
|
||||||
/// is adjusted at startup for the specific system and also dynamically.
|
/// is adjusted at startup for the specific system and also dynamically.
|
||||||
|
@ -24,3 +38,13 @@ pub fn automatic_width() -> usize {
|
||||||
debug_assert!(width <= WIDTH_LIMIT.1, "WIDTH is probably too large");
|
debug_assert!(width <= WIDTH_LIMIT.1, "WIDTH is probably too large");
|
||||||
width
|
width
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Used by stream operations where the amplification hasn't been manually
|
||||||
|
/// supplied by the caller. Instead we provide a computed value.
|
||||||
|
#[inline]
|
||||||
|
pub fn automatic_amplification() -> usize {
|
||||||
|
let amplification = AMPLIFICATION.load(Ordering::Relaxed);
|
||||||
|
debug_assert!(amplification >= AMPLIFICATION_LIMIT.0, "amplification is too low");
|
||||||
|
debug_assert!(amplification <= AMPLIFICATION_LIMIT.1, "amplification is too high");
|
||||||
|
amplification
|
||||||
|
}
|
||||||
|
|
|
@ -10,7 +10,10 @@ mod try_broadband;
|
||||||
mod try_ready;
|
mod try_ready;
|
||||||
mod wideband;
|
mod wideband;
|
||||||
|
|
||||||
pub use band::{automatic_width, set_width, WIDTH_LIMIT};
|
pub use band::{
|
||||||
|
automatic_amplification, automatic_width, set_amplification, set_width, AMPLIFICATION_LIMIT,
|
||||||
|
WIDTH_LIMIT,
|
||||||
|
};
|
||||||
pub use broadband::BroadbandExt;
|
pub use broadband::BroadbandExt;
|
||||||
pub use cloned::Cloned;
|
pub use cloned::Cloned;
|
||||||
pub use expect::TryExpect;
|
pub use expect::TryExpect;
|
||||||
|
|
|
@ -2,7 +2,7 @@ use std::{convert::AsRef, fmt::Debug, io::Write, sync::Arc};
|
||||||
|
|
||||||
use arrayvec::ArrayVec;
|
use arrayvec::ArrayVec;
|
||||||
use conduwuit::{err, implement, utils::result::MapExpect, Err, Result};
|
use conduwuit::{err, implement, utils::result::MapExpect, Err, Result};
|
||||||
use futures::{Future, FutureExt};
|
use futures::{future::ready, Future, FutureExt, TryFutureExt};
|
||||||
use serde::Serialize;
|
use serde::Serialize;
|
||||||
use tokio::task;
|
use tokio::task;
|
||||||
|
|
||||||
|
@ -79,11 +79,15 @@ where
|
||||||
debug_assert!(matches!(cached, Ok(None)), "expected status Incomplete");
|
debug_assert!(matches!(cached, Ok(None)), "expected status Incomplete");
|
||||||
let cmd = Get {
|
let cmd = Get {
|
||||||
map: self.clone(),
|
map: self.clone(),
|
||||||
key: key.as_ref().into(),
|
key: [key.as_ref().into()].into(),
|
||||||
res: None,
|
res: None,
|
||||||
};
|
};
|
||||||
|
|
||||||
self.db.pool.execute_get(cmd).boxed()
|
self.db
|
||||||
|
.pool
|
||||||
|
.execute_get(cmd)
|
||||||
|
.and_then(|mut res| ready(res.remove(0)))
|
||||||
|
.boxed()
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Fetch a value from the database into cache, returning a reference-handle.
|
/// Fetch a value from the database into cache, returning a reference-handle.
|
||||||
|
|
|
@ -2,42 +2,68 @@ use std::{convert::AsRef, fmt::Debug, sync::Arc};
|
||||||
|
|
||||||
use conduwuit::{
|
use conduwuit::{
|
||||||
err, implement,
|
err, implement,
|
||||||
utils::{stream::automatic_width, IterStream},
|
utils::{
|
||||||
|
stream::{automatic_amplification, automatic_width, WidebandExt},
|
||||||
|
IterStream,
|
||||||
|
},
|
||||||
Result,
|
Result,
|
||||||
};
|
};
|
||||||
use futures::{Stream, StreamExt};
|
use futures::{Stream, StreamExt, TryStreamExt};
|
||||||
use serde::Serialize;
|
use serde::Serialize;
|
||||||
|
|
||||||
use crate::{util::map_err, Handle};
|
use crate::{keyval::KeyBuf, ser, util::map_err, Handle};
|
||||||
|
|
||||||
#[implement(super::Map)]
|
#[implement(super::Map)]
|
||||||
#[tracing::instrument(skip(self, keys), level = "trace")]
|
#[tracing::instrument(skip(self, keys), level = "trace")]
|
||||||
pub fn aqry_batch<'b, 'a: 'b, const MAX: usize, I, K>(
|
pub fn qry_batch<'a, S, K>(
|
||||||
self: &'a Arc<Self>,
|
self: &'a Arc<Self>,
|
||||||
keys: I,
|
keys: S,
|
||||||
) -> impl Stream<Item = Result<Handle<'b>>> + Send + 'a
|
) -> impl Stream<Item = Result<Handle<'_>>> + Send + 'a
|
||||||
where
|
where
|
||||||
I: Iterator<Item = &'b K> + Send + 'a,
|
S: Stream<Item = K> + Send + 'a,
|
||||||
K: Serialize + ?Sized + Debug + 'b,
|
K: Serialize + Debug + 'a,
|
||||||
{
|
{
|
||||||
keys.stream()
|
use crate::pool::Get;
|
||||||
.map(move |key| self.aqry::<MAX, _>(&key))
|
|
||||||
.buffered(automatic_width())
|
keys.ready_chunks(automatic_amplification())
|
||||||
|
.widen_then(automatic_width(), |chunk| {
|
||||||
|
let keys = chunk
|
||||||
|
.iter()
|
||||||
|
.map(ser::serialize_to::<KeyBuf, _>)
|
||||||
|
.map(|result| result.expect("failed to serialize query key"))
|
||||||
|
.map(Into::into)
|
||||||
|
.collect();
|
||||||
|
|
||||||
|
self.db
|
||||||
|
.pool
|
||||||
|
.execute_get(Get { map: self.clone(), key: keys, res: None })
|
||||||
|
})
|
||||||
|
.map_ok(|results| results.into_iter().stream())
|
||||||
|
.try_flatten()
|
||||||
}
|
}
|
||||||
|
|
||||||
#[implement(super::Map)]
|
#[implement(super::Map)]
|
||||||
#[tracing::instrument(skip(self, keys), level = "trace")]
|
#[tracing::instrument(skip(self, keys), level = "trace")]
|
||||||
pub fn get_batch<'a, I, K>(
|
pub fn get_batch<'a, S, K>(
|
||||||
self: &'a Arc<Self>,
|
self: &'a Arc<Self>,
|
||||||
keys: I,
|
keys: S,
|
||||||
) -> impl Stream<Item = Result<Handle<'_>>> + Send + 'a
|
) -> impl Stream<Item = Result<Handle<'_>>> + Send + 'a
|
||||||
where
|
where
|
||||||
I: Iterator<Item = &'a K> + Debug + Send + 'a,
|
S: Stream<Item = K> + Send + 'a,
|
||||||
K: AsRef<[u8]> + Debug + Send + ?Sized + Sync + 'a,
|
K: AsRef<[u8]> + Send + Sync + 'a,
|
||||||
{
|
{
|
||||||
keys.stream()
|
use crate::pool::Get;
|
||||||
.map(move |key| self.get(key))
|
|
||||||
.buffered(automatic_width())
|
keys.ready_chunks(automatic_amplification())
|
||||||
|
.widen_then(automatic_width(), |chunk| {
|
||||||
|
self.db.pool.execute_get(Get {
|
||||||
|
map: self.clone(),
|
||||||
|
key: chunk.iter().map(AsRef::as_ref).map(Into::into).collect(),
|
||||||
|
res: None,
|
||||||
|
})
|
||||||
|
})
|
||||||
|
.map_ok(|results| results.into_iter().stream())
|
||||||
|
.try_flatten()
|
||||||
}
|
}
|
||||||
|
|
||||||
#[implement(super::Map)]
|
#[implement(super::Map)]
|
||||||
|
@ -47,8 +73,8 @@ pub(crate) fn get_batch_blocking<'a, I, K>(
|
||||||
keys: I,
|
keys: I,
|
||||||
) -> impl Iterator<Item = Result<Handle<'_>>> + Send
|
) -> impl Iterator<Item = Result<Handle<'_>>> + Send
|
||||||
where
|
where
|
||||||
I: Iterator<Item = &'a K> + ExactSizeIterator + Debug + Send,
|
I: Iterator<Item = &'a K> + ExactSizeIterator + Send,
|
||||||
K: AsRef<[u8]> + Debug + Send + ?Sized + Sync + 'a,
|
K: AsRef<[u8]> + Send + ?Sized + Sync + 'a,
|
||||||
{
|
{
|
||||||
// Optimization can be `true` if key vector is pre-sorted **by the column
|
// Optimization can be `true` if key vector is pre-sorted **by the column
|
||||||
// comparator**.
|
// comparator**.
|
||||||
|
|
|
@ -19,6 +19,7 @@ use conduwuit::{
|
||||||
use futures::{channel::oneshot, TryFutureExt};
|
use futures::{channel::oneshot, TryFutureExt};
|
||||||
use oneshot::Sender as ResultSender;
|
use oneshot::Sender as ResultSender;
|
||||||
use rocksdb::Direction;
|
use rocksdb::Direction;
|
||||||
|
use smallvec::SmallVec;
|
||||||
use tokio::task::JoinSet;
|
use tokio::task::JoinSet;
|
||||||
|
|
||||||
use self::configure::configure;
|
use self::configure::configure;
|
||||||
|
@ -42,11 +43,11 @@ pub(crate) enum Cmd {
|
||||||
Iter(Seek),
|
Iter(Seek),
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Point-query
|
/// Multi-point-query
|
||||||
pub(crate) struct Get {
|
pub(crate) struct Get {
|
||||||
pub(crate) map: Arc<Map>,
|
pub(crate) map: Arc<Map>,
|
||||||
pub(crate) key: KeyBuf,
|
pub(crate) key: BatchQuery<'static>,
|
||||||
pub(crate) res: Option<ResultSender<Result<Handle<'static>>>>,
|
pub(crate) res: Option<ResultSender<BatchResult<'static>>>,
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Iterator-seek.
|
/// Iterator-seek.
|
||||||
|
@ -60,8 +61,13 @@ pub(crate) struct Seek {
|
||||||
pub(crate) res: Option<ResultSender<stream::State<'static>>>,
|
pub(crate) res: Option<ResultSender<stream::State<'static>>>,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
pub(crate) type BatchQuery<'a> = SmallVec<[KeyBuf; BATCH_INLINE]>;
|
||||||
|
pub(crate) type BatchResult<'a> = SmallVec<[ResultHandle<'a>; BATCH_INLINE]>;
|
||||||
|
pub(crate) type ResultHandle<'a> = Result<Handle<'a>>;
|
||||||
|
|
||||||
const WORKER_LIMIT: (usize, usize) = (1, 1024);
|
const WORKER_LIMIT: (usize, usize) = (1, 1024);
|
||||||
const QUEUE_LIMIT: (usize, usize) = (1, 2048);
|
const QUEUE_LIMIT: (usize, usize) = (1, 2048);
|
||||||
|
const BATCH_INLINE: usize = 1;
|
||||||
|
|
||||||
#[implement(Pool)]
|
#[implement(Pool)]
|
||||||
pub(crate) async fn new(server: &Arc<Server>) -> Result<Arc<Self>> {
|
pub(crate) async fn new(server: &Arc<Server>) -> Result<Arc<Self>> {
|
||||||
|
@ -179,22 +185,24 @@ fn spawn_one(self: &Arc<Self>, workers: &mut JoinSet<()>, recv: &[Receiver<Cmd>]
|
||||||
|
|
||||||
#[implement(Pool)]
|
#[implement(Pool)]
|
||||||
#[tracing::instrument(level = "trace", name = "get", skip(self, cmd))]
|
#[tracing::instrument(level = "trace", name = "get", skip(self, cmd))]
|
||||||
pub(crate) async fn execute_get(&self, mut cmd: Get) -> Result<Handle<'_>> {
|
pub(crate) async fn execute_get(self: &Arc<Self>, mut cmd: Get) -> Result<BatchResult<'_>> {
|
||||||
let (send, recv) = oneshot::channel();
|
let (send, recv) = oneshot::channel();
|
||||||
_ = cmd.res.insert(send);
|
_ = cmd.res.insert(send);
|
||||||
|
|
||||||
let queue = self.select_queue();
|
let queue = self.select_queue();
|
||||||
self.execute(queue, Cmd::Get(cmd))
|
self.execute(queue, Cmd::Get(cmd))
|
||||||
.and_then(|()| {
|
.and_then(move |()| {
|
||||||
recv.map_ok(into_recv_get_result)
|
recv.map_ok(into_recv_get)
|
||||||
.map_err(|e| err!(error!("recv failed {e:?}")))
|
.map_err(|e| err!(error!("recv failed {e:?}")))
|
||||||
})
|
})
|
||||||
.await?
|
.await
|
||||||
|
.map(Into::into)
|
||||||
|
.map_err(Into::into)
|
||||||
}
|
}
|
||||||
|
|
||||||
#[implement(Pool)]
|
#[implement(Pool)]
|
||||||
#[tracing::instrument(level = "trace", name = "iter", skip(self, cmd))]
|
#[tracing::instrument(level = "trace", name = "iter", skip(self, cmd))]
|
||||||
pub(crate) async fn execute_iter(&self, mut cmd: Seek) -> Result<stream::State<'_>> {
|
pub(crate) async fn execute_iter(self: &Arc<Self>, mut cmd: Seek) -> Result<stream::State<'_>> {
|
||||||
let (send, recv) = oneshot::channel();
|
let (send, recv) = oneshot::channel();
|
||||||
_ = cmd.res.insert(send);
|
_ = cmd.res.insert(send);
|
||||||
|
|
||||||
|
@ -282,7 +290,7 @@ fn worker_init(&self, id: usize) {
|
||||||
}
|
}
|
||||||
|
|
||||||
#[implement(Pool)]
|
#[implement(Pool)]
|
||||||
fn worker_loop(&self, recv: &Receiver<Cmd>) {
|
fn worker_loop(self: &Arc<Self>, recv: &Receiver<Cmd>) {
|
||||||
// initial +1 needed prior to entering wait
|
// initial +1 needed prior to entering wait
|
||||||
self.busy.fetch_add(1, Ordering::Relaxed);
|
self.busy.fetch_add(1, Ordering::Relaxed);
|
||||||
|
|
||||||
|
@ -302,18 +310,19 @@ fn worker_loop(&self, recv: &Receiver<Cmd>) {
|
||||||
busy = self.busy.fetch_sub(1, Ordering::Relaxed) - 1,
|
busy = self.busy.fetch_sub(1, Ordering::Relaxed) - 1,
|
||||||
),
|
),
|
||||||
)]
|
)]
|
||||||
fn worker_wait(&self, recv: &Receiver<Cmd>) -> Result<Cmd, RecvError> {
|
fn worker_wait(self: &Arc<Self>, recv: &Receiver<Cmd>) -> Result<Cmd, RecvError> {
|
||||||
recv.recv_blocking().debug_inspect(|_| {
|
recv.recv_blocking().debug_inspect(|_| {
|
||||||
self.busy.fetch_add(1, Ordering::Relaxed);
|
self.busy.fetch_add(1, Ordering::Relaxed);
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
#[implement(Pool)]
|
#[implement(Pool)]
|
||||||
fn worker_handle(&self, cmd: Cmd) {
|
fn worker_handle(self: &Arc<Self>, cmd: Cmd) {
|
||||||
match cmd {
|
match cmd {
|
||||||
| Cmd::Get(cmd) => self.handle_get(cmd),
|
| Cmd::Get(cmd) if cmd.key.len() == 1 => self.handle_get(cmd),
|
||||||
|
| Cmd::Get(cmd) => self.handle_batch(cmd),
|
||||||
| Cmd::Iter(cmd) => self.handle_iter(cmd),
|
| Cmd::Iter(cmd) => self.handle_iter(cmd),
|
||||||
}
|
};
|
||||||
}
|
}
|
||||||
|
|
||||||
#[implement(Pool)]
|
#[implement(Pool)]
|
||||||
|
@ -331,12 +340,43 @@ fn handle_iter(&self, mut cmd: Seek) {
|
||||||
}
|
}
|
||||||
|
|
||||||
let from = cmd.key.as_deref().map(Into::into);
|
let from = cmd.key.as_deref().map(Into::into);
|
||||||
|
|
||||||
let result = match cmd.dir {
|
let result = match cmd.dir {
|
||||||
| Direction::Forward => cmd.state.init_fwd(from),
|
| Direction::Forward => cmd.state.init_fwd(from),
|
||||||
| Direction::Reverse => cmd.state.init_rev(from),
|
| Direction::Reverse => cmd.state.init_rev(from),
|
||||||
};
|
};
|
||||||
|
|
||||||
let chan_result = chan.send(into_send_seek(result));
|
let chan_result = chan.send(into_send_seek(result));
|
||||||
|
|
||||||
|
let _chan_sent = chan_result.is_ok();
|
||||||
|
}
|
||||||
|
|
||||||
|
#[implement(Pool)]
|
||||||
|
#[tracing::instrument(
|
||||||
|
name = "batch",
|
||||||
|
level = "trace",
|
||||||
|
skip_all,
|
||||||
|
fields(
|
||||||
|
%cmd.map,
|
||||||
|
keys = %cmd.key.len(),
|
||||||
|
),
|
||||||
|
)]
|
||||||
|
fn handle_batch(self: &Arc<Self>, mut cmd: Get) {
|
||||||
|
debug_assert!(cmd.key.len() > 1, "should have more than one key");
|
||||||
|
debug_assert!(!cmd.key.iter().any(SmallVec::is_empty), "querying for empty key");
|
||||||
|
|
||||||
|
let chan = cmd.res.take().expect("missing result channel");
|
||||||
|
|
||||||
|
if chan.is_canceled() {
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
let keys = cmd.key.iter().map(Into::into);
|
||||||
|
|
||||||
|
let result: SmallVec<_> = cmd.map.get_batch_blocking(keys).collect();
|
||||||
|
|
||||||
|
let chan_result = chan.send(into_send_get(result));
|
||||||
|
|
||||||
let _chan_sent = chan_result.is_ok();
|
let _chan_sent = chan_result.is_ok();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -348,7 +388,7 @@ fn handle_iter(&self, mut cmd: Seek) {
|
||||||
fields(%cmd.map),
|
fields(%cmd.map),
|
||||||
)]
|
)]
|
||||||
fn handle_get(&self, mut cmd: Get) {
|
fn handle_get(&self, mut cmd: Get) {
|
||||||
debug_assert!(!cmd.key.is_empty(), "querying for empty key");
|
debug_assert!(!cmd.key[0].is_empty(), "querying for empty key");
|
||||||
|
|
||||||
// Obtain the result channel.
|
// Obtain the result channel.
|
||||||
let chan = cmd.res.take().expect("missing result channel");
|
let chan = cmd.res.take().expect("missing result channel");
|
||||||
|
@ -362,16 +402,16 @@ fn handle_get(&self, mut cmd: Get) {
|
||||||
// Perform the actual database query. We reuse our database::Map interface but
|
// Perform the actual database query. We reuse our database::Map interface but
|
||||||
// limited to the blocking calls, rather than creating another surface directly
|
// limited to the blocking calls, rather than creating another surface directly
|
||||||
// with rocksdb here.
|
// with rocksdb here.
|
||||||
let result = cmd.map.get_blocking(&cmd.key);
|
let result = cmd.map.get_blocking(&cmd.key[0]);
|
||||||
|
|
||||||
// Send the result back to the submitter.
|
// Send the result back to the submitter.
|
||||||
let chan_result = chan.send(into_send_get_result(result));
|
let chan_result = chan.send(into_send_get([result].into()));
|
||||||
|
|
||||||
// If the future was dropped during the query this will fail acceptably.
|
// If the future was dropped during the query this will fail acceptably.
|
||||||
let _chan_sent = chan_result.is_ok();
|
let _chan_sent = chan_result.is_ok();
|
||||||
}
|
}
|
||||||
|
|
||||||
fn into_send_get_result(result: Result<Handle<'_>>) -> Result<Handle<'static>> {
|
fn into_send_get(result: BatchResult<'_>) -> BatchResult<'static> {
|
||||||
// SAFETY: Necessary to send the Handle (rust_rocksdb::PinnableSlice) through
|
// SAFETY: Necessary to send the Handle (rust_rocksdb::PinnableSlice) through
|
||||||
// the channel. The lifetime on the handle is a device by rust-rocksdb to
|
// the channel. The lifetime on the handle is a device by rust-rocksdb to
|
||||||
// associate a database lifetime with its assets. The Handle must be dropped
|
// associate a database lifetime with its assets. The Handle must be dropped
|
||||||
|
@ -379,7 +419,7 @@ fn into_send_get_result(result: Result<Handle<'_>>) -> Result<Handle<'static>> {
|
||||||
unsafe { std::mem::transmute(result) }
|
unsafe { std::mem::transmute(result) }
|
||||||
}
|
}
|
||||||
|
|
||||||
fn into_recv_get_result(result: Result<Handle<'static>>) -> Result<Handle<'_>> {
|
fn into_recv_get<'a>(result: BatchResult<'static>) -> BatchResult<'a> {
|
||||||
// SAFETY: This is to receive the Handle from the channel.
|
// SAFETY: This is to receive the Handle from the channel.
|
||||||
unsafe { std::mem::transmute(result) }
|
unsafe { std::mem::transmute(result) }
|
||||||
}
|
}
|
||||||
|
|
|
@ -6,7 +6,7 @@ use conduwuit::{
|
||||||
math::usize_from_f64,
|
math::usize_from_f64,
|
||||||
result::LogDebugErr,
|
result::LogDebugErr,
|
||||||
stream,
|
stream,
|
||||||
stream::WIDTH_LIMIT,
|
stream::{AMPLIFICATION_LIMIT, WIDTH_LIMIT},
|
||||||
sys::{compute::is_core_available, storage},
|
sys::{compute::is_core_available, storage},
|
||||||
BoolExt,
|
BoolExt,
|
||||||
},
|
},
|
||||||
|
@ -124,19 +124,28 @@ pub(super) fn configure(server: &Arc<Server>) -> (usize, Vec<usize>, Vec<usize>)
|
||||||
fn update_stream_width(server: &Arc<Server>, num_queues: usize, total_workers: usize) {
|
fn update_stream_width(server: &Arc<Server>, num_queues: usize, total_workers: usize) {
|
||||||
let config = &server.config;
|
let config = &server.config;
|
||||||
let scale: f64 = config.stream_width_scale.min(100.0).into();
|
let scale: f64 = config.stream_width_scale.min(100.0).into();
|
||||||
|
|
||||||
let req_width = expected!(total_workers / num_queues).next_multiple_of(2);
|
let req_width = expected!(total_workers / num_queues).next_multiple_of(2);
|
||||||
let req_width = req_width as f64;
|
let req_width = req_width as f64;
|
||||||
let req_width = usize_from_f64(req_width * scale)
|
let req_width = usize_from_f64(req_width * scale)
|
||||||
.expect("failed to convert f64 to usize")
|
.expect("failed to convert f64 to usize")
|
||||||
.clamp(WIDTH_LIMIT.0, WIDTH_LIMIT.1);
|
.clamp(WIDTH_LIMIT.0, WIDTH_LIMIT.1);
|
||||||
|
|
||||||
|
let req_amp = config.stream_amplification as f64;
|
||||||
|
let req_amp = usize_from_f64(req_amp * scale)
|
||||||
|
.expect("failed to convert f64 to usize")
|
||||||
|
.clamp(AMPLIFICATION_LIMIT.0, AMPLIFICATION_LIMIT.1);
|
||||||
|
|
||||||
let (old_width, new_width) = stream::set_width(req_width);
|
let (old_width, new_width) = stream::set_width(req_width);
|
||||||
|
let (old_amp, new_amp) = stream::set_amplification(req_amp);
|
||||||
debug!(
|
debug!(
|
||||||
scale = ?config.stream_width_scale,
|
scale = ?config.stream_width_scale,
|
||||||
?num_queues,
|
?num_queues,
|
||||||
?req_width,
|
?req_width,
|
||||||
?old_width,
|
?old_width,
|
||||||
?new_width,
|
?new_width,
|
||||||
|
?old_amp,
|
||||||
|
?new_amp,
|
||||||
"Updated global stream width"
|
"Updated global stream width"
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|
|
@ -52,6 +52,7 @@ impl Server {
|
||||||
.expect("Unable to increase maximum soft and hard file descriptor limit");
|
.expect("Unable to increase maximum soft and hard file descriptor limit");
|
||||||
|
|
||||||
let (_old_width, _new_width) = stream::set_width(config.stream_width_default);
|
let (_old_width, _new_width) = stream::set_width(config.stream_width_default);
|
||||||
|
let (_old_amp, _new_amp) = stream::set_amplification(config.stream_amplification);
|
||||||
|
|
||||||
info!(
|
info!(
|
||||||
server_name = %config.server_name,
|
server_name = %config.server_name,
|
||||||
|
|
|
@ -71,7 +71,7 @@ impl Service {
|
||||||
let event_ids = self
|
let event_ids = self
|
||||||
.services
|
.services
|
||||||
.short
|
.short
|
||||||
.multi_get_eventid_from_short(chain.iter())
|
.multi_get_eventid_from_short(chain.into_iter().stream())
|
||||||
.ready_filter_map(Result::ok)
|
.ready_filter_map(Result::ok)
|
||||||
.collect()
|
.collect()
|
||||||
.await;
|
.await;
|
||||||
|
|
|
@ -69,7 +69,7 @@ where
|
||||||
{
|
{
|
||||||
self.db
|
self.db
|
||||||
.eventid_shorteventid
|
.eventid_shorteventid
|
||||||
.get_batch(event_ids.clone())
|
.get_batch(event_ids.clone().stream())
|
||||||
.zip(event_ids.into_iter().stream())
|
.zip(event_ids.into_iter().stream())
|
||||||
.map(|(result, event_id)| match result {
|
.map(|(result, event_id)| match result {
|
||||||
| Ok(ref short) => utils::u64_from_u8(short),
|
| Ok(ref short) => utils::u64_from_u8(short),
|
||||||
|
@ -162,20 +162,18 @@ where
|
||||||
}
|
}
|
||||||
|
|
||||||
#[implement(Service)]
|
#[implement(Service)]
|
||||||
pub fn multi_get_eventid_from_short<'a, Id, I>(
|
pub fn multi_get_eventid_from_short<'a, Id, S>(
|
||||||
&'a self,
|
&'a self,
|
||||||
shorteventid: I,
|
shorteventid: S,
|
||||||
) -> impl Stream<Item = Result<Id>> + Send + 'a
|
) -> impl Stream<Item = Result<Id>> + Send + 'a
|
||||||
where
|
where
|
||||||
I: Iterator<Item = &'a ShortEventId> + Send + 'a,
|
S: Stream<Item = ShortEventId> + Send + 'a,
|
||||||
Id: for<'de> Deserialize<'de> + Sized + ToOwned + 'a,
|
Id: for<'de> Deserialize<'de> + Sized + ToOwned + 'a,
|
||||||
<Id as ToOwned>::Owned: Borrow<EventId>,
|
<Id as ToOwned>::Owned: Borrow<EventId>,
|
||||||
{
|
{
|
||||||
const BUFSIZE: usize = size_of::<ShortEventId>();
|
|
||||||
|
|
||||||
self.db
|
self.db
|
||||||
.shorteventid_eventid
|
.shorteventid_eventid
|
||||||
.aqry_batch::<BUFSIZE, _, _>(shorteventid)
|
.qry_batch(shorteventid)
|
||||||
.map(Deserialized::deserialized)
|
.map(Deserialized::deserialized)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -467,7 +467,7 @@ impl Service {
|
||||||
let auth_pdus = self
|
let auth_pdus = self
|
||||||
.services
|
.services
|
||||||
.short
|
.short
|
||||||
.multi_get_eventid_from_short(event_ids.iter())
|
.multi_get_eventid_from_short(event_ids.into_iter().stream())
|
||||||
.zip(state_keys.into_iter().stream())
|
.zip(state_keys.into_iter().stream())
|
||||||
.ready_filter_map(|(event_id, tsk)| Some((tsk, event_id.ok()?)))
|
.ready_filter_map(|(event_id, tsk)| Some((tsk, event_id.ok()?)))
|
||||||
.broad_filter_map(|(tsk, event_id): (_, OwnedEventId)| async move {
|
.broad_filter_map(|(tsk, event_id): (_, OwnedEventId)| async move {
|
||||||
|
|
|
@ -1,7 +1,7 @@
|
||||||
use std::{borrow::Borrow, collections::HashMap, sync::Arc};
|
use std::{borrow::Borrow, collections::HashMap, sync::Arc};
|
||||||
|
|
||||||
use conduwuit::{
|
use conduwuit::{
|
||||||
at, err, ref_at,
|
at, err,
|
||||||
utils::stream::{BroadbandExt, IterStream, ReadyExt},
|
utils::stream::{BroadbandExt, IterStream, ReadyExt},
|
||||||
PduEvent, Result,
|
PduEvent, Result,
|
||||||
};
|
};
|
||||||
|
@ -69,7 +69,7 @@ impl Data {
|
||||||
let full_pdus = self
|
let full_pdus = self
|
||||||
.services
|
.services
|
||||||
.short
|
.short
|
||||||
.multi_get_eventid_from_short(short_ids.iter().map(ref_at!(1)))
|
.multi_get_eventid_from_short(short_ids.into_iter().map(at!(1)).stream())
|
||||||
.ready_filter_map(Result::ok)
|
.ready_filter_map(Result::ok)
|
||||||
.broad_filter_map(|event_id: OwnedEventId| async move {
|
.broad_filter_map(|event_id: OwnedEventId| async move {
|
||||||
self.services.timeline.get_pdu(&event_id).await.ok()
|
self.services.timeline.get_pdu(&event_id).await.ok()
|
||||||
|
@ -93,7 +93,7 @@ impl Data {
|
||||||
let full_ids = self
|
let full_ids = self
|
||||||
.services
|
.services
|
||||||
.short
|
.short
|
||||||
.multi_get_eventid_from_short(short_ids.iter().map(ref_at!(1)))
|
.multi_get_eventid_from_short(short_ids.iter().map(at!(1)).stream())
|
||||||
.zip(short_ids.iter().stream().map(at!(0)))
|
.zip(short_ids.iter().stream().map(at!(0)))
|
||||||
.ready_filter_map(|(event_id, shortstatekey)| Some((shortstatekey, event_id.ok()?)))
|
.ready_filter_map(|(event_id, shortstatekey)| Some((shortstatekey, event_id.ok()?)))
|
||||||
.collect()
|
.collect()
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue