chore: Fix typos across the codebase
This commit is contained in:
parent
d78fc53577
commit
01594a6243
21 changed files with 55 additions and 43 deletions
5
.typos.toml
Normal file
5
.typos.toml
Normal file
|
@ -0,0 +1,5 @@
|
|||
[default.extend-words]
|
||||
"allocatedp" = "allocatedp"
|
||||
"conduwuit" = "conduwuit"
|
||||
"continuwuity" = "continuwuity"
|
||||
"execuse" = "execuse"
|
|
@ -28,7 +28,7 @@ services:
|
|||
#CONDUWUIT_CONFIG: '/etc/conduwuit.toml' # Uncomment if you mapped config toml above
|
||||
|
||||
# We need some way to serve the client and server .well-known json. The simplest way is via the CONDUWUIT_WELL_KNOWN
|
||||
# variable / config option, there are multiple ways to do this, e.g. in the conduwuit.toml file, and in a seperate
|
||||
# variable / config option, there are multiple ways to do this, e.g. in the conduwuit.toml file, and in a separate
|
||||
# see the override file for more information about delegation
|
||||
CONDUWUIT_WELL_KNOWN: |
|
||||
{
|
||||
|
|
|
@ -36,7 +36,7 @@ services:
|
|||
# CONDUWUIT_NEW_USER_DISPLAYNAME_SUFFIX = "🏳<200d>⚧"
|
||||
|
||||
# We need some way to serve the client and server .well-known json. The simplest way is via the CONDUWUIT_WELL_KNOWN
|
||||
# variable / config option, there are multiple ways to do this, e.g. in the conduwuit.toml file, and in a seperate
|
||||
# variable / config option, there are multiple ways to do this, e.g. in the conduwuit.toml file, and in a separate
|
||||
# reverse proxy, but since you do not have a reverse proxy and following this guide, this example is included
|
||||
CONDUWUIT_WELL_KNOWN: |
|
||||
{
|
||||
|
|
|
@ -71,7 +71,7 @@ related to WAL tracking.
|
|||
|
||||
The only safe files that can be deleted are the `LOG` files (all caps). These
|
||||
are the real RocksDB telemetry/log files, however Continuwuity has already
|
||||
configured to only store up to 3 RocksDB `LOG` files due to generall being
|
||||
configured to only store up to 3 RocksDB `LOG` files due to generally being
|
||||
useless for average users unless troubleshooting something low-level. If you
|
||||
would like to store nearly none at all, see the `rocksdb_max_log_files`
|
||||
config option.
|
||||
|
|
|
@ -15,7 +15,7 @@ use crate::Ruma;
|
|||
|
||||
/// # `GET /_matrix/client/v3/capabilities`
|
||||
///
|
||||
/// Get information on the supported feature set and other relevent capabilities
|
||||
/// Get information on the supported feature set and other relevant capabilities
|
||||
/// of this server.
|
||||
pub(crate) async fn get_capabilities_route(
|
||||
State(services): State<crate::State>,
|
||||
|
|
|
@ -60,7 +60,7 @@ pub(crate) async fn well_known_support(
|
|||
return Err(Error::BadRequest(ErrorKind::NotFound, "Not found."));
|
||||
}
|
||||
|
||||
// TOOD: support defining multiple contacts in the config
|
||||
// TODO: support defining multiple contacts in the config
|
||||
let mut contacts: Vec<Contact> = vec![];
|
||||
|
||||
if let Some(role) = role {
|
||||
|
|
|
@ -36,7 +36,7 @@ const MAIN_MANIFEST: &'static str = ();
|
|||
/// For *enabled* features see the info::rustc module instead.
|
||||
static FEATURES: OnceLock<Vec<String>> = OnceLock::new();
|
||||
|
||||
/// Processed list of dependencies. This is generated from the datas captured in
|
||||
/// Processed list of dependencies. This is generated from the data captured in
|
||||
/// the MANIFEST.
|
||||
static DEPENDENCIES: OnceLock<DepsSet> = OnceLock::new();
|
||||
|
||||
|
|
|
@ -16,9 +16,9 @@ use crate::{Result, error};
|
|||
/// pulling in a version of tracing that's incompatible with the rest of our
|
||||
/// deps.
|
||||
///
|
||||
/// To work around this, we define an trait without the S paramter that forwards
|
||||
/// to the reload::Handle::reload method, and then store the handle as a trait
|
||||
/// object.
|
||||
/// To work around this, we define an trait without the S parameter that
|
||||
/// forwards to the reload::Handle::reload method, and then store the handle as
|
||||
/// a trait object.
|
||||
///
|
||||
/// [1]: <https://github.com/tokio-rs/tracing/pull/1035/commits/8a87ea52425098d3ef8f56d92358c2f6c144a28f>
|
||||
pub trait ReloadHandle<L> {
|
||||
|
|
|
@ -1,12 +1,12 @@
|
|||
type Delim<'a> = (&'a str, &'a str);
|
||||
|
||||
/// Slice a string between a pair of delimeters.
|
||||
/// Slice a string between a pair of delimiters.
|
||||
pub trait Between<'a> {
|
||||
/// Extract a string between the delimeters. If the delimeters were not
|
||||
/// Extract a string between the delimiters. If the delimiters were not
|
||||
/// found None is returned, otherwise the first extraction is returned.
|
||||
fn between(&self, delim: Delim<'_>) -> Option<&'a str>;
|
||||
|
||||
/// Extract a string between the delimeters. If the delimeters were not
|
||||
/// Extract a string between the delimiters. If the delimiters were not
|
||||
/// found the original string is returned; take note of this behavior,
|
||||
/// if an empty slice is desired for this case use the fallible version and
|
||||
/// unwrap to EMPTY.
|
||||
|
|
|
@ -193,7 +193,7 @@ fn get_cache(ctx: &Context, desc: &Descriptor) -> Option<Cache> {
|
|||
return None;
|
||||
}
|
||||
|
||||
// Some cache capacities are overriden by server config in a strange but
|
||||
// Some cache capacities are overridden by server config in a strange but
|
||||
// legacy-compat way
|
||||
let config = &ctx.server.config;
|
||||
let cap = match desc.name {
|
||||
|
|
|
@ -73,7 +73,7 @@ async fn async_main(server: &Arc<Server>) -> Result<(), Error> {
|
|||
.lock()
|
||||
.await
|
||||
.take()
|
||||
.expect("services initialied"),
|
||||
.expect("services initialized"),
|
||||
)
|
||||
.await
|
||||
{
|
||||
|
|
|
@ -13,8 +13,8 @@ pub(super) fn restart() -> ! {
|
|||
//
|
||||
// We can (and do) prevent that panic by checking the result of current_exe()
|
||||
// prior to committing to restart, returning an error to the user without any
|
||||
// unexpected shutdown. In a nutshell that is the execuse for this unsafety.
|
||||
// Nevertheless, we still want a way to override the restart preventation (i.e.
|
||||
// unexpected shutdown. In a nutshell that is the excuse for this unsafety.
|
||||
// Nevertheless, we still want a way to override the restart presentation (i.e.
|
||||
// admin server restart --force).
|
||||
let exe = unsafe { utils::sys::current_exe().expect("program path must be available") };
|
||||
let envs = env::vars();
|
||||
|
|
|
@ -25,7 +25,7 @@ pub(super) async fn console_auto_stop(&self) {
|
|||
/// Execute admin commands after startup
|
||||
#[implement(super::Service)]
|
||||
pub(super) async fn startup_execute(&self) -> Result {
|
||||
// List of comamnds to execute
|
||||
// List of commands to execute
|
||||
let commands = &self.services.server.config.admin_execute;
|
||||
|
||||
// Determine if we're running in smoketest-mode which will change some behaviors
|
||||
|
@ -64,7 +64,7 @@ pub(super) async fn startup_execute(&self) -> Result {
|
|||
/// Execute admin commands after signal
|
||||
#[implement(super::Service)]
|
||||
pub(super) async fn signal_execute(&self) -> Result {
|
||||
// List of comamnds to execute
|
||||
// List of commands to execute
|
||||
let commands = self.services.server.config.admin_signal_execute.clone();
|
||||
|
||||
// When true, errors are ignored and execution continues.
|
||||
|
|
|
@ -166,7 +166,7 @@ impl Service {
|
|||
.map_err(|e| err!("Failed to enqueue admin command: {e:?}"))
|
||||
}
|
||||
|
||||
/// Dispatches a comamnd to the processor on the current task and waits for
|
||||
/// Dispatches a command to the processor on the current task and waits for
|
||||
/// completion.
|
||||
pub async fn command_in_place(
|
||||
&self,
|
||||
|
|
|
@ -26,7 +26,7 @@ impl NamespaceRegex {
|
|||
false
|
||||
}
|
||||
|
||||
/// Checks if this namespace has exlusive rights to a namespace
|
||||
/// Checks if this namespace has exclusive rights to a namespace
|
||||
#[inline]
|
||||
#[must_use]
|
||||
pub fn is_exclusive_match(&self, heystack: &str) -> bool {
|
||||
|
|
|
@ -338,7 +338,7 @@ fn handle_federation_error(
|
|||
return fallback();
|
||||
}
|
||||
|
||||
// Reached for 5xx errors. This is where we don't fallback given the likelyhood
|
||||
// Reached for 5xx errors. This is where we don't fallback given the likelihood
|
||||
// the other endpoint will also be a 5xx and we're wasting time.
|
||||
error
|
||||
}
|
||||
|
@ -356,7 +356,7 @@ pub async fn fetch_remote_thumbnail_legacy(
|
|||
|
||||
self.check_legacy_freeze()?;
|
||||
self.check_fetch_authorized(&mxc)?;
|
||||
let reponse = self
|
||||
let response = self
|
||||
.services
|
||||
.sending
|
||||
.send_federation_request(mxc.server_name, media::get_content_thumbnail::v3::Request {
|
||||
|
@ -373,10 +373,17 @@ pub async fn fetch_remote_thumbnail_legacy(
|
|||
.await?;
|
||||
|
||||
let dim = Dim::from_ruma(body.width, body.height, body.method.clone())?;
|
||||
self.upload_thumbnail(&mxc, None, None, reponse.content_type.as_deref(), &dim, &reponse.file)
|
||||
.await?;
|
||||
self.upload_thumbnail(
|
||||
&mxc,
|
||||
None,
|
||||
None,
|
||||
response.content_type.as_deref(),
|
||||
&dim,
|
||||
&response.file,
|
||||
)
|
||||
.await?;
|
||||
|
||||
Ok(reponse)
|
||||
Ok(response)
|
||||
}
|
||||
|
||||
#[implement(super::Service)]
|
||||
|
|
|
@ -296,7 +296,7 @@ impl super::Service {
|
|||
expire: CachedOverride::default_expire(),
|
||||
overriding: (hostname != untername)
|
||||
.then_some(hostname.into())
|
||||
.inspect(|_| debug_info!("{untername:?} overriden by {hostname:?}")),
|
||||
.inspect(|_| debug_info!("{untername:?} overridden by {hostname:?}")),
|
||||
});
|
||||
|
||||
Ok(())
|
||||
|
|
|
@ -399,7 +399,7 @@ async fn get_room_summary(
|
|||
Ok(summary)
|
||||
}
|
||||
|
||||
/// With the given identifier, checks if a room is accessable
|
||||
/// With the given identifier, checks if a room is accessible
|
||||
#[implement(Service)]
|
||||
async fn is_accessible_child<'a, I>(
|
||||
&self,
|
||||
|
|
|
@ -267,15 +267,15 @@ impl Service {
|
|||
///
|
||||
/// Returns pdu id
|
||||
#[tracing::instrument(level = "debug", skip_all)]
|
||||
pub async fn append_pdu<'a, Leafs>(
|
||||
pub async fn append_pdu<'a, Leaves>(
|
||||
&'a self,
|
||||
pdu: &'a PduEvent,
|
||||
mut pdu_json: CanonicalJsonObject,
|
||||
leafs: Leafs,
|
||||
leaves: Leaves,
|
||||
state_lock: &'a RoomMutexGuard,
|
||||
) -> Result<RawPduId>
|
||||
where
|
||||
Leafs: Iterator<Item = &'a EventId> + Send + 'a,
|
||||
Leaves: Iterator<Item = &'a EventId> + Send + 'a,
|
||||
{
|
||||
// Coalesce database writes for the remainder of this scope.
|
||||
let _cork = self.db.db.cork_and_flush();
|
||||
|
@ -344,7 +344,7 @@ impl Service {
|
|||
|
||||
self.services
|
||||
.state
|
||||
.set_forward_extremities(&pdu.room_id, leafs, state_lock)
|
||||
.set_forward_extremities(&pdu.room_id, leaves, state_lock)
|
||||
.await;
|
||||
|
||||
let insert_lock = self.mutex_insert.lock(&pdu.room_id).await;
|
||||
|
@ -951,17 +951,17 @@ impl Service {
|
|||
/// Append the incoming event setting the state snapshot to the state from
|
||||
/// the server that sent the event.
|
||||
#[tracing::instrument(level = "debug", skip_all)]
|
||||
pub async fn append_incoming_pdu<'a, Leafs>(
|
||||
pub async fn append_incoming_pdu<'a, Leaves>(
|
||||
&'a self,
|
||||
pdu: &'a PduEvent,
|
||||
pdu_json: CanonicalJsonObject,
|
||||
new_room_leafs: Leafs,
|
||||
new_room_leaves: Leaves,
|
||||
state_ids_compressed: Arc<CompressedState>,
|
||||
soft_fail: bool,
|
||||
state_lock: &'a RoomMutexGuard,
|
||||
) -> Result<Option<RawPduId>>
|
||||
where
|
||||
Leafs: Iterator<Item = &'a EventId> + Send + 'a,
|
||||
Leaves: Iterator<Item = &'a EventId> + Send + 'a,
|
||||
{
|
||||
// We append to state before appending the pdu, so we don't have a moment in
|
||||
// time with the pdu without it's state. This is okay because append_pdu can't
|
||||
|
@ -978,14 +978,14 @@ impl Service {
|
|||
|
||||
self.services
|
||||
.state
|
||||
.set_forward_extremities(&pdu.room_id, new_room_leafs, state_lock)
|
||||
.set_forward_extremities(&pdu.room_id, new_room_leaves, state_lock)
|
||||
.await;
|
||||
|
||||
return Ok(None);
|
||||
}
|
||||
|
||||
let pdu_id = self
|
||||
.append_pdu(pdu, pdu_json, new_room_leafs, state_lock)
|
||||
.append_pdu(pdu, pdu_json, new_room_leaves, state_lock)
|
||||
.await?;
|
||||
|
||||
Ok(Some(pdu_id))
|
||||
|
|
|
@ -495,7 +495,7 @@ ul#searchresults span.teaser em {
|
|||
|
||||
.chapter li {
|
||||
display: flex;
|
||||
color: var(--sidebar-non-existant);
|
||||
color: var(--sidebar-non-existent);
|
||||
}
|
||||
.chapter li a {
|
||||
display: block;
|
||||
|
|
|
@ -20,7 +20,7 @@
|
|||
|
||||
--sidebar-bg: #14191f;
|
||||
--sidebar-fg: #c8c9db;
|
||||
--sidebar-non-existant: #5c6773;
|
||||
--sidebar-non-existent: #5c6773;
|
||||
--sidebar-active: #ffb454;
|
||||
--sidebar-spacer: #2d334f;
|
||||
|
||||
|
@ -64,7 +64,7 @@
|
|||
|
||||
--sidebar-bg: #292c2f;
|
||||
--sidebar-fg: #a1adb8;
|
||||
--sidebar-non-existant: #505254;
|
||||
--sidebar-non-existent: #505254;
|
||||
--sidebar-active: #3473ad;
|
||||
--sidebar-spacer: #393939;
|
||||
|
||||
|
@ -108,7 +108,7 @@
|
|||
|
||||
--sidebar-bg: #fafafa;
|
||||
--sidebar-fg: #AE518E;
|
||||
--sidebar-non-existant: #aaaaaa;
|
||||
--sidebar-non-existent: #aaaaaa;
|
||||
--sidebar-active: #2F7E86;
|
||||
--sidebar-spacer: #f4f4f4;
|
||||
|
||||
|
@ -152,7 +152,7 @@
|
|||
|
||||
--sidebar-bg: #282d3f;
|
||||
--sidebar-fg: #fdcbec;
|
||||
--sidebar-non-existant: #505274;
|
||||
--sidebar-non-existent: #505274;
|
||||
--sidebar-active: #5BCEFA;
|
||||
--sidebar-spacer: #2d334f;
|
||||
|
||||
|
@ -196,7 +196,7 @@
|
|||
|
||||
--sidebar-bg: #3b2e2a;
|
||||
--sidebar-fg: #c8c9db;
|
||||
--sidebar-non-existant: #505254;
|
||||
--sidebar-non-existent: #505254;
|
||||
--sidebar-active: #e69f67;
|
||||
--sidebar-spacer: #45373a;
|
||||
|
||||
|
@ -241,7 +241,7 @@
|
|||
|
||||
--sidebar-bg: #292c2f;
|
||||
--sidebar-fg: #a1adb8;
|
||||
--sidebar-non-existant: #505254;
|
||||
--sidebar-non-existent: #505254;
|
||||
--sidebar-active: #3473ad;
|
||||
--sidebar-spacer: #393939;
|
||||
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue