fix edition 2024 lints

Signed-off-by: Jason Volk <jason@zemos.net>
This commit is contained in:
Jason Volk 2025-02-25 18:38:12 +00:00 committed by June Clementine Strawberry
parent cbf207bd1f
commit a67ab75417
36 changed files with 60 additions and 72 deletions

View file

@ -905,6 +905,7 @@ missing_docs_in_private_items = { level = "allow", priority = 1 }
missing_errors_doc = { level = "allow", priority = 1 } missing_errors_doc = { level = "allow", priority = 1 }
missing_panics_doc = { level = "allow", priority = 1 } missing_panics_doc = { level = "allow", priority = 1 }
module_name_repetitions = { level = "allow", priority = 1 } module_name_repetitions = { level = "allow", priority = 1 }
needless_continue = { level = "allow", priority = 1 }
no_effect_underscore_binding = { level = "allow", priority = 1 } no_effect_underscore_binding = { level = "allow", priority = 1 }
similar_names = { level = "allow", priority = 1 } similar_names = { level = "allow", priority = 1 }
single_match_else = { level = "allow", priority = 1 } single_match_else = { level = "allow", priority = 1 }
@ -969,6 +970,7 @@ style = { level = "warn", priority = -1 }
# trivial assertions are quite alright # trivial assertions are quite alright
assertions_on_constants = { level = "allow", priority = 1 } assertions_on_constants = { level = "allow", priority = 1 }
module_inception = { level = "allow", priority = 1 } module_inception = { level = "allow", priority = 1 }
obfuscated_if_else = { level = "allow", priority = 1 }
################### ###################
suspicious = { level = "warn", priority = -1 } suspicious = { level = "warn", priority = -1 }

View file

@ -62,7 +62,7 @@ pub(super) async fn process(command: AdminCommand, context: &Command<'_>) -> Res
| Debug(command) => debug::process(command, context).await?, | Debug(command) => debug::process(command, context).await?,
| Query(command) => query::process(command, context).await?, | Query(command) => query::process(command, context).await?,
| Check(command) => check::process(command, context).await?, | Check(command) => check::process(command, context).await?,
}; }
Ok(()) Ok(())
} }

View file

@ -42,7 +42,7 @@ pub(super) async fn list_rooms(
if rooms.is_empty() { if rooms.is_empty() {
return Ok(RoomMessageEventContent::text_plain("No more rooms.")); return Ok(RoomMessageEventContent::text_plain("No more rooms."));
}; }
let output_plain = format!( let output_plain = format!(
"Rooms ({}):\n```\n{}\n```", "Rooms ({}):\n```\n{}\n```",

View file

@ -67,7 +67,7 @@ pub(super) async fn reprocess(
if rooms.is_empty() { if rooms.is_empty() {
return Ok(RoomMessageEventContent::text_plain("No more rooms.")); return Ok(RoomMessageEventContent::text_plain("No more rooms."));
}; }
let output = format!( let output = format!(
"Rooms (page {page}):\n```\n{}\n```", "Rooms (page {page}):\n```\n{}\n```",

View file

@ -166,7 +166,7 @@ pub(super) async fn create_user(
"Failed to automatically join room {room} for user {user_id}: {e}" "Failed to automatically join room {room} for user {user_id}: {e}"
); );
}, },
}; }
} }
} }
} }
@ -550,7 +550,7 @@ pub(super) async fn force_join_list_of_local_users(
debug_warn!("Failed force joining {user_id} to {room_id} during bulk join: {e}"); debug_warn!("Failed force joining {user_id} to {room_id} during bulk join: {e}");
failed_joins = failed_joins.saturating_add(1); failed_joins = failed_joins.saturating_add(1);
}, },
}; }
} }
Ok(RoomMessageEventContent::notice_markdown(format!( Ok(RoomMessageEventContent::notice_markdown(format!(
@ -646,7 +646,7 @@ pub(super) async fn force_join_all_local_users(
debug_warn!("Failed force joining {user_id} to {room_id} during bulk join: {e}"); debug_warn!("Failed force joining {user_id} to {room_id} during bulk join: {e}");
failed_joins = failed_joins.saturating_add(1); failed_joins = failed_joins.saturating_add(1);
}, },
}; }
} }
Ok(RoomMessageEventContent::notice_markdown(format!( Ok(RoomMessageEventContent::notice_markdown(format!(

View file

@ -499,7 +499,7 @@ pub(crate) async fn register_route(
| _ => { | _ => {
info!("Automatically joined room {room} for user {user_id}"); info!("Automatically joined room {room} for user {user_id}");
}, },
}; }
} }
} }
} }

View file

@ -131,7 +131,7 @@ pub(crate) async fn set_room_visibility_route(
if !services.rooms.metadata.exists(&body.room_id).await { if !services.rooms.metadata.exists(&body.room_id).await {
// Return 404 if the room doesn't exist // Return 404 if the room doesn't exist
return Err(Error::BadRequest(ErrorKind::NotFound, "Room not found")); return Err!(Request(NotFound("Room not found")));
} }
if services if services
@ -145,10 +145,7 @@ pub(crate) async fn set_room_visibility_route(
} }
if !user_can_publish_room(&services, sender_user, &body.room_id).await? { if !user_can_publish_room(&services, sender_user, &body.room_id).await? {
return Err(Error::BadRequest( return Err!(Request(Forbidden("User is not allowed to publish this room")));
ErrorKind::forbidden(),
"User is not allowed to publish this room",
));
} }
match &body.visibility { match &body.visibility {
@ -386,12 +383,7 @@ async fn user_can_publish_room(
.await .await
{ {
| Ok(event) => Ok(event.sender == user_id), | Ok(event) => Ok(event.sender == user_id),
| _ => { | _ => Err!(Request(Forbidden("User is not allowed to publish this room"))),
return Err(Error::BadRequest(
ErrorKind::forbidden(),
"User is not allowed to publish this room",
));
},
} }
}, },
} }

View file

@ -993,7 +993,7 @@ async fn join_room_by_id_helper_remote(
| _ => { | _ => {
join_event_stub.remove("event_id"); join_event_stub.remove("event_id");
}, },
}; }
// In order to create a compatible ref hash (EventID) the `hashes` field needs // In order to create a compatible ref hash (EventID) the `hashes` field needs
// to be present // to be present
@ -1420,7 +1420,7 @@ async fn join_room_by_id_helper_local(
| _ => { | _ => {
join_event_stub.remove("event_id"); join_event_stub.remove("event_id");
}, },
}; }
// In order to create a compatible ref hash (EventID) the `hashes` field needs // In order to create a compatible ref hash (EventID) the `hashes` field needs
// to be present // to be present
@ -1947,7 +1947,7 @@ async fn remote_leave_room(
| _ => { | _ => {
leave_event_stub.remove("event_id"); leave_event_stub.remove("event_id");
}, },
}; }
// In order to create a compatible ref hash (EventID) the `hashes` field needs // In order to create a compatible ref hash (EventID) the `hashes` field needs
// to be present // to be present

View file

@ -43,7 +43,7 @@ pub(crate) async fn report_room_route(
ErrorKind::InvalidParam, ErrorKind::InvalidParam,
"Reason too long, should be 750 characters or fewer", "Reason too long, should be 750 characters or fewer",
)); ));
}; }
delay_response().await; delay_response().await;
@ -164,14 +164,14 @@ async fn is_event_report_valid(
ErrorKind::InvalidParam, ErrorKind::InvalidParam,
"Invalid score, must be within 0 to -100", "Invalid score, must be within 0 to -100",
)); ));
}; }
if reason.as_ref().is_some_and(|s| s.len() > 750) { if reason.as_ref().is_some_and(|s| s.len() > 750) {
return Err(Error::BadRequest( return Err(Error::BadRequest(
ErrorKind::InvalidParam, ErrorKind::InvalidParam,
"Reason too long, should be 750 characters or fewer", "Reason too long, should be 750 characters or fewer",
)); ));
}; }
if !services if !services
.rooms .rooms

View file

@ -110,7 +110,7 @@ pub(super) async fn auth(
} }
}, },
| _ => {}, | _ => {},
}; }
} }
match (metadata.authentication, token) { match (metadata.authentication, token) {

View file

@ -135,7 +135,7 @@ async fn create_join_event(
if state_key != sender { if state_key != sender {
return Err!(Request(BadJson("State key does not match sender user."))); return Err!(Request(BadJson("State key does not match sender user.")));
}; }
if let Some(authorising_user) = content.join_authorized_via_users_server { if let Some(authorising_user) = content.join_authorized_via_users_server {
use ruma::RoomVersionId::*; use ruma::RoomVersionId::*;

View file

@ -137,7 +137,7 @@ pub(crate) async fn create_knock_event_v1_route(
if state_key != sender { if state_key != sender {
return Err!(Request(InvalidParam("state_key does not match sender user of event."))); return Err!(Request(InvalidParam("state_key does not match sender user of event.")));
}; }
let origin: OwnedServerName = serde_json::from_value( let origin: OwnedServerName = serde_json::from_value(
value value

View file

@ -241,7 +241,7 @@ impl<'a, 'de: 'a> de::Deserializer<'de> for &'a mut Deserializer<'de> {
| "Ignore" => self.record_ignore(), | "Ignore" => self.record_ignore(),
| "IgnoreAll" => self.record_ignore_all(), | "IgnoreAll" => self.record_ignore_all(),
| _ => unhandled!("Unrecognized deserialization Directive {name:?}"), | _ => unhandled!("Unrecognized deserialization Directive {name:?}"),
}; }
visitor.visit_unit() visitor.visit_unit()
} }

View file

@ -18,5 +18,5 @@ pub(crate) fn handle(level: LogLevel, msg: &str) {
| LogLevel::Error | LogLevel::Fatal => error!("{msg}"), | LogLevel::Error | LogLevel::Fatal => error!("{msg}"),
| LogLevel::Info => debug!("{msg}"), | LogLevel::Info => debug!("{msg}"),
| LogLevel::Warn => warn!("{msg}"), | LogLevel::Warn => warn!("{msg}"),
}; }
} }

View file

@ -52,7 +52,7 @@ pub fn compact_blocking(&self, opts: Options) -> Result {
co.set_target_level(level.try_into()?); co.set_target_level(level.try_into()?);
}, },
| (Some(_), Some(_)) => return Err!("compacting between specific levels not supported"), | (Some(_), Some(_)) => return Err!("compacting between specific levels not supported"),
}; }
self.db self.db
.db .db

View file

@ -50,7 +50,6 @@ where
.iter() .iter()
.map(ser::serialize_to::<KeyBuf, _>) .map(ser::serialize_to::<KeyBuf, _>)
.map(|result| result.expect("failed to serialize query key")) .map(|result| result.expect("failed to serialize query key"))
.map(Into::into)
.collect(); .collect();
self.db self.db

View file

@ -40,7 +40,7 @@ pub fn rev_raw_stream(self: &Arc<Self>) -> impl Stream<Item = Result<KeyVal<'_>>
.into_stream() .into_stream()
.flatten() .flatten()
.boxed(); .boxed();
}; }
let seek = Seek { let seek = Seek {
map: self.clone(), map: self.clone(),

View file

@ -89,7 +89,7 @@ where
.into_stream() .into_stream()
.flatten() .flatten()
.boxed(); .boxed();
}; }
let seek = Seek { let seek = Seek {
map: self.clone(), map: self.clone(),

View file

@ -39,7 +39,7 @@ pub fn raw_stream(self: &Arc<Self>) -> impl Stream<Item = Result<KeyVal<'_>>> +
.into_stream() .into_stream()
.flatten() .flatten()
.boxed(); .boxed();
}; }
let seek = Seek { let seek = Seek {
map: self.clone(), map: self.clone(),

View file

@ -86,7 +86,7 @@ where
.into_stream() .into_stream()
.flatten() .flatten()
.boxed(); .boxed();
}; }
let seek = Seek { let seek = Seek {
map: self.clone(), map: self.clone(),

View file

@ -146,11 +146,9 @@ pub(crate) fn close(&self) {
.map(JoinHandle::join) .map(JoinHandle::join)
.map(|result| result.map_err(Error::from_panic)) .map(|result| result.map_err(Error::from_panic))
.enumerate() .enumerate()
.for_each(|(id, result)| { .for_each(|(id, result)| match result {
match result {
| Ok(()) => trace!(?id, "worker joined"), | Ok(()) => trace!(?id, "worker joined"),
| Err(error) => error!(?id, "worker joined with error: {error}"), | Err(error) => error!(?id, "worker joined with error: {error}"),
};
}); });
} }
@ -345,7 +343,7 @@ fn worker_handle(self: &Arc<Self>, cmd: Cmd) {
| Cmd::Get(cmd) if cmd.key.len() == 1 => self.handle_get(cmd), | Cmd::Get(cmd) if cmd.key.len() == 1 => self.handle_get(cmd),
| Cmd::Get(cmd) => self.handle_batch(cmd), | Cmd::Get(cmd) => self.handle_batch(cmd),
| Cmd::Iter(cmd) => self.handle_iter(cmd), | Cmd::Iter(cmd) => self.handle_iter(cmd),
}; }
} }
#[implement(Pool)] #[implement(Pool)]
@ -362,7 +360,7 @@ fn handle_iter(&self, mut cmd: Seek) {
return; return;
} }
let from = cmd.key.as_deref().map(Into::into); let from = cmd.key.as_deref();
let result = match cmd.dir { let result = match cmd.dir {
| Direction::Forward => cmd.state.init_fwd(from), | Direction::Forward => cmd.state.init_fwd(from),
@ -394,7 +392,7 @@ fn handle_batch(self: &Arc<Self>, mut cmd: Get) {
return; return;
} }
let keys = cmd.key.iter().map(Into::into); let keys = cmd.key.iter();
let result: SmallVec<_> = cmd.map.get_batch_blocking(keys).collect(); let result: SmallVec<_> = cmd.map.get_batch_blocking(keys).collect();

View file

@ -224,7 +224,7 @@ impl<W: Write> ser::Serializer for &mut Serializer<'_, W> {
self.separator()?; self.separator()?;
}, },
| _ => unhandled!("Unrecognized serialization directive: {name:?}"), | _ => unhandled!("Unrecognized serialization directive: {name:?}"),
}; }
Ok(()) Ok(())
} }

View file

@ -113,13 +113,13 @@ impl<'a> State<'a> {
} }
#[inline] #[inline]
fn fetch_key(&self) -> Option<Key<'_>> { self.inner.key().map(Key::from) } fn fetch_key(&self) -> Option<Key<'_>> { self.inner.key() }
#[inline] #[inline]
fn _fetch_val(&self) -> Option<Val<'_>> { self.inner.value().map(Val::from) } fn _fetch_val(&self) -> Option<Val<'_>> { self.inner.value() }
#[inline] #[inline]
fn fetch(&self) -> Option<KeyVal<'_>> { self.inner.item().map(KeyVal::from) } fn fetch(&self) -> Option<KeyVal<'_>> { self.inner.item() }
#[inline] #[inline]
pub(super) fn status(&self) -> Option<rocksdb::Error> { self.inner.status().err() } pub(super) fn status(&self) -> Option<rocksdb::Error> { self.inner.status().err() }

View file

@ -53,6 +53,6 @@ impl Watchers {
tx.0.send(()).expect("channel should still be open"); tx.0.send(()).expect("channel should still be open");
} }
} }
}; }
} }
} }

View file

@ -205,7 +205,7 @@ fn get_default(field: &Field) -> Option<String> {
}, },
| Meta::Path { .. } => return Some("false".to_owned()), | Meta::Path { .. } => return Some("false".to_owned()),
| _ => return None, | _ => return None,
}; }
} }
None None

View file

@ -172,7 +172,7 @@ impl std::fmt::Display for BlurhashingError {
#[cfg(feature = "blurhashing")] #[cfg(feature = "blurhashing")]
| Self::ImageError(e) => | Self::ImageError(e) =>
write!(f, "There was an error with the image loading library => {e}")?, write!(f, "There was an error with the image loading library => {e}")?,
}; }
Ok(()) Ok(())
} }

View file

@ -283,7 +283,7 @@ async fn location_request(&self, location: &str) -> Result<FileMeta> {
.map_err(Into::into) .map_err(Into::into)
.map(|content| FileMeta { .map(|content| FileMeta {
content: Some(content), content: Some(content),
content_type: content_type.clone().map(Into::into), content_type: content_type.clone(),
content_disposition: Some(make_content_disposition( content_disposition: Some(make_content_disposition(
content_disposition.as_ref(), content_disposition.as_ref(),
content_type.as_deref(), content_type.as_deref(),

View file

@ -170,7 +170,7 @@ pub async fn handle_incoming_pdu<'a>(
| Entry::Occupied(mut e) => { | Entry::Occupied(mut e) => {
*e.get_mut() = (now, e.get().1.saturating_add(1)); *e.get_mut() = (now, e.get().1.saturating_add(1));
}, },
}; }
} }
} }

View file

@ -125,7 +125,7 @@ pub async fn get_summary_and_children_local(
SummaryAccessibility::Inaccessible SummaryAccessibility::Inaccessible
}, },
)), )),
}; }
let children_pdus: Vec<_> = self let children_pdus: Vec<_> = self
.get_stripped_space_child_events(current_room) .get_stripped_space_child_events(current_room)

View file

@ -218,7 +218,7 @@ impl Service {
) )
.await .await
.ok(); .ok();
}; }
// Copy direct chat flag // Copy direct chat flag
if let Ok(mut direct_event) = self if let Ok(mut direct_event) = self
@ -250,7 +250,7 @@ impl Service {
) )
.await?; .await?;
} }
}; }
} }
} }

View file

@ -303,7 +303,7 @@ impl Service {
}); });
return Ok(()); return Ok(());
}; }
// Else we have two options. // Else we have two options.
// 1. We add the current diff on top of the parent layer. // 1. We add the current diff on top of the parent layer.
@ -419,7 +419,7 @@ impl Service {
2, // every state change is 2 event changes on average 2, // every state change is 2 event changes on average
states_parents, states_parents,
)?; )?;
}; }
Ok(HashSetCompressStateEvent { Ok(HashSetCompressStateEvent {
shortstatehash: new_shortstatehash, shortstatehash: new_shortstatehash,

View file

@ -422,7 +422,7 @@ impl Service {
highlight = true; highlight = true;
}, },
| _ => {}, | _ => {},
}; }
// Break early if both conditions are true // Break early if both conditions are true
if notify && highlight { if notify && highlight {
@ -484,7 +484,7 @@ impl Service {
} }
} }
}, },
}; }
}, },
| TimelineEventType::SpaceChild => | TimelineEventType::SpaceChild =>
if let Some(_state_key) = &pdu.state_key { if let Some(_state_key) = &pdu.state_key {
@ -776,7 +776,7 @@ impl Service {
| _ => { | _ => {
pdu_json.remove("event_id"); pdu_json.remove("event_id");
}, },
}; }
// Add origin because synapse likes that (and it's required in the spec) // Add origin because synapse likes that (and it's required in the spec)
pdu_json.insert( pdu_json.insert(
@ -847,7 +847,7 @@ impl Service {
{ {
return Err!(Request(Forbidden("User cannot redact this event."))); return Err!(Request(Forbidden("User cannot redact this event.")));
} }
}; }
}, },
| _ => { | _ => {
let content: RoomRedactionEventContent = pdu.get_content()?; let content: RoomRedactionEventContent = pdu.get_content()?;
@ -863,7 +863,7 @@ impl Service {
} }
}, },
} }
}; }
if pdu.kind == TimelineEventType::RoomMember { if pdu.kind == TimelineEventType::RoomMember {
let content: RoomMemberEventContent = pdu.get_content()?; let content: RoomMemberEventContent = pdu.get_content()?;
@ -1293,10 +1293,10 @@ async fn check_pdu_for_admin_room(&self, pdu: &PduEvent, sender: &UserId) -> Res
} }
}, },
| _ => {}, | _ => {},
}; }
}, },
| _ => {}, | _ => {},
}; }
Ok(()) Ok(())
} }

View file

@ -131,7 +131,7 @@ impl crate::Service for Service {
| Err(error) => { | Err(error) => {
error!(id = ?error.id(), ?error, "sender worker finished"); error!(id = ?error.id(), ?error, "sender worker finished");
}, },
}; }
} }
Ok(()) Ok(())

View file

@ -138,7 +138,7 @@ impl Service {
match response { match response {
| Ok(dest) => self.handle_response_ok(&dest, futures, statuses).await, | Ok(dest) => self.handle_response_ok(&dest, futures, statuses).await,
| Err((dest, e)) => Self::handle_response_err(dest, statuses, &e), | Err((dest, e)) => Self::handle_response_err(dest, statuses, &e),
}; }
} }
fn handle_response_err(dest: Destination, statuses: &mut CurTransactionStatus, e: &Error) { fn handle_response_err(dest: Destination, statuses: &mut CurTransactionStatus, e: &Error) {
@ -319,10 +319,7 @@ impl Service {
if let Destination::Federation(server_name) = dest { if let Destination::Federation(server_name) = dest {
if let Ok((select_edus, last_count)) = self.select_edus(server_name).await { if let Ok((select_edus, last_count)) = self.select_edus(server_name).await {
debug_assert!(select_edus.len() <= EDU_LIMIT, "exceeded edus limit"); debug_assert!(select_edus.len() <= EDU_LIMIT, "exceeded edus limit");
let select_edus = select_edus let select_edus = select_edus.into_iter().map(SendingEvent::Edu);
.into_iter()
.map(Into::into)
.map(SendingEvent::Edu);
events.extend(select_edus); events.extend(select_edus);
self.db.set_latest_educount(server_name, last_count); self.db.set_latest_educount(server_name, last_count);

View file

@ -43,7 +43,7 @@ where
.keys() .keys()
.rev() .rev()
.take(self.services.server.config.trusted_server_batch_size) .take(self.services.server.config.trusted_server_batch_size)
.last() .next_back()
.cloned() .cloned()
{ {
let request = Request { let request = Request {

View file

@ -69,7 +69,7 @@ pub async fn read_tokens(&self) -> Result<HashSet<String>> {
}, },
| Err(e) => error!("Failed to read the registration token file: {e}"), | Err(e) => error!("Failed to read the registration token file: {e}"),
} }
}; }
if let Some(token) = &self.services.config.registration_token { if let Some(token) = &self.services.config.registration_token {
tokens.insert(token.to_owned()); tokens.insert(token.to_owned());
} }