540 lines
18 KiB
Rust
540 lines
18 KiB
Rust
mod data;
|
|
|
|
use std::{
|
|
collections::{BTreeMap, BTreeSet},
|
|
mem,
|
|
sync::{Arc, Mutex, Mutex as StdMutex},
|
|
};
|
|
|
|
use conduit::{Error, Result};
|
|
use ruma::{
|
|
api::client::{
|
|
device::Device,
|
|
filter::FilterDefinition,
|
|
sync::sync_events::{
|
|
self,
|
|
v4::{ExtensionsConfig, SyncRequestList},
|
|
},
|
|
},
|
|
encryption::{CrossSigningKey, DeviceKeys, OneTimeKey},
|
|
events::AnyToDeviceEvent,
|
|
serde::Raw,
|
|
DeviceId, DeviceKeyAlgorithm, DeviceKeyId, OwnedDeviceId, OwnedDeviceKeyId, OwnedMxcUri, OwnedRoomId, OwnedUserId,
|
|
UInt, UserId,
|
|
};
|
|
|
|
use self::data::Data;
|
|
use crate::{admin, rooms, Dep};
|
|
|
|
pub struct Service {
|
|
connections: DbConnections,
|
|
pub db: Data,
|
|
services: Services,
|
|
}
|
|
|
|
struct Services {
|
|
admin: Dep<admin::Service>,
|
|
state_cache: Dep<rooms::state_cache::Service>,
|
|
}
|
|
|
|
impl crate::Service for Service {
|
|
fn build(args: crate::Args<'_>) -> Result<Arc<Self>> {
|
|
Ok(Arc::new(Self {
|
|
connections: StdMutex::new(BTreeMap::new()),
|
|
db: Data::new(&args),
|
|
services: Services {
|
|
admin: args.depend::<admin::Service>("admin"),
|
|
state_cache: args.depend::<rooms::state_cache::Service>("rooms::state_cache"),
|
|
},
|
|
}))
|
|
}
|
|
|
|
fn name(&self) -> &str { crate::service::make_name(std::module_path!()) }
|
|
}
|
|
|
|
type DbConnections = Mutex<BTreeMap<DbConnectionsKey, DbConnectionsVal>>;
|
|
type DbConnectionsKey = (OwnedUserId, OwnedDeviceId, String);
|
|
type DbConnectionsVal = Arc<Mutex<SlidingSyncCache>>;
|
|
|
|
struct SlidingSyncCache {
|
|
lists: BTreeMap<String, SyncRequestList>,
|
|
subscriptions: BTreeMap<OwnedRoomId, sync_events::v4::RoomSubscription>,
|
|
known_rooms: BTreeMap<String, BTreeMap<OwnedRoomId, u64>>, // For every room, the roomsince number
|
|
extensions: ExtensionsConfig,
|
|
}
|
|
|
|
impl Service {
|
|
/// Check if a user has an account on this homeserver.
|
|
#[inline]
|
|
pub fn exists(&self, user_id: &UserId) -> Result<bool> { self.db.exists(user_id) }
|
|
|
|
pub fn remembered(&self, user_id: OwnedUserId, device_id: OwnedDeviceId, conn_id: String) -> bool {
|
|
self.connections
|
|
.lock()
|
|
.unwrap()
|
|
.contains_key(&(user_id, device_id, conn_id))
|
|
}
|
|
|
|
pub fn forget_sync_request_connection(&self, user_id: OwnedUserId, device_id: OwnedDeviceId, conn_id: String) {
|
|
self.connections
|
|
.lock()
|
|
.unwrap()
|
|
.remove(&(user_id, device_id, conn_id));
|
|
}
|
|
|
|
pub fn update_sync_request_with_cache(
|
|
&self, user_id: OwnedUserId, device_id: OwnedDeviceId, request: &mut sync_events::v4::Request,
|
|
) -> BTreeMap<String, BTreeMap<OwnedRoomId, u64>> {
|
|
let Some(conn_id) = request.conn_id.clone() else {
|
|
return BTreeMap::new();
|
|
};
|
|
|
|
let mut cache = self.connections.lock().unwrap();
|
|
let cached = Arc::clone(
|
|
cache
|
|
.entry((user_id, device_id, conn_id))
|
|
.or_insert_with(|| {
|
|
Arc::new(Mutex::new(SlidingSyncCache {
|
|
lists: BTreeMap::new(),
|
|
subscriptions: BTreeMap::new(),
|
|
known_rooms: BTreeMap::new(),
|
|
extensions: ExtensionsConfig::default(),
|
|
}))
|
|
}),
|
|
);
|
|
let cached = &mut cached.lock().unwrap();
|
|
drop(cache);
|
|
|
|
for (list_id, list) in &mut request.lists {
|
|
if let Some(cached_list) = cached.lists.get(list_id) {
|
|
if list.sort.is_empty() {
|
|
list.sort.clone_from(&cached_list.sort);
|
|
};
|
|
if list.room_details.required_state.is_empty() {
|
|
list.room_details
|
|
.required_state
|
|
.clone_from(&cached_list.room_details.required_state);
|
|
};
|
|
list.room_details.timeline_limit = list
|
|
.room_details
|
|
.timeline_limit
|
|
.or(cached_list.room_details.timeline_limit);
|
|
list.include_old_rooms = list
|
|
.include_old_rooms
|
|
.clone()
|
|
.or_else(|| cached_list.include_old_rooms.clone());
|
|
match (&mut list.filters, cached_list.filters.clone()) {
|
|
(Some(list_filters), Some(cached_filters)) => {
|
|
list_filters.is_dm = list_filters.is_dm.or(cached_filters.is_dm);
|
|
if list_filters.spaces.is_empty() {
|
|
list_filters.spaces = cached_filters.spaces;
|
|
}
|
|
list_filters.is_encrypted = list_filters.is_encrypted.or(cached_filters.is_encrypted);
|
|
list_filters.is_invite = list_filters.is_invite.or(cached_filters.is_invite);
|
|
if list_filters.room_types.is_empty() {
|
|
list_filters.room_types = cached_filters.room_types;
|
|
}
|
|
if list_filters.not_room_types.is_empty() {
|
|
list_filters.not_room_types = cached_filters.not_room_types;
|
|
}
|
|
list_filters.room_name_like = list_filters
|
|
.room_name_like
|
|
.clone()
|
|
.or(cached_filters.room_name_like);
|
|
if list_filters.tags.is_empty() {
|
|
list_filters.tags = cached_filters.tags;
|
|
}
|
|
if list_filters.not_tags.is_empty() {
|
|
list_filters.not_tags = cached_filters.not_tags;
|
|
}
|
|
},
|
|
(_, Some(cached_filters)) => list.filters = Some(cached_filters),
|
|
(Some(list_filters), _) => list.filters = Some(list_filters.clone()),
|
|
(..) => {},
|
|
}
|
|
if list.bump_event_types.is_empty() {
|
|
list.bump_event_types
|
|
.clone_from(&cached_list.bump_event_types);
|
|
};
|
|
}
|
|
cached.lists.insert(list_id.clone(), list.clone());
|
|
}
|
|
|
|
cached
|
|
.subscriptions
|
|
.extend(request.room_subscriptions.clone());
|
|
request
|
|
.room_subscriptions
|
|
.extend(cached.subscriptions.clone());
|
|
|
|
request.extensions.e2ee.enabled = request
|
|
.extensions
|
|
.e2ee
|
|
.enabled
|
|
.or(cached.extensions.e2ee.enabled);
|
|
|
|
request.extensions.to_device.enabled = request
|
|
.extensions
|
|
.to_device
|
|
.enabled
|
|
.or(cached.extensions.to_device.enabled);
|
|
|
|
request.extensions.account_data.enabled = request
|
|
.extensions
|
|
.account_data
|
|
.enabled
|
|
.or(cached.extensions.account_data.enabled);
|
|
request.extensions.account_data.lists = request
|
|
.extensions
|
|
.account_data
|
|
.lists
|
|
.clone()
|
|
.or_else(|| cached.extensions.account_data.lists.clone());
|
|
request.extensions.account_data.rooms = request
|
|
.extensions
|
|
.account_data
|
|
.rooms
|
|
.clone()
|
|
.or_else(|| cached.extensions.account_data.rooms.clone());
|
|
|
|
cached.extensions = request.extensions.clone();
|
|
|
|
cached.known_rooms.clone()
|
|
}
|
|
|
|
pub fn update_sync_subscriptions(
|
|
&self, user_id: OwnedUserId, device_id: OwnedDeviceId, conn_id: String,
|
|
subscriptions: BTreeMap<OwnedRoomId, sync_events::v4::RoomSubscription>,
|
|
) {
|
|
let mut cache = self.connections.lock().unwrap();
|
|
let cached = Arc::clone(
|
|
cache
|
|
.entry((user_id, device_id, conn_id))
|
|
.or_insert_with(|| {
|
|
Arc::new(Mutex::new(SlidingSyncCache {
|
|
lists: BTreeMap::new(),
|
|
subscriptions: BTreeMap::new(),
|
|
known_rooms: BTreeMap::new(),
|
|
extensions: ExtensionsConfig::default(),
|
|
}))
|
|
}),
|
|
);
|
|
let cached = &mut cached.lock().unwrap();
|
|
drop(cache);
|
|
|
|
cached.subscriptions = subscriptions;
|
|
}
|
|
|
|
pub fn update_sync_known_rooms(
|
|
&self, user_id: OwnedUserId, device_id: OwnedDeviceId, conn_id: String, list_id: String,
|
|
new_cached_rooms: BTreeSet<OwnedRoomId>, globalsince: u64,
|
|
) {
|
|
let mut cache = self.connections.lock().unwrap();
|
|
let cached = Arc::clone(
|
|
cache
|
|
.entry((user_id, device_id, conn_id))
|
|
.or_insert_with(|| {
|
|
Arc::new(Mutex::new(SlidingSyncCache {
|
|
lists: BTreeMap::new(),
|
|
subscriptions: BTreeMap::new(),
|
|
known_rooms: BTreeMap::new(),
|
|
extensions: ExtensionsConfig::default(),
|
|
}))
|
|
}),
|
|
);
|
|
let cached = &mut cached.lock().unwrap();
|
|
drop(cache);
|
|
|
|
for (roomid, lastsince) in cached
|
|
.known_rooms
|
|
.entry(list_id.clone())
|
|
.or_default()
|
|
.iter_mut()
|
|
{
|
|
if !new_cached_rooms.contains(roomid) {
|
|
*lastsince = 0;
|
|
}
|
|
}
|
|
let list = cached.known_rooms.entry(list_id).or_default();
|
|
for roomid in new_cached_rooms {
|
|
list.insert(roomid, globalsince);
|
|
}
|
|
}
|
|
|
|
/// Check if account is deactivated
|
|
pub fn is_deactivated(&self, user_id: &UserId) -> Result<bool> { self.db.is_deactivated(user_id) }
|
|
|
|
/// Check if a user is an admin
|
|
pub fn is_admin(&self, user_id: &UserId) -> Result<bool> {
|
|
if let Some(admin_room_id) = self.services.admin.get_admin_room()? {
|
|
self.services.state_cache.is_joined(user_id, &admin_room_id)
|
|
} else {
|
|
Ok(false)
|
|
}
|
|
}
|
|
|
|
/// Create a new user account on this homeserver.
|
|
#[inline]
|
|
pub fn create(&self, user_id: &UserId, password: Option<&str>) -> Result<()> {
|
|
self.db.set_password(user_id, password)?;
|
|
Ok(())
|
|
}
|
|
|
|
/// Returns the number of users registered on this server.
|
|
#[inline]
|
|
pub fn count(&self) -> Result<usize> { self.db.count() }
|
|
|
|
/// Find out which user an access token belongs to.
|
|
pub fn find_from_token(&self, token: &str) -> Result<Option<(OwnedUserId, String)>> {
|
|
self.db.find_from_token(token)
|
|
}
|
|
|
|
/// Returns an iterator over all users on this homeserver.
|
|
pub fn iter(&self) -> impl Iterator<Item = Result<OwnedUserId>> + '_ { self.db.iter() }
|
|
|
|
/// Returns a list of local users as list of usernames.
|
|
///
|
|
/// A user account is considered `local` if the length of it's password is
|
|
/// greater then zero.
|
|
pub fn list_local_users(&self) -> Result<Vec<String>> { self.db.list_local_users() }
|
|
|
|
/// Returns the password hash for the given user.
|
|
pub fn password_hash(&self, user_id: &UserId) -> Result<Option<String>> { self.db.password_hash(user_id) }
|
|
|
|
/// Hash and set the user's password to the Argon2 hash
|
|
#[inline]
|
|
pub fn set_password(&self, user_id: &UserId, password: Option<&str>) -> Result<()> {
|
|
self.db.set_password(user_id, password)
|
|
}
|
|
|
|
/// Returns the displayname of a user on this homeserver.
|
|
pub fn displayname(&self, user_id: &UserId) -> Result<Option<String>> { self.db.displayname(user_id) }
|
|
|
|
/// Sets a new displayname or removes it if displayname is None. You still
|
|
/// need to nofify all rooms of this change.
|
|
pub async fn set_displayname(&self, user_id: &UserId, displayname: Option<String>) -> Result<()> {
|
|
self.db.set_displayname(user_id, displayname)
|
|
}
|
|
|
|
/// Get the avatar_url of a user.
|
|
pub fn avatar_url(&self, user_id: &UserId) -> Result<Option<OwnedMxcUri>> { self.db.avatar_url(user_id) }
|
|
|
|
/// Sets a new avatar_url or removes it if avatar_url is None.
|
|
pub async fn set_avatar_url(&self, user_id: &UserId, avatar_url: Option<OwnedMxcUri>) -> Result<()> {
|
|
self.db.set_avatar_url(user_id, avatar_url)
|
|
}
|
|
|
|
/// Get the blurhash of a user.
|
|
pub fn blurhash(&self, user_id: &UserId) -> Result<Option<String>> { self.db.blurhash(user_id) }
|
|
|
|
/// Sets a new blurhash or removes it if blurhash is None.
|
|
pub async fn set_blurhash(&self, user_id: &UserId, blurhash: Option<String>) -> Result<()> {
|
|
self.db.set_blurhash(user_id, blurhash)
|
|
}
|
|
|
|
/// Adds a new device to a user.
|
|
pub fn create_device(
|
|
&self, user_id: &UserId, device_id: &DeviceId, token: &str, initial_device_display_name: Option<String>,
|
|
client_ip: Option<String>,
|
|
) -> Result<()> {
|
|
self.db
|
|
.create_device(user_id, device_id, token, initial_device_display_name, client_ip)
|
|
}
|
|
|
|
/// Removes a device from a user.
|
|
pub fn remove_device(&self, user_id: &UserId, device_id: &DeviceId) -> Result<()> {
|
|
self.db.remove_device(user_id, device_id)
|
|
}
|
|
|
|
/// Returns an iterator over all device ids of this user.
|
|
pub fn all_device_ids<'a>(&'a self, user_id: &UserId) -> impl Iterator<Item = Result<OwnedDeviceId>> + 'a {
|
|
self.db.all_device_ids(user_id)
|
|
}
|
|
|
|
/// Replaces the access token of one device.
|
|
#[inline]
|
|
pub fn set_token(&self, user_id: &UserId, device_id: &DeviceId, token: &str) -> Result<()> {
|
|
self.db.set_token(user_id, device_id, token)
|
|
}
|
|
|
|
pub fn add_one_time_key(
|
|
&self, user_id: &UserId, device_id: &DeviceId, one_time_key_key: &DeviceKeyId,
|
|
one_time_key_value: &Raw<OneTimeKey>,
|
|
) -> Result<()> {
|
|
self.db
|
|
.add_one_time_key(user_id, device_id, one_time_key_key, one_time_key_value)
|
|
}
|
|
|
|
// TODO: use this ?
|
|
#[allow(dead_code)]
|
|
pub fn last_one_time_keys_update(&self, user_id: &UserId) -> Result<u64> {
|
|
self.db.last_one_time_keys_update(user_id)
|
|
}
|
|
|
|
pub fn take_one_time_key(
|
|
&self, user_id: &UserId, device_id: &DeviceId, key_algorithm: &DeviceKeyAlgorithm,
|
|
) -> Result<Option<(OwnedDeviceKeyId, Raw<OneTimeKey>)>> {
|
|
self.db.take_one_time_key(user_id, device_id, key_algorithm)
|
|
}
|
|
|
|
pub fn count_one_time_keys(
|
|
&self, user_id: &UserId, device_id: &DeviceId,
|
|
) -> Result<BTreeMap<DeviceKeyAlgorithm, UInt>> {
|
|
self.db.count_one_time_keys(user_id, device_id)
|
|
}
|
|
|
|
pub fn add_device_keys(&self, user_id: &UserId, device_id: &DeviceId, device_keys: &Raw<DeviceKeys>) -> Result<()> {
|
|
self.db.add_device_keys(user_id, device_id, device_keys)
|
|
}
|
|
|
|
pub fn add_cross_signing_keys(
|
|
&self, user_id: &UserId, master_key: &Raw<CrossSigningKey>, self_signing_key: &Option<Raw<CrossSigningKey>>,
|
|
user_signing_key: &Option<Raw<CrossSigningKey>>, notify: bool,
|
|
) -> Result<()> {
|
|
self.db
|
|
.add_cross_signing_keys(user_id, master_key, self_signing_key, user_signing_key, notify)
|
|
}
|
|
|
|
pub fn sign_key(
|
|
&self, target_id: &UserId, key_id: &str, signature: (String, String), sender_id: &UserId,
|
|
) -> Result<()> {
|
|
self.db.sign_key(target_id, key_id, signature, sender_id)
|
|
}
|
|
|
|
pub fn keys_changed<'a>(
|
|
&'a self, user_or_room_id: &str, from: u64, to: Option<u64>,
|
|
) -> impl Iterator<Item = Result<OwnedUserId>> + 'a {
|
|
self.db.keys_changed(user_or_room_id, from, to)
|
|
}
|
|
|
|
#[inline]
|
|
pub fn mark_device_key_update(&self, user_id: &UserId) -> Result<()> { self.db.mark_device_key_update(user_id) }
|
|
|
|
pub fn get_device_keys(&self, user_id: &UserId, device_id: &DeviceId) -> Result<Option<Raw<DeviceKeys>>> {
|
|
self.db.get_device_keys(user_id, device_id)
|
|
}
|
|
|
|
#[inline]
|
|
pub fn parse_master_key(
|
|
&self, user_id: &UserId, master_key: &Raw<CrossSigningKey>,
|
|
) -> Result<(Vec<u8>, CrossSigningKey)> {
|
|
Data::parse_master_key(user_id, master_key)
|
|
}
|
|
|
|
#[inline]
|
|
pub fn get_key(
|
|
&self, key: &[u8], sender_user: Option<&UserId>, user_id: &UserId, allowed_signatures: &dyn Fn(&UserId) -> bool,
|
|
) -> Result<Option<Raw<CrossSigningKey>>> {
|
|
self.db
|
|
.get_key(key, sender_user, user_id, allowed_signatures)
|
|
}
|
|
|
|
pub fn get_master_key(
|
|
&self, sender_user: Option<&UserId>, user_id: &UserId, allowed_signatures: &dyn Fn(&UserId) -> bool,
|
|
) -> Result<Option<Raw<CrossSigningKey>>> {
|
|
self.db
|
|
.get_master_key(sender_user, user_id, allowed_signatures)
|
|
}
|
|
|
|
pub fn get_self_signing_key(
|
|
&self, sender_user: Option<&UserId>, user_id: &UserId, allowed_signatures: &dyn Fn(&UserId) -> bool,
|
|
) -> Result<Option<Raw<CrossSigningKey>>> {
|
|
self.db
|
|
.get_self_signing_key(sender_user, user_id, allowed_signatures)
|
|
}
|
|
|
|
pub fn get_user_signing_key(&self, user_id: &UserId) -> Result<Option<Raw<CrossSigningKey>>> {
|
|
self.db.get_user_signing_key(user_id)
|
|
}
|
|
|
|
pub fn add_to_device_event(
|
|
&self, sender: &UserId, target_user_id: &UserId, target_device_id: &DeviceId, event_type: &str,
|
|
content: serde_json::Value,
|
|
) -> Result<()> {
|
|
self.db
|
|
.add_to_device_event(sender, target_user_id, target_device_id, event_type, content)
|
|
}
|
|
|
|
pub fn get_to_device_events(&self, user_id: &UserId, device_id: &DeviceId) -> Result<Vec<Raw<AnyToDeviceEvent>>> {
|
|
self.db.get_to_device_events(user_id, device_id)
|
|
}
|
|
|
|
pub fn remove_to_device_events(&self, user_id: &UserId, device_id: &DeviceId, until: u64) -> Result<()> {
|
|
self.db.remove_to_device_events(user_id, device_id, until)
|
|
}
|
|
|
|
pub fn update_device_metadata(&self, user_id: &UserId, device_id: &DeviceId, device: &Device) -> Result<()> {
|
|
self.db.update_device_metadata(user_id, device_id, device)
|
|
}
|
|
|
|
/// Get device metadata.
|
|
pub fn get_device_metadata(&self, user_id: &UserId, device_id: &DeviceId) -> Result<Option<Device>> {
|
|
self.db.get_device_metadata(user_id, device_id)
|
|
}
|
|
|
|
pub fn get_devicelist_version(&self, user_id: &UserId) -> Result<Option<u64>> {
|
|
self.db.get_devicelist_version(user_id)
|
|
}
|
|
|
|
pub fn all_devices_metadata<'a>(&'a self, user_id: &UserId) -> impl Iterator<Item = Result<Device>> + 'a {
|
|
self.db.all_devices_metadata(user_id)
|
|
}
|
|
|
|
/// Deactivate account
|
|
pub fn deactivate_account(&self, user_id: &UserId) -> Result<()> {
|
|
// Remove all associated devices
|
|
for device_id in self.all_device_ids(user_id) {
|
|
self.remove_device(user_id, &device_id?)?;
|
|
}
|
|
|
|
// Set the password to "" to indicate a deactivated account. Hashes will never
|
|
// result in an empty string, so the user will not be able to log in again.
|
|
// Systems like changing the password without logging in should check if the
|
|
// account is deactivated.
|
|
self.db.set_password(user_id, None)?;
|
|
|
|
// TODO: Unhook 3PID
|
|
Ok(())
|
|
}
|
|
|
|
/// Creates a new sync filter. Returns the filter id.
|
|
pub fn create_filter(&self, user_id: &UserId, filter: &FilterDefinition) -> Result<String> {
|
|
self.db.create_filter(user_id, filter)
|
|
}
|
|
|
|
pub fn get_filter(&self, user_id: &UserId, filter_id: &str) -> Result<Option<FilterDefinition>> {
|
|
self.db.get_filter(user_id, filter_id)
|
|
}
|
|
|
|
/// Creates an OpenID token, which can be used to prove that a user has
|
|
/// access to an account (primarily for integrations)
|
|
pub fn create_openid_token(&self, user_id: &UserId, token: &str) -> Result<u64> {
|
|
self.db.create_openid_token(user_id, token)
|
|
}
|
|
|
|
/// Find out which user an OpenID access token belongs to.
|
|
pub fn find_from_openid_token(&self, token: &str) -> Result<OwnedUserId> { self.db.find_from_openid_token(token) }
|
|
}
|
|
|
|
/// Ensure that a user only sees signatures from themselves and the target user
|
|
pub fn clean_signatures<F: Fn(&UserId) -> bool>(
|
|
cross_signing_key: &mut serde_json::Value, sender_user: Option<&UserId>, user_id: &UserId, allowed_signatures: F,
|
|
) -> Result<(), Error> {
|
|
if let Some(signatures) = cross_signing_key
|
|
.get_mut("signatures")
|
|
.and_then(|v| v.as_object_mut())
|
|
{
|
|
// Don't allocate for the full size of the current signatures, but require
|
|
// at most one resize if nothing is dropped
|
|
let new_capacity = signatures.len() / 2;
|
|
for (user, signature) in mem::replace(signatures, serde_json::Map::with_capacity(new_capacity)) {
|
|
let sid =
|
|
<&UserId>::try_from(user.as_str()).map_err(|_| Error::bad_database("Invalid user ID in database."))?;
|
|
if sender_user == Some(user_id) || sid == user_id || allowed_signatures(sid) {
|
|
signatures.insert(user, signature);
|
|
}
|
|
}
|
|
}
|
|
|
|
Ok(())
|
|
}
|