248 lines
6.4 KiB
Rust
248 lines
6.4 KiB
Rust
mod aggregate;
|
|
mod data;
|
|
// Write/update pipeline lives in pipeline.rs.
|
|
mod pipeline;
|
|
mod presence;
|
|
|
|
use std::{collections::HashMap, sync::Arc, time::Duration};
|
|
|
|
use async_trait::async_trait;
|
|
use futures::{Stream, StreamExt, TryFutureExt, future::{AbortHandle, Abortable}, stream::FuturesUnordered};
|
|
use loole::{Receiver, Sender};
|
|
use ruma::{
|
|
OwnedUserId, UserId, events::presence::PresenceEvent, presence::PresenceState,
|
|
};
|
|
use tokio::sync::RwLock;
|
|
use tuwunel_core::{
|
|
Result, checked, debug, debug_warn,
|
|
result::LogErr,
|
|
trace,
|
|
};
|
|
|
|
use self::{aggregate::PresenceAggregator, data::Data, presence::Presence};
|
|
|
|
pub struct Service {
|
|
timer_channel: (Sender<TimerType>, Receiver<TimerType>),
|
|
timeout_remote_users: bool,
|
|
idle_timeout: u64,
|
|
offline_timeout: u64,
|
|
db: Data,
|
|
services: Arc<crate::services::OnceServices>,
|
|
last_sync_seen: RwLock<HashMap<OwnedUserId, u64>>,
|
|
device_presence: PresenceAggregator,
|
|
}
|
|
|
|
type TimerType = (OwnedUserId, Duration, u64);
|
|
type TimerFired = (OwnedUserId, u64);
|
|
|
|
#[async_trait]
|
|
impl crate::Service for Service {
|
|
fn build(args: &crate::Args<'_>) -> Result<Arc<Self>> {
|
|
let config = &args.server.config;
|
|
let idle_timeout_s = config.presence_idle_timeout_s;
|
|
let offline_timeout_s = config.presence_offline_timeout_s;
|
|
Ok(Arc::new(Self {
|
|
timer_channel: loole::unbounded(),
|
|
timeout_remote_users: config.presence_timeout_remote_users,
|
|
idle_timeout: checked!(idle_timeout_s * 1_000)?,
|
|
offline_timeout: checked!(offline_timeout_s * 1_000)?,
|
|
db: Data::new(args),
|
|
services: args.services.clone(),
|
|
last_sync_seen: RwLock::new(HashMap::new()),
|
|
device_presence: PresenceAggregator::new(),
|
|
}))
|
|
}
|
|
|
|
async fn worker(self: Arc<Self>) -> Result {
|
|
// reset dormant online/away statuses to offline, and set the server user as
|
|
// online
|
|
self.unset_all_presence().await;
|
|
self.device_presence.clear().await;
|
|
_ = self
|
|
.maybe_ping_presence(&self.services.globals.server_user, None, &PresenceState::Online)
|
|
.await;
|
|
|
|
let receiver = self.timer_channel.1.clone();
|
|
|
|
let mut presence_timers: FuturesUnordered<_> = FuturesUnordered::new();
|
|
let mut timer_handles: HashMap<OwnedUserId, (u64, AbortHandle)> = HashMap::new();
|
|
while !receiver.is_closed() {
|
|
tokio::select! {
|
|
Some(result) = presence_timers.next() => {
|
|
let Ok((user_id, count)) = result else {
|
|
continue;
|
|
};
|
|
|
|
if let Some((current_count, _)) = timer_handles.get(&user_id)
|
|
&& *current_count != count {
|
|
trace!(?user_id, count, current_count, "Skipping stale presence timer");
|
|
continue;
|
|
}
|
|
|
|
timer_handles.remove(&user_id);
|
|
self.process_presence_timer(&user_id, count).await.log_err().ok();
|
|
},
|
|
event = receiver.recv_async() => match event {
|
|
Err(_) => break,
|
|
Ok((user_id, timeout, count)) => {
|
|
debug!(
|
|
"Adding timer {}: {user_id} timeout:{timeout:?} count:{count}",
|
|
presence_timers.len()
|
|
);
|
|
if let Some((_, handle)) = timer_handles.remove(&user_id) {
|
|
handle.abort();
|
|
}
|
|
|
|
let (handle, reg) = AbortHandle::new_pair();
|
|
presence_timers.push(Abortable::new(
|
|
pipeline::presence_timer(user_id.clone(), timeout, count),
|
|
reg,
|
|
));
|
|
timer_handles.insert(user_id, (count, handle));
|
|
},
|
|
},
|
|
}
|
|
}
|
|
|
|
Ok(())
|
|
}
|
|
|
|
async fn interrupt(&self) {
|
|
// set the server user as offline
|
|
_ = self
|
|
.maybe_ping_presence(
|
|
&self.services.globals.server_user,
|
|
None,
|
|
&PresenceState::Offline,
|
|
)
|
|
.await;
|
|
|
|
let (timer_sender, _) = &self.timer_channel;
|
|
if !timer_sender.is_closed() {
|
|
timer_sender.close();
|
|
}
|
|
}
|
|
|
|
fn name(&self) -> &str { crate::service::make_name(std::module_path!()) }
|
|
}
|
|
|
|
impl Service {
|
|
/// record that a user has just successfully completed a /sync (or
|
|
/// equivalent activity)
|
|
pub async fn note_sync(&self, user_id: &UserId) {
|
|
if !self.services.config.suppress_push_when_active {
|
|
return;
|
|
}
|
|
|
|
let now = tuwunel_core::utils::millis_since_unix_epoch();
|
|
self.last_sync_seen
|
|
.write()
|
|
.await
|
|
.insert(user_id.to_owned(), now);
|
|
}
|
|
|
|
/// Returns milliseconds since last observed sync for user (if any)
|
|
pub async fn last_sync_gap_ms(&self, user_id: &UserId) -> Option<u64> {
|
|
let now = tuwunel_core::utils::millis_since_unix_epoch();
|
|
self.last_sync_seen
|
|
.read()
|
|
.await
|
|
.get(user_id)
|
|
.map(|ts| now.saturating_sub(*ts))
|
|
}
|
|
|
|
/// Returns the latest presence event for the given user.
|
|
pub async fn get_presence(&self, user_id: &UserId) -> Result<PresenceEvent> {
|
|
self.db
|
|
.get_presence(user_id)
|
|
.map_ok(|(_, presence)| presence)
|
|
.await
|
|
}
|
|
|
|
/// Removes the presence record for the given user from the database.
|
|
///
|
|
/// TODO: Why is this not used?
|
|
#[allow(dead_code)]
|
|
pub async fn remove_presence(&self, user_id: &UserId) {
|
|
self.db.remove_presence(user_id).await;
|
|
}
|
|
|
|
// Unset online/unavailable presence to offline on startup
|
|
async fn unset_all_presence(&self) {
|
|
if !self.services.server.config.allow_local_presence || self.services.db.is_read_only() {
|
|
return;
|
|
}
|
|
|
|
let _cork = self.services.db.cork();
|
|
|
|
for user_id in &self
|
|
.services
|
|
.users
|
|
.list_local_users()
|
|
.map(UserId::to_owned)
|
|
.collect::<Vec<_>>()
|
|
.await
|
|
{
|
|
let presence = self.db.get_presence(user_id).await;
|
|
|
|
let presence = match presence {
|
|
| Ok((_, ref presence)) => &presence.content,
|
|
| _ => continue,
|
|
};
|
|
|
|
if !matches!(
|
|
presence.presence,
|
|
PresenceState::Unavailable | PresenceState::Online | PresenceState::Busy
|
|
) {
|
|
trace!(?user_id, ?presence, "Skipping user");
|
|
continue;
|
|
}
|
|
|
|
trace!(?user_id, ?presence, "Resetting presence to offline");
|
|
|
|
_ = self
|
|
.set_presence(
|
|
user_id,
|
|
&PresenceState::Offline,
|
|
Some(false),
|
|
presence.last_active_ago,
|
|
presence.status_msg.clone(),
|
|
)
|
|
.await
|
|
.inspect_err(|e| {
|
|
debug_warn!(
|
|
?presence,
|
|
"{user_id} has invalid presence in database and failed to reset it to \
|
|
offline: {e}"
|
|
);
|
|
});
|
|
}
|
|
}
|
|
|
|
/// Returns the most recent presence updates that happened after the event
|
|
/// with id `since`.
|
|
pub fn presence_since(
|
|
&self,
|
|
since: u64,
|
|
to: Option<u64>,
|
|
) -> impl Stream<Item = (&UserId, u64, &[u8])> + Send + '_ {
|
|
self.db.presence_since(since, to)
|
|
}
|
|
|
|
#[inline]
|
|
pub async fn from_json_bytes_to_event(
|
|
&self,
|
|
bytes: &[u8],
|
|
user_id: &UserId,
|
|
) -> Result<PresenceEvent> {
|
|
let presence = Presence::from_json_bytes(bytes)?;
|
|
let event = presence
|
|
.to_presence_event(user_id, &self.services.users)
|
|
.await;
|
|
|
|
Ok(event)
|
|
}
|
|
}
|
|
|
|
// presence_timer lives in pipeline.rs alongside the timer handling logic.
|