Refactor sender's push destination handler. Combine remnants of service::rooms::user with pusher service. Further split and reorg pusher service units. Signed-off-by: Jason Volk <jason@zemos.net>
239 lines
5.8 KiB
Rust
239 lines
5.8 KiB
Rust
mod append;
|
|
mod notification;
|
|
mod request;
|
|
mod send;
|
|
|
|
use std::sync::Arc;
|
|
|
|
use futures::{Stream, StreamExt};
|
|
use ipaddress::IPAddress;
|
|
use ruma::{
|
|
DeviceId, OwnedDeviceId, RoomId, UserId,
|
|
api::client::push::{Pusher, PusherKind, set_pusher},
|
|
events::{AnySyncTimelineEvent, room::power_levels::RoomPowerLevels},
|
|
push::{Action, PushConditionPowerLevelsCtx, PushConditionRoomCtx, Ruleset},
|
|
serde::Raw,
|
|
uint,
|
|
};
|
|
use tuwunel_core::{
|
|
Err, Result, err, implement,
|
|
utils::stream::{BroadbandExt, TryIgnore},
|
|
};
|
|
use tuwunel_database::{Database, Deserialized, Ignore, Interfix, Json, Map};
|
|
|
|
pub struct Service {
|
|
db: Data,
|
|
services: Arc<crate::services::OnceServices>,
|
|
}
|
|
|
|
struct Data {
|
|
senderkey_pusher: Arc<Map>,
|
|
pushkey_deviceid: Arc<Map>,
|
|
userroomid_highlightcount: Arc<Map>,
|
|
userroomid_notificationcount: Arc<Map>,
|
|
roomuserid_lastnotificationread: Arc<Map>,
|
|
db: Arc<Database>,
|
|
}
|
|
|
|
impl crate::Service for Service {
|
|
fn build(args: &crate::Args<'_>) -> Result<Arc<Self>> {
|
|
Ok(Arc::new(Self {
|
|
db: Data {
|
|
senderkey_pusher: args.db["senderkey_pusher"].clone(),
|
|
pushkey_deviceid: args.db["pushkey_deviceid"].clone(),
|
|
userroomid_highlightcount: args.db["userroomid_highlightcount"].clone(),
|
|
userroomid_notificationcount: args.db["userroomid_notificationcount"].clone(),
|
|
roomuserid_lastnotificationread: args.db["roomuserid_lastnotificationread"]
|
|
.clone(),
|
|
db: args.db.clone(),
|
|
},
|
|
services: args.services.clone(),
|
|
}))
|
|
}
|
|
|
|
fn name(&self) -> &str { crate::service::make_name(std::module_path!()) }
|
|
}
|
|
|
|
#[implement(Service)]
|
|
pub async fn set_pusher(
|
|
&self,
|
|
sender: &UserId,
|
|
sender_device: &DeviceId,
|
|
pusher: &set_pusher::v3::PusherAction,
|
|
) -> Result {
|
|
match pusher {
|
|
| set_pusher::v3::PusherAction::Post(data) => {
|
|
let pushkey = data.pusher.ids.pushkey.as_str();
|
|
|
|
if pushkey.len() > 512 {
|
|
return Err!(Request(InvalidParam(
|
|
"Push key length cannot be greater than 512 bytes."
|
|
)));
|
|
}
|
|
|
|
if data.pusher.ids.app_id.as_str().len() > 64 {
|
|
return Err!(Request(InvalidParam(
|
|
"App ID length cannot be greater than 64 bytes."
|
|
)));
|
|
}
|
|
|
|
// add some validation to the pusher URL
|
|
let pusher_kind = &data.pusher.kind;
|
|
if let PusherKind::Http(http) = pusher_kind {
|
|
let url = &http.url;
|
|
let url = url::Url::parse(&http.url).map_err(|e| {
|
|
err!(Request(InvalidParam(
|
|
warn!(%url, "HTTP pusher URL is not a valid URL: {e}")
|
|
)))
|
|
})?;
|
|
|
|
if ["http", "https"]
|
|
.iter()
|
|
.all(|&scheme| scheme != url.scheme().to_lowercase())
|
|
{
|
|
return Err!(Request(InvalidParam(
|
|
warn!(%url, "HTTP pusher URL is not a valid HTTP/HTTPS URL")
|
|
)));
|
|
}
|
|
|
|
if let Ok(ip) =
|
|
IPAddress::parse(url.host_str().expect("URL previously validated"))
|
|
{
|
|
if !self.services.client.valid_cidr_range(&ip) {
|
|
return Err!(Request(InvalidParam(
|
|
warn!(%url, "HTTP pusher URL is a forbidden remote address")
|
|
)));
|
|
}
|
|
}
|
|
}
|
|
|
|
let pushkey = data.pusher.ids.pushkey.as_str();
|
|
let key = (sender, pushkey);
|
|
self.db.senderkey_pusher.put(key, Json(pusher));
|
|
self.db
|
|
.pushkey_deviceid
|
|
.insert(pushkey, sender_device);
|
|
},
|
|
| set_pusher::v3::PusherAction::Delete(ids) => {
|
|
self.delete_pusher(sender, ids.pushkey.as_str())
|
|
.await;
|
|
},
|
|
}
|
|
|
|
Ok(())
|
|
}
|
|
|
|
#[implement(Service)]
|
|
pub async fn delete_pusher(&self, sender: &UserId, pushkey: &str) {
|
|
let key = (sender, pushkey);
|
|
self.db.senderkey_pusher.del(key);
|
|
self.db.pushkey_deviceid.remove(pushkey);
|
|
|
|
self.services
|
|
.sending
|
|
.cleanup_events(None, Some(sender), Some(pushkey))
|
|
.await
|
|
.ok();
|
|
}
|
|
|
|
#[implement(Service)]
|
|
pub async fn get_device_pushkeys(&self, sender: &UserId, device_id: &DeviceId) -> Vec<String> {
|
|
self.get_pushkeys(sender)
|
|
.map(ToOwned::to_owned)
|
|
.broad_filter_map(async |pushkey| {
|
|
self.get_pusher_device(&pushkey)
|
|
.await
|
|
.ok()
|
|
.filter(|pusher_device| pusher_device == device_id)
|
|
.is_some()
|
|
.then_some(pushkey)
|
|
})
|
|
.collect()
|
|
.await
|
|
}
|
|
|
|
#[implement(Service)]
|
|
pub async fn get_pusher_device(&self, pushkey: &str) -> Result<OwnedDeviceId> {
|
|
self.db
|
|
.pushkey_deviceid
|
|
.get(pushkey)
|
|
.await
|
|
.deserialized()
|
|
}
|
|
|
|
#[implement(Service)]
|
|
pub async fn get_pusher(&self, sender: &UserId, pushkey: &str) -> Result<Pusher> {
|
|
let senderkey = (sender, pushkey);
|
|
self.db
|
|
.senderkey_pusher
|
|
.qry(&senderkey)
|
|
.await
|
|
.deserialized()
|
|
}
|
|
|
|
#[implement(Service)]
|
|
pub async fn get_pushers(&self, sender: &UserId) -> Vec<Pusher> {
|
|
let prefix = (sender, Interfix);
|
|
self.db
|
|
.senderkey_pusher
|
|
.stream_prefix(&prefix)
|
|
.ignore_err()
|
|
.map(|(_, pusher): (Ignore, Pusher)| pusher)
|
|
.collect()
|
|
.await
|
|
}
|
|
|
|
#[implement(Service)]
|
|
pub fn get_pushkeys<'a>(&'a self, sender: &'a UserId) -> impl Stream<Item = &str> + Send + 'a {
|
|
let prefix = (sender, Interfix);
|
|
self.db
|
|
.senderkey_pusher
|
|
.keys_prefix(&prefix)
|
|
.ignore_err()
|
|
.map(|(_, pushkey): (Ignore, &str)| pushkey)
|
|
}
|
|
|
|
#[implement(Service)]
|
|
#[tracing::instrument(level = "debug", skip_all)]
|
|
pub async fn get_actions<'a>(
|
|
&self,
|
|
user: &UserId,
|
|
ruleset: &'a Ruleset,
|
|
power_levels: &RoomPowerLevels,
|
|
pdu: &Raw<AnySyncTimelineEvent>,
|
|
room_id: &RoomId,
|
|
) -> &'a [Action] {
|
|
let power_levels = PushConditionPowerLevelsCtx {
|
|
users: power_levels.users.clone(),
|
|
users_default: power_levels.users_default,
|
|
notifications: power_levels.notifications.clone(),
|
|
rules: power_levels.rules.clone(),
|
|
};
|
|
|
|
let room_joined_count = self
|
|
.services
|
|
.state_cache
|
|
.room_joined_count(room_id)
|
|
.await
|
|
.unwrap_or(1)
|
|
.try_into()
|
|
.unwrap_or_else(|_| uint!(0));
|
|
|
|
let user_display_name = self
|
|
.services
|
|
.users
|
|
.displayname(user)
|
|
.await
|
|
.unwrap_or_else(|_| user.localpart().to_owned());
|
|
|
|
let ctx = PushConditionRoomCtx {
|
|
room_id: room_id.to_owned(),
|
|
member_count: room_joined_count,
|
|
user_id: user.to_owned(),
|
|
user_display_name,
|
|
power_levels: Some(power_levels),
|
|
};
|
|
|
|
ruleset.get_actions(pdu, &ctx).await
|
|
}
|