combine service/users data w/ mod unit split sliding sync related out of service/users instrument database entry points remove increment crap from database interface de-wrap all database get() calls de-wrap all database insert() calls de-wrap all database remove() calls refactor database interface for async streaming add query key serializer for database implement Debug for result handle add query deserializer for database add deserialization trait for option handle start a stream utils suite de-wrap/asyncify/type-query count_one_time_keys() de-wrap/asyncify users count add admin query users command suite de-wrap/asyncify users exists de-wrap/partially asyncify user filter related asyncify/de-wrap users device/keys related asyncify/de-wrap user auth/misc related asyncify/de-wrap users blurhash asyncify/de-wrap account_data get; merge Data into Service partial asyncify/de-wrap uiaa; merge Data into Service partially asyncify/de-wrap transaction_ids get; merge Data into Service partially asyncify/de-wrap key_backups; merge Data into Service asyncify/de-wrap pusher service getters; merge Data into Service asyncify/de-wrap rooms alias getters/some iterators asyncify/de-wrap rooms directory getters/iterator partially asyncify/de-wrap rooms lazy-loading partially asyncify/de-wrap rooms metadata asyncify/dewrap rooms outlier asyncify/dewrap rooms pdu_metadata dewrap/partially asyncify rooms read receipt de-wrap rooms search service de-wrap/partially asyncify rooms user service partial de-wrap rooms state_compressor de-wrap rooms state_cache de-wrap room state et al de-wrap rooms timeline service additional users device/keys related de-wrap/asyncify sender asyncify services refactor database to TryFuture/TryStream refactor services for TryFuture/TryStream asyncify api handlers additional asyncification for admin module abstract stream related; support reverse streams additional stream conversions asyncify state-res related Signed-off-by: Jason Volk <jason@zemos.net>
656 lines
20 KiB
Rust
656 lines
20 KiB
Rust
use std::{
|
|
cmp,
|
|
collections::{BTreeMap, HashMap, HashSet},
|
|
fmt::Debug,
|
|
time::{Duration, Instant},
|
|
};
|
|
|
|
use base64::{engine::general_purpose, Engine as _};
|
|
use conduit::{
|
|
debug, debug_warn, err, trace,
|
|
utils::{calculate_hash, math::continue_exponential_backoff_secs, ReadyExt},
|
|
warn, Error, Result,
|
|
};
|
|
use futures::{future::BoxFuture, pin_mut, stream::FuturesUnordered, FutureExt, StreamExt};
|
|
use ruma::{
|
|
api::federation::transactions::{
|
|
edu::{DeviceListUpdateContent, Edu, PresenceContent, PresenceUpdate, ReceiptContent, ReceiptData, ReceiptMap},
|
|
send_transaction_message,
|
|
},
|
|
device_id,
|
|
events::{push_rules::PushRulesEvent, receipt::ReceiptType, AnySyncEphemeralRoomEvent, GlobalAccountDataEventType},
|
|
push, uint, CanonicalJsonObject, MilliSecondsSinceUnixEpoch, OwnedServerName, OwnedUserId, RoomId, RoomVersionId,
|
|
ServerName, UInt,
|
|
};
|
|
use serde_json::value::{to_raw_value, RawValue as RawJsonValue};
|
|
use tokio::time::sleep_until;
|
|
|
|
use super::{appservice, data::QueueItem, Destination, Msg, SendingEvent, Service};
|
|
|
|
#[derive(Debug)]
|
|
enum TransactionStatus {
|
|
Running,
|
|
Failed(u32, Instant), // number of times failed, time of last failure
|
|
Retrying(u32), // number of times failed
|
|
}
|
|
|
|
type SendingError = (Destination, Error);
|
|
type SendingResult = Result<Destination, SendingError>;
|
|
type SendingFuture<'a> = BoxFuture<'a, SendingResult>;
|
|
type SendingFutures<'a> = FuturesUnordered<SendingFuture<'a>>;
|
|
type CurTransactionStatus = HashMap<Destination, TransactionStatus>;
|
|
|
|
const DEQUEUE_LIMIT: usize = 48;
|
|
const SELECT_EDU_LIMIT: usize = 16;
|
|
const CLEANUP_TIMEOUT_MS: u64 = 3500;
|
|
|
|
impl Service {
|
|
#[tracing::instrument(skip_all, name = "sender")]
|
|
pub(super) async fn sender(&self) -> Result<()> {
|
|
let mut statuses: CurTransactionStatus = CurTransactionStatus::new();
|
|
let mut futures: SendingFutures<'_> = FuturesUnordered::new();
|
|
let receiver = self.receiver.lock().await;
|
|
|
|
self.initial_requests(&mut futures, &mut statuses).await;
|
|
loop {
|
|
debug_assert!(!receiver.is_closed(), "channel error");
|
|
tokio::select! {
|
|
request = receiver.recv_async() => match request {
|
|
Ok(request) => self.handle_request(request, &mut futures, &mut statuses).await,
|
|
Err(_) => break,
|
|
},
|
|
Some(response) = futures.next() => {
|
|
self.handle_response(response, &mut futures, &mut statuses).await;
|
|
},
|
|
}
|
|
}
|
|
self.finish_responses(&mut futures, &mut statuses).await;
|
|
|
|
Ok(())
|
|
}
|
|
|
|
async fn handle_response<'a>(
|
|
&'a self, response: SendingResult, futures: &mut SendingFutures<'a>, statuses: &mut CurTransactionStatus,
|
|
) {
|
|
match response {
|
|
Ok(dest) => self.handle_response_ok(&dest, futures, statuses).await,
|
|
Err((dest, e)) => Self::handle_response_err(dest, statuses, &e),
|
|
};
|
|
}
|
|
|
|
fn handle_response_err(dest: Destination, statuses: &mut CurTransactionStatus, e: &Error) {
|
|
debug!(dest = ?dest, "{e:?}");
|
|
statuses.entry(dest).and_modify(|e| {
|
|
*e = match e {
|
|
TransactionStatus::Running => TransactionStatus::Failed(1, Instant::now()),
|
|
TransactionStatus::Retrying(ref n) => TransactionStatus::Failed(n.saturating_add(1), Instant::now()),
|
|
TransactionStatus::Failed(..) => panic!("Request that was not even running failed?!"),
|
|
}
|
|
});
|
|
}
|
|
|
|
#[allow(clippy::needless_pass_by_ref_mut)]
|
|
async fn handle_response_ok<'a>(
|
|
&'a self, dest: &Destination, futures: &mut SendingFutures<'a>, statuses: &mut CurTransactionStatus,
|
|
) {
|
|
let _cork = self.db.db.cork();
|
|
self.db.delete_all_active_requests_for(dest).await;
|
|
|
|
// Find events that have been added since starting the last request
|
|
let new_events = self
|
|
.db
|
|
.queued_requests(dest)
|
|
.take(DEQUEUE_LIMIT)
|
|
.collect::<Vec<_>>()
|
|
.await;
|
|
|
|
// Insert any pdus we found
|
|
if !new_events.is_empty() {
|
|
self.db.mark_as_active(&new_events);
|
|
|
|
let new_events_vec = new_events.into_iter().map(|(_, event)| event).collect();
|
|
futures.push(self.send_events(dest.clone(), new_events_vec).boxed());
|
|
} else {
|
|
statuses.remove(dest);
|
|
}
|
|
}
|
|
|
|
#[allow(clippy::needless_pass_by_ref_mut)]
|
|
async fn handle_request<'a>(
|
|
&'a self, msg: Msg, futures: &mut SendingFutures<'a>, statuses: &mut CurTransactionStatus,
|
|
) {
|
|
let iv = vec![(msg.queue_id, msg.event)];
|
|
if let Ok(Some(events)) = self.select_events(&msg.dest, iv, statuses).await {
|
|
if !events.is_empty() {
|
|
futures.push(self.send_events(msg.dest, events).boxed());
|
|
} else {
|
|
statuses.remove(&msg.dest);
|
|
}
|
|
}
|
|
}
|
|
|
|
async fn finish_responses<'a>(&'a self, futures: &mut SendingFutures<'a>, statuses: &mut CurTransactionStatus) {
|
|
let now = Instant::now();
|
|
let timeout = Duration::from_millis(CLEANUP_TIMEOUT_MS);
|
|
let deadline = now.checked_add(timeout).unwrap_or(now);
|
|
loop {
|
|
trace!("Waiting for {} requests to complete...", futures.len());
|
|
tokio::select! {
|
|
() = sleep_until(deadline.into()) => break,
|
|
response = futures.next() => match response {
|
|
Some(response) => self.handle_response(response, futures, statuses).await,
|
|
None => return,
|
|
}
|
|
}
|
|
}
|
|
|
|
debug_warn!("Leaving with {} unfinished requests...", futures.len());
|
|
}
|
|
|
|
#[allow(clippy::needless_pass_by_ref_mut)]
|
|
async fn initial_requests<'a>(&'a self, futures: &mut SendingFutures<'a>, statuses: &mut CurTransactionStatus) {
|
|
let keep = usize::try_from(self.server.config.startup_netburst_keep).unwrap_or(usize::MAX);
|
|
let mut txns = HashMap::<Destination, Vec<SendingEvent>>::new();
|
|
let mut active = self.db.active_requests().boxed();
|
|
|
|
while let Some((key, event, dest)) = active.next().await {
|
|
let entry = txns.entry(dest.clone()).or_default();
|
|
if self.server.config.startup_netburst_keep >= 0 && entry.len() >= keep {
|
|
warn!("Dropping unsent event {dest:?} {:?}", String::from_utf8_lossy(&key));
|
|
self.db.delete_active_request(&key);
|
|
} else {
|
|
entry.push(event);
|
|
}
|
|
}
|
|
|
|
for (dest, events) in txns {
|
|
if self.server.config.startup_netburst && !events.is_empty() {
|
|
statuses.insert(dest.clone(), TransactionStatus::Running);
|
|
futures.push(self.send_events(dest.clone(), events).boxed());
|
|
}
|
|
}
|
|
}
|
|
|
|
#[tracing::instrument(skip_all, level = "debug")]
|
|
async fn select_events(
|
|
&self,
|
|
dest: &Destination,
|
|
new_events: Vec<QueueItem>, // Events we want to send: event and full key
|
|
statuses: &mut CurTransactionStatus,
|
|
) -> Result<Option<Vec<SendingEvent>>> {
|
|
let (allow, retry) = self.select_events_current(dest.clone(), statuses)?;
|
|
|
|
// Nothing can be done for this remote, bail out.
|
|
if !allow {
|
|
return Ok(None);
|
|
}
|
|
|
|
let _cork = self.db.db.cork();
|
|
let mut events = Vec::new();
|
|
|
|
// Must retry any previous transaction for this remote.
|
|
if retry {
|
|
self.db
|
|
.active_requests_for(dest)
|
|
.ready_for_each(|(_, e)| events.push(e))
|
|
.await;
|
|
|
|
return Ok(Some(events));
|
|
}
|
|
|
|
// Compose the next transaction
|
|
let _cork = self.db.db.cork();
|
|
if !new_events.is_empty() {
|
|
self.db.mark_as_active(&new_events);
|
|
for (_, e) in new_events {
|
|
events.push(e);
|
|
}
|
|
}
|
|
|
|
// Add EDU's into the transaction
|
|
if let Destination::Normal(server_name) = dest {
|
|
if let Ok((select_edus, last_count)) = self.select_edus(server_name).await {
|
|
events.extend(select_edus.into_iter().map(SendingEvent::Edu));
|
|
self.db.set_latest_educount(server_name, last_count);
|
|
}
|
|
}
|
|
|
|
Ok(Some(events))
|
|
}
|
|
|
|
#[tracing::instrument(skip_all, level = "debug")]
|
|
fn select_events_current(&self, dest: Destination, statuses: &mut CurTransactionStatus) -> Result<(bool, bool)> {
|
|
let (mut allow, mut retry) = (true, false);
|
|
statuses
|
|
.entry(dest)
|
|
.and_modify(|e| match e {
|
|
TransactionStatus::Failed(tries, time) => {
|
|
// Fail if a request has failed recently (exponential backoff)
|
|
let min = self.server.config.sender_timeout;
|
|
let max = self.server.config.sender_retry_backoff_limit;
|
|
if continue_exponential_backoff_secs(min, max, time.elapsed(), *tries) {
|
|
allow = false;
|
|
} else {
|
|
retry = true;
|
|
*e = TransactionStatus::Retrying(*tries);
|
|
}
|
|
},
|
|
TransactionStatus::Running | TransactionStatus::Retrying(_) => {
|
|
allow = false; // already running
|
|
},
|
|
})
|
|
.or_insert(TransactionStatus::Running);
|
|
|
|
Ok((allow, retry))
|
|
}
|
|
|
|
#[tracing::instrument(skip_all, level = "debug")]
|
|
async fn select_edus(&self, server_name: &ServerName) -> Result<(Vec<Vec<u8>>, u64)> {
|
|
// u64: count of last edu
|
|
let since = self.db.get_latest_educount(server_name).await;
|
|
let mut events = Vec::new();
|
|
let mut max_edu_count = since;
|
|
let mut device_list_changes = HashSet::new();
|
|
|
|
let server_rooms = self.services.state_cache.server_rooms(server_name);
|
|
|
|
pin_mut!(server_rooms);
|
|
while let Some(room_id) = server_rooms.next().await {
|
|
// Look for device list updates in this room
|
|
device_list_changes.extend(
|
|
self.services
|
|
.users
|
|
.keys_changed(room_id.as_str(), since, None)
|
|
.ready_filter(|user_id| self.services.globals.user_is_local(user_id))
|
|
.map(ToOwned::to_owned)
|
|
.collect::<Vec<_>>()
|
|
.await,
|
|
);
|
|
|
|
if self.server.config.allow_outgoing_read_receipts
|
|
&& !self
|
|
.select_edus_receipts(room_id, since, &mut max_edu_count, &mut events)
|
|
.await?
|
|
{
|
|
break;
|
|
}
|
|
}
|
|
|
|
for user_id in device_list_changes {
|
|
// Empty prev id forces synapse to resync; because synapse resyncs,
|
|
// we can just insert placeholder data
|
|
let edu = Edu::DeviceListUpdate(DeviceListUpdateContent {
|
|
user_id,
|
|
device_id: device_id!("placeholder").to_owned(),
|
|
device_display_name: Some("Placeholder".to_owned()),
|
|
stream_id: uint!(1),
|
|
prev_id: Vec::new(),
|
|
deleted: None,
|
|
keys: None,
|
|
});
|
|
|
|
events.push(serde_json::to_vec(&edu).expect("json can be serialized"));
|
|
}
|
|
|
|
if self.server.config.allow_outgoing_presence {
|
|
self.select_edus_presence(server_name, since, &mut max_edu_count, &mut events)
|
|
.await?;
|
|
}
|
|
|
|
Ok((events, max_edu_count))
|
|
}
|
|
|
|
/// Look for presence
|
|
async fn select_edus_presence(
|
|
&self, server_name: &ServerName, since: u64, max_edu_count: &mut u64, events: &mut Vec<Vec<u8>>,
|
|
) -> Result<bool> {
|
|
let presence_since = self.services.presence.presence_since(since);
|
|
|
|
pin_mut!(presence_since);
|
|
let mut presence_updates = Vec::new();
|
|
while let Some((user_id, count, presence_bytes)) = presence_since.next().await {
|
|
*max_edu_count = cmp::max(count, *max_edu_count);
|
|
|
|
if !self.services.globals.user_is_local(&user_id) {
|
|
continue;
|
|
}
|
|
|
|
if !self
|
|
.services
|
|
.state_cache
|
|
.server_sees_user(server_name, &user_id)
|
|
.await
|
|
{
|
|
continue;
|
|
}
|
|
|
|
let presence_event = self
|
|
.services
|
|
.presence
|
|
.from_json_bytes_to_event(&presence_bytes, &user_id)
|
|
.await?;
|
|
|
|
presence_updates.push(PresenceUpdate {
|
|
user_id,
|
|
presence: presence_event.content.presence,
|
|
currently_active: presence_event.content.currently_active.unwrap_or(false),
|
|
last_active_ago: presence_event
|
|
.content
|
|
.last_active_ago
|
|
.unwrap_or_else(|| uint!(0)),
|
|
status_msg: presence_event.content.status_msg,
|
|
});
|
|
|
|
if presence_updates.len() >= SELECT_EDU_LIMIT {
|
|
break;
|
|
}
|
|
}
|
|
|
|
if !presence_updates.is_empty() {
|
|
let presence_content = Edu::Presence(PresenceContent::new(presence_updates));
|
|
events.push(serde_json::to_vec(&presence_content).expect("PresenceEvent can be serialized"));
|
|
}
|
|
|
|
Ok(true)
|
|
}
|
|
|
|
/// Look for read receipts in this room
|
|
async fn select_edus_receipts(
|
|
&self, room_id: &RoomId, since: u64, max_edu_count: &mut u64, events: &mut Vec<Vec<u8>>,
|
|
) -> Result<bool> {
|
|
let receipts = self
|
|
.services
|
|
.read_receipt
|
|
.readreceipts_since(room_id, since);
|
|
|
|
pin_mut!(receipts);
|
|
while let Some((user_id, count, read_receipt)) = receipts.next().await {
|
|
*max_edu_count = cmp::max(count, *max_edu_count);
|
|
if !self.services.globals.user_is_local(&user_id) {
|
|
continue;
|
|
}
|
|
|
|
let event = serde_json::from_str(read_receipt.json().get())
|
|
.map_err(|_| Error::bad_database("Invalid edu event in read_receipts."))?;
|
|
|
|
let federation_event = if let AnySyncEphemeralRoomEvent::Receipt(r) = event {
|
|
let mut read = BTreeMap::new();
|
|
let (event_id, mut receipt) = r
|
|
.content
|
|
.0
|
|
.into_iter()
|
|
.next()
|
|
.expect("we only use one event per read receipt");
|
|
|
|
let receipt = receipt
|
|
.remove(&ReceiptType::Read)
|
|
.expect("our read receipts always set this")
|
|
.remove(&user_id)
|
|
.expect("our read receipts always have the user here");
|
|
|
|
read.insert(
|
|
user_id,
|
|
ReceiptData {
|
|
data: receipt.clone(),
|
|
event_ids: vec![event_id.clone()],
|
|
},
|
|
);
|
|
|
|
let receipt_map = ReceiptMap {
|
|
read,
|
|
};
|
|
|
|
let mut receipts = BTreeMap::new();
|
|
receipts.insert(room_id.to_owned(), receipt_map);
|
|
|
|
Edu::Receipt(ReceiptContent {
|
|
receipts,
|
|
})
|
|
} else {
|
|
Error::bad_database("Invalid event type in read_receipts");
|
|
continue;
|
|
};
|
|
|
|
events.push(serde_json::to_vec(&federation_event).expect("json can be serialized"));
|
|
|
|
if events.len() >= SELECT_EDU_LIMIT {
|
|
return Ok(false);
|
|
}
|
|
}
|
|
|
|
Ok(true)
|
|
}
|
|
|
|
async fn send_events(&self, dest: Destination, events: Vec<SendingEvent>) -> SendingResult {
|
|
//debug_assert!(!events.is_empty(), "sending empty transaction");
|
|
match dest {
|
|
Destination::Normal(ref server) => self.send_events_dest_normal(&dest, server, events).await,
|
|
Destination::Appservice(ref id) => self.send_events_dest_appservice(&dest, id, events).await,
|
|
Destination::Push(ref userid, ref pushkey) => {
|
|
self.send_events_dest_push(&dest, userid, pushkey, events)
|
|
.await
|
|
},
|
|
}
|
|
}
|
|
|
|
#[tracing::instrument(skip(self, dest, events), name = "appservice")]
|
|
async fn send_events_dest_appservice(
|
|
&self, dest: &Destination, id: &str, events: Vec<SendingEvent>,
|
|
) -> SendingResult {
|
|
let Some(appservice) = self.services.appservice.get_registration(id).await else {
|
|
return Err((dest.clone(), err!(Database(warn!(?id, "Missing appservice registration")))));
|
|
};
|
|
|
|
let mut pdu_jsons = Vec::new();
|
|
for event in &events {
|
|
match event {
|
|
SendingEvent::Pdu(pdu_id) => {
|
|
if let Ok(pdu) = self.services.timeline.get_pdu_from_id(pdu_id).await {
|
|
pdu_jsons.push(pdu.to_room_event());
|
|
}
|
|
},
|
|
SendingEvent::Edu(_) | SendingEvent::Flush => {
|
|
// Appservices don't need EDUs (?) and flush only;
|
|
// no new content
|
|
},
|
|
}
|
|
}
|
|
|
|
let txn_id = &*general_purpose::URL_SAFE_NO_PAD.encode(calculate_hash(
|
|
&events
|
|
.iter()
|
|
.map(|e| match e {
|
|
SendingEvent::Edu(b) | SendingEvent::Pdu(b) => &**b,
|
|
SendingEvent::Flush => &[],
|
|
})
|
|
.collect::<Vec<_>>(),
|
|
));
|
|
|
|
//debug_assert!(!pdu_jsons.is_empty(), "sending empty transaction");
|
|
let client = &self.services.client.appservice;
|
|
match appservice::send_request(
|
|
client,
|
|
appservice,
|
|
ruma::api::appservice::event::push_events::v1::Request {
|
|
events: pdu_jsons,
|
|
txn_id: txn_id.into(),
|
|
ephemeral: Vec::new(),
|
|
to_device: Vec::new(),
|
|
},
|
|
)
|
|
.await
|
|
{
|
|
Ok(_) => Ok(dest.clone()),
|
|
Err(e) => Err((dest.clone(), e)),
|
|
}
|
|
}
|
|
|
|
#[tracing::instrument(skip(self, dest, events), name = "push")]
|
|
async fn send_events_dest_push(
|
|
&self, dest: &Destination, userid: &OwnedUserId, pushkey: &str, events: Vec<SendingEvent>,
|
|
) -> SendingResult {
|
|
let Ok(pusher) = self.services.pusher.get_pusher(userid, pushkey).await else {
|
|
return Err((dest.clone(), err!(Database(error!(?userid, ?pushkey, "Missing pusher")))));
|
|
};
|
|
|
|
let mut pdus = Vec::new();
|
|
for event in &events {
|
|
match event {
|
|
SendingEvent::Pdu(pdu_id) => {
|
|
if let Ok(pdu) = self.services.timeline.get_pdu_from_id(pdu_id).await {
|
|
pdus.push(pdu);
|
|
}
|
|
},
|
|
SendingEvent::Edu(_) | SendingEvent::Flush => {
|
|
// Push gateways don't need EDUs (?) and flush only;
|
|
// no new content
|
|
},
|
|
}
|
|
}
|
|
|
|
for pdu in pdus {
|
|
// Redacted events are not notification targets (we don't send push for them)
|
|
if let Some(unsigned) = &pdu.unsigned {
|
|
if let Ok(unsigned) = serde_json::from_str::<serde_json::Value>(unsigned.get()) {
|
|
if unsigned.get("redacted_because").is_some() {
|
|
continue;
|
|
}
|
|
}
|
|
}
|
|
|
|
let rules_for_user = self
|
|
.services
|
|
.account_data
|
|
.get(None, userid, GlobalAccountDataEventType::PushRules.to_string().into())
|
|
.await
|
|
.and_then(|event| serde_json::from_str::<PushRulesEvent>(event.get()).map_err(Into::into))
|
|
.map_or_else(
|
|
|_| push::Ruleset::server_default(userid),
|
|
|ev: PushRulesEvent| ev.content.global,
|
|
);
|
|
|
|
let unread: UInt = self
|
|
.services
|
|
.user
|
|
.notification_count(userid, &pdu.room_id)
|
|
.await
|
|
.try_into()
|
|
.expect("notification count can't go that high");
|
|
|
|
let _response = self
|
|
.services
|
|
.pusher
|
|
.send_push_notice(userid, unread, &pusher, rules_for_user, &pdu)
|
|
.await
|
|
.map_err(|e| (dest.clone(), e));
|
|
}
|
|
|
|
Ok(dest.clone())
|
|
}
|
|
|
|
#[tracing::instrument(skip(self, dest, events), name = "", level = "debug")]
|
|
async fn send_events_dest_normal(
|
|
&self, dest: &Destination, server: &OwnedServerName, events: Vec<SendingEvent>,
|
|
) -> SendingResult {
|
|
let mut pdu_jsons = Vec::with_capacity(
|
|
events
|
|
.iter()
|
|
.filter(|event| matches!(event, SendingEvent::Pdu(_)))
|
|
.count(),
|
|
);
|
|
let mut edu_jsons = Vec::with_capacity(
|
|
events
|
|
.iter()
|
|
.filter(|event| matches!(event, SendingEvent::Edu(_)))
|
|
.count(),
|
|
);
|
|
|
|
for event in &events {
|
|
match event {
|
|
// TODO: check room version and remove event_id if needed
|
|
SendingEvent::Pdu(pdu_id) => {
|
|
if let Ok(pdu) = self.services.timeline.get_pdu_json_from_id(pdu_id).await {
|
|
pdu_jsons.push(self.convert_to_outgoing_federation_event(pdu).await);
|
|
}
|
|
},
|
|
SendingEvent::Edu(edu) => {
|
|
if let Ok(raw) = serde_json::from_slice(edu) {
|
|
edu_jsons.push(raw);
|
|
}
|
|
},
|
|
SendingEvent::Flush => {}, // flush only; no new content
|
|
}
|
|
}
|
|
|
|
//debug_assert!(pdu_jsons.len() + edu_jsons.len() > 0, "sending empty
|
|
// transaction");
|
|
let transaction_id = &*general_purpose::URL_SAFE_NO_PAD.encode(calculate_hash(
|
|
&events
|
|
.iter()
|
|
.map(|e| match e {
|
|
SendingEvent::Edu(b) | SendingEvent::Pdu(b) => &**b,
|
|
SendingEvent::Flush => &[],
|
|
})
|
|
.collect::<Vec<_>>(),
|
|
));
|
|
|
|
let request = send_transaction_message::v1::Request {
|
|
origin: self.server.config.server_name.clone(),
|
|
pdus: pdu_jsons,
|
|
edus: edu_jsons,
|
|
origin_server_ts: MilliSecondsSinceUnixEpoch::now(),
|
|
transaction_id: transaction_id.into(),
|
|
};
|
|
|
|
let client = &self.services.client.sender;
|
|
self.send(client, server, request)
|
|
.await
|
|
.inspect(|response| {
|
|
response
|
|
.pdus
|
|
.iter()
|
|
.filter(|(_, res)| res.is_err())
|
|
.for_each(
|
|
|(pdu_id, res)| warn!(%transaction_id, %server, "error sending PDU {pdu_id} to remote server: {res:?}"),
|
|
);
|
|
})
|
|
.map(|_| dest.clone())
|
|
.map_err(|e| (dest.clone(), e))
|
|
}
|
|
|
|
/// This does not return a full `Pdu` it is only to satisfy ruma's types.
|
|
pub async fn convert_to_outgoing_federation_event(&self, mut pdu_json: CanonicalJsonObject) -> Box<RawJsonValue> {
|
|
if let Some(unsigned) = pdu_json
|
|
.get_mut("unsigned")
|
|
.and_then(|val| val.as_object_mut())
|
|
{
|
|
unsigned.remove("transaction_id");
|
|
}
|
|
|
|
// room v3 and above removed the "event_id" field from remote PDU format
|
|
if let Some(room_id) = pdu_json
|
|
.get("room_id")
|
|
.and_then(|val| RoomId::parse(val.as_str()?).ok())
|
|
{
|
|
match self.services.state.get_room_version(&room_id).await {
|
|
Ok(room_version_id) => match room_version_id {
|
|
RoomVersionId::V1 | RoomVersionId::V2 => {},
|
|
_ => _ = pdu_json.remove("event_id"),
|
|
},
|
|
Err(_) => _ = pdu_json.remove("event_id"),
|
|
}
|
|
} else {
|
|
pdu_json.remove("event_id");
|
|
}
|
|
|
|
// TODO: another option would be to convert it to a canonical string to validate
|
|
// size and return a Result<Raw<...>>
|
|
// serde_json::from_str::<Raw<_>>(
|
|
// ruma::serde::to_canonical_json_string(pdu_json).expect("CanonicalJson is
|
|
// valid serde_json::Value"), )
|
|
// .expect("Raw::from_value always works")
|
|
|
|
to_raw_value(&pdu_json).expect("CanonicalJson is valid serde_json::Value")
|
|
}
|
|
}
|