Cleanup/optimize sliding sync types and arguments.

Signed-off-by: Jason Volk <jason@zemos.net>
This commit is contained in:
Jason Volk
2025-10-06 01:12:01 +00:00
parent b1ea7b101d
commit 6c724bbc2f
10 changed files with 436 additions and 392 deletions

View File

@@ -15,7 +15,7 @@ use std::{
use axum::extract::State;
use futures::{
FutureExt, Stream, StreamExt, TryFutureExt, TryStreamExt,
future::{OptionFuture, join4, join5, try_join},
future::{OptionFuture, join, join3, join5, try_join},
pin_mut,
};
use ruma::{
@@ -24,6 +24,7 @@ use ruma::{
Request, Response, request::ExtensionRoomConfig, response,
},
directory::RoomTypeFilter,
events::room::member::MembershipState,
uint,
};
use tokio::time::{Instant, timeout_at};
@@ -44,16 +45,29 @@ use tuwunel_core::{
};
use tuwunel_service::{
Services,
sync::{KnownRooms, into_snake_key},
sync::{KnownRooms, ListId, into_connection_key},
};
use super::share_encrypted_room;
use crate::{Ruma, client::DEFAULT_BUMP_TYPES};
type SyncInfo<'a> = (&'a UserId, &'a DeviceId, u64, &'a Request);
#[derive(Copy, Clone)]
struct SyncInfo<'a> {
sender_user: &'a UserId,
sender_device: &'a DeviceId,
request: &'a Request,
globalsince: u64,
}
struct TodoRoom {
membership: MembershipState,
requested_state: BTreeSet<TypeStateKey>,
timeline_limit: usize,
roomsince: u64,
}
type TodoRooms = BTreeMap<OwnedRoomId, TodoRoom>;
type TodoRoom = (BTreeSet<TypeStateKey>, usize, u64);
type ResponseLists = BTreeMap<String, response::List>;
type ResponseLists = BTreeMap<ListId, response::List>;
/// `POST /_matrix/client/unstable/org.matrix.simplified_msc3575/sync`
/// ([MSC4186])
@@ -88,8 +102,9 @@ pub(crate) async fn sync_events_v5_route(
.unwrap_or(0);
let (sender_user, sender_device) = body.sender();
let snake_key = into_snake_key(sender_user, sender_device, request.conn_id.as_deref());
if globalsince != 0 && !services.sync.snake_connection_cached(&snake_key) {
let conn_key = into_connection_key(sender_user, sender_device, request.conn_id.as_deref());
if globalsince != 0 && !services.sync.is_connection_cached(&conn_key) {
return Err!(Request(UnknownPos(
"Connection data unknown to server; restarting sync stream."
)));
@@ -97,15 +112,22 @@ pub(crate) async fn sync_events_v5_route(
// Client / User requested an initial sync
if globalsince == 0 {
services
.sync
.forget_snake_sync_connection(&snake_key);
services.sync.forget_connection(&conn_key);
}
// Get sticky parameters from cache
let known_rooms = services
.sync
.update_snake_sync_request_with_cache(&snake_key, &mut request);
.update_cache(&conn_key, &mut request);
let sync_info = SyncInfo {
sender_user,
sender_device,
globalsince,
request: &request,
};
let lists = handle_lists(services, sync_info, known_rooms);
let ping_presence = services
.presence
@@ -113,45 +135,7 @@ pub(crate) async fn sync_events_v5_route(
.inspect_err(inspect_log)
.ok();
let all_joined_rooms = services
.state_cache
.rooms_joined(sender_user)
.map(ToOwned::to_owned)
.collect::<Vec<OwnedRoomId>>();
let all_invited_rooms = services
.state_cache
.rooms_invited(sender_user)
.map(ToOwned::to_owned)
.collect::<Vec<OwnedRoomId>>();
let all_knocked_rooms = services
.state_cache
.rooms_knocked(sender_user)
.map(ToOwned::to_owned)
.collect::<Vec<OwnedRoomId>>();
let (all_joined_rooms, all_invited_rooms, all_knocked_rooms, _) =
join4(all_joined_rooms, all_invited_rooms, all_knocked_rooms, ping_presence).await;
let all_invited_rooms = all_invited_rooms.iter().map(AsRef::as_ref);
let all_knocked_rooms = all_knocked_rooms.iter().map(AsRef::as_ref);
let all_joined_rooms = all_joined_rooms.iter().map(AsRef::as_ref);
let all_rooms = all_joined_rooms
.clone()
.chain(all_invited_rooms.clone())
.chain(all_knocked_rooms.clone());
let sync_info: SyncInfo<'_> = (sender_user, sender_device, globalsince, &request);
let (known_rooms, todo_rooms, lists) = handle_lists(
services,
sync_info,
known_rooms,
all_invited_rooms.clone(),
all_joined_rooms.clone(),
all_rooms.clone(),
)
.await;
let ((known_rooms, todo_rooms, lists), _) = join(lists, ping_presence).await;
let timeout = request
.timeout
@@ -170,7 +154,7 @@ pub(crate) async fn sync_events_v5_route(
let mut response = Response {
txn_id: request.txn_id.clone(),
lists,
pos: String::new(),
pos: Default::default(),
rooms: Default::default(),
extensions: Default::default(),
};
@@ -181,15 +165,8 @@ pub(crate) async fn sync_events_v5_route(
debug_assert!(globalsince <= next_batch, "next_batch is monotonic");
if globalsince < next_batch {
let rooms = handle_rooms(
services,
&sync_info,
next_batch,
&known_rooms,
&todo_rooms,
all_invited_rooms.clone(),
)
.map_ok(|rooms| response.rooms = rooms);
let rooms = handle_rooms(services, sync_info, next_batch, &todo_rooms)
.map_ok(|rooms| response.rooms = rooms);
let extensions =
handle_extensions(services, sync_info, next_batch, &known_rooms, &todo_rooms)
@@ -199,14 +176,14 @@ pub(crate) async fn sync_events_v5_route(
if !is_empty_response(&response) {
trace!(globalsince, next_batch, "response {response:?}");
response.pos = next_batch.to_string();
response.pos = next_batch.to_string().into();
return Ok(response);
}
}
if timeout_at(stop_at, watchers).await.is_err() {
trace!(globalsince, next_batch, "timeout; empty response");
response.pos = next_batch.to_string();
response.pos = next_batch.to_string().into();
return Ok(response);
}
@@ -234,26 +211,50 @@ fn is_empty_response(response: &Response) -> bool {
level = "debug",
skip_all,
fields(
all_invited_rooms = all_invited_rooms.clone().count(),
all_joined_rooms = all_joined_rooms.clone().count(),
all_rooms = all_rooms.clone().count(),
known_rooms = known_rooms.len(),
)
)]
#[allow(clippy::too_many_arguments)]
async fn handle_lists<'a, Rooms, AllRooms>(
async fn handle_lists(
services: &Services,
sync_info: SyncInfo<'_>,
known_rooms: KnownRooms,
all_invited_rooms: Rooms,
all_joined_rooms: Rooms,
all_rooms: AllRooms,
) -> (KnownRooms, TodoRooms, ResponseLists)
where
Rooms: Iterator<Item = &'a RoomId> + Clone + Send + 'a,
AllRooms: Iterator<Item = &'a RoomId> + Clone + Send + 'a,
{
let &(sender_user, sender_device, globalsince, request) = &sync_info;
) -> (KnownRooms, TodoRooms, ResponseLists) {
let &SyncInfo {
sender_user,
sender_device,
request,
globalsince,
} = &sync_info;
let all_joined_rooms = services
.state_cache
.rooms_joined(sender_user)
.map(ToOwned::to_owned)
.collect::<Vec<OwnedRoomId>>();
let all_invited_rooms = services
.state_cache
.rooms_invited(sender_user)
.map(ToOwned::to_owned)
.collect::<Vec<OwnedRoomId>>();
let all_knocked_rooms = services
.state_cache
.rooms_knocked(sender_user)
.map(ToOwned::to_owned)
.collect::<Vec<OwnedRoomId>>();
let (all_joined_rooms, all_invited_rooms, all_knocked_rooms) =
join3(all_joined_rooms, all_invited_rooms, all_knocked_rooms).await;
let all_invited_rooms = all_invited_rooms.iter().map(AsRef::as_ref);
let all_knocked_rooms = all_knocked_rooms.iter().map(AsRef::as_ref);
let all_joined_rooms = all_joined_rooms.iter().map(AsRef::as_ref);
let all_rooms = all_joined_rooms
.clone()
.chain(all_invited_rooms.clone())
.chain(all_knocked_rooms.clone());
let mut todo_rooms: TodoRooms = BTreeMap::new();
let mut response_lists = ResponseLists::new();
@@ -298,13 +299,25 @@ where
new_known_rooms.extend(new_rooms);
for room_id in room_ids {
let todo_room = todo_rooms.entry(room_id.to_owned()).or_insert((
BTreeSet::new(),
0_usize,
u64::MAX,
));
let todo_room = todo_rooms
.entry(room_id.to_owned())
.or_insert(TodoRoom {
membership: MembershipState::Join,
requested_state: BTreeSet::new(),
timeline_limit: 0_usize,
roomsince: u64::MAX,
});
todo_room.0.extend(
todo_room.membership = if all_invited_rooms
.clone()
.any(is_equal_to!(room_id))
{
MembershipState::Invite
} else {
MembershipState::Join
};
todo_room.requested_state.extend(
list.room_details
.required_state
.iter()
@@ -312,10 +325,10 @@ where
);
let limit: usize = usize_from_ruma(list.room_details.timeline_limit).min(100);
todo_room.1 = todo_room.1.max(limit);
todo_room.timeline_limit = todo_room.timeline_limit.max(limit);
// 0 means unknown because it got out of date
todo_room.2 = todo_room.2.min(
todo_room.roomsince = todo_room.roomsince.min(
known_rooms
.get(list_id.as_str())
.and_then(|k| k.get(room_id))
@@ -326,14 +339,11 @@ where
}
if let Some(conn_id) = request.conn_id.as_deref() {
let snake_key = into_snake_key(sender_user, sender_device, conn_id.into());
let conn_key = into_connection_key(sender_user, sender_device, conn_id.into());
let list_id = list_id.as_str().into();
services.sync.update_snake_sync_known_rooms(
&snake_key,
list_id,
new_known_rooms,
globalsince,
);
services
.sync
.update_known_rooms(&conn_key, list_id, new_known_rooms, globalsince);
}
response_lists.insert(list_id.clone(), response::List {
@@ -358,7 +368,12 @@ where
)]
async fn fetch_subscriptions(
services: &Services,
(sender_user, sender_device, globalsince, request): SyncInfo<'_>,
SyncInfo {
sender_user,
sender_device,
globalsince,
request,
}: SyncInfo<'_>,
known_rooms: KnownRooms,
todo_rooms: TodoRooms,
) -> (KnownRooms, TodoRooms) {
@@ -381,22 +396,28 @@ async fn fetch_subscriptions(
.then_some((room_id, room))
})
.ready_fold(subs, |(mut todo_rooms, mut known_subs), (room_id, room)| {
let todo_room =
todo_rooms
.entry(room_id.clone())
.or_insert((BTreeSet::new(), 0_usize, u64::MAX));
let todo_room = todo_rooms
.entry(room_id.clone())
.or_insert(TodoRoom {
membership: MembershipState::Join,
requested_state: BTreeSet::new(),
timeline_limit: 0_usize,
roomsince: u64::MAX,
});
todo_room.0.extend(
todo_room.requested_state.extend(
room.required_state
.iter()
.map(|(ty, sk)| (ty.clone(), sk.as_str().into())),
);
let limit: UInt = room.timeline_limit;
todo_room.1 = todo_room.1.max(usize_from_ruma(limit));
todo_room.timeline_limit = todo_room
.timeline_limit
.max(usize_from_ruma(limit));
// 0 means unknown because it got out of date
todo_room.2 = todo_room.2.min(
todo_room.roomsince = todo_room.roomsince.min(
known_rooms
.get("subscriptions")
.and_then(|k| k.get(room_id))
@@ -410,11 +431,11 @@ async fn fetch_subscriptions(
.await;
if let Some(conn_id) = request.conn_id.as_deref() {
let snake_key = into_snake_key(sender_user, sender_device, conn_id.into());
let conn_key = into_connection_key(sender_user, sender_device, conn_id.into());
let list_id = "subscriptions".into();
services
.sync
.update_snake_sync_known_rooms(&snake_key, list_id, known_subs, globalsince);
.update_known_rooms(&conn_key, list_id, known_subs, globalsince);
}
(known_rooms, todo_rooms)
@@ -461,32 +482,20 @@ where
skip_all,
fields(
next_batch,
all_invited_rooms = all_invited_rooms.clone().count(),
todo_rooms = todo_rooms.len(),
)
)]
async fn handle_rooms<'a, Rooms>(
async fn handle_rooms(
services: &Services,
sync_info: &SyncInfo<'_>,
sync_info: SyncInfo<'_>,
next_batch: u64,
_known_rooms: &KnownRooms,
todo_rooms: &TodoRooms,
all_invited_rooms: Rooms,
) -> Result<BTreeMap<OwnedRoomId, response::Room>>
where
Rooms: Iterator<Item = &'a RoomId> + Clone + Send + Sync + 'a,
{
) -> Result<BTreeMap<OwnedRoomId, response::Room>> {
let rooms: BTreeMap<_, _> = todo_rooms
.iter()
.try_stream()
.broad_and_then(async |(room_id, todo_room)| {
let is_invited = all_invited_rooms
.clone()
.any(is_equal_to!(room_id));
let room =
room::handle(services, next_batch, sync_info, room_id, todo_room, is_invited)
.await?;
let room = room::handle(services, next_batch, sync_info, room_id, todo_room).await?;
Ok((room_id, room))
})
@@ -513,7 +522,7 @@ async fn handle_extensions(
known_rooms: &KnownRooms,
todo_rooms: &TodoRooms,
) -> Result<response::Extensions> {
let &(_, _, _, request) = &sync_info;
let SyncInfo { request, .. } = sync_info;
let account_data: OptionFuture<_> = request
.extensions
@@ -569,26 +578,31 @@ async fn handle_extensions(
})
}
fn extension_rooms_todo<'a>(
(_, _, _, request): SyncInfo<'a>,
fn extension_rooms_todo<'a, ListIter, ConfigIter>(
SyncInfo { request, .. }: SyncInfo<'a>,
known_rooms: &'a KnownRooms,
todo_rooms: &'a TodoRooms,
lists: Option<&'a Vec<String>>,
rooms: Option<&'a Vec<ExtensionRoomConfig>>,
) -> impl Iterator<Item = &'a RoomId> + Send + 'a {
let lists_explicit = lists.into_iter().flat_map(|vec| vec.iter());
lists: Option<ListIter>,
rooms: Option<ConfigIter>,
) -> impl Iterator<Item = &'a RoomId> + Send + Sync + 'a
where
ListIter: Iterator<Item = &'a ListId> + Clone + Send + Sync + 'a,
ConfigIter: Iterator<Item = &'a ExtensionRoomConfig> + Clone + Send + Sync + 'a,
{
let lists_explicit = lists.clone().into_iter().flatten();
let rooms_explicit = rooms
.clone()
.into_iter()
.flatten()
.filter_map(|erc| extract_variant!(erc, ExtensionRoomConfig::Room))
.map(AsRef::<RoomId>::as_ref);
let lists_requested = request
.lists
.keys()
.filter(move |_| lists.is_none());
let rooms_explicit = rooms
.into_iter()
.flat_map(|vec| vec.iter())
.filter_map(|erc| extract_variant!(erc, ExtensionRoomConfig::Room))
.map(AsRef::<RoomId>::as_ref);
let rooms_implicit = todo_rooms
.keys()
.map(AsRef::as_ref)

View File

@@ -6,7 +6,7 @@ use tuwunel_core::{
};
use tuwunel_service::Services;
use super::{KnownRooms, SyncInfo, TodoRooms, extension_rooms_todo};
use super::{KnownRooms, SyncInfo, TodoRoom, TodoRooms, extension_rooms_todo};
#[tracing::instrument(level = "trace", skip_all, fields(globalsince, next_batch))]
pub(super) async fn collect(
@@ -16,31 +16,39 @@ pub(super) async fn collect(
known_rooms: &KnownRooms,
todo_rooms: &TodoRooms,
) -> Result<response::AccountData> {
let (sender_user, _, globalsince, request) = sync_info;
let data = &request.extensions.account_data;
let rooms = extension_rooms_todo(
sync_info,
known_rooms,
todo_rooms,
data.lists.as_ref(),
data.rooms.as_ref(),
)
.stream()
.broad_filter_map(async |room_id| {
let &(_, _, roomsince) = todo_rooms.get(room_id)?;
let changes: Vec<_> = services
.account_data
.changes_since(Some(room_id), sender_user, roomsince, Some(next_batch))
.ready_filter_map(|e| extract_variant!(e, AnyRawAccountDataEvent::Room))
.collect()
.await;
let SyncInfo { sender_user, globalsince, request, .. } = sync_info;
changes
.is_empty()
.eq(&false)
.then(move || (room_id.to_owned(), changes))
})
.collect();
let lists = request
.extensions
.account_data
.lists
.as_deref()
.map(<[_]>::iter);
let rooms = request
.extensions
.account_data
.rooms
.as_deref()
.map(<[_]>::iter);
let rooms = extension_rooms_todo(sync_info, known_rooms, todo_rooms, lists, rooms)
.stream()
.broad_filter_map(async |room_id| {
let &TodoRoom { roomsince, .. } = todo_rooms.get(room_id)?;
let changes: Vec<_> = services
.account_data
.changes_since(Some(room_id), sender_user, roomsince, Some(next_batch))
.ready_filter_map(|e| extract_variant!(e, AnyRawAccountDataEvent::Room))
.collect()
.await;
changes
.is_empty()
.eq(&false)
.then(move || (room_id.to_owned(), changes))
})
.collect();
let global = services
.account_data

View File

@@ -29,10 +29,13 @@ use super::{SyncInfo, share_encrypted_room};
#[tracing::instrument(level = "trace", skip_all, fields(globalsince, next_batch,))]
pub(super) async fn collect(
services: &Services,
syncinfo: SyncInfo<'_>,
sync_info: SyncInfo<'_>,
next_batch: u64,
) -> Result<response::E2EE> {
let &(sender_user, sender_device, globalsince, _) = &syncinfo;
let SyncInfo {
sender_user, sender_device, globalsince, ..
} = sync_info;
let keys_changed = services
.users
.keys_changed(sender_user, globalsince, Some(next_batch))
@@ -46,7 +49,7 @@ pub(super) async fn collect(
.rooms_joined(sender_user)
.map(ToOwned::to_owned)
.broad_filter_map(async |room_id| {
collect_room(services, syncinfo, next_batch, &room_id)
collect_room(services, sync_info, next_batch, &room_id)
.await
.ok()
})
@@ -101,7 +104,7 @@ pub(super) async fn collect(
#[tracing::instrument(level = "trace", skip_all, fields(room_id))]
async fn collect_room(
services: &Services,
(sender_user, _, globalsince, _): SyncInfo<'_>,
SyncInfo { sender_user, globalsince, .. }: SyncInfo<'_>,
next_batch: u64,
room_id: &RoomId,
) -> Result<pair_of!(HashSet<OwnedUserId>)> {

View File

@@ -11,7 +11,7 @@ use tuwunel_core::{
};
use tuwunel_service::{Services, rooms::read_receipt::pack_receipts};
use super::{KnownRooms, SyncInfo, TodoRooms, extension_rooms_todo};
use super::{KnownRooms, SyncInfo, TodoRoom, TodoRooms, extension_rooms_todo};
#[tracing::instrument(level = "trace", skip_all)]
pub(super) async fn collect(
@@ -21,33 +21,41 @@ pub(super) async fn collect(
known_rooms: &KnownRooms,
todo_rooms: &TodoRooms,
) -> Result<response::Receipts> {
let (_, _, _, request) = sync_info;
let data = &request.extensions.receipts;
let rooms = extension_rooms_todo(
sync_info,
known_rooms,
todo_rooms,
data.lists.as_ref(),
data.rooms.as_ref(),
)
.stream()
.broad_filter_map(async |room_id| {
collect_room(services, sync_info, next_batch, todo_rooms, room_id).await
})
.collect()
.await;
let SyncInfo { request, .. } = sync_info;
let lists = request
.extensions
.receipts
.lists
.as_deref()
.map(<[_]>::iter);
let rooms = request
.extensions
.receipts
.rooms
.as_deref()
.map(<[_]>::iter);
let rooms = extension_rooms_todo(sync_info, known_rooms, todo_rooms, lists, rooms)
.stream()
.broad_filter_map(async |room_id| {
collect_room(services, sync_info, next_batch, todo_rooms, room_id).await
})
.collect()
.await;
Ok(response::Receipts { rooms })
}
async fn collect_room(
services: &Services,
(sender_user, ..): SyncInfo<'_>,
SyncInfo { sender_user, .. }: SyncInfo<'_>,
next_batch: u64,
todo_rooms: &TodoRooms,
room_id: &RoomId,
) -> Option<(OwnedRoomId, Raw<SyncReceiptEvent>)> {
let &(_, _, roomsince) = todo_rooms.get(room_id)?;
let &TodoRoom { roomsince, .. } = todo_rooms.get(room_id)?;
let private_receipt = services
.read_receipt
.last_privateread_update(sender_user, room_id)

View File

@@ -7,7 +7,7 @@ use futures::{
use ruma::{
JsOption, MxcUri, OwnedMxcUri, RoomId, UInt, UserId,
api::client::sync::sync_events::{UnreadNotificationsCount, v5::response},
events::StateEventType,
events::{StateEventType, room::member::MembershipState},
};
use tuwunel_core::{
Result, at, debug_error, is_equal_to,
@@ -25,21 +25,25 @@ use crate::client::{DEFAULT_BUMP_TYPES, ignored_filter, sync::load_timeline};
pub(super) async fn handle(
services: &Services,
next_batch: u64,
(sender_user, _, _globalsince, _): &SyncInfo<'_>,
SyncInfo { sender_user, .. }: SyncInfo<'_>,
room_id: &RoomId,
(required_state_request, timeline_limit, roomsince): &TodoRoom,
is_invited: bool,
&TodoRoom {
ref membership,
ref requested_state,
timeline_limit,
roomsince,
}: &TodoRoom,
) -> Result<Option<response::Room>> {
let timeline: OptionFuture<_> = is_invited
.eq(&false)
let timeline: OptionFuture<_> = membership
.ne(&MembershipState::Invite)
.then(|| {
load_timeline(
services,
sender_user,
room_id,
PduCount::Normal(*roomsince),
PduCount::Normal(roomsince),
Some(PduCount::from(next_batch)),
*timeline_limit,
timeline_limit,
)
})
.into();
@@ -52,7 +56,7 @@ pub(super) async fn handle(
let (timeline_pdus, limited, _lastcount) =
timeline.unwrap_or_else(|| (Vec::new(), true, PduCount::default()));
if *roomsince != 0 && timeline_pdus.is_empty() && !is_invited {
if roomsince != 0 && timeline_pdus.is_empty() && membership.ne(&MembershipState::Invite) {
return Ok(None);
}
@@ -60,7 +64,7 @@ pub(super) async fn handle(
.first()
.map(at!(0))
.map(PduCount::into_unsigned)
.or_else(|| roomsince.ne(&0).then_some(*roomsince))
.or_else(|| roomsince.ne(&0).then_some(roomsince))
.as_ref()
.map(ToString::to_string);
@@ -80,7 +84,7 @@ pub(super) async fn handle(
bump_stamp
});
let lazy = required_state_request
let lazy = requested_state
.iter()
.any(is_equal_to!(&(StateEventType::RoomMember, "$LAZY".into())));
@@ -97,7 +101,7 @@ pub(super) async fn handle(
.iter()
.map(|sender| (StateEventType::RoomMember, StateKey::from_str(sender.as_str())));
let required_state = required_state_request
let required_state = requested_state
.iter()
.cloned()
.chain(timeline_senders)
@@ -119,7 +123,8 @@ pub(super) async fn handle(
.collect();
// TODO: figure out a timestamp we can use for remote invites
let invite_state: OptionFuture<_> = is_invited
let invite_state: OptionFuture<_> = membership
.eq(&MembershipState::Invite)
.then(|| {
services
.state_cache
@@ -199,7 +204,7 @@ pub(super) async fn handle(
let num_live = None; // Count events in timeline greater than global sync counter
Ok(Some(response::Room {
initial: Some(*roomsince == 0),
initial: Some(roomsince == 0),
name: room_name.or(hero_name),
avatar: JsOption::from_option(room_avatar.or(heroes_avatar)),
invite_state: invite_state.flatten(),

View File

@@ -8,7 +8,9 @@ use super::SyncInfo;
#[tracing::instrument(level = "trace", skip_all, fields(globalsince, next_batch))]
pub(super) async fn collect(
services: &Services,
(sender_user, sender_device, globalsince, _request): SyncInfo<'_>,
SyncInfo {
sender_user, sender_device, globalsince, ..
}: SyncInfo<'_>,
next_batch: u64,
) -> Result<Option<response::ToDevice>> {
services

View File

@@ -2,7 +2,9 @@ use std::collections::BTreeMap;
use futures::{FutureExt, StreamExt, TryFutureExt};
use ruma::{
api::client::sync::sync_events::v5::response, events::typing::TypingEventContent, serde::Raw,
api::client::sync::sync_events::v5::response,
events::typing::{SyncTypingEvent, TypingEventContent},
serde::Raw,
};
use tuwunel_core::{
Result, debug_error,
@@ -21,37 +23,44 @@ pub(super) async fn collect(
todo_rooms: &TodoRooms,
) -> Result<response::Typing> {
use response::Typing;
use ruma::events::typing::SyncTypingEvent;
let (sender_user, _, _, request) = sync_info;
let data = &request.extensions.typing;
extension_rooms_todo(
sync_info,
known_rooms,
todo_rooms,
data.lists.as_ref(),
data.rooms.as_ref(),
)
.stream()
.filter_map(async |room_id| {
services
.typing
.typing_users_for_user(room_id, sender_user)
.inspect_err(|e| debug_error!(%room_id, "Failed to get typing events: {e}"))
.await
.ok()
.filter(|users| !users.is_empty())
.map(|users| (room_id, users))
})
.ready_filter_map(|(room_id, users)| {
let content = TypingEventContent::new(users);
let event = SyncTypingEvent { content };
let event = Raw::new(&event);
let SyncInfo { sender_user, request, .. } = sync_info;
Some((room_id.to_owned(), event.ok()?))
})
.collect::<BTreeMap<_, _>>()
.map(|rooms| Typing { rooms })
.map(Ok)
.await
let lists = request
.extensions
.typing
.lists
.as_deref()
.map(<[_]>::iter);
let rooms = request
.extensions
.typing
.rooms
.as_deref()
.map(<[_]>::iter);
extension_rooms_todo(sync_info, known_rooms, todo_rooms, lists, rooms)
.stream()
.filter_map(async |room_id| {
services
.typing
.typing_users_for_user(room_id, sender_user)
.inspect_err(|e| debug_error!(%room_id, "Failed to get typing events: {e}"))
.await
.ok()
.filter(|users| !users.is_empty())
.map(|users| (room_id, users))
})
.ready_filter_map(|(room_id, users)| {
let content = TypingEventContent::new(users);
let event = SyncTypingEvent { content };
let event = Raw::new(&event);
Some((room_id.to_owned(), event.ok()?))
})
.collect::<BTreeMap<_, _>>()
.map(|rooms| Typing { rooms })
.map(Ok)
.await
}