Fix sync handling for appservices calling without device_id
This commit is contained in:
@@ -32,8 +32,8 @@ pub(crate) async fn get_context_route(
|
||||
State(services): State<crate::State>,
|
||||
body: Ruma<get_context::v3::Request>,
|
||||
) -> Result<get_context::v3::Response> {
|
||||
let sender = body.sender();
|
||||
let (sender_user, sender_device) = sender;
|
||||
let sender_user = body.sender_user();
|
||||
let sender_device = body.sender_device.as_deref();
|
||||
let room_id = &body.room_id;
|
||||
let event_id = &body.event_id;
|
||||
let filter = &body.filter;
|
||||
@@ -110,7 +110,7 @@ pub(crate) async fn get_context_route(
|
||||
|
||||
let lazy_loading_context = lazy_loading::Context {
|
||||
user_id: sender_user,
|
||||
device_id: Some(sender_device),
|
||||
device_id: sender_device,
|
||||
room_id,
|
||||
token: Some(base_count.into_unsigned()),
|
||||
options: Some(&filter.lazy_load_options),
|
||||
|
||||
@@ -23,7 +23,7 @@ pub(crate) async fn events_route(
|
||||
State(services): State<crate::State>,
|
||||
body: Ruma<Request>,
|
||||
) -> Result<Response> {
|
||||
let (sender_user, sender_device) = body.sender();
|
||||
let sender_user = body.sender_user();
|
||||
|
||||
let from = body
|
||||
.body
|
||||
@@ -62,9 +62,11 @@ pub(crate) async fn events_route(
|
||||
.expect("configuration must limit maximum timeout");
|
||||
|
||||
loop {
|
||||
let watchers = services
|
||||
.sync
|
||||
.watch(sender_user, sender_device, once(room_id).stream());
|
||||
let watchers = services.sync.watch(
|
||||
sender_user,
|
||||
body.sender_device.as_deref(),
|
||||
once(room_id).stream(),
|
||||
);
|
||||
|
||||
let next_batch = services.globals.wait_pending().await?;
|
||||
|
||||
|
||||
@@ -35,7 +35,8 @@ pub(crate) async fn upload_keys_route(
|
||||
State(services): State<crate::State>,
|
||||
body: Ruma<upload_keys::v3::Request>,
|
||||
) -> Result<upload_keys::v3::Response> {
|
||||
let (sender_user, sender_device) = body.sender();
|
||||
let sender_user = body.sender_user();
|
||||
let sender_device = body.sender_device()?;
|
||||
|
||||
let one_time_keys = body
|
||||
.one_time_keys
|
||||
|
||||
@@ -589,7 +589,7 @@ pub(crate) async fn set_pushers_route(
|
||||
|
||||
services
|
||||
.pusher
|
||||
.set_pusher(sender_user, body.sender_device(), &body.action)
|
||||
.set_pusher(sender_user, body.sender_device()?, &body.action)
|
||||
.await?;
|
||||
|
||||
Ok(set_pusher::v3::Response::new())
|
||||
|
||||
@@ -23,7 +23,7 @@ pub(crate) async fn logout_route(
|
||||
) -> Result<logout::v3::Response> {
|
||||
services
|
||||
.users
|
||||
.remove_device(body.sender_user(), body.sender_device())
|
||||
.remove_device(body.sender_user(), body.sender_device()?)
|
||||
.await;
|
||||
|
||||
Ok(logout::v3::Response::new())
|
||||
|
||||
@@ -50,7 +50,8 @@ pub(crate) async fn login_token_route(
|
||||
|
||||
// This route SHOULD have UIA
|
||||
// TODO: How do we make only UIA sessions that have not been used before valid?
|
||||
let (sender_user, sender_device) = body.sender();
|
||||
let sender_user = body.sender_user();
|
||||
let sender_device = body.sender_device()?;
|
||||
|
||||
let password_flow = uiaa::AuthFlow { stages: vec![uiaa::AuthType::Password] };
|
||||
|
||||
|
||||
@@ -120,13 +120,15 @@ type PresenceUpdates = HashMap<OwnedUserId, PresenceEventContent>;
|
||||
skip_all,
|
||||
fields(
|
||||
user_id = %body.sender_user(),
|
||||
device_id = %body.sender_device.as_deref().map_or("<no device>", |x| x.as_str()),
|
||||
)
|
||||
)]
|
||||
pub(crate) async fn sync_events_route(
|
||||
State(services): State<crate::State>,
|
||||
body: Ruma<sync_events::v3::Request>,
|
||||
) -> Result<sync_events::v3::Response> {
|
||||
let (sender_user, sender_device) = body.sender();
|
||||
let sender_user = body.sender_user();
|
||||
let sender_device = body.sender_device.as_deref();
|
||||
|
||||
let filter: OptionFuture<_> = body
|
||||
.body
|
||||
@@ -195,7 +197,8 @@ pub(crate) async fn sync_events_route(
|
||||
if since < next_batch || full_state {
|
||||
let response = build_sync_events(
|
||||
&services,
|
||||
body.sender(),
|
||||
sender_user,
|
||||
sender_device,
|
||||
since,
|
||||
next_batch,
|
||||
full_state,
|
||||
@@ -216,7 +219,8 @@ pub(crate) async fn sync_events_route(
|
||||
|
||||
// Wait for activity
|
||||
if time::timeout_at(stop_at, watchers).await.is_err() || services.server.is_stopping() {
|
||||
let response = build_empty_response(&services, body.sender(), next_batch).await;
|
||||
let response =
|
||||
build_empty_response(&services, sender_user, sender_device, next_batch).await;
|
||||
trace!(since, next_batch, "empty response");
|
||||
return Ok(response);
|
||||
}
|
||||
@@ -235,14 +239,22 @@ pub(crate) async fn sync_events_route(
|
||||
|
||||
async fn build_empty_response(
|
||||
services: &Services,
|
||||
(sender_user, sender_device): (&UserId, &DeviceId),
|
||||
sender_user: &UserId,
|
||||
sender_device: Option<&DeviceId>,
|
||||
next_batch: u64,
|
||||
) -> sync_events::v3::Response {
|
||||
let device_one_time_keys_count: OptionFuture<_> = sender_device
|
||||
.map(|sender_device| {
|
||||
services
|
||||
.users
|
||||
.count_one_time_keys(sender_user, sender_device)
|
||||
})
|
||||
.into();
|
||||
|
||||
sync_events::v3::Response {
|
||||
device_one_time_keys_count: services
|
||||
.users
|
||||
.count_one_time_keys(sender_user, sender_device)
|
||||
.await,
|
||||
device_one_time_keys_count: device_one_time_keys_count
|
||||
.await
|
||||
.unwrap_or_default(),
|
||||
|
||||
..sync_events::v3::Response::new(to_small_string(next_batch))
|
||||
}
|
||||
@@ -261,7 +273,8 @@ async fn build_empty_response(
|
||||
)]
|
||||
async fn build_sync_events(
|
||||
services: &Services,
|
||||
(sender_user, sender_device): (&UserId, &DeviceId),
|
||||
sender_user: &UserId,
|
||||
sender_device: Option<&DeviceId>,
|
||||
since: u64,
|
||||
next_batch: u64,
|
||||
full_state: bool,
|
||||
@@ -387,27 +400,38 @@ async fn build_sync_events(
|
||||
.map(ToOwned::to_owned)
|
||||
.collect::<HashSet<_>>();
|
||||
|
||||
let to_device_events = services
|
||||
.users
|
||||
.get_to_device_events(sender_user, sender_device, Some(since), Some(next_batch))
|
||||
.map(at!(1))
|
||||
.collect::<Vec<_>>();
|
||||
let to_device_events: OptionFuture<_> = sender_device
|
||||
.map(|sender_device| {
|
||||
services
|
||||
.users
|
||||
.get_to_device_events(sender_user, sender_device, Some(since), Some(next_batch))
|
||||
.map(at!(1))
|
||||
.collect::<Vec<_>>()
|
||||
})
|
||||
.into();
|
||||
|
||||
let device_one_time_keys_count = services
|
||||
.users
|
||||
.count_one_time_keys(sender_user, sender_device);
|
||||
let device_one_time_keys_count: OptionFuture<_> = sender_device
|
||||
.map(|sender_device| {
|
||||
services
|
||||
.users
|
||||
.count_one_time_keys(sender_user, sender_device)
|
||||
})
|
||||
.into();
|
||||
|
||||
// Remove all to-device events the device received *last time*
|
||||
let remove_to_device_events =
|
||||
services
|
||||
.users
|
||||
.remove_to_device_events(sender_user, sender_device, since);
|
||||
let remove_to_device_events: OptionFuture<_> = sender_device
|
||||
.map(|sender_device| {
|
||||
services
|
||||
.users
|
||||
.remove_to_device_events(sender_user, sender_device, since)
|
||||
})
|
||||
.into();
|
||||
|
||||
let (
|
||||
account_data,
|
||||
keys_changed,
|
||||
device_one_time_keys_count,
|
||||
((), to_device_events, presence_updates),
|
||||
presence_updates,
|
||||
(_, to_device_events, device_one_time_keys_count),
|
||||
(
|
||||
(joined_rooms, mut device_list_updates, left_encrypted_users),
|
||||
left_rooms,
|
||||
@@ -417,8 +441,8 @@ async fn build_sync_events(
|
||||
) = join5(
|
||||
account_data,
|
||||
keys_changed,
|
||||
device_one_time_keys_count,
|
||||
join3(remove_to_device_events, to_device_events, presence_updates),
|
||||
presence_updates,
|
||||
join3(remove_to_device_events, to_device_events, device_one_time_keys_count),
|
||||
join4(joined_rooms, left_rooms, invited_rooms, knocked_rooms),
|
||||
)
|
||||
.boxed()
|
||||
@@ -454,7 +478,7 @@ async fn build_sync_events(
|
||||
left: device_list_left,
|
||||
changed: device_list_updates.into_iter().collect(),
|
||||
},
|
||||
device_one_time_keys_count,
|
||||
device_one_time_keys_count: device_one_time_keys_count.unwrap_or_default(),
|
||||
// Fallback keys are not yet supported
|
||||
device_unused_fallback_key_types: None,
|
||||
next_batch: to_small_string(next_batch),
|
||||
@@ -465,7 +489,9 @@ async fn build_sync_events(
|
||||
invite: invited_rooms,
|
||||
knock: knocked_rooms,
|
||||
},
|
||||
to_device: ToDevice { events: to_device_events },
|
||||
to_device: ToDevice {
|
||||
events: to_device_events.unwrap_or_default(),
|
||||
},
|
||||
})
|
||||
}
|
||||
|
||||
@@ -732,7 +758,7 @@ async fn load_left_room(
|
||||
async fn load_joined_room(
|
||||
services: &Services,
|
||||
sender_user: &UserId,
|
||||
sender_device: &DeviceId,
|
||||
sender_device: Option<&DeviceId>,
|
||||
ref room_id: OwnedRoomId,
|
||||
since: u64,
|
||||
next_batch: u64,
|
||||
@@ -841,7 +867,7 @@ async fn load_joined_room(
|
||||
|
||||
let lazy_loading_context = &lazy_loading::Context {
|
||||
user_id: sender_user,
|
||||
device_id: Some(sender_device),
|
||||
device_id: sender_device,
|
||||
room_id,
|
||||
token: Some(since),
|
||||
options: Some(&filter.room.state.lazy_load_options),
|
||||
|
||||
@@ -37,7 +37,7 @@ use crate::Ruma;
|
||||
struct SyncInfo<'a> {
|
||||
services: &'a Services,
|
||||
sender_user: &'a UserId,
|
||||
sender_device: &'a DeviceId,
|
||||
sender_device: Option<&'a DeviceId>,
|
||||
}
|
||||
|
||||
#[derive(Clone, Debug)]
|
||||
@@ -69,7 +69,7 @@ type ListIds = SmallVec<[ListId; 1]>;
|
||||
skip_all,
|
||||
fields(
|
||||
user_id = %body.sender_user().localpart(),
|
||||
device_id = %body.sender_device(),
|
||||
device_id = %body.sender_device.as_deref().map_or("<no device>", |x| x.as_str()),
|
||||
conn_id = ?body.body.conn_id.clone().unwrap_or_default(),
|
||||
since = ?body.body.pos.clone().unwrap_or_default(),
|
||||
)
|
||||
@@ -78,7 +78,8 @@ pub(crate) async fn sync_events_v5_route(
|
||||
State(ref services): State<crate::State>,
|
||||
body: Ruma<Request>,
|
||||
) -> Result<Response> {
|
||||
let (sender_user, sender_device) = body.sender();
|
||||
let sender_user = body.sender_user();
|
||||
let sender_device = body.sender_device.as_deref();
|
||||
let request = &body.body;
|
||||
let since = request
|
||||
.pos
|
||||
@@ -108,7 +109,7 @@ pub(crate) async fn sync_events_v5_route(
|
||||
let conn = conn_val.lock();
|
||||
let ping_presence = services
|
||||
.presence
|
||||
.maybe_ping_presence(sender_user, Some(sender_device), &request.set_presence)
|
||||
.maybe_ping_presence(sender_user, sender_device, &request.set_presence)
|
||||
.inspect_err(inspect_log)
|
||||
.ok();
|
||||
|
||||
|
||||
@@ -32,6 +32,9 @@ pub(super) async fn collect(
|
||||
conn: &Connection,
|
||||
) -> Result<response::E2EE> {
|
||||
let SyncInfo { services, sender_user, sender_device, .. } = sync_info;
|
||||
let Some(sender_device) = sender_device else {
|
||||
return Ok(response::E2EE::default());
|
||||
};
|
||||
|
||||
let keys_changed = services
|
||||
.users
|
||||
|
||||
@@ -9,6 +9,10 @@ pub(super) async fn collect(
|
||||
SyncInfo { services, sender_user, sender_device, .. }: SyncInfo<'_>,
|
||||
conn: &Connection,
|
||||
) -> Result<Option<response::ToDevice>> {
|
||||
let Some(sender_device) = sender_device else {
|
||||
return Ok(None);
|
||||
};
|
||||
|
||||
services
|
||||
.users
|
||||
.remove_to_device_events(sender_user, sender_device, conn.globalsince)
|
||||
|
||||
Reference in New Issue
Block a user