Merge remote-tracking branch 'origin/develop' into matrix-org-hotfixes

This commit is contained in:
Erik Johnston 2024-08-14 12:42:35 +01:00
commit 829f2d0605
9 changed files with 397 additions and 109 deletions

1
changelog.d/17538.bugfix Normal file
View file

@ -0,0 +1 @@
Better exclude partially stated rooms if we must await full state in experimental [MSC3575](https://github.com/matrix-org/matrix-spec-proposals/pull/3575) Sliding Sync `/sync` endpoint.

1
changelog.d/17545.bugfix Normal file
View file

@ -0,0 +1 @@
Handle lower-case http headers in `_Mulitpart_Parser_Protocol`.

1
changelog.d/17567.misc Normal file
View file

@ -0,0 +1 @@
Speed up responding to media requests.

View file

@ -24,6 +24,7 @@ from itertools import chain
from typing import ( from typing import (
TYPE_CHECKING, TYPE_CHECKING,
Any, Any,
Callable,
Dict, Dict,
Final, Final,
List, List,
@ -366,6 +367,73 @@ class RoomSyncConfig:
else: else:
self.required_state_map[state_type].add(state_key) self.required_state_map[state_type].add(state_key)
def must_await_full_state(
self,
is_mine_id: Callable[[str], bool],
) -> bool:
"""
Check if we have a we're only requesting `required_state` which is completely
satisfied even with partial state, then we don't need to `await_full_state` before
we can return it.
Also see `StateFilter.must_await_full_state(...)` for comparison
Partially-stated rooms should have all state events except for remote membership
events so if we require a remote membership event anywhere, then we need to
return `True` (requires full state).
Args:
is_mine_id: a callable which confirms if a given state_key matches a mxid
of a local user
"""
wildcard_state_keys = self.required_state_map.get(StateValues.WILDCARD)
# Requesting *all* state in the room so we have to wait
if (
wildcard_state_keys is not None
and StateValues.WILDCARD in wildcard_state_keys
):
return True
# If the wildcards don't refer to remote user IDs, then we don't need to wait
# for full state.
if wildcard_state_keys is not None:
for possible_user_id in wildcard_state_keys:
if not possible_user_id[0].startswith(UserID.SIGIL):
# Not a user ID
continue
localpart_hostname = possible_user_id.split(":", 1)
if len(localpart_hostname) < 2:
# Not a user ID
continue
if not is_mine_id(possible_user_id):
return True
membership_state_keys = self.required_state_map.get(EventTypes.Member)
# We aren't requesting any membership events at all so the partial state will
# cover us.
if membership_state_keys is None:
return False
# If we're requesting entirely local users, the partial state will cover us.
for user_id in membership_state_keys:
if user_id == StateValues.ME:
continue
# We're lazy-loading membership so we can just return the state we have.
# Lazy-loading means we include membership for any event `sender` in the
# timeline but since we had to auth those timeline events, we will have the
# membership state for them (including from remote senders).
elif user_id == StateValues.LAZY:
continue
elif user_id == StateValues.WILDCARD:
return False
elif not is_mine_id(user_id):
return True
# Local users only so the partial state will cover us.
return False
class StateValues: class StateValues:
""" """
@ -395,6 +463,7 @@ class SlidingSyncHandler:
self.device_handler = hs.get_device_handler() self.device_handler = hs.get_device_handler()
self.push_rules_handler = hs.get_push_rules_handler() self.push_rules_handler = hs.get_push_rules_handler()
self.rooms_to_exclude_globally = hs.config.server.rooms_to_exclude_from_sync self.rooms_to_exclude_globally = hs.config.server.rooms_to_exclude_from_sync
self.is_mine_id = hs.is_mine_id
self.connection_store = SlidingSyncConnectionStore() self.connection_store = SlidingSyncConnectionStore()
@ -575,19 +644,10 @@ class SlidingSyncHandler:
# Since creating the `RoomSyncConfig` takes some work, let's just do it # Since creating the `RoomSyncConfig` takes some work, let's just do it
# once and make a copy whenever we need it. # once and make a copy whenever we need it.
room_sync_config = RoomSyncConfig.from_room_config(list_config) room_sync_config = RoomSyncConfig.from_room_config(list_config)
membership_state_keys = room_sync_config.required_state_map.get(
EventTypes.Member
)
# Also see `StateFilter.must_await_full_state(...)` for comparison
lazy_loading = (
membership_state_keys is not None
and StateValues.LAZY in membership_state_keys
)
if not lazy_loading: # Exclude partially-stated rooms if we must wait for the room to be
# Exclude partially-stated rooms unless the `required_state` # fully-stated
# only has `["m.room.member", "$LAZY"]` for membership if room_sync_config.must_await_full_state(self.is_mine_id):
# (lazy-loading room members).
filtered_sync_room_map = { filtered_sync_room_map = {
room_id: room room_id: room
for room_id, room in filtered_sync_room_map.items() for room_id, room in filtered_sync_room_map.items()
@ -654,6 +714,12 @@ class SlidingSyncHandler:
# Handle room subscriptions # Handle room subscriptions
if has_room_subscriptions and sync_config.room_subscriptions is not None: if has_room_subscriptions and sync_config.room_subscriptions is not None:
with start_active_span("assemble_room_subscriptions"): with start_active_span("assemble_room_subscriptions"):
# Find which rooms are partially stated and may need to be filtered out
# depending on the `required_state` requested (see below).
partial_state_room_map = await self.store.is_partial_state_room_batched(
sync_config.room_subscriptions.keys()
)
for ( for (
room_id, room_id,
room_subscription, room_subscription,
@ -677,12 +743,20 @@ class SlidingSyncHandler:
) )
# Take the superset of the `RoomSyncConfig` for each room. # Take the superset of the `RoomSyncConfig` for each room.
#
# Update our `relevant_room_map` with the room we're going to display
# and need to fetch more info about.
room_sync_config = RoomSyncConfig.from_room_config( room_sync_config = RoomSyncConfig.from_room_config(
room_subscription room_subscription
) )
# Exclude partially-stated rooms if we must wait for the room to be
# fully-stated
if room_sync_config.must_await_full_state(self.is_mine_id):
if partial_state_room_map.get(room_id):
continue
all_rooms.add(room_id)
# Update our `relevant_room_map` with the room we're going to display
# and need to fetch more info about.
existing_room_sync_config = relevant_room_map.get(room_id) existing_room_sync_config = relevant_room_map.get(room_id)
if existing_room_sync_config is not None: if existing_room_sync_config is not None:
existing_room_sync_config.combine_room_sync_config( existing_room_sync_config.combine_room_sync_config(

View file

@ -1057,11 +1057,11 @@ class _MultipartParserProtocol(protocol.Protocol):
if not self.parser: if not self.parser:
def on_header_field(data: bytes, start: int, end: int) -> None: def on_header_field(data: bytes, start: int, end: int) -> None:
if data[start:end] == b"Location": if data[start:end].lower() == b"location":
self.has_redirect = True self.has_redirect = True
if data[start:end] == b"Content-Disposition": if data[start:end].lower() == b"content-disposition":
self.in_disposition = True self.in_disposition = True
if data[start:end] == b"Content-Type": if data[start:end].lower() == b"content-type":
self.in_content_type = True self.in_content_type = True
def on_header_value(data: bytes, start: int, end: int) -> None: def on_header_value(data: bytes, start: int, end: int) -> None:

View file

@ -22,7 +22,6 @@
import logging import logging
import os import os
import threading
import urllib import urllib
from abc import ABC, abstractmethod from abc import ABC, abstractmethod
from types import TracebackType from types import TracebackType
@ -56,6 +55,7 @@ from synapse.logging.context import (
run_in_background, run_in_background,
) )
from synapse.util import Clock from synapse.util import Clock
from synapse.util.async_helpers import DeferredEvent
from synapse.util.stringutils import is_ascii from synapse.util.stringutils import is_ascii
if TYPE_CHECKING: if TYPE_CHECKING:
@ -620,10 +620,13 @@ class ThreadedFileSender:
A producer that sends the contents of a file to a consumer, reading from the A producer that sends the contents of a file to a consumer, reading from the
file on a thread. file on a thread.
This works by spawning a loop in a threadpool that repeatedly reads from the This works by having a loop in a threadpool repeatedly reading from the
file and sends it to the consumer. The main thread communicates with the file, until the consumer pauses the producer. There is then a loop in the
loop via two `threading.Event`, which controls when to start/pause reading main thread that waits until the consumer resumes the producer and then
and when to terminate. starts reading in the threadpool again.
This is done to ensure that we're never waiting in the threadpool, as
otherwise its easy to starve it of threads.
""" """
# How much data to read in one go. # How much data to read in one go.
@ -643,12 +646,11 @@ class ThreadedFileSender:
# Signals if the thread should keep reading/sending data. Set means # Signals if the thread should keep reading/sending data. Set means
# continue, clear means pause. # continue, clear means pause.
self.wakeup_event = threading.Event() self.wakeup_event = DeferredEvent(self.reactor)
# Signals if the thread should terminate, e.g. because the consumer has # Signals if the thread should terminate, e.g. because the consumer has
# gone away. Both this and `wakeup_event` should be set to terminate the # gone away.
# loop (otherwise the thread will block on `wakeup_event`). self.stop_writing = False
self.stop_event = threading.Event()
def beginFileTransfer( def beginFileTransfer(
self, file: BinaryIO, consumer: interfaces.IConsumer self, file: BinaryIO, consumer: interfaces.IConsumer
@ -663,12 +665,7 @@ class ThreadedFileSender:
# We set the wakeup signal as we should start producing immediately. # We set the wakeup signal as we should start producing immediately.
self.wakeup_event.set() self.wakeup_event.set()
run_in_background( run_in_background(self.start_read_loop)
defer_to_threadpool,
self.reactor,
self.thread_pool,
self._on_thread_read_loop,
)
return make_deferred_yieldable(self.deferred) return make_deferred_yieldable(self.deferred)
@ -686,42 +683,52 @@ class ThreadedFileSender:
# Unregister the consumer so we don't try and interact with it again. # Unregister the consumer so we don't try and interact with it again.
self.consumer = None self.consumer = None
# Terminate the thread loop. # Terminate the loop.
self.stop_writing = True
self.wakeup_event.set() self.wakeup_event.set()
self.stop_event.set()
if not self.deferred.called: if not self.deferred.called:
self.deferred.errback(Exception("Consumer asked us to stop producing")) self.deferred.errback(Exception("Consumer asked us to stop producing"))
def _on_thread_read_loop(self) -> None: async def start_read_loop(self) -> None:
"""This is the loop that happens on a thread.""" """This is the loop that drives reading/writing"""
try: try:
while not self.stop_event.is_set(): while not self.stop_writing:
# We wait for the producer to signal that the consumer wants # Start the loop in the threadpool to read data.
# more data (or we should abort) more_data = await defer_to_threadpool(
if not self.wakeup_event.is_set(): self.reactor, self.thread_pool, self._on_thread_read_loop
ret = self.wakeup_event.wait(self.TIMEOUT_SECONDS) )
if not ret: if not more_data:
raise Exception("Timed out waiting to resume") # Reached EOF, we can just return.
# Check if we were woken up so that we abort the download
if self.stop_event.is_set():
return return
if not self.wakeup_event.is_set():
ret = await self.wakeup_event.wait(self.TIMEOUT_SECONDS)
if not ret:
raise Exception("Timed out waiting to resume")
except Exception:
self._error(Failure())
finally:
self._finish()
def _on_thread_read_loop(self) -> bool:
"""This is the loop that happens on a thread.
Returns:
Whether there is more data to send.
"""
while not self.stop_writing and self.wakeup_event.is_set():
# The file should always have been set before we get here. # The file should always have been set before we get here.
assert self.file is not None assert self.file is not None
chunk = self.file.read(self.CHUNK_SIZE) chunk = self.file.read(self.CHUNK_SIZE)
if not chunk: if not chunk:
return return False
self.reactor.callFromThread(self._write, chunk) self.reactor.callFromThread(self._write, chunk)
except Exception: return True
self.reactor.callFromThread(self._error, Failure())
finally:
self.reactor.callFromThread(self._finish)
def _write(self, chunk: bytes) -> None: def _write(self, chunk: bytes) -> None:
"""Called from the thread to write a chunk of data""" """Called from the thread to write a chunk of data"""
@ -729,7 +736,7 @@ class ThreadedFileSender:
self.consumer.write(chunk) self.consumer.write(chunk)
def _error(self, failure: Failure) -> None: def _error(self, failure: Failure) -> None:
"""Called from the thread when there was a fatal error""" """Called when there was a fatal error"""
if self.consumer: if self.consumer:
self.consumer.unregisterProducer() self.consumer.unregisterProducer()
self.consumer = None self.consumer = None
@ -738,7 +745,7 @@ class ThreadedFileSender:
self.deferred.errback(failure) self.deferred.errback(failure)
def _finish(self) -> None: def _finish(self) -> None:
"""Called from the thread when it finishes (either on success or """Called when we have finished writing (either on success or
failure).""" failure)."""
if self.file: if self.file:
self.file.close() self.file.close()

View file

@ -885,3 +885,46 @@ class AwakenableSleeper:
# Cancel the sleep if we were woken up # Cancel the sleep if we were woken up
if call.active(): if call.active():
call.cancel() call.cancel()
class DeferredEvent:
"""Like threading.Event but for async code"""
def __init__(self, reactor: IReactorTime) -> None:
self._reactor = reactor
self._deferred: "defer.Deferred[None]" = defer.Deferred()
def set(self) -> None:
if not self._deferred.called:
self._deferred.callback(None)
def clear(self) -> None:
if self._deferred.called:
self._deferred = defer.Deferred()
def is_set(self) -> bool:
return self._deferred.called
async def wait(self, timeout_seconds: float) -> bool:
if self.is_set():
return True
# Create a deferred that gets called in N seconds
sleep_deferred: "defer.Deferred[None]" = defer.Deferred()
call = self._reactor.callLater(timeout_seconds, sleep_deferred.callback, None)
try:
await make_deferred_yieldable(
defer.DeferredList(
[sleep_deferred, self._deferred],
fireOnOneCallback=True,
fireOnOneErrback=True,
consumeErrors=True,
)
)
finally:
# Cancel the sleep if we were woken up
if call.active():
call.cancel()
return self.is_set()

View file

@ -49,8 +49,11 @@ from tests.unittest import TestCase
class ReadMultipartResponseTests(TestCase): class ReadMultipartResponseTests(TestCase):
data1 = b"\r\n\r\n--6067d4698f8d40a0a794ea7d7379d53a\r\nContent-Type: application/json\r\n\r\n{}\r\n--6067d4698f8d40a0a794ea7d7379d53a\r\nContent-Type: text/plain\r\nContent-Disposition: inline; filename=test_upload\r\n\r\nfile_" multipart_response_data1 = b"\r\n\r\n--6067d4698f8d40a0a794ea7d7379d53a\r\nContent-Type: application/json\r\n\r\n{}\r\n--6067d4698f8d40a0a794ea7d7379d53a\r\nContent-Type: text/plain\r\nContent-Disposition: inline; filename=test_upload\r\n\r\nfile_"
data2 = b"to_stream\r\n--6067d4698f8d40a0a794ea7d7379d53a--\r\n\r\n" multipart_response_data2 = (
b"to_stream\r\n--6067d4698f8d40a0a794ea7d7379d53a--\r\n\r\n"
)
multipart_response_data_cased = b"\r\n\r\n--6067d4698f8d40a0a794ea7d7379d53a\r\ncOntEnt-type: application/json\r\n\r\n{}\r\n--6067d4698f8d40a0a794ea7d7379d53a\r\nContent-tyPe: text/plain\r\nconTent-dispOsition: inline; filename=test_upload\r\n\r\nfile_"
redirect_data = b"\r\n\r\n--6067d4698f8d40a0a794ea7d7379d53a\r\nContent-Type: application/json\r\n\r\n{}\r\n--6067d4698f8d40a0a794ea7d7379d53a\r\nLocation: https://cdn.example.org/ab/c1/2345.txt\r\n\r\n--6067d4698f8d40a0a794ea7d7379d53a--\r\n\r\n" redirect_data = b"\r\n\r\n--6067d4698f8d40a0a794ea7d7379d53a\r\nContent-Type: application/json\r\n\r\n{}\r\n--6067d4698f8d40a0a794ea7d7379d53a\r\nLocation: https://cdn.example.org/ab/c1/2345.txt\r\n\r\n--6067d4698f8d40a0a794ea7d7379d53a--\r\n\r\n"
@ -103,8 +106,31 @@ class ReadMultipartResponseTests(TestCase):
result, deferred, protocol = self._build_multipart_response(249, 250) result, deferred, protocol = self._build_multipart_response(249, 250)
# Start sending data. # Start sending data.
protocol.dataReceived(self.data1) protocol.dataReceived(self.multipart_response_data1)
protocol.dataReceived(self.data2) protocol.dataReceived(self.multipart_response_data2)
# Close the connection.
protocol.connectionLost(Failure(ResponseDone()))
multipart_response: MultipartResponse = deferred.result # type: ignore[assignment]
self.assertEqual(multipart_response.json, b"{}")
self.assertEqual(result.getvalue(), b"file_to_stream")
self.assertEqual(multipart_response.length, len(b"file_to_stream"))
self.assertEqual(multipart_response.content_type, b"text/plain")
self.assertEqual(
multipart_response.disposition, b"inline; filename=test_upload"
)
def test_parse_file_lowercase_headers(self) -> None:
"""
Check that a multipart response containing a file is properly parsed
into the json/file parts, and the json and file are properly captured if the http headers are lowercased
"""
result, deferred, protocol = self._build_multipart_response(249, 250)
# Start sending data.
protocol.dataReceived(self.multipart_response_data_cased)
protocol.dataReceived(self.multipart_response_data2)
# Close the connection. # Close the connection.
protocol.connectionLost(Failure(ResponseDone())) protocol.connectionLost(Failure(ResponseDone()))
@ -143,7 +169,7 @@ class ReadMultipartResponseTests(TestCase):
result, deferred, protocol = self._build_multipart_response(UNKNOWN_LENGTH, 180) result, deferred, protocol = self._build_multipart_response(UNKNOWN_LENGTH, 180)
# Start sending data. # Start sending data.
protocol.dataReceived(self.data1) protocol.dataReceived(self.multipart_response_data1)
self.assertEqual(result.getvalue(), b"file_") self.assertEqual(result.getvalue(), b"file_")
self._assert_error(deferred, protocol) self._assert_error(deferred, protocol)
@ -154,11 +180,11 @@ class ReadMultipartResponseTests(TestCase):
result, deferred, protocol = self._build_multipart_response(UNKNOWN_LENGTH, 180) result, deferred, protocol = self._build_multipart_response(UNKNOWN_LENGTH, 180)
# Start sending data. # Start sending data.
protocol.dataReceived(self.data1) protocol.dataReceived(self.multipart_response_data1)
self._assert_error(deferred, protocol) self._assert_error(deferred, protocol)
# More data might have come in. # More data might have come in.
protocol.dataReceived(self.data2) protocol.dataReceived(self.multipart_response_data2)
self.assertEqual(result.getvalue(), b"file_") self.assertEqual(result.getvalue(), b"file_")
self._assert_error(deferred, protocol) self._assert_error(deferred, protocol)
@ -172,7 +198,7 @@ class ReadMultipartResponseTests(TestCase):
self.assertFalse(deferred.called) self.assertFalse(deferred.called)
# Start sending data. # Start sending data.
protocol.dataReceived(self.data1) protocol.dataReceived(self.multipart_response_data1)
self._assert_error(deferred, protocol) self._assert_error(deferred, protocol)
self._cleanup_error(deferred) self._cleanup_error(deferred)

View file

@ -631,8 +631,7 @@ class SlidingSyncRoomsRequiredStateTestCase(SlidingSyncBase):
def test_rooms_required_state_partial_state(self) -> None: def test_rooms_required_state_partial_state(self) -> None:
""" """
Test partially-stated room are excluded unless `rooms.required_state` is Test partially-stated room are excluded if they require full state.
lazy-loading room members.
""" """
user1_id = self.register_user("user1", "pass") user1_id = self.register_user("user1", "pass")
user1_tok = self.login(user1_id, "pass") user1_tok = self.login(user1_id, "pass")
@ -649,59 +648,195 @@ class SlidingSyncRoomsRequiredStateTestCase(SlidingSyncBase):
mark_event_as_partial_state(self.hs, join_response2["event_id"], room_id2) mark_event_as_partial_state(self.hs, join_response2["event_id"], room_id2)
) )
# Make the Sliding Sync request (NOT lazy-loading room members) # Make the Sliding Sync request with examples where `must_await_full_state()` is
# `False`
sync_body = { sync_body = {
"lists": { "lists": {
"foo-list": { "no-state-list": {
"ranges": [[0, 1]],
"required_state": [],
"timeline_limit": 0,
},
"other-state-list": {
"ranges": [[0, 1]], "ranges": [[0, 1]],
"required_state": [ "required_state": [
[EventTypes.Create, ""], [EventTypes.Create, ""],
], ],
"timeline_limit": 0, "timeline_limit": 0,
}, },
"lazy-load-list": {
"ranges": [[0, 1]],
"required_state": [
[EventTypes.Create, ""],
# Lazy-load room members
[EventTypes.Member, StateValues.LAZY],
# Local member
[EventTypes.Member, user2_id],
],
"timeline_limit": 0,
},
"local-members-only-list": {
"ranges": [[0, 1]],
"required_state": [
# Own user ID
[EventTypes.Member, user1_id],
# Local member
[EventTypes.Member, user2_id],
],
"timeline_limit": 0,
},
"me-list": {
"ranges": [[0, 1]],
"required_state": [
# Own user ID
[EventTypes.Member, StateValues.ME],
# Local member
[EventTypes.Member, user2_id],
],
"timeline_limit": 0,
},
"wildcard-type-local-state-key-list": {
"ranges": [[0, 1]],
"required_state": [
["*", user1_id],
# Not a user ID
["*", "foobarbaz"],
# Not a user ID
["*", "foo.bar.baz"],
# Not a user ID
["*", "@foo"],
],
"timeline_limit": 0,
},
}
}
response_body, _ = self.do_sync(sync_body, tok=user1_tok)
# The list should include both rooms now because we don't need full state
for list_key in response_body["lists"].keys():
self.assertIncludes(
set(response_body["lists"][list_key]["ops"][0]["room_ids"]),
{room_id2, room_id1},
exact=True,
message=f"Expected all rooms to show up for list_key={list_key}. Response "
+ str(response_body["lists"][list_key]),
)
# Take each of the list variants and apply them to room subscriptions to make
# sure the same rules apply
for list_key in sync_body["lists"].keys():
sync_body_for_subscriptions = {
"room_subscriptions": {
room_id1: {
"required_state": sync_body["lists"][list_key][
"required_state"
],
"timeline_limit": 0,
},
room_id2: {
"required_state": sync_body["lists"][list_key][
"required_state"
],
"timeline_limit": 0,
},
}
}
response_body, _ = self.do_sync(sync_body_for_subscriptions, tok=user1_tok)
self.assertIncludes(
set(response_body["rooms"].keys()),
{room_id2, room_id1},
exact=True,
message=f"Expected all rooms to show up for test_key={list_key}.",
)
# =====================================================================
# Make the Sliding Sync request with examples where `must_await_full_state()` is
# `True`
sync_body = {
"lists": {
"wildcard-list": {
"ranges": [[0, 1]],
"required_state": [
["*", "*"],
],
"timeline_limit": 0,
},
"wildcard-type-remote-state-key-list": {
"ranges": [[0, 1]],
"required_state": [
["*", "@some:remote"],
# Not a user ID
["*", "foobarbaz"],
# Not a user ID
["*", "foo.bar.baz"],
# Not a user ID
["*", "@foo"],
],
"timeline_limit": 0,
},
"remote-member-list": {
"ranges": [[0, 1]],
"required_state": [
# Own user ID
[EventTypes.Member, user1_id],
# Remote member
[EventTypes.Member, "@some:remote"],
# Local member
[EventTypes.Member, user2_id],
],
"timeline_limit": 0,
},
"lazy-but-remote-member-list": {
"ranges": [[0, 1]],
"required_state": [
# Lazy-load room members
[EventTypes.Member, StateValues.LAZY],
# Remote member
[EventTypes.Member, "@some:remote"],
],
"timeline_limit": 0,
},
} }
} }
response_body, _ = self.do_sync(sync_body, tok=user1_tok) response_body, _ = self.do_sync(sync_body, tok=user1_tok)
# Make sure the list includes room1 but room2 is excluded because it's still # Make sure the list includes room1 but room2 is excluded because it's still
# partially-stated # partially-stated
self.assertListEqual( for list_key in response_body["lists"].keys():
list(response_body["lists"]["foo-list"]["ops"]), self.assertIncludes(
[ set(response_body["lists"][list_key]["ops"][0]["room_ids"]),
{ {room_id1},
"op": "SYNC", exact=True,
"range": [0, 1], message=f"Expected only fully-stated rooms to show up for list_key={list_key}. Response "
"room_ids": [room_id1], + str(response_body["lists"][list_key]),
}
],
response_body["lists"]["foo-list"],
) )
# Make the Sliding Sync request (with lazy-loading room members) # Take each of the list variants and apply them to room subscriptions to make
sync_body = { # sure the same rules apply
"lists": { for list_key in sync_body["lists"].keys():
"foo-list": { sync_body_for_subscriptions = {
"ranges": [[0, 1]], "room_subscriptions": {
"required_state": [ room_id1: {
[EventTypes.Create, ""], "required_state": sync_body["lists"][list_key][
# Lazy-load room members "required_state"
[EventTypes.Member, StateValues.LAZY], ],
"timeline_limit": 0,
},
room_id2: {
"required_state": sync_body["lists"][list_key][
"required_state"
], ],
"timeline_limit": 0, "timeline_limit": 0,
}, },
} }
} }
response_body, _ = self.do_sync(sync_body, tok=user1_tok) response_body, _ = self.do_sync(sync_body_for_subscriptions, tok=user1_tok)
# The list should include both rooms now because we're lazy-loading room members self.assertIncludes(
self.assertListEqual( set(response_body["rooms"].keys()),
list(response_body["lists"]["foo-list"]["ops"]), {room_id1},
[ exact=True,
{ message=f"Expected only fully-stated rooms to show up for test_key={list_key}.",
"op": "SYNC",
"range": [0, 1],
"room_ids": [room_id2, room_id1],
}
],
response_body["lists"]["foo-list"],
) )