mirror of
https://github.com/element-hq/synapse.git
synced 2024-11-22 17:46:08 +03:00
75dff3dc98
The `latest_event` field of the bundled aggregations for `m.thread` relations did not include bundled aggregations itself. This resulted in clients needing to immediately request the event from the server (and thus making it useless that the latest event itself was serialized instead of just including an event ID).
1614 lines
61 KiB
Python
1614 lines
61 KiB
Python
# Copyright 2019 New Vector Ltd
|
|
# Copyright 2021 The Matrix.org Foundation C.I.C.
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
# you may not use this file except in compliance with the License.
|
|
# You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
|
|
import urllib.parse
|
|
from typing import Any, Callable, Dict, List, Optional, Tuple
|
|
from unittest.mock import patch
|
|
|
|
from twisted.test.proto_helpers import MemoryReactor
|
|
|
|
from synapse.api.constants import AccountDataTypes, EventTypes, RelationTypes
|
|
from synapse.rest import admin
|
|
from synapse.rest.client import login, register, relations, room, sync
|
|
from synapse.server import HomeServer
|
|
from synapse.types import JsonDict
|
|
from synapse.util import Clock
|
|
|
|
from tests import unittest
|
|
from tests.server import FakeChannel
|
|
from tests.test_utils import make_awaitable
|
|
from tests.test_utils.event_injection import inject_event
|
|
|
|
|
|
class BaseRelationsTestCase(unittest.HomeserverTestCase):
|
|
servlets = [
|
|
relations.register_servlets,
|
|
room.register_servlets,
|
|
sync.register_servlets,
|
|
login.register_servlets,
|
|
register.register_servlets,
|
|
admin.register_servlets_for_client_rest_resource,
|
|
]
|
|
hijack_auth = False
|
|
|
|
def default_config(self) -> Dict[str, Any]:
|
|
# We need to enable msc1849 support for aggregations
|
|
config = super().default_config()
|
|
|
|
# We enable frozen dicts as relations/edits change event contents, so we
|
|
# want to test that we don't modify the events in the caches.
|
|
config["use_frozen_dicts"] = True
|
|
|
|
return config
|
|
|
|
def prepare(self, reactor: MemoryReactor, clock: Clock, hs: HomeServer) -> None:
|
|
self.store = hs.get_datastores().main
|
|
|
|
self.user_id, self.user_token = self._create_user("alice")
|
|
self.user2_id, self.user2_token = self._create_user("bob")
|
|
|
|
self.room = self.helper.create_room_as(self.user_id, tok=self.user_token)
|
|
self.helper.join(self.room, user=self.user2_id, tok=self.user2_token)
|
|
res = self.helper.send(self.room, body="Hi!", tok=self.user_token)
|
|
self.parent_id = res["event_id"]
|
|
|
|
def _create_user(self, localpart: str) -> Tuple[str, str]:
|
|
user_id = self.register_user(localpart, "abc123")
|
|
access_token = self.login(localpart, "abc123")
|
|
|
|
return user_id, access_token
|
|
|
|
def _send_relation(
|
|
self,
|
|
relation_type: str,
|
|
event_type: str,
|
|
key: Optional[str] = None,
|
|
content: Optional[dict] = None,
|
|
access_token: Optional[str] = None,
|
|
parent_id: Optional[str] = None,
|
|
expected_response_code: int = 200,
|
|
) -> FakeChannel:
|
|
"""Helper function to send a relation pointing at `self.parent_id`
|
|
|
|
Args:
|
|
relation_type: One of `RelationTypes`
|
|
event_type: The type of the event to create
|
|
key: The aggregation key used for m.annotation relation type.
|
|
content: The content of the created event. Will be modified to configure
|
|
the m.relates_to key based on the other provided parameters.
|
|
access_token: The access token used to send the relation, defaults
|
|
to `self.user_token`
|
|
parent_id: The event_id this relation relates to. If None, then self.parent_id
|
|
|
|
Returns:
|
|
FakeChannel
|
|
"""
|
|
if not access_token:
|
|
access_token = self.user_token
|
|
|
|
original_id = parent_id if parent_id else self.parent_id
|
|
|
|
if content is None:
|
|
content = {}
|
|
content["m.relates_to"] = {
|
|
"event_id": original_id,
|
|
"rel_type": relation_type,
|
|
}
|
|
if key is not None:
|
|
content["m.relates_to"]["key"] = key
|
|
|
|
channel = self.make_request(
|
|
"POST",
|
|
f"/_matrix/client/v3/rooms/{self.room}/send/{event_type}",
|
|
content,
|
|
access_token=access_token,
|
|
)
|
|
self.assertEqual(expected_response_code, channel.code, channel.json_body)
|
|
return channel
|
|
|
|
def _get_related_events(self) -> List[str]:
|
|
"""
|
|
Requests /relations on the parent ID and returns a list of event IDs.
|
|
"""
|
|
# Request the relations of the event.
|
|
channel = self.make_request(
|
|
"GET",
|
|
f"/_matrix/client/v1/rooms/{self.room}/relations/{self.parent_id}",
|
|
access_token=self.user_token,
|
|
)
|
|
self.assertEquals(200, channel.code, channel.json_body)
|
|
return [ev["event_id"] for ev in channel.json_body["chunk"]]
|
|
|
|
def _get_bundled_aggregations(self) -> JsonDict:
|
|
"""
|
|
Requests /event on the parent ID and returns the m.relations field (from unsigned), if it exists.
|
|
"""
|
|
# Fetch the bundled aggregations of the event.
|
|
channel = self.make_request(
|
|
"GET",
|
|
f"/_matrix/client/v3/rooms/{self.room}/event/{self.parent_id}",
|
|
access_token=self.user_token,
|
|
)
|
|
self.assertEquals(200, channel.code, channel.json_body)
|
|
return channel.json_body["unsigned"].get("m.relations", {})
|
|
|
|
def _find_event_in_chunk(self, events: List[JsonDict]) -> JsonDict:
|
|
"""
|
|
Find the parent event in a chunk of events and assert that it has the proper bundled aggregations.
|
|
"""
|
|
for event in events:
|
|
if event["event_id"] == self.parent_id:
|
|
return event
|
|
|
|
raise AssertionError(f"Event {self.parent_id} not found in chunk")
|
|
|
|
|
|
class RelationsTestCase(BaseRelationsTestCase):
|
|
def test_send_relation(self) -> None:
|
|
"""Tests that sending a relation works."""
|
|
channel = self._send_relation(RelationTypes.ANNOTATION, "m.reaction", key="👍")
|
|
event_id = channel.json_body["event_id"]
|
|
|
|
channel = self.make_request(
|
|
"GET",
|
|
f"/rooms/{self.room}/event/{event_id}",
|
|
access_token=self.user_token,
|
|
)
|
|
self.assertEqual(200, channel.code, channel.json_body)
|
|
|
|
self.assert_dict(
|
|
{
|
|
"type": "m.reaction",
|
|
"sender": self.user_id,
|
|
"content": {
|
|
"m.relates_to": {
|
|
"event_id": self.parent_id,
|
|
"key": "👍",
|
|
"rel_type": RelationTypes.ANNOTATION,
|
|
}
|
|
},
|
|
},
|
|
channel.json_body,
|
|
)
|
|
|
|
def test_deny_invalid_event(self) -> None:
|
|
"""Test that we deny relations on non-existant events"""
|
|
self._send_relation(
|
|
RelationTypes.ANNOTATION,
|
|
EventTypes.Message,
|
|
parent_id="foo",
|
|
content={"body": "foo", "msgtype": "m.text"},
|
|
expected_response_code=400,
|
|
)
|
|
|
|
# Unless that event is referenced from another event!
|
|
self.get_success(
|
|
self.hs.get_datastores().main.db_pool.simple_insert(
|
|
table="event_relations",
|
|
values={
|
|
"event_id": "bar",
|
|
"relates_to_id": "foo",
|
|
"relation_type": RelationTypes.THREAD,
|
|
},
|
|
desc="test_deny_invalid_event",
|
|
)
|
|
)
|
|
self._send_relation(
|
|
RelationTypes.THREAD,
|
|
EventTypes.Message,
|
|
parent_id="foo",
|
|
content={"body": "foo", "msgtype": "m.text"},
|
|
)
|
|
|
|
def test_deny_invalid_room(self) -> None:
|
|
"""Test that we deny relations on non-existant events"""
|
|
# Create another room and send a message in it.
|
|
room2 = self.helper.create_room_as(self.user_id, tok=self.user_token)
|
|
res = self.helper.send(room2, body="Hi!", tok=self.user_token)
|
|
parent_id = res["event_id"]
|
|
|
|
# Attempt to send an annotation to that event.
|
|
self._send_relation(
|
|
RelationTypes.ANNOTATION,
|
|
"m.reaction",
|
|
parent_id=parent_id,
|
|
key="A",
|
|
expected_response_code=400,
|
|
)
|
|
|
|
def test_deny_double_react(self) -> None:
|
|
"""Test that we deny relations on membership events"""
|
|
self._send_relation(RelationTypes.ANNOTATION, "m.reaction", key="a")
|
|
self._send_relation(
|
|
RelationTypes.ANNOTATION, "m.reaction", "a", expected_response_code=400
|
|
)
|
|
|
|
def test_deny_forked_thread(self) -> None:
|
|
"""It is invalid to start a thread off a thread."""
|
|
channel = self._send_relation(
|
|
RelationTypes.THREAD,
|
|
"m.room.message",
|
|
content={"msgtype": "m.text", "body": "foo"},
|
|
parent_id=self.parent_id,
|
|
)
|
|
parent_id = channel.json_body["event_id"]
|
|
|
|
self._send_relation(
|
|
RelationTypes.THREAD,
|
|
"m.room.message",
|
|
content={"msgtype": "m.text", "body": "foo"},
|
|
parent_id=parent_id,
|
|
expected_response_code=400,
|
|
)
|
|
|
|
def test_ignore_invalid_room(self) -> None:
|
|
"""Test that we ignore invalid relations over federation."""
|
|
# Create another room and send a message in it.
|
|
room2 = self.helper.create_room_as(self.user_id, tok=self.user_token)
|
|
res = self.helper.send(room2, body="Hi!", tok=self.user_token)
|
|
parent_id = res["event_id"]
|
|
|
|
# Disable the validation to pretend this came over federation.
|
|
with patch(
|
|
"synapse.handlers.message.EventCreationHandler._validate_event_relation",
|
|
new=lambda self, event: make_awaitable(None),
|
|
):
|
|
# Generate a various relations from a different room.
|
|
self.get_success(
|
|
inject_event(
|
|
self.hs,
|
|
room_id=self.room,
|
|
type="m.reaction",
|
|
sender=self.user_id,
|
|
content={
|
|
"m.relates_to": {
|
|
"rel_type": RelationTypes.ANNOTATION,
|
|
"event_id": parent_id,
|
|
"key": "A",
|
|
}
|
|
},
|
|
)
|
|
)
|
|
|
|
self.get_success(
|
|
inject_event(
|
|
self.hs,
|
|
room_id=self.room,
|
|
type="m.room.message",
|
|
sender=self.user_id,
|
|
content={
|
|
"body": "foo",
|
|
"msgtype": "m.text",
|
|
"m.relates_to": {
|
|
"rel_type": RelationTypes.REFERENCE,
|
|
"event_id": parent_id,
|
|
},
|
|
},
|
|
)
|
|
)
|
|
|
|
self.get_success(
|
|
inject_event(
|
|
self.hs,
|
|
room_id=self.room,
|
|
type="m.room.message",
|
|
sender=self.user_id,
|
|
content={
|
|
"body": "foo",
|
|
"msgtype": "m.text",
|
|
"m.relates_to": {
|
|
"rel_type": RelationTypes.THREAD,
|
|
"event_id": parent_id,
|
|
},
|
|
},
|
|
)
|
|
)
|
|
|
|
self.get_success(
|
|
inject_event(
|
|
self.hs,
|
|
room_id=self.room,
|
|
type="m.room.message",
|
|
sender=self.user_id,
|
|
content={
|
|
"body": "foo",
|
|
"msgtype": "m.text",
|
|
"new_content": {
|
|
"body": "new content",
|
|
"msgtype": "m.text",
|
|
},
|
|
"m.relates_to": {
|
|
"rel_type": RelationTypes.REPLACE,
|
|
"event_id": parent_id,
|
|
},
|
|
},
|
|
)
|
|
)
|
|
|
|
# They should be ignored when fetching relations.
|
|
channel = self.make_request(
|
|
"GET",
|
|
f"/_matrix/client/v1/rooms/{room2}/relations/{parent_id}",
|
|
access_token=self.user_token,
|
|
)
|
|
self.assertEqual(200, channel.code, channel.json_body)
|
|
self.assertEqual(channel.json_body["chunk"], [])
|
|
|
|
# And for bundled aggregations.
|
|
channel = self.make_request(
|
|
"GET",
|
|
f"/rooms/{room2}/event/{parent_id}",
|
|
access_token=self.user_token,
|
|
)
|
|
self.assertEqual(200, channel.code, channel.json_body)
|
|
self.assertNotIn("m.relations", channel.json_body["unsigned"])
|
|
|
|
def test_edit(self) -> None:
|
|
"""Test that a simple edit works."""
|
|
|
|
new_body = {"msgtype": "m.text", "body": "I've been edited!"}
|
|
channel = self._send_relation(
|
|
RelationTypes.REPLACE,
|
|
"m.room.message",
|
|
content={"msgtype": "m.text", "body": "foo", "m.new_content": new_body},
|
|
)
|
|
edit_event_id = channel.json_body["event_id"]
|
|
|
|
def assert_bundle(event_json: JsonDict) -> None:
|
|
"""Assert the expected values of the bundled aggregations."""
|
|
relations_dict = event_json["unsigned"].get("m.relations")
|
|
self.assertIn(RelationTypes.REPLACE, relations_dict)
|
|
|
|
m_replace_dict = relations_dict[RelationTypes.REPLACE]
|
|
for key in ["event_id", "sender", "origin_server_ts"]:
|
|
self.assertIn(key, m_replace_dict)
|
|
|
|
self.assert_dict(
|
|
{"event_id": edit_event_id, "sender": self.user_id}, m_replace_dict
|
|
)
|
|
|
|
# /event should return the *original* event
|
|
channel = self.make_request(
|
|
"GET",
|
|
f"/rooms/{self.room}/event/{self.parent_id}",
|
|
access_token=self.user_token,
|
|
)
|
|
self.assertEqual(200, channel.code, channel.json_body)
|
|
self.assertEqual(
|
|
channel.json_body["content"], {"body": "Hi!", "msgtype": "m.text"}
|
|
)
|
|
assert_bundle(channel.json_body)
|
|
|
|
# Request the room messages.
|
|
channel = self.make_request(
|
|
"GET",
|
|
f"/rooms/{self.room}/messages?dir=b",
|
|
access_token=self.user_token,
|
|
)
|
|
self.assertEqual(200, channel.code, channel.json_body)
|
|
assert_bundle(self._find_event_in_chunk(channel.json_body["chunk"]))
|
|
|
|
# Request the room context.
|
|
# /context should return the edited event.
|
|
channel = self.make_request(
|
|
"GET",
|
|
f"/rooms/{self.room}/context/{self.parent_id}",
|
|
access_token=self.user_token,
|
|
)
|
|
self.assertEqual(200, channel.code, channel.json_body)
|
|
assert_bundle(channel.json_body["event"])
|
|
self.assertEqual(channel.json_body["event"]["content"], new_body)
|
|
|
|
# Request sync, but limit the timeline so it becomes limited (and includes
|
|
# bundled aggregations).
|
|
filter = urllib.parse.quote_plus(b'{"room": {"timeline": {"limit": 2}}}')
|
|
channel = self.make_request(
|
|
"GET", f"/sync?filter={filter}", access_token=self.user_token
|
|
)
|
|
self.assertEqual(200, channel.code, channel.json_body)
|
|
room_timeline = channel.json_body["rooms"]["join"][self.room]["timeline"]
|
|
self.assertTrue(room_timeline["limited"])
|
|
assert_bundle(self._find_event_in_chunk(room_timeline["events"]))
|
|
|
|
# Request search.
|
|
channel = self.make_request(
|
|
"POST",
|
|
"/search",
|
|
# Search term matches the parent message.
|
|
content={"search_categories": {"room_events": {"search_term": "Hi"}}},
|
|
access_token=self.user_token,
|
|
)
|
|
self.assertEqual(200, channel.code, channel.json_body)
|
|
chunk = [
|
|
result["result"]
|
|
for result in channel.json_body["search_categories"]["room_events"][
|
|
"results"
|
|
]
|
|
]
|
|
assert_bundle(self._find_event_in_chunk(chunk))
|
|
|
|
def test_multi_edit(self) -> None:
|
|
"""Test that multiple edits, including attempts by people who
|
|
shouldn't be allowed, are correctly handled.
|
|
"""
|
|
|
|
self._send_relation(
|
|
RelationTypes.REPLACE,
|
|
"m.room.message",
|
|
content={
|
|
"msgtype": "m.text",
|
|
"body": "Wibble",
|
|
"m.new_content": {"msgtype": "m.text", "body": "First edit"},
|
|
},
|
|
)
|
|
|
|
new_body = {"msgtype": "m.text", "body": "I've been edited!"}
|
|
channel = self._send_relation(
|
|
RelationTypes.REPLACE,
|
|
"m.room.message",
|
|
content={"msgtype": "m.text", "body": "foo", "m.new_content": new_body},
|
|
)
|
|
edit_event_id = channel.json_body["event_id"]
|
|
|
|
self._send_relation(
|
|
RelationTypes.REPLACE,
|
|
"m.room.message.WRONG_TYPE",
|
|
content={
|
|
"msgtype": "m.text",
|
|
"body": "Wibble",
|
|
"m.new_content": {"msgtype": "m.text", "body": "Edit, but wrong type"},
|
|
},
|
|
)
|
|
|
|
channel = self.make_request(
|
|
"GET",
|
|
f"/rooms/{self.room}/context/{self.parent_id}",
|
|
access_token=self.user_token,
|
|
)
|
|
self.assertEqual(200, channel.code, channel.json_body)
|
|
|
|
self.assertEqual(channel.json_body["event"]["content"], new_body)
|
|
|
|
relations_dict = channel.json_body["event"]["unsigned"].get("m.relations")
|
|
self.assertIn(RelationTypes.REPLACE, relations_dict)
|
|
|
|
m_replace_dict = relations_dict[RelationTypes.REPLACE]
|
|
for key in ["event_id", "sender", "origin_server_ts"]:
|
|
self.assertIn(key, m_replace_dict)
|
|
|
|
self.assert_dict(
|
|
{"event_id": edit_event_id, "sender": self.user_id}, m_replace_dict
|
|
)
|
|
|
|
def test_edit_reply(self) -> None:
|
|
"""Test that editing a reply works."""
|
|
|
|
# Create a reply to edit.
|
|
original_body = {"msgtype": "m.text", "body": "A reply!"}
|
|
channel = self._send_relation(
|
|
RelationTypes.REFERENCE, "m.room.message", content=original_body
|
|
)
|
|
reply = channel.json_body["event_id"]
|
|
|
|
new_body = {"msgtype": "m.text", "body": "I've been edited!"}
|
|
channel = self._send_relation(
|
|
RelationTypes.REPLACE,
|
|
"m.room.message",
|
|
content={"msgtype": "m.text", "body": "foo", "m.new_content": new_body},
|
|
parent_id=reply,
|
|
)
|
|
edit_event_id = channel.json_body["event_id"]
|
|
|
|
# /event returns the original event
|
|
channel = self.make_request(
|
|
"GET",
|
|
f"/rooms/{self.room}/event/{reply}",
|
|
access_token=self.user_token,
|
|
)
|
|
self.assertEqual(200, channel.code, channel.json_body)
|
|
event_result = channel.json_body
|
|
self.assertDictContainsSubset(original_body, event_result["content"])
|
|
|
|
# also check /context, which returns the *edited* event
|
|
channel = self.make_request(
|
|
"GET",
|
|
f"/rooms/{self.room}/context/{reply}",
|
|
access_token=self.user_token,
|
|
)
|
|
self.assertEqual(200, channel.code, channel.json_body)
|
|
context_result = channel.json_body["event"]
|
|
|
|
# check that the relations are correct for both APIs
|
|
for result_event_dict, desc in (
|
|
(event_result, "/event"),
|
|
(context_result, "/context"),
|
|
):
|
|
# The reference metadata should still be intact.
|
|
self.assertDictContainsSubset(
|
|
{
|
|
"m.relates_to": {
|
|
"event_id": self.parent_id,
|
|
"rel_type": "m.reference",
|
|
}
|
|
},
|
|
result_event_dict["content"],
|
|
desc,
|
|
)
|
|
|
|
# We expect that the edit relation appears in the unsigned relations
|
|
# section.
|
|
relations_dict = result_event_dict["unsigned"].get("m.relations")
|
|
self.assertIn(RelationTypes.REPLACE, relations_dict, desc)
|
|
|
|
m_replace_dict = relations_dict[RelationTypes.REPLACE]
|
|
for key in ["event_id", "sender", "origin_server_ts"]:
|
|
self.assertIn(key, m_replace_dict, desc)
|
|
|
|
self.assert_dict(
|
|
{"event_id": edit_event_id, "sender": self.user_id}, m_replace_dict
|
|
)
|
|
|
|
def test_edit_edit(self) -> None:
|
|
"""Test that an edit cannot be edited."""
|
|
new_body = {"msgtype": "m.text", "body": "Initial edit"}
|
|
channel = self._send_relation(
|
|
RelationTypes.REPLACE,
|
|
"m.room.message",
|
|
content={
|
|
"msgtype": "m.text",
|
|
"body": "Wibble",
|
|
"m.new_content": new_body,
|
|
},
|
|
)
|
|
edit_event_id = channel.json_body["event_id"]
|
|
|
|
# Edit the edit event.
|
|
self._send_relation(
|
|
RelationTypes.REPLACE,
|
|
"m.room.message",
|
|
content={
|
|
"msgtype": "m.text",
|
|
"body": "foo",
|
|
"m.new_content": {"msgtype": "m.text", "body": "Ignored edit"},
|
|
},
|
|
parent_id=edit_event_id,
|
|
)
|
|
|
|
# Request the original event.
|
|
# /event should return the original event.
|
|
channel = self.make_request(
|
|
"GET",
|
|
f"/rooms/{self.room}/event/{self.parent_id}",
|
|
access_token=self.user_token,
|
|
)
|
|
self.assertEqual(200, channel.code, channel.json_body)
|
|
self.assertEqual(
|
|
channel.json_body["content"], {"body": "Hi!", "msgtype": "m.text"}
|
|
)
|
|
|
|
# The relations information should not include the edit to the edit.
|
|
relations_dict = channel.json_body["unsigned"].get("m.relations")
|
|
self.assertIn(RelationTypes.REPLACE, relations_dict)
|
|
|
|
# /context should return the event updated for the *first* edit
|
|
# (The edit to the edit should be ignored.)
|
|
channel = self.make_request(
|
|
"GET",
|
|
f"/rooms/{self.room}/context/{self.parent_id}",
|
|
access_token=self.user_token,
|
|
)
|
|
self.assertEqual(200, channel.code, channel.json_body)
|
|
self.assertEqual(channel.json_body["event"]["content"], new_body)
|
|
|
|
m_replace_dict = relations_dict[RelationTypes.REPLACE]
|
|
for key in ["event_id", "sender", "origin_server_ts"]:
|
|
self.assertIn(key, m_replace_dict)
|
|
|
|
self.assert_dict(
|
|
{"event_id": edit_event_id, "sender": self.user_id}, m_replace_dict
|
|
)
|
|
|
|
def test_unknown_relations(self) -> None:
|
|
"""Unknown relations should be accepted."""
|
|
channel = self._send_relation("m.relation.test", "m.room.test")
|
|
event_id = channel.json_body["event_id"]
|
|
|
|
channel = self.make_request(
|
|
"GET",
|
|
f"/_matrix/client/v1/rooms/{self.room}/relations/{self.parent_id}?limit=1",
|
|
access_token=self.user_token,
|
|
)
|
|
self.assertEqual(200, channel.code, channel.json_body)
|
|
|
|
# We expect to get back a single pagination result, which is the full
|
|
# relation event we sent above.
|
|
self.assertEqual(len(channel.json_body["chunk"]), 1, channel.json_body)
|
|
self.assert_dict(
|
|
{"event_id": event_id, "sender": self.user_id, "type": "m.room.test"},
|
|
channel.json_body["chunk"][0],
|
|
)
|
|
|
|
# We also expect to get the original event (the id of which is self.parent_id)
|
|
self.assertEqual(
|
|
channel.json_body["original_event"]["event_id"], self.parent_id
|
|
)
|
|
|
|
# When bundling the unknown relation is not included.
|
|
channel = self.make_request(
|
|
"GET",
|
|
f"/rooms/{self.room}/event/{self.parent_id}",
|
|
access_token=self.user_token,
|
|
)
|
|
self.assertEqual(200, channel.code, channel.json_body)
|
|
self.assertNotIn("m.relations", channel.json_body["unsigned"])
|
|
|
|
def test_background_update(self) -> None:
|
|
"""Test the event_arbitrary_relations background update."""
|
|
channel = self._send_relation(RelationTypes.ANNOTATION, "m.reaction", key="👍")
|
|
annotation_event_id_good = channel.json_body["event_id"]
|
|
|
|
channel = self._send_relation(RelationTypes.ANNOTATION, "m.reaction", key="A")
|
|
annotation_event_id_bad = channel.json_body["event_id"]
|
|
|
|
channel = self._send_relation(RelationTypes.THREAD, "m.room.test")
|
|
thread_event_id = channel.json_body["event_id"]
|
|
|
|
# Clean-up the table as if the inserts did not happen during event creation.
|
|
self.get_success(
|
|
self.store.db_pool.simple_delete_many(
|
|
table="event_relations",
|
|
column="event_id",
|
|
iterable=(annotation_event_id_bad, thread_event_id),
|
|
keyvalues={},
|
|
desc="RelationsTestCase.test_background_update",
|
|
)
|
|
)
|
|
|
|
# Only the "good" annotation should be found.
|
|
channel = self.make_request(
|
|
"GET",
|
|
f"/_matrix/client/v1/rooms/{self.room}/relations/{self.parent_id}?limit=10",
|
|
access_token=self.user_token,
|
|
)
|
|
self.assertEqual(200, channel.code, channel.json_body)
|
|
self.assertEqual(
|
|
[ev["event_id"] for ev in channel.json_body["chunk"]],
|
|
[annotation_event_id_good],
|
|
)
|
|
|
|
# Insert and run the background update.
|
|
self.get_success(
|
|
self.store.db_pool.simple_insert(
|
|
"background_updates",
|
|
{"update_name": "event_arbitrary_relations", "progress_json": "{}"},
|
|
)
|
|
)
|
|
|
|
# Ugh, have to reset this flag
|
|
self.store.db_pool.updates._all_done = False
|
|
self.wait_for_background_updates()
|
|
|
|
# The "good" annotation and the thread should be found, but not the "bad"
|
|
# annotation.
|
|
channel = self.make_request(
|
|
"GET",
|
|
f"/_matrix/client/v1/rooms/{self.room}/relations/{self.parent_id}?limit=10",
|
|
access_token=self.user_token,
|
|
)
|
|
self.assertEqual(200, channel.code, channel.json_body)
|
|
self.assertCountEqual(
|
|
[ev["event_id"] for ev in channel.json_body["chunk"]],
|
|
[annotation_event_id_good, thread_event_id],
|
|
)
|
|
|
|
|
|
class RelationPaginationTestCase(BaseRelationsTestCase):
|
|
def test_basic_paginate_relations(self) -> None:
|
|
"""Tests that calling pagination API correctly the latest relations."""
|
|
channel = self._send_relation(RelationTypes.ANNOTATION, "m.reaction", "a")
|
|
first_annotation_id = channel.json_body["event_id"]
|
|
|
|
channel = self._send_relation(RelationTypes.ANNOTATION, "m.reaction", "b")
|
|
second_annotation_id = channel.json_body["event_id"]
|
|
|
|
channel = self.make_request(
|
|
"GET",
|
|
f"/_matrix/client/v1/rooms/{self.room}/relations/{self.parent_id}?limit=1",
|
|
access_token=self.user_token,
|
|
)
|
|
self.assertEqual(200, channel.code, channel.json_body)
|
|
|
|
# We expect to get back a single pagination result, which is the latest
|
|
# full relation event we sent above.
|
|
self.assertEqual(len(channel.json_body["chunk"]), 1, channel.json_body)
|
|
self.assert_dict(
|
|
{
|
|
"event_id": second_annotation_id,
|
|
"sender": self.user_id,
|
|
"type": "m.reaction",
|
|
},
|
|
channel.json_body["chunk"][0],
|
|
)
|
|
|
|
# We also expect to get the original event (the id of which is self.parent_id)
|
|
self.assertEqual(
|
|
channel.json_body["original_event"]["event_id"], self.parent_id
|
|
)
|
|
|
|
# Make sure next_batch has something in it that looks like it could be a
|
|
# valid token.
|
|
self.assertIsInstance(
|
|
channel.json_body.get("next_batch"), str, channel.json_body
|
|
)
|
|
|
|
# Request the relations again, but with a different direction.
|
|
channel = self.make_request(
|
|
"GET",
|
|
f"/_matrix/client/v1/rooms/{self.room}/relations"
|
|
f"/{self.parent_id}?limit=1&org.matrix.msc3715.dir=f",
|
|
access_token=self.user_token,
|
|
)
|
|
self.assertEqual(200, channel.code, channel.json_body)
|
|
|
|
# We expect to get back a single pagination result, which is the earliest
|
|
# full relation event we sent above.
|
|
self.assertEqual(len(channel.json_body["chunk"]), 1, channel.json_body)
|
|
self.assert_dict(
|
|
{
|
|
"event_id": first_annotation_id,
|
|
"sender": self.user_id,
|
|
"type": "m.reaction",
|
|
},
|
|
channel.json_body["chunk"][0],
|
|
)
|
|
|
|
def test_repeated_paginate_relations(self) -> None:
|
|
"""Test that if we paginate using a limit and tokens then we get the
|
|
expected events.
|
|
"""
|
|
|
|
expected_event_ids = []
|
|
for idx in range(10):
|
|
channel = self._send_relation(
|
|
RelationTypes.ANNOTATION, "m.reaction", chr(ord("a") + idx)
|
|
)
|
|
expected_event_ids.append(channel.json_body["event_id"])
|
|
|
|
prev_token = ""
|
|
found_event_ids: List[str] = []
|
|
for _ in range(20):
|
|
from_token = ""
|
|
if prev_token:
|
|
from_token = "&from=" + prev_token
|
|
|
|
channel = self.make_request(
|
|
"GET",
|
|
f"/_matrix/client/v1/rooms/{self.room}/relations/{self.parent_id}?limit=1{from_token}",
|
|
access_token=self.user_token,
|
|
)
|
|
self.assertEqual(200, channel.code, channel.json_body)
|
|
|
|
found_event_ids.extend(e["event_id"] for e in channel.json_body["chunk"])
|
|
next_batch = channel.json_body.get("next_batch")
|
|
|
|
self.assertNotEqual(prev_token, next_batch)
|
|
prev_token = next_batch
|
|
|
|
if not prev_token:
|
|
break
|
|
|
|
# We paginated backwards, so reverse
|
|
found_event_ids.reverse()
|
|
self.assertEqual(found_event_ids, expected_event_ids)
|
|
|
|
def test_pagination_from_sync_and_messages(self) -> None:
|
|
"""Pagination tokens from /sync and /messages can be used to paginate /relations."""
|
|
channel = self._send_relation(RelationTypes.ANNOTATION, "m.reaction", "A")
|
|
annotation_id = channel.json_body["event_id"]
|
|
# Send an event after the relation events.
|
|
self.helper.send(self.room, body="Latest event", tok=self.user_token)
|
|
|
|
# Request /sync, limiting it such that only the latest event is returned
|
|
# (and not the relation).
|
|
filter = urllib.parse.quote_plus(b'{"room": {"timeline": {"limit": 1}}}')
|
|
channel = self.make_request(
|
|
"GET", f"/sync?filter={filter}", access_token=self.user_token
|
|
)
|
|
self.assertEqual(200, channel.code, channel.json_body)
|
|
room_timeline = channel.json_body["rooms"]["join"][self.room]["timeline"]
|
|
sync_prev_batch = room_timeline["prev_batch"]
|
|
self.assertIsNotNone(sync_prev_batch)
|
|
# Ensure the relation event is not in the batch returned from /sync.
|
|
self.assertNotIn(
|
|
annotation_id, [ev["event_id"] for ev in room_timeline["events"]]
|
|
)
|
|
|
|
# Request /messages, limiting it such that only the latest event is
|
|
# returned (and not the relation).
|
|
channel = self.make_request(
|
|
"GET",
|
|
f"/rooms/{self.room}/messages?dir=b&limit=1",
|
|
access_token=self.user_token,
|
|
)
|
|
self.assertEqual(200, channel.code, channel.json_body)
|
|
messages_end = channel.json_body["end"]
|
|
self.assertIsNotNone(messages_end)
|
|
# Ensure the relation event is not in the chunk returned from /messages.
|
|
self.assertNotIn(
|
|
annotation_id, [ev["event_id"] for ev in channel.json_body["chunk"]]
|
|
)
|
|
|
|
# Request /relations with the pagination tokens received from both the
|
|
# /sync and /messages responses above, in turn.
|
|
#
|
|
# This is a tiny bit silly since the client wouldn't know the parent ID
|
|
# from the requests above; consider the parent ID to be known from a
|
|
# previous /sync.
|
|
for from_token in (sync_prev_batch, messages_end):
|
|
channel = self.make_request(
|
|
"GET",
|
|
f"/_matrix/client/v1/rooms/{self.room}/relations/{self.parent_id}?from={from_token}",
|
|
access_token=self.user_token,
|
|
)
|
|
self.assertEqual(200, channel.code, channel.json_body)
|
|
|
|
# The relation should be in the returned chunk.
|
|
self.assertIn(
|
|
annotation_id, [ev["event_id"] for ev in channel.json_body["chunk"]]
|
|
)
|
|
|
|
|
|
class BundledAggregationsTestCase(BaseRelationsTestCase):
|
|
"""
|
|
See RelationsTestCase.test_edit for a similar test for edits.
|
|
|
|
Note that this doesn't test against /relations since only thread relations
|
|
get bundled via that API. See test_aggregation_get_event_for_thread.
|
|
"""
|
|
|
|
def _test_bundled_aggregations(
|
|
self,
|
|
relation_type: str,
|
|
assertion_callable: Callable[[JsonDict], None],
|
|
expected_db_txn_for_event: int,
|
|
) -> None:
|
|
"""
|
|
Makes requests to various endpoints which should include bundled aggregations
|
|
and then calls an assertion function on the bundled aggregations.
|
|
|
|
Args:
|
|
relation_type: The field to search for in the `m.relations` field in unsigned.
|
|
assertion_callable: Called with the contents of unsigned["m.relations"][relation_type]
|
|
for relation-specific assertions.
|
|
expected_db_txn_for_event: The number of database transactions which
|
|
are expected for a call to /event/.
|
|
"""
|
|
|
|
def assert_bundle(event_json: JsonDict) -> None:
|
|
"""Assert the expected values of the bundled aggregations."""
|
|
relations_dict = event_json["unsigned"].get("m.relations")
|
|
|
|
# Ensure the fields are as expected.
|
|
self.assertCountEqual(relations_dict.keys(), (relation_type,))
|
|
assertion_callable(relations_dict[relation_type])
|
|
|
|
# Request the event directly.
|
|
channel = self.make_request(
|
|
"GET",
|
|
f"/rooms/{self.room}/event/{self.parent_id}",
|
|
access_token=self.user_token,
|
|
)
|
|
self.assertEqual(200, channel.code, channel.json_body)
|
|
assert_bundle(channel.json_body)
|
|
assert channel.resource_usage is not None
|
|
self.assertEqual(channel.resource_usage.db_txn_count, expected_db_txn_for_event)
|
|
|
|
# Request the room messages.
|
|
channel = self.make_request(
|
|
"GET",
|
|
f"/rooms/{self.room}/messages?dir=b",
|
|
access_token=self.user_token,
|
|
)
|
|
self.assertEqual(200, channel.code, channel.json_body)
|
|
assert_bundle(self._find_event_in_chunk(channel.json_body["chunk"]))
|
|
|
|
# Request the room context.
|
|
channel = self.make_request(
|
|
"GET",
|
|
f"/rooms/{self.room}/context/{self.parent_id}",
|
|
access_token=self.user_token,
|
|
)
|
|
self.assertEqual(200, channel.code, channel.json_body)
|
|
assert_bundle(channel.json_body["event"])
|
|
|
|
# Request sync.
|
|
filter = urllib.parse.quote_plus(b'{"room": {"timeline": {"limit": 4}}}')
|
|
channel = self.make_request(
|
|
"GET", f"/sync?filter={filter}", access_token=self.user_token
|
|
)
|
|
self.assertEqual(200, channel.code, channel.json_body)
|
|
room_timeline = channel.json_body["rooms"]["join"][self.room]["timeline"]
|
|
self.assertTrue(room_timeline["limited"])
|
|
assert_bundle(self._find_event_in_chunk(room_timeline["events"]))
|
|
|
|
# Request search.
|
|
channel = self.make_request(
|
|
"POST",
|
|
"/search",
|
|
# Search term matches the parent message.
|
|
content={"search_categories": {"room_events": {"search_term": "Hi"}}},
|
|
access_token=self.user_token,
|
|
)
|
|
self.assertEqual(200, channel.code, channel.json_body)
|
|
chunk = [
|
|
result["result"]
|
|
for result in channel.json_body["search_categories"]["room_events"][
|
|
"results"
|
|
]
|
|
]
|
|
assert_bundle(self._find_event_in_chunk(chunk))
|
|
|
|
def test_annotation(self) -> None:
|
|
"""
|
|
Test that annotations get correctly bundled.
|
|
"""
|
|
# Setup by sending a variety of relations.
|
|
self._send_relation(RelationTypes.ANNOTATION, "m.reaction", "a")
|
|
self._send_relation(
|
|
RelationTypes.ANNOTATION, "m.reaction", "a", access_token=self.user2_token
|
|
)
|
|
self._send_relation(RelationTypes.ANNOTATION, "m.reaction", "b")
|
|
|
|
def assert_annotations(bundled_aggregations: JsonDict) -> None:
|
|
self.assertEqual(
|
|
{
|
|
"chunk": [
|
|
{"type": "m.reaction", "key": "a", "count": 2},
|
|
{"type": "m.reaction", "key": "b", "count": 1},
|
|
]
|
|
},
|
|
bundled_aggregations,
|
|
)
|
|
|
|
self._test_bundled_aggregations(RelationTypes.ANNOTATION, assert_annotations, 7)
|
|
|
|
def test_reference(self) -> None:
|
|
"""
|
|
Test that references get correctly bundled.
|
|
"""
|
|
channel = self._send_relation(RelationTypes.REFERENCE, "m.room.test")
|
|
reply_1 = channel.json_body["event_id"]
|
|
|
|
channel = self._send_relation(RelationTypes.REFERENCE, "m.room.test")
|
|
reply_2 = channel.json_body["event_id"]
|
|
|
|
def assert_annotations(bundled_aggregations: JsonDict) -> None:
|
|
self.assertEqual(
|
|
{"chunk": [{"event_id": reply_1}, {"event_id": reply_2}]},
|
|
bundled_aggregations,
|
|
)
|
|
|
|
self._test_bundled_aggregations(RelationTypes.REFERENCE, assert_annotations, 7)
|
|
|
|
def test_thread(self) -> None:
|
|
"""
|
|
Test that threads get correctly bundled.
|
|
"""
|
|
self._send_relation(RelationTypes.THREAD, "m.room.test")
|
|
channel = self._send_relation(RelationTypes.THREAD, "m.room.test")
|
|
thread_2 = channel.json_body["event_id"]
|
|
|
|
def assert_thread(bundled_aggregations: JsonDict) -> None:
|
|
self.assertEqual(2, bundled_aggregations.get("count"))
|
|
self.assertTrue(bundled_aggregations.get("current_user_participated"))
|
|
# The latest thread event has some fields that don't matter.
|
|
self.assert_dict(
|
|
{
|
|
"content": {
|
|
"m.relates_to": {
|
|
"event_id": self.parent_id,
|
|
"rel_type": RelationTypes.THREAD,
|
|
}
|
|
},
|
|
"event_id": thread_2,
|
|
"sender": self.user_id,
|
|
"type": "m.room.test",
|
|
},
|
|
bundled_aggregations.get("latest_event"),
|
|
)
|
|
|
|
self._test_bundled_aggregations(RelationTypes.THREAD, assert_thread, 10)
|
|
|
|
def test_thread_with_bundled_aggregations_for_latest(self) -> None:
|
|
"""
|
|
Bundled aggregations should get applied to the latest thread event.
|
|
"""
|
|
self._send_relation(RelationTypes.THREAD, "m.room.test")
|
|
channel = self._send_relation(RelationTypes.THREAD, "m.room.test")
|
|
thread_2 = channel.json_body["event_id"]
|
|
|
|
self._send_relation(
|
|
RelationTypes.ANNOTATION, "m.reaction", "a", parent_id=thread_2
|
|
)
|
|
|
|
def assert_thread(bundled_aggregations: JsonDict) -> None:
|
|
self.assertEqual(2, bundled_aggregations.get("count"))
|
|
self.assertTrue(bundled_aggregations.get("current_user_participated"))
|
|
# The latest thread event has some fields that don't matter.
|
|
self.assert_dict(
|
|
{
|
|
"content": {
|
|
"m.relates_to": {
|
|
"event_id": self.parent_id,
|
|
"rel_type": RelationTypes.THREAD,
|
|
}
|
|
},
|
|
"event_id": thread_2,
|
|
"sender": self.user_id,
|
|
"type": "m.room.test",
|
|
},
|
|
bundled_aggregations.get("latest_event"),
|
|
)
|
|
# Check the unsigned field on the latest event.
|
|
self.assert_dict(
|
|
{
|
|
"m.relations": {
|
|
RelationTypes.ANNOTATION: {
|
|
"chunk": [
|
|
{"type": "m.reaction", "key": "a", "count": 1},
|
|
]
|
|
},
|
|
}
|
|
},
|
|
bundled_aggregations["latest_event"].get("unsigned"),
|
|
)
|
|
|
|
self._test_bundled_aggregations(RelationTypes.THREAD, assert_thread, 10)
|
|
|
|
def test_nested_thread(self) -> None:
|
|
"""
|
|
Ensure that a nested thread gets ignored by bundled aggregations, as
|
|
those are forbidden.
|
|
"""
|
|
|
|
# Start a thread.
|
|
channel = self._send_relation(RelationTypes.THREAD, "m.room.test")
|
|
reply_event_id = channel.json_body["event_id"]
|
|
|
|
# Disable the validation to pretend this came over federation, since it is
|
|
# not an event the Client-Server API will allow..
|
|
with patch(
|
|
"synapse.handlers.message.EventCreationHandler._validate_event_relation",
|
|
new=lambda self, event: make_awaitable(None),
|
|
):
|
|
# Create a sub-thread off the thread, which is not allowed.
|
|
self._send_relation(
|
|
RelationTypes.THREAD, "m.room.test", parent_id=reply_event_id
|
|
)
|
|
|
|
# Fetch the thread root, to get the bundled aggregation for the thread.
|
|
relations_from_event = self._get_bundled_aggregations()
|
|
|
|
# Ensure that requesting the room messages also does not return the sub-thread.
|
|
channel = self.make_request(
|
|
"GET",
|
|
f"/rooms/{self.room}/messages?dir=b",
|
|
access_token=self.user_token,
|
|
)
|
|
self.assertEqual(200, channel.code, channel.json_body)
|
|
event = self._find_event_in_chunk(channel.json_body["chunk"])
|
|
relations_from_messages = event["unsigned"]["m.relations"]
|
|
|
|
# Check the bundled aggregations from each point.
|
|
for aggregations, desc in (
|
|
(relations_from_event, "/event"),
|
|
(relations_from_messages, "/messages"),
|
|
):
|
|
# The latest event should have bundled aggregations.
|
|
self.assertIn(RelationTypes.THREAD, aggregations, desc)
|
|
thread_summary = aggregations[RelationTypes.THREAD]
|
|
self.assertIn("latest_event", thread_summary, desc)
|
|
self.assertEqual(
|
|
thread_summary["latest_event"]["event_id"], reply_event_id, desc
|
|
)
|
|
|
|
# The latest event should not have any bundled aggregations (since the
|
|
# only relation to it is another thread, which is invalid).
|
|
self.assertNotIn(
|
|
"m.relations", thread_summary["latest_event"]["unsigned"], desc
|
|
)
|
|
|
|
def test_thread_edit_latest_event(self) -> None:
|
|
"""Test that editing the latest event in a thread works."""
|
|
|
|
# Create a thread and edit the last event.
|
|
channel = self._send_relation(
|
|
RelationTypes.THREAD,
|
|
"m.room.message",
|
|
content={"msgtype": "m.text", "body": "A threaded reply!"},
|
|
)
|
|
threaded_event_id = channel.json_body["event_id"]
|
|
|
|
new_body = {"msgtype": "m.text", "body": "I've been edited!"}
|
|
channel = self._send_relation(
|
|
RelationTypes.REPLACE,
|
|
"m.room.message",
|
|
content={"msgtype": "m.text", "body": "foo", "m.new_content": new_body},
|
|
parent_id=threaded_event_id,
|
|
)
|
|
edit_event_id = channel.json_body["event_id"]
|
|
|
|
# Fetch the thread root, to get the bundled aggregation for the thread.
|
|
relations_dict = self._get_bundled_aggregations()
|
|
|
|
# We expect that the edit message appears in the thread summary in the
|
|
# unsigned relations section.
|
|
self.assertIn(RelationTypes.THREAD, relations_dict)
|
|
|
|
thread_summary = relations_dict[RelationTypes.THREAD]
|
|
self.assertIn("latest_event", thread_summary)
|
|
latest_event_in_thread = thread_summary["latest_event"]
|
|
self.assertEqual(latest_event_in_thread["content"]["body"], "I've been edited!")
|
|
# The latest event in the thread should have the edit appear under the
|
|
# bundled aggregations.
|
|
self.assertDictContainsSubset(
|
|
{"event_id": edit_event_id, "sender": "@alice:test"},
|
|
latest_event_in_thread["unsigned"]["m.relations"][RelationTypes.REPLACE],
|
|
)
|
|
|
|
def test_aggregation_get_event_for_annotation(self) -> None:
|
|
"""Test that annotations do not get bundled aggregations included
|
|
when directly requested.
|
|
"""
|
|
channel = self._send_relation(RelationTypes.ANNOTATION, "m.reaction", "a")
|
|
annotation_id = channel.json_body["event_id"]
|
|
|
|
# Annotate the annotation.
|
|
self._send_relation(
|
|
RelationTypes.ANNOTATION, "m.reaction", "a", parent_id=annotation_id
|
|
)
|
|
|
|
channel = self.make_request(
|
|
"GET",
|
|
f"/rooms/{self.room}/event/{annotation_id}",
|
|
access_token=self.user_token,
|
|
)
|
|
self.assertEqual(200, channel.code, channel.json_body)
|
|
self.assertIsNone(channel.json_body["unsigned"].get("m.relations"))
|
|
|
|
def test_aggregation_get_event_for_thread(self) -> None:
|
|
"""Test that threads get bundled aggregations included when directly requested."""
|
|
channel = self._send_relation(RelationTypes.THREAD, "m.room.test")
|
|
thread_id = channel.json_body["event_id"]
|
|
|
|
# Annotate the thread.
|
|
self._send_relation(
|
|
RelationTypes.ANNOTATION, "m.reaction", "a", parent_id=thread_id
|
|
)
|
|
|
|
channel = self.make_request(
|
|
"GET",
|
|
f"/rooms/{self.room}/event/{thread_id}",
|
|
access_token=self.user_token,
|
|
)
|
|
self.assertEqual(200, channel.code, channel.json_body)
|
|
self.assertEqual(
|
|
channel.json_body["unsigned"].get("m.relations"),
|
|
{
|
|
RelationTypes.ANNOTATION: {
|
|
"chunk": [{"count": 1, "key": "a", "type": "m.reaction"}]
|
|
},
|
|
},
|
|
)
|
|
|
|
# It should also be included when the entire thread is requested.
|
|
channel = self.make_request(
|
|
"GET",
|
|
f"/_matrix/client/v1/rooms/{self.room}/relations/{self.parent_id}?limit=1",
|
|
access_token=self.user_token,
|
|
)
|
|
self.assertEqual(200, channel.code, channel.json_body)
|
|
self.assertEqual(len(channel.json_body["chunk"]), 1)
|
|
|
|
thread_message = channel.json_body["chunk"][0]
|
|
self.assertEqual(
|
|
thread_message["unsigned"].get("m.relations"),
|
|
{
|
|
RelationTypes.ANNOTATION: {
|
|
"chunk": [{"count": 1, "key": "a", "type": "m.reaction"}]
|
|
},
|
|
},
|
|
)
|
|
|
|
def test_bundled_aggregations_with_filter(self) -> None:
|
|
"""
|
|
If "unsigned" is an omitted field (due to filtering), adding the bundled
|
|
aggregations should not break.
|
|
|
|
Note that the spec allows for a server to return additional fields beyond
|
|
what is specified.
|
|
"""
|
|
self._send_relation(RelationTypes.ANNOTATION, "m.reaction", "a")
|
|
|
|
# Note that the sync filter does not include "unsigned" as a field.
|
|
filter = urllib.parse.quote_plus(
|
|
b'{"event_fields": ["content", "event_id"], "room": {"timeline": {"limit": 3}}}'
|
|
)
|
|
channel = self.make_request(
|
|
"GET", f"/sync?filter={filter}", access_token=self.user_token
|
|
)
|
|
self.assertEqual(200, channel.code, channel.json_body)
|
|
|
|
# Ensure the timeline is limited, find the parent event.
|
|
room_timeline = channel.json_body["rooms"]["join"][self.room]["timeline"]
|
|
self.assertTrue(room_timeline["limited"])
|
|
parent_event = self._find_event_in_chunk(room_timeline["events"])
|
|
|
|
# Ensure there's bundled aggregations on it.
|
|
self.assertIn("unsigned", parent_event)
|
|
self.assertIn("m.relations", parent_event["unsigned"])
|
|
|
|
|
|
class RelationIgnoredUserTestCase(BaseRelationsTestCase):
|
|
"""Relations sent from an ignored user should be ignored."""
|
|
|
|
def _test_ignored_user(
|
|
self,
|
|
relation_type: str,
|
|
allowed_event_ids: List[str],
|
|
ignored_event_ids: List[str],
|
|
) -> Tuple[JsonDict, JsonDict]:
|
|
"""
|
|
Fetch the relations and ensure they're all there, then ignore user2, and
|
|
repeat.
|
|
|
|
Returns:
|
|
A tuple of two JSON dictionaries, each are bundled aggregations, the
|
|
first is from before the user is ignored, and the second is after.
|
|
"""
|
|
# Get the relations.
|
|
event_ids = self._get_related_events()
|
|
self.assertCountEqual(event_ids, allowed_event_ids + ignored_event_ids)
|
|
|
|
# And the bundled aggregations.
|
|
before_aggregations = self._get_bundled_aggregations()
|
|
self.assertIn(relation_type, before_aggregations)
|
|
|
|
# Ignore user2 and re-do the requests.
|
|
self.get_success(
|
|
self.store.add_account_data_for_user(
|
|
self.user_id,
|
|
AccountDataTypes.IGNORED_USER_LIST,
|
|
{"ignored_users": {self.user2_id: {}}},
|
|
)
|
|
)
|
|
|
|
# Get the relations.
|
|
event_ids = self._get_related_events()
|
|
self.assertCountEqual(event_ids, allowed_event_ids)
|
|
|
|
# And the bundled aggregations.
|
|
after_aggregations = self._get_bundled_aggregations()
|
|
self.assertIn(relation_type, after_aggregations)
|
|
|
|
return before_aggregations[relation_type], after_aggregations[relation_type]
|
|
|
|
def test_annotation(self) -> None:
|
|
"""Annotations should ignore"""
|
|
# Send 2 from us, 2 from the to be ignored user.
|
|
allowed_event_ids = []
|
|
ignored_event_ids = []
|
|
channel = self._send_relation(RelationTypes.ANNOTATION, "m.reaction", key="a")
|
|
allowed_event_ids.append(channel.json_body["event_id"])
|
|
channel = self._send_relation(RelationTypes.ANNOTATION, "m.reaction", key="b")
|
|
allowed_event_ids.append(channel.json_body["event_id"])
|
|
channel = self._send_relation(
|
|
RelationTypes.ANNOTATION,
|
|
"m.reaction",
|
|
key="a",
|
|
access_token=self.user2_token,
|
|
)
|
|
ignored_event_ids.append(channel.json_body["event_id"])
|
|
channel = self._send_relation(
|
|
RelationTypes.ANNOTATION,
|
|
"m.reaction",
|
|
key="c",
|
|
access_token=self.user2_token,
|
|
)
|
|
ignored_event_ids.append(channel.json_body["event_id"])
|
|
|
|
before_aggregations, after_aggregations = self._test_ignored_user(
|
|
RelationTypes.ANNOTATION, allowed_event_ids, ignored_event_ids
|
|
)
|
|
|
|
self.assertCountEqual(
|
|
before_aggregations["chunk"],
|
|
[
|
|
{"type": "m.reaction", "key": "a", "count": 2},
|
|
{"type": "m.reaction", "key": "b", "count": 1},
|
|
{"type": "m.reaction", "key": "c", "count": 1},
|
|
],
|
|
)
|
|
|
|
self.assertCountEqual(
|
|
after_aggregations["chunk"],
|
|
[
|
|
{"type": "m.reaction", "key": "a", "count": 1},
|
|
{"type": "m.reaction", "key": "b", "count": 1},
|
|
],
|
|
)
|
|
|
|
def test_reference(self) -> None:
|
|
"""Annotations should ignore"""
|
|
channel = self._send_relation(RelationTypes.REFERENCE, "m.room.test")
|
|
allowed_event_ids = [channel.json_body["event_id"]]
|
|
|
|
channel = self._send_relation(
|
|
RelationTypes.REFERENCE, "m.room.test", access_token=self.user2_token
|
|
)
|
|
ignored_event_ids = [channel.json_body["event_id"]]
|
|
|
|
before_aggregations, after_aggregations = self._test_ignored_user(
|
|
RelationTypes.REFERENCE, allowed_event_ids, ignored_event_ids
|
|
)
|
|
|
|
self.assertCountEqual(
|
|
[e["event_id"] for e in before_aggregations["chunk"]],
|
|
allowed_event_ids + ignored_event_ids,
|
|
)
|
|
|
|
self.assertCountEqual(
|
|
[e["event_id"] for e in after_aggregations["chunk"]], allowed_event_ids
|
|
)
|
|
|
|
def test_thread(self) -> None:
|
|
"""Annotations should ignore"""
|
|
channel = self._send_relation(RelationTypes.THREAD, "m.room.test")
|
|
allowed_event_ids = [channel.json_body["event_id"]]
|
|
|
|
channel = self._send_relation(
|
|
RelationTypes.THREAD, "m.room.test", access_token=self.user2_token
|
|
)
|
|
ignored_event_ids = [channel.json_body["event_id"]]
|
|
|
|
before_aggregations, after_aggregations = self._test_ignored_user(
|
|
RelationTypes.THREAD, allowed_event_ids, ignored_event_ids
|
|
)
|
|
|
|
self.assertEqual(before_aggregations["count"], 2)
|
|
self.assertTrue(before_aggregations["current_user_participated"])
|
|
# The latest thread event has some fields that don't matter.
|
|
self.assertEqual(
|
|
before_aggregations["latest_event"]["event_id"], ignored_event_ids[0]
|
|
)
|
|
|
|
self.assertEqual(after_aggregations["count"], 1)
|
|
self.assertTrue(after_aggregations["current_user_participated"])
|
|
# The latest thread event has some fields that don't matter.
|
|
self.assertEqual(
|
|
after_aggregations["latest_event"]["event_id"], allowed_event_ids[0]
|
|
)
|
|
|
|
|
|
class RelationRedactionTestCase(BaseRelationsTestCase):
|
|
"""
|
|
Test the behaviour of relations when the parent or child event is redacted.
|
|
|
|
The behaviour of each relation type is subtly different which causes the tests
|
|
to be a bit repetitive, they follow a naming scheme of:
|
|
|
|
test_redact_(relation|parent)_{relation_type}
|
|
|
|
The first bit of "relation" means that the event with the relation defined
|
|
on it (the child event) is to be redacted. A "parent" means that the target
|
|
of the relation (the parent event) is to be redacted.
|
|
|
|
The relation_type describes which type of relation is under test (i.e. it is
|
|
related to the value of rel_type in the event content).
|
|
"""
|
|
|
|
def _redact(self, event_id: str) -> None:
|
|
channel = self.make_request(
|
|
"POST",
|
|
f"/_matrix/client/r0/rooms/{self.room}/redact/{event_id}",
|
|
access_token=self.user_token,
|
|
content={},
|
|
)
|
|
self.assertEqual(200, channel.code, channel.json_body)
|
|
|
|
def test_redact_relation_annotation(self) -> None:
|
|
"""
|
|
Test that annotations of an event are properly handled after the
|
|
annotation is redacted.
|
|
|
|
The redacted relation should not be included in bundled aggregations or
|
|
the response to relations.
|
|
"""
|
|
channel = self._send_relation(RelationTypes.ANNOTATION, "m.reaction", "a")
|
|
to_redact_event_id = channel.json_body["event_id"]
|
|
|
|
channel = self._send_relation(
|
|
RelationTypes.ANNOTATION, "m.reaction", "a", access_token=self.user2_token
|
|
)
|
|
unredacted_event_id = channel.json_body["event_id"]
|
|
|
|
# Both relations should exist.
|
|
event_ids = self._get_related_events()
|
|
relations = self._get_bundled_aggregations()
|
|
self.assertCountEqual(event_ids, [to_redact_event_id, unredacted_event_id])
|
|
self.assertEquals(
|
|
relations["m.annotation"],
|
|
{"chunk": [{"type": "m.reaction", "key": "a", "count": 2}]},
|
|
)
|
|
|
|
# Redact one of the reactions.
|
|
self._redact(to_redact_event_id)
|
|
|
|
# The unredacted relation should still exist.
|
|
event_ids = self._get_related_events()
|
|
relations = self._get_bundled_aggregations()
|
|
self.assertEquals(event_ids, [unredacted_event_id])
|
|
self.assertEquals(
|
|
relations["m.annotation"],
|
|
{"chunk": [{"type": "m.reaction", "key": "a", "count": 1}]},
|
|
)
|
|
|
|
def test_redact_relation_thread(self) -> None:
|
|
"""
|
|
Test that thread replies are properly handled after the thread reply redacted.
|
|
|
|
The redacted event should not be included in bundled aggregations or
|
|
the response to relations.
|
|
"""
|
|
channel = self._send_relation(
|
|
RelationTypes.THREAD,
|
|
EventTypes.Message,
|
|
content={"body": "reply 1", "msgtype": "m.text"},
|
|
)
|
|
unredacted_event_id = channel.json_body["event_id"]
|
|
|
|
# Note that the *last* event in the thread is redacted, as that gets
|
|
# included in the bundled aggregation.
|
|
channel = self._send_relation(
|
|
RelationTypes.THREAD,
|
|
EventTypes.Message,
|
|
content={"body": "reply 2", "msgtype": "m.text"},
|
|
)
|
|
to_redact_event_id = channel.json_body["event_id"]
|
|
|
|
# Both relations exist.
|
|
event_ids = self._get_related_events()
|
|
relations = self._get_bundled_aggregations()
|
|
self.assertEquals(event_ids, [to_redact_event_id, unredacted_event_id])
|
|
self.assertDictContainsSubset(
|
|
{
|
|
"count": 2,
|
|
"current_user_participated": True,
|
|
},
|
|
relations[RelationTypes.THREAD],
|
|
)
|
|
# And the latest event returned is the event that will be redacted.
|
|
self.assertEqual(
|
|
relations[RelationTypes.THREAD]["latest_event"]["event_id"],
|
|
to_redact_event_id,
|
|
)
|
|
|
|
# Redact one of the reactions.
|
|
self._redact(to_redact_event_id)
|
|
|
|
# The unredacted relation should still exist.
|
|
event_ids = self._get_related_events()
|
|
relations = self._get_bundled_aggregations()
|
|
self.assertEquals(event_ids, [unredacted_event_id])
|
|
self.assertDictContainsSubset(
|
|
{
|
|
"count": 1,
|
|
"current_user_participated": True,
|
|
},
|
|
relations[RelationTypes.THREAD],
|
|
)
|
|
# And the latest event is now the unredacted event.
|
|
self.assertEqual(
|
|
relations[RelationTypes.THREAD]["latest_event"]["event_id"],
|
|
unredacted_event_id,
|
|
)
|
|
|
|
def test_redact_parent_edit(self) -> None:
|
|
"""Test that edits of an event are redacted when the original event
|
|
is redacted.
|
|
"""
|
|
# Add a relation
|
|
self._send_relation(
|
|
RelationTypes.REPLACE,
|
|
"m.room.message",
|
|
parent_id=self.parent_id,
|
|
content={
|
|
"msgtype": "m.text",
|
|
"body": "Wibble",
|
|
"m.new_content": {"msgtype": "m.text", "body": "First edit"},
|
|
},
|
|
)
|
|
|
|
# Check the relation is returned
|
|
event_ids = self._get_related_events()
|
|
relations = self._get_bundled_aggregations()
|
|
self.assertEqual(len(event_ids), 1)
|
|
self.assertIn(RelationTypes.REPLACE, relations)
|
|
|
|
# Redact the original event
|
|
self._redact(self.parent_id)
|
|
|
|
# The relations are not returned.
|
|
event_ids = self._get_related_events()
|
|
relations = self._get_bundled_aggregations()
|
|
self.assertEqual(len(event_ids), 0)
|
|
self.assertEqual(relations, {})
|
|
|
|
def test_redact_parent_annotation(self) -> None:
|
|
"""Test that annotations of an event are viewable when the original event
|
|
is redacted.
|
|
"""
|
|
# Add a relation
|
|
channel = self._send_relation(RelationTypes.ANNOTATION, "m.reaction", key="👍")
|
|
related_event_id = channel.json_body["event_id"]
|
|
|
|
# The relations should exist.
|
|
event_ids = self._get_related_events()
|
|
relations = self._get_bundled_aggregations()
|
|
self.assertEqual(len(event_ids), 1)
|
|
self.assertIn(RelationTypes.ANNOTATION, relations)
|
|
|
|
# Redact the original event.
|
|
self._redact(self.parent_id)
|
|
|
|
# The relations are returned.
|
|
event_ids = self._get_related_events()
|
|
relations = self._get_bundled_aggregations()
|
|
self.assertEquals(event_ids, [related_event_id])
|
|
self.assertEquals(
|
|
relations["m.annotation"],
|
|
{"chunk": [{"type": "m.reaction", "key": "👍", "count": 1}]},
|
|
)
|
|
|
|
@unittest.override_config({"experimental_features": {"msc3440_enabled": True}})
|
|
def test_redact_parent_thread(self) -> None:
|
|
"""
|
|
Test that thread replies are still available when the root event is redacted.
|
|
"""
|
|
channel = self._send_relation(
|
|
RelationTypes.THREAD,
|
|
EventTypes.Message,
|
|
content={"body": "reply 1", "msgtype": "m.text"},
|
|
)
|
|
related_event_id = channel.json_body["event_id"]
|
|
|
|
# Redact one of the reactions.
|
|
self._redact(self.parent_id)
|
|
|
|
# The unredacted relation should still exist.
|
|
event_ids = self._get_related_events()
|
|
relations = self._get_bundled_aggregations()
|
|
self.assertEquals(len(event_ids), 1)
|
|
self.assertDictContainsSubset(
|
|
{
|
|
"count": 1,
|
|
"current_user_participated": True,
|
|
},
|
|
relations[RelationTypes.THREAD],
|
|
)
|
|
self.assertEqual(
|
|
relations[RelationTypes.THREAD]["latest_event"]["event_id"],
|
|
related_event_id,
|
|
)
|