2019-05-29 16:19:11 +03:00
|
|
|
#
|
2023-11-21 23:29:58 +03:00
|
|
|
# This file is licensed under the Affero General Public License (AGPL) version 3.
|
|
|
|
#
|
2024-01-23 14:26:48 +03:00
|
|
|
# Copyright 2019 The Matrix.org Foundation C.I.C.
|
2023-11-21 23:29:58 +03:00
|
|
|
# Copyright (C) 2023 New Vector, Ltd
|
|
|
|
#
|
|
|
|
# This program is free software: you can redistribute it and/or modify
|
|
|
|
# it under the terms of the GNU Affero General Public License as
|
|
|
|
# published by the Free Software Foundation, either version 3 of the
|
|
|
|
# License, or (at your option) any later version.
|
|
|
|
#
|
|
|
|
# See the GNU Affero General Public License for more details:
|
|
|
|
# <https://www.gnu.org/licenses/agpl-3.0.html>.
|
|
|
|
#
|
|
|
|
# Originally licensed under the Apache License, Version 2.0:
|
|
|
|
# <http://www.apache.org/licenses/LICENSE-2.0>.
|
|
|
|
#
|
|
|
|
# [This file includes modifications made by New Vector Limited]
|
2019-05-29 16:19:11 +03:00
|
|
|
#
|
|
|
|
#
|
|
|
|
|
|
|
|
import os.path
|
2021-04-09 20:44:38 +03:00
|
|
|
from unittest.mock import Mock, patch
|
2019-09-26 13:47:53 +03:00
|
|
|
|
2022-12-09 20:36:32 +03:00
|
|
|
from twisted.test.proto_helpers import MemoryReactor
|
|
|
|
|
2019-09-26 13:47:53 +03:00
|
|
|
import synapse.rest.admin
|
|
|
|
from synapse.api.constants import EventTypes
|
2021-08-17 14:57:58 +03:00
|
|
|
from synapse.rest.client import login, room
|
2022-12-09 20:36:32 +03:00
|
|
|
from synapse.server import HomeServer
|
2019-05-29 16:19:11 +03:00
|
|
|
from synapse.storage import prepare_database
|
2022-12-09 20:36:32 +03:00
|
|
|
from synapse.storage.types import Cursor
|
2020-10-22 12:11:06 +03:00
|
|
|
from synapse.types import UserID, create_requester
|
2022-12-09 20:36:32 +03:00
|
|
|
from synapse.util import Clock
|
2019-05-29 16:19:11 +03:00
|
|
|
|
|
|
|
from tests.unittest import HomeserverTestCase
|
|
|
|
|
|
|
|
|
|
|
|
class CleanupExtremBackgroundUpdateStoreTestCase(HomeserverTestCase):
|
|
|
|
"""
|
2019-06-13 15:40:52 +03:00
|
|
|
Test the background update to clean forward extremities table.
|
|
|
|
"""
|
2019-05-29 16:19:11 +03:00
|
|
|
|
2022-12-09 20:36:32 +03:00
|
|
|
def prepare(
|
|
|
|
self, reactor: MemoryReactor, clock: Clock, homeserver: HomeServer
|
|
|
|
) -> None:
|
2022-02-23 14:04:02 +03:00
|
|
|
self.store = homeserver.get_datastores().main
|
2019-05-29 16:19:11 +03:00
|
|
|
self.room_creator = homeserver.get_room_creation_handler()
|
|
|
|
|
|
|
|
# Create a test user and room
|
|
|
|
self.user = UserID("alice", "test")
|
2020-10-22 12:11:06 +03:00
|
|
|
self.requester = create_requester(self.user)
|
2023-02-22 14:07:28 +03:00
|
|
|
self.room_id, _, _ = self.get_success(
|
|
|
|
self.room_creator.create_room(self.requester, {})
|
|
|
|
)
|
2019-05-29 16:19:11 +03:00
|
|
|
|
2022-12-09 20:36:32 +03:00
|
|
|
def run_background_update(self) -> None:
|
2019-05-29 16:19:11 +03:00
|
|
|
"""Re run the background update to clean up the extremities."""
|
|
|
|
# Make sure we don't clash with in progress updates.
|
2019-12-04 18:09:36 +03:00
|
|
|
self.assertTrue(
|
2020-08-05 23:38:57 +03:00
|
|
|
self.store.db_pool.updates._all_done, "Background updates are still ongoing"
|
2019-12-04 18:09:36 +03:00
|
|
|
)
|
2019-05-29 16:19:11 +03:00
|
|
|
|
|
|
|
schema_path = os.path.join(
|
2021-05-07 12:22:05 +03:00
|
|
|
prepare_database.schema_path,
|
2019-10-21 14:56:42 +03:00
|
|
|
"main",
|
2019-05-29 16:19:11 +03:00
|
|
|
"delta",
|
|
|
|
"54",
|
|
|
|
"delete_forward_extremities.sql",
|
|
|
|
)
|
|
|
|
|
2022-12-09 20:36:32 +03:00
|
|
|
def run_delta_file(txn: Cursor) -> None:
|
2019-05-29 16:19:11 +03:00
|
|
|
prepare_database.executescript(txn, schema_path)
|
|
|
|
|
|
|
|
self.get_success(
|
2020-08-05 23:38:57 +03:00
|
|
|
self.store.db_pool.runInteraction(
|
2019-12-04 16:52:46 +03:00
|
|
|
"test_delete_forward_extremities", run_delta_file
|
|
|
|
)
|
2019-05-29 16:19:11 +03:00
|
|
|
)
|
|
|
|
|
|
|
|
# Ugh, have to reset this flag
|
2020-08-05 23:38:57 +03:00
|
|
|
self.store.db_pool.updates._all_done = False
|
2019-05-29 16:19:11 +03:00
|
|
|
|
2021-10-06 15:56:45 +03:00
|
|
|
self.wait_for_background_updates()
|
2019-05-29 16:19:11 +03:00
|
|
|
|
2022-04-01 15:48:08 +03:00
|
|
|
def add_extremity(self, room_id: str, event_id: str) -> None:
|
|
|
|
"""
|
|
|
|
Add the given event as an extremity to the room.
|
|
|
|
"""
|
|
|
|
self.get_success(
|
|
|
|
self.hs.get_datastores().main.db_pool.simple_insert(
|
|
|
|
table="event_forward_extremities",
|
|
|
|
values={"room_id": room_id, "event_id": event_id},
|
|
|
|
desc="test_add_extremity",
|
|
|
|
)
|
|
|
|
)
|
|
|
|
|
|
|
|
self.hs.get_datastores().main.get_latest_event_ids_in_room.invalidate(
|
|
|
|
(room_id,)
|
|
|
|
)
|
|
|
|
|
2022-12-09 20:36:32 +03:00
|
|
|
def test_soft_failed_extremities_handled_correctly(self) -> None:
|
2019-05-29 16:19:11 +03:00
|
|
|
"""Test that extremities are correctly calculated in the presence of
|
|
|
|
soft failed events.
|
|
|
|
|
|
|
|
Tests a graph like:
|
|
|
|
|
|
|
|
A <- SF1 <- SF2 <- B
|
|
|
|
|
|
|
|
Where SF* are soft failed.
|
|
|
|
"""
|
|
|
|
|
|
|
|
# Create the room graph
|
2019-06-13 15:40:52 +03:00
|
|
|
event_id_1 = self.create_and_send_event(self.room_id, self.user)
|
|
|
|
event_id_2 = self.create_and_send_event(
|
|
|
|
self.room_id, self.user, True, [event_id_1]
|
|
|
|
)
|
|
|
|
event_id_3 = self.create_and_send_event(
|
|
|
|
self.room_id, self.user, True, [event_id_2]
|
|
|
|
)
|
|
|
|
event_id_4 = self.create_and_send_event(
|
|
|
|
self.room_id, self.user, False, [event_id_3]
|
|
|
|
)
|
2019-05-29 16:19:11 +03:00
|
|
|
|
|
|
|
# Check the latest events are as expected
|
|
|
|
latest_event_ids = self.get_success(
|
|
|
|
self.store.get_latest_event_ids_in_room(self.room_id)
|
|
|
|
)
|
|
|
|
|
2023-09-18 16:29:05 +03:00
|
|
|
self.assertEqual(latest_event_ids, {event_id_4})
|
2019-05-29 16:19:11 +03:00
|
|
|
|
2022-12-09 20:36:32 +03:00
|
|
|
def test_basic_cleanup(self) -> None:
|
2019-05-29 16:19:11 +03:00
|
|
|
"""Test that extremities are correctly calculated in the presence of
|
|
|
|
soft failed events.
|
|
|
|
|
|
|
|
Tests a graph like:
|
|
|
|
|
|
|
|
A <- SF1 <- B
|
|
|
|
|
|
|
|
Where SF* are soft failed, and with extremities of A and B
|
|
|
|
"""
|
|
|
|
# Create the room graph
|
2019-06-13 15:40:52 +03:00
|
|
|
event_id_a = self.create_and_send_event(self.room_id, self.user)
|
|
|
|
event_id_sf1 = self.create_and_send_event(
|
|
|
|
self.room_id, self.user, True, [event_id_a]
|
|
|
|
)
|
|
|
|
event_id_b = self.create_and_send_event(
|
|
|
|
self.room_id, self.user, False, [event_id_sf1]
|
|
|
|
)
|
2019-05-29 16:19:11 +03:00
|
|
|
|
|
|
|
# Add the new extremity and check the latest events are as expected
|
2019-06-13 15:40:52 +03:00
|
|
|
self.add_extremity(self.room_id, event_id_a)
|
2019-05-29 16:19:11 +03:00
|
|
|
|
|
|
|
latest_event_ids = self.get_success(
|
|
|
|
self.store.get_latest_event_ids_in_room(self.room_id)
|
|
|
|
)
|
2023-09-18 16:29:05 +03:00
|
|
|
self.assertEqual(latest_event_ids, {event_id_a, event_id_b})
|
2019-05-29 16:19:11 +03:00
|
|
|
|
|
|
|
# Run the background update and check it did the right thing
|
|
|
|
self.run_background_update()
|
|
|
|
|
|
|
|
latest_event_ids = self.get_success(
|
|
|
|
self.store.get_latest_event_ids_in_room(self.room_id)
|
|
|
|
)
|
2023-09-18 16:29:05 +03:00
|
|
|
self.assertEqual(latest_event_ids, {event_id_b})
|
2019-05-29 16:19:11 +03:00
|
|
|
|
2022-12-09 20:36:32 +03:00
|
|
|
def test_chain_of_fail_cleanup(self) -> None:
|
2019-05-29 16:19:11 +03:00
|
|
|
"""Test that extremities are correctly calculated in the presence of
|
|
|
|
soft failed events.
|
|
|
|
|
|
|
|
Tests a graph like:
|
|
|
|
|
|
|
|
A <- SF1 <- SF2 <- B
|
|
|
|
|
|
|
|
Where SF* are soft failed, and with extremities of A and B
|
|
|
|
"""
|
|
|
|
# Create the room graph
|
2019-06-13 15:40:52 +03:00
|
|
|
event_id_a = self.create_and_send_event(self.room_id, self.user)
|
|
|
|
event_id_sf1 = self.create_and_send_event(
|
|
|
|
self.room_id, self.user, True, [event_id_a]
|
|
|
|
)
|
|
|
|
event_id_sf2 = self.create_and_send_event(
|
|
|
|
self.room_id, self.user, True, [event_id_sf1]
|
|
|
|
)
|
|
|
|
event_id_b = self.create_and_send_event(
|
|
|
|
self.room_id, self.user, False, [event_id_sf2]
|
|
|
|
)
|
2019-05-29 16:19:11 +03:00
|
|
|
|
|
|
|
# Add the new extremity and check the latest events are as expected
|
2019-06-13 15:40:52 +03:00
|
|
|
self.add_extremity(self.room_id, event_id_a)
|
2019-05-29 16:19:11 +03:00
|
|
|
|
|
|
|
latest_event_ids = self.get_success(
|
|
|
|
self.store.get_latest_event_ids_in_room(self.room_id)
|
|
|
|
)
|
2023-09-18 16:29:05 +03:00
|
|
|
self.assertEqual(latest_event_ids, {event_id_a, event_id_b})
|
2019-05-29 16:19:11 +03:00
|
|
|
|
|
|
|
# Run the background update and check it did the right thing
|
|
|
|
self.run_background_update()
|
|
|
|
|
|
|
|
latest_event_ids = self.get_success(
|
|
|
|
self.store.get_latest_event_ids_in_room(self.room_id)
|
|
|
|
)
|
2023-09-18 16:29:05 +03:00
|
|
|
self.assertEqual(latest_event_ids, {event_id_b})
|
2019-05-29 16:19:11 +03:00
|
|
|
|
2022-12-09 20:36:32 +03:00
|
|
|
def test_forked_graph_cleanup(self) -> None:
|
2019-05-29 16:19:11 +03:00
|
|
|
r"""Test that extremities are correctly calculated in the presence of
|
|
|
|
soft failed events.
|
|
|
|
|
|
|
|
Tests a graph like, where time flows down the page:
|
|
|
|
|
|
|
|
A B
|
|
|
|
/ \ /
|
|
|
|
/ \ /
|
|
|
|
SF1 SF2
|
|
|
|
| |
|
|
|
|
SF3 |
|
|
|
|
/ \ |
|
|
|
|
| \ |
|
|
|
|
C SF4
|
|
|
|
|
|
|
|
Where SF* are soft failed, and with them A, B and C marked as
|
|
|
|
extremities. This should resolve to B and C being marked as extremity.
|
|
|
|
"""
|
2019-06-06 22:13:47 +03:00
|
|
|
|
2019-05-29 16:19:11 +03:00
|
|
|
# Create the room graph
|
2019-06-13 15:40:52 +03:00
|
|
|
event_id_a = self.create_and_send_event(self.room_id, self.user)
|
|
|
|
event_id_b = self.create_and_send_event(self.room_id, self.user)
|
|
|
|
event_id_sf1 = self.create_and_send_event(
|
|
|
|
self.room_id, self.user, True, [event_id_a]
|
|
|
|
)
|
|
|
|
event_id_sf2 = self.create_and_send_event(
|
|
|
|
self.room_id, self.user, True, [event_id_a, event_id_b]
|
|
|
|
)
|
|
|
|
event_id_sf3 = self.create_and_send_event(
|
|
|
|
self.room_id, self.user, True, [event_id_sf1]
|
|
|
|
)
|
|
|
|
self.create_and_send_event(
|
|
|
|
self.room_id, self.user, True, [event_id_sf2, event_id_sf3]
|
|
|
|
) # SF4
|
|
|
|
event_id_c = self.create_and_send_event(
|
|
|
|
self.room_id, self.user, False, [event_id_sf3]
|
|
|
|
)
|
2019-05-29 16:19:11 +03:00
|
|
|
|
|
|
|
# Add the new extremity and check the latest events are as expected
|
2019-06-13 15:40:52 +03:00
|
|
|
self.add_extremity(self.room_id, event_id_a)
|
2019-05-29 16:19:11 +03:00
|
|
|
|
|
|
|
latest_event_ids = self.get_success(
|
|
|
|
self.store.get_latest_event_ids_in_room(self.room_id)
|
|
|
|
)
|
2023-09-18 16:29:05 +03:00
|
|
|
self.assertEqual(latest_event_ids, {event_id_a, event_id_b, event_id_c})
|
2019-05-29 16:19:11 +03:00
|
|
|
|
|
|
|
# Run the background update and check it did the right thing
|
|
|
|
self.run_background_update()
|
|
|
|
|
|
|
|
latest_event_ids = self.get_success(
|
|
|
|
self.store.get_latest_event_ids_in_room(self.room_id)
|
|
|
|
)
|
2023-09-18 16:29:05 +03:00
|
|
|
self.assertEqual(latest_event_ids, {event_id_b, event_id_c})
|
2019-06-17 20:04:42 +03:00
|
|
|
|
|
|
|
|
|
|
|
class CleanupExtremDummyEventsTestCase(HomeserverTestCase):
|
2019-09-26 13:47:53 +03:00
|
|
|
CONSENT_VERSION = "1"
|
|
|
|
EXTREMITIES_COUNT = 50
|
|
|
|
servlets = [
|
|
|
|
synapse.rest.admin.register_servlets_for_client_rest_resource,
|
|
|
|
login.register_servlets,
|
|
|
|
room.register_servlets,
|
|
|
|
]
|
|
|
|
|
2022-12-09 20:36:32 +03:00
|
|
|
def make_homeserver(self, reactor: MemoryReactor, clock: Clock) -> HomeServer:
|
2019-06-17 20:04:42 +03:00
|
|
|
config = self.default_config()
|
|
|
|
config["cleanup_extremities_with_dummy_events"] = True
|
|
|
|
return self.setup_test_homeserver(config=config)
|
|
|
|
|
2022-12-09 20:36:32 +03:00
|
|
|
def prepare(
|
|
|
|
self, reactor: MemoryReactor, clock: Clock, homeserver: HomeServer
|
|
|
|
) -> None:
|
2022-02-23 14:04:02 +03:00
|
|
|
self.store = homeserver.get_datastores().main
|
2019-06-17 20:04:42 +03:00
|
|
|
self.room_creator = homeserver.get_room_creation_handler()
|
2019-09-26 13:47:53 +03:00
|
|
|
self.event_creator_handler = homeserver.get_event_creation_handler()
|
2019-06-17 20:04:42 +03:00
|
|
|
|
|
|
|
# Create a test user and room
|
2019-09-26 13:47:53 +03:00
|
|
|
self.user = UserID.from_string(self.register_user("user1", "password"))
|
|
|
|
self.token1 = self.login("user1", "password")
|
2020-10-22 12:11:06 +03:00
|
|
|
self.requester = create_requester(self.user)
|
2023-02-22 14:07:28 +03:00
|
|
|
self.room_id, _, _ = self.get_success(
|
2022-04-01 17:55:09 +03:00
|
|
|
self.room_creator.create_room(self.requester, {"visibility": "public"})
|
|
|
|
)
|
2019-09-26 13:47:53 +03:00
|
|
|
self.event_creator = homeserver.get_event_creation_handler()
|
2021-09-23 14:13:34 +03:00
|
|
|
homeserver.config.consent.user_consent_version = self.CONSENT_VERSION
|
2019-06-17 20:04:42 +03:00
|
|
|
|
2022-12-09 20:36:32 +03:00
|
|
|
def test_send_dummy_event(self) -> None:
|
2019-09-26 13:47:53 +03:00
|
|
|
self._create_extremity_rich_graph()
|
2019-06-17 20:04:42 +03:00
|
|
|
|
2019-09-26 13:47:53 +03:00
|
|
|
# Pump the reactor repeatedly so that the background updates have a
|
|
|
|
# chance to run.
|
2020-08-27 13:39:53 +03:00
|
|
|
self.pump(20)
|
2019-06-17 20:04:42 +03:00
|
|
|
|
|
|
|
latest_event_ids = self.get_success(
|
|
|
|
self.store.get_latest_event_ids_in_room(self.room_id)
|
|
|
|
)
|
2019-09-26 13:47:53 +03:00
|
|
|
self.assertTrue(len(latest_event_ids) < 10, len(latest_event_ids))
|
2019-06-17 20:04:42 +03:00
|
|
|
|
2019-09-26 13:47:53 +03:00
|
|
|
@patch("synapse.handlers.message._DUMMY_EVENT_ROOM_EXCLUSION_EXPIRY", new=0)
|
2022-12-09 20:36:32 +03:00
|
|
|
def test_send_dummy_events_when_insufficient_power(self) -> None:
|
2019-09-26 13:47:53 +03:00
|
|
|
self._create_extremity_rich_graph()
|
|
|
|
# Criple power levels
|
|
|
|
self.helper.send_state(
|
|
|
|
self.room_id,
|
|
|
|
EventTypes.PowerLevels,
|
|
|
|
body={"users": {str(self.user): -1}},
|
|
|
|
tok=self.token1,
|
|
|
|
)
|
2019-06-19 13:33:03 +03:00
|
|
|
# Pump the reactor repeatedly so that the background updates have a
|
2019-06-17 20:04:42 +03:00
|
|
|
# chance to run.
|
|
|
|
self.pump(10 * 60)
|
|
|
|
|
2019-09-26 13:47:53 +03:00
|
|
|
latest_event_ids = self.get_success(
|
|
|
|
self.store.get_latest_event_ids_in_room(self.room_id)
|
|
|
|
)
|
|
|
|
# Check that the room has not been pruned
|
|
|
|
self.assertTrue(len(latest_event_ids) > 10)
|
|
|
|
|
|
|
|
# New user with regular levels
|
|
|
|
user2 = self.register_user("user2", "password")
|
|
|
|
token2 = self.login("user2", "password")
|
|
|
|
self.helper.join(self.room_id, user2, tok=token2)
|
|
|
|
self.pump(10 * 60)
|
|
|
|
|
|
|
|
latest_event_ids = self.get_success(
|
|
|
|
self.store.get_latest_event_ids_in_room(self.room_id)
|
|
|
|
)
|
|
|
|
self.assertTrue(len(latest_event_ids) < 10, len(latest_event_ids))
|
|
|
|
|
|
|
|
@patch("synapse.handlers.message._DUMMY_EVENT_ROOM_EXCLUSION_EXPIRY", new=250)
|
2022-12-09 20:36:32 +03:00
|
|
|
def test_expiry_logic(self) -> None:
|
2019-09-26 13:47:53 +03:00
|
|
|
"""Simple test to ensure that _expire_rooms_to_exclude_from_dummy_event_insertion()
|
|
|
|
expires old entries correctly.
|
|
|
|
"""
|
|
|
|
self.event_creator_handler._rooms_to_exclude_from_dummy_event_insertion[
|
|
|
|
"1"
|
|
|
|
] = 100000
|
|
|
|
self.event_creator_handler._rooms_to_exclude_from_dummy_event_insertion[
|
|
|
|
"2"
|
|
|
|
] = 200000
|
|
|
|
self.event_creator_handler._rooms_to_exclude_from_dummy_event_insertion[
|
|
|
|
"3"
|
|
|
|
] = 300000
|
2020-08-19 15:07:57 +03:00
|
|
|
|
2019-09-26 13:47:53 +03:00
|
|
|
self.event_creator_handler._expire_rooms_to_exclude_from_dummy_event_insertion()
|
|
|
|
# All entries within time frame
|
|
|
|
self.assertEqual(
|
|
|
|
len(
|
|
|
|
self.event_creator_handler._rooms_to_exclude_from_dummy_event_insertion
|
|
|
|
),
|
|
|
|
3,
|
|
|
|
)
|
|
|
|
# Oldest room to expire
|
2020-08-19 15:07:57 +03:00
|
|
|
self.pump(1.01)
|
2019-09-26 13:47:53 +03:00
|
|
|
self.event_creator_handler._expire_rooms_to_exclude_from_dummy_event_insertion()
|
|
|
|
self.assertEqual(
|
|
|
|
len(
|
|
|
|
self.event_creator_handler._rooms_to_exclude_from_dummy_event_insertion
|
|
|
|
),
|
|
|
|
2,
|
|
|
|
)
|
|
|
|
# All rooms to expire
|
|
|
|
self.pump(2)
|
|
|
|
self.assertEqual(
|
|
|
|
len(
|
|
|
|
self.event_creator_handler._rooms_to_exclude_from_dummy_event_insertion
|
|
|
|
),
|
|
|
|
0,
|
|
|
|
)
|
|
|
|
|
2022-12-09 20:36:32 +03:00
|
|
|
def _create_extremity_rich_graph(self) -> None:
|
2019-09-26 13:47:53 +03:00
|
|
|
"""Helper method to create bushy graph on demand"""
|
|
|
|
|
|
|
|
event_id_start = self.create_and_send_event(self.room_id, self.user)
|
|
|
|
|
|
|
|
for _ in range(self.EXTREMITIES_COUNT):
|
|
|
|
self.create_and_send_event(
|
|
|
|
self.room_id, self.user, prev_event_ids=[event_id_start]
|
|
|
|
)
|
|
|
|
|
|
|
|
latest_event_ids = self.get_success(
|
|
|
|
self.store.get_latest_event_ids_in_room(self.room_id)
|
|
|
|
)
|
|
|
|
self.assertEqual(len(latest_event_ids), 50)
|
|
|
|
|
2022-12-09 20:36:32 +03:00
|
|
|
def _enable_consent_checking(self) -> None:
|
2019-09-26 13:47:53 +03:00
|
|
|
"""Helper method to enable consent checking"""
|
|
|
|
self.event_creator._block_events_without_consent_error = "No consent from user"
|
|
|
|
consent_uri_builder = Mock()
|
|
|
|
consent_uri_builder.build_user_consent_uri.return_value = "http://example.com"
|
|
|
|
self.event_creator._consent_uri_builder = consent_uri_builder
|