mirror of
https://github.com/element-hq/synapse.git
synced 2024-12-18 00:43:30 +03:00
Update code to refer to "workers". (#15606)
A bunch of comments and variables are out of date and use obsolete terms.
This commit is contained in:
parent
7148c2a0d6
commit
375b0a8a11
17 changed files with 38 additions and 69 deletions
1
changelog.d/15606.misc
Normal file
1
changelog.d/15606.misc
Normal file
|
@ -0,0 +1 @@
|
||||||
|
Update internal terminology for workers.
|
|
@ -30,12 +30,6 @@ minimal.
|
||||||
|
|
||||||
See [the TCP replication documentation](tcp_replication.md).
|
See [the TCP replication documentation](tcp_replication.md).
|
||||||
|
|
||||||
### The Slaved DataStore
|
|
||||||
|
|
||||||
There are read-only version of the synapse storage layer in
|
|
||||||
`synapse/replication/slave/storage` that use the response of the
|
|
||||||
replication API to invalidate their caches.
|
|
||||||
|
|
||||||
### The TCP Replication Module
|
### The TCP Replication Module
|
||||||
Information about how the tcp replication module is structured, including how
|
Information about how the tcp replication module is structured, including how
|
||||||
the classes interact, can be found in
|
the classes interact, can be found in
|
||||||
|
|
|
@ -64,7 +64,7 @@ from synapse.util.logcontext import LoggingContext
|
||||||
logger = logging.getLogger("synapse.app.admin_cmd")
|
logger = logging.getLogger("synapse.app.admin_cmd")
|
||||||
|
|
||||||
|
|
||||||
class AdminCmdSlavedStore(
|
class AdminCmdStore(
|
||||||
FilteringWorkerStore,
|
FilteringWorkerStore,
|
||||||
ClientIpWorkerStore,
|
ClientIpWorkerStore,
|
||||||
DeviceWorkerStore,
|
DeviceWorkerStore,
|
||||||
|
@ -103,7 +103,7 @@ class AdminCmdSlavedStore(
|
||||||
|
|
||||||
|
|
||||||
class AdminCmdServer(HomeServer):
|
class AdminCmdServer(HomeServer):
|
||||||
DATASTORE_CLASS = AdminCmdSlavedStore # type: ignore
|
DATASTORE_CLASS = AdminCmdStore # type: ignore
|
||||||
|
|
||||||
|
|
||||||
async def export_data_command(hs: HomeServer, args: argparse.Namespace) -> None:
|
async def export_data_command(hs: HomeServer, args: argparse.Namespace) -> None:
|
||||||
|
|
|
@ -102,7 +102,7 @@ from synapse.util.httpresourcetree import create_resource_tree
|
||||||
logger = logging.getLogger("synapse.app.generic_worker")
|
logger = logging.getLogger("synapse.app.generic_worker")
|
||||||
|
|
||||||
|
|
||||||
class GenericWorkerSlavedStore(
|
class GenericWorkerStore(
|
||||||
# FIXME(#3714): We need to add UserDirectoryStore as we write directly
|
# FIXME(#3714): We need to add UserDirectoryStore as we write directly
|
||||||
# rather than going via the correct worker.
|
# rather than going via the correct worker.
|
||||||
UserDirectoryStore,
|
UserDirectoryStore,
|
||||||
|
@ -154,7 +154,7 @@ class GenericWorkerSlavedStore(
|
||||||
|
|
||||||
|
|
||||||
class GenericWorkerServer(HomeServer):
|
class GenericWorkerServer(HomeServer):
|
||||||
DATASTORE_CLASS = GenericWorkerSlavedStore # type: ignore
|
DATASTORE_CLASS = GenericWorkerStore # type: ignore
|
||||||
|
|
||||||
def _listen_http(self, listener_config: ListenerConfig) -> None:
|
def _listen_http(self, listener_config: ListenerConfig) -> None:
|
||||||
assert listener_config.http_options is not None
|
assert listener_config.http_options is not None
|
||||||
|
|
|
@ -134,7 +134,7 @@ from synapse.util.caches.descriptors import CachedFunction, cached as _cached
|
||||||
from synapse.util.frozenutils import freeze
|
from synapse.util.frozenutils import freeze
|
||||||
|
|
||||||
if TYPE_CHECKING:
|
if TYPE_CHECKING:
|
||||||
from synapse.app.generic_worker import GenericWorkerSlavedStore
|
from synapse.app.generic_worker import GenericWorkerStore
|
||||||
from synapse.server import HomeServer
|
from synapse.server import HomeServer
|
||||||
|
|
||||||
|
|
||||||
|
@ -237,9 +237,7 @@ class ModuleApi:
|
||||||
|
|
||||||
# TODO: Fix this type hint once the types for the data stores have been ironed
|
# TODO: Fix this type hint once the types for the data stores have been ironed
|
||||||
# out.
|
# out.
|
||||||
self._store: Union[
|
self._store: Union[DataStore, "GenericWorkerStore"] = hs.get_datastores().main
|
||||||
DataStore, "GenericWorkerSlavedStore"
|
|
||||||
] = hs.get_datastores().main
|
|
||||||
self._storage_controllers = hs.get_storage_controllers()
|
self._storage_controllers = hs.get_storage_controllers()
|
||||||
self._auth = hs.get_auth()
|
self._auth = hs.get_auth()
|
||||||
self._auth_handler = auth_handler
|
self._auth_handler = auth_handler
|
||||||
|
|
|
@ -60,7 +60,7 @@ _WAIT_FOR_REPLICATION_TIMEOUT_SECONDS = 5
|
||||||
class ReplicationDataHandler:
|
class ReplicationDataHandler:
|
||||||
"""Handles incoming stream updates from replication.
|
"""Handles incoming stream updates from replication.
|
||||||
|
|
||||||
This instance notifies the slave data store about updates. Can be subclassed
|
This instance notifies the data store about updates. Can be subclassed
|
||||||
to handle updates in additional ways.
|
to handle updates in additional ways.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
|
@ -91,7 +91,7 @@ class ReplicationDataHandler:
|
||||||
) -> None:
|
) -> None:
|
||||||
"""Called to handle a batch of replication data with a given stream token.
|
"""Called to handle a batch of replication data with a given stream token.
|
||||||
|
|
||||||
By default this just pokes the slave store. Can be overridden in subclasses to
|
By default, this just pokes the data store. Can be overridden in subclasses to
|
||||||
handle more.
|
handle more.
|
||||||
|
|
||||||
Args:
|
Args:
|
||||||
|
|
|
@ -85,13 +85,10 @@ class AccountDataWorkerStore(PushRulesWorkerStore, CacheInvalidationWorkerStore)
|
||||||
writers=hs.config.worker.writers.account_data,
|
writers=hs.config.worker.writers.account_data,
|
||||||
)
|
)
|
||||||
else:
|
else:
|
||||||
|
# Multiple writers are not supported for SQLite.
|
||||||
|
#
|
||||||
# We shouldn't be running in worker mode with SQLite, but its useful
|
# We shouldn't be running in worker mode with SQLite, but its useful
|
||||||
# to support it for unit tests.
|
# to support it for unit tests.
|
||||||
#
|
|
||||||
# If this process is the writer than we need to use
|
|
||||||
# `StreamIdGenerator`, otherwise we use `SlavedIdTracker` which gets
|
|
||||||
# updated over replication. (Multiple writers are not supported for
|
|
||||||
# SQLite).
|
|
||||||
self._account_data_id_gen = StreamIdGenerator(
|
self._account_data_id_gen = StreamIdGenerator(
|
||||||
db_conn,
|
db_conn,
|
||||||
hs.get_replication_notifier(),
|
hs.get_replication_notifier(),
|
||||||
|
|
|
@ -274,11 +274,11 @@ class CacheInvalidationWorkerStore(SQLBaseStore):
|
||||||
async def invalidate_cache_and_stream(
|
async def invalidate_cache_and_stream(
|
||||||
self, cache_name: str, keys: Tuple[Any, ...]
|
self, cache_name: str, keys: Tuple[Any, ...]
|
||||||
) -> None:
|
) -> None:
|
||||||
"""Invalidates the cache and adds it to the cache stream so slaves
|
"""Invalidates the cache and adds it to the cache stream so other workers
|
||||||
will know to invalidate their caches.
|
will know to invalidate their caches.
|
||||||
|
|
||||||
This should only be used to invalidate caches where slaves won't
|
This should only be used to invalidate caches where other workers won't
|
||||||
otherwise know from other replication streams that the cache should
|
otherwise have known from other replication streams that the cache should
|
||||||
be invalidated.
|
be invalidated.
|
||||||
"""
|
"""
|
||||||
cache_func = getattr(self, cache_name, None)
|
cache_func = getattr(self, cache_name, None)
|
||||||
|
@ -297,11 +297,11 @@ class CacheInvalidationWorkerStore(SQLBaseStore):
|
||||||
cache_func: CachedFunction,
|
cache_func: CachedFunction,
|
||||||
keys: Tuple[Any, ...],
|
keys: Tuple[Any, ...],
|
||||||
) -> None:
|
) -> None:
|
||||||
"""Invalidates the cache and adds it to the cache stream so slaves
|
"""Invalidates the cache and adds it to the cache stream so other workers
|
||||||
will know to invalidate their caches.
|
will know to invalidate their caches.
|
||||||
|
|
||||||
This should only be used to invalidate caches where slaves won't
|
This should only be used to invalidate caches where other workers won't
|
||||||
otherwise know from other replication streams that the cache should
|
otherwise have known from other replication streams that the cache should
|
||||||
be invalidated.
|
be invalidated.
|
||||||
"""
|
"""
|
||||||
txn.call_after(cache_func.invalidate, keys)
|
txn.call_after(cache_func.invalidate, keys)
|
||||||
|
@ -310,7 +310,7 @@ class CacheInvalidationWorkerStore(SQLBaseStore):
|
||||||
def _invalidate_all_cache_and_stream(
|
def _invalidate_all_cache_and_stream(
|
||||||
self, txn: LoggingTransaction, cache_func: CachedFunction
|
self, txn: LoggingTransaction, cache_func: CachedFunction
|
||||||
) -> None:
|
) -> None:
|
||||||
"""Invalidates the entire cache and adds it to the cache stream so slaves
|
"""Invalidates the entire cache and adds it to the cache stream so other workers
|
||||||
will know to invalidate their caches.
|
will know to invalidate their caches.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
|
|
|
@ -105,8 +105,6 @@ class DeviceWorkerStore(RoomMemberWorkerStore, EndToEndKeyWorkerStore):
|
||||||
is_writer=hs.config.worker.worker_app is None,
|
is_writer=hs.config.worker.worker_app is None,
|
||||||
)
|
)
|
||||||
|
|
||||||
# Type-ignore: _device_list_id_gen is mixed in from either DataStore (as a
|
|
||||||
# StreamIdGenerator) or SlavedDataStore (as a SlavedIdTracker).
|
|
||||||
device_list_max = self._device_list_id_gen.get_current_token()
|
device_list_max = self._device_list_id_gen.get_current_token()
|
||||||
device_list_prefill, min_device_list_id = self.db_pool.get_cache_dict(
|
device_list_prefill, min_device_list_id = self.db_pool.get_cache_dict(
|
||||||
db_conn,
|
db_conn,
|
||||||
|
|
|
@ -213,13 +213,10 @@ class EventsWorkerStore(SQLBaseStore):
|
||||||
writers=hs.config.worker.writers.events,
|
writers=hs.config.worker.writers.events,
|
||||||
)
|
)
|
||||||
else:
|
else:
|
||||||
|
# Multiple writers are not supported for SQLite.
|
||||||
|
#
|
||||||
# We shouldn't be running in worker mode with SQLite, but its useful
|
# We shouldn't be running in worker mode with SQLite, but its useful
|
||||||
# to support it for unit tests.
|
# to support it for unit tests.
|
||||||
#
|
|
||||||
# If this process is the writer than we need to use
|
|
||||||
# `StreamIdGenerator`, otherwise we use `SlavedIdTracker` which gets
|
|
||||||
# updated over replication. (Multiple writers are not supported for
|
|
||||||
# SQLite).
|
|
||||||
self._stream_id_gen = StreamIdGenerator(
|
self._stream_id_gen = StreamIdGenerator(
|
||||||
db_conn,
|
db_conn,
|
||||||
hs.get_replication_notifier(),
|
hs.get_replication_notifier(),
|
||||||
|
|
|
@ -85,13 +85,10 @@ class ReceiptsWorkerStore(SQLBaseStore):
|
||||||
else:
|
else:
|
||||||
self._can_write_to_receipts = True
|
self._can_write_to_receipts = True
|
||||||
|
|
||||||
|
# Multiple writers are not supported for SQLite.
|
||||||
|
#
|
||||||
# We shouldn't be running in worker mode with SQLite, but its useful
|
# We shouldn't be running in worker mode with SQLite, but its useful
|
||||||
# to support it for unit tests.
|
# to support it for unit tests.
|
||||||
#
|
|
||||||
# If this process is the writer than we need to use
|
|
||||||
# `StreamIdGenerator`, otherwise we use `SlavedIdTracker` which gets
|
|
||||||
# updated over replication. (Multiple writers are not supported for
|
|
||||||
# SQLite).
|
|
||||||
self._receipts_id_gen = StreamIdGenerator(
|
self._receipts_id_gen = StreamIdGenerator(
|
||||||
db_conn,
|
db_conn,
|
||||||
hs.get_replication_notifier(),
|
hs.get_replication_notifier(),
|
||||||
|
|
|
@ -21,7 +21,7 @@ from synapse.storage.prepare_database import get_statements
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
# This stream is used to notify replication slaves that some caches have
|
# This stream is used to notify workers over replication that some caches have
|
||||||
# been invalidated that they cannot infer from the other streams.
|
# been invalidated that they cannot infer from the other streams.
|
||||||
CREATE_TABLE = """
|
CREATE_TABLE = """
|
||||||
CREATE TABLE cache_invalidation_stream (
|
CREATE TABLE cache_invalidation_stream (
|
||||||
|
|
|
@ -38,7 +38,7 @@ class FederationReaderOpenIDListenerTests(HomeserverTestCase):
|
||||||
|
|
||||||
def default_config(self) -> JsonDict:
|
def default_config(self) -> JsonDict:
|
||||||
conf = super().default_config()
|
conf = super().default_config()
|
||||||
# we're using FederationReaderServer, which uses a SlavedStore, so we
|
# we're using GenericWorkerServer, which uses a GenericWorkerStore, so we
|
||||||
# have to tell the FederationHandler not to try to access stuff that is only
|
# have to tell the FederationHandler not to try to access stuff that is only
|
||||||
# in the primary store.
|
# in the primary store.
|
||||||
conf["worker_app"] = "yes"
|
conf["worker_app"] = "yes"
|
||||||
|
|
|
@ -1,13 +0,0 @@
|
||||||
# Copyright 2016 OpenMarket Ltd
|
|
||||||
#
|
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
||||||
# you may not use this file except in compliance with the License.
|
|
||||||
# You may obtain a copy of the License at
|
|
||||||
#
|
|
||||||
# http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
#
|
|
||||||
# Unless required by applicable law or agreed to in writing, software
|
|
||||||
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
# See the License for the specific language governing permissions and
|
|
||||||
# limitations under the License.
|
|
|
@ -24,7 +24,7 @@ from synapse.util import Clock
|
||||||
from tests.replication._base import BaseStreamTestCase
|
from tests.replication._base import BaseStreamTestCase
|
||||||
|
|
||||||
|
|
||||||
class BaseSlavedStoreTestCase(BaseStreamTestCase):
|
class BaseWorkerStoreTestCase(BaseStreamTestCase):
|
||||||
def make_homeserver(self, reactor: MemoryReactor, clock: Clock) -> HomeServer:
|
def make_homeserver(self, reactor: MemoryReactor, clock: Clock) -> HomeServer:
|
||||||
return self.setup_test_homeserver(federation_client=Mock())
|
return self.setup_test_homeserver(federation_client=Mock())
|
||||||
|
|
||||||
|
@ -34,7 +34,7 @@ class BaseSlavedStoreTestCase(BaseStreamTestCase):
|
||||||
self.reconnect()
|
self.reconnect()
|
||||||
|
|
||||||
self.master_store = hs.get_datastores().main
|
self.master_store = hs.get_datastores().main
|
||||||
self.slaved_store = self.worker_hs.get_datastores().main
|
self.worker_store = self.worker_hs.get_datastores().main
|
||||||
persistence = hs.get_storage_controllers().persistence
|
persistence = hs.get_storage_controllers().persistence
|
||||||
assert persistence is not None
|
assert persistence is not None
|
||||||
self.persistance = persistence
|
self.persistance = persistence
|
||||||
|
@ -50,7 +50,7 @@ class BaseSlavedStoreTestCase(BaseStreamTestCase):
|
||||||
self, method: str, args: Iterable[Any], expected_result: Optional[Any] = None
|
self, method: str, args: Iterable[Any], expected_result: Optional[Any] = None
|
||||||
) -> None:
|
) -> None:
|
||||||
master_result = self.get_success(getattr(self.master_store, method)(*args))
|
master_result = self.get_success(getattr(self.master_store, method)(*args))
|
||||||
slaved_result = self.get_success(getattr(self.slaved_store, method)(*args))
|
worker_result = self.get_success(getattr(self.worker_store, method)(*args))
|
||||||
if expected_result is not None:
|
if expected_result is not None:
|
||||||
self.assertEqual(
|
self.assertEqual(
|
||||||
master_result,
|
master_result,
|
||||||
|
@ -59,14 +59,14 @@ class BaseSlavedStoreTestCase(BaseStreamTestCase):
|
||||||
% (expected_result, master_result),
|
% (expected_result, master_result),
|
||||||
)
|
)
|
||||||
self.assertEqual(
|
self.assertEqual(
|
||||||
slaved_result,
|
worker_result,
|
||||||
expected_result,
|
expected_result,
|
||||||
"Expected slave result to be %r but was %r"
|
"Expected worker result to be %r but was %r"
|
||||||
% (expected_result, slaved_result),
|
% (expected_result, worker_result),
|
||||||
)
|
)
|
||||||
self.assertEqual(
|
self.assertEqual(
|
||||||
master_result,
|
master_result,
|
||||||
slaved_result,
|
worker_result,
|
||||||
"Slave result %r does not match master result %r"
|
"Worker result %r does not match master result %r"
|
||||||
% (slaved_result, master_result),
|
% (worker_result, master_result),
|
||||||
)
|
)
|
|
@ -36,7 +36,7 @@ from synapse.util import Clock
|
||||||
|
|
||||||
from tests.server import FakeTransport
|
from tests.server import FakeTransport
|
||||||
|
|
||||||
from ._base import BaseSlavedStoreTestCase
|
from ._base import BaseWorkerStoreTestCase
|
||||||
|
|
||||||
USER_ID = "@feeling:test"
|
USER_ID = "@feeling:test"
|
||||||
USER_ID_2 = "@bright:test"
|
USER_ID_2 = "@bright:test"
|
||||||
|
@ -63,7 +63,7 @@ def patch__eq__(cls: object) -> Callable[[], None]:
|
||||||
return unpatch
|
return unpatch
|
||||||
|
|
||||||
|
|
||||||
class EventsWorkerStoreTestCase(BaseSlavedStoreTestCase):
|
class EventsWorkerStoreTestCase(BaseWorkerStoreTestCase):
|
||||||
STORE_TYPE = EventsWorkerStore
|
STORE_TYPE = EventsWorkerStore
|
||||||
|
|
||||||
def setUp(self) -> None:
|
def setUp(self) -> None:
|
||||||
|
@ -294,7 +294,7 @@ class EventsWorkerStoreTestCase(BaseSlavedStoreTestCase):
|
||||||
assert j2.internal_metadata.stream_ordering is not None
|
assert j2.internal_metadata.stream_ordering is not None
|
||||||
|
|
||||||
event_source = RoomEventSource(self.hs)
|
event_source = RoomEventSource(self.hs)
|
||||||
event_source.store = self.slaved_store
|
event_source.store = self.worker_store
|
||||||
current_token = event_source.get_current_key()
|
current_token = event_source.get_current_key()
|
||||||
|
|
||||||
# gradually stream out the replication
|
# gradually stream out the replication
|
||||||
|
@ -310,12 +310,12 @@ class EventsWorkerStoreTestCase(BaseSlavedStoreTestCase):
|
||||||
#
|
#
|
||||||
# First, we get a list of the rooms we are joined to
|
# First, we get a list of the rooms we are joined to
|
||||||
joined_rooms = self.get_success(
|
joined_rooms = self.get_success(
|
||||||
self.slaved_store.get_rooms_for_user_with_stream_ordering(USER_ID_2)
|
self.worker_store.get_rooms_for_user_with_stream_ordering(USER_ID_2)
|
||||||
)
|
)
|
||||||
|
|
||||||
# Then, we get a list of the events since the last sync
|
# Then, we get a list of the events since the last sync
|
||||||
membership_changes = self.get_success(
|
membership_changes = self.get_success(
|
||||||
self.slaved_store.get_membership_changes_for_user(
|
self.worker_store.get_membership_changes_for_user(
|
||||||
USER_ID_2, prev_token, current_token
|
USER_ID_2, prev_token, current_token
|
||||||
)
|
)
|
||||||
)
|
)
|
Loading…
Reference in a new issue