2017-03-22 16:54:20 +03:00
|
|
|
# Copyright 2016 OpenMarket Ltd
|
2018-04-05 18:24:04 +03:00
|
|
|
# Copyright 2018 New Vector Ltd
|
2017-03-22 16:54:20 +03:00
|
|
|
#
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
2017-03-30 15:22:24 +03:00
|
|
|
import logging
|
2023-05-24 17:18:52 +03:00
|
|
|
from typing import (
|
|
|
|
Any,
|
|
|
|
Generator,
|
|
|
|
Iterable,
|
|
|
|
List,
|
2023-09-20 14:48:55 +03:00
|
|
|
Mapping,
|
2023-05-24 17:18:52 +03:00
|
|
|
NoReturn,
|
|
|
|
Optional,
|
|
|
|
Set,
|
|
|
|
Tuple,
|
|
|
|
cast,
|
|
|
|
)
|
2021-04-09 20:44:38 +03:00
|
|
|
from unittest import mock
|
2018-07-09 09:09:20 +03:00
|
|
|
|
|
|
|
from twisted.internet import defer, reactor
|
2022-03-14 22:04:29 +03:00
|
|
|
from twisted.internet.defer import CancelledError, Deferred
|
2022-11-23 01:35:54 +03:00
|
|
|
from twisted.internet.interfaces import IReactorTime
|
2018-07-09 09:09:20 +03:00
|
|
|
|
2017-03-30 15:22:24 +03:00
|
|
|
from synapse.api.errors import SynapseError
|
2019-07-03 17:07:04 +03:00
|
|
|
from synapse.logging.context import (
|
2020-03-24 17:45:33 +03:00
|
|
|
SENTINEL_CONTEXT,
|
2019-07-03 17:07:04 +03:00
|
|
|
LoggingContext,
|
|
|
|
PreserveLoggingContext,
|
2020-03-24 17:45:33 +03:00
|
|
|
current_context,
|
2019-07-03 17:07:04 +03:00
|
|
|
make_deferred_yieldable,
|
|
|
|
)
|
2017-03-22 16:54:20 +03:00
|
|
|
from synapse.util.caches import descriptors
|
2023-05-24 17:18:52 +03:00
|
|
|
from synapse.util.caches.descriptors import _CacheContext, cached, cachedList
|
2018-07-09 09:09:20 +03:00
|
|
|
|
2017-03-22 16:54:20 +03:00
|
|
|
from tests import unittest
|
2020-10-30 14:43:17 +03:00
|
|
|
from tests.test_utils import get_awaitable_result
|
2017-03-22 16:54:20 +03:00
|
|
|
|
2017-03-30 15:22:24 +03:00
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
2017-03-22 16:54:20 +03:00
|
|
|
|
2023-05-24 17:18:52 +03:00
|
|
|
def run_on_reactor() -> "Deferred[int]":
|
|
|
|
d: "Deferred[int]" = Deferred()
|
2022-11-23 01:35:54 +03:00
|
|
|
cast(IReactorTime, reactor).callLater(0, d.callback, 0)
|
2019-07-03 17:07:04 +03:00
|
|
|
return make_deferred_yieldable(d)
|
2018-07-04 11:35:40 +03:00
|
|
|
|
|
|
|
|
2017-03-22 16:54:20 +03:00
|
|
|
class DescriptorTestCase(unittest.TestCase):
|
|
|
|
@defer.inlineCallbacks
|
2023-05-24 17:18:52 +03:00
|
|
|
def test_cache(self) -> Generator["Deferred[Any]", object, None]:
|
2020-09-04 13:54:56 +03:00
|
|
|
class Cls:
|
2023-05-24 17:18:52 +03:00
|
|
|
def __init__(self) -> None:
|
2017-03-22 16:54:20 +03:00
|
|
|
self.mock = mock.Mock()
|
|
|
|
|
|
|
|
@descriptors.cached()
|
2023-05-24 17:18:52 +03:00
|
|
|
def fn(self, arg1: int, arg2: int) -> str:
|
2017-03-22 16:54:20 +03:00
|
|
|
return self.mock(arg1, arg2)
|
|
|
|
|
|
|
|
obj = Cls()
|
|
|
|
|
|
|
|
obj.mock.return_value = "fish"
|
|
|
|
r = yield obj.fn(1, 2)
|
|
|
|
self.assertEqual(r, "fish")
|
|
|
|
obj.mock.assert_called_once_with(1, 2)
|
|
|
|
obj.mock.reset_mock()
|
|
|
|
|
|
|
|
# a call with different params should call the mock again
|
|
|
|
obj.mock.return_value = "chips"
|
|
|
|
r = yield obj.fn(1, 3)
|
|
|
|
self.assertEqual(r, "chips")
|
|
|
|
obj.mock.assert_called_once_with(1, 3)
|
|
|
|
obj.mock.reset_mock()
|
|
|
|
|
|
|
|
# the two values should now be cached
|
|
|
|
r = yield obj.fn(1, 2)
|
|
|
|
self.assertEqual(r, "fish")
|
|
|
|
r = yield obj.fn(1, 3)
|
|
|
|
self.assertEqual(r, "chips")
|
|
|
|
obj.mock.assert_not_called()
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
2023-05-24 17:18:52 +03:00
|
|
|
def test_cache_num_args(self) -> Generator["Deferred[Any]", object, None]:
|
2017-03-22 16:54:20 +03:00
|
|
|
"""Only the first num_args arguments should matter to the cache"""
|
|
|
|
|
2020-09-04 13:54:56 +03:00
|
|
|
class Cls:
|
2023-05-24 17:18:52 +03:00
|
|
|
def __init__(self) -> None:
|
2017-03-22 16:54:20 +03:00
|
|
|
self.mock = mock.Mock()
|
|
|
|
|
|
|
|
@descriptors.cached(num_args=1)
|
2023-09-20 14:48:55 +03:00
|
|
|
def fn(self, arg1: int, arg2: int) -> str:
|
2017-03-22 16:54:20 +03:00
|
|
|
return self.mock(arg1, arg2)
|
|
|
|
|
|
|
|
obj = Cls()
|
|
|
|
obj.mock.return_value = "fish"
|
|
|
|
r = yield obj.fn(1, 2)
|
|
|
|
self.assertEqual(r, "fish")
|
|
|
|
obj.mock.assert_called_once_with(1, 2)
|
|
|
|
obj.mock.reset_mock()
|
|
|
|
|
|
|
|
# a call with different params should call the mock again
|
|
|
|
obj.mock.return_value = "chips"
|
|
|
|
r = yield obj.fn(2, 3)
|
|
|
|
self.assertEqual(r, "chips")
|
|
|
|
obj.mock.assert_called_once_with(2, 3)
|
|
|
|
obj.mock.reset_mock()
|
|
|
|
|
|
|
|
# the two values should now be cached; we should be able to vary
|
|
|
|
# the second argument and still get the cached result.
|
|
|
|
r = yield obj.fn(1, 4)
|
|
|
|
self.assertEqual(r, "fish")
|
|
|
|
r = yield obj.fn(2, 5)
|
|
|
|
self.assertEqual(r, "chips")
|
|
|
|
obj.mock.assert_not_called()
|
2017-03-30 15:22:24 +03:00
|
|
|
|
2022-03-09 21:07:41 +03:00
|
|
|
@defer.inlineCallbacks
|
2023-05-24 17:18:52 +03:00
|
|
|
def test_cache_uncached_args(self) -> Generator["Deferred[Any]", object, None]:
|
2022-03-09 21:07:41 +03:00
|
|
|
"""
|
|
|
|
Only the arguments not named in uncached_args should matter to the cache
|
|
|
|
|
|
|
|
Note that this is identical to test_cache_num_args, but provides the
|
|
|
|
arguments differently.
|
|
|
|
"""
|
|
|
|
|
|
|
|
class Cls:
|
|
|
|
# Note that it is important that this is not the last argument to
|
|
|
|
# test behaviour of skipping arguments properly.
|
|
|
|
@descriptors.cached(uncached_args=("arg2",))
|
2023-05-24 17:18:52 +03:00
|
|
|
def fn(self, arg1: int, arg2: int, arg3: int) -> str:
|
2022-03-09 21:07:41 +03:00
|
|
|
return self.mock(arg1, arg2, arg3)
|
|
|
|
|
2023-05-24 17:18:52 +03:00
|
|
|
def __init__(self) -> None:
|
2022-03-09 21:07:41 +03:00
|
|
|
self.mock = mock.Mock()
|
|
|
|
|
|
|
|
obj = Cls()
|
|
|
|
obj.mock.return_value = "fish"
|
|
|
|
r = yield obj.fn(1, 2, 3)
|
|
|
|
self.assertEqual(r, "fish")
|
|
|
|
obj.mock.assert_called_once_with(1, 2, 3)
|
|
|
|
obj.mock.reset_mock()
|
|
|
|
|
|
|
|
# a call with different params should call the mock again
|
|
|
|
obj.mock.return_value = "chips"
|
|
|
|
r = yield obj.fn(2, 3, 4)
|
|
|
|
self.assertEqual(r, "chips")
|
|
|
|
obj.mock.assert_called_once_with(2, 3, 4)
|
|
|
|
obj.mock.reset_mock()
|
|
|
|
|
|
|
|
# the two values should now be cached; we should be able to vary
|
|
|
|
# the second argument and still get the cached result.
|
|
|
|
r = yield obj.fn(1, 4, 3)
|
|
|
|
self.assertEqual(r, "fish")
|
|
|
|
r = yield obj.fn(2, 5, 4)
|
|
|
|
self.assertEqual(r, "chips")
|
|
|
|
obj.mock.assert_not_called()
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
2023-05-24 17:18:52 +03:00
|
|
|
def test_cache_kwargs(self) -> Generator["Deferred[Any]", object, None]:
|
2022-03-09 21:07:41 +03:00
|
|
|
"""Test that keyword arguments are treated properly"""
|
|
|
|
|
|
|
|
class Cls:
|
2023-05-24 17:18:52 +03:00
|
|
|
def __init__(self) -> None:
|
2022-03-09 21:07:41 +03:00
|
|
|
self.mock = mock.Mock()
|
|
|
|
|
|
|
|
@descriptors.cached()
|
2023-05-24 17:18:52 +03:00
|
|
|
def fn(self, arg1: int, kwarg1: int = 2) -> str:
|
2022-03-09 21:07:41 +03:00
|
|
|
return self.mock(arg1, kwarg1=kwarg1)
|
|
|
|
|
|
|
|
obj = Cls()
|
|
|
|
obj.mock.return_value = "fish"
|
|
|
|
r = yield obj.fn(1, kwarg1=2)
|
|
|
|
self.assertEqual(r, "fish")
|
|
|
|
obj.mock.assert_called_once_with(1, kwarg1=2)
|
|
|
|
obj.mock.reset_mock()
|
|
|
|
|
|
|
|
# a call with different params should call the mock again
|
|
|
|
obj.mock.return_value = "chips"
|
|
|
|
r = yield obj.fn(1, kwarg1=3)
|
|
|
|
self.assertEqual(r, "chips")
|
|
|
|
obj.mock.assert_called_once_with(1, kwarg1=3)
|
|
|
|
obj.mock.reset_mock()
|
|
|
|
|
|
|
|
# the values should now be cached.
|
|
|
|
r = yield obj.fn(1, kwarg1=2)
|
|
|
|
self.assertEqual(r, "fish")
|
|
|
|
# We should be able to not provide kwarg1 and get the cached value back.
|
|
|
|
r = yield obj.fn(1)
|
|
|
|
self.assertEqual(r, "fish")
|
|
|
|
# Keyword arguments can be in any order.
|
|
|
|
r = yield obj.fn(kwarg1=2, arg1=1)
|
|
|
|
self.assertEqual(r, "fish")
|
|
|
|
obj.mock.assert_not_called()
|
|
|
|
|
2023-05-24 17:18:52 +03:00
|
|
|
def test_cache_with_sync_exception(self) -> None:
|
2019-07-25 17:59:45 +03:00
|
|
|
"""If the wrapped function throws synchronously, things should continue to work"""
|
|
|
|
|
2020-09-04 13:54:56 +03:00
|
|
|
class Cls:
|
2019-07-25 17:59:45 +03:00
|
|
|
@cached()
|
2023-05-24 17:18:52 +03:00
|
|
|
def fn(self, arg1: int) -> NoReturn:
|
2019-07-25 17:59:45 +03:00
|
|
|
raise SynapseError(100, "mai spoon iz too big!!1")
|
|
|
|
|
|
|
|
obj = Cls()
|
|
|
|
|
|
|
|
# this should fail immediately
|
|
|
|
d = obj.fn(1)
|
|
|
|
self.failureResultOf(d, SynapseError)
|
|
|
|
|
|
|
|
# ... leaving the cache empty
|
|
|
|
self.assertEqual(len(obj.fn.cache.cache), 0)
|
|
|
|
|
|
|
|
# and a second call should result in a second exception
|
|
|
|
d = obj.fn(1)
|
|
|
|
self.failureResultOf(d, SynapseError)
|
|
|
|
|
2023-05-24 17:18:52 +03:00
|
|
|
def test_cache_with_async_exception(self) -> None:
|
2020-10-16 23:32:52 +03:00
|
|
|
"""The wrapped function returns a failure"""
|
|
|
|
|
|
|
|
class Cls:
|
2023-05-24 17:18:52 +03:00
|
|
|
result: Optional[Deferred] = None
|
2020-10-16 23:32:52 +03:00
|
|
|
call_count = 0
|
|
|
|
|
|
|
|
@cached()
|
2023-09-20 14:48:55 +03:00
|
|
|
def fn(self, arg1: int) -> Deferred:
|
2020-10-16 23:32:52 +03:00
|
|
|
self.call_count += 1
|
2023-09-20 14:48:55 +03:00
|
|
|
assert self.result is not None
|
2020-10-16 23:32:52 +03:00
|
|
|
return self.result
|
|
|
|
|
|
|
|
obj = Cls()
|
2021-07-13 13:52:58 +03:00
|
|
|
callbacks: Set[str] = set()
|
2020-10-16 23:32:52 +03:00
|
|
|
|
|
|
|
# set off an asynchronous request
|
2023-05-24 17:18:52 +03:00
|
|
|
origin_d: Deferred = Deferred()
|
2022-11-23 01:35:54 +03:00
|
|
|
obj.result = origin_d
|
2020-10-16 23:32:52 +03:00
|
|
|
|
|
|
|
d1 = obj.fn(1, on_invalidate=lambda: callbacks.add("d1"))
|
|
|
|
self.assertFalse(d1.called)
|
|
|
|
|
|
|
|
# a second request should also return a deferred, but should not call the
|
|
|
|
# function itself.
|
|
|
|
d2 = obj.fn(1, on_invalidate=lambda: callbacks.add("d2"))
|
|
|
|
self.assertFalse(d2.called)
|
|
|
|
self.assertEqual(obj.call_count, 1)
|
|
|
|
|
|
|
|
# no callbacks yet
|
|
|
|
self.assertEqual(callbacks, set())
|
|
|
|
|
|
|
|
# the original request fails
|
|
|
|
e = Exception("bzz")
|
|
|
|
origin_d.errback(e)
|
|
|
|
|
|
|
|
# ... which should cause the lookups to fail similarly
|
|
|
|
self.assertIs(self.failureResultOf(d1, Exception).value, e)
|
|
|
|
self.assertIs(self.failureResultOf(d2, Exception).value, e)
|
|
|
|
|
|
|
|
# ... and the callbacks to have been, uh, called.
|
|
|
|
self.assertEqual(callbacks, {"d1", "d2"})
|
|
|
|
|
|
|
|
# ... leaving the cache empty
|
|
|
|
self.assertEqual(len(obj.fn.cache.cache), 0)
|
|
|
|
|
|
|
|
# and a second call should work as normal
|
|
|
|
obj.result = defer.succeed(100)
|
|
|
|
d3 = obj.fn(1)
|
|
|
|
self.assertEqual(self.successResultOf(d3), 100)
|
|
|
|
self.assertEqual(obj.call_count, 2)
|
|
|
|
|
2023-05-24 17:18:52 +03:00
|
|
|
def test_cache_logcontexts(self) -> Deferred:
|
2017-03-30 15:22:24 +03:00
|
|
|
"""Check that logcontexts are set and restored correctly when
|
|
|
|
using the cache."""
|
|
|
|
|
2023-05-24 17:18:52 +03:00
|
|
|
complete_lookup: Deferred = Deferred()
|
2017-03-30 15:22:24 +03:00
|
|
|
|
2020-09-04 13:54:56 +03:00
|
|
|
class Cls:
|
2017-03-30 15:22:24 +03:00
|
|
|
@descriptors.cached()
|
2023-05-24 17:18:52 +03:00
|
|
|
def fn(self, arg1: int) -> "Deferred[int]":
|
2017-03-30 15:22:24 +03:00
|
|
|
@defer.inlineCallbacks
|
2023-05-24 17:18:52 +03:00
|
|
|
def inner_fn() -> Generator["Deferred[object]", object, int]:
|
2019-07-03 17:07:04 +03:00
|
|
|
with PreserveLoggingContext():
|
2017-03-30 15:22:24 +03:00
|
|
|
yield complete_lookup
|
2019-07-23 16:00:55 +03:00
|
|
|
return 1
|
2017-03-30 15:22:24 +03:00
|
|
|
|
|
|
|
return inner_fn()
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
2023-05-24 17:18:52 +03:00
|
|
|
def do_lookup() -> Generator["Deferred[Any]", object, int]:
|
2021-04-20 16:19:00 +03:00
|
|
|
with LoggingContext("c1") as c1:
|
2017-03-30 15:22:24 +03:00
|
|
|
r = yield obj.fn(1)
|
2020-03-24 17:45:33 +03:00
|
|
|
self.assertEqual(current_context(), c1)
|
2023-05-24 17:18:52 +03:00
|
|
|
return cast(int, r)
|
2017-03-30 15:22:24 +03:00
|
|
|
|
2023-05-24 17:18:52 +03:00
|
|
|
def check_result(r: int) -> None:
|
2017-03-30 15:22:24 +03:00
|
|
|
self.assertEqual(r, 1)
|
|
|
|
|
|
|
|
obj = Cls()
|
|
|
|
|
|
|
|
# set off a deferred which will do a cache lookup
|
|
|
|
d1 = do_lookup()
|
2020-03-24 17:45:33 +03:00
|
|
|
self.assertEqual(current_context(), SENTINEL_CONTEXT)
|
2017-03-30 15:22:24 +03:00
|
|
|
d1.addCallback(check_result)
|
|
|
|
|
|
|
|
# and another
|
|
|
|
d2 = do_lookup()
|
2020-03-24 17:45:33 +03:00
|
|
|
self.assertEqual(current_context(), SENTINEL_CONTEXT)
|
2017-03-30 15:22:24 +03:00
|
|
|
d2.addCallback(check_result)
|
|
|
|
|
|
|
|
# let the lookup complete
|
|
|
|
complete_lookup.callback(None)
|
|
|
|
|
|
|
|
return defer.gatherResults([d1, d2])
|
|
|
|
|
2023-05-24 17:18:52 +03:00
|
|
|
def test_cache_logcontexts_with_exception(self) -> "Deferred[None]":
|
2017-03-30 15:22:24 +03:00
|
|
|
"""Check that the cache sets and restores logcontexts correctly when
|
|
|
|
the lookup function throws an exception"""
|
|
|
|
|
2020-09-04 13:54:56 +03:00
|
|
|
class Cls:
|
2017-03-30 15:22:24 +03:00
|
|
|
@descriptors.cached()
|
2023-05-24 17:18:52 +03:00
|
|
|
def fn(self, arg1: int) -> Deferred:
|
2017-03-30 15:22:24 +03:00
|
|
|
@defer.inlineCallbacks
|
2023-05-24 17:18:52 +03:00
|
|
|
def inner_fn() -> Generator["Deferred[Any]", object, NoReturn]:
|
2018-07-04 11:35:40 +03:00
|
|
|
# we want this to behave like an asynchronous function
|
|
|
|
yield run_on_reactor()
|
2017-03-30 15:22:24 +03:00
|
|
|
raise SynapseError(400, "blah")
|
|
|
|
|
|
|
|
return inner_fn()
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
2023-05-24 17:18:52 +03:00
|
|
|
def do_lookup() -> Generator["Deferred[object]", object, None]:
|
2021-04-20 16:19:00 +03:00
|
|
|
with LoggingContext("c1") as c1:
|
2017-03-30 15:22:24 +03:00
|
|
|
try:
|
2018-07-04 11:35:40 +03:00
|
|
|
d = obj.fn(1)
|
|
|
|
self.assertEqual(
|
2020-03-24 17:45:33 +03:00
|
|
|
current_context(),
|
|
|
|
SENTINEL_CONTEXT,
|
2018-07-04 11:35:40 +03:00
|
|
|
)
|
|
|
|
yield d
|
2017-03-30 15:22:24 +03:00
|
|
|
self.fail("No exception thrown")
|
|
|
|
except SynapseError:
|
|
|
|
pass
|
|
|
|
|
2020-03-24 17:45:33 +03:00
|
|
|
self.assertEqual(current_context(), c1)
|
2017-03-30 15:22:24 +03:00
|
|
|
|
2019-07-25 17:59:45 +03:00
|
|
|
# the cache should now be empty
|
|
|
|
self.assertEqual(len(obj.fn.cache.cache), 0)
|
|
|
|
|
2017-03-30 15:22:24 +03:00
|
|
|
obj = Cls()
|
|
|
|
|
|
|
|
# set off a deferred which will do a cache lookup
|
|
|
|
d1 = do_lookup()
|
2020-03-24 17:45:33 +03:00
|
|
|
self.assertEqual(current_context(), SENTINEL_CONTEXT)
|
2017-03-30 15:22:24 +03:00
|
|
|
|
|
|
|
return d1
|
2017-03-28 13:19:15 +03:00
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
2023-05-24 17:18:52 +03:00
|
|
|
def test_cache_default_args(self) -> Generator["Deferred[Any]", object, None]:
|
2020-09-04 13:54:56 +03:00
|
|
|
class Cls:
|
2023-05-24 17:18:52 +03:00
|
|
|
def __init__(self) -> None:
|
2017-03-28 13:19:15 +03:00
|
|
|
self.mock = mock.Mock()
|
|
|
|
|
|
|
|
@descriptors.cached()
|
2023-05-24 17:18:52 +03:00
|
|
|
def fn(self, arg1: int, arg2: int = 2, arg3: int = 3) -> str:
|
2017-03-28 13:19:15 +03:00
|
|
|
return self.mock(arg1, arg2, arg3)
|
|
|
|
|
|
|
|
obj = Cls()
|
|
|
|
|
|
|
|
obj.mock.return_value = "fish"
|
|
|
|
r = yield obj.fn(1, 2, 3)
|
|
|
|
self.assertEqual(r, "fish")
|
|
|
|
obj.mock.assert_called_once_with(1, 2, 3)
|
|
|
|
obj.mock.reset_mock()
|
|
|
|
|
|
|
|
# a call with same params shouldn't call the mock again
|
|
|
|
r = yield obj.fn(1, 2)
|
|
|
|
self.assertEqual(r, "fish")
|
|
|
|
obj.mock.assert_not_called()
|
|
|
|
obj.mock.reset_mock()
|
|
|
|
|
|
|
|
# a call with different params should call the mock again
|
|
|
|
obj.mock.return_value = "chips"
|
|
|
|
r = yield obj.fn(2, 3)
|
|
|
|
self.assertEqual(r, "chips")
|
|
|
|
obj.mock.assert_called_once_with(2, 3, 3)
|
|
|
|
obj.mock.reset_mock()
|
|
|
|
|
|
|
|
# the two values should now be cached
|
|
|
|
r = yield obj.fn(1, 2)
|
|
|
|
self.assertEqual(r, "fish")
|
|
|
|
r = yield obj.fn(2, 3)
|
|
|
|
self.assertEqual(r, "chips")
|
|
|
|
obj.mock.assert_not_called()
|
2018-06-11 00:38:50 +03:00
|
|
|
|
2023-05-24 17:18:52 +03:00
|
|
|
def test_cache_iterable(self) -> None:
|
2020-09-04 13:54:56 +03:00
|
|
|
class Cls:
|
2023-05-24 17:18:52 +03:00
|
|
|
def __init__(self) -> None:
|
2019-07-25 17:59:45 +03:00
|
|
|
self.mock = mock.Mock()
|
|
|
|
|
|
|
|
@descriptors.cached(iterable=True)
|
2023-09-20 14:48:55 +03:00
|
|
|
def fn(self, arg1: int, arg2: int) -> Tuple[str, ...]:
|
2019-07-25 17:59:45 +03:00
|
|
|
return self.mock(arg1, arg2)
|
|
|
|
|
|
|
|
obj = Cls()
|
|
|
|
|
2023-09-20 14:48:55 +03:00
|
|
|
obj.mock.return_value = ("spam", "eggs")
|
2019-07-25 17:59:45 +03:00
|
|
|
r = obj.fn(1, 2)
|
2023-09-20 14:48:55 +03:00
|
|
|
self.assertEqual(r.result, ("spam", "eggs"))
|
2019-07-25 17:59:45 +03:00
|
|
|
obj.mock.assert_called_once_with(1, 2)
|
|
|
|
obj.mock.reset_mock()
|
|
|
|
|
|
|
|
# a call with different params should call the mock again
|
2023-09-20 14:48:55 +03:00
|
|
|
obj.mock.return_value = ("chips",)
|
2019-07-25 17:59:45 +03:00
|
|
|
r = obj.fn(1, 3)
|
2023-09-20 14:48:55 +03:00
|
|
|
self.assertEqual(r.result, ("chips",))
|
2019-07-25 17:59:45 +03:00
|
|
|
obj.mock.assert_called_once_with(1, 3)
|
|
|
|
obj.mock.reset_mock()
|
|
|
|
|
|
|
|
# the two values should now be cached
|
|
|
|
self.assertEqual(len(obj.fn.cache.cache), 3)
|
|
|
|
|
|
|
|
r = obj.fn(1, 2)
|
2023-09-20 14:48:55 +03:00
|
|
|
self.assertEqual(r.result, ("spam", "eggs"))
|
2019-07-25 17:59:45 +03:00
|
|
|
r = obj.fn(1, 3)
|
2023-09-20 14:48:55 +03:00
|
|
|
self.assertEqual(r.result, ("chips",))
|
2019-07-25 17:59:45 +03:00
|
|
|
obj.mock.assert_not_called()
|
|
|
|
|
2023-05-24 17:18:52 +03:00
|
|
|
def test_cache_iterable_with_sync_exception(self) -> None:
|
2019-07-25 17:59:45 +03:00
|
|
|
"""If the wrapped function throws synchronously, things should continue to work"""
|
|
|
|
|
2020-09-04 13:54:56 +03:00
|
|
|
class Cls:
|
2019-07-25 17:59:45 +03:00
|
|
|
@descriptors.cached(iterable=True)
|
2023-05-24 17:18:52 +03:00
|
|
|
def fn(self, arg1: int) -> NoReturn:
|
2019-07-25 17:59:45 +03:00
|
|
|
raise SynapseError(100, "mai spoon iz too big!!1")
|
|
|
|
|
|
|
|
obj = Cls()
|
|
|
|
|
|
|
|
# this should fail immediately
|
|
|
|
d = obj.fn(1)
|
|
|
|
self.failureResultOf(d, SynapseError)
|
|
|
|
|
|
|
|
# ... leaving the cache empty
|
|
|
|
self.assertEqual(len(obj.fn.cache.cache), 0)
|
|
|
|
|
|
|
|
# and a second call should result in a second exception
|
|
|
|
d = obj.fn(1)
|
|
|
|
self.failureResultOf(d, SynapseError)
|
|
|
|
|
2023-05-24 17:18:52 +03:00
|
|
|
def test_invalidate_cascade(self) -> None:
|
2020-10-30 14:43:17 +03:00
|
|
|
"""Invalidations should cascade up through cache contexts"""
|
|
|
|
|
|
|
|
class Cls:
|
|
|
|
@cached(cache_context=True)
|
2023-05-24 17:18:52 +03:00
|
|
|
async def func1(self, key: str, cache_context: _CacheContext) -> int:
|
2020-10-30 14:43:17 +03:00
|
|
|
return await self.func2(key, on_invalidate=cache_context.invalidate)
|
|
|
|
|
|
|
|
@cached(cache_context=True)
|
2023-05-24 17:18:52 +03:00
|
|
|
async def func2(self, key: str, cache_context: _CacheContext) -> int:
|
2022-10-25 13:39:25 +03:00
|
|
|
return await self.func3(key, on_invalidate=cache_context.invalidate)
|
2020-10-30 14:43:17 +03:00
|
|
|
|
2022-10-25 13:39:25 +03:00
|
|
|
@cached(cache_context=True)
|
2023-05-24 17:18:52 +03:00
|
|
|
async def func3(self, key: str, cache_context: _CacheContext) -> int:
|
2020-10-30 14:43:17 +03:00
|
|
|
self.invalidate = cache_context.invalidate
|
|
|
|
return 42
|
|
|
|
|
|
|
|
obj = Cls()
|
|
|
|
|
|
|
|
top_invalidate = mock.Mock()
|
|
|
|
r = get_awaitable_result(obj.func1("k1", on_invalidate=top_invalidate))
|
|
|
|
self.assertEqual(r, 42)
|
|
|
|
obj.invalidate()
|
|
|
|
top_invalidate.assert_called_once()
|
|
|
|
|
2023-05-24 17:18:52 +03:00
|
|
|
def test_cancel(self) -> None:
|
2022-03-14 22:04:29 +03:00
|
|
|
"""Test that cancelling a lookup does not cancel other lookups"""
|
|
|
|
complete_lookup: "Deferred[None]" = Deferred()
|
|
|
|
|
|
|
|
class Cls:
|
|
|
|
@cached()
|
2023-05-24 17:18:52 +03:00
|
|
|
async def fn(self, arg1: int) -> str:
|
2022-03-14 22:04:29 +03:00
|
|
|
await complete_lookup
|
|
|
|
return str(arg1)
|
|
|
|
|
|
|
|
obj = Cls()
|
|
|
|
|
|
|
|
d1 = obj.fn(123)
|
|
|
|
d2 = obj.fn(123)
|
|
|
|
self.assertFalse(d1.called)
|
|
|
|
self.assertFalse(d2.called)
|
|
|
|
|
|
|
|
# Cancel `d1`, which is the lookup that caused `fn` to run.
|
|
|
|
d1.cancel()
|
|
|
|
|
|
|
|
# `d2` should complete normally.
|
|
|
|
complete_lookup.callback(None)
|
|
|
|
self.failureResultOf(d1, CancelledError)
|
|
|
|
self.assertEqual(d2.result, "123")
|
|
|
|
|
2023-05-24 17:18:52 +03:00
|
|
|
def test_cancel_logcontexts(self) -> None:
|
2022-03-14 22:04:29 +03:00
|
|
|
"""Test that cancellation does not break logcontexts.
|
|
|
|
|
|
|
|
* The `CancelledError` must be raised with the correct logcontext.
|
|
|
|
* The inner lookup must not resume with a finished logcontext.
|
|
|
|
* The inner lookup must not restore a finished logcontext when done.
|
|
|
|
"""
|
|
|
|
complete_lookup: "Deferred[None]" = Deferred()
|
|
|
|
|
|
|
|
class Cls:
|
|
|
|
inner_context_was_finished = False
|
|
|
|
|
|
|
|
@cached()
|
2023-05-24 17:18:52 +03:00
|
|
|
async def fn(self, arg1: int) -> str:
|
2022-03-14 22:04:29 +03:00
|
|
|
await make_deferred_yieldable(complete_lookup)
|
|
|
|
self.inner_context_was_finished = current_context().finished
|
|
|
|
return str(arg1)
|
|
|
|
|
|
|
|
obj = Cls()
|
|
|
|
|
2023-05-24 17:18:52 +03:00
|
|
|
async def do_lookup() -> None:
|
2022-03-14 22:04:29 +03:00
|
|
|
with LoggingContext("c1") as c1:
|
|
|
|
try:
|
|
|
|
await obj.fn(123)
|
|
|
|
self.fail("No CancelledError thrown")
|
|
|
|
except CancelledError:
|
|
|
|
self.assertEqual(
|
|
|
|
current_context(),
|
|
|
|
c1,
|
|
|
|
"CancelledError was not raised with the correct logcontext",
|
|
|
|
)
|
|
|
|
# suppress the error and succeed
|
|
|
|
|
|
|
|
d = defer.ensureDeferred(do_lookup())
|
|
|
|
d.cancel()
|
|
|
|
|
|
|
|
complete_lookup.callback(None)
|
|
|
|
self.successResultOf(d)
|
|
|
|
self.assertFalse(
|
|
|
|
obj.inner_context_was_finished, "Tried to restart a finished logcontext"
|
|
|
|
)
|
|
|
|
self.assertEqual(current_context(), SENTINEL_CONTEXT)
|
|
|
|
|
2018-06-11 00:38:50 +03:00
|
|
|
|
2020-10-17 00:31:16 +03:00
|
|
|
class CacheDecoratorTestCase(unittest.HomeserverTestCase):
|
|
|
|
"""More tests for @cached
|
|
|
|
|
|
|
|
The following is a set of tests that got lost in a different file for a while.
|
|
|
|
|
|
|
|
There are probably duplicates of the tests in DescriptorTestCase. Ideally the
|
|
|
|
duplicates would be removed and the two sets of classes combined.
|
|
|
|
"""
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
2023-05-24 17:18:52 +03:00
|
|
|
def test_passthrough(self) -> Generator["Deferred[Any]", object, None]:
|
2020-10-17 00:31:16 +03:00
|
|
|
class A:
|
|
|
|
@cached()
|
2023-05-24 17:18:52 +03:00
|
|
|
def func(self, key: str) -> str:
|
2020-10-17 00:31:16 +03:00
|
|
|
return key
|
|
|
|
|
|
|
|
a = A()
|
|
|
|
|
2022-02-28 15:12:29 +03:00
|
|
|
self.assertEqual((yield a.func("foo")), "foo")
|
|
|
|
self.assertEqual((yield a.func("bar")), "bar")
|
2020-10-17 00:31:16 +03:00
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
2023-05-24 17:18:52 +03:00
|
|
|
def test_hit(self) -> Generator["Deferred[Any]", object, None]:
|
2020-10-17 00:31:16 +03:00
|
|
|
callcount = [0]
|
|
|
|
|
|
|
|
class A:
|
|
|
|
@cached()
|
2023-05-24 17:18:52 +03:00
|
|
|
def func(self, key: str) -> str:
|
2020-10-17 00:31:16 +03:00
|
|
|
callcount[0] += 1
|
|
|
|
return key
|
|
|
|
|
|
|
|
a = A()
|
|
|
|
yield a.func("foo")
|
|
|
|
|
2022-02-28 15:12:29 +03:00
|
|
|
self.assertEqual(callcount[0], 1)
|
2020-10-17 00:31:16 +03:00
|
|
|
|
2022-02-28 15:12:29 +03:00
|
|
|
self.assertEqual((yield a.func("foo")), "foo")
|
|
|
|
self.assertEqual(callcount[0], 1)
|
2020-10-17 00:31:16 +03:00
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
2023-05-24 17:18:52 +03:00
|
|
|
def test_invalidate(self) -> Generator["Deferred[Any]", object, None]:
|
2020-10-17 00:31:16 +03:00
|
|
|
callcount = [0]
|
|
|
|
|
|
|
|
class A:
|
|
|
|
@cached()
|
2023-05-24 17:18:52 +03:00
|
|
|
def func(self, key: str) -> str:
|
2020-10-17 00:31:16 +03:00
|
|
|
callcount[0] += 1
|
|
|
|
return key
|
|
|
|
|
|
|
|
a = A()
|
|
|
|
yield a.func("foo")
|
|
|
|
|
2022-02-28 15:12:29 +03:00
|
|
|
self.assertEqual(callcount[0], 1)
|
2020-10-17 00:31:16 +03:00
|
|
|
|
|
|
|
a.func.invalidate(("foo",))
|
|
|
|
|
|
|
|
yield a.func("foo")
|
|
|
|
|
2022-02-28 15:12:29 +03:00
|
|
|
self.assertEqual(callcount[0], 2)
|
2020-10-17 00:31:16 +03:00
|
|
|
|
2023-05-24 17:18:52 +03:00
|
|
|
def test_invalidate_missing(self) -> None:
|
2020-10-17 00:31:16 +03:00
|
|
|
class A:
|
|
|
|
@cached()
|
2023-05-24 17:18:52 +03:00
|
|
|
def func(self, key: str) -> str:
|
2020-10-17 00:31:16 +03:00
|
|
|
return key
|
|
|
|
|
|
|
|
A().func.invalidate(("what",))
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
2023-05-24 17:18:52 +03:00
|
|
|
def test_max_entries(self) -> Generator["Deferred[Any]", object, None]:
|
2020-10-17 00:31:16 +03:00
|
|
|
callcount = [0]
|
|
|
|
|
|
|
|
class A:
|
|
|
|
@cached(max_entries=10)
|
2023-05-24 17:18:52 +03:00
|
|
|
def func(self, key: int) -> int:
|
2020-10-17 00:31:16 +03:00
|
|
|
callcount[0] += 1
|
|
|
|
return key
|
|
|
|
|
|
|
|
a = A()
|
|
|
|
|
2023-09-08 18:24:36 +03:00
|
|
|
for k in range(12):
|
2020-10-17 00:31:16 +03:00
|
|
|
yield a.func(k)
|
|
|
|
|
2022-02-28 15:12:29 +03:00
|
|
|
self.assertEqual(callcount[0], 12)
|
2020-10-17 00:31:16 +03:00
|
|
|
|
|
|
|
# There must have been at least 2 evictions, meaning if we calculate
|
|
|
|
# all 12 values again, we must get called at least 2 more times
|
2023-09-08 18:24:36 +03:00
|
|
|
for k in range(12):
|
2020-10-17 00:31:16 +03:00
|
|
|
yield a.func(k)
|
|
|
|
|
|
|
|
self.assertTrue(
|
|
|
|
callcount[0] >= 14, msg="Expected callcount >= 14, got %d" % (callcount[0])
|
|
|
|
)
|
|
|
|
|
2023-05-24 17:18:52 +03:00
|
|
|
def test_prefill(self) -> None:
|
2020-10-17 00:31:16 +03:00
|
|
|
callcount = [0]
|
|
|
|
|
|
|
|
d = defer.succeed(123)
|
|
|
|
|
|
|
|
class A:
|
|
|
|
@cached()
|
2023-05-24 17:18:52 +03:00
|
|
|
def func(self, key: str) -> "Deferred[int]":
|
2020-10-17 00:31:16 +03:00
|
|
|
callcount[0] += 1
|
|
|
|
return d
|
|
|
|
|
|
|
|
a = A()
|
|
|
|
|
2020-10-16 14:34:55 +03:00
|
|
|
a.func.prefill(("foo",), 456)
|
2020-10-17 00:31:16 +03:00
|
|
|
|
2022-02-28 15:12:29 +03:00
|
|
|
self.assertEqual(a.func("foo").result, 456)
|
|
|
|
self.assertEqual(callcount[0], 0)
|
2020-10-17 00:31:16 +03:00
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
2023-05-24 17:18:52 +03:00
|
|
|
def test_invalidate_context(self) -> Generator["Deferred[Any]", object, None]:
|
2020-10-17 00:31:16 +03:00
|
|
|
callcount = [0]
|
|
|
|
callcount2 = [0]
|
|
|
|
|
|
|
|
class A:
|
|
|
|
@cached()
|
2023-05-24 17:18:52 +03:00
|
|
|
def func(self, key: str) -> str:
|
2020-10-17 00:31:16 +03:00
|
|
|
callcount[0] += 1
|
|
|
|
return key
|
|
|
|
|
|
|
|
@cached(cache_context=True)
|
2023-05-24 17:18:52 +03:00
|
|
|
def func2(self, key: str, cache_context: _CacheContext) -> "Deferred[str]":
|
2020-10-17 00:31:16 +03:00
|
|
|
callcount2[0] += 1
|
|
|
|
return self.func(key, on_invalidate=cache_context.invalidate)
|
|
|
|
|
|
|
|
a = A()
|
|
|
|
yield a.func2("foo")
|
|
|
|
|
2022-02-28 15:12:29 +03:00
|
|
|
self.assertEqual(callcount[0], 1)
|
|
|
|
self.assertEqual(callcount2[0], 1)
|
2020-10-17 00:31:16 +03:00
|
|
|
|
|
|
|
a.func.invalidate(("foo",))
|
|
|
|
yield a.func("foo")
|
|
|
|
|
2022-02-28 15:12:29 +03:00
|
|
|
self.assertEqual(callcount[0], 2)
|
|
|
|
self.assertEqual(callcount2[0], 1)
|
2020-10-17 00:31:16 +03:00
|
|
|
|
|
|
|
yield a.func2("foo")
|
|
|
|
|
2022-02-28 15:12:29 +03:00
|
|
|
self.assertEqual(callcount[0], 2)
|
|
|
|
self.assertEqual(callcount2[0], 2)
|
2020-10-17 00:31:16 +03:00
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
2023-05-24 17:18:52 +03:00
|
|
|
def test_eviction_context(self) -> Generator["Deferred[Any]", object, None]:
|
2020-10-17 00:31:16 +03:00
|
|
|
callcount = [0]
|
|
|
|
callcount2 = [0]
|
|
|
|
|
|
|
|
class A:
|
|
|
|
@cached(max_entries=2)
|
2023-05-24 17:18:52 +03:00
|
|
|
def func(self, key: str) -> str:
|
2020-10-17 00:31:16 +03:00
|
|
|
callcount[0] += 1
|
|
|
|
return key
|
|
|
|
|
|
|
|
@cached(cache_context=True)
|
2023-05-24 17:18:52 +03:00
|
|
|
def func2(self, key: str, cache_context: _CacheContext) -> "Deferred[str]":
|
2020-10-17 00:31:16 +03:00
|
|
|
callcount2[0] += 1
|
|
|
|
return self.func(key, on_invalidate=cache_context.invalidate)
|
|
|
|
|
|
|
|
a = A()
|
|
|
|
yield a.func2("foo")
|
|
|
|
yield a.func2("foo2")
|
|
|
|
|
2022-02-28 15:12:29 +03:00
|
|
|
self.assertEqual(callcount[0], 2)
|
|
|
|
self.assertEqual(callcount2[0], 2)
|
2020-10-17 00:31:16 +03:00
|
|
|
|
|
|
|
yield a.func2("foo")
|
2022-02-28 15:12:29 +03:00
|
|
|
self.assertEqual(callcount[0], 2)
|
|
|
|
self.assertEqual(callcount2[0], 2)
|
2020-10-17 00:31:16 +03:00
|
|
|
|
|
|
|
yield a.func("foo3")
|
|
|
|
|
2022-02-28 15:12:29 +03:00
|
|
|
self.assertEqual(callcount[0], 3)
|
|
|
|
self.assertEqual(callcount2[0], 2)
|
2020-10-17 00:31:16 +03:00
|
|
|
|
|
|
|
yield a.func2("foo")
|
|
|
|
|
2022-02-28 15:12:29 +03:00
|
|
|
self.assertEqual(callcount[0], 4)
|
|
|
|
self.assertEqual(callcount2[0], 3)
|
2020-10-17 00:31:16 +03:00
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
2023-05-24 17:18:52 +03:00
|
|
|
def test_double_get(self) -> Generator["Deferred[Any]", object, None]:
|
2020-10-17 00:31:16 +03:00
|
|
|
callcount = [0]
|
|
|
|
callcount2 = [0]
|
|
|
|
|
|
|
|
class A:
|
|
|
|
@cached()
|
2023-05-24 17:18:52 +03:00
|
|
|
def func(self, key: str) -> str:
|
2020-10-17 00:31:16 +03:00
|
|
|
callcount[0] += 1
|
|
|
|
return key
|
|
|
|
|
|
|
|
@cached(cache_context=True)
|
2023-05-24 17:18:52 +03:00
|
|
|
def func2(self, key: str, cache_context: _CacheContext) -> "Deferred[str]":
|
2020-10-17 00:31:16 +03:00
|
|
|
callcount2[0] += 1
|
|
|
|
return self.func(key, on_invalidate=cache_context.invalidate)
|
|
|
|
|
|
|
|
a = A()
|
|
|
|
a.func2.cache.cache = mock.Mock(wraps=a.func2.cache.cache)
|
|
|
|
|
|
|
|
yield a.func2("foo")
|
|
|
|
|
2022-02-28 15:12:29 +03:00
|
|
|
self.assertEqual(callcount[0], 1)
|
|
|
|
self.assertEqual(callcount2[0], 1)
|
2020-10-17 00:31:16 +03:00
|
|
|
|
|
|
|
a.func2.invalidate(("foo",))
|
2022-02-28 15:12:29 +03:00
|
|
|
self.assertEqual(a.func2.cache.cache.del_multi.call_count, 1)
|
2020-10-17 00:31:16 +03:00
|
|
|
|
|
|
|
yield a.func2("foo")
|
|
|
|
a.func2.invalidate(("foo",))
|
2022-02-28 15:12:29 +03:00
|
|
|
self.assertEqual(a.func2.cache.cache.del_multi.call_count, 2)
|
2020-10-17 00:31:16 +03:00
|
|
|
|
2022-02-28 15:12:29 +03:00
|
|
|
self.assertEqual(callcount[0], 1)
|
|
|
|
self.assertEqual(callcount2[0], 2)
|
2020-10-17 00:31:16 +03:00
|
|
|
|
|
|
|
a.func.invalidate(("foo",))
|
2022-02-28 15:12:29 +03:00
|
|
|
self.assertEqual(a.func2.cache.cache.del_multi.call_count, 3)
|
2020-10-17 00:31:16 +03:00
|
|
|
yield a.func("foo")
|
|
|
|
|
2022-02-28 15:12:29 +03:00
|
|
|
self.assertEqual(callcount[0], 2)
|
|
|
|
self.assertEqual(callcount2[0], 2)
|
2020-10-17 00:31:16 +03:00
|
|
|
|
|
|
|
yield a.func2("foo")
|
|
|
|
|
2022-02-28 15:12:29 +03:00
|
|
|
self.assertEqual(callcount[0], 2)
|
|
|
|
self.assertEqual(callcount2[0], 3)
|
2020-10-17 00:31:16 +03:00
|
|
|
|
|
|
|
|
2018-06-11 00:38:50 +03:00
|
|
|
class CachedListDescriptorTestCase(unittest.TestCase):
|
|
|
|
@defer.inlineCallbacks
|
2023-05-24 17:18:52 +03:00
|
|
|
def test_cache(self) -> Generator["Deferred[Any]", object, None]:
|
2020-09-04 13:54:56 +03:00
|
|
|
class Cls:
|
2023-05-24 17:18:52 +03:00
|
|
|
def __init__(self) -> None:
|
2018-06-11 00:38:50 +03:00
|
|
|
self.mock = mock.Mock()
|
|
|
|
|
|
|
|
@descriptors.cached()
|
2023-05-24 17:18:52 +03:00
|
|
|
def fn(self, arg1: int, arg2: int) -> None:
|
2018-06-11 00:38:50 +03:00
|
|
|
pass
|
|
|
|
|
2022-03-09 21:07:41 +03:00
|
|
|
@descriptors.cachedList(cached_method_name="fn", list_name="args1")
|
2023-09-20 14:48:55 +03:00
|
|
|
async def list_fn(
|
|
|
|
self, args1: Iterable[int], arg2: int
|
|
|
|
) -> Mapping[int, str]:
|
2022-11-23 01:35:54 +03:00
|
|
|
context = current_context()
|
|
|
|
assert isinstance(context, LoggingContext)
|
|
|
|
assert context.name == "c1"
|
2018-06-11 00:38:50 +03:00
|
|
|
# we want this to behave like an asynchronous function
|
2020-08-19 14:09:07 +03:00
|
|
|
await run_on_reactor()
|
2022-11-23 01:35:54 +03:00
|
|
|
context = current_context()
|
|
|
|
assert isinstance(context, LoggingContext)
|
|
|
|
assert context.name == "c1"
|
2019-07-23 16:00:55 +03:00
|
|
|
return self.mock(args1, arg2)
|
2018-06-11 00:38:50 +03:00
|
|
|
|
2021-04-08 15:01:14 +03:00
|
|
|
with LoggingContext("c1") as c1:
|
2018-06-11 00:38:50 +03:00
|
|
|
obj = Cls()
|
|
|
|
obj.mock.return_value = {10: "fish", 20: "chips"}
|
2021-05-14 13:12:36 +03:00
|
|
|
|
|
|
|
# start the lookup off
|
2018-06-11 00:38:50 +03:00
|
|
|
d1 = obj.list_fn([10, 20], 2)
|
2020-03-24 17:45:33 +03:00
|
|
|
self.assertEqual(current_context(), SENTINEL_CONTEXT)
|
2018-06-11 00:38:50 +03:00
|
|
|
r = yield d1
|
2020-03-24 17:45:33 +03:00
|
|
|
self.assertEqual(current_context(), c1)
|
2022-03-01 12:51:38 +03:00
|
|
|
obj.mock.assert_called_once_with({10, 20}, 2)
|
2018-06-11 00:38:50 +03:00
|
|
|
self.assertEqual(r, {10: "fish", 20: "chips"})
|
|
|
|
obj.mock.reset_mock()
|
|
|
|
|
|
|
|
# a call with different params should call the mock again
|
|
|
|
obj.mock.return_value = {30: "peas"}
|
|
|
|
r = yield obj.list_fn([20, 30], 2)
|
2022-03-01 12:51:38 +03:00
|
|
|
obj.mock.assert_called_once_with({30}, 2)
|
2018-06-11 00:38:50 +03:00
|
|
|
self.assertEqual(r, {20: "chips", 30: "peas"})
|
|
|
|
obj.mock.reset_mock()
|
|
|
|
|
|
|
|
# all the values should now be cached
|
|
|
|
r = yield obj.fn(10, 2)
|
|
|
|
self.assertEqual(r, "fish")
|
|
|
|
r = yield obj.fn(20, 2)
|
|
|
|
self.assertEqual(r, "chips")
|
|
|
|
r = yield obj.fn(30, 2)
|
|
|
|
self.assertEqual(r, "peas")
|
|
|
|
r = yield obj.list_fn([10, 20, 30], 2)
|
|
|
|
obj.mock.assert_not_called()
|
|
|
|
self.assertEqual(r, {10: "fish", 20: "chips", 30: "peas"})
|
2018-07-27 18:17:17 +03:00
|
|
|
|
2021-05-14 13:12:36 +03:00
|
|
|
# we should also be able to use a (single-use) iterable, and should
|
|
|
|
# deduplicate the keys
|
|
|
|
obj.mock.reset_mock()
|
|
|
|
obj.mock.return_value = {40: "gravy"}
|
|
|
|
iterable = (x for x in [10, 40, 40])
|
|
|
|
r = yield obj.list_fn(iterable, 2)
|
2022-03-01 12:51:38 +03:00
|
|
|
obj.mock.assert_called_once_with({40}, 2)
|
2021-05-14 13:12:36 +03:00
|
|
|
self.assertEqual(r, {10: "fish", 40: "gravy"})
|
|
|
|
|
2023-05-24 17:18:52 +03:00
|
|
|
def test_concurrent_lookups(self) -> None:
|
2021-11-04 17:45:34 +03:00
|
|
|
"""All concurrent lookups should get the same result"""
|
|
|
|
|
|
|
|
class Cls:
|
2023-05-24 17:18:52 +03:00
|
|
|
def __init__(self) -> None:
|
2021-11-04 17:45:34 +03:00
|
|
|
self.mock = mock.Mock()
|
|
|
|
|
|
|
|
@descriptors.cached()
|
2023-05-24 17:18:52 +03:00
|
|
|
def fn(self, arg1: int) -> None:
|
2021-11-04 17:45:34 +03:00
|
|
|
pass
|
|
|
|
|
2022-03-09 21:07:41 +03:00
|
|
|
@descriptors.cachedList(cached_method_name="fn", list_name="args1")
|
2023-09-20 14:48:55 +03:00
|
|
|
def list_fn(self, args1: List[int]) -> "Deferred[Mapping[int, str]]":
|
2021-11-04 17:45:34 +03:00
|
|
|
return self.mock(args1)
|
|
|
|
|
|
|
|
obj = Cls()
|
2023-09-20 14:48:55 +03:00
|
|
|
deferred_result: "Deferred[Mapping[int, str]]" = Deferred()
|
2021-11-04 17:45:34 +03:00
|
|
|
obj.mock.return_value = deferred_result
|
|
|
|
|
|
|
|
# start off several concurrent lookups of the same key
|
|
|
|
d1 = obj.list_fn([10])
|
|
|
|
d2 = obj.list_fn([10])
|
|
|
|
d3 = obj.list_fn([10])
|
|
|
|
|
|
|
|
# the mock should have been called exactly once
|
2022-03-01 12:51:38 +03:00
|
|
|
obj.mock.assert_called_once_with({10})
|
2021-11-04 17:45:34 +03:00
|
|
|
obj.mock.reset_mock()
|
|
|
|
|
|
|
|
# ... and none of the calls should yet be complete
|
|
|
|
self.assertFalse(d1.called)
|
|
|
|
self.assertFalse(d2.called)
|
|
|
|
self.assertFalse(d3.called)
|
|
|
|
|
|
|
|
# complete the lookup. @cachedList functions need to complete with a map
|
|
|
|
# of input->result
|
|
|
|
deferred_result.callback({10: "peas"})
|
|
|
|
|
|
|
|
# ... which should give the right result to all the callers
|
|
|
|
self.assertEqual(self.successResultOf(d1), {10: "peas"})
|
|
|
|
self.assertEqual(self.successResultOf(d2), {10: "peas"})
|
|
|
|
self.assertEqual(self.successResultOf(d3), {10: "peas"})
|
|
|
|
|
2018-07-27 18:17:17 +03:00
|
|
|
@defer.inlineCallbacks
|
2023-05-24 17:18:52 +03:00
|
|
|
def test_invalidate(self) -> Generator["Deferred[Any]", object, None]:
|
2018-07-27 18:17:17 +03:00
|
|
|
"""Make sure that invalidation callbacks are called."""
|
2018-08-10 16:54:09 +03:00
|
|
|
|
2020-09-04 13:54:56 +03:00
|
|
|
class Cls:
|
2023-05-24 17:18:52 +03:00
|
|
|
def __init__(self) -> None:
|
2018-07-27 18:17:17 +03:00
|
|
|
self.mock = mock.Mock()
|
|
|
|
|
|
|
|
@descriptors.cached()
|
2023-05-24 17:18:52 +03:00
|
|
|
def fn(self, arg1: int, arg2: int) -> None:
|
2018-07-27 18:17:17 +03:00
|
|
|
pass
|
|
|
|
|
2022-03-09 21:07:41 +03:00
|
|
|
@descriptors.cachedList(cached_method_name="fn", list_name="args1")
|
2023-09-20 14:48:55 +03:00
|
|
|
async def list_fn(self, args1: List[int], arg2: int) -> Mapping[int, str]:
|
2018-07-27 18:17:17 +03:00
|
|
|
# we want this to behave like an asynchronous function
|
2020-08-19 14:09:07 +03:00
|
|
|
await run_on_reactor()
|
2019-07-23 16:00:55 +03:00
|
|
|
return self.mock(args1, arg2)
|
2018-07-27 18:17:17 +03:00
|
|
|
|
|
|
|
obj = Cls()
|
|
|
|
invalidate0 = mock.Mock()
|
|
|
|
invalidate1 = mock.Mock()
|
|
|
|
|
|
|
|
# cache miss
|
|
|
|
obj.mock.return_value = {10: "fish", 20: "chips"}
|
|
|
|
r1 = yield obj.list_fn([10, 20], 2, on_invalidate=invalidate0)
|
2022-03-01 12:51:38 +03:00
|
|
|
obj.mock.assert_called_once_with({10, 20}, 2)
|
2018-07-27 18:17:17 +03:00
|
|
|
self.assertEqual(r1, {10: "fish", 20: "chips"})
|
|
|
|
obj.mock.reset_mock()
|
|
|
|
|
|
|
|
# cache hit
|
|
|
|
r2 = yield obj.list_fn([10, 20], 2, on_invalidate=invalidate1)
|
|
|
|
obj.mock.assert_not_called()
|
|
|
|
self.assertEqual(r2, {10: "fish", 20: "chips"})
|
|
|
|
|
|
|
|
invalidate0.assert_not_called()
|
|
|
|
invalidate1.assert_not_called()
|
|
|
|
|
|
|
|
# now if we invalidate the keys, both invalidations should get called
|
|
|
|
obj.fn.invalidate((10, 2))
|
|
|
|
invalidate0.assert_called_once()
|
|
|
|
invalidate1.assert_called_once()
|
2022-03-14 22:04:29 +03:00
|
|
|
|
2023-05-24 17:18:52 +03:00
|
|
|
def test_cancel(self) -> None:
|
2022-03-14 22:04:29 +03:00
|
|
|
"""Test that cancelling a lookup does not cancel other lookups"""
|
|
|
|
complete_lookup: "Deferred[None]" = Deferred()
|
|
|
|
|
|
|
|
class Cls:
|
|
|
|
@cached()
|
2023-05-24 17:18:52 +03:00
|
|
|
def fn(self, arg1: int) -> None:
|
2022-03-14 22:04:29 +03:00
|
|
|
pass
|
|
|
|
|
|
|
|
@cachedList(cached_method_name="fn", list_name="args")
|
2023-09-20 14:48:55 +03:00
|
|
|
async def list_fn(self, args: List[int]) -> Mapping[int, str]:
|
2022-03-14 22:04:29 +03:00
|
|
|
await complete_lookup
|
|
|
|
return {arg: str(arg) for arg in args}
|
|
|
|
|
|
|
|
obj = Cls()
|
|
|
|
|
|
|
|
d1 = obj.list_fn([123, 456])
|
|
|
|
d2 = obj.list_fn([123, 456, 789])
|
|
|
|
self.assertFalse(d1.called)
|
|
|
|
self.assertFalse(d2.called)
|
|
|
|
|
|
|
|
d1.cancel()
|
|
|
|
|
|
|
|
# `d2` should complete normally.
|
|
|
|
complete_lookup.callback(None)
|
|
|
|
self.failureResultOf(d1, CancelledError)
|
|
|
|
self.assertEqual(d2.result, {123: "123", 456: "456", 789: "789"})
|
|
|
|
|
2023-05-24 17:18:52 +03:00
|
|
|
def test_cancel_logcontexts(self) -> None:
|
2022-03-14 22:04:29 +03:00
|
|
|
"""Test that cancellation does not break logcontexts.
|
|
|
|
|
|
|
|
* The `CancelledError` must be raised with the correct logcontext.
|
|
|
|
* The inner lookup must not resume with a finished logcontext.
|
|
|
|
* The inner lookup must not restore a finished logcontext when done.
|
|
|
|
"""
|
|
|
|
complete_lookup: "Deferred[None]" = Deferred()
|
|
|
|
|
|
|
|
class Cls:
|
|
|
|
inner_context_was_finished = False
|
|
|
|
|
|
|
|
@cached()
|
2023-05-24 17:18:52 +03:00
|
|
|
def fn(self, arg1: int) -> None:
|
2022-03-14 22:04:29 +03:00
|
|
|
pass
|
|
|
|
|
|
|
|
@cachedList(cached_method_name="fn", list_name="args")
|
2023-09-20 14:48:55 +03:00
|
|
|
async def list_fn(self, args: List[int]) -> Mapping[int, str]:
|
2022-03-14 22:04:29 +03:00
|
|
|
await make_deferred_yieldable(complete_lookup)
|
|
|
|
self.inner_context_was_finished = current_context().finished
|
|
|
|
return {arg: str(arg) for arg in args}
|
|
|
|
|
|
|
|
obj = Cls()
|
|
|
|
|
2023-05-24 17:18:52 +03:00
|
|
|
async def do_lookup() -> None:
|
2022-03-14 22:04:29 +03:00
|
|
|
with LoggingContext("c1") as c1:
|
|
|
|
try:
|
|
|
|
await obj.list_fn([123])
|
|
|
|
self.fail("No CancelledError thrown")
|
|
|
|
except CancelledError:
|
|
|
|
self.assertEqual(
|
|
|
|
current_context(),
|
|
|
|
c1,
|
|
|
|
"CancelledError was not raised with the correct logcontext",
|
|
|
|
)
|
|
|
|
# suppress the error and succeed
|
|
|
|
|
|
|
|
d = defer.ensureDeferred(do_lookup())
|
|
|
|
d.cancel()
|
|
|
|
|
|
|
|
complete_lookup.callback(None)
|
|
|
|
self.successResultOf(d)
|
|
|
|
self.assertFalse(
|
|
|
|
obj.inner_context_was_finished, "Tried to restart a finished logcontext"
|
|
|
|
)
|
|
|
|
self.assertEqual(current_context(), SENTINEL_CONTEXT)
|
2022-09-27 23:55:43 +03:00
|
|
|
|
2023-05-24 17:18:52 +03:00
|
|
|
def test_num_args_mismatch(self) -> None:
|
2022-09-27 23:55:43 +03:00
|
|
|
"""
|
|
|
|
Make sure someone does not accidentally use @cachedList on a method with
|
|
|
|
a mismatch in the number args to the underlying single cache method.
|
|
|
|
"""
|
|
|
|
|
|
|
|
class Cls:
|
|
|
|
@descriptors.cached(tree=True)
|
2023-05-24 17:18:52 +03:00
|
|
|
def fn(self, room_id: str, event_id: str) -> None:
|
2022-09-27 23:55:43 +03:00
|
|
|
pass
|
|
|
|
|
|
|
|
# This is wrong ❌. `@cachedList` expects to be given the same number
|
|
|
|
# of arguments as the underlying cached function, just with one of
|
|
|
|
# the arguments being an iterable
|
|
|
|
@descriptors.cachedList(cached_method_name="fn", list_name="keys")
|
2023-05-24 17:18:52 +03:00
|
|
|
def list_fn(self, keys: Iterable[Tuple[str, str]]) -> None:
|
2022-09-27 23:55:43 +03:00
|
|
|
pass
|
|
|
|
|
|
|
|
# Corrected syntax ✅
|
|
|
|
#
|
|
|
|
# @cachedList(cached_method_name="fn", list_name="event_ids")
|
|
|
|
# async def list_fn(
|
|
|
|
# self, room_id: str, event_ids: Collection[str],
|
|
|
|
# )
|
|
|
|
|
|
|
|
obj = Cls()
|
|
|
|
|
|
|
|
# Make sure this raises an error about the arg mismatch
|
2022-10-19 22:38:24 +03:00
|
|
|
with self.assertRaises(TypeError):
|
2022-09-27 23:55:43 +03:00
|
|
|
obj.list_fn([("foo", "bar")])
|