2022-04-26 03:27:11 -06:00
|
|
|
# Copyright 2014-2022 The Matrix.org Foundation C.I.C.
|
2021-06-09 12:39:51 -06:00
|
|
|
# Copyright 2020 Sorunome
|
2014-08-12 08:10:52 -06:00
|
|
|
#
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
2014-08-12 20:14:34 -06:00
|
|
|
|
2014-08-12 08:10:52 -06:00
|
|
|
"""Contains handlers for federation events."""
|
2018-04-17 15:11:19 -06:00
|
|
|
|
2022-04-26 03:27:11 -06:00
|
|
|
import enum
|
|
|
|
import itertools
|
2018-04-17 15:11:19 -06:00
|
|
|
import logging
|
2022-04-26 03:27:11 -06:00
|
|
|
from enum import Enum
|
2020-06-16 06:51:47 -06:00
|
|
|
from http import HTTPStatus
|
2023-02-03 08:39:59 -07:00
|
|
|
from typing import (
|
|
|
|
TYPE_CHECKING,
|
|
|
|
AbstractSet,
|
|
|
|
Dict,
|
|
|
|
Iterable,
|
|
|
|
List,
|
|
|
|
Optional,
|
|
|
|
Set,
|
|
|
|
Tuple,
|
|
|
|
Union,
|
|
|
|
)
|
2018-04-17 15:11:19 -06:00
|
|
|
|
2022-04-26 03:27:11 -06:00
|
|
|
import attr
|
2022-08-17 03:33:19 -06:00
|
|
|
from prometheus_client import Histogram
|
2018-07-09 00:09:20 -06:00
|
|
|
from signedjson.key import decode_verify_key_bytes
|
|
|
|
from signedjson.sign import verify_signed_json
|
2016-02-23 08:11:25 -07:00
|
|
|
from unpaddedbase64 import decode_base64
|
2014-08-12 08:10:52 -06:00
|
|
|
|
2019-10-18 11:43:36 -06:00
|
|
|
from synapse import event_auth
|
2022-09-28 14:26:16 -06:00
|
|
|
from synapse.api.constants import MAX_DEPTH, EventContentFields, EventTypes, Membership
|
2014-11-26 09:06:20 -07:00
|
|
|
from synapse.api.errors import (
|
2018-07-09 00:09:20 -06:00
|
|
|
AuthError,
|
|
|
|
CodeMessageException,
|
2019-06-12 03:31:37 -06:00
|
|
|
Codes,
|
2018-01-22 11:11:18 -07:00
|
|
|
FederationDeniedError,
|
2022-05-31 08:50:29 -06:00
|
|
|
FederationError,
|
2022-10-14 23:36:49 -06:00
|
|
|
FederationPullAttemptBackoffError,
|
2020-05-22 04:39:20 -06:00
|
|
|
HttpResponseException,
|
2020-07-16 08:17:31 -06:00
|
|
|
NotFoundError,
|
2023-02-10 16:31:05 -07:00
|
|
|
PartialStateConflictError,
|
2019-06-03 02:56:45 -06:00
|
|
|
RequestSendFailed,
|
2018-07-09 00:09:20 -06:00
|
|
|
SynapseError,
|
2014-11-26 09:06:20 -07:00
|
|
|
)
|
2021-10-19 03:24:09 -06:00
|
|
|
from synapse.api.room_versions import KNOWN_ROOM_VERSIONS, RoomVersion
|
2019-01-23 13:05:44 -07:00
|
|
|
from synapse.crypto.event_signing import compute_event_signature
|
2021-10-19 03:24:09 -06:00
|
|
|
from synapse.event_auth import validate_event_for_room_version
|
2020-01-28 07:18:29 -07:00
|
|
|
from synapse.events import EventBase
|
2023-02-09 14:05:02 -07:00
|
|
|
from synapse.events.snapshot import EventContext, UnpersistedEventContextBase
|
2018-07-09 00:09:20 -06:00
|
|
|
from synapse.events.validator import EventValidator
|
2021-08-26 11:34:57 -06:00
|
|
|
from synapse.federation.federation_client import InvalidResponseError
|
2023-08-23 02:23:41 -06:00
|
|
|
from synapse.handlers.pagination import PURGE_PAGINATION_LOCK_NAME
|
2020-11-19 03:05:33 -07:00
|
|
|
from synapse.http.servlet import assert_params_in_dict
|
2022-03-09 04:00:48 -07:00
|
|
|
from synapse.logging.context import nested_logging_context
|
2022-08-16 11:39:40 -06:00
|
|
|
from synapse.logging.opentracing import SynapseTags, set_tag, tag_args, trace
|
2022-02-22 06:33:22 -07:00
|
|
|
from synapse.metrics.background_process_metrics import run_as_background_process
|
2022-06-13 12:16:16 -06:00
|
|
|
from synapse.module_api import NOT_SPAM
|
2018-07-26 04:44:22 -06:00
|
|
|
from synapse.replication.http.federation import (
|
2018-08-09 03:29:48 -06:00
|
|
|
ReplicationCleanRoomRestServlet,
|
2020-11-13 09:24:04 -07:00
|
|
|
ReplicationStoreRoomOnOutlierMembershipRestServlet,
|
2018-07-26 04:44:22 -06:00
|
|
|
)
|
2020-08-05 14:38:57 -06:00
|
|
|
from synapse.storage.databases.main.events_worker import EventRedactBehaviour
|
2023-01-26 10:31:58 -07:00
|
|
|
from synapse.types import JsonDict, StrCollection, get_domain_from_id
|
2022-12-12 09:19:30 -07:00
|
|
|
from synapse.types.state import StateFilter
|
2021-08-26 14:41:44 -06:00
|
|
|
from synapse.util.async_helpers import Linearizer
|
2015-05-12 03:35:45 -06:00
|
|
|
from synapse.util.retryutils import NotRetryingDestination
|
2018-07-16 04:38:45 -06:00
|
|
|
from synapse.visibility import filter_events_for_server
|
2015-05-12 03:35:45 -06:00
|
|
|
|
2020-09-28 08:20:02 -06:00
|
|
|
if TYPE_CHECKING:
|
|
|
|
from synapse.server import HomeServer
|
|
|
|
|
2014-08-12 08:10:52 -06:00
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
2022-08-17 03:33:19 -06:00
|
|
|
# Added to debug performance and track progress on optimizations
|
|
|
|
backfill_processing_before_timer = Histogram(
|
|
|
|
"synapse_federation_backfill_processing_before_time_seconds",
|
|
|
|
"sec",
|
|
|
|
[],
|
|
|
|
buckets=(
|
2022-08-23 01:47:30 -06:00
|
|
|
0.1,
|
|
|
|
0.5,
|
2022-08-17 03:33:19 -06:00
|
|
|
1.0,
|
2022-08-23 01:47:30 -06:00
|
|
|
2.5,
|
2022-08-17 03:33:19 -06:00
|
|
|
5.0,
|
2022-08-23 01:47:30 -06:00
|
|
|
7.5,
|
2022-08-17 03:33:19 -06:00
|
|
|
10.0,
|
2022-08-23 01:47:30 -06:00
|
|
|
15.0,
|
2022-08-17 03:33:19 -06:00
|
|
|
20.0,
|
|
|
|
30.0,
|
|
|
|
40.0,
|
|
|
|
60.0,
|
|
|
|
80.0,
|
|
|
|
"+Inf",
|
|
|
|
),
|
|
|
|
)
|
|
|
|
|
2019-12-05 08:02:35 -07:00
|
|
|
|
2023-06-16 13:12:24 -06:00
|
|
|
# TODO: We can refactor this away now that there is only one backfill point again
|
2022-04-26 03:27:11 -06:00
|
|
|
class _BackfillPointType(Enum):
|
|
|
|
# a regular backwards extremity (ie, an event which we don't yet have, but which
|
|
|
|
# is referred to by other events in the DAG)
|
|
|
|
BACKWARDS_EXTREMITY = enum.auto()
|
|
|
|
|
|
|
|
|
|
|
|
@attr.s(slots=True, auto_attribs=True, frozen=True)
|
|
|
|
class _BackfillPoint:
|
|
|
|
"""A potential point we might backfill from"""
|
|
|
|
|
|
|
|
event_id: str
|
|
|
|
depth: int
|
|
|
|
type: _BackfillPointType
|
|
|
|
|
|
|
|
|
2021-10-08 05:44:43 -06:00
|
|
|
class FederationHandler:
|
2021-08-26 14:41:44 -06:00
|
|
|
"""Handles general incoming federation requests
|
|
|
|
|
|
|
|
Incoming events are *not* handled here, for which see FederationEventHandler.
|
2014-08-26 12:49:42 -06:00
|
|
|
"""
|
2014-08-12 08:10:52 -06:00
|
|
|
|
2020-09-28 08:20:02 -06:00
|
|
|
def __init__(self, hs: "HomeServer"):
|
2015-11-05 09:43:19 -07:00
|
|
|
self.hs = hs
|
|
|
|
|
2022-08-17 03:33:19 -06:00
|
|
|
self.clock = hs.get_clock()
|
2022-02-23 04:04:02 -07:00
|
|
|
self.store = hs.get_datastores().main
|
2022-05-31 06:17:50 -06:00
|
|
|
self._storage_controllers = hs.get_storage_controllers()
|
|
|
|
self._state_storage_controller = self._storage_controllers.state
|
2018-07-31 08:44:05 -06:00
|
|
|
self.federation_client = hs.get_federation_client()
|
2014-08-26 12:49:42 -06:00
|
|
|
self.state_handler = hs.get_state_handler()
|
|
|
|
self.server_name = hs.hostname
|
2014-11-14 09:45:39 -07:00
|
|
|
self.keyring = hs.get_keyring()
|
2017-06-30 09:20:30 -06:00
|
|
|
self.is_mine_id = hs.is_mine_id
|
2023-05-05 08:06:22 -06:00
|
|
|
self.is_mine_server_name = hs.is_mine_server_name
|
2023-04-17 18:57:40 -06:00
|
|
|
self._spam_checker_module_callbacks = hs.get_module_api_callbacks().spam_checker
|
2018-01-15 09:52:07 -07:00
|
|
|
self.event_creation_handler = hs.get_event_creation_handler()
|
2021-10-08 05:44:43 -06:00
|
|
|
self.event_builder_factory = hs.get_event_builder_factory()
|
2021-04-23 05:05:51 -06:00
|
|
|
self._event_auth_handler = hs.get_event_auth_handler()
|
2021-09-24 05:25:21 -06:00
|
|
|
self._server_notices_mxid = hs.config.servernotices.server_notices_mxid
|
2018-07-25 09:32:05 -06:00
|
|
|
self.config = hs.config
|
2023-05-19 06:25:25 -06:00
|
|
|
self.http_client = hs.get_proxied_blocklisted_http_client()
|
2020-05-22 07:21:54 -06:00
|
|
|
self._replication = hs.get_replication_data_handler()
|
2021-08-26 14:41:44 -06:00
|
|
|
self._federation_event_handler = hs.get_federation_event_handler()
|
2022-09-28 07:42:43 -06:00
|
|
|
self._device_handler = hs.get_device_handler()
|
2022-09-28 06:31:53 -06:00
|
|
|
self._bulk_push_rule_evaluator = hs.get_bulk_push_rule_evaluator()
|
2022-12-05 06:07:55 -07:00
|
|
|
self._notifier = hs.get_notifier()
|
2023-08-23 02:23:41 -06:00
|
|
|
self._worker_locks = hs.get_worker_locks_handler()
|
2014-08-26 12:49:42 -06:00
|
|
|
|
2018-08-09 03:29:48 -06:00
|
|
|
self._clean_room_for_join_client = ReplicationCleanRoomRestServlet.make_client(
|
|
|
|
hs
|
|
|
|
)
|
2014-08-26 12:49:42 -06:00
|
|
|
|
2021-09-13 11:07:12 -06:00
|
|
|
if hs.config.worker.worker_app:
|
2020-11-13 09:24:04 -07:00
|
|
|
self._maybe_store_room_on_outlier_membership = (
|
|
|
|
ReplicationStoreRoomOnOutlierMembershipRestServlet.make_client(hs)
|
2020-02-26 09:58:33 -07:00
|
|
|
)
|
2020-01-30 10:06:38 -07:00
|
|
|
else:
|
2020-11-13 09:24:04 -07:00
|
|
|
self._maybe_store_room_on_outlier_membership = (
|
|
|
|
self.store.maybe_store_room_on_outlier_membership
|
|
|
|
)
|
2020-01-30 10:06:38 -07:00
|
|
|
|
2021-06-04 03:47:58 -06:00
|
|
|
self._room_backfill = Linearizer("room_backfill")
|
|
|
|
|
2023-05-04 08:18:22 -06:00
|
|
|
self._third_party_event_rules = (
|
|
|
|
hs.get_module_api_callbacks().third_party_event_rules
|
|
|
|
)
|
2019-06-12 03:31:37 -06:00
|
|
|
|
2023-01-20 05:06:19 -07:00
|
|
|
# Tracks running partial state syncs by room ID.
|
|
|
|
# Partial state syncs currently only run on the main process, so it's okay to
|
|
|
|
# track them in-memory for now.
|
|
|
|
self._active_partial_state_syncs: Set[str] = set()
|
|
|
|
# Tracks partial state syncs we may want to restart.
|
|
|
|
# A dictionary mapping room IDs to (initial destination, other destinations)
|
|
|
|
# tuples.
|
|
|
|
self._partial_state_syncs_maybe_needing_restart: Dict[
|
2023-02-03 08:39:59 -07:00
|
|
|
str, Tuple[Optional[str], AbstractSet[str]]
|
2023-01-20 05:06:19 -07:00
|
|
|
] = {}
|
2023-01-22 12:19:31 -07:00
|
|
|
# A lock guarding the partial state flag for rooms.
|
|
|
|
# When the lock is held for a given room, no other concurrent code may
|
|
|
|
# partial state or un-partial state the room.
|
|
|
|
self._is_partial_state_room_linearizer = Linearizer(
|
|
|
|
name="_is_partial_state_room_linearizer"
|
|
|
|
)
|
2023-01-20 05:06:19 -07:00
|
|
|
|
2022-05-31 09:15:08 -06:00
|
|
|
# if this is the main process, fire off a background process to resume
|
|
|
|
# any partial-state-resync operations which were in flight when we
|
|
|
|
# were shut down.
|
|
|
|
if not hs.config.worker.worker_app:
|
|
|
|
run_as_background_process(
|
2023-01-20 05:06:19 -07:00
|
|
|
"resume_sync_partial_state_room", self._resume_partial_state_room_sync
|
2022-05-31 09:15:08 -06:00
|
|
|
)
|
|
|
|
|
2022-08-03 09:57:38 -06:00
|
|
|
@trace
|
2023-06-05 22:38:52 -06:00
|
|
|
@tag_args
|
2020-09-18 07:25:52 -06:00
|
|
|
async def maybe_backfill(
|
2023-08-23 02:23:41 -06:00
|
|
|
self, room_id: str, current_depth: int, limit: int, record_time: bool = True
|
2020-09-18 07:25:52 -06:00
|
|
|
) -> bool:
|
2015-05-12 03:35:45 -06:00
|
|
|
"""Checks the database to see if we should backfill before paginating,
|
|
|
|
and if so do.
|
2020-09-18 07:25:52 -06:00
|
|
|
|
|
|
|
Args:
|
|
|
|
room_id
|
|
|
|
current_depth: The depth from which we're paginating from. This is
|
|
|
|
used to decide if we should backfill and what extremities to
|
|
|
|
use.
|
|
|
|
limit: The number of events that the pagination request will
|
|
|
|
return. This is used as part of the heuristic to decide if we
|
|
|
|
should back paginate.
|
2023-08-23 02:23:41 -06:00
|
|
|
record_time: Whether to record the time it takes to backfill.
|
2023-06-05 22:38:52 -06:00
|
|
|
|
|
|
|
Returns:
|
|
|
|
True if we actually tried to backfill something, otherwise False.
|
2015-05-11 11:01:31 -06:00
|
|
|
"""
|
2022-08-17 03:33:19 -06:00
|
|
|
# Starting the processing time here so we can include the room backfill
|
|
|
|
# linearizer lock queue in the timing
|
2023-08-23 02:23:41 -06:00
|
|
|
processing_start_time = self.clock.time_msec() if record_time else 0
|
2022-08-17 03:33:19 -06:00
|
|
|
|
2022-04-05 08:43:52 -06:00
|
|
|
async with self._room_backfill.queue(room_id):
|
2023-08-23 02:23:41 -06:00
|
|
|
async with self._worker_locks.acquire_read_write_lock(
|
|
|
|
PURGE_PAGINATION_LOCK_NAME, room_id, write=False
|
|
|
|
):
|
|
|
|
return await self._maybe_backfill_inner(
|
|
|
|
room_id,
|
|
|
|
current_depth,
|
|
|
|
limit,
|
|
|
|
processing_start_time=processing_start_time,
|
|
|
|
)
|
2021-06-04 03:47:58 -06:00
|
|
|
|
2023-06-05 22:38:52 -06:00
|
|
|
@trace
|
|
|
|
@tag_args
|
2021-06-04 03:47:58 -06:00
|
|
|
async def _maybe_backfill_inner(
|
2022-08-17 03:33:19 -06:00
|
|
|
self,
|
|
|
|
room_id: str,
|
|
|
|
current_depth: int,
|
|
|
|
limit: int,
|
|
|
|
*,
|
2022-09-28 14:26:16 -06:00
|
|
|
processing_start_time: Optional[int],
|
2021-06-04 03:47:58 -06:00
|
|
|
) -> bool:
|
2022-08-17 03:33:19 -06:00
|
|
|
"""
|
|
|
|
Checks whether the `current_depth` is at or approaching any backfill
|
|
|
|
points in the room and if so, will backfill. We only care about
|
|
|
|
checking backfill points that happened before the `current_depth`
|
|
|
|
(meaning less than or equal to the `current_depth`).
|
|
|
|
|
|
|
|
Args:
|
|
|
|
room_id: The room to backfill in.
|
|
|
|
current_depth: The depth to check at for any upcoming backfill points.
|
|
|
|
limit: The max number of events to request from the remote federated server.
|
2022-09-28 14:26:16 -06:00
|
|
|
processing_start_time: The time when `maybe_backfill` started processing.
|
|
|
|
Only used for timing. If `None`, no timing observation will be made.
|
2023-06-05 22:38:52 -06:00
|
|
|
|
|
|
|
Returns:
|
|
|
|
True if we actually tried to backfill something, otherwise False.
|
2022-08-17 03:33:19 -06:00
|
|
|
"""
|
2022-04-26 03:27:11 -06:00
|
|
|
backwards_extremities = [
|
|
|
|
_BackfillPoint(event_id, depth, _BackfillPointType.BACKWARDS_EXTREMITY)
|
2022-09-28 14:26:16 -06:00
|
|
|
for event_id, depth in await self.store.get_backfill_points_in_room(
|
|
|
|
room_id=room_id,
|
|
|
|
current_depth=current_depth,
|
|
|
|
# We only need to end up with 5 extremities combined with the
|
|
|
|
# insertion event extremities to make the `/backfill` request
|
|
|
|
# but fetch an order of magnitude more to make sure there is
|
|
|
|
# enough even after we filter them by whether visible in the
|
|
|
|
# history. This isn't fool-proof as all backfill points within
|
|
|
|
# our limit could be filtered out but seems like a good amount
|
|
|
|
# to try with at least.
|
|
|
|
limit=50,
|
|
|
|
)
|
2022-04-26 03:27:11 -06:00
|
|
|
]
|
Fix historical messages backfilling in random order on remote homeservers (MSC2716) (#11114)
Fix https://github.com/matrix-org/synapse/issues/11091
Fix https://github.com/matrix-org/synapse/issues/10764 (side-stepping the issue because we no longer have to deal with `fake_prev_event_id`)
1. Made the `/backfill` response return messages in `(depth, stream_ordering)` order (previously only sorted by `depth`)
- Technically, it shouldn't really matter how `/backfill` returns things but I'm just trying to make the `stream_ordering` a little more consistent from the origin to the remote homeservers in order to get the order of messages from `/messages` consistent ([sorted by `(topological_ordering, stream_ordering)`](https://github.com/matrix-org/synapse/blob/develop/docs/development/room-dag-concepts.md#depth-and-stream-ordering)).
- Even now that we return backfilled messages in order, it still doesn't guarantee the same `stream_ordering` (and more importantly the [`/messages` order](https://github.com/matrix-org/synapse/blob/develop/docs/development/room-dag-concepts.md#depth-and-stream-ordering)) on the other server. For example, if a room has a bunch of history imported and someone visits a permalink to a historical message back in time, their homeserver will skip over the historical messages in between and insert the permalink as the next message in the `stream_order` and totally throw off the sort.
- This will be even more the case when we add the [MSC3030 jump to date API endpoint](https://github.com/matrix-org/matrix-doc/pull/3030) so the static archives can navigate and jump to a certain date.
- We're solving this in the future by switching to [online topological ordering](https://github.com/matrix-org/gomatrixserverlib/issues/187) and [chunking](https://github.com/matrix-org/synapse/issues/3785) which by its nature will apply retroactively to fix any inconsistencies introduced by people permalinking
2. As we're navigating `prev_events` to return in `/backfill`, we order by `depth` first (newest -> oldest) and now also tie-break based on the `stream_ordering` (newest -> oldest). This is technically important because MSC2716 inserts a bunch of historical messages at the same `depth` so it's best to be prescriptive about which ones we should process first. In reality, I think the code already looped over the historical messages as expected because the database is already in order.
3. Making the historical state chain and historical event chain float on their own by having no `prev_events` instead of a fake `prev_event` which caused backfill to get clogged with an unresolvable event. Fixes https://github.com/matrix-org/synapse/issues/11091 and https://github.com/matrix-org/synapse/issues/10764
4. We no longer find connected insertion events by finding a potential `prev_event` connection to the current event we're iterating over. We now solely rely on marker events which when processed, add the insertion event as an extremity and the federating homeserver can ask about it when time calls.
- Related discussion, https://github.com/matrix-org/synapse/pull/11114#discussion_r741514793
Before | After
--- | ---
![](https://user-images.githubusercontent.com/558581/139218681-b465c862-5c49-4702-a59e-466733b0cf45.png) | ![](https://user-images.githubusercontent.com/558581/146453159-a1609e0a-8324-439d-ae44-e4bce43ac6d1.png)
#### Why aren't we sorting topologically when receiving backfill events?
> The main reason we're going to opt to not sort topologically when receiving backfill events is because it's probably best to do whatever is easiest to make it just work. People will probably have opinions once they look at [MSC2716](https://github.com/matrix-org/matrix-doc/pull/2716) which could change whatever implementation anyway.
>
> As mentioned, ideally we would do this but code necessary to make the fake edges but it gets confusing and gives an impression of “just whyyyy” (feels icky). This problem also dissolves with online topological ordering.
>
> -- https://github.com/matrix-org/synapse/pull/11114#discussion_r741517138
See https://github.com/matrix-org/synapse/pull/11114#discussion_r739610091 for the technical difficulties
2022-02-07 14:54:13 -07:00
|
|
|
|
2022-04-26 03:27:11 -06:00
|
|
|
# we now have a list of potential places to backpaginate from. We prefer to
|
|
|
|
# start with the most recent (ie, max depth), so let's sort the list.
|
|
|
|
sorted_backfill_points: List[_BackfillPoint] = sorted(
|
2023-06-16 13:12:24 -06:00
|
|
|
backwards_extremities,
|
2022-04-26 03:27:11 -06:00
|
|
|
key=lambda e: -int(e.depth),
|
2019-02-20 09:54:35 -07:00
|
|
|
)
|
|
|
|
|
Add support for MSC2716 marker events (#10498)
* Make historical messages available to federated servers
Part of MSC2716: https://github.com/matrix-org/matrix-doc/pull/2716
Follow-up to https://github.com/matrix-org/synapse/pull/9247
* Debug message not available on federation
* Add base starting insertion point when no chunk ID is provided
* Fix messages from multiple senders in historical chunk
Follow-up to https://github.com/matrix-org/synapse/pull/9247
Part of MSC2716: https://github.com/matrix-org/matrix-doc/pull/2716
---
Previously, Synapse would throw a 403,
`Cannot force another user to join.`,
because we were trying to use `?user_id` from a single virtual user
which did not match with messages from other users in the chunk.
* Remove debug lines
* Messing with selecting insertion event extremeties
* Move db schema change to new version
* Add more better comments
* Make a fake requester with just what we need
See https://github.com/matrix-org/synapse/pull/10276#discussion_r660999080
* Store insertion events in table
* Make base insertion event float off on its own
See https://github.com/matrix-org/synapse/pull/10250#issuecomment-875711889
Conflicts:
synapse/rest/client/v1/room.py
* Validate that the app service can actually control the given user
See https://github.com/matrix-org/synapse/pull/10276#issuecomment-876316455
Conflicts:
synapse/rest/client/v1/room.py
* Add some better comments on what we're trying to check for
* Continue debugging
* Share validation logic
* Add inserted historical messages to /backfill response
* Remove debug sql queries
* Some marker event implemntation trials
* Clean up PR
* Rename insertion_event_id to just event_id
* Add some better sql comments
* More accurate description
* Add changelog
* Make it clear what MSC the change is part of
* Add more detail on which insertion event came through
* Address review and improve sql queries
* Only use event_id as unique constraint
* Fix test case where insertion event is already in the normal DAG
* Remove debug changes
* Add support for MSC2716 marker events
* Process markers when we receive it over federation
* WIP: make hs2 backfill historical messages after marker event
* hs2 to better ask for insertion event extremity
But running into the `sqlite3.IntegrityError: NOT NULL constraint failed: event_to_state_groups.state_group`
error
* Add insertion_event_extremities table
* Switch to chunk events so we can auth via power_levels
Previously, we were using `content.chunk_id` to connect one
chunk to another. But these events can be from any `sender`
and we can't tell who should be able to send historical events.
We know we only want the application service to do it but these
events have the sender of a real historical message, not the
application service user ID as the sender. Other federated homeservers
also have no indicator which senders are an application service on
the originating homeserver.
So we want to auth all of the MSC2716 events via power_levels
and have them be sent by the application service with proper
PL levels in the room.
* Switch to chunk events for federation
* Add unstable room version to support new historical PL
* Messy: Fix undefined state_group for federated historical events
```
2021-07-13 02:27:57,810 - synapse.handlers.federation - 1248 - ERROR - GET-4 - Failed to backfill from hs1 because NOT NULL constraint failed: event_to_state_groups.state_group
Traceback (most recent call last):
File "/usr/local/lib/python3.8/site-packages/synapse/handlers/federation.py", line 1216, in try_backfill
await self.backfill(
File "/usr/local/lib/python3.8/site-packages/synapse/handlers/federation.py", line 1035, in backfill
await self._auth_and_persist_event(dest, event, context, backfilled=True)
File "/usr/local/lib/python3.8/site-packages/synapse/handlers/federation.py", line 2222, in _auth_and_persist_event
await self._run_push_actions_and_persist_event(event, context, backfilled)
File "/usr/local/lib/python3.8/site-packages/synapse/handlers/federation.py", line 2244, in _run_push_actions_and_persist_event
await self.persist_events_and_notify(
File "/usr/local/lib/python3.8/site-packages/synapse/handlers/federation.py", line 3290, in persist_events_and_notify
events, max_stream_token = await self.storage.persistence.persist_events(
File "/usr/local/lib/python3.8/site-packages/synapse/logging/opentracing.py", line 774, in _trace_inner
return await func(*args, **kwargs)
File "/usr/local/lib/python3.8/site-packages/synapse/storage/persist_events.py", line 320, in persist_events
ret_vals = await yieldable_gather_results(enqueue, partitioned.items())
File "/usr/local/lib/python3.8/site-packages/synapse/storage/persist_events.py", line 237, in handle_queue_loop
ret = await self._per_item_callback(
File "/usr/local/lib/python3.8/site-packages/synapse/storage/persist_events.py", line 577, in _persist_event_batch
await self.persist_events_store._persist_events_and_state_updates(
File "/usr/local/lib/python3.8/site-packages/synapse/storage/databases/main/events.py", line 176, in _persist_events_and_state_updates
await self.db_pool.runInteraction(
File "/usr/local/lib/python3.8/site-packages/synapse/storage/database.py", line 681, in runInteraction
result = await self.runWithConnection(
File "/usr/local/lib/python3.8/site-packages/synapse/storage/database.py", line 770, in runWithConnection
return await make_deferred_yieldable(
File "/usr/local/lib/python3.8/site-packages/twisted/python/threadpool.py", line 238, in inContext
result = inContext.theWork() # type: ignore[attr-defined]
File "/usr/local/lib/python3.8/site-packages/twisted/python/threadpool.py", line 254, in <lambda>
inContext.theWork = lambda: context.call( # type: ignore[attr-defined]
File "/usr/local/lib/python3.8/site-packages/twisted/python/context.py", line 118, in callWithContext
return self.currentContext().callWithContext(ctx, func, *args, **kw)
File "/usr/local/lib/python3.8/site-packages/twisted/python/context.py", line 83, in callWithContext
return func(*args, **kw)
File "/usr/local/lib/python3.8/site-packages/twisted/enterprise/adbapi.py", line 293, in _runWithConnection
compat.reraise(excValue, excTraceback)
File "/usr/local/lib/python3.8/site-packages/twisted/python/deprecate.py", line 298, in deprecatedFunction
return function(*args, **kwargs)
File "/usr/local/lib/python3.8/site-packages/twisted/python/compat.py", line 403, in reraise
raise exception.with_traceback(traceback)
File "/usr/local/lib/python3.8/site-packages/twisted/enterprise/adbapi.py", line 284, in _runWithConnection
result = func(conn, *args, **kw)
File "/usr/local/lib/python3.8/site-packages/synapse/storage/database.py", line 765, in inner_func
return func(db_conn, *args, **kwargs)
File "/usr/local/lib/python3.8/site-packages/synapse/storage/database.py", line 549, in new_transaction
r = func(cursor, *args, **kwargs)
File "/usr/local/lib/python3.8/site-packages/synapse/logging/utils.py", line 69, in wrapped
return f(*args, **kwargs)
File "/usr/local/lib/python3.8/site-packages/synapse/storage/databases/main/events.py", line 385, in _persist_events_txn
self._store_event_state_mappings_txn(txn, events_and_contexts)
File "/usr/local/lib/python3.8/site-packages/synapse/storage/databases/main/events.py", line 2065, in _store_event_state_mappings_txn
self.db_pool.simple_insert_many_txn(
File "/usr/local/lib/python3.8/site-packages/synapse/storage/database.py", line 923, in simple_insert_many_txn
txn.execute_batch(sql, vals)
File "/usr/local/lib/python3.8/site-packages/synapse/storage/database.py", line 280, in execute_batch
self.executemany(sql, args)
File "/usr/local/lib/python3.8/site-packages/synapse/storage/database.py", line 300, in executemany
self._do_execute(self.txn.executemany, sql, *args)
File "/usr/local/lib/python3.8/site-packages/synapse/storage/database.py", line 330, in _do_execute
return func(sql, *args)
sqlite3.IntegrityError: NOT NULL constraint failed: event_to_state_groups.state_group
```
* Revert "Messy: Fix undefined state_group for federated historical events"
This reverts commit 187ab28611546321e02770944c86f30ee2bc742a.
* Fix federated events being rejected for no state_groups
Add fix from https://github.com/matrix-org/synapse/pull/10439
until it merges.
* Adapting to experimental room version
* Some log cleanup
* Add better comments around extremity fetching code and why
* Rename to be more accurate to what the function returns
* Add changelog
* Ignore rejected events
* Use simplified upsert
* Add Erik's explanation of extra event checks
See https://github.com/matrix-org/synapse/pull/10498#discussion_r680880332
* Clarify that the depth is not directly correlated to the backwards extremity that we return
See https://github.com/matrix-org/synapse/pull/10498#discussion_r681725404
* lock only matters for sqlite
See https://github.com/matrix-org/synapse/pull/10498#discussion_r681728061
* Move new SQL changes to its own delta file
* Clean up upsert docstring
* Bump database schema version (62)
2021-08-04 11:07:57 -06:00
|
|
|
logger.debug(
|
2022-04-26 03:27:11 -06:00
|
|
|
"_maybe_backfill_inner: room_id: %s: current_depth: %s, limit: %s, "
|
|
|
|
"backfill points (%d): %s",
|
|
|
|
room_id,
|
|
|
|
current_depth,
|
|
|
|
limit,
|
|
|
|
len(sorted_backfill_points),
|
|
|
|
sorted_backfill_points,
|
Add support for MSC2716 marker events (#10498)
* Make historical messages available to federated servers
Part of MSC2716: https://github.com/matrix-org/matrix-doc/pull/2716
Follow-up to https://github.com/matrix-org/synapse/pull/9247
* Debug message not available on federation
* Add base starting insertion point when no chunk ID is provided
* Fix messages from multiple senders in historical chunk
Follow-up to https://github.com/matrix-org/synapse/pull/9247
Part of MSC2716: https://github.com/matrix-org/matrix-doc/pull/2716
---
Previously, Synapse would throw a 403,
`Cannot force another user to join.`,
because we were trying to use `?user_id` from a single virtual user
which did not match with messages from other users in the chunk.
* Remove debug lines
* Messing with selecting insertion event extremeties
* Move db schema change to new version
* Add more better comments
* Make a fake requester with just what we need
See https://github.com/matrix-org/synapse/pull/10276#discussion_r660999080
* Store insertion events in table
* Make base insertion event float off on its own
See https://github.com/matrix-org/synapse/pull/10250#issuecomment-875711889
Conflicts:
synapse/rest/client/v1/room.py
* Validate that the app service can actually control the given user
See https://github.com/matrix-org/synapse/pull/10276#issuecomment-876316455
Conflicts:
synapse/rest/client/v1/room.py
* Add some better comments on what we're trying to check for
* Continue debugging
* Share validation logic
* Add inserted historical messages to /backfill response
* Remove debug sql queries
* Some marker event implemntation trials
* Clean up PR
* Rename insertion_event_id to just event_id
* Add some better sql comments
* More accurate description
* Add changelog
* Make it clear what MSC the change is part of
* Add more detail on which insertion event came through
* Address review and improve sql queries
* Only use event_id as unique constraint
* Fix test case where insertion event is already in the normal DAG
* Remove debug changes
* Add support for MSC2716 marker events
* Process markers when we receive it over federation
* WIP: make hs2 backfill historical messages after marker event
* hs2 to better ask for insertion event extremity
But running into the `sqlite3.IntegrityError: NOT NULL constraint failed: event_to_state_groups.state_group`
error
* Add insertion_event_extremities table
* Switch to chunk events so we can auth via power_levels
Previously, we were using `content.chunk_id` to connect one
chunk to another. But these events can be from any `sender`
and we can't tell who should be able to send historical events.
We know we only want the application service to do it but these
events have the sender of a real historical message, not the
application service user ID as the sender. Other federated homeservers
also have no indicator which senders are an application service on
the originating homeserver.
So we want to auth all of the MSC2716 events via power_levels
and have them be sent by the application service with proper
PL levels in the room.
* Switch to chunk events for federation
* Add unstable room version to support new historical PL
* Messy: Fix undefined state_group for federated historical events
```
2021-07-13 02:27:57,810 - synapse.handlers.federation - 1248 - ERROR - GET-4 - Failed to backfill from hs1 because NOT NULL constraint failed: event_to_state_groups.state_group
Traceback (most recent call last):
File "/usr/local/lib/python3.8/site-packages/synapse/handlers/federation.py", line 1216, in try_backfill
await self.backfill(
File "/usr/local/lib/python3.8/site-packages/synapse/handlers/federation.py", line 1035, in backfill
await self._auth_and_persist_event(dest, event, context, backfilled=True)
File "/usr/local/lib/python3.8/site-packages/synapse/handlers/federation.py", line 2222, in _auth_and_persist_event
await self._run_push_actions_and_persist_event(event, context, backfilled)
File "/usr/local/lib/python3.8/site-packages/synapse/handlers/federation.py", line 2244, in _run_push_actions_and_persist_event
await self.persist_events_and_notify(
File "/usr/local/lib/python3.8/site-packages/synapse/handlers/federation.py", line 3290, in persist_events_and_notify
events, max_stream_token = await self.storage.persistence.persist_events(
File "/usr/local/lib/python3.8/site-packages/synapse/logging/opentracing.py", line 774, in _trace_inner
return await func(*args, **kwargs)
File "/usr/local/lib/python3.8/site-packages/synapse/storage/persist_events.py", line 320, in persist_events
ret_vals = await yieldable_gather_results(enqueue, partitioned.items())
File "/usr/local/lib/python3.8/site-packages/synapse/storage/persist_events.py", line 237, in handle_queue_loop
ret = await self._per_item_callback(
File "/usr/local/lib/python3.8/site-packages/synapse/storage/persist_events.py", line 577, in _persist_event_batch
await self.persist_events_store._persist_events_and_state_updates(
File "/usr/local/lib/python3.8/site-packages/synapse/storage/databases/main/events.py", line 176, in _persist_events_and_state_updates
await self.db_pool.runInteraction(
File "/usr/local/lib/python3.8/site-packages/synapse/storage/database.py", line 681, in runInteraction
result = await self.runWithConnection(
File "/usr/local/lib/python3.8/site-packages/synapse/storage/database.py", line 770, in runWithConnection
return await make_deferred_yieldable(
File "/usr/local/lib/python3.8/site-packages/twisted/python/threadpool.py", line 238, in inContext
result = inContext.theWork() # type: ignore[attr-defined]
File "/usr/local/lib/python3.8/site-packages/twisted/python/threadpool.py", line 254, in <lambda>
inContext.theWork = lambda: context.call( # type: ignore[attr-defined]
File "/usr/local/lib/python3.8/site-packages/twisted/python/context.py", line 118, in callWithContext
return self.currentContext().callWithContext(ctx, func, *args, **kw)
File "/usr/local/lib/python3.8/site-packages/twisted/python/context.py", line 83, in callWithContext
return func(*args, **kw)
File "/usr/local/lib/python3.8/site-packages/twisted/enterprise/adbapi.py", line 293, in _runWithConnection
compat.reraise(excValue, excTraceback)
File "/usr/local/lib/python3.8/site-packages/twisted/python/deprecate.py", line 298, in deprecatedFunction
return function(*args, **kwargs)
File "/usr/local/lib/python3.8/site-packages/twisted/python/compat.py", line 403, in reraise
raise exception.with_traceback(traceback)
File "/usr/local/lib/python3.8/site-packages/twisted/enterprise/adbapi.py", line 284, in _runWithConnection
result = func(conn, *args, **kw)
File "/usr/local/lib/python3.8/site-packages/synapse/storage/database.py", line 765, in inner_func
return func(db_conn, *args, **kwargs)
File "/usr/local/lib/python3.8/site-packages/synapse/storage/database.py", line 549, in new_transaction
r = func(cursor, *args, **kwargs)
File "/usr/local/lib/python3.8/site-packages/synapse/logging/utils.py", line 69, in wrapped
return f(*args, **kwargs)
File "/usr/local/lib/python3.8/site-packages/synapse/storage/databases/main/events.py", line 385, in _persist_events_txn
self._store_event_state_mappings_txn(txn, events_and_contexts)
File "/usr/local/lib/python3.8/site-packages/synapse/storage/databases/main/events.py", line 2065, in _store_event_state_mappings_txn
self.db_pool.simple_insert_many_txn(
File "/usr/local/lib/python3.8/site-packages/synapse/storage/database.py", line 923, in simple_insert_many_txn
txn.execute_batch(sql, vals)
File "/usr/local/lib/python3.8/site-packages/synapse/storage/database.py", line 280, in execute_batch
self.executemany(sql, args)
File "/usr/local/lib/python3.8/site-packages/synapse/storage/database.py", line 300, in executemany
self._do_execute(self.txn.executemany, sql, *args)
File "/usr/local/lib/python3.8/site-packages/synapse/storage/database.py", line 330, in _do_execute
return func(sql, *args)
sqlite3.IntegrityError: NOT NULL constraint failed: event_to_state_groups.state_group
```
* Revert "Messy: Fix undefined state_group for federated historical events"
This reverts commit 187ab28611546321e02770944c86f30ee2bc742a.
* Fix federated events being rejected for no state_groups
Add fix from https://github.com/matrix-org/synapse/pull/10439
until it merges.
* Adapting to experimental room version
* Some log cleanup
* Add better comments around extremity fetching code and why
* Rename to be more accurate to what the function returns
* Add changelog
* Ignore rejected events
* Use simplified upsert
* Add Erik's explanation of extra event checks
See https://github.com/matrix-org/synapse/pull/10498#discussion_r680880332
* Clarify that the depth is not directly correlated to the backwards extremity that we return
See https://github.com/matrix-org/synapse/pull/10498#discussion_r681725404
* lock only matters for sqlite
See https://github.com/matrix-org/synapse/pull/10498#discussion_r681728061
* Move new SQL changes to its own delta file
* Clean up upsert docstring
* Bump database schema version (62)
2021-08-04 11:07:57 -06:00
|
|
|
)
|
2023-06-05 22:38:52 -06:00
|
|
|
set_tag(
|
|
|
|
SynapseTags.RESULT_PREFIX + "sorted_backfill_points",
|
|
|
|
str(sorted_backfill_points),
|
|
|
|
)
|
|
|
|
set_tag(
|
|
|
|
SynapseTags.RESULT_PREFIX + "sorted_backfill_points.length",
|
|
|
|
str(len(sorted_backfill_points)),
|
|
|
|
)
|
2019-02-20 09:54:35 -07:00
|
|
|
|
2023-06-09 14:39:49 -06:00
|
|
|
# If we have no backfill points lower than the `current_depth` then either we
|
|
|
|
# can a) bail or b) still attempt to backfill. We opt to try backfilling anyway
|
|
|
|
# just in case we do get relevant events. This is good for eventual consistency
|
|
|
|
# sake but we don't need to block the client for something that is just as
|
|
|
|
# likely not to return anything relevant so we backfill in the background. The
|
|
|
|
# only way, this could return something relevant is if we discover a new branch
|
|
|
|
# of history that extends all the way back to where we are currently paginating
|
|
|
|
# and it's within the 100 events that are returned from `/backfill`.
|
2022-09-28 14:26:16 -06:00
|
|
|
if not sorted_backfill_points and current_depth != MAX_DEPTH:
|
2023-08-23 02:23:41 -06:00
|
|
|
# Check that we actually have later backfill points, if not just return.
|
|
|
|
have_later_backfill_points = await self.store.get_backfill_points_in_room(
|
|
|
|
room_id=room_id,
|
|
|
|
current_depth=MAX_DEPTH,
|
|
|
|
limit=1,
|
|
|
|
)
|
|
|
|
if not have_later_backfill_points:
|
|
|
|
return False
|
|
|
|
|
2022-09-28 14:26:16 -06:00
|
|
|
logger.debug(
|
|
|
|
"_maybe_backfill_inner: all backfill points are *after* current depth. Trying again with later backfill points."
|
|
|
|
)
|
2023-06-09 14:39:49 -06:00
|
|
|
run_as_background_process(
|
|
|
|
"_maybe_backfill_inner_anyway_with_max_depth",
|
2023-08-23 02:23:41 -06:00
|
|
|
self.maybe_backfill,
|
2022-09-28 14:26:16 -06:00
|
|
|
room_id=room_id,
|
|
|
|
# We use `MAX_DEPTH` so that we find all backfill points next
|
|
|
|
# time (all events are below the `MAX_DEPTH`)
|
|
|
|
current_depth=MAX_DEPTH,
|
|
|
|
limit=limit,
|
|
|
|
# We don't want to start another timing observation from this
|
|
|
|
# nested recursive call. The top-most call can record the time
|
|
|
|
# overall otherwise the smaller one will throw off the results.
|
2023-08-23 02:23:41 -06:00
|
|
|
record_time=False,
|
2022-09-28 14:26:16 -06:00
|
|
|
)
|
2023-06-09 14:39:49 -06:00
|
|
|
# We return `False` because we're backfilling in the background and there is
|
|
|
|
# no new events immediately for the caller to know about yet.
|
|
|
|
return False
|
2022-09-28 14:26:16 -06:00
|
|
|
|
|
|
|
# Even after recursing with `MAX_DEPTH`, we didn't find any
|
|
|
|
# backward extremities to backfill from.
|
|
|
|
if not sorted_backfill_points:
|
|
|
|
logger.debug(
|
|
|
|
"_maybe_backfill_inner: Not backfilling as no backward extremeties found."
|
|
|
|
)
|
|
|
|
return False
|
|
|
|
|
2020-09-18 07:25:52 -06:00
|
|
|
# If we're approaching an extremity we trigger a backfill, otherwise we
|
|
|
|
# no-op.
|
|
|
|
#
|
|
|
|
# We chose twice the limit here as then clients paginating backwards
|
|
|
|
# will send pagination requests that trigger backfill at least twice
|
|
|
|
# using the most recent extremity before it gets removed (see below). We
|
|
|
|
# chose more than one times the limit in case of failure, but choosing a
|
|
|
|
# much larger factor will result in triggering a backfill request much
|
|
|
|
# earlier than necessary.
|
2022-09-28 14:26:16 -06:00
|
|
|
max_depth_of_backfill_points = sorted_backfill_points[0].depth
|
|
|
|
if current_depth - 2 * limit > max_depth_of_backfill_points:
|
2020-09-18 07:25:52 -06:00
|
|
|
logger.debug(
|
|
|
|
"Not backfilling as we don't need to. %d < %d - 2 * %d",
|
2022-09-28 14:26:16 -06:00
|
|
|
max_depth_of_backfill_points,
|
2020-09-18 07:25:52 -06:00
|
|
|
current_depth,
|
|
|
|
limit,
|
|
|
|
)
|
|
|
|
return False
|
|
|
|
|
2022-04-26 03:27:11 -06:00
|
|
|
# For performance's sake, we only want to paginate from a particular extremity
|
|
|
|
# if we can actually see the events we'll get. Otherwise, we'd just spend a lot
|
|
|
|
# of resources to get redacted events. We check each extremity in turn and
|
|
|
|
# ignore those which users on our server wouldn't be able to see.
|
|
|
|
#
|
|
|
|
# Additionally, we limit ourselves to backfilling from at most 5 extremities,
|
|
|
|
# for two reasons:
|
|
|
|
#
|
|
|
|
# - The check which determines if we can see an extremity's events can be
|
|
|
|
# expensive (we load the full state for the room at each of the backfill
|
|
|
|
# points, or (worse) their successors)
|
|
|
|
# - We want to avoid the server-server API request URI becoming too long.
|
|
|
|
#
|
|
|
|
# *Note*: the spec wants us to keep backfilling until we reach the start
|
|
|
|
# of the room in case we are allowed to see some of the history. However,
|
|
|
|
# in practice that causes more issues than its worth, as (a) it's
|
|
|
|
# relatively rare for there to be any visible history and (b) even when
|
|
|
|
# there is it's often sufficiently long ago that clients would stop
|
|
|
|
# attempting to paginate before backfill reached the visible history.
|
|
|
|
|
|
|
|
extremities_to_request: List[str] = []
|
|
|
|
for bp in sorted_backfill_points:
|
|
|
|
if len(extremities_to_request) >= 5:
|
|
|
|
break
|
|
|
|
|
|
|
|
# For regular backwards extremities, we don't have the extremity events
|
|
|
|
# themselves, so we need to actually check the events that reference them -
|
|
|
|
# their "successor" events.
|
|
|
|
#
|
|
|
|
# TODO: Correctly handle the case where we are allowed to see the
|
|
|
|
# successor event but not the backward extremity, e.g. in the case of
|
|
|
|
# initial join of the server where we are allowed to see the join
|
|
|
|
# event but not anything before it. This would require looking at the
|
|
|
|
# state *before* the event, ignoring the special casing certain event
|
|
|
|
# types have.
|
2023-06-16 13:12:24 -06:00
|
|
|
event_ids_to_check = await self.store.get_successor_events(bp.event_id)
|
2022-04-26 03:27:11 -06:00
|
|
|
|
|
|
|
events_to_check = await self.store.get_events_as_list(
|
|
|
|
event_ids_to_check,
|
2022-05-04 05:26:11 -06:00
|
|
|
redact_behaviour=EventRedactBehaviour.as_is,
|
2022-04-26 03:27:11 -06:00
|
|
|
get_prev_content=False,
|
|
|
|
)
|
|
|
|
|
2023-03-10 08:31:25 -07:00
|
|
|
# We unset `filter_out_erased_senders` as we might otherwise get false
|
2022-04-26 03:27:11 -06:00
|
|
|
# positives from users having been erased.
|
|
|
|
filtered_extremities = await filter_events_for_server(
|
2022-05-31 06:17:50 -06:00
|
|
|
self._storage_controllers,
|
2022-04-26 03:27:11 -06:00
|
|
|
self.server_name,
|
2022-11-21 08:46:14 -07:00
|
|
|
self.server_name,
|
2022-04-26 03:27:11 -06:00
|
|
|
events_to_check,
|
|
|
|
redact=False,
|
2023-03-10 08:31:25 -07:00
|
|
|
filter_out_erased_senders=False,
|
|
|
|
filter_out_remote_partial_state_events=False,
|
2022-04-26 03:27:11 -06:00
|
|
|
)
|
|
|
|
if filtered_extremities:
|
|
|
|
extremities_to_request.append(bp.event_id)
|
|
|
|
else:
|
|
|
|
logger.debug(
|
|
|
|
"_maybe_backfill_inner: skipping extremity %s as it would not be visible",
|
|
|
|
bp,
|
|
|
|
)
|
|
|
|
|
|
|
|
if not extremities_to_request:
|
|
|
|
logger.debug(
|
|
|
|
"_maybe_backfill_inner: found no extremities which would be visible"
|
|
|
|
)
|
|
|
|
return False
|
2020-09-18 07:25:52 -06:00
|
|
|
|
2022-04-26 03:27:11 -06:00
|
|
|
logger.debug(
|
|
|
|
"_maybe_backfill_inner: extremities_to_request %s", extremities_to_request
|
|
|
|
)
|
2022-08-16 11:39:40 -06:00
|
|
|
set_tag(
|
|
|
|
SynapseTags.RESULT_PREFIX + "extremities_to_request",
|
|
|
|
str(extremities_to_request),
|
|
|
|
)
|
|
|
|
set_tag(
|
|
|
|
SynapseTags.RESULT_PREFIX + "extremities_to_request.length",
|
|
|
|
str(len(extremities_to_request)),
|
|
|
|
)
|
2016-08-16 04:34:36 -06:00
|
|
|
|
2015-05-11 11:01:31 -06:00
|
|
|
# Now we need to decide which hosts to hit first.
|
2022-08-30 00:38:14 -06:00
|
|
|
# First we try hosts that are already in the room.
|
2015-05-12 03:35:45 -06:00
|
|
|
# TODO: HEURISTIC ALERT.
|
2022-08-30 00:38:14 -06:00
|
|
|
likely_domains = (
|
2022-09-30 06:15:32 -06:00
|
|
|
await self._storage_controllers.state.get_current_hosts_in_room_ordered(
|
|
|
|
room_id
|
|
|
|
)
|
2022-08-30 00:38:14 -06:00
|
|
|
)
|
2015-05-11 11:01:31 -06:00
|
|
|
|
2023-01-26 10:31:58 -07:00
|
|
|
async def try_backfill(domains: StrCollection) -> bool:
|
2015-05-11 11:01:31 -06:00
|
|
|
# TODO: Should we try multiple of these at a time?
|
2022-09-29 04:00:02 -06:00
|
|
|
|
|
|
|
# Number of contacted remote homeservers that have denied our backfill
|
|
|
|
# request with a 4xx code.
|
|
|
|
denied_count = 0
|
|
|
|
|
|
|
|
# Maximum number of contacted remote homeservers that can deny our
|
|
|
|
# backfill request with 4xx codes before we give up.
|
|
|
|
max_denied_count = 5
|
|
|
|
|
2015-05-11 11:01:31 -06:00
|
|
|
for dom in domains:
|
2022-08-30 00:38:14 -06:00
|
|
|
# We don't want to ask our own server for information we don't have
|
2023-05-05 08:06:22 -06:00
|
|
|
if self.is_mine_server_name(dom):
|
2022-08-30 00:38:14 -06:00
|
|
|
continue
|
|
|
|
|
2015-05-12 03:35:45 -06:00
|
|
|
try:
|
2021-08-26 14:41:44 -06:00
|
|
|
await self._federation_event_handler.backfill(
|
2022-04-26 03:27:11 -06:00
|
|
|
dom, room_id, limit=100, extremities=extremities_to_request
|
2015-05-12 03:35:45 -06:00
|
|
|
)
|
2016-04-12 05:04:19 -06:00
|
|
|
# If this succeeded then we probably already have the
|
|
|
|
# appropriate stuff.
|
2016-04-12 05:48:30 -06:00
|
|
|
# TODO: We can probably do something more intelligent here.
|
2019-07-23 07:00:55 -06:00
|
|
|
return True
|
2022-10-26 15:10:55 -06:00
|
|
|
except NotRetryingDestination as e:
|
|
|
|
logger.info("_maybe_backfill_inner: %s", e)
|
|
|
|
continue
|
|
|
|
except FederationDeniedError:
|
|
|
|
logger.info(
|
|
|
|
"_maybe_backfill_inner: Not attempting to backfill from %s because the homeserver is not on our federation whitelist",
|
|
|
|
dom,
|
|
|
|
)
|
|
|
|
continue
|
2021-08-26 11:34:57 -06:00
|
|
|
except (SynapseError, InvalidResponseError) as e:
|
2020-05-22 04:39:20 -06:00
|
|
|
logger.info("Failed to backfill from %s because %s", dom, e)
|
|
|
|
continue
|
|
|
|
except HttpResponseException as e:
|
|
|
|
if 400 <= e.code < 500:
|
2022-09-29 04:00:02 -06:00
|
|
|
logger.warning(
|
|
|
|
"Backfill denied from %s because %s [%d/%d]",
|
|
|
|
dom,
|
|
|
|
e,
|
|
|
|
denied_count,
|
|
|
|
max_denied_count,
|
|
|
|
)
|
|
|
|
denied_count += 1
|
|
|
|
if denied_count >= max_denied_count:
|
|
|
|
return False
|
|
|
|
continue
|
2020-05-22 04:39:20 -06:00
|
|
|
|
2015-05-12 07:09:54 -06:00
|
|
|
logger.info("Failed to backfill from %s because %s", dom, e)
|
2015-05-12 03:35:45 -06:00
|
|
|
continue
|
|
|
|
except CodeMessageException as e:
|
|
|
|
if 400 <= e.code < 500:
|
2022-09-29 04:00:02 -06:00
|
|
|
logger.warning(
|
|
|
|
"Backfill denied from %s because %s [%d/%d]",
|
|
|
|
dom,
|
|
|
|
e,
|
|
|
|
denied_count,
|
|
|
|
max_denied_count,
|
|
|
|
)
|
|
|
|
denied_count += 1
|
|
|
|
if denied_count >= max_denied_count:
|
|
|
|
return False
|
|
|
|
continue
|
2015-05-12 03:35:45 -06:00
|
|
|
|
2015-05-12 07:09:54 -06:00
|
|
|
logger.info("Failed to backfill from %s because %s", dom, e)
|
2015-05-12 03:35:45 -06:00
|
|
|
continue
|
2019-07-30 06:19:22 -06:00
|
|
|
except RequestSendFailed as e:
|
2020-10-23 10:38:40 -06:00
|
|
|
logger.info("Failed to get backfill from %s because %s", dom, e)
|
2019-07-30 06:19:22 -06:00
|
|
|
continue
|
2015-05-12 03:35:45 -06:00
|
|
|
except Exception as e:
|
2015-05-12 07:09:54 -06:00
|
|
|
logger.exception("Failed to backfill from %s because %s", dom, e)
|
2015-05-12 03:35:45 -06:00
|
|
|
continue
|
2015-05-11 11:01:31 -06:00
|
|
|
|
2019-07-23 07:00:55 -06:00
|
|
|
return False
|
2015-05-11 11:01:31 -06:00
|
|
|
|
2022-09-28 14:26:16 -06:00
|
|
|
# If we have the `processing_start_time`, then we can make an
|
|
|
|
# observation. We wouldn't have the `processing_start_time` in the case
|
|
|
|
# where `_maybe_backfill_inner` is recursively called to find any
|
|
|
|
# backfill points regardless of `current_depth`.
|
|
|
|
if processing_start_time is not None:
|
|
|
|
processing_end_time = self.clock.time_msec()
|
|
|
|
backfill_processing_before_timer.observe(
|
|
|
|
(processing_end_time - processing_start_time) / 1000
|
|
|
|
)
|
2022-08-17 03:33:19 -06:00
|
|
|
|
2019-12-10 09:54:34 -07:00
|
|
|
success = await try_backfill(likely_domains)
|
2015-05-11 11:01:31 -06:00
|
|
|
if success:
|
2019-07-23 07:00:55 -06:00
|
|
|
return True
|
2015-05-11 11:01:31 -06:00
|
|
|
|
2022-03-09 04:00:48 -07:00
|
|
|
# TODO: we could also try servers which were previously in the room, but
|
|
|
|
# are no longer.
|
2015-05-11 11:01:31 -06:00
|
|
|
|
2021-08-26 14:41:44 -06:00
|
|
|
return False
|
2018-04-17 16:41:12 -06:00
|
|
|
|
2021-04-06 05:21:57 -06:00
|
|
|
async def send_invite(self, target_host: str, event: EventBase) -> EventBase:
|
2014-11-12 09:20:21 -07:00
|
|
|
"""Sends the invite to the remote server for signing.
|
|
|
|
|
|
|
|
Invites must be signed by the invitee's server before distribution.
|
|
|
|
"""
|
2021-07-15 03:35:46 -06:00
|
|
|
try:
|
|
|
|
pdu = await self.federation_client.send_invite(
|
|
|
|
destination=target_host,
|
|
|
|
room_id=event.room_id,
|
|
|
|
event_id=event.event_id,
|
|
|
|
pdu=event,
|
|
|
|
)
|
|
|
|
except RequestSendFailed:
|
|
|
|
raise SynapseError(502, f"Can't connect to server {target_host}")
|
2014-11-07 06:41:00 -07:00
|
|
|
|
2019-07-23 07:00:55 -06:00
|
|
|
return pdu
|
2014-11-07 06:41:00 -07:00
|
|
|
|
2020-02-03 09:06:46 -07:00
|
|
|
async def on_event_auth(self, event_id: str) -> List[EventBase]:
|
|
|
|
event = await self.store.get_event(event_id)
|
|
|
|
auth = await self.store.get_auth_chain(
|
2021-03-10 07:57:59 -07:00
|
|
|
event.room_id, list(event.auth_event_ids()), include_given=True
|
2017-05-24 07:22:41 -06:00
|
|
|
)
|
2020-02-03 09:06:46 -07:00
|
|
|
return list(auth)
|
2014-11-07 08:35:53 -07:00
|
|
|
|
2020-02-03 09:13:13 -07:00
|
|
|
async def do_invite_join(
|
|
|
|
self, target_hosts: Iterable[str], room_id: str, joinee: str, content: JsonDict
|
2020-05-22 07:21:54 -06:00
|
|
|
) -> Tuple[str, int]:
|
2014-11-12 09:20:21 -07:00
|
|
|
"""Attempts to join the `joinee` to the room `room_id` via the
|
2019-11-01 04:28:09 -06:00
|
|
|
servers contained in `target_hosts`.
|
2014-11-12 09:20:21 -07:00
|
|
|
|
|
|
|
This first triggers a /make_join/ request that returns a partial
|
|
|
|
event that we can fill out and sign. This is then sent to the
|
|
|
|
remote server via /send_join/ which responds with the state at that
|
|
|
|
event and the auth_chains.
|
|
|
|
|
|
|
|
We suspend processing of any received events from this room until we
|
|
|
|
have finished processing the join.
|
2019-11-01 04:28:09 -06:00
|
|
|
|
|
|
|
Args:
|
2020-02-03 09:13:13 -07:00
|
|
|
target_hosts: List of servers to attempt to join the room with.
|
2019-11-01 04:28:09 -06:00
|
|
|
|
2020-02-03 09:13:13 -07:00
|
|
|
room_id: The ID of the room to join.
|
2019-11-01 04:28:09 -06:00
|
|
|
|
2020-02-03 09:13:13 -07:00
|
|
|
joinee: The User ID of the joining user.
|
2019-11-01 04:28:09 -06:00
|
|
|
|
2020-02-03 09:13:13 -07:00
|
|
|
content: The event content to use for the join event.
|
2014-11-12 09:20:21 -07:00
|
|
|
"""
|
2020-05-22 09:11:35 -06:00
|
|
|
# TODO: We should be able to call this on workers, but the upgrading of
|
|
|
|
# room stuff after join currently doesn't work on workers.
|
2022-05-31 09:15:08 -06:00
|
|
|
# TODO: Before we relax this condition, we need to allow re-syncing of
|
|
|
|
# partial room state to happen on workers.
|
2020-05-22 09:11:35 -06:00
|
|
|
assert self.config.worker.worker_app is None
|
|
|
|
|
2014-11-25 04:31:18 -07:00
|
|
|
logger.debug("Joining %s to %s", joinee, room_id)
|
|
|
|
|
2020-02-03 09:13:13 -07:00
|
|
|
origin, event, room_version_obj = await self._make_and_verify_event(
|
2015-02-05 06:43:28 -07:00
|
|
|
target_hosts,
|
2014-10-17 08:04:17 -06:00
|
|
|
room_id,
|
2015-10-01 10:49:52 -06:00
|
|
|
joinee,
|
2015-11-12 09:19:55 -07:00
|
|
|
"join",
|
|
|
|
content,
|
2018-08-06 06:45:37 -06:00
|
|
|
params={"ver": KNOWN_ROOM_VERSIONS},
|
2014-08-20 07:42:36 -06:00
|
|
|
)
|
|
|
|
|
2017-03-14 05:26:57 -06:00
|
|
|
# This shouldn't happen, because the RoomMemberHandler has a
|
|
|
|
# linearizer lock which only allows one operation per user per room
|
|
|
|
# at a time - so this is just paranoia.
|
2021-08-26 14:41:44 -06:00
|
|
|
assert room_id not in self._federation_event_handler.room_queues
|
2017-03-14 05:26:57 -06:00
|
|
|
|
2021-08-26 14:41:44 -06:00
|
|
|
self._federation_event_handler.room_queues[room_id] = []
|
2017-03-14 05:26:57 -06:00
|
|
|
|
2023-01-22 12:19:31 -07:00
|
|
|
is_host_joined = await self.store.is_host_joined(room_id, self.server_name)
|
|
|
|
|
|
|
|
if not is_host_joined:
|
|
|
|
# We may have old forward extremities lying around if the homeserver left
|
|
|
|
# the room completely in the past. Clear them out.
|
|
|
|
#
|
|
|
|
# Note that this check-then-clear is subject to races where
|
|
|
|
# * the homeserver is in the room and stops being in the room just after
|
|
|
|
# the check. We won't reset the forward extremities, but that's okay,
|
|
|
|
# since they will be almost up to date.
|
|
|
|
# * the homeserver is not in the room and starts being in the room just
|
|
|
|
# after the check. This can't happen, since `RoomMemberHandler` has a
|
|
|
|
# linearizer lock which prevents concurrent remote joins into the same
|
|
|
|
# room.
|
|
|
|
# In short, the races either have an acceptable outcome or should be
|
|
|
|
# impossible.
|
|
|
|
await self._clean_room_for_join(room_id)
|
2017-03-14 05:26:57 -06:00
|
|
|
|
2014-10-29 10:59:24 -06:00
|
|
|
try:
|
2015-02-05 06:43:28 -07:00
|
|
|
# Try the host we successfully got a response to /make_join/
|
|
|
|
# request first.
|
2020-07-01 09:21:02 -06:00
|
|
|
host_list = list(target_hosts)
|
2015-02-06 03:53:18 -07:00
|
|
|
try:
|
2020-07-01 09:21:02 -06:00
|
|
|
host_list.remove(origin)
|
|
|
|
host_list.insert(0, origin)
|
2015-02-06 03:53:18 -07:00
|
|
|
except ValueError:
|
|
|
|
pass
|
2020-01-27 07:30:57 -07:00
|
|
|
|
2023-01-22 12:19:31 -07:00
|
|
|
async with self._is_partial_state_room_linearizer.queue(room_id):
|
|
|
|
already_partial_state_room = await self.store.is_partial_state_room(
|
|
|
|
room_id
|
|
|
|
)
|
2014-10-29 10:59:24 -06:00
|
|
|
|
2023-01-22 12:19:31 -07:00
|
|
|
ret = await self.federation_client.send_join(
|
|
|
|
host_list,
|
|
|
|
event,
|
|
|
|
room_version_obj,
|
|
|
|
# Perform a full join when we are already in the room and it is a
|
|
|
|
# full state room, since we are not allowed to persist a partial
|
|
|
|
# state join event in a full state room. In the future, we could
|
|
|
|
# optimize this by always performing a partial state join and
|
|
|
|
# computing the state ourselves or retrieving it from the remote
|
|
|
|
# homeserver if necessary.
|
|
|
|
#
|
|
|
|
# There's a race where we leave the room, then perform a full join
|
|
|
|
# anyway. This should end up being fast anyway, since we would
|
|
|
|
# already have the full room state and auth chain persisted.
|
|
|
|
partial_state=not is_host_joined or already_partial_state_room,
|
|
|
|
)
|
2014-10-29 10:59:24 -06:00
|
|
|
|
2023-01-22 12:19:31 -07:00
|
|
|
event = ret.event
|
|
|
|
origin = ret.origin
|
|
|
|
state = ret.state
|
|
|
|
auth_chain = ret.auth_chain
|
|
|
|
auth_chain.sort(key=lambda e: e.depth)
|
|
|
|
|
|
|
|
logger.debug("do_invite_join auth_chain: %s", auth_chain)
|
|
|
|
logger.debug("do_invite_join state: %s", state)
|
|
|
|
|
|
|
|
logger.debug("do_invite_join event: %s", event)
|
|
|
|
|
|
|
|
# if this is the first time we've joined this room, it's time to add
|
|
|
|
# a row to `rooms` with the correct room version. If there's already a
|
|
|
|
# row there, we should override it, since it may have been populated
|
|
|
|
# based on an invite request which lied about the room version.
|
|
|
|
#
|
|
|
|
# federation_client.send_join has already checked that the room
|
|
|
|
# version in the received create event is the same as room_version_obj,
|
|
|
|
# so we can rely on it now.
|
|
|
|
#
|
|
|
|
await self.store.upsert_room_on_join(
|
2022-09-27 10:26:35 -06:00
|
|
|
room_id=room_id,
|
2023-01-22 12:19:31 -07:00
|
|
|
room_version=room_version_obj,
|
|
|
|
state_events=state,
|
2022-09-27 10:26:35 -06:00
|
|
|
)
|
2022-03-01 05:49:54 -07:00
|
|
|
|
2023-01-22 12:19:31 -07:00
|
|
|
if ret.partial_state and not already_partial_state_room:
|
|
|
|
# Mark the room as having partial state.
|
|
|
|
# The background process is responsible for unmarking this flag,
|
|
|
|
# even if the join fails.
|
|
|
|
# TODO(faster_joins):
|
|
|
|
# We may want to reset the partial state info if it's from an
|
|
|
|
# old, failed partial state join.
|
|
|
|
# https://github.com/matrix-org/synapse/issues/13000
|
|
|
|
await self.store.store_partial_state_room(
|
|
|
|
room_id=room_id,
|
|
|
|
servers=ret.servers_in_room,
|
|
|
|
device_lists_stream_id=self.store.get_device_stream_token(),
|
|
|
|
joined_via=origin,
|
Handle race between persisting an event and un-partial stating a room (#13100)
Whenever we want to persist an event, we first compute an event context,
which includes the state at the event and a flag indicating whether the
state is partial. After a lot of processing, we finally try to store the
event in the database, which can fail for partial state events when the
containing room has been un-partial stated in the meantime.
We detect the race as a foreign key constraint failure in the data store
layer and turn it into a special `PartialStateConflictError` exception,
which makes its way up to the method in which we computed the event
context.
To make things difficult, the exception needs to cross a replication
request: `/fed_send_events` for events coming over federation and
`/send_event` for events from clients. We transport the
`PartialStateConflictError` as a `409 Conflict` over replication and
turn `409`s back into `PartialStateConflictError`s on the worker making
the request.
All client events go through
`EventCreationHandler.handle_new_client_event`, which is called in
*a lot* of places. Instead of trying to update all the code which
creates client events, we turn the `PartialStateConflictError` into a
`429 Too Many Requests` in
`EventCreationHandler.handle_new_client_event` and hope that clients
take it as a hint to retry their request.
On the federation event side, there are 7 places which compute event
contexts. 4 of them use outlier event contexts:
`FederationEventHandler._auth_and_persist_outliers_inner`,
`FederationHandler.do_knock`, `FederationHandler.on_invite_request` and
`FederationHandler.do_remotely_reject_invite`. These events won't have
the partial state flag, so we do not need to do anything for then.
The remaining 3 paths which create events are
`FederationEventHandler.process_remote_join`,
`FederationEventHandler.on_send_membership_event` and
`FederationEventHandler._process_received_pdu`.
We can't experience the race in `process_remote_join`, unless we're
handling an additional join into a partial state room, which currently
blocks, so we make no attempt to handle it correctly.
`on_send_membership_event` is only called by
`FederationServer._on_send_membership_event`, so we catch the
`PartialStateConflictError` there and retry just once.
`_process_received_pdu` is called by `on_receive_pdu` for incoming
events and `_process_pulled_event` for backfill. The latter should never
try to persist partial state events, so we ignore it. We catch the
`PartialStateConflictError` in `on_receive_pdu` and retry just once.
Refering to the graph of code paths in
https://github.com/matrix-org/synapse/issues/12988#issuecomment-1156857648
may make the above make more sense.
Signed-off-by: Sean Quah <seanq@matrix.org>
2022-07-05 09:12:52 -06:00
|
|
|
)
|
2023-01-22 12:19:31 -07:00
|
|
|
|
|
|
|
try:
|
|
|
|
max_stream_id = (
|
|
|
|
await self._federation_event_handler.process_remote_join(
|
|
|
|
origin,
|
|
|
|
room_id,
|
|
|
|
auth_chain,
|
|
|
|
state,
|
|
|
|
event,
|
|
|
|
room_version_obj,
|
|
|
|
partial_state=ret.partial_state,
|
|
|
|
)
|
2022-09-27 10:26:35 -06:00
|
|
|
)
|
2023-01-22 12:19:31 -07:00
|
|
|
except PartialStateConflictError:
|
|
|
|
# This should be impossible, since we hold the lock on the room's
|
|
|
|
# partial statedness.
|
|
|
|
logger.error(
|
|
|
|
"Room %s was un-partial stated while processing remote join.",
|
|
|
|
room_id,
|
2022-08-01 10:45:39 -06:00
|
|
|
)
|
2023-01-22 12:19:31 -07:00
|
|
|
raise
|
|
|
|
else:
|
|
|
|
# Record the join event id for future use (when we finish the full
|
|
|
|
# join). We have to do this after persisting the event to keep
|
|
|
|
# foreign key constraints intact.
|
|
|
|
if ret.partial_state and not already_partial_state_room:
|
|
|
|
# TODO(faster_joins):
|
|
|
|
# We may want to reset the partial state info if it's from
|
|
|
|
# an old, failed partial state join.
|
|
|
|
# https://github.com/matrix-org/synapse/issues/13000
|
|
|
|
await self.store.write_partial_state_rooms_join_event_id(
|
|
|
|
room_id, event.event_id
|
|
|
|
)
|
|
|
|
finally:
|
|
|
|
# Always kick off the background process that asynchronously fetches
|
|
|
|
# state for the room.
|
|
|
|
# If the join failed, the background process is responsible for
|
|
|
|
# cleaning up — including unmarking the room as a partial state
|
|
|
|
# room.
|
|
|
|
if ret.partial_state:
|
|
|
|
# Kick off the process of asynchronously fetching the state for
|
|
|
|
# this room.
|
|
|
|
self._start_partial_state_room_sync(
|
|
|
|
initial_destination=origin,
|
|
|
|
other_destinations=ret.servers_in_room,
|
|
|
|
room_id=room_id,
|
|
|
|
)
|
2022-04-12 07:23:43 -06:00
|
|
|
|
2020-05-22 07:21:54 -06:00
|
|
|
# We wait here until this instance has seen the events come down
|
|
|
|
# replication (if we're using replication) as the below uses caches.
|
|
|
|
await self._replication.wait_for_stream_position(
|
2020-09-14 03:16:41 -06:00
|
|
|
self.config.worker.events_shard_config.get_instance(room_id),
|
|
|
|
"events",
|
|
|
|
max_stream_id,
|
2020-05-22 07:21:54 -06:00
|
|
|
)
|
|
|
|
|
2019-11-01 04:28:09 -06:00
|
|
|
# Check whether this room is the result of an upgrade of a room we already know
|
|
|
|
# about. If so, migrate over user information
|
2020-02-03 09:13:13 -07:00
|
|
|
predecessor = await self.store.get_room_predecessor(room_id)
|
2019-12-11 06:07:25 -07:00
|
|
|
if not predecessor or not isinstance(predecessor.get("room_id"), str):
|
2020-05-22 07:21:54 -06:00
|
|
|
return event.event_id, max_stream_id
|
2019-11-01 04:28:09 -06:00
|
|
|
old_room_id = predecessor["room_id"]
|
|
|
|
logger.debug(
|
|
|
|
"Found predecessor for %s during remote join: %s", room_id, old_room_id
|
|
|
|
)
|
|
|
|
|
|
|
|
# We retrieve the room member handler here as to not cause a cyclic dependency
|
|
|
|
member_handler = self.hs.get_room_member_handler()
|
2020-02-03 09:13:13 -07:00
|
|
|
await member_handler.transfer_room_state_on_room_upgrade(
|
2019-11-01 04:28:09 -06:00
|
|
|
old_room_id, room_id
|
|
|
|
)
|
|
|
|
|
2014-11-25 04:31:18 -07:00
|
|
|
logger.debug("Finished joining %s to %s", joinee, room_id)
|
2020-05-22 07:21:54 -06:00
|
|
|
return event.event_id, max_stream_id
|
2014-10-29 10:59:24 -06:00
|
|
|
finally:
|
2021-08-26 14:41:44 -06:00
|
|
|
room_queue = self._federation_event_handler.room_queues[room_id]
|
|
|
|
del self._federation_event_handler.room_queues[room_id]
|
2014-10-17 11:56:42 -06:00
|
|
|
|
2017-03-14 05:26:57 -06:00
|
|
|
# we don't need to wait for the queued events to be processed -
|
|
|
|
# it's just a best-effort thing at this point. We do want to do
|
|
|
|
# them roughly in order, though, otherwise we'll end up making
|
|
|
|
# lots of requests for missing prev_events which we do actually
|
2020-07-24 08:53:25 -06:00
|
|
|
# have. Hence we fire off the background task, but don't wait for it.
|
2014-12-10 08:55:03 -07:00
|
|
|
|
2022-02-22 06:33:22 -07:00
|
|
|
run_as_background_process(
|
|
|
|
"handle_queued_pdus", self._handle_queued_pdus, room_queue
|
|
|
|
)
|
2014-10-17 11:56:42 -06:00
|
|
|
|
2021-06-09 12:39:51 -06:00
|
|
|
async def do_knock(
|
|
|
|
self,
|
|
|
|
target_hosts: List[str],
|
|
|
|
room_id: str,
|
|
|
|
knockee: str,
|
|
|
|
content: JsonDict,
|
|
|
|
) -> Tuple[str, int]:
|
|
|
|
"""Sends the knock to the remote server.
|
|
|
|
|
|
|
|
This first triggers a make_knock request that returns a partial
|
|
|
|
event that we can fill out and sign. This is then sent to the
|
|
|
|
remote server via send_knock.
|
|
|
|
|
|
|
|
Knock events must be signed by the knockee's server before distributing.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
target_hosts: A list of hosts that we want to try knocking through.
|
|
|
|
room_id: The ID of the room to knock on.
|
|
|
|
knockee: The ID of the user who is knocking.
|
|
|
|
content: The content of the knock event.
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
A tuple of (event ID, stream ID).
|
|
|
|
|
|
|
|
Raises:
|
|
|
|
SynapseError: If the chosen remote server returns a 3xx/4xx code.
|
|
|
|
RuntimeError: If no servers were reachable.
|
|
|
|
"""
|
|
|
|
logger.debug("Knocking on room %s on behalf of user %s", room_id, knockee)
|
|
|
|
|
|
|
|
# Inform the remote server of the room versions we support
|
|
|
|
supported_room_versions = list(KNOWN_ROOM_VERSIONS.keys())
|
|
|
|
|
|
|
|
# Ask the remote server to create a valid knock event for us. Once received,
|
|
|
|
# we sign the event
|
2021-07-16 11:22:36 -06:00
|
|
|
params: Dict[str, Iterable[str]] = {"ver": supported_room_versions}
|
2021-06-09 12:39:51 -06:00
|
|
|
origin, event, event_format_version = await self._make_and_verify_event(
|
|
|
|
target_hosts, room_id, knockee, Membership.KNOCK, content, params=params
|
|
|
|
)
|
|
|
|
|
2021-09-22 08:20:18 -06:00
|
|
|
# Mark the knock as an outlier as we don't yet have the state at this point in
|
|
|
|
# the DAG.
|
|
|
|
event.internal_metadata.outlier = True
|
|
|
|
|
|
|
|
# ... but tell /sync to send it to clients anyway.
|
|
|
|
event.internal_metadata.out_of_band_membership = True
|
|
|
|
|
2021-06-09 12:39:51 -06:00
|
|
|
# Record the room ID and its version so that we have a record of the room
|
|
|
|
await self._maybe_store_room_on_outlier_membership(
|
|
|
|
room_id=event.room_id, room_version=event_format_version
|
|
|
|
)
|
|
|
|
|
|
|
|
# Initially try the host that we successfully called /make_knock on
|
|
|
|
try:
|
|
|
|
target_hosts.remove(origin)
|
|
|
|
target_hosts.insert(0, origin)
|
|
|
|
except ValueError:
|
|
|
|
pass
|
|
|
|
|
|
|
|
# Send the signed event back to the room, and potentially receive some
|
|
|
|
# further information about the room in the form of partial state events
|
2022-10-12 07:37:20 -06:00
|
|
|
knock_response = await self.federation_client.send_knock(target_hosts, event)
|
2021-06-09 12:39:51 -06:00
|
|
|
|
|
|
|
# Store any stripped room state events in the "unsigned" key of the event.
|
|
|
|
# This is a bit of a hack and is cribbing off of invites. Basically we
|
|
|
|
# store the room state here and retrieve it again when this event appears
|
|
|
|
# in the invitee's sync stream. It is stripped out for all other local users.
|
2023-10-06 05:27:35 -06:00
|
|
|
stripped_room_state = knock_response.get("knock_room_state")
|
2022-10-12 07:37:20 -06:00
|
|
|
|
|
|
|
if stripped_room_state is None:
|
2023-10-06 05:27:35 -06:00
|
|
|
raise KeyError("Missing 'knock_room_state' field in send_knock response")
|
2022-10-12 07:37:20 -06:00
|
|
|
|
|
|
|
event.unsigned["knock_room_state"] = stripped_room_state
|
2021-06-09 12:39:51 -06:00
|
|
|
|
2022-05-31 06:17:50 -06:00
|
|
|
context = EventContext.for_outlier(self._storage_controllers)
|
2021-08-26 14:41:44 -06:00
|
|
|
stream_id = await self._federation_event_handler.persist_events_and_notify(
|
2021-06-09 12:39:51 -06:00
|
|
|
event.room_id, [(event, context)]
|
|
|
|
)
|
|
|
|
return event.event_id, stream_id
|
|
|
|
|
2021-04-06 05:21:57 -06:00
|
|
|
async def _handle_queued_pdus(
|
|
|
|
self, room_queue: List[Tuple[EventBase, str]]
|
|
|
|
) -> None:
|
2017-03-14 05:26:57 -06:00
|
|
|
"""Process PDUs which got queued up while we were busy send_joining.
|
|
|
|
|
|
|
|
Args:
|
2021-04-06 05:21:57 -06:00
|
|
|
room_queue: list of PDUs to be processed and the servers that sent them
|
2017-03-14 05:26:57 -06:00
|
|
|
"""
|
|
|
|
for p, origin in room_queue:
|
|
|
|
try:
|
|
|
|
logger.info(
|
2021-08-16 06:19:02 -06:00
|
|
|
"Processing queued PDU %s which was received while we were joining",
|
|
|
|
p,
|
2017-03-14 05:26:57 -06:00
|
|
|
)
|
2019-07-03 08:07:04 -06:00
|
|
|
with nested_logging_context(p.event_id):
|
2021-08-26 14:41:44 -06:00
|
|
|
await self._federation_event_handler.on_receive_pdu(origin, p)
|
2017-03-14 05:26:57 -06:00
|
|
|
except Exception as e:
|
2019-10-31 04:23:24 -06:00
|
|
|
logger.warning(
|
2017-03-14 05:26:57 -06:00
|
|
|
"Error handling queued PDU %s from %s: %s", p.event_id, origin, e
|
|
|
|
)
|
|
|
|
|
2020-02-03 08:35:30 -07:00
|
|
|
async def on_make_join_request(
|
|
|
|
self, origin: str, room_id: str, user_id: str
|
|
|
|
) -> EventBase:
|
2014-11-12 09:20:21 -07:00
|
|
|
"""We've received a /make_join/ request, so we create a partial
|
2015-10-20 04:58:58 -06:00
|
|
|
join event for the room and return that. We do *not* persist or
|
2014-11-12 09:20:21 -07:00
|
|
|
process it until the other server has signed it and sent it back.
|
2019-07-26 03:08:22 -06:00
|
|
|
|
|
|
|
Args:
|
2020-02-03 08:35:30 -07:00
|
|
|
origin: The (verified) server name of the requesting server.
|
|
|
|
room_id: Room to create join event in
|
|
|
|
user_id: The user to create the join for
|
2014-11-12 09:20:21 -07:00
|
|
|
"""
|
2019-07-26 03:08:22 -06:00
|
|
|
if get_domain_from_id(user_id) != origin:
|
|
|
|
logger.info(
|
|
|
|
"Got /make_join request for user %r from different origin %s, ignoring",
|
|
|
|
user_id,
|
|
|
|
origin,
|
|
|
|
)
|
|
|
|
raise SynapseError(403, "User not from origin", Codes.FORBIDDEN)
|
|
|
|
|
2020-07-16 08:17:31 -06:00
|
|
|
# checking the room version will check that we've actually heard of the room
|
|
|
|
# (and return a 404 otherwise)
|
2021-07-26 10:17:00 -06:00
|
|
|
room_version = await self.store.get_room_version(room_id)
|
2019-01-23 13:21:33 -07:00
|
|
|
|
2022-08-04 09:27:04 -06:00
|
|
|
if await self.store.is_partial_state_room(room_id):
|
|
|
|
# If our server is still only partially joined, we can't give a complete
|
|
|
|
# response to /make_join, so return a 404 as we would if we weren't in the
|
|
|
|
# room at all.
|
|
|
|
# The main reason we can't respond properly is that we need to know about
|
|
|
|
# the auth events for the join event that we would return.
|
|
|
|
# We also should not bother entertaining the /make_join since we cannot
|
|
|
|
# handle the /send_join.
|
|
|
|
logger.info(
|
|
|
|
"Rejecting /make_join to %s because it's a partial state room", room_id
|
|
|
|
)
|
|
|
|
raise SynapseError(
|
|
|
|
404,
|
|
|
|
"Unable to handle /make_join right now; this server is not fully joined.",
|
|
|
|
errcode=Codes.NOT_FOUND,
|
|
|
|
)
|
|
|
|
|
2020-07-16 08:17:31 -06:00
|
|
|
# now check that we are *still* in the room
|
2022-09-23 04:47:16 -06:00
|
|
|
is_in_room = await self._event_auth_handler.is_host_in_room(
|
2021-07-01 12:25:37 -06:00
|
|
|
room_id, self.server_name
|
|
|
|
)
|
2020-07-16 08:17:31 -06:00
|
|
|
if not is_in_room:
|
|
|
|
logger.info(
|
|
|
|
"Got /make_join request for room %s we are no longer in",
|
|
|
|
room_id,
|
|
|
|
)
|
|
|
|
raise NotFoundError("Not an active room on this server")
|
|
|
|
|
|
|
|
event_content = {"membership": Membership.JOIN}
|
|
|
|
|
2021-07-26 10:17:00 -06:00
|
|
|
# If the current room is using restricted join rules, additional information
|
|
|
|
# may need to be included in the event content in order to efficiently
|
|
|
|
# validate the event.
|
|
|
|
#
|
|
|
|
# Note that this requires the /send_join request to come back to the
|
|
|
|
# same server.
|
Mitigate a race where /make_join could 403 for restricted rooms (#15080)
Previously, when creating a join event in /make_join, we would decide
whether to include additional fields to satisfy restricted room checks
based on the current state of the room. Then, when building the event,
we would capture the forward extremities of the room to use as prev
events.
This is subject to race conditions. For example, when leaving and
rejoining a room, the following sequence of events leads to a misleading
403 response:
1. /make_join reads the current state of the room and sees that the user
is still in the room. It decides to omit the field required for
restricted room joins.
2. The leave event is persisted and the room's forward extremities are
updated.
3. /make_join builds the event, using the post-leave forward extremities.
The event then fails the restricted room checks.
To mitigate the race, we move the read of the forward extremities closer
to the read of the current state. Ideally, we would compute the state
based off the chosen prev events, but that can involve state resolution,
which is expensive.
Signed-off-by: Sean Quah <seanq@matrix.org>
2023-02-17 02:40:32 -07:00
|
|
|
prev_event_ids = None
|
2023-07-18 06:44:59 -06:00
|
|
|
if room_version.restricted_join_rule:
|
Mitigate a race where /make_join could 403 for restricted rooms (#15080)
Previously, when creating a join event in /make_join, we would decide
whether to include additional fields to satisfy restricted room checks
based on the current state of the room. Then, when building the event,
we would capture the forward extremities of the room to use as prev
events.
This is subject to race conditions. For example, when leaving and
rejoining a room, the following sequence of events leads to a misleading
403 response:
1. /make_join reads the current state of the room and sees that the user
is still in the room. It decides to omit the field required for
restricted room joins.
2. The leave event is persisted and the room's forward extremities are
updated.
3. /make_join builds the event, using the post-leave forward extremities.
The event then fails the restricted room checks.
To mitigate the race, we move the read of the forward extremities closer
to the read of the current state. Ideally, we would compute the state
based off the chosen prev events, but that can involve state resolution,
which is expensive.
Signed-off-by: Sean Quah <seanq@matrix.org>
2023-02-17 02:40:32 -07:00
|
|
|
# Note that the room's state can change out from under us and render our
|
|
|
|
# nice join rules-conformant event non-conformant by the time we build the
|
|
|
|
# event. When this happens, our validation at the end fails and we respond
|
|
|
|
# to the requesting server with a 403, which is misleading — it indicates
|
|
|
|
# that the user is not allowed to join the room and the joining server
|
|
|
|
# should not bother retrying via this homeserver or any others, when
|
|
|
|
# in fact we've just messed up with building the event.
|
|
|
|
#
|
|
|
|
# To reduce the likelihood of this race, we capture the forward extremities
|
|
|
|
# of the room (prev_event_ids) just before fetching the current state, and
|
|
|
|
# hope that the state we fetch corresponds to the prev events we chose.
|
|
|
|
prev_event_ids = await self.store.get_prev_events_for_room(room_id)
|
2022-06-01 09:02:53 -06:00
|
|
|
state_ids = await self._state_storage_controller.get_current_state_ids(
|
|
|
|
room_id
|
|
|
|
)
|
2021-07-26 10:17:00 -06:00
|
|
|
if await self._event_auth_handler.has_restricted_join_rules(
|
|
|
|
state_ids, room_version
|
|
|
|
):
|
|
|
|
prev_member_event_id = state_ids.get((EventTypes.Member, user_id), None)
|
|
|
|
# If the user is invited or joined to the room already, then
|
|
|
|
# no additional info is needed.
|
|
|
|
include_auth_user_id = True
|
|
|
|
if prev_member_event_id:
|
|
|
|
prev_member_event = await self.store.get_event(prev_member_event_id)
|
|
|
|
include_auth_user_id = prev_member_event.membership not in (
|
|
|
|
Membership.JOIN,
|
|
|
|
Membership.INVITE,
|
|
|
|
)
|
|
|
|
|
|
|
|
if include_auth_user_id:
|
|
|
|
event_content[
|
2021-09-30 09:13:59 -06:00
|
|
|
EventContentFields.AUTHORISING_USER
|
2021-07-26 10:17:00 -06:00
|
|
|
] = await self._event_auth_handler.get_user_which_could_invite(
|
|
|
|
room_id,
|
|
|
|
state_ids,
|
|
|
|
)
|
|
|
|
|
2021-09-29 03:57:10 -06:00
|
|
|
builder = self.event_builder_factory.for_room_version(
|
|
|
|
room_version,
|
2019-01-23 13:21:33 -07:00
|
|
|
{
|
|
|
|
"type": EventTypes.Member,
|
|
|
|
"content": event_content,
|
|
|
|
"room_id": room_id,
|
|
|
|
"sender": user_id,
|
|
|
|
"state_key": user_id,
|
|
|
|
},
|
|
|
|
)
|
2014-12-04 08:50:01 -07:00
|
|
|
|
2016-04-13 04:11:46 -06:00
|
|
|
try:
|
2023-02-09 14:05:02 -07:00
|
|
|
(
|
|
|
|
event,
|
|
|
|
unpersisted_context,
|
|
|
|
) = await self.event_creation_handler.create_new_client_event(
|
Mitigate a race where /make_join could 403 for restricted rooms (#15080)
Previously, when creating a join event in /make_join, we would decide
whether to include additional fields to satisfy restricted room checks
based on the current state of the room. Then, when building the event,
we would capture the forward extremities of the room to use as prev
events.
This is subject to race conditions. For example, when leaving and
rejoining a room, the following sequence of events leads to a misleading
403 response:
1. /make_join reads the current state of the room and sees that the user
is still in the room. It decides to omit the field required for
restricted room joins.
2. The leave event is persisted and the room's forward extremities are
updated.
3. /make_join builds the event, using the post-leave forward extremities.
The event then fails the restricted room checks.
To mitigate the race, we move the read of the forward extremities closer
to the read of the current state. Ideally, we would compute the state
based off the chosen prev events, but that can involve state resolution,
which is expensive.
Signed-off-by: Sean Quah <seanq@matrix.org>
2023-02-17 02:40:32 -07:00
|
|
|
builder=builder,
|
|
|
|
prev_event_ids=prev_event_ids,
|
2016-04-13 04:11:46 -06:00
|
|
|
)
|
2020-10-13 08:44:54 -06:00
|
|
|
except SynapseError as e:
|
2019-10-31 04:23:24 -06:00
|
|
|
logger.warning("Failed to create join to %s because %s", room_id, e)
|
2020-10-13 08:44:54 -06:00
|
|
|
raise
|
2019-06-12 03:31:37 -06:00
|
|
|
|
2021-07-26 10:17:00 -06:00
|
|
|
# Ensure the user can even join the room.
|
2023-02-09 14:05:02 -07:00
|
|
|
await self._federation_event_handler.check_join_restrictions(
|
|
|
|
unpersisted_context, event
|
|
|
|
)
|
2021-07-26 10:17:00 -06:00
|
|
|
|
2016-07-15 02:29:54 -06:00
|
|
|
# The remote hasn't signed it yet, obviously. We'll do the full checks
|
|
|
|
# when we get the event back in `on_send_join_request`
|
2022-10-18 10:58:47 -06:00
|
|
|
await self._event_auth_handler.check_auth_rules_from_context(event)
|
2019-07-23 07:00:55 -06:00
|
|
|
return event
|
2014-10-16 09:56:51 -06:00
|
|
|
|
2020-02-03 08:33:42 -07:00
|
|
|
async def on_invite_request(
|
2020-01-30 15:13:02 -07:00
|
|
|
self, origin: str, event: EventBase, room_version: RoomVersion
|
2021-04-06 05:21:57 -06:00
|
|
|
) -> EventBase:
|
2014-11-12 09:20:21 -07:00
|
|
|
"""We've got an invite event. Process and persist it. Sign it.
|
|
|
|
|
|
|
|
Respond with the now signed event.
|
|
|
|
"""
|
2017-10-05 07:02:28 -06:00
|
|
|
if event.state_key is None:
|
|
|
|
raise SynapseError(400, "The invite event did not have a state key")
|
|
|
|
|
2020-02-03 08:33:42 -07:00
|
|
|
is_blocked = await self.store.is_room_blocked(event.room_id)
|
2017-06-19 05:36:28 -06:00
|
|
|
if is_blocked:
|
|
|
|
raise SynapseError(403, "This room has been blocked on this server")
|
|
|
|
|
2021-09-29 04:44:15 -06:00
|
|
|
if self.hs.config.server.block_non_admin_invites:
|
2017-09-19 09:08:14 -06:00
|
|
|
raise SynapseError(403, "This server does not accept room invites")
|
|
|
|
|
2023-04-17 18:57:40 -06:00
|
|
|
spam_check = await self._spam_checker_module_callbacks.user_may_invite(
|
2017-10-05 07:02:28 -06:00
|
|
|
event.sender, event.state_key, event.room_id
|
2022-06-13 12:16:16 -06:00
|
|
|
)
|
|
|
|
if spam_check != NOT_SPAM:
|
2017-10-03 07:04:10 -06:00
|
|
|
raise SynapseError(
|
2022-06-13 12:16:16 -06:00
|
|
|
403,
|
|
|
|
"This user is not permitted to send invites to this server/user",
|
2022-07-11 10:52:10 -06:00
|
|
|
errcode=spam_check[0],
|
|
|
|
additional_fields=spam_check[1],
|
2017-10-03 07:04:10 -06:00
|
|
|
)
|
2017-10-03 06:53:09 -06:00
|
|
|
|
2017-06-30 09:20:30 -06:00
|
|
|
membership = event.content.get("membership")
|
|
|
|
if event.type != EventTypes.Member or membership != Membership.INVITE:
|
|
|
|
raise SynapseError(400, "The event was not an m.room.member invite event")
|
|
|
|
|
|
|
|
sender_domain = get_domain_from_id(event.sender)
|
|
|
|
if sender_domain != origin:
|
|
|
|
raise SynapseError(
|
|
|
|
400, "The invite event was not from the server sending it"
|
|
|
|
)
|
|
|
|
|
|
|
|
if not self.is_mine_id(event.state_key):
|
|
|
|
raise SynapseError(400, "The invite event must be for this server")
|
|
|
|
|
2018-05-18 04:18:39 -06:00
|
|
|
# block any attempts to invite the server notices mxid
|
|
|
|
if event.state_key == self._server_notices_mxid:
|
2020-06-16 06:51:47 -06:00
|
|
|
raise SynapseError(HTTPStatus.FORBIDDEN, "Cannot invite this user")
|
2018-05-18 04:18:39 -06:00
|
|
|
|
2021-01-29 09:38:29 -07:00
|
|
|
# We retrieve the room member handler here as to not cause a cyclic dependency
|
|
|
|
member_handler = self.hs.get_room_member_handler()
|
2021-02-03 03:17:37 -07:00
|
|
|
# We don't rate limit based on room ID, as that should be done by
|
|
|
|
# sending server.
|
2021-03-30 05:06:09 -06:00
|
|
|
await member_handler.ratelimit_invite(None, None, event.state_key)
|
2021-01-29 09:38:29 -07:00
|
|
|
|
2020-02-26 09:58:33 -07:00
|
|
|
# keep a record of the room version, if we don't yet know it.
|
|
|
|
# (this may get overwritten if we later get a different room version in a
|
|
|
|
# join dance).
|
2020-11-13 09:24:04 -07:00
|
|
|
await self._maybe_store_room_on_outlier_membership(
|
2020-02-26 09:58:33 -07:00
|
|
|
room_id=event.room_id, room_version=room_version
|
|
|
|
)
|
|
|
|
|
2014-12-05 09:20:48 -07:00
|
|
|
event.internal_metadata.outlier = True
|
2019-01-24 10:33:19 -07:00
|
|
|
event.internal_metadata.out_of_band_membership = True
|
2014-11-07 06:41:00 -07:00
|
|
|
|
|
|
|
event.signatures.update(
|
|
|
|
compute_event_signature(
|
2020-01-31 06:47:43 -07:00
|
|
|
room_version,
|
|
|
|
event.get_pdu_json(),
|
|
|
|
self.hs.hostname,
|
2020-07-08 10:51:56 -06:00
|
|
|
self.hs.signing_key,
|
2014-11-07 06:41:00 -07:00
|
|
|
)
|
|
|
|
)
|
|
|
|
|
2022-05-31 06:17:50 -06:00
|
|
|
context = EventContext.for_outlier(self._storage_controllers)
|
2022-09-28 06:31:53 -06:00
|
|
|
|
2022-10-21 11:46:22 -06:00
|
|
|
await self._bulk_push_rule_evaluator.action_for_events_by_user(
|
|
|
|
[(event, context)]
|
|
|
|
)
|
2022-09-28 06:31:53 -06:00
|
|
|
try:
|
|
|
|
await self._federation_event_handler.persist_events_and_notify(
|
|
|
|
event.room_id, [(event, context)]
|
|
|
|
)
|
|
|
|
except Exception:
|
|
|
|
await self.store.remove_push_actions_from_staging(event.event_id)
|
|
|
|
raise
|
2014-11-07 06:41:00 -07:00
|
|
|
|
2019-07-23 07:00:55 -06:00
|
|
|
return event
|
2014-11-07 06:41:00 -07:00
|
|
|
|
2020-02-03 09:19:18 -07:00
|
|
|
async def do_remotely_reject_invite(
|
|
|
|
self, target_hosts: Iterable[str], room_id: str, user_id: str, content: JsonDict
|
2020-05-22 07:21:54 -06:00
|
|
|
) -> Tuple[EventBase, int]:
|
2020-02-03 09:19:18 -07:00
|
|
|
origin, event, room_version = await self._make_and_verify_event(
|
2019-12-11 06:07:25 -07:00
|
|
|
target_hosts, room_id, user_id, "leave", content=content
|
2017-04-20 18:24:17 -06:00
|
|
|
)
|
2017-06-09 06:05:05 -06:00
|
|
|
# Mark as outlier as we don't have any state for this event; we're not
|
|
|
|
# even in the room.
|
2017-06-09 03:08:18 -06:00
|
|
|
event.internal_metadata.outlier = True
|
2019-01-24 10:33:19 -07:00
|
|
|
event.internal_metadata.out_of_band_membership = True
|
2015-10-20 04:58:58 -06:00
|
|
|
|
2020-10-23 10:38:40 -06:00
|
|
|
# Try the host that we successfully called /make_leave/ on first for
|
2017-04-07 07:39:32 -06:00
|
|
|
# the /send_leave/ request.
|
2020-07-01 09:21:02 -06:00
|
|
|
host_list = list(target_hosts)
|
2015-10-20 04:58:58 -06:00
|
|
|
try:
|
2020-07-01 09:21:02 -06:00
|
|
|
host_list.remove(origin)
|
|
|
|
host_list.insert(0, origin)
|
2015-10-20 04:58:58 -06:00
|
|
|
except ValueError:
|
|
|
|
pass
|
|
|
|
|
2020-07-01 09:21:02 -06:00
|
|
|
await self.federation_client.send_leave(host_list, event)
|
2016-03-15 07:24:31 -06:00
|
|
|
|
2022-05-31 06:17:50 -06:00
|
|
|
context = EventContext.for_outlier(self._storage_controllers)
|
2021-08-26 14:41:44 -06:00
|
|
|
stream_id = await self._federation_event_handler.persist_events_and_notify(
|
2020-09-14 03:16:41 -06:00
|
|
|
event.room_id, [(event, context)]
|
|
|
|
)
|
2016-03-15 07:24:31 -06:00
|
|
|
|
2020-05-22 07:21:54 -06:00
|
|
|
return event, stream_id
|
2015-10-20 04:58:58 -06:00
|
|
|
|
2020-02-03 09:22:30 -07:00
|
|
|
async def _make_and_verify_event(
|
|
|
|
self,
|
|
|
|
target_hosts: Iterable[str],
|
|
|
|
room_id: str,
|
|
|
|
user_id: str,
|
|
|
|
membership: str,
|
2021-04-08 15:38:54 -06:00
|
|
|
content: JsonDict,
|
2020-07-01 09:21:02 -06:00
|
|
|
params: Optional[Dict[str, Union[str, Iterable[str]]]] = None,
|
2020-02-03 09:22:30 -07:00
|
|
|
) -> Tuple[str, EventBase, RoomVersion]:
|
2020-01-27 07:30:57 -07:00
|
|
|
(
|
|
|
|
origin,
|
|
|
|
event,
|
|
|
|
room_version,
|
2020-02-03 09:22:30 -07:00
|
|
|
) = await self.federation_client.make_membership_event(
|
2018-08-06 06:45:37 -06:00
|
|
|
target_hosts, room_id, user_id, membership, content, params=params
|
2015-10-20 04:58:58 -06:00
|
|
|
)
|
|
|
|
|
2019-01-23 13:21:33 -07:00
|
|
|
logger.debug("Got response to make_%s: %s", membership, event)
|
2015-10-20 04:58:58 -06:00
|
|
|
|
|
|
|
# We should assert some things.
|
|
|
|
# FIXME: Do this in a nicer way
|
|
|
|
assert event.type == EventTypes.Member
|
|
|
|
assert event.user_id == user_id
|
|
|
|
assert event.state_key == user_id
|
|
|
|
assert event.room_id == room_id
|
2020-01-27 07:30:57 -07:00
|
|
|
return origin, event, room_version
|
2015-10-20 04:58:58 -06:00
|
|
|
|
2020-02-03 08:40:41 -07:00
|
|
|
async def on_make_leave_request(
|
|
|
|
self, origin: str, room_id: str, user_id: str
|
|
|
|
) -> EventBase:
|
2015-10-20 04:58:58 -06:00
|
|
|
"""We've received a /make_leave/ request, so we create a partial
|
2018-07-25 15:44:41 -06:00
|
|
|
leave event for the room and return that. We do *not* persist or
|
2015-10-20 04:58:58 -06:00
|
|
|
process it until the other server has signed it and sent it back.
|
2019-07-26 03:08:22 -06:00
|
|
|
|
|
|
|
Args:
|
2020-02-03 08:40:41 -07:00
|
|
|
origin: The (verified) server name of the requesting server.
|
|
|
|
room_id: Room to create leave event in
|
|
|
|
user_id: The user to create the leave for
|
2015-10-20 04:58:58 -06:00
|
|
|
"""
|
2019-07-26 03:08:22 -06:00
|
|
|
if get_domain_from_id(user_id) != origin:
|
|
|
|
logger.info(
|
|
|
|
"Got /make_leave request for user %r from different origin %s, ignoring",
|
|
|
|
user_id,
|
|
|
|
origin,
|
|
|
|
)
|
|
|
|
raise SynapseError(403, "User not from origin", Codes.FORBIDDEN)
|
|
|
|
|
2021-09-29 03:57:10 -06:00
|
|
|
room_version_obj = await self.store.get_room_version(room_id)
|
|
|
|
builder = self.event_builder_factory.for_room_version(
|
|
|
|
room_version_obj,
|
2019-01-23 13:21:33 -07:00
|
|
|
{
|
|
|
|
"type": EventTypes.Member,
|
|
|
|
"content": {"membership": Membership.LEAVE},
|
|
|
|
"room_id": room_id,
|
|
|
|
"sender": user_id,
|
|
|
|
"state_key": user_id,
|
|
|
|
},
|
|
|
|
)
|
2015-10-20 04:58:58 -06:00
|
|
|
|
2023-02-09 14:05:02 -07:00
|
|
|
event, _ = await self.event_creation_handler.create_new_client_event(
|
2015-10-20 04:58:58 -06:00
|
|
|
builder=builder
|
|
|
|
)
|
|
|
|
|
2016-04-13 04:11:46 -06:00
|
|
|
try:
|
2016-07-15 02:29:54 -06:00
|
|
|
# The remote hasn't signed it yet, obviously. We'll do the full checks
|
|
|
|
# when we get the event back in `on_send_leave_request`
|
2022-10-18 10:58:47 -06:00
|
|
|
await self._event_auth_handler.check_auth_rules_from_context(event)
|
2016-04-13 04:11:46 -06:00
|
|
|
except AuthError as e:
|
2019-10-31 04:23:24 -06:00
|
|
|
logger.warning("Failed to create new leave %r because %s", event, e)
|
2016-04-13 04:11:46 -06:00
|
|
|
raise e
|
2015-10-20 04:58:58 -06:00
|
|
|
|
2019-07-23 07:00:55 -06:00
|
|
|
return event
|
2015-10-20 04:58:58 -06:00
|
|
|
|
2021-06-09 12:39:51 -06:00
|
|
|
async def on_make_knock_request(
|
|
|
|
self, origin: str, room_id: str, user_id: str
|
|
|
|
) -> EventBase:
|
|
|
|
"""We've received a make_knock request, so we create a partial
|
|
|
|
knock event for the room and return that. We do *not* persist or
|
|
|
|
process it until the other server has signed it and sent it back.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
origin: The (verified) server name of the requesting server.
|
|
|
|
room_id: The room to create the knock event in.
|
|
|
|
user_id: The user to create the knock for.
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
The partial knock event.
|
|
|
|
"""
|
|
|
|
if get_domain_from_id(user_id) != origin:
|
|
|
|
logger.info(
|
2021-06-15 05:45:14 -06:00
|
|
|
"Get /make_knock request for user %r from different origin %s, ignoring",
|
2021-06-09 12:39:51 -06:00
|
|
|
user_id,
|
|
|
|
origin,
|
|
|
|
)
|
|
|
|
raise SynapseError(403, "User not from origin", Codes.FORBIDDEN)
|
|
|
|
|
2021-09-29 03:57:10 -06:00
|
|
|
room_version_obj = await self.store.get_room_version(room_id)
|
2021-06-09 12:39:51 -06:00
|
|
|
|
2021-09-29 03:57:10 -06:00
|
|
|
builder = self.event_builder_factory.for_room_version(
|
|
|
|
room_version_obj,
|
2021-06-09 12:39:51 -06:00
|
|
|
{
|
|
|
|
"type": EventTypes.Member,
|
|
|
|
"content": {"membership": Membership.KNOCK},
|
|
|
|
"room_id": room_id,
|
|
|
|
"sender": user_id,
|
|
|
|
"state_key": user_id,
|
|
|
|
},
|
|
|
|
)
|
|
|
|
|
2023-02-09 14:05:02 -07:00
|
|
|
(
|
|
|
|
event,
|
|
|
|
unpersisted_context,
|
|
|
|
) = await self.event_creation_handler.create_new_client_event(builder=builder)
|
2021-06-09 12:39:51 -06:00
|
|
|
|
2023-05-04 08:18:22 -06:00
|
|
|
event_allowed, _ = await self._third_party_event_rules.check_event_allowed(
|
2023-02-09 14:05:02 -07:00
|
|
|
event, unpersisted_context
|
2021-06-09 12:39:51 -06:00
|
|
|
)
|
|
|
|
if not event_allowed:
|
|
|
|
logger.warning("Creation of knock %s forbidden by third-party rules", event)
|
|
|
|
raise SynapseError(
|
|
|
|
403, "This event is not allowed in this context", Codes.FORBIDDEN
|
|
|
|
)
|
|
|
|
|
|
|
|
try:
|
|
|
|
# The remote hasn't signed it yet, obviously. We'll do the full checks
|
|
|
|
# when we get the event back in `on_send_knock_request`
|
2022-10-18 10:58:47 -06:00
|
|
|
await self._event_auth_handler.check_auth_rules_from_context(event)
|
2021-06-09 12:39:51 -06:00
|
|
|
except AuthError as e:
|
|
|
|
logger.warning("Failed to create new knock %r because %s", event, e)
|
|
|
|
raise e
|
|
|
|
|
|
|
|
return event
|
|
|
|
|
2022-08-15 12:41:23 -06:00
|
|
|
@trace
|
|
|
|
@tag_args
|
2020-04-24 12:36:38 -06:00
|
|
|
async def get_state_ids_for_pdu(self, room_id: str, event_id: str) -> List[str]:
|
2016-09-02 07:19:22 -06:00
|
|
|
"""Returns the state at the event. i.e. not including said event."""
|
2020-08-18 14:20:49 -06:00
|
|
|
event = await self.store.get_event(event_id, check_room_id=room_id)
|
2022-03-21 05:23:32 -06:00
|
|
|
if event.internal_metadata.outlier:
|
|
|
|
raise NotFoundError("State not known at event %s" % (event_id,))
|
2018-08-02 04:53:52 -06:00
|
|
|
|
2022-05-31 06:17:50 -06:00
|
|
|
state_groups = await self._state_storage_controller.get_state_groups_ids(
|
2022-05-25 05:59:04 -06:00
|
|
|
room_id, [event_id]
|
|
|
|
)
|
2016-09-02 03:49:43 -06:00
|
|
|
|
2022-03-21 05:23:32 -06:00
|
|
|
# get_state_groups_ids should return exactly one result
|
|
|
|
assert len(state_groups) == 1
|
2016-09-02 03:49:43 -06:00
|
|
|
|
2022-03-21 05:23:32 -06:00
|
|
|
state_map = next(iter(state_groups.values()))
|
2016-09-02 03:49:43 -06:00
|
|
|
|
2022-03-21 05:23:32 -06:00
|
|
|
state_key = event.get_state_key()
|
|
|
|
if state_key is not None:
|
|
|
|
# the event was not rejected (get_event raises a NotFoundError for rejected
|
|
|
|
# events) so the state at the event should include the event itself.
|
|
|
|
assert (
|
|
|
|
state_map.get((event.type, state_key)) == event.event_id
|
|
|
|
), "State at event did not include event itself"
|
|
|
|
|
|
|
|
# ... but we need the state *before* that event
|
|
|
|
if "replaces_state" in event.unsigned:
|
|
|
|
prev_id = event.unsigned["replaces_state"]
|
|
|
|
state_map[(event.type, state_key)] = prev_id
|
|
|
|
else:
|
|
|
|
del state_map[(event.type, state_key)]
|
|
|
|
|
|
|
|
return list(state_map.values())
|
2016-09-02 03:49:43 -06:00
|
|
|
|
2020-04-24 12:36:38 -06:00
|
|
|
async def on_backfill_request(
|
|
|
|
self, origin: str, room_id: str, pdu_list: List[str], limit: int
|
|
|
|
) -> List[EventBase]:
|
2022-11-21 08:46:14 -07:00
|
|
|
# We allow partially joined rooms since in this case we are filtering out
|
|
|
|
# non-local events in `filter_events_for_server`.
|
|
|
|
await self._event_auth_handler.assert_host_in_room(room_id, origin, True)
|
2014-10-31 03:59:02 -06:00
|
|
|
|
2020-02-06 11:25:24 -07:00
|
|
|
# Synapse asks for 100 events per backfill request. Do not allow more.
|
|
|
|
limit = min(limit, 100)
|
|
|
|
|
2020-04-24 12:36:38 -06:00
|
|
|
events = await self.store.get_backfill_events(room_id, pdu_list, limit)
|
Fix historical messages backfilling in random order on remote homeservers (MSC2716) (#11114)
Fix https://github.com/matrix-org/synapse/issues/11091
Fix https://github.com/matrix-org/synapse/issues/10764 (side-stepping the issue because we no longer have to deal with `fake_prev_event_id`)
1. Made the `/backfill` response return messages in `(depth, stream_ordering)` order (previously only sorted by `depth`)
- Technically, it shouldn't really matter how `/backfill` returns things but I'm just trying to make the `stream_ordering` a little more consistent from the origin to the remote homeservers in order to get the order of messages from `/messages` consistent ([sorted by `(topological_ordering, stream_ordering)`](https://github.com/matrix-org/synapse/blob/develop/docs/development/room-dag-concepts.md#depth-and-stream-ordering)).
- Even now that we return backfilled messages in order, it still doesn't guarantee the same `stream_ordering` (and more importantly the [`/messages` order](https://github.com/matrix-org/synapse/blob/develop/docs/development/room-dag-concepts.md#depth-and-stream-ordering)) on the other server. For example, if a room has a bunch of history imported and someone visits a permalink to a historical message back in time, their homeserver will skip over the historical messages in between and insert the permalink as the next message in the `stream_order` and totally throw off the sort.
- This will be even more the case when we add the [MSC3030 jump to date API endpoint](https://github.com/matrix-org/matrix-doc/pull/3030) so the static archives can navigate and jump to a certain date.
- We're solving this in the future by switching to [online topological ordering](https://github.com/matrix-org/gomatrixserverlib/issues/187) and [chunking](https://github.com/matrix-org/synapse/issues/3785) which by its nature will apply retroactively to fix any inconsistencies introduced by people permalinking
2. As we're navigating `prev_events` to return in `/backfill`, we order by `depth` first (newest -> oldest) and now also tie-break based on the `stream_ordering` (newest -> oldest). This is technically important because MSC2716 inserts a bunch of historical messages at the same `depth` so it's best to be prescriptive about which ones we should process first. In reality, I think the code already looped over the historical messages as expected because the database is already in order.
3. Making the historical state chain and historical event chain float on their own by having no `prev_events` instead of a fake `prev_event` which caused backfill to get clogged with an unresolvable event. Fixes https://github.com/matrix-org/synapse/issues/11091 and https://github.com/matrix-org/synapse/issues/10764
4. We no longer find connected insertion events by finding a potential `prev_event` connection to the current event we're iterating over. We now solely rely on marker events which when processed, add the insertion event as an extremity and the federating homeserver can ask about it when time calls.
- Related discussion, https://github.com/matrix-org/synapse/pull/11114#discussion_r741514793
Before | After
--- | ---
![](https://user-images.githubusercontent.com/558581/139218681-b465c862-5c49-4702-a59e-466733b0cf45.png) | ![](https://user-images.githubusercontent.com/558581/146453159-a1609e0a-8324-439d-ae44-e4bce43ac6d1.png)
#### Why aren't we sorting topologically when receiving backfill events?
> The main reason we're going to opt to not sort topologically when receiving backfill events is because it's probably best to do whatever is easiest to make it just work. People will probably have opinions once they look at [MSC2716](https://github.com/matrix-org/matrix-doc/pull/2716) which could change whatever implementation anyway.
>
> As mentioned, ideally we would do this but code necessary to make the fake edges but it gets confusing and gives an impression of “just whyyyy” (feels icky). This problem also dissolves with online topological ordering.
>
> -- https://github.com/matrix-org/synapse/pull/11114#discussion_r741517138
See https://github.com/matrix-org/synapse/pull/11114#discussion_r739610091 for the technical difficulties
2022-02-07 14:54:13 -07:00
|
|
|
logger.debug(
|
|
|
|
"on_backfill_request: backfill events=%s",
|
|
|
|
[
|
|
|
|
"event_id=%s,depth=%d,body=%s,prevs=%s\n"
|
|
|
|
% (
|
|
|
|
event.event_id,
|
|
|
|
event.depth,
|
|
|
|
event.content.get("body", event.type),
|
|
|
|
event.prev_event_ids(),
|
|
|
|
)
|
|
|
|
for event in events
|
|
|
|
],
|
|
|
|
)
|
2014-10-31 03:59:02 -06:00
|
|
|
|
2022-05-31 06:17:50 -06:00
|
|
|
events = await filter_events_for_server(
|
2023-03-10 08:31:25 -07:00
|
|
|
self._storage_controllers,
|
|
|
|
origin,
|
|
|
|
self.server_name,
|
|
|
|
events,
|
|
|
|
redact=True,
|
|
|
|
filter_out_erased_senders=True,
|
|
|
|
filter_out_remote_partial_state_events=True,
|
2022-05-31 06:17:50 -06:00
|
|
|
)
|
2015-07-03 10:52:57 -06:00
|
|
|
|
2019-07-23 07:00:55 -06:00
|
|
|
return events
|
2014-10-31 03:59:02 -06:00
|
|
|
|
2020-04-24 12:36:38 -06:00
|
|
|
async def get_persisted_pdu(
|
|
|
|
self, origin: str, event_id: str
|
|
|
|
) -> Optional[EventBase]:
|
2018-06-07 09:18:57 -06:00
|
|
|
"""Get an event from the database for the given server.
|
|
|
|
|
|
|
|
Args:
|
2020-04-24 12:36:38 -06:00
|
|
|
origin: hostname of server which is requesting the event; we
|
2018-06-07 09:18:57 -06:00
|
|
|
will check that the server is allowed to see it.
|
2020-04-24 12:36:38 -06:00
|
|
|
event_id: id of the event being requested
|
2014-10-31 04:47:34 -06:00
|
|
|
|
|
|
|
Returns:
|
2020-04-24 12:36:38 -06:00
|
|
|
None if we know nothing about the event; otherwise the (possibly-redacted) event.
|
2018-06-07 09:18:57 -06:00
|
|
|
|
|
|
|
Raises:
|
|
|
|
AuthError if the server is not currently in the room
|
2014-10-31 04:47:34 -06:00
|
|
|
"""
|
2020-04-24 12:36:38 -06:00
|
|
|
event = await self.store.get_event(
|
2015-02-03 03:40:14 -07:00
|
|
|
event_id, allow_none=True, allow_rejected=True
|
2014-10-31 04:47:34 -06:00
|
|
|
)
|
|
|
|
|
2022-09-23 04:47:16 -06:00
|
|
|
if not event:
|
2019-07-23 07:00:55 -06:00
|
|
|
return None
|
2014-10-31 04:47:34 -06:00
|
|
|
|
2022-09-23 04:47:16 -06:00
|
|
|
await self._event_auth_handler.assert_host_in_room(event.room_id, origin)
|
|
|
|
|
|
|
|
events = await filter_events_for_server(
|
2023-03-10 08:31:25 -07:00
|
|
|
self._storage_controllers,
|
|
|
|
origin,
|
|
|
|
self.server_name,
|
|
|
|
[event],
|
|
|
|
redact=True,
|
|
|
|
filter_out_erased_senders=True,
|
|
|
|
filter_out_remote_partial_state_events=True,
|
2022-09-23 04:47:16 -06:00
|
|
|
)
|
|
|
|
event = events[0]
|
|
|
|
return event
|
|
|
|
|
2020-02-03 12:15:08 -07:00
|
|
|
async def on_get_missing_events(
|
2021-04-06 05:21:57 -06:00
|
|
|
self,
|
|
|
|
origin: str,
|
|
|
|
room_id: str,
|
|
|
|
earliest_events: List[str],
|
|
|
|
latest_events: List[str],
|
|
|
|
limit: int,
|
|
|
|
) -> List[EventBase]:
|
2022-11-21 08:46:14 -07:00
|
|
|
# We allow partially joined rooms since in this case we are filtering out
|
|
|
|
# non-local events in `filter_events_for_server`.
|
|
|
|
await self._event_auth_handler.assert_host_in_room(room_id, origin, True)
|
2015-02-23 06:58:02 -07:00
|
|
|
|
2020-02-06 11:25:24 -07:00
|
|
|
# Only allow up to 20 events to be retrieved per request.
|
2015-02-23 06:58:02 -07:00
|
|
|
limit = min(limit, 20)
|
|
|
|
|
2020-02-03 12:15:08 -07:00
|
|
|
missing_events = await self.store.get_missing_events(
|
2015-02-23 06:58:02 -07:00
|
|
|
room_id=room_id,
|
|
|
|
earliest_events=earliest_events,
|
|
|
|
latest_events=latest_events,
|
|
|
|
limit=limit,
|
|
|
|
)
|
|
|
|
|
2020-02-03 12:15:08 -07:00
|
|
|
missing_events = await filter_events_for_server(
|
2023-03-10 08:31:25 -07:00
|
|
|
self._storage_controllers,
|
|
|
|
origin,
|
|
|
|
self.server_name,
|
|
|
|
missing_events,
|
|
|
|
redact=True,
|
|
|
|
filter_out_erased_senders=True,
|
|
|
|
filter_out_remote_partial_state_events=True,
|
2018-06-08 04:34:46 -06:00
|
|
|
)
|
|
|
|
|
2019-07-23 07:00:55 -06:00
|
|
|
return missing_events
|
2015-02-23 06:58:02 -07:00
|
|
|
|
2020-05-01 08:15:36 -06:00
|
|
|
async def exchange_third_party_invite(
|
2021-04-06 05:21:57 -06:00
|
|
|
self, sender_user_id: str, target_user_id: str, room_id: str, signed: JsonDict
|
|
|
|
) -> None:
|
2015-12-17 10:09:51 -07:00
|
|
|
third_party_invite = {"signed": signed}
|
|
|
|
|
2015-11-05 09:43:19 -07:00
|
|
|
event_dict = {
|
|
|
|
"type": EventTypes.Member,
|
|
|
|
"content": {
|
|
|
|
"membership": Membership.INVITE,
|
2015-12-17 10:09:51 -07:00
|
|
|
"third_party_invite": third_party_invite,
|
2015-11-05 09:43:19 -07:00
|
|
|
},
|
|
|
|
"room_id": room_id,
|
2016-02-23 08:11:25 -07:00
|
|
|
"sender": sender_user_id,
|
|
|
|
"state_key": target_user_id,
|
2015-11-05 09:43:19 -07:00
|
|
|
}
|
|
|
|
|
2022-09-23 04:47:16 -06:00
|
|
|
if await self._event_auth_handler.is_host_in_room(room_id, self.hs.hostname):
|
2021-09-29 03:57:10 -06:00
|
|
|
room_version_obj = await self.store.get_room_version(room_id)
|
|
|
|
builder = self.event_builder_factory.for_room_version(
|
|
|
|
room_version_obj, event_dict
|
|
|
|
)
|
2019-01-23 13:21:33 -07:00
|
|
|
|
2019-01-28 10:00:14 -07:00
|
|
|
EventValidator().validate_builder(builder)
|
2015-12-17 10:31:20 -07:00
|
|
|
|
2022-12-15 09:04:23 -07:00
|
|
|
# Try several times, it could fail with PartialStateConflictError
|
|
|
|
# in send_membership_event, cf comment in except block.
|
|
|
|
max_retries = 5
|
|
|
|
for i in range(max_retries):
|
|
|
|
try:
|
|
|
|
(
|
|
|
|
event,
|
2023-02-09 14:05:02 -07:00
|
|
|
unpersisted_context,
|
2022-12-15 09:04:23 -07:00
|
|
|
) = await self.event_creation_handler.create_new_client_event(
|
|
|
|
builder=builder
|
|
|
|
)
|
2015-12-17 10:31:20 -07:00
|
|
|
|
2023-02-09 14:05:02 -07:00
|
|
|
(
|
|
|
|
event,
|
|
|
|
unpersisted_context,
|
|
|
|
) = await self.add_display_name_to_third_party_invite(
|
|
|
|
room_version_obj, event_dict, event, unpersisted_context
|
2022-12-15 09:04:23 -07:00
|
|
|
)
|
2019-01-28 10:00:14 -07:00
|
|
|
|
2023-02-09 14:05:02 -07:00
|
|
|
context = await unpersisted_context.persist(event)
|
|
|
|
|
2022-12-15 09:04:23 -07:00
|
|
|
EventValidator().validate_new(event, self.config)
|
2019-01-29 09:15:16 -07:00
|
|
|
|
2022-12-15 09:04:23 -07:00
|
|
|
# We need to tell the transaction queue to send this out, even
|
|
|
|
# though the sender isn't a local user.
|
|
|
|
event.internal_metadata.send_on_behalf_of = self.hs.hostname
|
2016-04-13 04:11:46 -06:00
|
|
|
|
2022-12-15 09:04:23 -07:00
|
|
|
try:
|
|
|
|
validate_event_for_room_version(event)
|
|
|
|
await self._event_auth_handler.check_auth_rules_from_context(
|
|
|
|
event
|
|
|
|
)
|
|
|
|
except AuthError as e:
|
|
|
|
logger.warning(
|
|
|
|
"Denying new third party invite %r because %s", event, e
|
|
|
|
)
|
|
|
|
raise e
|
2019-11-01 04:28:09 -06:00
|
|
|
|
2022-12-15 09:04:23 -07:00
|
|
|
await self._check_signature(event, context)
|
|
|
|
|
|
|
|
# We retrieve the room member handler here as to not cause a cyclic dependency
|
|
|
|
member_handler = self.hs.get_room_member_handler()
|
|
|
|
await member_handler.send_membership_event(None, event, context)
|
|
|
|
|
|
|
|
break
|
|
|
|
except PartialStateConflictError as e:
|
|
|
|
# Persisting couldn't happen because the room got un-partial stated
|
|
|
|
# in the meantime and context needs to be recomputed, so let's do so.
|
|
|
|
if i == max_retries - 1:
|
|
|
|
raise e
|
2015-11-05 09:43:19 -07:00
|
|
|
else:
|
2020-02-21 05:15:07 -07:00
|
|
|
destinations = {x.split(":", 1)[-1] for x in (sender_user_id, room_id)}
|
2021-07-15 03:35:46 -06:00
|
|
|
|
|
|
|
try:
|
|
|
|
await self.federation_client.forward_third_party_invite(
|
|
|
|
destinations, room_id, event_dict
|
|
|
|
)
|
|
|
|
except (RequestSendFailed, HttpResponseException):
|
|
|
|
raise SynapseError(502, "Failed to forward third party invite")
|
2015-11-05 09:43:19 -07:00
|
|
|
|
2020-02-03 09:02:50 -07:00
|
|
|
async def on_exchange_third_party_invite_request(
|
2020-11-19 03:05:33 -07:00
|
|
|
self, event_dict: JsonDict
|
2020-02-03 09:02:50 -07:00
|
|
|
) -> None:
|
2017-09-19 05:18:01 -06:00
|
|
|
"""Handle an exchange_third_party_invite request from a remote server
|
|
|
|
|
|
|
|
The remote server will call this when it wants to turn a 3pid invite
|
|
|
|
into a normal m.room.member invite.
|
|
|
|
|
2019-09-11 03:37:17 -06:00
|
|
|
Args:
|
2020-11-19 03:05:33 -07:00
|
|
|
event_dict: Dictionary containing the event body.
|
2019-09-11 03:37:17 -06:00
|
|
|
|
2017-09-19 05:18:01 -06:00
|
|
|
"""
|
2020-11-19 03:05:33 -07:00
|
|
|
assert_params_in_dict(event_dict, ["room_id"])
|
2021-09-29 03:57:10 -06:00
|
|
|
room_version_obj = await self.store.get_room_version(event_dict["room_id"])
|
2019-01-23 13:21:33 -07:00
|
|
|
|
|
|
|
# NB: event_dict has a particular specced format we might need to fudge
|
|
|
|
# if we change event formats too much.
|
2021-09-29 03:57:10 -06:00
|
|
|
builder = self.event_builder_factory.for_room_version(
|
|
|
|
room_version_obj, event_dict
|
|
|
|
)
|
2015-11-05 09:43:19 -07:00
|
|
|
|
2022-12-15 09:04:23 -07:00
|
|
|
# Try several times, it could fail with PartialStateConflictError
|
|
|
|
# in send_membership_event, cf comment in except block.
|
|
|
|
max_retries = 5
|
|
|
|
for i in range(max_retries):
|
|
|
|
try:
|
|
|
|
(
|
|
|
|
event,
|
2023-02-09 14:05:02 -07:00
|
|
|
unpersisted_context,
|
2022-12-15 09:04:23 -07:00
|
|
|
) = await self.event_creation_handler.create_new_client_event(
|
|
|
|
builder=builder
|
|
|
|
)
|
2023-02-09 14:05:02 -07:00
|
|
|
(
|
|
|
|
event,
|
|
|
|
unpersisted_context,
|
|
|
|
) = await self.add_display_name_to_third_party_invite(
|
|
|
|
room_version_obj, event_dict, event, unpersisted_context
|
2022-12-15 09:04:23 -07:00
|
|
|
)
|
2015-12-17 10:31:20 -07:00
|
|
|
|
2023-02-09 14:05:02 -07:00
|
|
|
context = await unpersisted_context.persist(event)
|
|
|
|
|
2022-12-15 09:04:23 -07:00
|
|
|
try:
|
|
|
|
validate_event_for_room_version(event)
|
|
|
|
await self._event_auth_handler.check_auth_rules_from_context(event)
|
|
|
|
except AuthError as e:
|
|
|
|
logger.warning("Denying third party invite %r because %s", event, e)
|
|
|
|
raise e
|
|
|
|
await self._check_signature(event, context)
|
|
|
|
|
|
|
|
# We need to tell the transaction queue to send this out, even
|
|
|
|
# though the sender isn't a local user.
|
|
|
|
event.internal_metadata.send_on_behalf_of = get_domain_from_id(
|
|
|
|
event.sender
|
|
|
|
)
|
2015-11-05 09:43:19 -07:00
|
|
|
|
2022-12-15 09:04:23 -07:00
|
|
|
# We retrieve the room member handler here as to not cause a cyclic dependency
|
|
|
|
member_handler = self.hs.get_room_member_handler()
|
|
|
|
await member_handler.send_membership_event(None, event, context)
|
2019-01-29 09:15:16 -07:00
|
|
|
|
2022-12-15 09:04:23 -07:00
|
|
|
break
|
|
|
|
except PartialStateConflictError as e:
|
|
|
|
# Persisting couldn't happen because the room got un-partial stated
|
|
|
|
# in the meantime and context needs to be recomputed, so let's do so.
|
|
|
|
if i == max_retries - 1:
|
|
|
|
raise e
|
2015-11-05 09:43:19 -07:00
|
|
|
|
2020-05-11 13:12:46 -06:00
|
|
|
async def add_display_name_to_third_party_invite(
|
2021-04-06 05:21:57 -06:00
|
|
|
self,
|
2021-09-29 03:57:10 -06:00
|
|
|
room_version_obj: RoomVersion,
|
2021-04-06 05:21:57 -06:00
|
|
|
event_dict: JsonDict,
|
|
|
|
event: EventBase,
|
2023-02-09 14:05:02 -07:00
|
|
|
context: UnpersistedEventContextBase,
|
|
|
|
) -> Tuple[EventBase, UnpersistedEventContextBase]:
|
2015-12-17 10:31:20 -07:00
|
|
|
key = (
|
|
|
|
EventTypes.ThirdPartyInvite,
|
|
|
|
event.content["third_party_invite"]["signed"]["token"],
|
|
|
|
)
|
2016-08-25 10:32:22 -06:00
|
|
|
original_invite = None
|
2023-07-20 03:46:37 -06:00
|
|
|
prev_state_ids = await context.get_prev_state_ids(StateFilter.from_types([key]))
|
2018-07-23 06:00:22 -06:00
|
|
|
original_invite_id = prev_state_ids.get(key)
|
2016-08-25 10:32:22 -06:00
|
|
|
if original_invite_id:
|
2020-05-11 13:12:46 -06:00
|
|
|
original_invite = await self.store.get_event(
|
2016-08-25 10:32:22 -06:00
|
|
|
original_invite_id, allow_none=True
|
|
|
|
)
|
2016-09-22 03:56:53 -06:00
|
|
|
if original_invite:
|
2019-10-02 04:16:38 -06:00
|
|
|
# If the m.room.third_party_invite event's content is empty, it means the
|
2019-10-04 04:16:19 -06:00
|
|
|
# invite has been revoked. In this case, we don't have to raise an error here
|
|
|
|
# because the auth check will fail on the invite (because it's not able to
|
|
|
|
# fetch public keys from the m.room.third_party_invite event's content, which
|
2019-10-04 04:21:24 -06:00
|
|
|
# is empty).
|
2019-10-04 04:16:19 -06:00
|
|
|
display_name = original_invite.content.get("display_name")
|
2019-10-04 04:18:28 -06:00
|
|
|
event_dict["content"]["third_party_invite"]["display_name"] = display_name
|
2016-09-22 03:56:53 -06:00
|
|
|
else:
|
2015-12-17 10:31:20 -07:00
|
|
|
logger.info(
|
2016-09-22 03:56:53 -06:00
|
|
|
"Could not find invite event for third_party_invite: %r", event_dict
|
2015-12-17 10:31:20 -07:00
|
|
|
)
|
2016-09-22 04:59:46 -06:00
|
|
|
# We don't discard here as this is not the appropriate place to do
|
|
|
|
# auth checks. If we need the invite and don't have it then the
|
|
|
|
# auth check code will explode appropriately.
|
2015-12-17 10:31:20 -07:00
|
|
|
|
2021-09-29 03:57:10 -06:00
|
|
|
builder = self.event_builder_factory.for_room_version(
|
|
|
|
room_version_obj, event_dict
|
|
|
|
)
|
2019-01-28 10:00:14 -07:00
|
|
|
EventValidator().validate_builder(builder)
|
2023-02-09 14:05:02 -07:00
|
|
|
|
|
|
|
(
|
|
|
|
event,
|
|
|
|
unpersisted_context,
|
|
|
|
) = await self.event_creation_handler.create_new_client_event(builder=builder)
|
|
|
|
|
2019-11-19 07:07:39 -07:00
|
|
|
EventValidator().validate_new(event, self.config)
|
2023-02-09 14:05:02 -07:00
|
|
|
return event, unpersisted_context
|
2015-12-17 10:31:20 -07:00
|
|
|
|
2021-04-06 05:21:57 -06:00
|
|
|
async def _check_signature(self, event: EventBase, context: EventContext) -> None:
|
2016-02-23 08:11:25 -07:00
|
|
|
"""
|
|
|
|
Checks that the signature in the event is consistent with its invite.
|
|
|
|
|
2016-04-01 09:08:59 -06:00
|
|
|
Args:
|
2021-04-06 05:21:57 -06:00
|
|
|
event: The m.room.member event to check
|
|
|
|
context:
|
2016-04-01 09:08:59 -06:00
|
|
|
|
|
|
|
Raises:
|
|
|
|
AuthError: if signature didn't match any keys, or key has been
|
2016-02-23 08:11:25 -07:00
|
|
|
revoked,
|
2016-04-01 09:08:59 -06:00
|
|
|
SynapseError: if a transient error meant a key couldn't be checked
|
2016-02-23 08:11:25 -07:00
|
|
|
for revocation.
|
|
|
|
"""
|
|
|
|
signed = event.content["third_party_invite"]["signed"]
|
|
|
|
token = signed["token"]
|
2015-11-05 09:43:19 -07:00
|
|
|
|
2022-05-20 02:54:12 -06:00
|
|
|
prev_state_ids = await context.get_prev_state_ids(
|
2023-07-20 03:46:37 -06:00
|
|
|
StateFilter.from_types([(EventTypes.ThirdPartyInvite, token)])
|
2022-05-20 02:54:12 -06:00
|
|
|
)
|
2018-07-23 06:00:22 -06:00
|
|
|
invite_event_id = prev_state_ids.get((EventTypes.ThirdPartyInvite, token))
|
2015-11-05 09:43:19 -07:00
|
|
|
|
2016-08-25 10:32:22 -06:00
|
|
|
invite_event = None
|
|
|
|
if invite_event_id:
|
2020-05-11 13:12:46 -06:00
|
|
|
invite_event = await self.store.get_event(invite_event_id, allow_none=True)
|
2016-08-25 10:32:22 -06:00
|
|
|
|
2016-02-23 08:11:25 -07:00
|
|
|
if not invite_event:
|
|
|
|
raise AuthError(403, "Could not find invite")
|
|
|
|
|
2019-06-18 15:51:24 -06:00
|
|
|
logger.debug("Checking auth on event %r", event.content)
|
|
|
|
|
2021-07-16 11:22:36 -06:00
|
|
|
last_exception: Optional[Exception] = None
|
2020-07-01 09:21:02 -06:00
|
|
|
|
2019-06-18 15:51:24 -06:00
|
|
|
# for each public key in the 3pid invite event
|
2021-07-01 12:25:37 -06:00
|
|
|
for public_key_object in event_auth.get_public_keys(invite_event):
|
2016-02-23 08:11:25 -07:00
|
|
|
try:
|
2019-06-18 15:51:24 -06:00
|
|
|
# for each sig on the third_party_invite block of the actual invite
|
2016-02-23 08:11:25 -07:00
|
|
|
for server, signature_block in signed["signatures"].items():
|
2021-04-20 04:50:49 -06:00
|
|
|
for key_name in signature_block.keys():
|
2016-02-23 08:11:25 -07:00
|
|
|
if not key_name.startswith("ed25519:"):
|
|
|
|
continue
|
|
|
|
|
2019-06-18 15:51:24 -06:00
|
|
|
logger.debug(
|
|
|
|
"Attempting to verify sig with key %s from %r "
|
|
|
|
"against pubkey %r",
|
|
|
|
key_name,
|
|
|
|
server,
|
|
|
|
public_key_object,
|
2016-02-23 08:11:25 -07:00
|
|
|
)
|
2019-06-18 15:51:24 -06:00
|
|
|
|
|
|
|
try:
|
|
|
|
public_key = public_key_object["public_key"]
|
|
|
|
verify_key = decode_verify_key_bytes(
|
|
|
|
key_name, decode_base64(public_key)
|
|
|
|
)
|
|
|
|
verify_signed_json(signed, server, verify_key)
|
|
|
|
logger.debug(
|
|
|
|
"Successfully verified sig with key %s from %r "
|
|
|
|
"against pubkey %r",
|
|
|
|
key_name,
|
|
|
|
server,
|
|
|
|
public_key_object,
|
|
|
|
)
|
|
|
|
except Exception:
|
|
|
|
logger.info(
|
|
|
|
"Failed to verify sig with key %s from %r "
|
|
|
|
"against pubkey %r",
|
|
|
|
key_name,
|
|
|
|
server,
|
|
|
|
public_key_object,
|
|
|
|
)
|
|
|
|
raise
|
|
|
|
try:
|
|
|
|
if "key_validity_url" in public_key_object:
|
2020-05-11 13:12:46 -06:00
|
|
|
await self._check_key_revocation(
|
2019-06-18 15:51:24 -06:00
|
|
|
public_key, public_key_object["key_validity_url"]
|
|
|
|
)
|
|
|
|
except Exception:
|
|
|
|
logger.info(
|
|
|
|
"Failed to query key_validity_url %s",
|
2016-02-23 08:11:25 -07:00
|
|
|
public_key_object["key_validity_url"],
|
|
|
|
)
|
2019-06-18 15:51:24 -06:00
|
|
|
raise
|
2016-02-23 08:11:25 -07:00
|
|
|
return
|
|
|
|
except Exception as e:
|
|
|
|
last_exception = e
|
2020-07-01 09:21:02 -06:00
|
|
|
|
|
|
|
if last_exception is None:
|
|
|
|
# we can only get here if get_public_keys() returned an empty list
|
|
|
|
# TODO: make this better
|
|
|
|
raise RuntimeError("no public key in invite event")
|
|
|
|
|
2016-02-23 08:11:25 -07:00
|
|
|
raise last_exception
|
|
|
|
|
2021-04-06 05:21:57 -06:00
|
|
|
async def _check_key_revocation(self, public_key: str, url: str) -> None:
|
2016-02-23 08:11:25 -07:00
|
|
|
"""
|
|
|
|
Checks whether public_key has been revoked.
|
|
|
|
|
2016-04-01 09:08:59 -06:00
|
|
|
Args:
|
2021-04-06 05:21:57 -06:00
|
|
|
public_key: base-64 encoded public key.
|
|
|
|
url: Key revocation URL.
|
2016-02-23 08:11:25 -07:00
|
|
|
|
2016-04-01 09:08:59 -06:00
|
|
|
Raises:
|
|
|
|
AuthError: if they key has been revoked.
|
|
|
|
SynapseError: if a transient error meant a key couldn't be checked
|
2016-02-23 08:11:25 -07:00
|
|
|
for revocation.
|
|
|
|
"""
|
2015-11-05 09:43:19 -07:00
|
|
|
try:
|
2020-05-11 13:12:46 -06:00
|
|
|
response = await self.http_client.get_json(url, {"public_key": public_key})
|
2015-11-05 09:43:19 -07:00
|
|
|
except Exception:
|
|
|
|
raise SynapseError(502, "Third party certificate could not be checked")
|
|
|
|
if "valid" not in response or not response["valid"]:
|
|
|
|
raise AuthError(403, "Third party certificate was invalid")
|
2018-07-25 09:00:38 -06:00
|
|
|
|
2020-02-03 09:27:05 -07:00
|
|
|
async def _clean_room_for_join(self, room_id: str) -> None:
|
2018-08-09 03:29:48 -06:00
|
|
|
"""Called to clean up any data in DB for a given room, ready for the
|
|
|
|
server to join the room.
|
|
|
|
|
|
|
|
Args:
|
2020-02-03 09:27:05 -07:00
|
|
|
room_id
|
2018-08-09 03:29:48 -06:00
|
|
|
"""
|
2021-09-13 11:07:12 -06:00
|
|
|
if self.config.worker.worker_app:
|
2020-02-03 09:27:05 -07:00
|
|
|
await self._clean_room_for_join_client(room_id)
|
2018-08-09 03:29:48 -06:00
|
|
|
else:
|
2020-02-03 09:27:05 -07:00
|
|
|
await self.store.clean_room_for_join(room_id)
|
2018-07-25 09:00:38 -06:00
|
|
|
|
2020-07-24 08:53:25 -06:00
|
|
|
async def get_room_complexity(
|
|
|
|
self, remote_room_hosts: List[str], room_id: str
|
|
|
|
) -> Optional[dict]:
|
2019-07-29 10:47:27 -06:00
|
|
|
"""
|
|
|
|
Fetch the complexity of a remote room over federation.
|
|
|
|
|
|
|
|
Args:
|
2022-11-16 08:25:24 -07:00
|
|
|
remote_room_hosts: The remote servers to ask.
|
|
|
|
room_id: The room ID to ask about.
|
2019-07-29 10:47:27 -06:00
|
|
|
|
|
|
|
Returns:
|
2020-07-24 08:53:25 -06:00
|
|
|
Dict contains the complexity
|
2019-07-29 10:47:27 -06:00
|
|
|
metric versions, while None means we could not fetch the complexity.
|
|
|
|
"""
|
|
|
|
|
|
|
|
for host in remote_room_hosts:
|
2020-05-11 13:12:46 -06:00
|
|
|
res = await self.federation_client.get_room_complexity(host, room_id)
|
2019-07-29 10:47:27 -06:00
|
|
|
|
|
|
|
# We got a result, return it.
|
|
|
|
if res:
|
2020-05-11 13:12:46 -06:00
|
|
|
return res
|
2019-07-29 10:47:27 -06:00
|
|
|
|
|
|
|
# We fell off the bottom, couldn't get the complexity from anyone. Oh
|
|
|
|
# well.
|
2020-05-11 13:12:46 -06:00
|
|
|
return None
|
2022-04-12 07:23:43 -06:00
|
|
|
|
2023-01-20 05:06:19 -07:00
|
|
|
async def _resume_partial_state_room_sync(self) -> None:
|
2022-05-31 09:15:08 -06:00
|
|
|
"""Resumes resyncing of all partial-state rooms after a restart."""
|
|
|
|
assert not self.config.worker.worker_app
|
|
|
|
|
2022-10-18 05:33:18 -06:00
|
|
|
partial_state_rooms = await self.store.get_partial_state_room_resync_info()
|
|
|
|
for room_id, resync_info in partial_state_rooms.items():
|
2023-01-20 05:06:19 -07:00
|
|
|
self._start_partial_state_room_sync(
|
2022-10-18 05:33:18 -06:00
|
|
|
initial_destination=resync_info.joined_via,
|
|
|
|
other_destinations=resync_info.servers_in_room,
|
2022-05-31 09:15:08 -06:00
|
|
|
room_id=room_id,
|
|
|
|
)
|
|
|
|
|
2023-01-20 05:06:19 -07:00
|
|
|
def _start_partial_state_room_sync(
|
|
|
|
self,
|
|
|
|
initial_destination: Optional[str],
|
2023-02-03 08:39:59 -07:00
|
|
|
other_destinations: AbstractSet[str],
|
2023-01-20 05:06:19 -07:00
|
|
|
room_id: str,
|
|
|
|
) -> None:
|
|
|
|
"""Starts the background process to resync the state of a partial state room,
|
|
|
|
if it is not already running.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
initial_destination: the initial homeserver to pull the state from
|
|
|
|
other_destinations: other homeservers to try to pull the state from, if
|
|
|
|
`initial_destination` is unavailable
|
|
|
|
room_id: room to be resynced
|
|
|
|
"""
|
|
|
|
|
|
|
|
async def _sync_partial_state_room_wrapper() -> None:
|
|
|
|
if room_id in self._active_partial_state_syncs:
|
|
|
|
# Another local user has joined the room while there is already a
|
|
|
|
# partial state sync running. This implies that there is a new join
|
|
|
|
# event to un-partial state. We might find ourselves in one of a few
|
|
|
|
# scenarios:
|
|
|
|
# 1. There is an existing partial state sync. The partial state sync
|
|
|
|
# un-partial states the new join event before completing and all is
|
|
|
|
# well.
|
|
|
|
# 2. Before the latest join, the homeserver was no longer in the room
|
|
|
|
# and there is an existing partial state sync from our previous
|
|
|
|
# membership of the room. The partial state sync may have:
|
|
|
|
# a) succeeded, but not yet terminated. The room will not be
|
|
|
|
# un-partial stated again unless we restart the partial state
|
|
|
|
# sync.
|
|
|
|
# b) failed, because we were no longer in the room and remote
|
|
|
|
# homeservers were refusing our requests, but not yet
|
|
|
|
# terminated. After the latest join, remote homeservers may
|
|
|
|
# start answering our requests again, so we should restart the
|
|
|
|
# partial state sync.
|
|
|
|
# In the cases where we would want to restart the partial state sync,
|
|
|
|
# the room would have the partial state flag when the partial state sync
|
|
|
|
# terminates.
|
|
|
|
self._partial_state_syncs_maybe_needing_restart[room_id] = (
|
|
|
|
initial_destination,
|
|
|
|
other_destinations,
|
|
|
|
)
|
|
|
|
return
|
|
|
|
|
|
|
|
self._active_partial_state_syncs.add(room_id)
|
|
|
|
|
|
|
|
try:
|
|
|
|
await self._sync_partial_state_room(
|
|
|
|
initial_destination=initial_destination,
|
|
|
|
other_destinations=other_destinations,
|
|
|
|
room_id=room_id,
|
|
|
|
)
|
|
|
|
finally:
|
|
|
|
# Read the room's partial state flag while we still hold the claim to
|
|
|
|
# being the active partial state sync (so that another partial state
|
|
|
|
# sync can't come along and mess with it under us).
|
|
|
|
# Normally, the partial state flag will be gone. If it isn't, then we
|
|
|
|
# may find ourselves in scenario 2a or 2b as described in the comment
|
|
|
|
# above, where we want to restart the partial state sync.
|
|
|
|
is_still_partial_state_room = await self.store.is_partial_state_room(
|
|
|
|
room_id
|
|
|
|
)
|
|
|
|
self._active_partial_state_syncs.remove(room_id)
|
|
|
|
|
|
|
|
if room_id in self._partial_state_syncs_maybe_needing_restart:
|
|
|
|
(
|
|
|
|
restart_initial_destination,
|
|
|
|
restart_other_destinations,
|
|
|
|
) = self._partial_state_syncs_maybe_needing_restart.pop(room_id)
|
|
|
|
|
|
|
|
if is_still_partial_state_room:
|
|
|
|
self._start_partial_state_room_sync(
|
|
|
|
initial_destination=restart_initial_destination,
|
|
|
|
other_destinations=restart_other_destinations,
|
|
|
|
room_id=room_id,
|
|
|
|
)
|
|
|
|
|
|
|
|
run_as_background_process(
|
|
|
|
desc="sync_partial_state_room", func=_sync_partial_state_room_wrapper
|
|
|
|
)
|
|
|
|
|
2022-04-12 07:23:43 -06:00
|
|
|
async def _sync_partial_state_room(
|
|
|
|
self,
|
2022-05-31 08:50:29 -06:00
|
|
|
initial_destination: Optional[str],
|
2023-02-03 08:39:59 -07:00
|
|
|
other_destinations: AbstractSet[str],
|
2022-04-12 07:23:43 -06:00
|
|
|
room_id: str,
|
|
|
|
) -> None:
|
|
|
|
"""Background process to resync the state of a partial-state room
|
|
|
|
|
|
|
|
Args:
|
2022-05-31 08:50:29 -06:00
|
|
|
initial_destination: the initial homeserver to pull the state from
|
|
|
|
other_destinations: other homeservers to try to pull the state from, if
|
|
|
|
`initial_destination` is unavailable
|
2022-04-12 07:23:43 -06:00
|
|
|
room_id: room to be resynced
|
|
|
|
"""
|
2023-01-22 12:19:31 -07:00
|
|
|
# Assume that we run on the main process for now.
|
|
|
|
# TODO(faster_joins,multiple workers)
|
|
|
|
# When moving the sync to workers, we need to ensure that
|
|
|
|
# * `_start_partial_state_room_sync` still prevents duplicate resyncs
|
|
|
|
# * `_is_partial_state_room_linearizer` correctly guards partial state flags
|
|
|
|
# for rooms between the workers doing remote joins and resync.
|
2022-05-31 09:15:08 -06:00
|
|
|
assert not self.config.worker.worker_app
|
2022-04-12 07:23:43 -06:00
|
|
|
|
|
|
|
# TODO(faster_joins): do we need to lock to avoid races? What happens if other
|
|
|
|
# worker processes kick off a resync in parallel? Perhaps we should just elect
|
|
|
|
# a single worker to do the resync.
|
2022-06-09 04:13:03 -06:00
|
|
|
# https://github.com/matrix-org/synapse/issues/12994
|
2022-04-12 07:23:43 -06:00
|
|
|
#
|
|
|
|
# TODO(faster_joins): what happens if we leave the room during a resync? if we
|
|
|
|
# really leave, that might mean we have difficulty getting the room state over
|
|
|
|
# federation.
|
2022-06-09 04:13:03 -06:00
|
|
|
# https://github.com/matrix-org/synapse/issues/12802
|
2022-04-12 07:23:43 -06:00
|
|
|
|
2022-05-31 08:50:29 -06:00
|
|
|
# Make an infinite iterator of destinations to try. Once we find a working
|
|
|
|
# destination, we'll stick with it until it flakes.
|
2022-10-18 05:33:18 -06:00
|
|
|
destinations = _prioritise_destinations_for_partial_state_resync(
|
|
|
|
initial_destination, other_destinations, room_id
|
|
|
|
)
|
2022-08-02 05:12:44 -06:00
|
|
|
destination_iter = itertools.cycle(destinations)
|
2022-05-31 08:50:29 -06:00
|
|
|
|
|
|
|
# `destination` is the current remote homeserver we're pulling from.
|
|
|
|
destination = next(destination_iter)
|
2022-04-12 07:23:43 -06:00
|
|
|
logger.info("Syncing state for room %s via %s", room_id, destination)
|
|
|
|
|
|
|
|
# we work through the queue in order of increasing stream ordering.
|
|
|
|
while True:
|
|
|
|
batch = await self.store.get_partial_state_events_batch(room_id)
|
|
|
|
if not batch:
|
|
|
|
# all the events are updated, so we can update current state and
|
|
|
|
# clear the lazy-loading flag.
|
|
|
|
logger.info("Updating current state for %s", room_id)
|
2022-07-07 06:19:31 -06:00
|
|
|
# TODO(faster_joins): notify workers in notify_room_un_partial_stated
|
2022-06-09 04:13:03 -06:00
|
|
|
# https://github.com/matrix-org/synapse/issues/12994
|
2023-02-14 16:42:29 -07:00
|
|
|
#
|
|
|
|
# NB: there's a potential race here. If room is purged just before we
|
|
|
|
# call this, we _might_ end up inserting rows into current_state_events.
|
|
|
|
# (The logic is hard to chase through.) We think this is fine, but if
|
|
|
|
# not the HS admin should purge the room again.
|
2022-07-07 06:19:31 -06:00
|
|
|
await self.state_handler.update_current_state(room_id)
|
2022-04-12 07:23:43 -06:00
|
|
|
|
2022-09-28 07:42:43 -06:00
|
|
|
logger.info("Handling any pending device list updates")
|
|
|
|
await self._device_handler.handle_room_un_partial_stated(room_id)
|
|
|
|
|
2023-01-22 12:19:31 -07:00
|
|
|
async with self._is_partial_state_room_linearizer.queue(room_id):
|
|
|
|
logger.info("Clearing partial-state flag for %s", room_id)
|
2023-01-23 08:44:39 -07:00
|
|
|
new_stream_id = await self.store.clear_partial_state_room(room_id)
|
2023-01-22 12:19:31 -07:00
|
|
|
|
2023-01-23 08:44:39 -07:00
|
|
|
if new_stream_id is not None:
|
2022-04-12 07:23:43 -06:00
|
|
|
logger.info("State resync complete for %s", room_id)
|
2022-06-01 09:02:53 -06:00
|
|
|
self._storage_controllers.state.notify_room_un_partial_stated(
|
|
|
|
room_id
|
|
|
|
)
|
2022-04-12 07:23:43 -06:00
|
|
|
|
2023-01-23 08:44:39 -07:00
|
|
|
await self._notifier.on_un_partial_stated_room(
|
|
|
|
room_id, new_stream_id
|
|
|
|
)
|
2022-04-12 07:23:43 -06:00
|
|
|
return
|
|
|
|
|
|
|
|
# we raced against more events arriving with partial state. Go round
|
|
|
|
# the loop again. We've already logged a warning, so no need for more.
|
|
|
|
continue
|
|
|
|
|
|
|
|
events = await self.store.get_events_as_list(
|
|
|
|
batch,
|
2022-05-04 05:26:11 -06:00
|
|
|
redact_behaviour=EventRedactBehaviour.as_is,
|
2022-04-12 07:23:43 -06:00
|
|
|
allow_rejected=True,
|
|
|
|
)
|
|
|
|
for event in events:
|
2022-05-31 08:50:29 -06:00
|
|
|
for attempt in itertools.count():
|
2023-03-30 06:36:41 -06:00
|
|
|
# We try a new destination on every iteration.
|
2022-05-31 08:50:29 -06:00
|
|
|
try:
|
2023-03-30 06:36:41 -06:00
|
|
|
while True:
|
|
|
|
try:
|
|
|
|
await self._federation_event_handler.update_state_for_partial_state_event(
|
|
|
|
destination, event
|
|
|
|
)
|
|
|
|
break
|
|
|
|
except FederationPullAttemptBackoffError as e:
|
|
|
|
# We are in the backoff period for one of the event's
|
|
|
|
# prev_events. Wait it out and try again after.
|
|
|
|
logger.warning(
|
|
|
|
"%s; waiting for %d ms...", e, e.retry_after_ms
|
|
|
|
)
|
|
|
|
await self.clock.sleep(e.retry_after_ms / 1000)
|
|
|
|
|
|
|
|
# Success, no need to try the rest of the destinations.
|
2022-05-31 08:50:29 -06:00
|
|
|
break
|
|
|
|
except FederationError as e:
|
|
|
|
if attempt == len(destinations) - 1:
|
|
|
|
# We have tried every remote server for this event. Give up.
|
|
|
|
# TODO(faster_joins) giving up isn't the right thing to do
|
|
|
|
# if there's a temporary network outage. retrying
|
|
|
|
# indefinitely is also not the right thing to do if we can
|
|
|
|
# reach all homeservers and they all claim they don't have
|
|
|
|
# the state we want.
|
2022-06-09 04:13:03 -06:00
|
|
|
# https://github.com/matrix-org/synapse/issues/13000
|
2022-05-31 08:50:29 -06:00
|
|
|
logger.error(
|
|
|
|
"Failed to get state for %s at %s from %s because %s, "
|
|
|
|
"giving up!",
|
|
|
|
room_id,
|
|
|
|
event,
|
|
|
|
destination,
|
|
|
|
e,
|
|
|
|
)
|
2023-03-30 06:36:41 -06:00
|
|
|
# TODO: We should `record_event_failed_pull_attempt` here,
|
|
|
|
# see https://github.com/matrix-org/synapse/issues/13700
|
2022-05-31 08:50:29 -06:00
|
|
|
raise
|
|
|
|
|
|
|
|
# Try the next remote server.
|
|
|
|
logger.info(
|
|
|
|
"Failed to get state for %s at %s from %s because %s",
|
|
|
|
room_id,
|
|
|
|
event,
|
|
|
|
destination,
|
|
|
|
e,
|
|
|
|
)
|
|
|
|
destination = next(destination_iter)
|
|
|
|
logger.info(
|
|
|
|
"Syncing state for room %s via %s instead",
|
|
|
|
room_id,
|
|
|
|
destination,
|
|
|
|
)
|
2022-10-18 05:33:18 -06:00
|
|
|
|
|
|
|
|
|
|
|
def _prioritise_destinations_for_partial_state_resync(
|
|
|
|
initial_destination: Optional[str],
|
2023-02-03 08:39:59 -07:00
|
|
|
other_destinations: AbstractSet[str],
|
2022-10-18 05:33:18 -06:00
|
|
|
room_id: str,
|
2023-01-26 10:31:58 -07:00
|
|
|
) -> StrCollection:
|
2022-10-18 05:33:18 -06:00
|
|
|
"""Work out the order in which we should ask servers to resync events.
|
|
|
|
|
|
|
|
If an `initial_destination` is given, it takes top priority. Otherwise
|
|
|
|
all servers are treated equally.
|
|
|
|
|
|
|
|
:raises ValueError: if no destination is provided at all.
|
|
|
|
"""
|
|
|
|
if initial_destination is None and len(other_destinations) == 0:
|
|
|
|
raise ValueError(f"Cannot resync state of {room_id}: no destinations provided")
|
|
|
|
|
|
|
|
if initial_destination is None:
|
|
|
|
return other_destinations
|
|
|
|
|
|
|
|
# Move `initial_destination` to the front of the list.
|
|
|
|
destinations = list(other_destinations)
|
|
|
|
if initial_destination in destinations:
|
|
|
|
destinations.remove(initial_destination)
|
|
|
|
destinations = [initial_destination] + destinations
|
|
|
|
return destinations
|