2014-08-12 08:10:52 -06:00
|
|
|
# -*- coding: utf-8 -*-
|
2016-01-06 21:26:29 -07:00
|
|
|
# Copyright 2014-2016 OpenMarket Ltd
|
2019-06-12 03:31:37 -06:00
|
|
|
# Copyright 2017-2018 New Vector Ltd
|
|
|
|
# Copyright 2019 The Matrix.org Foundation C.I.C.
|
2014-08-12 08:10:52 -06:00
|
|
|
#
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
2014-08-12 20:14:34 -06:00
|
|
|
|
2014-08-12 08:10:52 -06:00
|
|
|
"""Contains handlers for federation events."""
|
2018-04-17 15:11:19 -06:00
|
|
|
|
|
|
|
import itertools
|
|
|
|
import logging
|
2020-07-20 11:33:04 -06:00
|
|
|
from collections.abc import Container
|
2020-06-16 06:51:47 -06:00
|
|
|
from http import HTTPStatus
|
2020-09-28 08:20:02 -06:00
|
|
|
from typing import TYPE_CHECKING, Dict, Iterable, List, Optional, Sequence, Tuple, Union
|
2018-04-17 15:11:19 -06:00
|
|
|
|
2019-12-05 08:02:35 -07:00
|
|
|
import attr
|
2018-07-09 00:09:20 -06:00
|
|
|
from signedjson.key import decode_verify_key_bytes
|
|
|
|
from signedjson.sign import verify_signed_json
|
2016-02-23 08:11:25 -07:00
|
|
|
from unpaddedbase64 import decode_base64
|
2014-08-12 08:10:52 -06:00
|
|
|
|
2018-07-09 00:09:20 -06:00
|
|
|
from twisted.internet import defer
|
2014-08-12 08:10:52 -06:00
|
|
|
|
2019-10-18 11:43:36 -06:00
|
|
|
from synapse import event_auth
|
2020-06-10 10:44:34 -06:00
|
|
|
from synapse.api.constants import (
|
|
|
|
EventTypes,
|
|
|
|
Membership,
|
|
|
|
RejectedReason,
|
|
|
|
RoomEncryptionAlgorithms,
|
|
|
|
)
|
2014-11-26 09:06:20 -07:00
|
|
|
from synapse.api.errors import (
|
2018-07-09 00:09:20 -06:00
|
|
|
AuthError,
|
|
|
|
CodeMessageException,
|
2019-06-12 03:31:37 -06:00
|
|
|
Codes,
|
2018-01-22 11:11:18 -07:00
|
|
|
FederationDeniedError,
|
2018-07-09 00:09:20 -06:00
|
|
|
FederationError,
|
2020-05-22 04:39:20 -06:00
|
|
|
HttpResponseException,
|
2020-07-16 08:17:31 -06:00
|
|
|
NotFoundError,
|
2019-06-03 02:56:45 -06:00
|
|
|
RequestSendFailed,
|
2018-07-09 00:09:20 -06:00
|
|
|
SynapseError,
|
2014-11-26 09:06:20 -07:00
|
|
|
)
|
2020-01-27 07:30:57 -07:00
|
|
|
from synapse.api.room_versions import KNOWN_ROOM_VERSIONS, RoomVersion, RoomVersions
|
2019-01-23 13:05:44 -07:00
|
|
|
from synapse.crypto.event_signing import compute_event_signature
|
2019-02-12 03:31:21 -07:00
|
|
|
from synapse.event_auth import auth_types_for_event
|
2020-01-28 07:18:29 -07:00
|
|
|
from synapse.events import EventBase
|
2019-11-01 10:19:09 -06:00
|
|
|
from synapse.events.snapshot import EventContext
|
2018-07-09 00:09:20 -06:00
|
|
|
from synapse.events.validator import EventValidator
|
2020-03-30 12:06:52 -06:00
|
|
|
from synapse.handlers._base import BaseHandler
|
2020-11-19 03:05:33 -07:00
|
|
|
from synapse.http.servlet import assert_params_in_dict
|
2019-07-03 08:07:04 -06:00
|
|
|
from synapse.logging.context import (
|
|
|
|
make_deferred_yieldable,
|
|
|
|
nested_logging_context,
|
|
|
|
preserve_fn,
|
|
|
|
run_in_background,
|
|
|
|
)
|
|
|
|
from synapse.logging.utils import log_function
|
2020-07-10 11:23:17 -06:00
|
|
|
from synapse.metrics.background_process_metrics import run_as_background_process
|
2020-01-30 10:06:38 -07:00
|
|
|
from synapse.replication.http.devices import ReplicationUserDevicesResyncRestServlet
|
2018-07-26 04:44:22 -06:00
|
|
|
from synapse.replication.http.federation import (
|
2018-08-09 03:29:48 -06:00
|
|
|
ReplicationCleanRoomRestServlet,
|
2018-07-26 04:44:22 -06:00
|
|
|
ReplicationFederationSendEventsRestServlet,
|
2020-11-13 09:24:04 -07:00
|
|
|
ReplicationStoreRoomOnOutlierMembershipRestServlet,
|
2018-07-26 04:44:22 -06:00
|
|
|
)
|
2020-09-28 08:20:02 -06:00
|
|
|
from synapse.state import StateResolutionStore
|
2020-08-05 14:38:57 -06:00
|
|
|
from synapse.storage.databases.main.events_worker import EventRedactBehaviour
|
2020-08-28 05:28:53 -06:00
|
|
|
from synapse.types import (
|
|
|
|
JsonDict,
|
|
|
|
MutableStateMap,
|
2020-09-24 06:24:17 -06:00
|
|
|
PersistedEventPosition,
|
|
|
|
RoomStreamToken,
|
2020-08-28 05:28:53 -06:00
|
|
|
StateMap,
|
|
|
|
UserID,
|
|
|
|
get_domain_from_id,
|
|
|
|
)
|
2019-12-16 05:26:28 -07:00
|
|
|
from synapse.util.async_helpers import Linearizer, concurrently_execute
|
2015-05-12 03:35:45 -06:00
|
|
|
from synapse.util.retryutils import NotRetryingDestination
|
2020-03-30 12:06:52 -06:00
|
|
|
from synapse.util.stringutils import shortstr
|
2018-07-16 04:38:45 -06:00
|
|
|
from synapse.visibility import filter_events_for_server
|
2015-05-12 03:35:45 -06:00
|
|
|
|
2020-09-28 08:20:02 -06:00
|
|
|
if TYPE_CHECKING:
|
|
|
|
from synapse.server import HomeServer
|
|
|
|
|
2014-08-12 08:10:52 -06:00
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
|
|
|
|
2020-09-14 10:50:06 -06:00
|
|
|
@attr.s(slots=True)
|
2019-12-05 08:02:35 -07:00
|
|
|
class _NewEventInfo:
|
|
|
|
"""Holds information about a received event, ready for passing to _handle_new_events
|
|
|
|
|
|
|
|
Attributes:
|
|
|
|
event: the received event
|
|
|
|
|
|
|
|
state: the state at that event
|
|
|
|
|
|
|
|
auth_events: the auth_event map for that event
|
|
|
|
"""
|
|
|
|
|
|
|
|
event = attr.ib(type=EventBase)
|
|
|
|
state = attr.ib(type=Optional[Sequence[EventBase]], default=None)
|
2020-08-28 05:28:53 -06:00
|
|
|
auth_events = attr.ib(type=Optional[MutableStateMap[EventBase]], default=None)
|
2019-12-05 08:02:35 -07:00
|
|
|
|
|
|
|
|
2014-08-12 08:10:52 -06:00
|
|
|
class FederationHandler(BaseHandler):
|
2014-08-26 12:49:42 -06:00
|
|
|
"""Handles events that originated from federation.
|
|
|
|
Responsible for:
|
|
|
|
a) handling received Pdus before handing them on as Events to the rest
|
2020-10-23 10:38:40 -06:00
|
|
|
of the homeserver (including auth and state conflict resolutions)
|
2014-08-26 12:49:42 -06:00
|
|
|
b) converting events that were produced by local clients that may need
|
2019-11-12 06:08:12 -07:00
|
|
|
to be sent to remote homeservers.
|
2014-11-12 09:20:21 -07:00
|
|
|
c) doing the necessary dances to invite remote users and join remote
|
|
|
|
rooms.
|
2014-08-26 12:49:42 -06:00
|
|
|
"""
|
2014-08-12 08:10:52 -06:00
|
|
|
|
2020-09-28 08:20:02 -06:00
|
|
|
def __init__(self, hs: "HomeServer"):
|
2020-09-18 07:56:44 -06:00
|
|
|
super().__init__(hs)
|
2014-08-21 07:38:22 -06:00
|
|
|
|
2015-11-05 09:43:19 -07:00
|
|
|
self.hs = hs
|
|
|
|
|
2019-01-30 03:53:17 -07:00
|
|
|
self.store = hs.get_datastore()
|
2019-10-23 05:02:36 -06:00
|
|
|
self.storage = hs.get_storage()
|
2019-10-23 10:25:54 -06:00
|
|
|
self.state_store = self.storage.state
|
2018-07-31 08:44:05 -06:00
|
|
|
self.federation_client = hs.get_federation_client()
|
2014-08-26 12:49:42 -06:00
|
|
|
self.state_handler = hs.get_state_handler()
|
2020-09-28 08:20:02 -06:00
|
|
|
self._state_resolution_handler = hs.get_state_resolution_handler()
|
2014-08-26 12:49:42 -06:00
|
|
|
self.server_name = hs.hostname
|
2014-11-14 09:45:39 -07:00
|
|
|
self.keyring = hs.get_keyring()
|
2017-05-18 11:17:40 -06:00
|
|
|
self.action_generator = hs.get_action_generator()
|
2017-06-30 09:20:30 -06:00
|
|
|
self.is_mine_id = hs.is_mine_id
|
2017-10-03 06:53:09 -06:00
|
|
|
self.spam_checker = hs.get_spam_checker()
|
2018-01-15 09:52:07 -07:00
|
|
|
self.event_creation_handler = hs.get_event_creation_handler()
|
2019-12-03 12:19:45 -07:00
|
|
|
self._message_handler = hs.get_message_handler()
|
2018-05-18 04:18:39 -06:00
|
|
|
self._server_notices_mxid = hs.config.server_notices_mxid
|
2018-07-25 09:32:05 -06:00
|
|
|
self.config = hs.config
|
|
|
|
self.http_client = hs.get_simple_http_client()
|
2020-05-22 09:11:35 -06:00
|
|
|
self._instance_name = hs.get_instance_name()
|
2020-05-22 07:21:54 -06:00
|
|
|
self._replication = hs.get_replication_data_handler()
|
2014-08-26 12:49:42 -06:00
|
|
|
|
2020-05-22 09:11:35 -06:00
|
|
|
self._send_events = ReplicationFederationSendEventsRestServlet.make_client(hs)
|
2019-06-20 03:32:02 -06:00
|
|
|
self._clean_room_for_join_client = ReplicationCleanRoomRestServlet.make_client(
|
|
|
|
hs
|
2018-08-09 03:29:48 -06:00
|
|
|
)
|
2014-08-26 12:49:42 -06:00
|
|
|
|
2020-01-30 10:06:38 -07:00
|
|
|
if hs.config.worker_app:
|
|
|
|
self._user_device_resync = ReplicationUserDevicesResyncRestServlet.make_client(
|
|
|
|
hs
|
|
|
|
)
|
2020-11-13 09:24:04 -07:00
|
|
|
self._maybe_store_room_on_outlier_membership = ReplicationStoreRoomOnOutlierMembershipRestServlet.make_client(
|
2020-02-26 09:58:33 -07:00
|
|
|
hs
|
|
|
|
)
|
2020-01-30 10:06:38 -07:00
|
|
|
else:
|
|
|
|
self._device_list_updater = hs.get_device_handler().device_list_updater
|
2020-11-13 09:24:04 -07:00
|
|
|
self._maybe_store_room_on_outlier_membership = (
|
|
|
|
self.store.maybe_store_room_on_outlier_membership
|
|
|
|
)
|
2020-01-30 10:06:38 -07:00
|
|
|
|
2020-09-29 08:57:36 -06:00
|
|
|
# When joining a room we need to queue any events for that room up.
|
|
|
|
# For each room, a list of (pdu, origin) tuples.
|
|
|
|
self.room_queues = {} # type: Dict[str, List[Tuple[EventBase, str]]]
|
2017-03-09 09:20:13 -07:00
|
|
|
self._room_pdu_linearizer = Linearizer("fed_room_pdu")
|
|
|
|
|
2019-06-12 03:31:37 -06:00
|
|
|
self.third_party_event_rules = hs.get_third_party_event_rules()
|
|
|
|
|
2019-12-03 12:19:45 -07:00
|
|
|
self._ephemeral_messages_enabled = hs.config.enable_ephemeral_messages
|
|
|
|
|
2019-12-10 10:01:37 -07:00
|
|
|
async def on_receive_pdu(self, origin, pdu, sent_to_us_directly=False) -> None:
|
2017-03-09 09:20:13 -07:00
|
|
|
""" Process a PDU received via a federation /send/ transaction, or
|
|
|
|
via backfill of missing prev_events
|
|
|
|
|
|
|
|
Args:
|
|
|
|
origin (str): server which initiated the /send/ transaction. Will
|
|
|
|
be used to fetch missing events or state.
|
|
|
|
pdu (FrozenEvent): received PDU
|
2018-09-20 06:06:55 -06:00
|
|
|
sent_to_us_directly (bool): True if this event was pushed to us; False if
|
|
|
|
we pulled it as the result of a missing prev_event.
|
2017-03-09 09:20:13 -07:00
|
|
|
"""
|
|
|
|
|
2018-09-19 10:28:18 -06:00
|
|
|
room_id = pdu.room_id
|
|
|
|
event_id = pdu.event_id
|
|
|
|
|
2019-12-11 07:32:25 -07:00
|
|
|
logger.info("handling received PDU: %s", pdu)
|
2018-09-19 10:28:18 -06:00
|
|
|
|
2017-03-09 09:20:13 -07:00
|
|
|
# We reprocess pdus when we have seen them only as outliers
|
2019-12-10 10:01:37 -07:00
|
|
|
existing = await self.store.get_event(
|
2019-06-20 03:32:02 -06:00
|
|
|
event_id, allow_none=True, allow_rejected=True
|
2017-03-09 09:20:13 -07:00
|
|
|
)
|
|
|
|
|
|
|
|
# FIXME: Currently we fetch an event again when we already have it
|
|
|
|
# if it has been marked as an outlier.
|
|
|
|
|
2019-06-20 03:32:02 -06:00
|
|
|
already_seen = existing and (
|
|
|
|
not existing.internal_metadata.is_outlier()
|
|
|
|
or pdu.internal_metadata.is_outlier()
|
2017-03-09 09:20:13 -07:00
|
|
|
)
|
|
|
|
if already_seen:
|
2018-09-19 10:28:18 -06:00
|
|
|
logger.debug("[%s %s]: Already seen pdu", room_id, event_id)
|
2017-03-09 09:20:13 -07:00
|
|
|
return
|
|
|
|
|
2018-04-17 15:11:19 -06:00
|
|
|
# do some initial sanity-checking of the event. In particular, make
|
|
|
|
# sure it doesn't have hundreds of prev_events or auth_events, which
|
2018-04-17 16:41:12 -06:00
|
|
|
# could cause a huge state resolution or cascade of event fetches.
|
|
|
|
try:
|
|
|
|
self._sanity_check_event(pdu)
|
|
|
|
except SynapseError as err:
|
2019-10-31 04:23:24 -06:00
|
|
|
logger.warning(
|
2019-06-20 03:32:02 -06:00
|
|
|
"[%s %s] Received event failed sanity checks", room_id, event_id
|
2018-04-17 15:11:19 -06:00
|
|
|
)
|
2019-06-20 03:32:02 -06:00
|
|
|
raise FederationError("ERROR", err.code, err.msg, affected=pdu.event_id)
|
2018-04-17 15:11:19 -06:00
|
|
|
|
2017-03-14 05:26:57 -06:00
|
|
|
# If we are currently in the process of joining this room, then we
|
|
|
|
# queue up events for later processing.
|
2018-09-19 10:28:18 -06:00
|
|
|
if room_id in self.room_queues:
|
|
|
|
logger.info(
|
|
|
|
"[%s %s] Queuing PDU from %s for now: join in progress",
|
2019-06-20 03:32:02 -06:00
|
|
|
room_id,
|
|
|
|
event_id,
|
|
|
|
origin,
|
2018-09-19 10:28:18 -06:00
|
|
|
)
|
|
|
|
self.room_queues[room_id].append((pdu, origin))
|
2017-03-14 05:26:57 -06:00
|
|
|
return
|
|
|
|
|
2018-11-09 04:34:45 -07:00
|
|
|
# If we're not in the room just ditch the event entirely. This is
|
|
|
|
# probably an old server that has come back and thinks we're still in
|
|
|
|
# the room (or we've been rejoined to the room by a state reset).
|
2017-10-03 04:09:51 -06:00
|
|
|
#
|
2018-11-09 04:34:45 -07:00
|
|
|
# Note that if we were never in the room then we would have already
|
|
|
|
# dropped the event, since we wouldn't know the room version.
|
2019-12-10 10:01:37 -07:00
|
|
|
is_in_room = await self.auth.check_host_in_room(room_id, self.server_name)
|
2017-10-03 04:09:51 -06:00
|
|
|
if not is_in_room:
|
2018-11-09 04:34:45 -07:00
|
|
|
logger.info(
|
|
|
|
"[%s %s] Ignoring PDU from %s as we're not in the room",
|
2019-06-20 03:32:02 -06:00
|
|
|
room_id,
|
|
|
|
event_id,
|
|
|
|
origin,
|
2017-10-03 04:09:51 -06:00
|
|
|
)
|
2019-07-23 07:00:55 -06:00
|
|
|
return None
|
2017-10-03 04:09:51 -06:00
|
|
|
|
2017-03-09 09:20:13 -07:00
|
|
|
state = None
|
|
|
|
|
|
|
|
# Get missing pdus if necessary.
|
|
|
|
if not pdu.internal_metadata.is_outlier():
|
|
|
|
# We only backfill backwards to the min depth.
|
2019-12-10 10:01:37 -07:00
|
|
|
min_depth = await self.get_min_depth_for_context(pdu.room_id)
|
2017-03-09 09:20:13 -07:00
|
|
|
|
2019-06-20 03:32:02 -06:00
|
|
|
logger.debug("[%s %s] min_depth: %d", room_id, event_id, min_depth)
|
2017-03-09 09:20:13 -07:00
|
|
|
|
2018-11-05 06:35:15 -07:00
|
|
|
prevs = set(pdu.prev_event_ids())
|
2020-07-01 09:59:31 -06:00
|
|
|
seen = await self.store.have_events_in_timeline(prevs)
|
2017-03-09 09:20:13 -07:00
|
|
|
|
2020-01-07 05:08:58 -07:00
|
|
|
if min_depth is not None and pdu.depth < min_depth:
|
2017-03-09 09:20:13 -07:00
|
|
|
# This is so that we don't notify the user about this
|
|
|
|
# message, to work around the fact that some events will
|
|
|
|
# reference really really old events we really don't want to
|
|
|
|
# send to the clients.
|
|
|
|
pdu.internal_metadata.outlier = True
|
2020-01-07 05:08:58 -07:00
|
|
|
elif min_depth is not None and pdu.depth > min_depth:
|
2018-09-19 10:28:18 -06:00
|
|
|
missing_prevs = prevs - seen
|
2018-09-20 06:06:55 -06:00
|
|
|
if sent_to_us_directly and missing_prevs:
|
2017-03-09 09:20:13 -07:00
|
|
|
# If we're missing stuff, ensure we only fetch stuff one
|
|
|
|
# at a time.
|
|
|
|
logger.info(
|
2018-09-19 10:28:18 -06:00
|
|
|
"[%s %s] Acquiring room lock to fetch %d missing prev_events: %s",
|
2019-06-20 03:32:02 -06:00
|
|
|
room_id,
|
|
|
|
event_id,
|
|
|
|
len(missing_prevs),
|
|
|
|
shortstr(missing_prevs),
|
2017-03-09 09:20:13 -07:00
|
|
|
)
|
2019-12-10 10:01:37 -07:00
|
|
|
with (await self._room_pdu_linearizer.queue(pdu.room_id)):
|
2017-03-09 09:20:13 -07:00
|
|
|
logger.info(
|
2018-09-19 10:28:18 -06:00
|
|
|
"[%s %s] Acquired room lock to fetch %d missing prev_events",
|
2019-06-20 03:32:02 -06:00
|
|
|
room_id,
|
|
|
|
event_id,
|
|
|
|
len(missing_prevs),
|
2017-03-09 09:20:13 -07:00
|
|
|
)
|
|
|
|
|
2019-12-11 07:32:25 -07:00
|
|
|
try:
|
2019-12-10 10:01:37 -07:00
|
|
|
await self._get_missing_events_for_pdu(
|
2019-12-11 07:32:25 -07:00
|
|
|
origin, pdu, prevs, min_depth
|
|
|
|
)
|
|
|
|
except Exception as e:
|
|
|
|
raise Exception(
|
|
|
|
"Error fetching missing prev_events for %s: %s"
|
|
|
|
% (event_id, e)
|
2020-09-28 06:15:00 -06:00
|
|
|
) from e
|
2017-04-28 04:26:46 -06:00
|
|
|
|
|
|
|
# Update the set of things we've seen after trying to
|
|
|
|
# fetch the missing stuff
|
2020-07-01 09:59:31 -06:00
|
|
|
seen = await self.store.have_events_in_timeline(prevs)
|
2017-05-03 03:06:43 -06:00
|
|
|
|
|
|
|
if not prevs - seen:
|
2017-04-28 04:55:25 -06:00
|
|
|
logger.info(
|
2018-09-19 10:28:18 -06:00
|
|
|
"[%s %s] Found all missing prev_events",
|
2019-06-20 03:32:02 -06:00
|
|
|
room_id,
|
|
|
|
event_id,
|
2017-04-28 04:55:25 -06:00
|
|
|
)
|
2017-03-09 09:20:13 -07:00
|
|
|
|
2018-09-20 06:06:55 -06:00
|
|
|
if prevs - seen:
|
|
|
|
# We've still not been able to get all of the prev_events for this event.
|
|
|
|
#
|
|
|
|
# In this case, we need to fall back to asking another server in the
|
|
|
|
# federation for the state at this event. That's ok provided we then
|
|
|
|
# resolve the state against other bits of the DAG before using it (which
|
|
|
|
# will ensure that you can't just take over a room by sending an event,
|
|
|
|
# withholding its prev_events, and declaring yourself to be an admin in
|
|
|
|
# the subsequent state request).
|
|
|
|
#
|
|
|
|
# Now, if we're pulling this event as a missing prev_event, then clearly
|
|
|
|
# this event is not going to become the only forward-extremity and we are
|
|
|
|
# guaranteed to resolve its state against our existing forward
|
|
|
|
# extremities, so that should be fine.
|
|
|
|
#
|
|
|
|
# On the other hand, if this event was pushed to us, it is possible for
|
|
|
|
# it to become the only forward-extremity in the room, and we would then
|
|
|
|
# trust its state to be the state for the whole room. This is very bad.
|
|
|
|
# Further, if the event was pushed to us, there is no excuse for us not to
|
|
|
|
# have all the prev_events. We therefore reject any such events.
|
|
|
|
#
|
|
|
|
# XXX this really feels like it could/should be merged with the above,
|
|
|
|
# but there is an interaction with min_depth that I'm not really
|
|
|
|
# following.
|
|
|
|
|
|
|
|
if sent_to_us_directly:
|
2019-10-31 04:23:24 -06:00
|
|
|
logger.warning(
|
2018-10-16 13:37:16 -06:00
|
|
|
"[%s %s] Rejecting: failed to fetch %d prev events: %s",
|
2019-06-20 03:32:02 -06:00
|
|
|
room_id,
|
|
|
|
event_id,
|
|
|
|
len(prevs - seen),
|
|
|
|
shortstr(prevs - seen),
|
2018-09-20 06:06:55 -06:00
|
|
|
)
|
|
|
|
raise FederationError(
|
|
|
|
"ERROR",
|
|
|
|
403,
|
|
|
|
(
|
|
|
|
"Your server isn't divulging details about prev_events "
|
|
|
|
"referenced in this event."
|
|
|
|
),
|
|
|
|
affected=pdu.event_id,
|
|
|
|
)
|
|
|
|
|
2019-12-11 07:32:25 -07:00
|
|
|
logger.info(
|
|
|
|
"Event %s is missing prev_events: calculating state for a "
|
|
|
|
"backwards extremity",
|
|
|
|
event_id,
|
|
|
|
)
|
|
|
|
|
2018-09-26 12:49:59 -06:00
|
|
|
# Calculate the state after each of the previous events, and
|
|
|
|
# resolve them to find the correct state at the current event.
|
2019-06-20 03:32:02 -06:00
|
|
|
event_map = {event_id: pdu}
|
2018-06-27 04:27:32 -06:00
|
|
|
try:
|
2018-06-27 04:36:03 -06:00
|
|
|
# Get the state of the events we know about
|
2019-12-10 10:01:37 -07:00
|
|
|
ours = await self.state_store.get_state_groups_ids(room_id, seen)
|
2018-09-26 02:52:56 -06:00
|
|
|
|
|
|
|
# state_maps is a list of mappings from (type, state_key) to event_id
|
2020-04-24 12:36:38 -06:00
|
|
|
state_maps = list(ours.values()) # type: List[StateMap[str]]
|
2018-09-26 02:52:56 -06:00
|
|
|
|
|
|
|
# we don't need this any more, let's delete it.
|
|
|
|
del ours
|
2018-06-27 04:27:32 -06:00
|
|
|
|
2018-06-27 04:36:03 -06:00
|
|
|
# Ask the remote server for the states we don't
|
|
|
|
# know about
|
2018-06-27 04:27:32 -06:00
|
|
|
for p in prevs - seen:
|
2018-09-19 10:28:18 -06:00
|
|
|
logger.info(
|
2019-12-11 07:32:25 -07:00
|
|
|
"Requesting state at missing prev_event %s", event_id,
|
2018-09-19 10:28:18 -06:00
|
|
|
)
|
2018-09-27 04:25:34 -06:00
|
|
|
|
2019-07-03 08:07:04 -06:00
|
|
|
with nested_logging_context(p):
|
2018-09-26 12:17:36 -06:00
|
|
|
# note that if any of the missing prevs share missing state or
|
|
|
|
# auth events, the requests to fetch those events are deduped
|
|
|
|
# by the get_pdu_cache in federation_client.
|
2019-12-16 05:26:28 -07:00
|
|
|
(remote_state, _,) = await self._get_state_for_room(
|
2019-12-11 09:37:51 -07:00
|
|
|
origin, room_id, p, include_event_in_state=True
|
2018-09-26 12:49:59 -06:00
|
|
|
)
|
|
|
|
|
2018-09-26 02:52:56 -06:00
|
|
|
remote_state_map = {
|
2018-09-26 01:09:07 -06:00
|
|
|
(x.type, x.state_key): x.event_id for x in remote_state
|
2018-09-27 04:25:34 -06:00
|
|
|
}
|
2018-09-26 02:52:56 -06:00
|
|
|
state_maps.append(remote_state_map)
|
2018-06-27 04:27:32 -06:00
|
|
|
|
2018-09-26 05:16:13 -06:00
|
|
|
for x in remote_state:
|
|
|
|
event_map[x.event_id] = x
|
|
|
|
|
2020-01-31 03:06:21 -07:00
|
|
|
room_version = await self.store.get_room_version_id(room_id)
|
2020-09-28 08:20:02 -06:00
|
|
|
state_map = await self._state_resolution_handler.resolve_events_with_store(
|
2019-12-13 05:55:32 -07:00
|
|
|
room_id,
|
2019-06-20 03:32:02 -06:00
|
|
|
room_version,
|
|
|
|
state_maps,
|
|
|
|
event_map,
|
2018-10-16 06:58:37 -06:00
|
|
|
state_res_store=StateResolutionStore(self.store),
|
2018-06-27 04:36:03 -06:00
|
|
|
)
|
2018-06-27 04:27:32 -06:00
|
|
|
|
2018-10-24 02:47:49 -06:00
|
|
|
# We need to give _process_received_pdu the actual state events
|
2018-09-26 05:16:13 -06:00
|
|
|
# rather than event ids, so generate that now.
|
2018-10-24 02:47:49 -06:00
|
|
|
|
|
|
|
# First though we need to fetch all the events that are in
|
|
|
|
# state_map, so we can build up the state below.
|
2019-12-10 10:01:37 -07:00
|
|
|
evs = await self.store.get_events(
|
2018-10-16 06:58:37 -06:00
|
|
|
list(state_map.values()),
|
|
|
|
get_prev_content=False,
|
2019-12-11 06:39:47 -07:00
|
|
|
redact_behaviour=EventRedactBehaviour.AS_IS,
|
2018-06-27 04:36:03 -06:00
|
|
|
)
|
2018-10-16 06:58:37 -06:00
|
|
|
event_map.update(evs)
|
2018-06-27 04:27:32 -06:00
|
|
|
|
2020-06-15 05:03:36 -06:00
|
|
|
state = [event_map[e] for e in state_map.values()]
|
2018-06-27 04:27:32 -06:00
|
|
|
except Exception:
|
2019-10-31 04:23:24 -06:00
|
|
|
logger.warning(
|
2018-09-19 10:28:18 -06:00
|
|
|
"[%s %s] Error attempting to resolve state at missing "
|
|
|
|
"prev_events",
|
2019-06-20 03:32:02 -06:00
|
|
|
room_id,
|
|
|
|
event_id,
|
|
|
|
exc_info=True,
|
2018-09-19 10:28:18 -06:00
|
|
|
)
|
2018-06-27 04:27:32 -06:00
|
|
|
raise FederationError(
|
|
|
|
"ERROR",
|
|
|
|
403,
|
|
|
|
"We can't get valid state history.",
|
2018-09-19 10:28:18 -06:00
|
|
|
affected=event_id,
|
2018-06-27 04:27:32 -06:00
|
|
|
)
|
2017-03-09 09:20:13 -07:00
|
|
|
|
2019-12-16 05:26:28 -07:00
|
|
|
await self._process_received_pdu(origin, pdu, state=state)
|
2017-03-09 09:20:13 -07:00
|
|
|
|
2019-12-10 10:01:37 -07:00
|
|
|
async def _get_missing_events_for_pdu(self, origin, pdu, prevs, min_depth):
|
2017-03-09 09:20:13 -07:00
|
|
|
"""
|
|
|
|
Args:
|
|
|
|
origin (str): Origin of the pdu. Will be called to get the missing events
|
|
|
|
pdu: received pdu
|
2017-04-28 05:46:53 -06:00
|
|
|
prevs (set(str)): List of event ids which we are missing
|
2017-03-09 09:20:13 -07:00
|
|
|
min_depth (int): Minimum depth of events to return.
|
|
|
|
"""
|
2018-09-19 10:28:18 -06:00
|
|
|
|
|
|
|
room_id = pdu.room_id
|
|
|
|
event_id = pdu.event_id
|
|
|
|
|
2020-07-01 09:59:31 -06:00
|
|
|
seen = await self.store.have_events_in_timeline(prevs)
|
2017-03-09 09:20:13 -07:00
|
|
|
|
|
|
|
if not prevs - seen:
|
2017-04-28 05:46:53 -06:00
|
|
|
return
|
2017-03-09 09:20:13 -07:00
|
|
|
|
2020-09-03 08:38:32 -06:00
|
|
|
latest_list = await self.store.get_latest_event_ids_in_room(room_id)
|
2017-03-09 09:20:13 -07:00
|
|
|
|
|
|
|
# We add the prev events that we have seen to the latest
|
|
|
|
# list to ensure the remote server doesn't give them to us
|
2020-09-03 08:38:32 -06:00
|
|
|
latest = set(latest_list)
|
2017-03-09 09:20:13 -07:00
|
|
|
latest |= seen
|
|
|
|
|
|
|
|
logger.info(
|
2018-10-16 13:37:16 -06:00
|
|
|
"[%s %s]: Requesting missing events between %s and %s",
|
2019-06-20 03:32:02 -06:00
|
|
|
room_id,
|
|
|
|
event_id,
|
|
|
|
shortstr(latest),
|
|
|
|
event_id,
|
2017-03-09 09:20:13 -07:00
|
|
|
)
|
|
|
|
|
|
|
|
# XXX: we set timeout to 10s to help workaround
|
|
|
|
# https://github.com/matrix-org/synapse/issues/1733.
|
|
|
|
# The reason is to avoid holding the linearizer lock
|
|
|
|
# whilst processing inbound /send transactions, causing
|
|
|
|
# FDs to stack up and block other inbound transactions
|
|
|
|
# which empirically can currently take up to 30 minutes.
|
|
|
|
#
|
|
|
|
# N.B. this explicitly disables retry attempts.
|
|
|
|
#
|
|
|
|
# N.B. this also increases our chances of falling back to
|
|
|
|
# fetching fresh state for the room if the missing event
|
|
|
|
# can't be found, which slightly reduces our security.
|
|
|
|
# it may also increase our DAG extremity count for the room,
|
|
|
|
# causing additional state resolution? See #1760.
|
|
|
|
# However, fetching state doesn't hold the linearizer lock
|
|
|
|
# apparently.
|
|
|
|
#
|
|
|
|
# see https://github.com/matrix-org/synapse/pull/1744
|
2018-09-18 08:02:51 -06:00
|
|
|
#
|
|
|
|
# ----
|
|
|
|
#
|
|
|
|
# Update richvdh 2018/09/18: There are a number of problems with timing this
|
2020-10-23 10:38:40 -06:00
|
|
|
# request out aggressively on the client side:
|
2018-09-18 08:02:51 -06:00
|
|
|
#
|
|
|
|
# - it plays badly with the server-side rate-limiter, which starts tarpitting you
|
|
|
|
# if you send too many requests at once, so you end up with the server carefully
|
|
|
|
# working through the backlog of your requests, which you have already timed
|
|
|
|
# out.
|
|
|
|
#
|
|
|
|
# - for this request in particular, we now (as of
|
|
|
|
# https://github.com/matrix-org/synapse/pull/3456) reject any PDUs where the
|
|
|
|
# server can't produce a plausible-looking set of prev_events - so we becone
|
|
|
|
# much more likely to reject the event.
|
|
|
|
#
|
|
|
|
# - contrary to what it says above, we do *not* fall back to fetching fresh state
|
|
|
|
# for the room if get_missing_events times out. Rather, we give up processing
|
|
|
|
# the PDU whose prevs we are missing, which then makes it much more likely that
|
|
|
|
# we'll end up back here for the *next* PDU in the list, which exacerbates the
|
|
|
|
# problem.
|
|
|
|
#
|
2020-10-23 10:38:40 -06:00
|
|
|
# - the aggressive 10s timeout was introduced to deal with incoming federation
|
2018-09-18 08:02:51 -06:00
|
|
|
# requests taking 8 hours to process. It's not entirely clear why that was going
|
|
|
|
# on; certainly there were other issues causing traffic storms which are now
|
|
|
|
# resolved, and I think in any case we may be more sensible about our locking
|
|
|
|
# now. We're *certainly* more sensible about our logging.
|
|
|
|
#
|
2020-10-23 10:38:40 -06:00
|
|
|
# All that said: Let's try increasing the timeout to 60s and see what happens.
|
2017-03-09 09:20:13 -07:00
|
|
|
|
2019-04-10 03:39:54 -06:00
|
|
|
try:
|
2019-12-10 10:01:37 -07:00
|
|
|
missing_events = await self.federation_client.get_missing_events(
|
2019-04-10 03:39:54 -06:00
|
|
|
origin,
|
|
|
|
room_id,
|
|
|
|
earliest_events_ids=list(latest),
|
|
|
|
latest_events=[pdu],
|
|
|
|
limit=10,
|
|
|
|
min_depth=min_depth,
|
|
|
|
timeout=60000,
|
|
|
|
)
|
2020-06-03 07:12:13 -06:00
|
|
|
except (RequestSendFailed, HttpResponseException, NotRetryingDestination) as e:
|
2019-04-10 03:39:54 -06:00
|
|
|
# We failed to get the missing events, but since we need to handle
|
|
|
|
# the case of `get_missing_events` not returning the necessary
|
|
|
|
# events anyway, it is safe to simply log the error and continue.
|
2019-10-31 04:23:24 -06:00
|
|
|
logger.warning(
|
|
|
|
"[%s %s]: Failed to get prev_events: %s", room_id, event_id, e
|
|
|
|
)
|
2019-04-10 03:39:54 -06:00
|
|
|
return
|
2017-03-09 09:20:13 -07:00
|
|
|
|
2017-04-27 07:38:21 -06:00
|
|
|
logger.info(
|
2018-09-19 10:28:18 -06:00
|
|
|
"[%s %s]: Got %d prev_events: %s",
|
2019-06-20 03:32:02 -06:00
|
|
|
room_id,
|
|
|
|
event_id,
|
|
|
|
len(missing_events),
|
|
|
|
shortstr(missing_events),
|
2017-04-27 07:38:21 -06:00
|
|
|
)
|
|
|
|
|
2017-03-09 09:20:13 -07:00
|
|
|
# We want to sort these by depth so we process them and
|
|
|
|
# tell clients about them in order.
|
|
|
|
missing_events.sort(key=lambda x: x.depth)
|
|
|
|
|
2018-09-19 10:28:18 -06:00
|
|
|
for ev in missing_events:
|
|
|
|
logger.info(
|
|
|
|
"[%s %s] Handling received prev_event %s",
|
2019-06-20 03:32:02 -06:00
|
|
|
room_id,
|
|
|
|
event_id,
|
|
|
|
ev.event_id,
|
2018-09-19 10:28:18 -06:00
|
|
|
)
|
2019-07-03 08:07:04 -06:00
|
|
|
with nested_logging_context(ev.event_id):
|
2018-09-27 04:25:34 -06:00
|
|
|
try:
|
2019-12-10 10:01:37 -07:00
|
|
|
await self.on_receive_pdu(origin, ev, sent_to_us_directly=False)
|
2018-09-27 04:25:34 -06:00
|
|
|
except FederationError as e:
|
|
|
|
if e.code == 403:
|
2019-10-31 04:23:24 -06:00
|
|
|
logger.warning(
|
2018-09-27 04:25:34 -06:00
|
|
|
"[%s %s] Received prev_event %s failed history check.",
|
2019-06-20 03:32:02 -06:00
|
|
|
room_id,
|
|
|
|
event_id,
|
|
|
|
ev.event_id,
|
2018-09-27 04:25:34 -06:00
|
|
|
)
|
|
|
|
else:
|
|
|
|
raise
|
2017-03-09 09:20:13 -07:00
|
|
|
|
2019-12-10 10:25:18 -07:00
|
|
|
async def _get_state_for_room(
|
2019-12-11 09:37:51 -07:00
|
|
|
self,
|
|
|
|
destination: str,
|
|
|
|
room_id: str,
|
|
|
|
event_id: str,
|
|
|
|
include_event_in_state: bool = False,
|
2019-12-10 10:25:18 -07:00
|
|
|
) -> Tuple[List[EventBase], List[EventBase]]:
|
2019-12-10 10:42:46 -07:00
|
|
|
"""Requests all of the room state at a given event from a remote homeserver.
|
|
|
|
|
|
|
|
Args:
|
2019-12-11 09:37:51 -07:00
|
|
|
destination: The remote homeserver to query for the state.
|
2019-12-10 10:25:18 -07:00
|
|
|
room_id: The id of the room we're interested in.
|
|
|
|
event_id: The id of the event we want the state at.
|
2019-12-11 09:37:51 -07:00
|
|
|
include_event_in_state: if true, the event itself will be included in the
|
|
|
|
returned state event list.
|
2019-12-10 10:42:46 -07:00
|
|
|
|
|
|
|
Returns:
|
2019-12-11 09:37:51 -07:00
|
|
|
A list of events in the state, possibly including the event itself, and
|
|
|
|
a list of events in the auth chain for the given event.
|
2019-12-10 10:42:46 -07:00
|
|
|
"""
|
|
|
|
(
|
|
|
|
state_event_ids,
|
|
|
|
auth_event_ids,
|
2019-12-10 10:25:18 -07:00
|
|
|
) = await self.federation_client.get_room_state_ids(
|
2019-12-10 10:42:46 -07:00
|
|
|
destination, room_id, event_id=event_id
|
|
|
|
)
|
|
|
|
|
|
|
|
desired_events = set(state_event_ids + auth_event_ids)
|
2019-12-11 09:37:51 -07:00
|
|
|
|
|
|
|
if include_event_in_state:
|
|
|
|
desired_events.add(event_id)
|
|
|
|
|
2019-12-10 10:25:18 -07:00
|
|
|
event_map = await self._get_events_from_store_or_dest(
|
2019-12-10 10:42:46 -07:00
|
|
|
destination, room_id, desired_events
|
|
|
|
)
|
|
|
|
|
|
|
|
failed_to_fetch = desired_events - event_map.keys()
|
|
|
|
if failed_to_fetch:
|
|
|
|
logger.warning(
|
2019-12-11 07:32:25 -07:00
|
|
|
"Failed to fetch missing state/auth events for %s %s",
|
|
|
|
event_id,
|
2019-12-10 10:42:46 -07:00
|
|
|
failed_to_fetch,
|
|
|
|
)
|
|
|
|
|
2019-12-11 09:37:51 -07:00
|
|
|
remote_state = [
|
|
|
|
event_map[e_id] for e_id in state_event_ids if e_id in event_map
|
|
|
|
]
|
2019-12-10 10:42:46 -07:00
|
|
|
|
2019-12-11 09:37:51 -07:00
|
|
|
if include_event_in_state:
|
|
|
|
remote_event = event_map.get(event_id)
|
|
|
|
if not remote_event:
|
|
|
|
raise Exception("Unable to get missing prev_event %s" % (event_id,))
|
2019-12-16 06:14:37 -07:00
|
|
|
if remote_event.is_state() and remote_event.rejected_reason is None:
|
2019-12-11 09:37:51 -07:00
|
|
|
remote_state.append(remote_event)
|
|
|
|
|
|
|
|
auth_chain = [event_map[e_id] for e_id in auth_event_ids if e_id in event_map]
|
2019-12-10 10:42:46 -07:00
|
|
|
auth_chain.sort(key=lambda e: e.depth)
|
|
|
|
|
2019-12-11 09:37:51 -07:00
|
|
|
return remote_state, auth_chain
|
2019-12-10 10:42:46 -07:00
|
|
|
|
2019-12-10 10:25:18 -07:00
|
|
|
async def _get_events_from_store_or_dest(
|
|
|
|
self, destination: str, room_id: str, event_ids: Iterable[str]
|
|
|
|
) -> Dict[str, EventBase]:
|
2019-12-10 10:42:46 -07:00
|
|
|
"""Fetch events from a remote destination, checking if we already have them.
|
|
|
|
|
2019-12-16 05:26:28 -07:00
|
|
|
Persists any events we don't already have as outliers.
|
2019-12-10 10:42:46 -07:00
|
|
|
|
2019-12-12 05:57:45 -07:00
|
|
|
If we fail to fetch any of the events, a warning will be logged, and the event
|
|
|
|
will be omitted from the result. Likewise, any events which turn out not to
|
|
|
|
be in the given room.
|
|
|
|
|
2020-07-10 11:15:35 -06:00
|
|
|
This function *does not* automatically get missing auth events of the
|
|
|
|
newly fetched events. Callers must include the full auth chain of
|
|
|
|
of the missing events in the `event_ids` argument, to ensure that any
|
|
|
|
missing auth events are correctly fetched.
|
|
|
|
|
2019-12-10 10:42:46 -07:00
|
|
|
Returns:
|
2019-12-10 10:25:18 -07:00
|
|
|
map from event_id to event
|
2019-12-10 10:42:46 -07:00
|
|
|
"""
|
2019-12-10 10:25:18 -07:00
|
|
|
fetched_events = await self.store.get_events(event_ids, allow_rejected=True)
|
2019-12-10 10:42:46 -07:00
|
|
|
|
|
|
|
missing_events = set(event_ids) - fetched_events.keys()
|
|
|
|
|
2019-12-12 05:57:45 -07:00
|
|
|
if missing_events:
|
|
|
|
logger.debug(
|
|
|
|
"Fetching unknown state/auth events %s for room %s",
|
|
|
|
missing_events,
|
|
|
|
room_id,
|
|
|
|
)
|
2019-12-10 10:42:46 -07:00
|
|
|
|
2019-12-16 05:26:28 -07:00
|
|
|
await self._get_events_and_persist(
|
|
|
|
destination=destination, room_id=room_id, events=missing_events
|
|
|
|
)
|
2019-12-10 10:42:46 -07:00
|
|
|
|
2019-12-16 05:26:28 -07:00
|
|
|
# we need to make sure we re-load from the database to get the rejected
|
|
|
|
# state correct.
|
|
|
|
fetched_events.update(
|
|
|
|
(await self.store.get_events(missing_events, allow_rejected=True))
|
|
|
|
)
|
2019-12-12 05:57:45 -07:00
|
|
|
|
|
|
|
# check for events which were in the wrong room.
|
|
|
|
#
|
|
|
|
# this can happen if a remote server claims that the state or
|
|
|
|
# auth_events at an event in room A are actually events in room B
|
|
|
|
|
2020-02-21 05:15:07 -07:00
|
|
|
bad_events = [
|
2019-12-12 05:57:45 -07:00
|
|
|
(event_id, event.room_id)
|
|
|
|
for event_id, event in fetched_events.items()
|
|
|
|
if event.room_id != room_id
|
2020-02-21 05:15:07 -07:00
|
|
|
]
|
2019-12-12 05:57:45 -07:00
|
|
|
|
|
|
|
for bad_event_id, bad_room_id in bad_events:
|
|
|
|
# This is a bogus situation, but since we may only discover it a long time
|
|
|
|
# after it happened, we try our best to carry on, by just omitting the
|
|
|
|
# bad events from the returned auth/state set.
|
|
|
|
logger.warning(
|
|
|
|
"Remote server %s claims event %s in room %s is an auth/state "
|
|
|
|
"event in room %s",
|
|
|
|
destination,
|
|
|
|
bad_event_id,
|
|
|
|
bad_room_id,
|
|
|
|
room_id,
|
2019-12-10 10:42:46 -07:00
|
|
|
)
|
2019-12-18 02:51:51 -07:00
|
|
|
|
2019-12-12 05:57:45 -07:00
|
|
|
del fetched_events[bad_event_id]
|
2019-12-10 10:42:46 -07:00
|
|
|
|
|
|
|
return fetched_events
|
|
|
|
|
2019-12-16 05:26:28 -07:00
|
|
|
async def _process_received_pdu(
|
|
|
|
self, origin: str, event: EventBase, state: Optional[Iterable[EventBase]],
|
|
|
|
):
|
2017-03-09 09:20:13 -07:00
|
|
|
""" Called when we have a new pdu. We need to do auth checks and put it
|
|
|
|
through the StateHandler.
|
2019-12-16 05:26:28 -07:00
|
|
|
|
|
|
|
Args:
|
|
|
|
origin: server sending the event
|
|
|
|
|
|
|
|
event: event to be persisted
|
|
|
|
|
|
|
|
state: Normally None, but if we are handling a gap in the graph
|
|
|
|
(ie, we are missing one or more prev_events), the resolved state at the
|
|
|
|
event
|
2014-08-26 12:49:42 -06:00
|
|
|
"""
|
2018-09-19 10:28:18 -06:00
|
|
|
room_id = event.room_id
|
|
|
|
event_id = event.event_id
|
2014-08-26 12:49:42 -06:00
|
|
|
|
2019-06-20 03:32:02 -06:00
|
|
|
logger.debug("[%s %s] Processing event: %s", room_id, event_id, event)
|
2014-10-16 09:56:51 -06:00
|
|
|
|
2018-11-09 07:58:09 -07:00
|
|
|
try:
|
2020-09-09 10:22:00 -06:00
|
|
|
await self._handle_new_event(origin, event, state=state)
|
2018-11-09 07:58:09 -07:00
|
|
|
except AuthError as e:
|
2019-06-20 03:32:02 -06:00
|
|
|
raise FederationError("ERROR", e.code, e.msg, affected=event.event_id)
|
2014-12-16 08:24:03 -07:00
|
|
|
|
2020-01-28 07:43:21 -07:00
|
|
|
# For encrypted messages we check that we know about the sending device,
|
|
|
|
# if we don't then we mark the device cache for that user as stale.
|
2020-02-04 10:25:54 -07:00
|
|
|
if event.type == EventTypes.Encrypted:
|
2020-01-28 07:43:21 -07:00
|
|
|
device_id = event.content.get("device_id")
|
2020-02-05 07:02:39 -07:00
|
|
|
sender_key = event.content.get("sender_key")
|
|
|
|
|
|
|
|
cached_devices = await self.store.get_cached_devices_for_user(event.sender)
|
|
|
|
|
|
|
|
resync = False # Whether we should resync device lists.
|
|
|
|
|
|
|
|
device = None
|
2020-01-28 07:43:21 -07:00
|
|
|
if device_id is not None:
|
2020-02-05 07:02:39 -07:00
|
|
|
device = cached_devices.get(device_id)
|
|
|
|
if device is None:
|
2020-01-28 07:43:21 -07:00
|
|
|
logger.info(
|
|
|
|
"Received event from remote device not in our cache: %s %s",
|
|
|
|
event.sender,
|
|
|
|
device_id,
|
|
|
|
)
|
2020-02-05 07:02:39 -07:00
|
|
|
resync = True
|
|
|
|
|
|
|
|
# We also check if the `sender_key` matches what we expect.
|
|
|
|
if sender_key is not None:
|
|
|
|
# Figure out what sender key we're expecting. If we know the
|
|
|
|
# device and recognize the algorithm then we can work out the
|
|
|
|
# exact key to expect. Otherwise check it matches any key we
|
|
|
|
# have for that device.
|
2020-07-01 09:21:02 -06:00
|
|
|
|
|
|
|
current_keys = [] # type: Container[str]
|
|
|
|
|
2020-02-05 07:02:39 -07:00
|
|
|
if device:
|
|
|
|
keys = device.get("keys", {}).get("keys", {})
|
|
|
|
|
2020-06-10 10:44:34 -06:00
|
|
|
if (
|
|
|
|
event.content.get("algorithm")
|
|
|
|
== RoomEncryptionAlgorithms.MEGOLM_V1_AES_SHA2
|
|
|
|
):
|
2020-02-05 07:02:39 -07:00
|
|
|
# For this algorithm we expect a curve25519 key.
|
|
|
|
key_name = "curve25519:%s" % (device_id,)
|
|
|
|
current_keys = [keys.get(key_name)]
|
|
|
|
else:
|
|
|
|
# We don't know understand the algorithm, so we just
|
|
|
|
# check it matches a key for the device.
|
|
|
|
current_keys = keys.values()
|
|
|
|
elif device_id:
|
|
|
|
# We don't have any keys for the device ID.
|
2020-07-01 09:21:02 -06:00
|
|
|
pass
|
2020-02-05 07:02:39 -07:00
|
|
|
else:
|
|
|
|
# The event didn't include a device ID, so we just look for
|
|
|
|
# keys across all devices.
|
2020-07-01 09:21:02 -06:00
|
|
|
current_keys = [
|
2020-02-05 07:02:39 -07:00
|
|
|
key
|
2020-09-03 08:38:32 -06:00
|
|
|
for device in cached_devices.values()
|
2020-02-05 07:02:39 -07:00
|
|
|
for key in device.get("keys", {}).get("keys", {}).values()
|
2020-07-01 09:21:02 -06:00
|
|
|
]
|
2020-01-30 10:06:38 -07:00
|
|
|
|
2020-02-05 07:02:39 -07:00
|
|
|
# We now check that the sender key matches (one of) the expected
|
|
|
|
# keys.
|
|
|
|
if sender_key not in current_keys:
|
|
|
|
logger.info(
|
|
|
|
"Received event from remote device with unexpected sender key: %s %s: %s",
|
|
|
|
event.sender,
|
|
|
|
device_id or "<no device_id>",
|
|
|
|
sender_key,
|
|
|
|
)
|
|
|
|
resync = True
|
|
|
|
|
|
|
|
if resync:
|
2020-07-10 11:23:17 -06:00
|
|
|
run_as_background_process(
|
|
|
|
"resync_device_due_to_pdu", self._resync_device, event.sender
|
|
|
|
)
|
2020-02-05 07:02:39 -07:00
|
|
|
|
2020-07-10 11:23:17 -06:00
|
|
|
async def _resync_device(self, sender: str) -> None:
|
|
|
|
"""We have detected that the device list for the given user may be out
|
|
|
|
of sync, so we try and resync them.
|
|
|
|
"""
|
|
|
|
|
|
|
|
try:
|
|
|
|
await self.store.mark_remote_user_device_cache_as_stale(sender)
|
|
|
|
|
|
|
|
# Immediately attempt a resync in the background
|
|
|
|
if self.config.worker_app:
|
|
|
|
await self._user_device_resync(user_id=sender)
|
|
|
|
else:
|
|
|
|
await self._device_list_updater.user_device_resync(sender)
|
|
|
|
except Exception:
|
|
|
|
logger.exception("Failed to resync device for %s", sender)
|
2020-01-28 07:43:21 -07:00
|
|
|
|
2014-08-19 07:20:03 -06:00
|
|
|
@log_function
|
2019-12-10 09:54:34 -07:00
|
|
|
async def backfill(self, dest, room_id, limit, extremities):
|
2014-11-12 09:20:21 -07:00
|
|
|
""" Trigger a backfill request to `dest` for the given `room_id`
|
2016-04-12 05:04:19 -06:00
|
|
|
|
2018-04-17 16:41:36 -06:00
|
|
|
This will attempt to get more events from the remote. If the other side
|
|
|
|
has no new events to offer, this will return an empty list.
|
|
|
|
|
|
|
|
As the events are received, we check their signatures, and also do some
|
|
|
|
sanity-checking on them. If any of the backfilled events are invalid,
|
|
|
|
this method throws a SynapseError.
|
|
|
|
|
|
|
|
TODO: make this more useful to distinguish failures of the remote
|
|
|
|
server from invalid events (there is probably no point in trying to
|
|
|
|
re-fetch invalid events from every other HS in the room.)
|
2014-11-12 09:20:21 -07:00
|
|
|
"""
|
2016-04-05 05:56:29 -06:00
|
|
|
if dest == self.server_name:
|
|
|
|
raise SynapseError(400, "Can't backfill from self.")
|
|
|
|
|
2019-12-10 09:54:34 -07:00
|
|
|
events = await self.federation_client.backfill(
|
2019-06-20 03:32:02 -06:00
|
|
|
dest, room_id, limit=limit, extremities=extremities
|
2014-10-31 03:59:02 -06:00
|
|
|
)
|
2014-08-26 12:49:42 -06:00
|
|
|
|
2020-09-29 08:57:36 -06:00
|
|
|
if not events:
|
|
|
|
return []
|
|
|
|
|
2018-04-20 04:41:03 -06:00
|
|
|
# ideally we'd sanity check the events here for excess prev_events etc,
|
|
|
|
# but it's hard to reject events at this point without completely
|
|
|
|
# breaking backfill in the same way that it is currently broken by
|
|
|
|
# events whose signature we cannot verify (#3121).
|
|
|
|
#
|
|
|
|
# So for now we accept the events anyway. #3124 tracks this.
|
|
|
|
#
|
|
|
|
# for ev in events:
|
|
|
|
# self._sanity_check_event(ev)
|
2018-04-17 16:41:12 -06:00
|
|
|
|
2016-04-12 04:54:41 -06:00
|
|
|
# Don't bother processing events we already have.
|
2019-12-10 09:54:34 -07:00
|
|
|
seen_events = await self.store.have_events_in_timeline(
|
2020-02-21 05:15:07 -07:00
|
|
|
{e.event_id for e in events}
|
2016-04-12 04:19:10 -06:00
|
|
|
)
|
|
|
|
|
|
|
|
events = [e for e in events if e.event_id not in seen_events]
|
|
|
|
|
|
|
|
if not events:
|
2019-07-23 07:00:55 -06:00
|
|
|
return []
|
2016-04-12 04:19:10 -06:00
|
|
|
|
2015-05-20 04:59:02 -06:00
|
|
|
event_map = {e.event_id: e for e in events}
|
2014-08-26 12:49:42 -06:00
|
|
|
|
2020-02-21 05:15:07 -07:00
|
|
|
event_ids = {e.event_id for e in events}
|
2014-10-15 09:06:59 -06:00
|
|
|
|
2019-12-16 05:26:28 -07:00
|
|
|
# build a list of events whose prev_events weren't in the batch.
|
|
|
|
# (XXX: this will include events whose prev_events we already have; that doesn't
|
|
|
|
# sound right?)
|
2019-06-20 03:32:02 -06:00
|
|
|
edges = [ev.event_id for ev in events if set(ev.prev_event_ids()) - event_ids]
|
2014-10-15 09:06:59 -06:00
|
|
|
|
2019-06-20 03:32:02 -06:00
|
|
|
logger.info("backfill: Got %d events with %d edges", len(events), len(edges))
|
2015-06-02 03:28:14 -06:00
|
|
|
|
2015-05-20 04:59:02 -06:00
|
|
|
# For each edge get the current state.
|
2014-10-15 09:06:59 -06:00
|
|
|
|
2015-05-20 04:59:02 -06:00
|
|
|
auth_events = {}
|
2015-06-02 03:11:32 -06:00
|
|
|
state_events = {}
|
2015-05-20 04:59:02 -06:00
|
|
|
events_to_state = {}
|
|
|
|
for e_id in edges:
|
2019-12-10 09:54:34 -07:00
|
|
|
state, auth = await self._get_state_for_room(
|
2019-12-20 02:55:45 -07:00
|
|
|
destination=dest,
|
|
|
|
room_id=room_id,
|
|
|
|
event_id=e_id,
|
|
|
|
include_event_in_state=False,
|
2015-05-20 04:59:02 -06:00
|
|
|
)
|
|
|
|
auth_events.update({a.event_id: a for a in auth})
|
2015-06-02 03:58:35 -06:00
|
|
|
auth_events.update({s.event_id: s for s in state})
|
2015-06-02 03:11:32 -06:00
|
|
|
state_events.update({s.event_id: s for s in state})
|
2015-05-20 04:59:02 -06:00
|
|
|
events_to_state[e_id] = state
|
|
|
|
|
2020-02-21 05:15:07 -07:00
|
|
|
required_auth = {
|
2016-08-05 05:59:04 -06:00
|
|
|
a_id
|
2019-06-20 03:32:02 -06:00
|
|
|
for event in events
|
|
|
|
+ list(state_events.values())
|
|
|
|
+ list(auth_events.values())
|
2018-11-05 06:35:15 -07:00
|
|
|
for a_id in event.auth_event_ids()
|
2020-02-21 05:15:07 -07:00
|
|
|
}
|
2019-06-20 03:32:02 -06:00
|
|
|
auth_events.update(
|
|
|
|
{e_id: event_map[e_id] for e_id in required_auth if e_id in event_map}
|
|
|
|
)
|
2019-02-25 03:02:12 -07:00
|
|
|
|
2015-06-25 10:18:19 -06:00
|
|
|
ev_infos = []
|
2019-02-22 04:33:51 -07:00
|
|
|
|
2019-12-16 05:26:28 -07:00
|
|
|
# Step 1: persist the events in the chunk we fetched state for (i.e.
|
|
|
|
# the backwards extremities), with custom auth events and state
|
2015-06-25 10:18:19 -06:00
|
|
|
for e_id in events_to_state:
|
2019-02-25 03:02:12 -07:00
|
|
|
# For paranoia we ensure that these events are marked as
|
|
|
|
# non-outliers
|
|
|
|
ev = event_map[e_id]
|
2019-06-20 03:32:02 -06:00
|
|
|
assert not ev.internal_metadata.is_outlier()
|
|
|
|
|
|
|
|
ev_infos.append(
|
2019-12-05 08:02:35 -07:00
|
|
|
_NewEventInfo(
|
|
|
|
event=ev,
|
|
|
|
state=events_to_state[e_id],
|
|
|
|
auth_events={
|
2019-06-20 03:32:02 -06:00
|
|
|
(
|
|
|
|
auth_events[a_id].type,
|
|
|
|
auth_events[a_id].state_key,
|
|
|
|
): auth_events[a_id]
|
|
|
|
for a_id in ev.auth_event_ids()
|
|
|
|
if a_id in auth_events
|
|
|
|
},
|
2019-12-05 08:02:35 -07:00
|
|
|
)
|
2019-06-20 03:32:02 -06:00
|
|
|
)
|
2015-05-20 04:59:02 -06:00
|
|
|
|
2020-09-14 03:16:41 -06:00
|
|
|
if ev_infos:
|
|
|
|
await self._handle_new_events(dest, room_id, ev_infos, backfilled=True)
|
2016-04-12 04:19:10 -06:00
|
|
|
|
2019-02-25 07:45:02 -07:00
|
|
|
# Step 2: Persist the rest of the events in the chunk one by one
|
2015-05-20 04:59:02 -06:00
|
|
|
events.sort(key=lambda e: e.depth)
|
|
|
|
|
|
|
|
for event in events:
|
|
|
|
if event in events_to_state:
|
|
|
|
continue
|
|
|
|
|
2019-02-25 03:02:12 -07:00
|
|
|
# For paranoia we ensure that these events are marked as
|
|
|
|
# non-outliers
|
2019-06-20 03:32:02 -06:00
|
|
|
assert not event.internal_metadata.is_outlier()
|
2019-02-25 03:02:12 -07:00
|
|
|
|
2016-04-12 05:48:30 -06:00
|
|
|
# We store these one at a time since each event depends on the
|
|
|
|
# previous to work out the state.
|
|
|
|
# TODO: We can probably do something more clever here.
|
2019-12-10 09:54:34 -07:00
|
|
|
await self._handle_new_event(dest, event, backfilled=True)
|
2014-08-26 12:49:42 -06:00
|
|
|
|
2019-07-23 07:00:55 -06:00
|
|
|
return events
|
2014-08-26 12:49:42 -06:00
|
|
|
|
2020-09-18 07:25:52 -06:00
|
|
|
async def maybe_backfill(
|
|
|
|
self, room_id: str, current_depth: int, limit: int
|
|
|
|
) -> bool:
|
2015-05-12 03:35:45 -06:00
|
|
|
"""Checks the database to see if we should backfill before paginating,
|
|
|
|
and if so do.
|
2020-09-18 07:25:52 -06:00
|
|
|
|
|
|
|
Args:
|
|
|
|
room_id
|
|
|
|
current_depth: The depth from which we're paginating from. This is
|
|
|
|
used to decide if we should backfill and what extremities to
|
|
|
|
use.
|
|
|
|
limit: The number of events that the pagination request will
|
|
|
|
return. This is used as part of the heuristic to decide if we
|
|
|
|
should back paginate.
|
2015-05-11 11:01:31 -06:00
|
|
|
"""
|
2019-12-10 09:54:34 -07:00
|
|
|
extremities = await self.store.get_oldest_events_with_depth_in_room(room_id)
|
2015-05-11 11:01:31 -06:00
|
|
|
|
|
|
|
if not extremities:
|
2015-05-12 03:35:45 -06:00
|
|
|
logger.debug("Not backfilling as no extremeties found.")
|
2020-09-18 07:25:52 -06:00
|
|
|
return False
|
2015-05-11 11:01:31 -06:00
|
|
|
|
2019-02-20 09:54:35 -07:00
|
|
|
# We only want to paginate if we can actually see the events we'll get,
|
|
|
|
# as otherwise we'll just spend a lot of resources to get redacted
|
|
|
|
# events.
|
|
|
|
#
|
2019-03-05 02:16:35 -07:00
|
|
|
# We do this by filtering all the backwards extremities and seeing if
|
|
|
|
# any remain. Given we don't have the extremity events themselves, we
|
|
|
|
# need to actually check the events that reference them.
|
2019-02-27 06:06:10 -07:00
|
|
|
#
|
|
|
|
# *Note*: the spec wants us to keep backfilling until we reach the start
|
|
|
|
# of the room in case we are allowed to see some of the history. However
|
|
|
|
# in practice that causes more issues than its worth, as a) its
|
|
|
|
# relatively rare for there to be any visible history and b) even when
|
|
|
|
# there is its often sufficiently long ago that clients would stop
|
|
|
|
# attempting to paginate before backfill reached the visible history.
|
|
|
|
#
|
2019-03-05 02:16:35 -07:00
|
|
|
# TODO: If we do do a backfill then we should filter the backwards
|
|
|
|
# extremities to only include those that point to visible portions of
|
|
|
|
# history.
|
2019-02-20 09:54:35 -07:00
|
|
|
#
|
|
|
|
# TODO: Correctly handle the case where we are allowed to see the
|
2019-03-05 02:16:35 -07:00
|
|
|
# forward event but not the backward extremity, e.g. in the case of
|
|
|
|
# initial join of the server where we are allowed to see the join
|
|
|
|
# event but not anything before it. This would require looking at the
|
|
|
|
# state *before* the event, ignoring the special casing certain event
|
|
|
|
# types have.
|
2019-02-20 09:54:35 -07:00
|
|
|
|
2019-12-10 09:54:34 -07:00
|
|
|
forward_events = await self.store.get_successor_events(list(extremities))
|
2019-02-20 09:54:35 -07:00
|
|
|
|
2019-12-10 09:54:34 -07:00
|
|
|
extremities_events = await self.store.get_events(
|
2019-12-11 06:39:47 -07:00
|
|
|
forward_events,
|
|
|
|
redact_behaviour=EventRedactBehaviour.AS_IS,
|
|
|
|
get_prev_content=False,
|
2019-02-20 09:54:35 -07:00
|
|
|
)
|
|
|
|
|
2019-03-04 07:34:34 -07:00
|
|
|
# We set `check_history_visibility_only` as we might otherwise get false
|
|
|
|
# positives from users having been erased.
|
2019-12-10 09:54:34 -07:00
|
|
|
filtered_extremities = await filter_events_for_server(
|
2019-10-23 10:25:54 -06:00
|
|
|
self.storage,
|
2019-06-20 03:32:02 -06:00
|
|
|
self.server_name,
|
|
|
|
list(extremities_events.values()),
|
|
|
|
redact=False,
|
|
|
|
check_history_visibility_only=True,
|
2019-02-20 09:54:35 -07:00
|
|
|
)
|
|
|
|
|
|
|
|
if not filtered_extremities:
|
2019-07-23 07:00:55 -06:00
|
|
|
return False
|
2019-02-20 09:54:35 -07:00
|
|
|
|
2015-05-11 11:01:31 -06:00
|
|
|
# Check if we reached a point where we should start backfilling.
|
2019-06-20 03:32:02 -06:00
|
|
|
sorted_extremeties_tuple = sorted(extremities.items(), key=lambda e: -int(e[1]))
|
2015-05-11 11:01:31 -06:00
|
|
|
max_depth = sorted_extremeties_tuple[0][1]
|
|
|
|
|
2020-09-18 07:25:52 -06:00
|
|
|
# If we're approaching an extremity we trigger a backfill, otherwise we
|
|
|
|
# no-op.
|
|
|
|
#
|
|
|
|
# We chose twice the limit here as then clients paginating backwards
|
|
|
|
# will send pagination requests that trigger backfill at least twice
|
|
|
|
# using the most recent extremity before it gets removed (see below). We
|
|
|
|
# chose more than one times the limit in case of failure, but choosing a
|
|
|
|
# much larger factor will result in triggering a backfill request much
|
|
|
|
# earlier than necessary.
|
|
|
|
if current_depth - 2 * limit > max_depth:
|
|
|
|
logger.debug(
|
|
|
|
"Not backfilling as we don't need to. %d < %d - 2 * %d",
|
|
|
|
max_depth,
|
|
|
|
current_depth,
|
|
|
|
limit,
|
|
|
|
)
|
|
|
|
return False
|
|
|
|
|
|
|
|
logger.debug(
|
|
|
|
"room_id: %s, backfill: current_depth: %s, max_depth: %s, extrems: %s",
|
|
|
|
room_id,
|
|
|
|
current_depth,
|
|
|
|
max_depth,
|
|
|
|
sorted_extremeties_tuple,
|
|
|
|
)
|
|
|
|
|
|
|
|
# We ignore extremities that have a greater depth than our current depth
|
|
|
|
# as:
|
|
|
|
# 1. we don't really care about getting events that have happened
|
|
|
|
# before our current position; and
|
|
|
|
# 2. we have likely previously tried and failed to backfill from that
|
|
|
|
# extremity, so to avoid getting "stuck" requesting the same
|
|
|
|
# backfill repeatedly we drop those extremities.
|
|
|
|
filtered_sorted_extremeties_tuple = [
|
|
|
|
t for t in sorted_extremeties_tuple if int(t[1]) <= current_depth
|
|
|
|
]
|
|
|
|
|
|
|
|
# However, we need to check that the filtered extremities are non-empty.
|
|
|
|
# If they are empty then either we can a) bail or b) still attempt to
|
|
|
|
# backill. We opt to try backfilling anyway just in case we do get
|
|
|
|
# relevant events.
|
|
|
|
if filtered_sorted_extremeties_tuple:
|
|
|
|
sorted_extremeties_tuple = filtered_sorted_extremeties_tuple
|
|
|
|
|
2016-08-16 04:34:36 -06:00
|
|
|
# We don't want to specify too many extremities as it causes the backfill
|
|
|
|
# request URI to be too long.
|
|
|
|
extremities = dict(sorted_extremeties_tuple[:5])
|
|
|
|
|
2015-05-11 11:01:31 -06:00
|
|
|
# Now we need to decide which hosts to hit first.
|
|
|
|
|
2015-05-12 03:35:45 -06:00
|
|
|
# First we try hosts that are already in the room
|
|
|
|
# TODO: HEURISTIC ALERT.
|
2015-05-11 11:01:31 -06:00
|
|
|
|
2019-12-10 09:54:34 -07:00
|
|
|
curr_state = await self.state_handler.get_current_state(room_id)
|
2015-05-11 11:01:31 -06:00
|
|
|
|
|
|
|
def get_domains_from_state(state):
|
2018-05-22 09:23:39 -06:00
|
|
|
"""Get joined domains from state
|
|
|
|
|
|
|
|
Args:
|
|
|
|
state (dict[tuple, FrozenEvent]): State map from type/state
|
|
|
|
key to event.
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
list[tuple[str, int]]: Returns a list of servers with the
|
|
|
|
lowest depth of their joins. Sorted by lowest depth first.
|
|
|
|
"""
|
2015-05-11 11:01:31 -06:00
|
|
|
joined_users = [
|
|
|
|
(state_key, int(event.depth))
|
2020-06-15 05:03:36 -06:00
|
|
|
for (e_type, state_key), event in state.items()
|
2019-06-20 03:32:02 -06:00
|
|
|
if e_type == EventTypes.Member and event.membership == Membership.JOIN
|
2015-05-11 11:01:31 -06:00
|
|
|
]
|
|
|
|
|
2020-07-01 09:21:02 -06:00
|
|
|
joined_domains = {} # type: Dict[str, int]
|
2015-05-11 11:01:31 -06:00
|
|
|
for u, d in joined_users:
|
|
|
|
try:
|
2016-05-16 12:17:03 -06:00
|
|
|
dom = get_domain_from_id(u)
|
2015-05-11 11:01:31 -06:00
|
|
|
old_d = joined_domains.get(dom)
|
|
|
|
if old_d:
|
|
|
|
joined_domains[dom] = min(d, old_d)
|
|
|
|
else:
|
|
|
|
joined_domains[dom] = d
|
2017-10-23 08:52:32 -06:00
|
|
|
except Exception:
|
2015-05-11 11:01:31 -06:00
|
|
|
pass
|
|
|
|
|
2018-07-20 23:47:18 -06:00
|
|
|
return sorted(joined_domains.items(), key=lambda d: d[1])
|
2015-05-11 11:01:31 -06:00
|
|
|
|
|
|
|
curr_domains = get_domains_from_state(curr_state)
|
|
|
|
|
|
|
|
likely_domains = [
|
2019-06-20 03:32:02 -06:00
|
|
|
domain for domain, depth in curr_domains if domain != self.server_name
|
2015-05-11 11:01:31 -06:00
|
|
|
]
|
|
|
|
|
2019-12-10 09:54:34 -07:00
|
|
|
async def try_backfill(domains):
|
2015-05-11 11:01:31 -06:00
|
|
|
# TODO: Should we try multiple of these at a time?
|
|
|
|
for dom in domains:
|
2015-05-12 03:35:45 -06:00
|
|
|
try:
|
2019-12-10 09:54:34 -07:00
|
|
|
await self.backfill(
|
2019-06-20 03:32:02 -06:00
|
|
|
dom, room_id, limit=100, extremities=extremities
|
2015-05-12 03:35:45 -06:00
|
|
|
)
|
2016-04-12 05:04:19 -06:00
|
|
|
# If this succeeded then we probably already have the
|
|
|
|
# appropriate stuff.
|
2016-04-12 05:48:30 -06:00
|
|
|
# TODO: We can probably do something more intelligent here.
|
2019-07-23 07:00:55 -06:00
|
|
|
return True
|
2016-03-07 13:13:10 -07:00
|
|
|
except SynapseError as e:
|
2020-05-22 04:39:20 -06:00
|
|
|
logger.info("Failed to backfill from %s because %s", dom, e)
|
|
|
|
continue
|
|
|
|
except HttpResponseException as e:
|
|
|
|
if 400 <= e.code < 500:
|
|
|
|
raise e.to_synapse_error()
|
|
|
|
|
2019-06-20 03:32:02 -06:00
|
|
|
logger.info("Failed to backfill from %s because %s", dom, e)
|
2015-05-12 03:35:45 -06:00
|
|
|
continue
|
|
|
|
except CodeMessageException as e:
|
|
|
|
if 400 <= e.code < 500:
|
|
|
|
raise
|
|
|
|
|
2019-06-20 03:32:02 -06:00
|
|
|
logger.info("Failed to backfill from %s because %s", dom, e)
|
2015-05-12 03:35:45 -06:00
|
|
|
continue
|
|
|
|
except NotRetryingDestination as e:
|
2018-09-06 08:22:23 -06:00
|
|
|
logger.info(str(e))
|
2015-05-12 03:35:45 -06:00
|
|
|
continue
|
2019-07-30 06:19:22 -06:00
|
|
|
except RequestSendFailed as e:
|
2020-10-23 10:38:40 -06:00
|
|
|
logger.info("Failed to get backfill from %s because %s", dom, e)
|
2019-07-30 06:19:22 -06:00
|
|
|
continue
|
2018-01-22 11:11:18 -07:00
|
|
|
except FederationDeniedError as e:
|
|
|
|
logger.info(e)
|
|
|
|
continue
|
2015-05-12 03:35:45 -06:00
|
|
|
except Exception as e:
|
2019-06-20 03:32:02 -06:00
|
|
|
logger.exception("Failed to backfill from %s because %s", dom, e)
|
2015-05-12 03:35:45 -06:00
|
|
|
continue
|
2015-05-11 11:01:31 -06:00
|
|
|
|
2019-07-23 07:00:55 -06:00
|
|
|
return False
|
2015-05-11 11:01:31 -06:00
|
|
|
|
2019-12-10 09:54:34 -07:00
|
|
|
success = await try_backfill(likely_domains)
|
2015-05-11 11:01:31 -06:00
|
|
|
if success:
|
2019-07-23 07:00:55 -06:00
|
|
|
return True
|
2015-05-11 11:01:31 -06:00
|
|
|
|
|
|
|
# Huh, well *those* domains didn't work out. Lets try some domains
|
|
|
|
# from the time.
|
|
|
|
|
|
|
|
tried_domains = set(likely_domains)
|
2015-05-12 09:19:42 -06:00
|
|
|
tried_domains.add(self.server_name)
|
2015-05-11 11:01:31 -06:00
|
|
|
|
2018-07-20 23:47:18 -06:00
|
|
|
event_ids = list(extremities.keys())
|
2015-05-12 06:58:14 -06:00
|
|
|
|
2017-01-17 10:07:15 -07:00
|
|
|
logger.debug("calling resolve_state_groups in _maybe_backfill")
|
2019-07-03 08:07:04 -06:00
|
|
|
resolve = preserve_fn(self.state_handler.resolve_state_groups_for_events)
|
2019-12-10 09:54:34 -07:00
|
|
|
states = await make_deferred_yieldable(
|
2019-06-20 03:32:02 -06:00
|
|
|
defer.gatherResults(
|
|
|
|
[resolve(room_id, [e]) for e in event_ids], consumeErrors=True
|
|
|
|
)
|
|
|
|
)
|
2018-05-22 12:00:48 -06:00
|
|
|
|
|
|
|
# dict[str, dict[tuple, str]], a map from event_id to state map of
|
|
|
|
# event_ids.
|
2017-01-13 06:21:04 -07:00
|
|
|
states = dict(zip(event_ids, [s.state for s in states]))
|
2015-05-11 11:01:31 -06:00
|
|
|
|
2019-12-10 09:54:34 -07:00
|
|
|
state_map = await self.store.get_events(
|
2020-06-15 05:03:36 -06:00
|
|
|
[e_id for ids in states.values() for e_id in ids.values()],
|
2019-06-20 03:32:02 -06:00
|
|
|
get_prev_content=False,
|
2016-08-25 06:28:31 -06:00
|
|
|
)
|
|
|
|
states = {
|
|
|
|
key: {
|
|
|
|
k: state_map[e_id]
|
2020-06-15 05:03:36 -06:00
|
|
|
for k, e_id in state_dict.items()
|
2016-08-25 06:28:31 -06:00
|
|
|
if e_id in state_map
|
2019-06-20 03:32:02 -06:00
|
|
|
}
|
2020-06-15 05:03:36 -06:00
|
|
|
for key, state_dict in states.items()
|
2016-08-25 06:28:31 -06:00
|
|
|
}
|
|
|
|
|
2015-05-11 11:01:31 -06:00
|
|
|
for e_id, _ in sorted_extremeties_tuple:
|
2015-05-12 07:02:01 -06:00
|
|
|
likely_domains = get_domains_from_state(states[e_id])
|
2015-05-11 11:01:31 -06:00
|
|
|
|
2019-12-10 09:54:34 -07:00
|
|
|
success = await try_backfill(
|
2019-06-20 03:32:02 -06:00
|
|
|
[dom for dom, _ in likely_domains if dom not in tried_domains]
|
|
|
|
)
|
2015-05-11 11:01:31 -06:00
|
|
|
if success:
|
2019-07-23 07:00:55 -06:00
|
|
|
return True
|
2015-05-11 11:01:31 -06:00
|
|
|
|
2018-05-22 09:23:39 -06:00
|
|
|
tried_domains.update(dom for dom, _ in likely_domains)
|
2015-05-11 11:01:31 -06:00
|
|
|
|
2019-07-23 07:00:55 -06:00
|
|
|
return False
|
2015-05-11 11:01:31 -06:00
|
|
|
|
2019-12-16 05:26:28 -07:00
|
|
|
async def _get_events_and_persist(
|
|
|
|
self, destination: str, room_id: str, events: Iterable[str]
|
|
|
|
):
|
|
|
|
"""Fetch the given events from a server, and persist them as outliers.
|
|
|
|
|
2020-07-10 11:15:35 -06:00
|
|
|
This function *does not* recursively get missing auth events of the
|
|
|
|
newly fetched events. Callers must include in the `events` argument
|
|
|
|
any missing events from the auth chain.
|
|
|
|
|
2019-12-16 05:26:28 -07:00
|
|
|
Logs a warning if we can't find the given event.
|
|
|
|
"""
|
|
|
|
|
2020-01-31 07:07:31 -07:00
|
|
|
room_version = await self.store.get_room_version(room_id)
|
2019-12-16 05:26:28 -07:00
|
|
|
|
2020-07-10 11:15:35 -06:00
|
|
|
event_map = {} # type: Dict[str, EventBase]
|
2019-12-16 05:26:28 -07:00
|
|
|
|
|
|
|
async def get_event(event_id: str):
|
|
|
|
with nested_logging_context(event_id):
|
|
|
|
try:
|
|
|
|
event = await self.federation_client.get_pdu(
|
|
|
|
[destination], event_id, room_version, outlier=True,
|
|
|
|
)
|
|
|
|
if event is None:
|
|
|
|
logger.warning(
|
|
|
|
"Server %s didn't return event %s", destination, event_id,
|
|
|
|
)
|
|
|
|
return
|
|
|
|
|
2020-07-10 11:15:35 -06:00
|
|
|
event_map[event.event_id] = event
|
2019-12-16 05:26:28 -07:00
|
|
|
|
|
|
|
except Exception as e:
|
|
|
|
logger.warning(
|
|
|
|
"Error fetching missing state/auth event %s: %s %s",
|
|
|
|
event_id,
|
|
|
|
type(e),
|
|
|
|
e,
|
|
|
|
)
|
|
|
|
|
|
|
|
await concurrently_execute(get_event, events, 5)
|
|
|
|
|
2020-07-10 11:15:35 -06:00
|
|
|
# Make a map of auth events for each event. We do this after fetching
|
|
|
|
# all the events as some of the events' auth events will be in the list
|
|
|
|
# of requested events.
|
|
|
|
|
|
|
|
auth_events = [
|
|
|
|
aid
|
|
|
|
for event in event_map.values()
|
|
|
|
for aid in event.auth_event_ids()
|
|
|
|
if aid not in event_map
|
|
|
|
]
|
|
|
|
persisted_events = await self.store.get_events(
|
|
|
|
auth_events, allow_rejected=True,
|
|
|
|
)
|
|
|
|
|
|
|
|
event_infos = []
|
|
|
|
for event in event_map.values():
|
|
|
|
auth = {}
|
|
|
|
for auth_event_id in event.auth_event_ids():
|
|
|
|
ae = persisted_events.get(auth_event_id) or event_map.get(auth_event_id)
|
|
|
|
if ae:
|
|
|
|
auth[(ae.type, ae.state_key)] = ae
|
|
|
|
else:
|
|
|
|
logger.info("Missing auth event %s", auth_event_id)
|
|
|
|
|
|
|
|
event_infos.append(_NewEventInfo(event, None, auth))
|
|
|
|
|
2019-12-16 05:26:28 -07:00
|
|
|
await self._handle_new_events(
|
2020-09-14 03:16:41 -06:00
|
|
|
destination, room_id, event_infos,
|
2019-12-16 05:26:28 -07:00
|
|
|
)
|
|
|
|
|
2018-04-17 16:41:12 -06:00
|
|
|
def _sanity_check_event(self, ev):
|
|
|
|
"""
|
|
|
|
Do some early sanity checks of a received event
|
|
|
|
|
|
|
|
In particular, checks it doesn't have an excessive number of
|
|
|
|
prev_events or auth_events, which could cause a huge state resolution
|
|
|
|
or cascade of event fetches.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
ev (synapse.events.EventBase): event to be checked
|
|
|
|
|
|
|
|
Returns: None
|
|
|
|
|
|
|
|
Raises:
|
|
|
|
SynapseError if the event does not pass muster
|
|
|
|
"""
|
2018-11-05 06:35:15 -07:00
|
|
|
if len(ev.prev_event_ids()) > 20:
|
2019-10-31 04:23:24 -06:00
|
|
|
logger.warning(
|
2019-06-20 03:32:02 -06:00
|
|
|
"Rejecting event %s which has %i prev_events",
|
|
|
|
ev.event_id,
|
|
|
|
len(ev.prev_event_ids()),
|
2018-04-17 16:41:12 -06:00
|
|
|
)
|
2020-06-16 06:51:47 -06:00
|
|
|
raise SynapseError(HTTPStatus.BAD_REQUEST, "Too many prev_events")
|
2018-04-17 16:41:12 -06:00
|
|
|
|
2018-11-05 06:35:15 -07:00
|
|
|
if len(ev.auth_event_ids()) > 10:
|
2019-10-31 04:23:24 -06:00
|
|
|
logger.warning(
|
2019-06-20 03:32:02 -06:00
|
|
|
"Rejecting event %s which has %i auth_events",
|
|
|
|
ev.event_id,
|
|
|
|
len(ev.auth_event_ids()),
|
2018-04-17 16:41:12 -06:00
|
|
|
)
|
2020-06-16 06:51:47 -06:00
|
|
|
raise SynapseError(HTTPStatus.BAD_REQUEST, "Too many auth_events")
|
2018-04-17 16:41:12 -06:00
|
|
|
|
2020-02-03 15:28:11 -07:00
|
|
|
async def send_invite(self, target_host, event):
|
2014-11-12 09:20:21 -07:00
|
|
|
""" Sends the invite to the remote server for signing.
|
|
|
|
|
|
|
|
Invites must be signed by the invitee's server before distribution.
|
|
|
|
"""
|
2020-02-03 15:28:11 -07:00
|
|
|
pdu = await self.federation_client.send_invite(
|
2014-11-07 06:41:00 -07:00
|
|
|
destination=target_host,
|
2015-01-16 11:59:04 -07:00
|
|
|
room_id=event.room_id,
|
2014-11-07 06:41:00 -07:00
|
|
|
event_id=event.event_id,
|
2019-06-20 03:32:02 -06:00
|
|
|
pdu=event,
|
2014-11-07 06:41:00 -07:00
|
|
|
)
|
|
|
|
|
2019-07-23 07:00:55 -06:00
|
|
|
return pdu
|
2014-11-07 06:41:00 -07:00
|
|
|
|
2020-02-03 09:06:46 -07:00
|
|
|
async def on_event_auth(self, event_id: str) -> List[EventBase]:
|
|
|
|
event = await self.store.get_event(event_id)
|
|
|
|
auth = await self.store.get_auth_chain(
|
2020-02-21 05:15:07 -07:00
|
|
|
list(event.auth_event_ids()), include_given=True
|
2017-05-24 07:22:41 -06:00
|
|
|
)
|
2020-02-03 09:06:46 -07:00
|
|
|
return list(auth)
|
2014-11-07 08:35:53 -07:00
|
|
|
|
2020-02-03 09:13:13 -07:00
|
|
|
async def do_invite_join(
|
|
|
|
self, target_hosts: Iterable[str], room_id: str, joinee: str, content: JsonDict
|
2020-05-22 07:21:54 -06:00
|
|
|
) -> Tuple[str, int]:
|
2014-11-12 09:20:21 -07:00
|
|
|
""" Attempts to join the `joinee` to the room `room_id` via the
|
2019-11-01 04:28:09 -06:00
|
|
|
servers contained in `target_hosts`.
|
2014-11-12 09:20:21 -07:00
|
|
|
|
|
|
|
This first triggers a /make_join/ request that returns a partial
|
|
|
|
event that we can fill out and sign. This is then sent to the
|
|
|
|
remote server via /send_join/ which responds with the state at that
|
|
|
|
event and the auth_chains.
|
|
|
|
|
|
|
|
We suspend processing of any received events from this room until we
|
|
|
|
have finished processing the join.
|
2019-11-01 04:28:09 -06:00
|
|
|
|
|
|
|
Args:
|
2020-02-03 09:13:13 -07:00
|
|
|
target_hosts: List of servers to attempt to join the room with.
|
2019-11-01 04:28:09 -06:00
|
|
|
|
2020-02-03 09:13:13 -07:00
|
|
|
room_id: The ID of the room to join.
|
2019-11-01 04:28:09 -06:00
|
|
|
|
2020-02-03 09:13:13 -07:00
|
|
|
joinee: The User ID of the joining user.
|
2019-11-01 04:28:09 -06:00
|
|
|
|
2020-02-03 09:13:13 -07:00
|
|
|
content: The event content to use for the join event.
|
2014-11-12 09:20:21 -07:00
|
|
|
"""
|
2020-05-22 09:11:35 -06:00
|
|
|
# TODO: We should be able to call this on workers, but the upgrading of
|
|
|
|
# room stuff after join currently doesn't work on workers.
|
|
|
|
assert self.config.worker.worker_app is None
|
|
|
|
|
2014-11-25 04:31:18 -07:00
|
|
|
logger.debug("Joining %s to %s", joinee, room_id)
|
|
|
|
|
2020-02-03 09:13:13 -07:00
|
|
|
origin, event, room_version_obj = await self._make_and_verify_event(
|
2015-02-05 06:43:28 -07:00
|
|
|
target_hosts,
|
2014-10-17 08:04:17 -06:00
|
|
|
room_id,
|
2015-10-01 10:49:52 -06:00
|
|
|
joinee,
|
2015-11-12 09:19:55 -07:00
|
|
|
"join",
|
|
|
|
content,
|
2019-06-20 03:32:02 -06:00
|
|
|
params={"ver": KNOWN_ROOM_VERSIONS},
|
2014-08-20 07:42:36 -06:00
|
|
|
)
|
|
|
|
|
2017-03-14 05:26:57 -06:00
|
|
|
# This shouldn't happen, because the RoomMemberHandler has a
|
|
|
|
# linearizer lock which only allows one operation per user per room
|
|
|
|
# at a time - so this is just paranoia.
|
2019-06-20 03:32:02 -06:00
|
|
|
assert room_id not in self.room_queues
|
2017-03-14 05:26:57 -06:00
|
|
|
|
2014-10-29 10:59:24 -06:00
|
|
|
self.room_queues[room_id] = []
|
2017-03-14 05:26:57 -06:00
|
|
|
|
2020-02-03 09:13:13 -07:00
|
|
|
await self._clean_room_for_join(room_id)
|
2017-03-14 05:26:57 -06:00
|
|
|
|
2014-12-10 08:55:03 -07:00
|
|
|
handled_events = set()
|
|
|
|
|
2014-10-29 10:59:24 -06:00
|
|
|
try:
|
2015-02-05 06:43:28 -07:00
|
|
|
# Try the host we successfully got a response to /make_join/
|
|
|
|
# request first.
|
2020-07-01 09:21:02 -06:00
|
|
|
host_list = list(target_hosts)
|
2015-02-06 03:53:18 -07:00
|
|
|
try:
|
2020-07-01 09:21:02 -06:00
|
|
|
host_list.remove(origin)
|
|
|
|
host_list.insert(0, origin)
|
2015-02-06 03:53:18 -07:00
|
|
|
except ValueError:
|
|
|
|
pass
|
2020-01-27 07:30:57 -07:00
|
|
|
|
2020-02-03 09:13:13 -07:00
|
|
|
ret = await self.federation_client.send_join(
|
2020-07-01 09:21:02 -06:00
|
|
|
host_list, event, room_version_obj
|
2019-01-23 13:21:33 -07:00
|
|
|
)
|
2014-10-17 11:56:42 -06:00
|
|
|
|
2015-02-05 06:43:28 -07:00
|
|
|
origin = ret["origin"]
|
2014-11-25 04:31:18 -07:00
|
|
|
state = ret["state"]
|
|
|
|
auth_chain = ret["auth_chain"]
|
2014-11-27 09:02:26 -07:00
|
|
|
auth_chain.sort(key=lambda e: e.depth)
|
2014-08-20 07:42:36 -06:00
|
|
|
|
2014-12-10 08:55:03 -07:00
|
|
|
handled_events.update([s.event_id for s in state])
|
|
|
|
handled_events.update([a.event_id for a in auth_chain])
|
2015-12-10 10:08:21 -07:00
|
|
|
handled_events.add(event.event_id)
|
2014-12-10 08:55:03 -07:00
|
|
|
|
2014-11-25 04:31:18 -07:00
|
|
|
logger.debug("do_invite_join auth_chain: %s", auth_chain)
|
|
|
|
logger.debug("do_invite_join state: %s", state)
|
2014-10-17 11:56:42 -06:00
|
|
|
|
2015-12-10 10:08:21 -07:00
|
|
|
logger.debug("do_invite_join event: %s", event)
|
2014-10-29 10:59:24 -06:00
|
|
|
|
2020-02-24 08:46:41 -07:00
|
|
|
# if this is the first time we've joined this room, it's time to add
|
|
|
|
# a row to `rooms` with the correct room version. If there's already a
|
|
|
|
# row there, we should override it, since it may have been populated
|
|
|
|
# based on an invite request which lied about the room version.
|
|
|
|
#
|
|
|
|
# federation_client.send_join has already checked that the room
|
|
|
|
# version in the received create event is the same as room_version_obj,
|
|
|
|
# so we can rely on it now.
|
|
|
|
#
|
|
|
|
await self.store.upsert_room_on_join(
|
|
|
|
room_id=room_id, room_version=room_version_obj,
|
|
|
|
)
|
2014-10-29 10:59:24 -06:00
|
|
|
|
2020-05-22 07:21:54 -06:00
|
|
|
max_stream_id = await self._persist_auth_tree(
|
2020-09-14 03:16:41 -06:00
|
|
|
origin, room_id, auth_chain, state, event, room_version_obj
|
2020-01-27 07:30:57 -07:00
|
|
|
)
|
2014-10-29 10:59:24 -06:00
|
|
|
|
2020-05-22 07:21:54 -06:00
|
|
|
# We wait here until this instance has seen the events come down
|
|
|
|
# replication (if we're using replication) as the below uses caches.
|
|
|
|
await self._replication.wait_for_stream_position(
|
2020-09-14 03:16:41 -06:00
|
|
|
self.config.worker.events_shard_config.get_instance(room_id),
|
|
|
|
"events",
|
|
|
|
max_stream_id,
|
2020-05-22 07:21:54 -06:00
|
|
|
)
|
|
|
|
|
2019-11-01 04:28:09 -06:00
|
|
|
# Check whether this room is the result of an upgrade of a room we already know
|
|
|
|
# about. If so, migrate over user information
|
2020-02-03 09:13:13 -07:00
|
|
|
predecessor = await self.store.get_room_predecessor(room_id)
|
2019-12-11 06:07:25 -07:00
|
|
|
if not predecessor or not isinstance(predecessor.get("room_id"), str):
|
2020-05-22 07:21:54 -06:00
|
|
|
return event.event_id, max_stream_id
|
2019-11-01 04:28:09 -06:00
|
|
|
old_room_id = predecessor["room_id"]
|
|
|
|
logger.debug(
|
|
|
|
"Found predecessor for %s during remote join: %s", room_id, old_room_id
|
|
|
|
)
|
|
|
|
|
|
|
|
# We retrieve the room member handler here as to not cause a cyclic dependency
|
|
|
|
member_handler = self.hs.get_room_member_handler()
|
2020-02-03 09:13:13 -07:00
|
|
|
await member_handler.transfer_room_state_on_room_upgrade(
|
2019-11-01 04:28:09 -06:00
|
|
|
old_room_id, room_id
|
|
|
|
)
|
|
|
|
|
2014-11-25 04:31:18 -07:00
|
|
|
logger.debug("Finished joining %s to %s", joinee, room_id)
|
2020-05-22 07:21:54 -06:00
|
|
|
return event.event_id, max_stream_id
|
2014-10-29 10:59:24 -06:00
|
|
|
finally:
|
|
|
|
room_queue = self.room_queues[room_id]
|
|
|
|
del self.room_queues[room_id]
|
2014-10-17 11:56:42 -06:00
|
|
|
|
2017-03-14 05:26:57 -06:00
|
|
|
# we don't need to wait for the queued events to be processed -
|
|
|
|
# it's just a best-effort thing at this point. We do want to do
|
|
|
|
# them roughly in order, though, otherwise we'll end up making
|
|
|
|
# lots of requests for missing prev_events which we do actually
|
2020-07-24 08:53:25 -06:00
|
|
|
# have. Hence we fire off the background task, but don't wait for it.
|
2014-12-10 08:55:03 -07:00
|
|
|
|
2019-07-03 08:07:04 -06:00
|
|
|
run_in_background(self._handle_queued_pdus, room_queue)
|
2014-10-17 11:56:42 -06:00
|
|
|
|
2019-12-10 10:01:37 -07:00
|
|
|
async def _handle_queued_pdus(self, room_queue):
|
2017-03-14 05:26:57 -06:00
|
|
|
"""Process PDUs which got queued up while we were busy send_joining.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
room_queue (list[FrozenEvent, str]): list of PDUs to be processed
|
|
|
|
and the servers that sent them
|
|
|
|
"""
|
|
|
|
for p, origin in room_queue:
|
|
|
|
try:
|
2019-06-20 03:32:02 -06:00
|
|
|
logger.info(
|
|
|
|
"Processing queued PDU %s which was received "
|
|
|
|
"while we were joining %s",
|
|
|
|
p.event_id,
|
|
|
|
p.room_id,
|
|
|
|
)
|
2019-07-03 08:07:04 -06:00
|
|
|
with nested_logging_context(p.event_id):
|
2019-12-10 10:01:37 -07:00
|
|
|
await self.on_receive_pdu(origin, p, sent_to_us_directly=True)
|
2017-03-14 05:26:57 -06:00
|
|
|
except Exception as e:
|
2019-10-31 04:23:24 -06:00
|
|
|
logger.warning(
|
2019-06-20 03:32:02 -06:00
|
|
|
"Error handling queued PDU %s from %s: %s", p.event_id, origin, e
|
|
|
|
)
|
2017-03-14 05:26:57 -06:00
|
|
|
|
2020-02-03 08:35:30 -07:00
|
|
|
async def on_make_join_request(
|
|
|
|
self, origin: str, room_id: str, user_id: str
|
|
|
|
) -> EventBase:
|
2014-11-12 09:20:21 -07:00
|
|
|
""" We've received a /make_join/ request, so we create a partial
|
2015-10-20 04:58:58 -06:00
|
|
|
join event for the room and return that. We do *not* persist or
|
2014-11-12 09:20:21 -07:00
|
|
|
process it until the other server has signed it and sent it back.
|
2019-07-26 03:08:22 -06:00
|
|
|
|
|
|
|
Args:
|
2020-02-03 08:35:30 -07:00
|
|
|
origin: The (verified) server name of the requesting server.
|
|
|
|
room_id: Room to create join event in
|
|
|
|
user_id: The user to create the join for
|
2014-11-12 09:20:21 -07:00
|
|
|
"""
|
2019-07-26 03:08:22 -06:00
|
|
|
if get_domain_from_id(user_id) != origin:
|
|
|
|
logger.info(
|
|
|
|
"Got /make_join request for user %r from different origin %s, ignoring",
|
|
|
|
user_id,
|
|
|
|
origin,
|
|
|
|
)
|
|
|
|
raise SynapseError(403, "User not from origin", Codes.FORBIDDEN)
|
|
|
|
|
2020-07-16 08:17:31 -06:00
|
|
|
# checking the room version will check that we've actually heard of the room
|
|
|
|
# (and return a 404 otherwise)
|
2020-02-03 08:35:30 -07:00
|
|
|
room_version = await self.store.get_room_version_id(room_id)
|
2019-01-23 13:21:33 -07:00
|
|
|
|
2020-07-16 08:17:31 -06:00
|
|
|
# now check that we are *still* in the room
|
|
|
|
is_in_room = await self.auth.check_host_in_room(room_id, self.server_name)
|
|
|
|
if not is_in_room:
|
|
|
|
logger.info(
|
|
|
|
"Got /make_join request for room %s we are no longer in", room_id,
|
|
|
|
)
|
|
|
|
raise NotFoundError("Not an active room on this server")
|
|
|
|
|
|
|
|
event_content = {"membership": Membership.JOIN}
|
|
|
|
|
2019-01-23 13:21:33 -07:00
|
|
|
builder = self.event_builder_factory.new(
|
|
|
|
room_version,
|
|
|
|
{
|
|
|
|
"type": EventTypes.Member,
|
|
|
|
"content": event_content,
|
|
|
|
"room_id": room_id,
|
|
|
|
"sender": user_id,
|
|
|
|
"state_key": user_id,
|
2019-06-20 03:32:02 -06:00
|
|
|
},
|
2019-01-23 13:21:33 -07:00
|
|
|
)
|
2014-12-04 08:50:01 -07:00
|
|
|
|
2016-04-13 04:11:46 -06:00
|
|
|
try:
|
2020-02-03 08:35:30 -07:00
|
|
|
event, context = await self.event_creation_handler.create_new_client_event(
|
2019-06-20 03:32:02 -06:00
|
|
|
builder=builder
|
2016-04-13 04:11:46 -06:00
|
|
|
)
|
2020-10-13 08:44:54 -06:00
|
|
|
except SynapseError as e:
|
2019-10-31 04:23:24 -06:00
|
|
|
logger.warning("Failed to create join to %s because %s", room_id, e)
|
2020-10-13 08:44:54 -06:00
|
|
|
raise
|
2019-06-12 03:31:37 -06:00
|
|
|
|
2016-07-15 02:29:54 -06:00
|
|
|
# The remote hasn't signed it yet, obviously. We'll do the full checks
|
|
|
|
# when we get the event back in `on_send_join_request`
|
2020-02-03 08:35:30 -07:00
|
|
|
await self.auth.check_from_context(
|
2019-06-20 03:32:02 -06:00
|
|
|
room_version, event, context, do_sig_check=False
|
2019-01-25 11:31:41 -07:00
|
|
|
)
|
2014-10-17 12:37:41 -06:00
|
|
|
|
2019-07-23 07:00:55 -06:00
|
|
|
return event
|
2014-10-16 09:56:51 -06:00
|
|
|
|
2020-02-03 08:32:48 -07:00
|
|
|
async def on_send_join_request(self, origin, pdu):
|
2014-11-12 09:20:21 -07:00
|
|
|
""" We have received a join event for a room. Fully process it and
|
|
|
|
respond with the current state and auth chains.
|
|
|
|
"""
|
2014-11-14 14:25:02 -07:00
|
|
|
event = pdu
|
2014-10-16 09:56:51 -06:00
|
|
|
|
2014-12-10 03:06:12 -07:00
|
|
|
logger.debug(
|
2019-10-28 06:43:23 -06:00
|
|
|
"on_send_join_request from %s: Got event: %s, signatures: %s",
|
|
|
|
origin,
|
2014-12-10 03:06:12 -07:00
|
|
|
event.event_id,
|
|
|
|
event.signatures,
|
|
|
|
)
|
|
|
|
|
2019-10-28 06:43:23 -06:00
|
|
|
if get_domain_from_id(event.sender) != origin:
|
|
|
|
logger.info(
|
|
|
|
"Got /send_join request for user %r from different origin %s",
|
|
|
|
event.sender,
|
|
|
|
origin,
|
|
|
|
)
|
|
|
|
raise SynapseError(403, "User not from origin", Codes.FORBIDDEN)
|
|
|
|
|
2014-12-08 02:08:26 -07:00
|
|
|
event.internal_metadata.outlier = False
|
2017-04-03 08:58:07 -06:00
|
|
|
# Send this event on behalf of the origin server.
|
|
|
|
#
|
|
|
|
# The reasons we have the destination server rather than the origin
|
|
|
|
# server send it are slightly mysterious: the origin server should have
|
2020-10-23 10:38:40 -06:00
|
|
|
# all the necessary state once it gets the response to the send_join,
|
2017-04-03 08:58:07 -06:00
|
|
|
# so it could send the event itself if it wanted to. It may be that
|
|
|
|
# doing it this way reduces failure modes, or avoids certain attacks
|
|
|
|
# where a new server selectively tells a subset of the federation that
|
|
|
|
# it has joined.
|
|
|
|
#
|
|
|
|
# The fact is that, as of the current writing, Synapse doesn't send out
|
|
|
|
# the join event over federation after joining, and changing it now
|
|
|
|
# would introduce the danger of backwards-compatibility problems.
|
2017-01-05 04:26:30 -07:00
|
|
|
event.internal_metadata.send_on_behalf_of = origin
|
2014-10-17 12:37:41 -06:00
|
|
|
|
2020-02-03 08:32:48 -07:00
|
|
|
context = await self._handle_new_event(origin, event)
|
2014-10-16 09:56:51 -06:00
|
|
|
|
2014-12-10 03:06:12 -07:00
|
|
|
logger.debug(
|
|
|
|
"on_send_join_request: After _handle_new_event: %s, sigs: %s",
|
|
|
|
event.event_id,
|
|
|
|
event.signatures,
|
|
|
|
)
|
|
|
|
|
2020-02-03 08:32:48 -07:00
|
|
|
prev_state_ids = await context.get_prev_state_ids()
|
2018-07-23 06:00:22 -06:00
|
|
|
|
|
|
|
state_ids = list(prev_state_ids.values())
|
2020-02-03 08:32:48 -07:00
|
|
|
auth_chain = await self.store.get_auth_chain(state_ids)
|
2014-11-07 04:22:12 -07:00
|
|
|
|
2020-02-03 08:32:48 -07:00
|
|
|
state = await self.store.get_events(list(prev_state_ids.values()))
|
2016-08-25 10:32:22 -06:00
|
|
|
|
2019-07-23 07:00:55 -06:00
|
|
|
return {"state": list(state.values()), "auth_chain": auth_chain}
|
2014-10-16 09:56:51 -06:00
|
|
|
|
2020-02-03 08:33:42 -07:00
|
|
|
async def on_invite_request(
|
2020-01-30 15:13:02 -07:00
|
|
|
self, origin: str, event: EventBase, room_version: RoomVersion
|
|
|
|
):
|
2014-11-12 09:20:21 -07:00
|
|
|
""" We've got an invite event. Process and persist it. Sign it.
|
|
|
|
|
|
|
|
Respond with the now signed event.
|
|
|
|
"""
|
2017-10-05 07:02:28 -06:00
|
|
|
if event.state_key is None:
|
|
|
|
raise SynapseError(400, "The invite event did not have a state key")
|
|
|
|
|
2020-02-03 08:33:42 -07:00
|
|
|
is_blocked = await self.store.is_room_blocked(event.room_id)
|
2017-06-19 05:36:28 -06:00
|
|
|
if is_blocked:
|
|
|
|
raise SynapseError(403, "This room has been blocked on this server")
|
|
|
|
|
2017-09-19 09:08:14 -06:00
|
|
|
if self.hs.config.block_non_admin_invites:
|
|
|
|
raise SynapseError(403, "This server does not accept room invites")
|
|
|
|
|
2017-10-05 07:02:28 -06:00
|
|
|
if not self.spam_checker.user_may_invite(
|
2019-06-20 03:32:02 -06:00
|
|
|
event.sender, event.state_key, event.room_id
|
2017-10-05 07:02:28 -06:00
|
|
|
):
|
2017-10-03 07:04:10 -06:00
|
|
|
raise SynapseError(
|
2017-10-05 07:02:28 -06:00
|
|
|
403, "This user is not permitted to send invites to this server/user"
|
2017-10-03 07:04:10 -06:00
|
|
|
)
|
2017-10-03 06:53:09 -06:00
|
|
|
|
2017-06-30 09:20:30 -06:00
|
|
|
membership = event.content.get("membership")
|
|
|
|
if event.type != EventTypes.Member or membership != Membership.INVITE:
|
|
|
|
raise SynapseError(400, "The event was not an m.room.member invite event")
|
|
|
|
|
|
|
|
sender_domain = get_domain_from_id(event.sender)
|
|
|
|
if sender_domain != origin:
|
2019-06-20 03:32:02 -06:00
|
|
|
raise SynapseError(
|
|
|
|
400, "The invite event was not from the server sending it"
|
|
|
|
)
|
2017-06-30 09:20:30 -06:00
|
|
|
|
|
|
|
if not self.is_mine_id(event.state_key):
|
|
|
|
raise SynapseError(400, "The invite event must be for this server")
|
|
|
|
|
2018-05-18 04:18:39 -06:00
|
|
|
# block any attempts to invite the server notices mxid
|
|
|
|
if event.state_key == self._server_notices_mxid:
|
2020-06-16 06:51:47 -06:00
|
|
|
raise SynapseError(HTTPStatus.FORBIDDEN, "Cannot invite this user")
|
2018-05-18 04:18:39 -06:00
|
|
|
|
2020-02-26 09:58:33 -07:00
|
|
|
# keep a record of the room version, if we don't yet know it.
|
|
|
|
# (this may get overwritten if we later get a different room version in a
|
|
|
|
# join dance).
|
2020-11-13 09:24:04 -07:00
|
|
|
await self._maybe_store_room_on_outlier_membership(
|
2020-02-26 09:58:33 -07:00
|
|
|
room_id=event.room_id, room_version=room_version
|
|
|
|
)
|
|
|
|
|
2014-12-05 09:20:48 -07:00
|
|
|
event.internal_metadata.outlier = True
|
2019-01-24 10:33:19 -07:00
|
|
|
event.internal_metadata.out_of_band_membership = True
|
2014-11-07 06:41:00 -07:00
|
|
|
|
|
|
|
event.signatures.update(
|
|
|
|
compute_event_signature(
|
2020-01-31 06:47:43 -07:00
|
|
|
room_version,
|
|
|
|
event.get_pdu_json(),
|
|
|
|
self.hs.hostname,
|
2020-07-08 10:51:56 -06:00
|
|
|
self.hs.signing_key,
|
2014-11-07 06:41:00 -07:00
|
|
|
)
|
|
|
|
)
|
|
|
|
|
2020-02-03 08:33:42 -07:00
|
|
|
context = await self.state_handler.compute_event_context(event)
|
2020-09-14 03:16:41 -06:00
|
|
|
await self.persist_events_and_notify(event.room_id, [(event, context)])
|
2014-11-07 06:41:00 -07:00
|
|
|
|
2019-07-23 07:00:55 -06:00
|
|
|
return event
|
2014-11-07 06:41:00 -07:00
|
|
|
|
2020-02-03 09:19:18 -07:00
|
|
|
async def do_remotely_reject_invite(
|
|
|
|
self, target_hosts: Iterable[str], room_id: str, user_id: str, content: JsonDict
|
2020-05-22 07:21:54 -06:00
|
|
|
) -> Tuple[EventBase, int]:
|
2020-02-03 09:19:18 -07:00
|
|
|
origin, event, room_version = await self._make_and_verify_event(
|
2019-12-11 06:07:25 -07:00
|
|
|
target_hosts, room_id, user_id, "leave", content=content
|
2017-04-20 18:24:17 -06:00
|
|
|
)
|
2017-06-09 06:05:05 -06:00
|
|
|
# Mark as outlier as we don't have any state for this event; we're not
|
|
|
|
# even in the room.
|
2017-06-09 03:08:18 -06:00
|
|
|
event.internal_metadata.outlier = True
|
2019-01-24 10:33:19 -07:00
|
|
|
event.internal_metadata.out_of_band_membership = True
|
2015-10-20 04:58:58 -06:00
|
|
|
|
2020-10-23 10:38:40 -06:00
|
|
|
# Try the host that we successfully called /make_leave/ on first for
|
2017-04-07 07:39:32 -06:00
|
|
|
# the /send_leave/ request.
|
2020-07-01 09:21:02 -06:00
|
|
|
host_list = list(target_hosts)
|
2015-10-20 04:58:58 -06:00
|
|
|
try:
|
2020-07-01 09:21:02 -06:00
|
|
|
host_list.remove(origin)
|
|
|
|
host_list.insert(0, origin)
|
2015-10-20 04:58:58 -06:00
|
|
|
except ValueError:
|
|
|
|
pass
|
|
|
|
|
2020-07-01 09:21:02 -06:00
|
|
|
await self.federation_client.send_leave(host_list, event)
|
2016-03-15 07:24:31 -06:00
|
|
|
|
2020-02-03 09:19:18 -07:00
|
|
|
context = await self.state_handler.compute_event_context(event)
|
2020-09-14 03:16:41 -06:00
|
|
|
stream_id = await self.persist_events_and_notify(
|
|
|
|
event.room_id, [(event, context)]
|
|
|
|
)
|
2016-03-15 07:24:31 -06:00
|
|
|
|
2020-05-22 07:21:54 -06:00
|
|
|
return event, stream_id
|
2015-10-20 04:58:58 -06:00
|
|
|
|
2020-02-03 09:22:30 -07:00
|
|
|
async def _make_and_verify_event(
|
|
|
|
self,
|
|
|
|
target_hosts: Iterable[str],
|
|
|
|
room_id: str,
|
|
|
|
user_id: str,
|
|
|
|
membership: str,
|
|
|
|
content: JsonDict = {},
|
2020-07-01 09:21:02 -06:00
|
|
|
params: Optional[Dict[str, Union[str, Iterable[str]]]] = None,
|
2020-02-03 09:22:30 -07:00
|
|
|
) -> Tuple[str, EventBase, RoomVersion]:
|
2020-01-27 07:30:57 -07:00
|
|
|
(
|
|
|
|
origin,
|
|
|
|
event,
|
|
|
|
room_version,
|
2020-02-03 09:22:30 -07:00
|
|
|
) = await self.federation_client.make_membership_event(
|
2019-06-20 03:32:02 -06:00
|
|
|
target_hosts, room_id, user_id, membership, content, params=params
|
2015-10-20 04:58:58 -06:00
|
|
|
)
|
|
|
|
|
2019-01-23 13:21:33 -07:00
|
|
|
logger.debug("Got response to make_%s: %s", membership, event)
|
2015-10-20 04:58:58 -06:00
|
|
|
|
|
|
|
# We should assert some things.
|
|
|
|
# FIXME: Do this in a nicer way
|
2019-06-20 03:32:02 -06:00
|
|
|
assert event.type == EventTypes.Member
|
|
|
|
assert event.user_id == user_id
|
|
|
|
assert event.state_key == user_id
|
|
|
|
assert event.room_id == room_id
|
2020-01-27 07:30:57 -07:00
|
|
|
return origin, event, room_version
|
2015-10-20 04:58:58 -06:00
|
|
|
|
2020-02-03 08:40:41 -07:00
|
|
|
async def on_make_leave_request(
|
|
|
|
self, origin: str, room_id: str, user_id: str
|
|
|
|
) -> EventBase:
|
2015-10-20 04:58:58 -06:00
|
|
|
""" We've received a /make_leave/ request, so we create a partial
|
2018-07-25 15:44:41 -06:00
|
|
|
leave event for the room and return that. We do *not* persist or
|
2015-10-20 04:58:58 -06:00
|
|
|
process it until the other server has signed it and sent it back.
|
2019-07-26 03:08:22 -06:00
|
|
|
|
|
|
|
Args:
|
2020-02-03 08:40:41 -07:00
|
|
|
origin: The (verified) server name of the requesting server.
|
|
|
|
room_id: Room to create leave event in
|
|
|
|
user_id: The user to create the leave for
|
2015-10-20 04:58:58 -06:00
|
|
|
"""
|
2019-07-26 03:08:22 -06:00
|
|
|
if get_domain_from_id(user_id) != origin:
|
|
|
|
logger.info(
|
|
|
|
"Got /make_leave request for user %r from different origin %s, ignoring",
|
|
|
|
user_id,
|
|
|
|
origin,
|
|
|
|
)
|
|
|
|
raise SynapseError(403, "User not from origin", Codes.FORBIDDEN)
|
|
|
|
|
2020-02-03 08:40:41 -07:00
|
|
|
room_version = await self.store.get_room_version_id(room_id)
|
2019-01-23 13:21:33 -07:00
|
|
|
builder = self.event_builder_factory.new(
|
|
|
|
room_version,
|
|
|
|
{
|
|
|
|
"type": EventTypes.Member,
|
|
|
|
"content": {"membership": Membership.LEAVE},
|
|
|
|
"room_id": room_id,
|
|
|
|
"sender": user_id,
|
|
|
|
"state_key": user_id,
|
2019-06-20 03:32:02 -06:00
|
|
|
},
|
2019-01-23 13:21:33 -07:00
|
|
|
)
|
2015-10-20 04:58:58 -06:00
|
|
|
|
2020-02-03 08:40:41 -07:00
|
|
|
event, context = await self.event_creation_handler.create_new_client_event(
|
2019-06-20 03:32:02 -06:00
|
|
|
builder=builder
|
2015-10-20 04:58:58 -06:00
|
|
|
)
|
|
|
|
|
2016-04-13 04:11:46 -06:00
|
|
|
try:
|
2016-07-15 02:29:54 -06:00
|
|
|
# The remote hasn't signed it yet, obviously. We'll do the full checks
|
|
|
|
# when we get the event back in `on_send_leave_request`
|
2020-02-03 08:40:41 -07:00
|
|
|
await self.auth.check_from_context(
|
2019-06-20 03:32:02 -06:00
|
|
|
room_version, event, context, do_sig_check=False
|
2019-01-25 11:31:41 -07:00
|
|
|
)
|
2016-04-13 04:11:46 -06:00
|
|
|
except AuthError as e:
|
2019-10-31 04:23:24 -06:00
|
|
|
logger.warning("Failed to create new leave %r because %s", event, e)
|
2016-04-13 04:11:46 -06:00
|
|
|
raise e
|
2015-10-20 04:58:58 -06:00
|
|
|
|
2019-07-23 07:00:55 -06:00
|
|
|
return event
|
2015-10-20 04:58:58 -06:00
|
|
|
|
2020-02-03 08:39:24 -07:00
|
|
|
async def on_send_leave_request(self, origin, pdu):
|
2015-10-20 04:58:58 -06:00
|
|
|
""" We have received a leave event for a room. Fully process it."""
|
|
|
|
event = pdu
|
|
|
|
|
|
|
|
logger.debug(
|
|
|
|
"on_send_leave_request: Got event: %s, signatures: %s",
|
|
|
|
event.event_id,
|
|
|
|
event.signatures,
|
|
|
|
)
|
|
|
|
|
2019-10-28 06:43:23 -06:00
|
|
|
if get_domain_from_id(event.sender) != origin:
|
|
|
|
logger.info(
|
|
|
|
"Got /send_leave request for user %r from different origin %s",
|
|
|
|
event.sender,
|
|
|
|
origin,
|
|
|
|
)
|
|
|
|
raise SynapseError(403, "User not from origin", Codes.FORBIDDEN)
|
|
|
|
|
2015-10-20 04:58:58 -06:00
|
|
|
event.internal_metadata.outlier = False
|
|
|
|
|
2020-10-13 08:17:13 -06:00
|
|
|
await self._handle_new_event(origin, event)
|
2019-06-12 03:31:37 -06:00
|
|
|
|
2015-10-20 04:58:58 -06:00
|
|
|
logger.debug(
|
|
|
|
"on_send_leave_request: After _handle_new_event: %s, sigs: %s",
|
|
|
|
event.event_id,
|
|
|
|
event.signatures,
|
|
|
|
)
|
|
|
|
|
2019-07-23 07:00:55 -06:00
|
|
|
return None
|
2015-10-20 04:58:58 -06:00
|
|
|
|
2020-04-24 12:36:38 -06:00
|
|
|
async def get_state_for_pdu(self, room_id: str, event_id: str) -> List[EventBase]:
|
2016-09-02 07:19:22 -06:00
|
|
|
"""Returns the state at the event. i.e. not including said event.
|
|
|
|
"""
|
2018-08-02 04:53:52 -06:00
|
|
|
|
2020-08-18 14:20:49 -06:00
|
|
|
event = await self.store.get_event(event_id, check_room_id=room_id)
|
2018-08-02 04:53:52 -06:00
|
|
|
|
2020-04-24 12:36:38 -06:00
|
|
|
state_groups = await self.state_store.get_state_groups(room_id, [event_id])
|
2014-10-17 08:04:17 -06:00
|
|
|
|
|
|
|
if state_groups:
|
2020-06-15 05:03:36 -06:00
|
|
|
_, state = list(state_groups.items()).pop()
|
2019-06-20 03:32:02 -06:00
|
|
|
results = {(e.type, e.state_key): e for e in state}
|
2014-10-30 05:53:35 -06:00
|
|
|
|
2018-08-02 06:23:48 -06:00
|
|
|
if event.is_state():
|
2014-10-30 05:53:35 -06:00
|
|
|
# Get previous state
|
2014-12-11 08:56:01 -07:00
|
|
|
if "replaces_state" in event.unsigned:
|
|
|
|
prev_id = event.unsigned["replaces_state"]
|
|
|
|
if prev_id != event.event_id:
|
2020-04-24 12:36:38 -06:00
|
|
|
prev_event = await self.store.get_event(prev_id)
|
2014-12-11 08:56:01 -07:00
|
|
|
results[(event.type, event.state_key)] = prev_event
|
2014-10-30 05:53:35 -06:00
|
|
|
else:
|
|
|
|
del results[(event.type, event.state_key)]
|
|
|
|
|
2018-05-31 03:03:47 -06:00
|
|
|
res = list(results.values())
|
2019-07-23 07:00:55 -06:00
|
|
|
return res
|
2014-10-17 08:04:17 -06:00
|
|
|
else:
|
2019-07-23 07:00:55 -06:00
|
|
|
return []
|
2014-10-17 08:04:17 -06:00
|
|
|
|
2020-04-24 12:36:38 -06:00
|
|
|
async def get_state_ids_for_pdu(self, room_id: str, event_id: str) -> List[str]:
|
2016-09-02 07:19:22 -06:00
|
|
|
"""Returns the state at the event. i.e. not including said event.
|
|
|
|
"""
|
2020-08-18 14:20:49 -06:00
|
|
|
event = await self.store.get_event(event_id, check_room_id=room_id)
|
2018-08-02 04:53:52 -06:00
|
|
|
|
2020-04-24 12:36:38 -06:00
|
|
|
state_groups = await self.state_store.get_state_groups_ids(room_id, [event_id])
|
2016-09-02 03:49:43 -06:00
|
|
|
|
|
|
|
if state_groups:
|
2018-09-06 08:22:23 -06:00
|
|
|
_, state = list(state_groups.items()).pop()
|
2016-09-02 03:49:43 -06:00
|
|
|
results = state
|
|
|
|
|
2018-08-02 06:23:48 -06:00
|
|
|
if event.is_state():
|
2016-09-02 03:49:43 -06:00
|
|
|
# Get previous state
|
|
|
|
if "replaces_state" in event.unsigned:
|
|
|
|
prev_id = event.unsigned["replaces_state"]
|
|
|
|
if prev_id != event.event_id:
|
|
|
|
results[(event.type, event.state_key)] = prev_id
|
|
|
|
else:
|
2017-02-28 03:01:19 -07:00
|
|
|
results.pop((event.type, event.state_key), None)
|
2016-09-02 03:49:43 -06:00
|
|
|
|
2019-07-23 07:00:55 -06:00
|
|
|
return list(results.values())
|
2016-09-02 03:49:43 -06:00
|
|
|
else:
|
2019-07-23 07:00:55 -06:00
|
|
|
return []
|
2016-09-02 03:49:43 -06:00
|
|
|
|
2014-10-31 03:59:02 -06:00
|
|
|
@log_function
|
2020-04-24 12:36:38 -06:00
|
|
|
async def on_backfill_request(
|
|
|
|
self, origin: str, room_id: str, pdu_list: List[str], limit: int
|
|
|
|
) -> List[EventBase]:
|
|
|
|
in_room = await self.auth.check_host_in_room(room_id, origin)
|
2014-11-10 04:59:51 -07:00
|
|
|
if not in_room:
|
|
|
|
raise AuthError(403, "Host not in room.")
|
2014-10-31 03:59:02 -06:00
|
|
|
|
2020-02-06 11:25:24 -07:00
|
|
|
# Synapse asks for 100 events per backfill request. Do not allow more.
|
|
|
|
limit = min(limit, 100)
|
|
|
|
|
2020-04-24 12:36:38 -06:00
|
|
|
events = await self.store.get_backfill_events(room_id, pdu_list, limit)
|
2014-10-31 03:59:02 -06:00
|
|
|
|
2020-04-24 12:36:38 -06:00
|
|
|
events = await filter_events_for_server(self.storage, origin, events)
|
2015-07-03 10:52:57 -06:00
|
|
|
|
2019-07-23 07:00:55 -06:00
|
|
|
return events
|
2014-10-31 03:59:02 -06:00
|
|
|
|
2014-10-31 04:47:34 -06:00
|
|
|
@log_function
|
2020-04-24 12:36:38 -06:00
|
|
|
async def get_persisted_pdu(
|
|
|
|
self, origin: str, event_id: str
|
|
|
|
) -> Optional[EventBase]:
|
2018-06-07 09:18:57 -06:00
|
|
|
"""Get an event from the database for the given server.
|
|
|
|
|
|
|
|
Args:
|
2020-04-24 12:36:38 -06:00
|
|
|
origin: hostname of server which is requesting the event; we
|
2018-06-07 09:18:57 -06:00
|
|
|
will check that the server is allowed to see it.
|
2020-04-24 12:36:38 -06:00
|
|
|
event_id: id of the event being requested
|
2014-10-31 04:47:34 -06:00
|
|
|
|
|
|
|
Returns:
|
2020-04-24 12:36:38 -06:00
|
|
|
None if we know nothing about the event; otherwise the (possibly-redacted) event.
|
2018-06-07 09:18:57 -06:00
|
|
|
|
|
|
|
Raises:
|
|
|
|
AuthError if the server is not currently in the room
|
2014-10-31 04:47:34 -06:00
|
|
|
"""
|
2020-04-24 12:36:38 -06:00
|
|
|
event = await self.store.get_event(
|
2019-06-20 03:32:02 -06:00
|
|
|
event_id, allow_none=True, allow_rejected=True
|
2014-10-31 04:47:34 -06:00
|
|
|
)
|
|
|
|
|
|
|
|
if event:
|
2020-04-24 12:36:38 -06:00
|
|
|
in_room = await self.auth.check_host_in_room(event.room_id, origin)
|
2018-06-07 09:18:57 -06:00
|
|
|
if not in_room:
|
|
|
|
raise AuthError(403, "Host not in room.")
|
2016-08-10 06:22:20 -06:00
|
|
|
|
2020-04-24 12:36:38 -06:00
|
|
|
events = await filter_events_for_server(self.storage, origin, [event])
|
2018-06-07 09:18:57 -06:00
|
|
|
event = events[0]
|
2019-07-23 07:00:55 -06:00
|
|
|
return event
|
2014-10-31 04:47:34 -06:00
|
|
|
else:
|
2019-07-23 07:00:55 -06:00
|
|
|
return None
|
2014-10-31 04:47:34 -06:00
|
|
|
|
2020-08-28 05:54:27 -06:00
|
|
|
async def get_min_depth_for_context(self, context):
|
|
|
|
return await self.store.get_min_depth(context)
|
2014-10-31 04:47:34 -06:00
|
|
|
|
2020-02-03 08:48:33 -07:00
|
|
|
async def _handle_new_event(
|
2019-06-20 03:32:02 -06:00
|
|
|
self, origin, event, state=None, auth_events=None, backfilled=False
|
|
|
|
):
|
2020-02-03 08:48:33 -07:00
|
|
|
context = await self._prep_event(
|
2019-06-20 03:32:02 -06:00
|
|
|
origin, event, state=state, auth_events=auth_events, backfilled=backfilled
|
2015-06-25 10:18:19 -06:00
|
|
|
)
|
|
|
|
|
2018-02-20 04:36:56 -07:00
|
|
|
try:
|
2019-11-05 15:13:37 -07:00
|
|
|
if (
|
|
|
|
not event.internal_metadata.is_outlier()
|
|
|
|
and not backfilled
|
|
|
|
and not context.rejected
|
|
|
|
):
|
2020-02-03 08:48:33 -07:00
|
|
|
await self.action_generator.handle_push_actions_for_event(
|
2018-02-20 04:36:56 -07:00
|
|
|
event, context
|
|
|
|
)
|
2016-02-09 09:19:15 -07:00
|
|
|
|
2020-02-03 08:48:33 -07:00
|
|
|
await self.persist_events_and_notify(
|
2020-09-14 03:16:41 -06:00
|
|
|
event.room_id, [(event, context)], backfilled=backfilled
|
2018-02-20 04:36:56 -07:00
|
|
|
)
|
2020-07-27 06:35:56 -06:00
|
|
|
except Exception:
|
|
|
|
run_in_background(
|
|
|
|
self.store.remove_push_actions_from_staging, event.event_id
|
|
|
|
)
|
|
|
|
raise
|
2015-06-25 10:18:19 -06:00
|
|
|
|
2019-07-23 07:00:55 -06:00
|
|
|
return context
|
2015-06-25 10:18:19 -06:00
|
|
|
|
2020-02-03 08:43:51 -07:00
|
|
|
async def _handle_new_events(
|
2019-12-05 08:02:35 -07:00
|
|
|
self,
|
|
|
|
origin: str,
|
2020-09-14 03:16:41 -06:00
|
|
|
room_id: str,
|
2019-12-05 08:02:35 -07:00
|
|
|
event_infos: Iterable[_NewEventInfo],
|
|
|
|
backfilled: bool = False,
|
2020-02-03 08:43:51 -07:00
|
|
|
) -> None:
|
2016-04-12 05:48:30 -06:00
|
|
|
"""Creates the appropriate contexts and persists events. The events
|
|
|
|
should not depend on one another, e.g. this should be used to persist
|
|
|
|
a bunch of outliers, but not a chunk of individual events that depend
|
|
|
|
on each other for state calculations.
|
2018-08-01 06:39:14 -06:00
|
|
|
|
|
|
|
Notifies about the events where appropriate.
|
2016-04-12 05:48:30 -06:00
|
|
|
"""
|
2018-09-27 04:25:34 -06:00
|
|
|
|
2020-02-03 08:43:51 -07:00
|
|
|
async def prep(ev_info: _NewEventInfo):
|
2019-12-05 08:02:35 -07:00
|
|
|
event = ev_info.event
|
2019-07-03 08:07:04 -06:00
|
|
|
with nested_logging_context(suffix=event.event_id):
|
2020-02-03 08:43:51 -07:00
|
|
|
res = await self._prep_event(
|
2015-06-25 10:18:19 -06:00
|
|
|
origin,
|
2018-09-27 04:25:34 -06:00
|
|
|
event,
|
2019-12-05 08:02:35 -07:00
|
|
|
state=ev_info.state,
|
|
|
|
auth_events=ev_info.auth_events,
|
2019-02-12 03:31:21 -07:00
|
|
|
backfilled=backfilled,
|
2015-06-25 10:18:19 -06:00
|
|
|
)
|
2019-07-23 07:00:55 -06:00
|
|
|
return res
|
2018-09-27 04:25:34 -06:00
|
|
|
|
2020-02-03 08:43:51 -07:00
|
|
|
contexts = await make_deferred_yieldable(
|
2019-06-20 03:32:02 -06:00
|
|
|
defer.gatherResults(
|
2019-07-03 08:07:04 -06:00
|
|
|
[run_in_background(prep, ev_info) for ev_info in event_infos],
|
2019-06-20 03:32:02 -06:00
|
|
|
consumeErrors=True,
|
|
|
|
)
|
|
|
|
)
|
2015-06-25 10:18:19 -06:00
|
|
|
|
2020-02-03 08:43:51 -07:00
|
|
|
await self.persist_events_and_notify(
|
2020-09-14 03:16:41 -06:00
|
|
|
room_id,
|
2015-06-25 10:18:19 -06:00
|
|
|
[
|
2019-12-05 08:02:35 -07:00
|
|
|
(ev_info.event, context)
|
2018-07-20 23:47:18 -06:00
|
|
|
for ev_info, context in zip(event_infos, contexts)
|
2015-06-25 10:18:19 -06:00
|
|
|
],
|
|
|
|
backfilled=backfilled,
|
2014-12-10 03:06:12 -07:00
|
|
|
)
|
|
|
|
|
2020-02-03 09:14:58 -07:00
|
|
|
async def _persist_auth_tree(
|
2020-01-27 07:30:57 -07:00
|
|
|
self,
|
|
|
|
origin: str,
|
2020-09-14 03:16:41 -06:00
|
|
|
room_id: str,
|
2020-01-27 07:30:57 -07:00
|
|
|
auth_events: List[EventBase],
|
|
|
|
state: List[EventBase],
|
|
|
|
event: EventBase,
|
|
|
|
room_version: RoomVersion,
|
2020-05-22 07:21:54 -06:00
|
|
|
) -> int:
|
2015-10-06 08:58:21 -06:00
|
|
|
"""Checks the auth chain is valid (and passes auth checks) for the
|
|
|
|
state and event. Then persists the auth chain and state atomically.
|
2018-08-01 06:39:14 -06:00
|
|
|
Persists the event separately. Notifies about the persisted events
|
|
|
|
where appropriate.
|
2015-10-06 08:58:21 -06:00
|
|
|
|
2016-07-29 04:17:04 -06:00
|
|
|
Will attempt to fetch missing auth events.
|
|
|
|
|
2016-07-29 03:45:05 -06:00
|
|
|
Args:
|
2020-01-27 07:30:57 -07:00
|
|
|
origin: Where the events came from
|
2020-09-14 03:16:41 -06:00
|
|
|
room_id,
|
2020-01-27 07:30:57 -07:00
|
|
|
auth_events
|
|
|
|
state
|
|
|
|
event
|
|
|
|
room_version: The room version we expect this room to have, and
|
|
|
|
will raise if it doesn't match the version in the create event.
|
2015-10-06 08:58:21 -06:00
|
|
|
"""
|
|
|
|
events_to_context = {}
|
|
|
|
for e in itertools.chain(auth_events, state):
|
|
|
|
e.internal_metadata.outlier = True
|
2020-02-03 09:14:58 -07:00
|
|
|
ctx = await self.state_handler.compute_event_context(e)
|
2016-03-31 08:32:24 -06:00
|
|
|
events_to_context[e.event_id] = ctx
|
2015-10-06 08:58:21 -06:00
|
|
|
|
|
|
|
event_map = {
|
2019-06-20 03:32:02 -06:00
|
|
|
e.event_id: e for e in itertools.chain(auth_events, state, [event])
|
2015-10-06 08:58:21 -06:00
|
|
|
}
|
|
|
|
|
|
|
|
create_event = None
|
|
|
|
for e in auth_events:
|
|
|
|
if (e.type, e.state_key) == (EventTypes.Create, ""):
|
|
|
|
create_event = e
|
|
|
|
break
|
|
|
|
|
2019-01-24 11:31:23 -07:00
|
|
|
if create_event is None:
|
|
|
|
# If the state doesn't have a create event then the room is
|
|
|
|
# invalid, and it would fail auth checks anyway.
|
|
|
|
raise SynapseError(400, "No create event in state")
|
|
|
|
|
2020-01-27 07:30:57 -07:00
|
|
|
room_version_id = create_event.content.get(
|
2019-06-20 03:32:02 -06:00
|
|
|
"room_version", RoomVersions.V1.identifier
|
2019-04-01 03:24:38 -06:00
|
|
|
)
|
2019-01-23 10:19:58 -07:00
|
|
|
|
2020-01-27 07:30:57 -07:00
|
|
|
if room_version.identifier != room_version_id:
|
|
|
|
raise SynapseError(400, "Room version mismatch")
|
|
|
|
|
2016-07-28 09:08:33 -06:00
|
|
|
missing_auth_events = set()
|
|
|
|
for e in itertools.chain(auth_events, state, [event]):
|
2018-11-05 06:35:15 -07:00
|
|
|
for e_id in e.auth_event_ids():
|
2016-07-28 09:08:33 -06:00
|
|
|
if e_id not in event_map:
|
|
|
|
missing_auth_events.add(e_id)
|
|
|
|
|
|
|
|
for e_id in missing_auth_events:
|
2020-02-03 09:14:58 -07:00
|
|
|
m_ev = await self.federation_client.get_pdu(
|
2020-01-31 07:07:31 -07:00
|
|
|
[origin], e_id, room_version=room_version, outlier=True, timeout=10000,
|
2016-07-28 09:08:33 -06:00
|
|
|
)
|
|
|
|
if m_ev and m_ev.event_id == e_id:
|
|
|
|
event_map[e_id] = m_ev
|
|
|
|
else:
|
|
|
|
logger.info("Failed to find auth event %r", e_id)
|
|
|
|
|
2015-10-06 08:58:21 -06:00
|
|
|
for e in itertools.chain(auth_events, state, [event]):
|
|
|
|
auth_for_e = {
|
|
|
|
(event_map[e_id].type, event_map[e_id].state_key): event_map[e_id]
|
2018-11-05 06:35:15 -07:00
|
|
|
for e_id in e.auth_event_ids()
|
2016-07-28 09:08:33 -06:00
|
|
|
if e_id in event_map
|
2015-10-06 08:58:21 -06:00
|
|
|
}
|
|
|
|
if create_event:
|
|
|
|
auth_for_e[(EventTypes.Create, "")] = create_event
|
|
|
|
|
|
|
|
try:
|
2019-10-18 11:43:36 -06:00
|
|
|
event_auth.check(room_version, e, auth_events=auth_for_e)
|
2016-01-27 10:02:10 -07:00
|
|
|
except SynapseError as err:
|
|
|
|
# we may get SynapseErrors here as well as AuthErrors. For
|
|
|
|
# instance, there are a couple of (ancient) events in some
|
|
|
|
# rooms whose senders do not have the correct sigil; these
|
|
|
|
# cause SynapseErrors in auth.check. We don't want to give up
|
|
|
|
# the attempt to federate altogether in such cases.
|
|
|
|
|
2019-10-31 04:23:24 -06:00
|
|
|
logger.warning("Rejecting %s because %s", e.event_id, err.msg)
|
2015-10-06 08:58:21 -06:00
|
|
|
|
|
|
|
if e == event:
|
|
|
|
raise
|
|
|
|
events_to_context[e.event_id].rejected = RejectedReason.AUTH_ERROR
|
|
|
|
|
2020-02-03 09:14:58 -07:00
|
|
|
await self.persist_events_and_notify(
|
2020-09-14 03:16:41 -06:00
|
|
|
room_id,
|
2015-10-06 08:58:21 -06:00
|
|
|
[
|
|
|
|
(e, events_to_context[e.event_id])
|
|
|
|
for e in itertools.chain(auth_events, state)
|
2020-09-14 03:16:41 -06:00
|
|
|
],
|
2015-10-06 08:58:21 -06:00
|
|
|
)
|
|
|
|
|
2020-02-03 09:14:58 -07:00
|
|
|
new_event_context = await self.state_handler.compute_event_context(
|
2016-03-31 08:32:24 -06:00
|
|
|
event, old_state=state
|
2015-10-06 08:58:21 -06:00
|
|
|
)
|
|
|
|
|
2020-09-14 03:16:41 -06:00
|
|
|
return await self.persist_events_and_notify(
|
|
|
|
room_id, [(event, new_event_context)]
|
|
|
|
)
|
2015-10-06 08:58:21 -06:00
|
|
|
|
2020-02-03 08:49:32 -07:00
|
|
|
async def _prep_event(
|
2019-12-05 08:02:35 -07:00
|
|
|
self,
|
|
|
|
origin: str,
|
|
|
|
event: EventBase,
|
|
|
|
state: Optional[Iterable[EventBase]],
|
2020-08-28 05:28:53 -06:00
|
|
|
auth_events: Optional[MutableStateMap[EventBase]],
|
2019-12-05 08:02:35 -07:00
|
|
|
backfilled: bool,
|
2020-02-03 08:49:32 -07:00
|
|
|
) -> EventContext:
|
|
|
|
context = await self.state_handler.compute_event_context(event, old_state=state)
|
2014-11-25 04:31:18 -07:00
|
|
|
|
2015-01-29 09:50:23 -07:00
|
|
|
if not auth_events:
|
2020-02-03 08:49:32 -07:00
|
|
|
prev_state_ids = await context.get_prev_state_ids()
|
2020-08-06 06:30:06 -06:00
|
|
|
auth_events_ids = self.auth.compute_auth_events(
|
2019-06-20 03:32:02 -06:00
|
|
|
event, prev_state_ids, for_verification=True
|
2016-08-25 10:32:22 -06:00
|
|
|
)
|
2020-07-01 09:21:02 -06:00
|
|
|
auth_events_x = await self.store.get_events(auth_events_ids)
|
|
|
|
auth_events = {(e.type, e.state_key): e for e in auth_events_x.values()}
|
2015-01-29 09:50:23 -07:00
|
|
|
|
|
|
|
# This is a hack to fix some old rooms where the initial join event
|
|
|
|
# didn't reference the create event in its auth events.
|
2018-11-05 06:35:15 -07:00
|
|
|
if event.type == EventTypes.Member and not event.auth_event_ids():
|
|
|
|
if len(event.prev_event_ids()) == 1 and event.depth < 5:
|
2020-02-03 08:49:32 -07:00
|
|
|
c = await self.store.get_event(
|
2019-06-20 03:32:02 -06:00
|
|
|
event.prev_event_ids()[0], allow_none=True
|
2015-05-19 07:15:05 -06:00
|
|
|
)
|
|
|
|
if c and c.type == EventTypes.Create:
|
2015-01-29 09:50:23 -07:00
|
|
|
auth_events[(c.type, c.state_key)] = c
|
2014-12-10 03:06:12 -07:00
|
|
|
|
2020-02-03 08:49:32 -07:00
|
|
|
context = await self.do_auth(origin, event, context, auth_events=auth_events)
|
2015-01-28 09:16:53 -07:00
|
|
|
|
2019-03-08 04:26:33 -07:00
|
|
|
if not context.rejected:
|
2020-02-03 08:49:32 -07:00
|
|
|
await self._check_for_soft_fail(event, state, backfilled)
|
2019-03-08 04:26:33 -07:00
|
|
|
|
|
|
|
if event.type == EventTypes.GuestAccess and not context.rejected:
|
2020-02-03 08:49:32 -07:00
|
|
|
await self.maybe_kick_guest_users(event)
|
2019-03-08 04:26:33 -07:00
|
|
|
|
2019-07-23 07:00:55 -06:00
|
|
|
return context
|
2019-03-08 04:26:33 -07:00
|
|
|
|
2020-02-03 09:16:31 -07:00
|
|
|
async def _check_for_soft_fail(
|
2019-12-05 08:02:35 -07:00
|
|
|
self, event: EventBase, state: Optional[Iterable[EventBase]], backfilled: bool
|
2020-02-03 09:16:31 -07:00
|
|
|
) -> None:
|
|
|
|
"""Checks if we should soft fail the event; if so, marks the event as
|
2019-03-08 04:26:33 -07:00
|
|
|
such.
|
|
|
|
|
|
|
|
Args:
|
2019-12-05 08:02:35 -07:00
|
|
|
event
|
|
|
|
state: The state at the event if we don't have all the event's prev events
|
|
|
|
backfilled: Whether the event is from backfill
|
2019-03-08 04:26:33 -07:00
|
|
|
"""
|
2019-02-12 03:31:21 -07:00
|
|
|
# For new (non-backfilled and non-outlier) events we check if the event
|
|
|
|
# passes auth based on the current state. If it doesn't then we
|
|
|
|
# "soft-fail" the event.
|
2020-07-01 09:41:19 -06:00
|
|
|
if backfilled or event.internal_metadata.is_outlier():
|
|
|
|
return
|
2019-02-12 03:31:21 -07:00
|
|
|
|
2020-09-03 08:38:32 -06:00
|
|
|
extrem_ids_list = await self.store.get_latest_event_ids_in_room(event.room_id)
|
|
|
|
extrem_ids = set(extrem_ids_list)
|
2020-07-01 09:41:19 -06:00
|
|
|
prev_event_ids = set(event.prev_event_ids())
|
2019-02-12 03:31:21 -07:00
|
|
|
|
2020-07-01 09:41:19 -06:00
|
|
|
if extrem_ids == prev_event_ids:
|
|
|
|
# If they're the same then the current state is the same as the
|
|
|
|
# state at the event, so no point rechecking auth for soft fail.
|
|
|
|
return
|
|
|
|
|
|
|
|
room_version = await self.store.get_room_version_id(event.room_id)
|
|
|
|
room_version_obj = KNOWN_ROOM_VERSIONS[room_version]
|
|
|
|
|
|
|
|
# Calculate the "current state".
|
|
|
|
if state is not None:
|
|
|
|
# If we're explicitly given the state then we won't have all the
|
|
|
|
# prev events, and so we have a gap in the graph. In this case
|
|
|
|
# we want to be a little careful as we might have been down for
|
|
|
|
# a while and have an incorrect view of the current state,
|
|
|
|
# however we still want to do checks as gaps are easy to
|
|
|
|
# maliciously manufacture.
|
|
|
|
#
|
|
|
|
# So we use a "current state" that is actually a state
|
|
|
|
# resolution across the current forward extremities and the
|
|
|
|
# given state at the event. This should correctly handle cases
|
|
|
|
# like bans, especially with state res v2.
|
|
|
|
|
2020-09-29 08:57:36 -06:00
|
|
|
state_sets_d = await self.state_store.get_state_groups(
|
2020-07-01 09:41:19 -06:00
|
|
|
event.room_id, extrem_ids
|
|
|
|
)
|
2020-09-29 08:57:36 -06:00
|
|
|
state_sets = list(state_sets_d.values()) # type: List[Iterable[EventBase]]
|
2020-07-01 09:41:19 -06:00
|
|
|
state_sets.append(state)
|
2020-08-24 12:25:27 -06:00
|
|
|
current_states = await self.state_handler.resolve_events(
|
2020-07-01 09:41:19 -06:00
|
|
|
room_version, state_sets, event
|
|
|
|
)
|
2020-08-28 05:28:53 -06:00
|
|
|
current_state_ids = {
|
|
|
|
k: e.event_id for k, e in current_states.items()
|
|
|
|
} # type: StateMap[str]
|
2020-07-01 09:41:19 -06:00
|
|
|
else:
|
|
|
|
current_state_ids = await self.state_handler.get_current_state_ids(
|
|
|
|
event.room_id, latest_event_ids=extrem_ids
|
2019-05-21 09:10:54 -06:00
|
|
|
)
|
|
|
|
|
2020-07-01 09:41:19 -06:00
|
|
|
logger.debug(
|
|
|
|
"Doing soft-fail check for %s: state %s", event.event_id, current_state_ids,
|
|
|
|
)
|
2019-02-12 03:31:21 -07:00
|
|
|
|
2020-07-01 09:41:19 -06:00
|
|
|
# Now check if event pass auth against said current state
|
|
|
|
auth_types = auth_types_for_event(event)
|
2020-08-24 12:25:27 -06:00
|
|
|
current_state_ids_list = [
|
|
|
|
e for k, e in current_state_ids.items() if k in auth_types
|
|
|
|
]
|
2019-02-12 03:31:21 -07:00
|
|
|
|
2020-08-24 12:25:27 -06:00
|
|
|
auth_events_map = await self.store.get_events(current_state_ids_list)
|
2020-07-01 09:41:19 -06:00
|
|
|
current_auth_events = {
|
2020-08-18 14:20:49 -06:00
|
|
|
(e.type, e.state_key): e for e in auth_events_map.values()
|
2020-07-01 09:41:19 -06:00
|
|
|
}
|
|
|
|
|
|
|
|
try:
|
|
|
|
event_auth.check(room_version_obj, event, auth_events=current_auth_events)
|
|
|
|
except AuthError as e:
|
|
|
|
logger.warning("Soft-failing %r because %s", event, e)
|
|
|
|
event.internal_metadata.soft_failed = True
|
2019-02-12 03:31:21 -07:00
|
|
|
|
2020-02-03 08:30:23 -07:00
|
|
|
async def on_query_auth(
|
2019-06-20 03:32:02 -06:00
|
|
|
self, origin, event_id, room_id, remote_auth_chain, rejects, missing
|
|
|
|
):
|
2020-02-03 08:30:23 -07:00
|
|
|
in_room = await self.auth.check_host_in_room(room_id, origin)
|
2018-08-02 04:53:52 -06:00
|
|
|
if not in_room:
|
|
|
|
raise AuthError(403, "Host not in room.")
|
|
|
|
|
2020-08-18 14:20:49 -06:00
|
|
|
event = await self.store.get_event(event_id, check_room_id=room_id)
|
2018-08-02 04:53:52 -06:00
|
|
|
|
2015-01-29 09:50:23 -07:00
|
|
|
# Just go through and process each event in `remote_auth_chain`. We
|
|
|
|
# don't want to fall into the trap of `missing` being wrong.
|
|
|
|
for e in remote_auth_chain:
|
|
|
|
try:
|
2020-02-03 08:30:23 -07:00
|
|
|
await self._handle_new_event(origin, e)
|
2015-01-29 09:50:23 -07:00
|
|
|
except AuthError:
|
|
|
|
pass
|
|
|
|
|
|
|
|
# Now get the current auth_chain for the event.
|
2020-02-03 08:30:23 -07:00
|
|
|
local_auth_chain = await self.store.get_auth_chain(
|
2020-02-21 05:15:07 -07:00
|
|
|
list(event.auth_event_ids()), include_given=True
|
2017-05-24 07:22:41 -06:00
|
|
|
)
|
2015-01-29 09:50:23 -07:00
|
|
|
|
|
|
|
# TODO: Check if we would now reject event_id. If so we need to tell
|
|
|
|
# everyone.
|
|
|
|
|
2020-02-03 08:30:23 -07:00
|
|
|
ret = await self.construct_auth_difference(local_auth_chain, remote_auth_chain)
|
2014-12-10 03:06:12 -07:00
|
|
|
|
2015-02-03 06:23:58 -07:00
|
|
|
logger.debug("on_query_auth returning: %s", ret)
|
2015-01-29 09:50:23 -07:00
|
|
|
|
2019-07-23 07:00:55 -06:00
|
|
|
return ret
|
2015-01-28 09:16:53 -07:00
|
|
|
|
2020-02-03 12:15:08 -07:00
|
|
|
async def on_get_missing_events(
|
2019-06-20 03:32:02 -06:00
|
|
|
self, origin, room_id, earliest_events, latest_events, limit
|
|
|
|
):
|
2020-02-03 12:15:08 -07:00
|
|
|
in_room = await self.auth.check_host_in_room(room_id, origin)
|
2015-02-23 06:58:02 -07:00
|
|
|
if not in_room:
|
|
|
|
raise AuthError(403, "Host not in room.")
|
|
|
|
|
2020-02-06 11:25:24 -07:00
|
|
|
# Only allow up to 20 events to be retrieved per request.
|
2015-02-23 06:58:02 -07:00
|
|
|
limit = min(limit, 20)
|
|
|
|
|
2020-02-03 12:15:08 -07:00
|
|
|
missing_events = await self.store.get_missing_events(
|
2015-02-23 06:58:02 -07:00
|
|
|
room_id=room_id,
|
|
|
|
earliest_events=earliest_events,
|
|
|
|
latest_events=latest_events,
|
|
|
|
limit=limit,
|
|
|
|
)
|
|
|
|
|
2020-02-03 12:15:08 -07:00
|
|
|
missing_events = await filter_events_for_server(
|
2019-10-23 10:25:54 -06:00
|
|
|
self.storage, origin, missing_events
|
2018-06-08 04:34:46 -06:00
|
|
|
)
|
|
|
|
|
2019-07-23 07:00:55 -06:00
|
|
|
return missing_events
|
2015-02-23 06:58:02 -07:00
|
|
|
|
2020-02-03 08:51:24 -07:00
|
|
|
async def do_auth(
|
|
|
|
self,
|
|
|
|
origin: str,
|
|
|
|
event: EventBase,
|
|
|
|
context: EventContext,
|
2020-08-28 05:28:53 -06:00
|
|
|
auth_events: MutableStateMap[EventBase],
|
2020-02-03 08:51:24 -07:00
|
|
|
) -> EventContext:
|
2017-11-07 09:43:00 -07:00
|
|
|
"""
|
|
|
|
|
|
|
|
Args:
|
2020-02-03 08:51:24 -07:00
|
|
|
origin:
|
|
|
|
event:
|
|
|
|
context:
|
|
|
|
auth_events:
|
2019-05-23 04:17:42 -06:00
|
|
|
Map from (event_type, state_key) to event
|
|
|
|
|
2019-11-08 05:18:20 -07:00
|
|
|
Normally, our calculated auth_events based on the state of the room
|
|
|
|
at the event's position in the DAG, though occasionally (eg if the
|
|
|
|
event is an outlier), may be the auth events claimed by the remote
|
|
|
|
server.
|
2019-05-23 04:17:42 -06:00
|
|
|
|
|
|
|
Also NB that this function adds entries to it.
|
|
|
|
Returns:
|
2020-02-03 08:51:24 -07:00
|
|
|
updated context object
|
2019-05-23 04:17:42 -06:00
|
|
|
"""
|
2020-02-03 08:51:24 -07:00
|
|
|
room_version = await self.store.get_room_version_id(event.room_id)
|
2020-01-28 07:18:29 -07:00
|
|
|
room_version_obj = KNOWN_ROOM_VERSIONS[room_version]
|
2019-05-23 04:17:42 -06:00
|
|
|
|
2019-06-03 02:56:45 -06:00
|
|
|
try:
|
2020-02-03 08:51:24 -07:00
|
|
|
context = await self._update_auth_events_and_context_for_auth(
|
2019-06-03 02:56:45 -06:00
|
|
|
origin, event, context, auth_events
|
|
|
|
)
|
|
|
|
except Exception:
|
|
|
|
# We don't really mind if the above fails, so lets not fail
|
2019-06-05 04:31:27 -06:00
|
|
|
# processing if it does. However, it really shouldn't fail so
|
|
|
|
# let's still log as an exception since we'll still want to fix
|
|
|
|
# any bugs.
|
|
|
|
logger.exception(
|
|
|
|
"Failed to double check auth events for %s with remote. "
|
|
|
|
"Ignoring failure and continuing processing of event.",
|
|
|
|
event.event_id,
|
|
|
|
)
|
2019-06-03 02:56:45 -06:00
|
|
|
|
2019-05-23 04:17:42 -06:00
|
|
|
try:
|
2020-01-28 07:18:29 -07:00
|
|
|
event_auth.check(room_version_obj, event, auth_events=auth_events)
|
2019-05-23 04:17:42 -06:00
|
|
|
except AuthError as e:
|
2019-10-31 04:23:24 -06:00
|
|
|
logger.warning("Failed auth resolution for %r because %s", event, e)
|
2019-11-01 10:19:09 -06:00
|
|
|
context.rejected = RejectedReason.AUTH_ERROR
|
|
|
|
|
|
|
|
return context
|
2019-05-23 04:17:42 -06:00
|
|
|
|
2020-02-03 08:53:30 -07:00
|
|
|
async def _update_auth_events_and_context_for_auth(
|
|
|
|
self,
|
|
|
|
origin: str,
|
|
|
|
event: EventBase,
|
|
|
|
context: EventContext,
|
2020-08-28 05:28:53 -06:00
|
|
|
auth_events: MutableStateMap[EventBase],
|
2020-02-03 08:53:30 -07:00
|
|
|
) -> EventContext:
|
2019-05-23 04:17:42 -06:00
|
|
|
"""Helper for do_auth. See there for docs.
|
|
|
|
|
2019-06-03 02:56:45 -06:00
|
|
|
Checks whether a given event has the expected auth events. If it
|
|
|
|
doesn't then we talk to the remote server to compare state to see if
|
|
|
|
we can come to a consensus (e.g. if one server missed some valid
|
|
|
|
state).
|
|
|
|
|
2020-02-03 08:53:30 -07:00
|
|
|
This attempts to resolve any potential divergence of state between
|
2019-06-03 02:56:45 -06:00
|
|
|
servers, but is not essential and so failures should not block further
|
|
|
|
processing of the event.
|
|
|
|
|
2019-05-23 04:17:42 -06:00
|
|
|
Args:
|
2020-02-03 08:53:30 -07:00
|
|
|
origin:
|
|
|
|
event:
|
|
|
|
context:
|
2019-11-08 05:18:20 -07:00
|
|
|
|
2020-02-03 08:53:30 -07:00
|
|
|
auth_events:
|
2019-11-08 05:18:20 -07:00
|
|
|
Map from (event_type, state_key) to event
|
|
|
|
|
|
|
|
Normally, our calculated auth_events based on the state of the room
|
|
|
|
at the event's position in the DAG, though occasionally (eg if the
|
|
|
|
event is an outlier), may be the auth events claimed by the remote
|
|
|
|
server.
|
|
|
|
|
|
|
|
Also NB that this function adds entries to it.
|
2017-11-07 09:43:00 -07:00
|
|
|
|
|
|
|
Returns:
|
2020-02-03 08:53:30 -07:00
|
|
|
updated context
|
2017-11-07 09:43:00 -07:00
|
|
|
"""
|
2018-11-05 06:35:15 -07:00
|
|
|
event_auth_events = set(event.auth_event_ids())
|
2015-06-25 10:18:19 -06:00
|
|
|
|
2019-11-08 05:18:20 -07:00
|
|
|
# missing_auth is the set of the event's auth_events which we don't yet have
|
|
|
|
# in auth_events.
|
2019-05-23 04:17:42 -06:00
|
|
|
missing_auth = event_auth_events.difference(
|
|
|
|
e.event_id for e in auth_events.values()
|
|
|
|
)
|
|
|
|
|
2019-11-08 05:18:20 -07:00
|
|
|
# if we have missing events, we need to fetch those events from somewhere.
|
|
|
|
#
|
|
|
|
# we start by checking if they are in the store, and then try calling /event_auth/.
|
2019-05-23 04:17:42 -06:00
|
|
|
if missing_auth:
|
2020-02-03 08:53:30 -07:00
|
|
|
have_events = await self.store.have_seen_events(missing_auth)
|
2019-12-04 10:27:32 -07:00
|
|
|
logger.debug("Events %s are in the store", have_events)
|
|
|
|
missing_auth.difference_update(have_events)
|
2015-06-25 10:18:19 -06:00
|
|
|
|
2015-01-28 09:16:53 -07:00
|
|
|
if missing_auth:
|
2015-01-29 09:50:23 -07:00
|
|
|
# If we don't have all the auth events, we need to get them.
|
2019-06-20 03:32:02 -06:00
|
|
|
logger.info("auth_events contains unknown events: %s", missing_auth)
|
2015-02-03 06:57:54 -07:00
|
|
|
try:
|
2019-06-03 02:56:45 -06:00
|
|
|
try:
|
2020-02-03 08:53:30 -07:00
|
|
|
remote_auth_chain = await self.federation_client.get_event_auth(
|
2019-06-03 02:56:45 -06:00
|
|
|
origin, event.room_id, event.event_id
|
|
|
|
)
|
2020-07-01 09:21:02 -06:00
|
|
|
except RequestSendFailed as e1:
|
2019-06-03 02:56:45 -06:00
|
|
|
# The other side isn't around or doesn't implement the
|
|
|
|
# endpoint, so lets just bail out.
|
2020-07-01 09:21:02 -06:00
|
|
|
logger.info("Failed to get event auth from remote: %s", e1)
|
2019-11-01 10:19:09 -06:00
|
|
|
return context
|
2015-01-29 09:50:23 -07:00
|
|
|
|
2020-02-03 08:53:30 -07:00
|
|
|
seen_remotes = await self.store.have_seen_events(
|
2015-02-03 06:57:54 -07:00
|
|
|
[e.event_id for e in remote_auth_chain]
|
|
|
|
)
|
2015-01-30 09:51:58 -07:00
|
|
|
|
2015-02-03 06:57:54 -07:00
|
|
|
for e in remote_auth_chain:
|
2018-04-17 11:30:53 -06:00
|
|
|
if e.event_id in seen_remotes:
|
2015-02-03 06:57:54 -07:00
|
|
|
continue
|
2015-01-30 09:51:58 -07:00
|
|
|
|
2015-02-03 06:57:54 -07:00
|
|
|
if e.event_id == event.event_id:
|
|
|
|
continue
|
2015-01-30 09:51:58 -07:00
|
|
|
|
2015-02-03 06:57:54 -07:00
|
|
|
try:
|
2018-11-05 06:35:15 -07:00
|
|
|
auth_ids = e.auth_event_ids()
|
2015-02-03 06:57:54 -07:00
|
|
|
auth = {
|
2019-06-20 03:32:02 -06:00
|
|
|
(e.type, e.state_key): e
|
|
|
|
for e in remote_auth_chain
|
2015-10-02 06:11:49 -06:00
|
|
|
if e.event_id in auth_ids or e.type == EventTypes.Create
|
2015-02-03 06:57:54 -07:00
|
|
|
}
|
|
|
|
e.internal_metadata.outlier = True
|
2015-01-30 09:51:58 -07:00
|
|
|
|
2015-02-03 06:57:54 -07:00
|
|
|
logger.debug(
|
2019-06-20 03:32:02 -06:00
|
|
|
"do_auth %s missing_auth: %s", event.event_id, e.event_id
|
2015-02-03 06:57:54 -07:00
|
|
|
)
|
2020-02-03 08:53:30 -07:00
|
|
|
await self._handle_new_event(origin, e, auth_events=auth)
|
2015-01-30 09:51:58 -07:00
|
|
|
|
2015-02-03 06:57:54 -07:00
|
|
|
if e.event_id in event_auth_events:
|
|
|
|
auth_events[(e.type, e.state_key)] = e
|
|
|
|
except AuthError:
|
|
|
|
pass
|
2015-02-06 07:16:50 -07:00
|
|
|
|
2017-10-23 08:52:32 -06:00
|
|
|
except Exception:
|
2015-02-03 06:57:54 -07:00
|
|
|
logger.exception("Failed to get auth chain")
|
2015-01-29 09:50:23 -07:00
|
|
|
|
2019-05-23 04:17:42 -06:00
|
|
|
if event.internal_metadata.is_outlier():
|
2019-11-08 05:18:20 -07:00
|
|
|
# XXX: given that, for an outlier, we'll be working with the
|
|
|
|
# event's *claimed* auth events rather than those we calculated:
|
|
|
|
# (a) is there any point in this test, since different_auth below will
|
|
|
|
# obviously be empty
|
|
|
|
# (b) alternatively, why don't we do it earlier?
|
2019-05-23 04:17:42 -06:00
|
|
|
logger.info("Skipping auth_event fetch for outlier")
|
2019-11-01 10:19:09 -06:00
|
|
|
return context
|
2019-05-23 04:17:42 -06:00
|
|
|
|
|
|
|
different_auth = event_auth_events.difference(
|
|
|
|
e.event_id for e in auth_events.values()
|
|
|
|
)
|
2015-01-29 09:50:23 -07:00
|
|
|
|
2019-05-23 04:17:42 -06:00
|
|
|
if not different_auth:
|
2019-11-01 10:19:09 -06:00
|
|
|
return context
|
2019-01-25 11:31:41 -07:00
|
|
|
|
2019-05-23 04:17:42 -06:00
|
|
|
logger.info(
|
|
|
|
"auth_events refers to events which are not in our calculated auth "
|
|
|
|
"chain: %s",
|
|
|
|
different_auth,
|
|
|
|
)
|
|
|
|
|
2019-12-04 10:27:32 -07:00
|
|
|
# XXX: currently this checks for redactions but I'm not convinced that is
|
|
|
|
# necessary?
|
2020-02-03 08:53:30 -07:00
|
|
|
different_events = await self.store.get_events_as_list(different_auth)
|
2019-05-23 04:17:42 -06:00
|
|
|
|
2019-12-05 07:14:45 -07:00
|
|
|
for d in different_events:
|
|
|
|
if d.room_id != event.room_id:
|
|
|
|
logger.warning(
|
|
|
|
"Event %s refers to auth_event %s which is in a different room",
|
|
|
|
event.event_id,
|
|
|
|
d.event_id,
|
|
|
|
)
|
|
|
|
|
|
|
|
# don't attempt to resolve the claimed auth events against our own
|
|
|
|
# in this case: just use our own auth events.
|
|
|
|
#
|
|
|
|
# XXX: should we reject the event in this case? It feels like we should,
|
|
|
|
# but then shouldn't we also do so if we've failed to fetch any of the
|
|
|
|
# auth events?
|
|
|
|
return context
|
2015-02-13 07:20:05 -07:00
|
|
|
|
2019-12-05 07:14:45 -07:00
|
|
|
# now we state-resolve between our own idea of the auth events, and the remote's
|
|
|
|
# idea of them.
|
|
|
|
|
|
|
|
local_state = auth_events.values()
|
|
|
|
remote_auth_events = dict(auth_events)
|
|
|
|
remote_auth_events.update({(d.type, d.state_key): d for d in different_events})
|
|
|
|
remote_state = remote_auth_events.values()
|
|
|
|
|
2020-02-03 08:53:30 -07:00
|
|
|
room_version = await self.store.get_room_version_id(event.room_id)
|
|
|
|
new_state = await self.state_handler.resolve_events(
|
2019-12-05 07:14:45 -07:00
|
|
|
room_version, (local_state, remote_state), event
|
2019-12-04 10:27:32 -07:00
|
|
|
)
|
2015-02-13 07:20:05 -07:00
|
|
|
|
2019-12-04 10:27:32 -07:00
|
|
|
logger.info(
|
|
|
|
"After state res: updating auth_events with new state %s",
|
|
|
|
{
|
|
|
|
(d.type, d.state_key): d.event_id
|
|
|
|
for d in new_state.values()
|
|
|
|
if auth_events.get((d.type, d.state_key)) != d
|
|
|
|
},
|
|
|
|
)
|
2015-02-13 07:20:05 -07:00
|
|
|
|
2019-12-04 10:27:32 -07:00
|
|
|
auth_events.update(new_state)
|
2015-02-13 07:20:05 -07:00
|
|
|
|
2020-02-03 08:53:30 -07:00
|
|
|
context = await self._update_context_for_auth_events(
|
2019-12-04 10:27:32 -07:00
|
|
|
event, context, auth_events
|
|
|
|
)
|
2015-02-06 08:16:26 -07:00
|
|
|
|
2019-11-01 10:19:09 -06:00
|
|
|
return context
|
|
|
|
|
2020-02-03 08:55:35 -07:00
|
|
|
async def _update_context_for_auth_events(
|
|
|
|
self, event: EventBase, context: EventContext, auth_events: StateMap[EventBase]
|
|
|
|
) -> EventContext:
|
2018-02-06 07:31:24 -07:00
|
|
|
"""Update the state_ids in an event context after auth event resolution,
|
|
|
|
storing the changes as a new state group.
|
2017-11-07 09:43:00 -07:00
|
|
|
|
|
|
|
Args:
|
2020-02-03 08:55:35 -07:00
|
|
|
event: The event we're handling the context for
|
2018-02-06 07:31:24 -07:00
|
|
|
|
2020-02-03 08:55:35 -07:00
|
|
|
context: initial event context
|
2017-11-07 09:43:00 -07:00
|
|
|
|
2020-02-03 08:55:35 -07:00
|
|
|
auth_events: Events to update in the event context.
|
2017-11-07 09:43:00 -07:00
|
|
|
|
2019-11-01 10:19:09 -06:00
|
|
|
Returns:
|
2020-02-03 08:55:35 -07:00
|
|
|
new event context
|
2017-11-07 09:43:00 -07:00
|
|
|
"""
|
2019-10-11 09:05:21 -06:00
|
|
|
# exclude the state key of the new event from the current_state in the context.
|
|
|
|
if event.is_state():
|
2020-04-24 12:36:38 -06:00
|
|
|
event_key = (event.type, event.state_key) # type: Optional[Tuple[str, str]]
|
2019-10-11 09:05:21 -06:00
|
|
|
else:
|
|
|
|
event_key = None
|
2017-11-07 09:43:00 -07:00
|
|
|
state_updates = {
|
2020-06-15 05:03:36 -06:00
|
|
|
k: a.event_id for k, a in auth_events.items() if k != event_key
|
2017-11-07 09:43:00 -07:00
|
|
|
}
|
2019-10-11 09:05:21 -06:00
|
|
|
|
2020-02-03 08:55:35 -07:00
|
|
|
current_state_ids = await context.get_current_state_ids()
|
2020-07-27 11:40:22 -06:00
|
|
|
current_state_ids = dict(current_state_ids) # type: ignore
|
2018-07-23 06:02:09 -06:00
|
|
|
|
|
|
|
current_state_ids.update(state_updates)
|
|
|
|
|
2020-02-03 08:55:35 -07:00
|
|
|
prev_state_ids = await context.get_prev_state_ids()
|
2018-07-23 06:02:09 -06:00
|
|
|
prev_state_ids = dict(prev_state_ids)
|
|
|
|
|
2020-06-15 05:03:36 -06:00
|
|
|
prev_state_ids.update({k: a.event_id for k, a in auth_events.items()})
|
2018-07-23 06:02:09 -06:00
|
|
|
|
2018-07-23 15:06:50 -06:00
|
|
|
# create a new state group as a delta from the existing one.
|
|
|
|
prev_group = context.state_group
|
2020-02-03 08:55:35 -07:00
|
|
|
state_group = await self.state_store.store_state_group(
|
2018-02-06 07:31:24 -07:00
|
|
|
event.event_id,
|
|
|
|
event.room_id,
|
2018-07-23 15:06:50 -06:00
|
|
|
prev_group=prev_group,
|
|
|
|
delta_ids=state_updates,
|
2018-07-23 06:02:09 -06:00
|
|
|
current_state_ids=current_state_ids,
|
|
|
|
)
|
|
|
|
|
2019-11-01 10:19:09 -06:00
|
|
|
return EventContext.with_state(
|
2018-07-23 06:02:09 -06:00
|
|
|
state_group=state_group,
|
2019-11-06 03:01:39 -07:00
|
|
|
state_group_before_event=context.state_group_before_event,
|
2018-07-23 06:02:09 -06:00
|
|
|
current_state_ids=current_state_ids,
|
|
|
|
prev_state_ids=prev_state_ids,
|
2018-07-23 15:06:50 -06:00
|
|
|
prev_group=prev_group,
|
|
|
|
delta_ids=state_updates,
|
2018-02-06 07:31:24 -07:00
|
|
|
)
|
2017-11-07 09:43:00 -07:00
|
|
|
|
2020-02-03 09:00:46 -07:00
|
|
|
async def construct_auth_difference(
|
|
|
|
self, local_auth: Iterable[EventBase], remote_auth: Iterable[EventBase]
|
|
|
|
) -> Dict:
|
2015-01-28 09:16:53 -07:00
|
|
|
""" Given a local and remote auth chain, find the differences. This
|
|
|
|
assumes that we have already processed all events in remote_auth
|
|
|
|
|
|
|
|
Params:
|
|
|
|
local_auth (list)
|
|
|
|
remote_auth (list)
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
dict
|
|
|
|
"""
|
|
|
|
|
2015-01-29 09:50:23 -07:00
|
|
|
logger.debug("construct_auth_difference Start!")
|
|
|
|
|
2015-01-28 09:16:53 -07:00
|
|
|
# TODO: Make sure we are OK with local_auth or remote_auth having more
|
|
|
|
# auth events in them than strictly necessary.
|
|
|
|
|
|
|
|
def sort_fun(ev):
|
|
|
|
return ev.depth, ev.event_id
|
|
|
|
|
2015-01-29 09:50:23 -07:00
|
|
|
logger.debug("construct_auth_difference after sort_fun!")
|
|
|
|
|
2015-01-28 09:16:53 -07:00
|
|
|
# We find the differences by starting at the "bottom" of each list
|
|
|
|
# and iterating up on both lists. The lists are ordered by depth and
|
|
|
|
# then event_id, we iterate up both lists until we find the event ids
|
|
|
|
# don't match. Then we look at depth/event_id to see which side is
|
|
|
|
# missing that event, and iterate only up that list. Repeat.
|
|
|
|
|
|
|
|
remote_list = list(remote_auth)
|
|
|
|
remote_list.sort(key=sort_fun)
|
|
|
|
|
|
|
|
local_list = list(local_auth)
|
|
|
|
local_list.sort(key=sort_fun)
|
|
|
|
|
|
|
|
local_iter = iter(local_list)
|
|
|
|
remote_iter = iter(remote_list)
|
|
|
|
|
2015-01-29 09:50:23 -07:00
|
|
|
logger.debug("construct_auth_difference before get_next!")
|
2015-01-28 09:16:53 -07:00
|
|
|
|
|
|
|
def get_next(it, opt=None):
|
2015-01-29 09:50:23 -07:00
|
|
|
try:
|
2018-05-05 14:47:18 -06:00
|
|
|
return next(it)
|
2017-10-23 08:52:32 -06:00
|
|
|
except Exception:
|
2015-01-29 09:50:23 -07:00
|
|
|
return opt
|
|
|
|
|
|
|
|
current_local = get_next(local_iter)
|
|
|
|
current_remote = get_next(remote_iter)
|
|
|
|
|
|
|
|
logger.debug("construct_auth_difference before while")
|
2015-01-28 09:16:53 -07:00
|
|
|
|
|
|
|
missing_remotes = []
|
|
|
|
missing_locals = []
|
2015-01-30 03:48:47 -07:00
|
|
|
while current_local or current_remote:
|
2015-01-28 09:16:53 -07:00
|
|
|
if current_remote is None:
|
|
|
|
missing_locals.append(current_local)
|
|
|
|
current_local = get_next(local_iter)
|
|
|
|
continue
|
|
|
|
|
|
|
|
if current_local is None:
|
|
|
|
missing_remotes.append(current_remote)
|
|
|
|
current_remote = get_next(remote_iter)
|
|
|
|
continue
|
|
|
|
|
|
|
|
if current_local.event_id == current_remote.event_id:
|
|
|
|
current_local = get_next(local_iter)
|
|
|
|
current_remote = get_next(remote_iter)
|
|
|
|
continue
|
|
|
|
|
|
|
|
if current_local.depth < current_remote.depth:
|
|
|
|
missing_locals.append(current_local)
|
|
|
|
current_local = get_next(local_iter)
|
|
|
|
continue
|
|
|
|
|
|
|
|
if current_local.depth > current_remote.depth:
|
|
|
|
missing_remotes.append(current_remote)
|
|
|
|
current_remote = get_next(remote_iter)
|
|
|
|
continue
|
|
|
|
|
|
|
|
# They have the same depth, so we fall back to the event_id order
|
|
|
|
if current_local.event_id < current_remote.event_id:
|
|
|
|
missing_locals.append(current_local)
|
|
|
|
current_local = get_next(local_iter)
|
|
|
|
|
|
|
|
if current_local.event_id > current_remote.event_id:
|
|
|
|
missing_remotes.append(current_remote)
|
|
|
|
current_remote = get_next(remote_iter)
|
|
|
|
continue
|
|
|
|
|
2015-01-29 09:50:23 -07:00
|
|
|
logger.debug("construct_auth_difference after while")
|
|
|
|
|
2015-01-28 09:16:53 -07:00
|
|
|
# missing locals should be sent to the server
|
|
|
|
# We should find why we are missing remotes, as they will have been
|
|
|
|
# rejected.
|
|
|
|
|
|
|
|
# Remove events from missing_remotes if they are referencing a missing
|
|
|
|
# remote. We only care about the "root" rejected ones.
|
|
|
|
missing_remote_ids = [e.event_id for e in missing_remotes]
|
|
|
|
base_remote_rejected = list(missing_remotes)
|
|
|
|
for e in missing_remotes:
|
2018-11-05 06:35:15 -07:00
|
|
|
for e_id in e.auth_event_ids():
|
2015-01-28 09:16:53 -07:00
|
|
|
if e_id in missing_remote_ids:
|
2015-02-06 03:53:18 -07:00
|
|
|
try:
|
|
|
|
base_remote_rejected.remove(e)
|
|
|
|
except ValueError:
|
|
|
|
pass
|
2015-01-28 09:16:53 -07:00
|
|
|
|
|
|
|
reason_map = {}
|
|
|
|
|
|
|
|
for e in base_remote_rejected:
|
2020-02-03 09:00:46 -07:00
|
|
|
reason = await self.store.get_rejection_reason(e.event_id)
|
2015-01-28 09:16:53 -07:00
|
|
|
if reason is None:
|
2015-02-06 07:16:50 -07:00
|
|
|
# TODO: e is not in the current state, so we should
|
|
|
|
# construct some proof of that.
|
|
|
|
continue
|
2015-01-28 09:16:53 -07:00
|
|
|
|
|
|
|
reason_map[e.event_id] = reason
|
|
|
|
|
2015-01-29 09:50:23 -07:00
|
|
|
logger.debug("construct_auth_difference returning")
|
|
|
|
|
2019-07-23 07:00:55 -06:00
|
|
|
return {
|
|
|
|
"auth_chain": local_auth,
|
|
|
|
"rejects": {
|
|
|
|
e.event_id: {"reason": reason_map[e.event_id], "proof": None}
|
|
|
|
for e in base_remote_rejected
|
|
|
|
},
|
|
|
|
"missing": [e.event_id for e in missing_locals],
|
|
|
|
}
|
2015-11-05 09:43:19 -07:00
|
|
|
|
|
|
|
@log_function
|
2020-05-01 08:15:36 -06:00
|
|
|
async def exchange_third_party_invite(
|
2019-06-20 03:32:02 -06:00
|
|
|
self, sender_user_id, target_user_id, room_id, signed
|
2016-02-23 08:11:25 -07:00
|
|
|
):
|
2019-06-20 03:32:02 -06:00
|
|
|
third_party_invite = {"signed": signed}
|
2015-12-17 10:09:51 -07:00
|
|
|
|
2015-11-05 09:43:19 -07:00
|
|
|
event_dict = {
|
|
|
|
"type": EventTypes.Member,
|
|
|
|
"content": {
|
|
|
|
"membership": Membership.INVITE,
|
2015-12-17 10:09:51 -07:00
|
|
|
"third_party_invite": third_party_invite,
|
2015-11-05 09:43:19 -07:00
|
|
|
},
|
|
|
|
"room_id": room_id,
|
2016-02-23 08:11:25 -07:00
|
|
|
"sender": sender_user_id,
|
|
|
|
"state_key": target_user_id,
|
2015-11-05 09:43:19 -07:00
|
|
|
}
|
|
|
|
|
2020-05-01 08:15:36 -06:00
|
|
|
if await self.auth.check_host_in_room(room_id, self.hs.hostname):
|
|
|
|
room_version = await self.store.get_room_version_id(room_id)
|
2019-01-23 13:21:33 -07:00
|
|
|
builder = self.event_builder_factory.new(room_version, event_dict)
|
|
|
|
|
2019-01-28 10:00:14 -07:00
|
|
|
EventValidator().validate_builder(builder)
|
2020-05-01 08:15:36 -06:00
|
|
|
event, context = await self.event_creation_handler.create_new_client_event(
|
2016-05-11 02:09:20 -06:00
|
|
|
builder=builder
|
|
|
|
)
|
2015-12-17 10:31:20 -07:00
|
|
|
|
2020-05-01 08:15:36 -06:00
|
|
|
event, context = await self.add_display_name_to_third_party_invite(
|
2019-01-23 13:21:33 -07:00
|
|
|
room_version, event_dict, event, context
|
2015-12-17 10:31:20 -07:00
|
|
|
)
|
|
|
|
|
2019-11-04 10:09:22 -07:00
|
|
|
EventValidator().validate_new(event, self.config)
|
2019-01-28 10:00:14 -07:00
|
|
|
|
2019-01-29 09:15:16 -07:00
|
|
|
# We need to tell the transaction queue to send this out, even
|
|
|
|
# though the sender isn't a local user.
|
|
|
|
event.internal_metadata.send_on_behalf_of = self.hs.hostname
|
|
|
|
|
2016-04-13 04:11:46 -06:00
|
|
|
try:
|
2020-05-01 08:15:36 -06:00
|
|
|
await self.auth.check_from_context(room_version, event, context)
|
2016-04-13 04:11:46 -06:00
|
|
|
except AuthError as e:
|
2019-10-31 04:23:24 -06:00
|
|
|
logger.warning("Denying new third party invite %r because %s", event, e)
|
2016-04-13 04:11:46 -06:00
|
|
|
raise e
|
|
|
|
|
2020-05-01 08:15:36 -06:00
|
|
|
await self._check_signature(event, context)
|
2019-11-01 04:28:09 -06:00
|
|
|
|
|
|
|
# We retrieve the room member handler here as to not cause a cyclic dependency
|
2018-03-01 03:54:37 -07:00
|
|
|
member_handler = self.hs.get_room_member_handler()
|
2020-05-01 08:15:36 -06:00
|
|
|
await member_handler.send_membership_event(None, event, context)
|
2015-11-05 09:43:19 -07:00
|
|
|
else:
|
2020-02-21 05:15:07 -07:00
|
|
|
destinations = {x.split(":", 1)[-1] for x in (sender_user_id, room_id)}
|
2020-05-01 08:15:36 -06:00
|
|
|
await self.federation_client.forward_third_party_invite(
|
2019-06-20 03:32:02 -06:00
|
|
|
destinations, room_id, event_dict
|
2015-11-05 09:43:19 -07:00
|
|
|
)
|
|
|
|
|
2020-02-03 09:02:50 -07:00
|
|
|
async def on_exchange_third_party_invite_request(
|
2020-11-19 03:05:33 -07:00
|
|
|
self, event_dict: JsonDict
|
2020-02-03 09:02:50 -07:00
|
|
|
) -> None:
|
2017-09-19 05:18:01 -06:00
|
|
|
"""Handle an exchange_third_party_invite request from a remote server
|
|
|
|
|
|
|
|
The remote server will call this when it wants to turn a 3pid invite
|
|
|
|
into a normal m.room.member invite.
|
|
|
|
|
2019-09-11 03:37:17 -06:00
|
|
|
Args:
|
2020-11-19 03:05:33 -07:00
|
|
|
event_dict: Dictionary containing the event body.
|
2019-09-11 03:37:17 -06:00
|
|
|
|
2017-09-19 05:18:01 -06:00
|
|
|
"""
|
2020-11-19 03:05:33 -07:00
|
|
|
assert_params_in_dict(event_dict, ["room_id"])
|
|
|
|
room_version = await self.store.get_room_version_id(event_dict["room_id"])
|
2019-01-23 13:21:33 -07:00
|
|
|
|
|
|
|
# NB: event_dict has a particular specced format we might need to fudge
|
|
|
|
# if we change event formats too much.
|
|
|
|
builder = self.event_builder_factory.new(room_version, event_dict)
|
2015-11-05 09:43:19 -07:00
|
|
|
|
2020-02-03 09:02:50 -07:00
|
|
|
event, context = await self.event_creation_handler.create_new_client_event(
|
2019-06-20 03:32:02 -06:00
|
|
|
builder=builder
|
2015-11-05 09:43:19 -07:00
|
|
|
)
|
2020-02-03 09:02:50 -07:00
|
|
|
event, context = await self.add_display_name_to_third_party_invite(
|
2019-01-23 13:21:33 -07:00
|
|
|
room_version, event_dict, event, context
|
2015-12-17 10:31:20 -07:00
|
|
|
)
|
|
|
|
|
2016-04-13 04:11:46 -06:00
|
|
|
try:
|
2020-02-03 09:02:50 -07:00
|
|
|
await self.auth.check_from_context(room_version, event, context)
|
2016-04-13 04:11:46 -06:00
|
|
|
except AuthError as e:
|
2019-10-31 04:23:24 -06:00
|
|
|
logger.warning("Denying third party invite %r because %s", event, e)
|
2016-04-13 04:11:46 -06:00
|
|
|
raise e
|
2020-02-03 09:02:50 -07:00
|
|
|
await self._check_signature(event, context)
|
2015-11-05 09:43:19 -07:00
|
|
|
|
2019-01-29 09:15:16 -07:00
|
|
|
# We need to tell the transaction queue to send this out, even
|
|
|
|
# though the sender isn't a local user.
|
|
|
|
event.internal_metadata.send_on_behalf_of = get_domain_from_id(event.sender)
|
|
|
|
|
2019-11-01 04:28:09 -06:00
|
|
|
# We retrieve the room member handler here as to not cause a cyclic dependency
|
2018-03-01 03:54:37 -07:00
|
|
|
member_handler = self.hs.get_room_member_handler()
|
2020-02-03 09:02:50 -07:00
|
|
|
await member_handler.send_membership_event(None, event, context)
|
2015-11-05 09:43:19 -07:00
|
|
|
|
2020-05-11 13:12:46 -06:00
|
|
|
async def add_display_name_to_third_party_invite(
|
2019-06-20 03:32:02 -06:00
|
|
|
self, room_version, event_dict, event, context
|
|
|
|
):
|
2015-12-17 10:31:20 -07:00
|
|
|
key = (
|
|
|
|
EventTypes.ThirdPartyInvite,
|
2019-06-20 03:32:02 -06:00
|
|
|
event.content["third_party_invite"]["signed"]["token"],
|
2015-12-17 10:31:20 -07:00
|
|
|
)
|
2016-08-25 10:32:22 -06:00
|
|
|
original_invite = None
|
2020-05-11 13:12:46 -06:00
|
|
|
prev_state_ids = await context.get_prev_state_ids()
|
2018-07-23 06:00:22 -06:00
|
|
|
original_invite_id = prev_state_ids.get(key)
|
2016-08-25 10:32:22 -06:00
|
|
|
if original_invite_id:
|
2020-05-11 13:12:46 -06:00
|
|
|
original_invite = await self.store.get_event(
|
2016-08-25 10:32:22 -06:00
|
|
|
original_invite_id, allow_none=True
|
|
|
|
)
|
2016-09-22 03:56:53 -06:00
|
|
|
if original_invite:
|
2019-10-02 04:16:38 -06:00
|
|
|
# If the m.room.third_party_invite event's content is empty, it means the
|
2019-10-04 04:16:19 -06:00
|
|
|
# invite has been revoked. In this case, we don't have to raise an error here
|
|
|
|
# because the auth check will fail on the invite (because it's not able to
|
|
|
|
# fetch public keys from the m.room.third_party_invite event's content, which
|
2019-10-04 04:21:24 -06:00
|
|
|
# is empty).
|
2019-10-04 04:16:19 -06:00
|
|
|
display_name = original_invite.content.get("display_name")
|
2019-10-04 04:18:28 -06:00
|
|
|
event_dict["content"]["third_party_invite"]["display_name"] = display_name
|
2016-09-22 03:56:53 -06:00
|
|
|
else:
|
2015-12-17 10:31:20 -07:00
|
|
|
logger.info(
|
2019-06-20 03:32:02 -06:00
|
|
|
"Could not find invite event for third_party_invite: %r", event_dict
|
2015-12-17 10:31:20 -07:00
|
|
|
)
|
2016-09-22 04:59:46 -06:00
|
|
|
# We don't discard here as this is not the appropriate place to do
|
|
|
|
# auth checks. If we need the invite and don't have it then the
|
|
|
|
# auth check code will explode appropriately.
|
2015-12-17 10:31:20 -07:00
|
|
|
|
2019-01-23 13:21:33 -07:00
|
|
|
builder = self.event_builder_factory.new(room_version, event_dict)
|
2019-01-28 10:00:14 -07:00
|
|
|
EventValidator().validate_builder(builder)
|
2020-05-11 13:12:46 -06:00
|
|
|
event, context = await self.event_creation_handler.create_new_client_event(
|
2019-06-20 03:32:02 -06:00
|
|
|
builder=builder
|
2018-01-15 09:52:07 -07:00
|
|
|
)
|
2019-11-19 07:07:39 -07:00
|
|
|
EventValidator().validate_new(event, self.config)
|
2019-07-23 07:00:55 -06:00
|
|
|
return (event, context)
|
2015-12-17 10:31:20 -07:00
|
|
|
|
2020-05-11 13:12:46 -06:00
|
|
|
async def _check_signature(self, event, context):
|
2016-02-23 08:11:25 -07:00
|
|
|
"""
|
|
|
|
Checks that the signature in the event is consistent with its invite.
|
|
|
|
|
2016-04-01 09:08:59 -06:00
|
|
|
Args:
|
|
|
|
event (Event): The m.room.member event to check
|
2016-08-25 10:32:22 -06:00
|
|
|
context (EventContext):
|
2016-04-01 09:08:59 -06:00
|
|
|
|
|
|
|
Raises:
|
|
|
|
AuthError: if signature didn't match any keys, or key has been
|
2016-02-23 08:11:25 -07:00
|
|
|
revoked,
|
2016-04-01 09:08:59 -06:00
|
|
|
SynapseError: if a transient error meant a key couldn't be checked
|
2016-02-23 08:11:25 -07:00
|
|
|
for revocation.
|
|
|
|
"""
|
|
|
|
signed = event.content["third_party_invite"]["signed"]
|
|
|
|
token = signed["token"]
|
2015-11-05 09:43:19 -07:00
|
|
|
|
2020-05-11 13:12:46 -06:00
|
|
|
prev_state_ids = await context.get_prev_state_ids()
|
2019-06-20 03:32:02 -06:00
|
|
|
invite_event_id = prev_state_ids.get((EventTypes.ThirdPartyInvite, token))
|
2015-11-05 09:43:19 -07:00
|
|
|
|
2016-08-25 10:32:22 -06:00
|
|
|
invite_event = None
|
|
|
|
if invite_event_id:
|
2020-05-11 13:12:46 -06:00
|
|
|
invite_event = await self.store.get_event(invite_event_id, allow_none=True)
|
2016-08-25 10:32:22 -06:00
|
|
|
|
2016-02-23 08:11:25 -07:00
|
|
|
if not invite_event:
|
|
|
|
raise AuthError(403, "Could not find invite")
|
|
|
|
|
2019-06-18 15:51:24 -06:00
|
|
|
logger.debug("Checking auth on event %r", event.content)
|
|
|
|
|
2020-07-01 09:21:02 -06:00
|
|
|
last_exception = None # type: Optional[Exception]
|
|
|
|
|
2019-06-18 15:51:24 -06:00
|
|
|
# for each public key in the 3pid invite event
|
2016-02-23 08:11:25 -07:00
|
|
|
for public_key_object in self.hs.get_auth().get_public_keys(invite_event):
|
|
|
|
try:
|
2019-06-18 15:51:24 -06:00
|
|
|
# for each sig on the third_party_invite block of the actual invite
|
2016-02-23 08:11:25 -07:00
|
|
|
for server, signature_block in signed["signatures"].items():
|
|
|
|
for key_name, encoded_signature in signature_block.items():
|
|
|
|
if not key_name.startswith("ed25519:"):
|
|
|
|
continue
|
|
|
|
|
2019-06-18 15:51:24 -06:00
|
|
|
logger.debug(
|
|
|
|
"Attempting to verify sig with key %s from %r "
|
|
|
|
"against pubkey %r",
|
2019-06-20 03:32:02 -06:00
|
|
|
key_name,
|
|
|
|
server,
|
|
|
|
public_key_object,
|
2016-02-23 08:11:25 -07:00
|
|
|
)
|
2019-06-18 15:51:24 -06:00
|
|
|
|
|
|
|
try:
|
|
|
|
public_key = public_key_object["public_key"]
|
|
|
|
verify_key = decode_verify_key_bytes(
|
2019-06-20 03:32:02 -06:00
|
|
|
key_name, decode_base64(public_key)
|
2019-06-18 15:51:24 -06:00
|
|
|
)
|
|
|
|
verify_signed_json(signed, server, verify_key)
|
|
|
|
logger.debug(
|
|
|
|
"Successfully verified sig with key %s from %r "
|
|
|
|
"against pubkey %r",
|
2019-06-20 03:32:02 -06:00
|
|
|
key_name,
|
|
|
|
server,
|
|
|
|
public_key_object,
|
2019-06-18 15:51:24 -06:00
|
|
|
)
|
|
|
|
except Exception:
|
|
|
|
logger.info(
|
|
|
|
"Failed to verify sig with key %s from %r "
|
|
|
|
"against pubkey %r",
|
2019-06-20 03:32:02 -06:00
|
|
|
key_name,
|
|
|
|
server,
|
|
|
|
public_key_object,
|
2019-06-18 15:51:24 -06:00
|
|
|
)
|
|
|
|
raise
|
|
|
|
try:
|
|
|
|
if "key_validity_url" in public_key_object:
|
2020-05-11 13:12:46 -06:00
|
|
|
await self._check_key_revocation(
|
2019-06-20 03:32:02 -06:00
|
|
|
public_key, public_key_object["key_validity_url"]
|
2019-06-18 15:51:24 -06:00
|
|
|
)
|
|
|
|
except Exception:
|
|
|
|
logger.info(
|
|
|
|
"Failed to query key_validity_url %s",
|
2019-06-20 03:32:02 -06:00
|
|
|
public_key_object["key_validity_url"],
|
2016-02-23 08:11:25 -07:00
|
|
|
)
|
2019-06-18 15:51:24 -06:00
|
|
|
raise
|
2016-02-23 08:11:25 -07:00
|
|
|
return
|
|
|
|
except Exception as e:
|
|
|
|
last_exception = e
|
2020-07-01 09:21:02 -06:00
|
|
|
|
|
|
|
if last_exception is None:
|
|
|
|
# we can only get here if get_public_keys() returned an empty list
|
|
|
|
# TODO: make this better
|
|
|
|
raise RuntimeError("no public key in invite event")
|
|
|
|
|
2016-02-23 08:11:25 -07:00
|
|
|
raise last_exception
|
|
|
|
|
2020-05-11 13:12:46 -06:00
|
|
|
async def _check_key_revocation(self, public_key, url):
|
2016-02-23 08:11:25 -07:00
|
|
|
"""
|
|
|
|
Checks whether public_key has been revoked.
|
|
|
|
|
2016-04-01 09:08:59 -06:00
|
|
|
Args:
|
|
|
|
public_key (str): base-64 encoded public key.
|
|
|
|
url (str): Key revocation URL.
|
2016-02-23 08:11:25 -07:00
|
|
|
|
2016-04-01 09:08:59 -06:00
|
|
|
Raises:
|
|
|
|
AuthError: if they key has been revoked.
|
|
|
|
SynapseError: if a transient error meant a key couldn't be checked
|
2016-02-23 08:11:25 -07:00
|
|
|
for revocation.
|
|
|
|
"""
|
2015-11-05 09:43:19 -07:00
|
|
|
try:
|
2020-05-11 13:12:46 -06:00
|
|
|
response = await self.http_client.get_json(url, {"public_key": public_key})
|
2015-11-05 09:43:19 -07:00
|
|
|
except Exception:
|
2019-06-20 03:32:02 -06:00
|
|
|
raise SynapseError(502, "Third party certificate could not be checked")
|
2015-11-05 09:43:19 -07:00
|
|
|
if "valid" not in response or not response["valid"]:
|
|
|
|
raise AuthError(403, "Third party certificate was invalid")
|
2018-07-25 09:00:38 -06:00
|
|
|
|
2020-02-03 09:24:07 -07:00
|
|
|
async def persist_events_and_notify(
|
|
|
|
self,
|
2020-09-14 03:16:41 -06:00
|
|
|
room_id: str,
|
2020-02-03 09:24:07 -07:00
|
|
|
event_and_contexts: Sequence[Tuple[EventBase, EventContext]],
|
|
|
|
backfilled: bool = False,
|
2020-05-22 07:21:54 -06:00
|
|
|
) -> int:
|
2018-07-25 09:00:38 -06:00
|
|
|
"""Persists events and tells the notifier/pushers about them, if
|
|
|
|
necessary.
|
|
|
|
|
|
|
|
Args:
|
2020-09-14 03:16:41 -06:00
|
|
|
room_id: The room ID of events being persisted.
|
|
|
|
event_and_contexts: Sequence of events with their associated
|
|
|
|
context that should be persisted. All events must belong to
|
|
|
|
the same room.
|
2020-02-03 09:24:07 -07:00
|
|
|
backfilled: Whether these events are a result of
|
2018-07-25 09:00:38 -06:00
|
|
|
backfilling or not
|
|
|
|
"""
|
2020-09-14 03:16:41 -06:00
|
|
|
instance = self.config.worker.events_shard_config.get_instance(room_id)
|
|
|
|
if instance != self._instance_name:
|
2020-05-22 09:11:35 -06:00
|
|
|
result = await self._send_events(
|
2020-09-14 03:16:41 -06:00
|
|
|
instance_name=instance,
|
2018-07-25 09:32:05 -06:00
|
|
|
store=self.store,
|
2020-09-14 03:16:41 -06:00
|
|
|
room_id=room_id,
|
2018-07-25 09:32:05 -06:00
|
|
|
event_and_contexts=event_and_contexts,
|
2019-06-20 03:32:02 -06:00
|
|
|
backfilled=backfilled,
|
2018-07-25 09:32:05 -06:00
|
|
|
)
|
2020-05-22 07:21:54 -06:00
|
|
|
return result["max_stream_id"]
|
2018-07-25 09:32:05 -06:00
|
|
|
else:
|
2020-09-29 08:57:36 -06:00
|
|
|
assert self.storage.persistence
|
2020-10-13 05:07:56 -06:00
|
|
|
|
|
|
|
# Note that this returns the events that were persisted, which may not be
|
|
|
|
# the same as were passed in if some were deduplicated due to transaction IDs.
|
|
|
|
events, max_stream_token = await self.storage.persistence.persist_events(
|
2019-06-20 03:32:02 -06:00
|
|
|
event_and_contexts, backfilled=backfilled
|
2018-07-25 09:32:05 -06:00
|
|
|
)
|
2018-07-25 09:00:38 -06:00
|
|
|
|
2019-12-03 12:19:45 -07:00
|
|
|
if self._ephemeral_messages_enabled:
|
2020-10-13 05:07:56 -06:00
|
|
|
for event in events:
|
2019-12-03 12:19:45 -07:00
|
|
|
# If there's an expiry timestamp on the event, schedule its expiry.
|
|
|
|
self._message_handler.maybe_schedule_expiry(event)
|
|
|
|
|
2018-07-25 09:32:05 -06:00
|
|
|
if not backfilled: # Never notify for backfilled events
|
2020-10-13 05:07:56 -06:00
|
|
|
for event in events:
|
2020-09-24 06:24:17 -06:00
|
|
|
await self._notify_persisted_event(event, max_stream_token)
|
2018-07-25 09:00:38 -06:00
|
|
|
|
2020-09-24 06:24:17 -06:00
|
|
|
return max_stream_token.stream
|
2020-05-22 07:21:54 -06:00
|
|
|
|
2020-02-03 09:25:41 -07:00
|
|
|
async def _notify_persisted_event(
|
2020-09-24 06:24:17 -06:00
|
|
|
self, event: EventBase, max_stream_token: RoomStreamToken
|
2020-02-03 09:25:41 -07:00
|
|
|
) -> None:
|
2018-07-25 09:00:38 -06:00
|
|
|
"""Checks to see if notifier/pushers should be notified about the
|
|
|
|
event or not.
|
|
|
|
|
|
|
|
Args:
|
2020-02-03 09:25:41 -07:00
|
|
|
event:
|
|
|
|
max_stream_id: The max_stream_id returned by persist_events
|
2018-07-25 09:00:38 -06:00
|
|
|
"""
|
|
|
|
|
|
|
|
extra_users = []
|
|
|
|
if event.type == EventTypes.Member:
|
|
|
|
target_user_id = event.state_key
|
|
|
|
|
|
|
|
# We notify for memberships if its an invite for one of our
|
|
|
|
# users
|
|
|
|
if event.internal_metadata.is_outlier():
|
|
|
|
if event.membership != Membership.INVITE:
|
|
|
|
if not self.is_mine_id(target_user_id):
|
|
|
|
return
|
|
|
|
|
|
|
|
target_user = UserID.from_string(target_user_id)
|
|
|
|
extra_users.append(target_user)
|
|
|
|
elif event.internal_metadata.is_outlier():
|
|
|
|
return
|
|
|
|
|
2020-10-05 07:43:14 -06:00
|
|
|
# the event has been persisted so it should have a stream ordering.
|
|
|
|
assert event.internal_metadata.stream_ordering
|
|
|
|
|
2020-09-24 06:24:17 -06:00
|
|
|
event_pos = PersistedEventPosition(
|
|
|
|
self._instance_name, event.internal_metadata.stream_ordering
|
|
|
|
)
|
2018-07-25 09:00:38 -06:00
|
|
|
self.notifier.on_new_room_event(
|
2020-09-24 06:24:17 -06:00
|
|
|
event, event_pos, max_stream_token, extra_users=extra_users
|
2018-07-25 09:00:38 -06:00
|
|
|
)
|
|
|
|
|
2020-02-03 09:27:05 -07:00
|
|
|
async def _clean_room_for_join(self, room_id: str) -> None:
|
2018-08-09 03:29:48 -06:00
|
|
|
"""Called to clean up any data in DB for a given room, ready for the
|
|
|
|
server to join the room.
|
|
|
|
|
|
|
|
Args:
|
2020-02-03 09:27:05 -07:00
|
|
|
room_id
|
2018-08-09 03:29:48 -06:00
|
|
|
"""
|
|
|
|
if self.config.worker_app:
|
2020-02-03 09:27:05 -07:00
|
|
|
await self._clean_room_for_join_client(room_id)
|
2018-08-09 03:29:48 -06:00
|
|
|
else:
|
2020-02-03 09:27:05 -07:00
|
|
|
await self.store.clean_room_for_join(room_id)
|
2018-07-25 09:00:38 -06:00
|
|
|
|
2020-07-24 08:53:25 -06:00
|
|
|
async def get_room_complexity(
|
|
|
|
self, remote_room_hosts: List[str], room_id: str
|
|
|
|
) -> Optional[dict]:
|
2019-07-29 10:47:27 -06:00
|
|
|
"""
|
|
|
|
Fetch the complexity of a remote room over federation.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
remote_room_hosts (list[str]): The remote servers to ask.
|
|
|
|
room_id (str): The room ID to ask about.
|
|
|
|
|
|
|
|
Returns:
|
2020-07-24 08:53:25 -06:00
|
|
|
Dict contains the complexity
|
2019-07-29 10:47:27 -06:00
|
|
|
metric versions, while None means we could not fetch the complexity.
|
|
|
|
"""
|
|
|
|
|
|
|
|
for host in remote_room_hosts:
|
2020-05-11 13:12:46 -06:00
|
|
|
res = await self.federation_client.get_room_complexity(host, room_id)
|
2019-07-29 10:47:27 -06:00
|
|
|
|
|
|
|
# We got a result, return it.
|
|
|
|
if res:
|
2020-05-11 13:12:46 -06:00
|
|
|
return res
|
2019-07-29 10:47:27 -06:00
|
|
|
|
|
|
|
# We fell off the bottom, couldn't get the complexity from anyone. Oh
|
|
|
|
# well.
|
2020-05-11 13:12:46 -06:00
|
|
|
return None
|