2015-02-03 08:00:42 -07:00
|
|
|
# -*- coding: utf-8 -*-
|
2016-01-06 21:26:29 -07:00
|
|
|
# Copyright 2015, 2016 OpenMarket Ltd
|
2015-02-03 08:00:42 -07:00
|
|
|
#
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
2017-09-19 05:20:11 -06:00
|
|
|
import logging
|
2018-09-03 18:09:12 -06:00
|
|
|
from collections import namedtuple
|
2015-02-03 08:00:42 -07:00
|
|
|
|
2018-05-01 09:19:39 -06:00
|
|
|
import six
|
|
|
|
|
2018-07-09 00:09:20 -06:00
|
|
|
from twisted.internet import defer
|
2018-09-03 18:09:12 -06:00
|
|
|
from twisted.internet.defer import DeferredList
|
2018-07-09 00:09:20 -06:00
|
|
|
|
2018-09-03 18:09:12 -06:00
|
|
|
from synapse.api.constants import MAX_DEPTH, EventTypes, Membership
|
2018-07-09 00:09:20 -06:00
|
|
|
from synapse.api.errors import Codes, SynapseError
|
2017-09-19 05:20:11 -06:00
|
|
|
from synapse.crypto.event_signing import check_event_content_hash
|
2017-12-30 11:40:19 -07:00
|
|
|
from synapse.events import FrozenEvent
|
2017-09-19 05:20:11 -06:00
|
|
|
from synapse.events.utils import prune_event
|
2018-07-13 13:53:01 -06:00
|
|
|
from synapse.http.servlet import assert_params_in_dict
|
2018-09-03 18:09:12 -06:00
|
|
|
from synapse.types import get_domain_from_id
|
2018-07-09 00:09:20 -06:00
|
|
|
from synapse.util import logcontext, unwrapFirstError
|
2015-02-03 08:00:42 -07:00
|
|
|
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
|
|
|
|
|
|
|
class FederationBase(object):
|
2016-06-15 08:12:59 -06:00
|
|
|
def __init__(self, hs):
|
2018-03-12 08:07:39 -06:00
|
|
|
self.hs = hs
|
|
|
|
|
|
|
|
self.server_name = hs.hostname
|
|
|
|
self.keyring = hs.get_keyring()
|
2017-09-26 12:20:23 -06:00
|
|
|
self.spam_checker = hs.get_spam_checker()
|
2018-03-12 08:07:39 -06:00
|
|
|
self.store = hs.get_datastore()
|
|
|
|
self._clock = hs.get_clock()
|
2016-06-15 08:12:59 -06:00
|
|
|
|
2015-02-03 08:00:42 -07:00
|
|
|
@defer.inlineCallbacks
|
2015-06-26 02:52:24 -06:00
|
|
|
def _check_sigs_and_hash_and_fetch(self, origin, pdus, outlier=False,
|
|
|
|
include_none=False):
|
2015-02-03 08:00:42 -07:00
|
|
|
"""Takes a list of PDUs and checks the signatures and hashs of each
|
|
|
|
one. If a PDU fails its signature check then we check if we have it in
|
|
|
|
the database and if not then request if from the originating server of
|
|
|
|
that PDU.
|
|
|
|
|
|
|
|
If a PDU fails its content hash check then it is redacted.
|
|
|
|
|
|
|
|
The given list of PDUs are not modified, instead the function returns
|
|
|
|
a new list.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
pdu (list)
|
|
|
|
outlier (bool)
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
Deferred : A list of PDUs that have valid signatures and hashes.
|
|
|
|
"""
|
2015-06-26 02:52:24 -06:00
|
|
|
deferreds = self._check_sigs_and_hashes(pdus)
|
2015-02-12 11:35:36 -07:00
|
|
|
|
2017-09-19 18:32:42 -06:00
|
|
|
@defer.inlineCallbacks
|
|
|
|
def handle_check_result(pdu, deferred):
|
|
|
|
try:
|
|
|
|
res = yield logcontext.make_deferred_yieldable(deferred)
|
|
|
|
except SynapseError:
|
|
|
|
res = None
|
2015-02-12 11:35:36 -07:00
|
|
|
|
2015-06-26 02:52:24 -06:00
|
|
|
if not res:
|
2015-02-03 08:00:42 -07:00
|
|
|
# Check local db.
|
2017-09-19 18:32:42 -06:00
|
|
|
res = yield self.store.get_event(
|
2015-02-03 08:00:42 -07:00
|
|
|
pdu.event_id,
|
2015-02-12 11:17:11 -07:00
|
|
|
allow_rejected=True,
|
|
|
|
allow_none=True,
|
2015-02-03 08:00:42 -07:00
|
|
|
)
|
2015-06-26 02:52:24 -06:00
|
|
|
|
|
|
|
if not res and pdu.origin != origin:
|
2017-09-19 18:32:42 -06:00
|
|
|
try:
|
|
|
|
res = yield self.get_pdu(
|
|
|
|
destinations=[pdu.origin],
|
|
|
|
event_id=pdu.event_id,
|
|
|
|
outlier=outlier,
|
|
|
|
timeout=10000,
|
|
|
|
)
|
|
|
|
except SynapseError:
|
|
|
|
pass
|
|
|
|
|
2015-06-26 02:52:24 -06:00
|
|
|
if not res:
|
2015-02-12 12:29:43 -07:00
|
|
|
logger.warn(
|
|
|
|
"Failed to find copy of %s with valid signature",
|
|
|
|
pdu.event_id,
|
|
|
|
)
|
2015-06-26 02:52:24 -06:00
|
|
|
|
2017-09-19 18:32:42 -06:00
|
|
|
defer.returnValue(res)
|
2015-02-03 08:00:42 -07:00
|
|
|
|
2017-09-19 18:32:42 -06:00
|
|
|
handle = logcontext.preserve_fn(handle_check_result)
|
|
|
|
deferreds2 = [
|
|
|
|
handle(pdu, deferred)
|
|
|
|
for pdu, deferred in zip(pdus, deferreds)
|
|
|
|
]
|
|
|
|
|
|
|
|
valid_pdus = yield logcontext.make_deferred_yieldable(
|
|
|
|
defer.gatherResults(
|
|
|
|
deferreds2,
|
|
|
|
consumeErrors=True,
|
|
|
|
)
|
|
|
|
).addErrback(unwrapFirstError)
|
2015-02-12 11:35:36 -07:00
|
|
|
|
2015-06-26 02:52:24 -06:00
|
|
|
if include_none:
|
|
|
|
defer.returnValue(valid_pdus)
|
|
|
|
else:
|
|
|
|
defer.returnValue([p for p in valid_pdus if p])
|
2015-02-03 08:00:42 -07:00
|
|
|
|
|
|
|
def _check_sigs_and_hash(self, pdu):
|
2017-09-19 18:32:42 -06:00
|
|
|
return logcontext.make_deferred_yieldable(
|
|
|
|
self._check_sigs_and_hashes([pdu])[0],
|
|
|
|
)
|
2015-06-26 02:52:24 -06:00
|
|
|
|
|
|
|
def _check_sigs_and_hashes(self, pdus):
|
2017-09-19 05:20:11 -06:00
|
|
|
"""Checks that each of the received events is correctly signed by the
|
|
|
|
sending server.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
pdus (list[FrozenEvent]): the events to be checked
|
2015-02-03 08:00:42 -07:00
|
|
|
|
|
|
|
Returns:
|
2017-09-19 05:20:11 -06:00
|
|
|
list[Deferred]: for each input event, a deferred which:
|
|
|
|
* returns the original event if the checks pass
|
|
|
|
* returns a redacted version of the event (if the signature
|
|
|
|
matched but the hash did not)
|
|
|
|
* throws a SynapseError if the signature check failed.
|
2017-09-19 18:32:42 -06:00
|
|
|
The deferreds run their callbacks in the sentinel logcontext.
|
2015-02-03 08:00:42 -07:00
|
|
|
"""
|
2018-09-03 18:09:12 -06:00
|
|
|
deferreds = _check_sigs_on_pdus(self.keyring, pdus)
|
2015-06-26 02:52:24 -06:00
|
|
|
|
2017-09-19 18:32:42 -06:00
|
|
|
ctx = logcontext.LoggingContext.current_context()
|
2017-09-19 05:20:11 -06:00
|
|
|
|
2018-09-03 18:09:12 -06:00
|
|
|
def callback(_, pdu):
|
2017-09-19 18:32:42 -06:00
|
|
|
with logcontext.PreserveLoggingContext(ctx):
|
|
|
|
if not check_event_content_hash(pdu):
|
2018-09-06 16:56:47 -06:00
|
|
|
# let's try to distinguish between failures because the event was
|
|
|
|
# redacted (which are somewhat expected) vs actual ball-tampering
|
|
|
|
# incidents.
|
|
|
|
#
|
|
|
|
# This is just a heuristic, so we just assume that if the keys are
|
|
|
|
# about the same between the redacted and received events, then the
|
|
|
|
# received event was probably a redacted copy (but we then use our
|
|
|
|
# *actual* redacted copy to be on the safe side.)
|
|
|
|
redacted_event = prune_event(pdu)
|
|
|
|
if (
|
2018-09-13 08:33:16 -06:00
|
|
|
set(redacted_event.keys()) == set(pdu.keys()) and
|
2018-09-06 16:56:47 -06:00
|
|
|
set(six.iterkeys(redacted_event.content))
|
|
|
|
== set(six.iterkeys(pdu.content))
|
|
|
|
):
|
|
|
|
logger.info(
|
|
|
|
"Event %s seems to have been redacted; using our redacted "
|
|
|
|
"copy",
|
|
|
|
pdu.event_id,
|
|
|
|
)
|
|
|
|
else:
|
|
|
|
logger.warning(
|
|
|
|
"Event %s content has been tampered, redacting",
|
|
|
|
pdu.event_id, pdu.get_pdu_json(),
|
|
|
|
)
|
|
|
|
return redacted_event
|
2017-09-19 18:32:42 -06:00
|
|
|
|
2017-09-26 12:20:23 -06:00
|
|
|
if self.spam_checker.check_event_for_spam(pdu):
|
2017-09-19 18:32:42 -06:00
|
|
|
logger.warn(
|
|
|
|
"Event contains spam, redacting %s: %s",
|
|
|
|
pdu.event_id, pdu.get_pdu_json()
|
|
|
|
)
|
2018-09-03 18:09:12 -06:00
|
|
|
return prune_event(pdu)
|
2017-09-19 18:32:42 -06:00
|
|
|
|
|
|
|
return pdu
|
2015-06-26 02:52:24 -06:00
|
|
|
|
|
|
|
def errback(failure, pdu):
|
|
|
|
failure.trap(SynapseError)
|
2017-09-19 18:32:42 -06:00
|
|
|
with logcontext.PreserveLoggingContext(ctx):
|
|
|
|
logger.warn(
|
2018-09-06 16:56:47 -06:00
|
|
|
"Signature check failed for %s: %s",
|
|
|
|
pdu.event_id, failure.getErrorMessage(),
|
2017-09-19 18:32:42 -06:00
|
|
|
)
|
2015-06-26 02:52:24 -06:00
|
|
|
return failure
|
2015-02-03 08:00:42 -07:00
|
|
|
|
2018-09-03 18:09:12 -06:00
|
|
|
for deferred, pdu in zip(deferreds, pdus):
|
2015-06-26 02:52:24 -06:00
|
|
|
deferred.addCallbacks(
|
|
|
|
callback, errback,
|
2018-09-03 18:09:12 -06:00
|
|
|
callbackArgs=[pdu],
|
2015-06-26 02:52:24 -06:00
|
|
|
errbackArgs=[pdu],
|
2015-02-03 08:00:42 -07:00
|
|
|
)
|
|
|
|
|
2015-06-26 02:52:24 -06:00
|
|
|
return deferreds
|
2017-12-30 11:40:19 -07:00
|
|
|
|
|
|
|
|
2018-09-03 18:09:12 -06:00
|
|
|
class PduToCheckSig(namedtuple("PduToCheckSig", [
|
|
|
|
"pdu", "redacted_pdu_json", "event_id_domain", "sender_domain", "deferreds",
|
|
|
|
])):
|
|
|
|
pass
|
|
|
|
|
|
|
|
|
|
|
|
def _check_sigs_on_pdus(keyring, pdus):
|
|
|
|
"""Check that the given events are correctly signed
|
|
|
|
|
|
|
|
Args:
|
|
|
|
keyring (synapse.crypto.Keyring): keyring object to do the checks
|
|
|
|
pdus (Collection[EventBase]): the events to be checked
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
List[Deferred]: a Deferred for each event in pdus, which will either succeed if
|
|
|
|
the signatures are valid, or fail (with a SynapseError) if not.
|
|
|
|
"""
|
|
|
|
|
|
|
|
# (currently this is written assuming the v1 room structure; we'll probably want a
|
|
|
|
# separate function for checking v2 rooms)
|
|
|
|
|
|
|
|
# we want to check that the event is signed by:
|
|
|
|
#
|
|
|
|
# (a) the server which created the event_id
|
|
|
|
#
|
|
|
|
# (b) the sender's server.
|
|
|
|
#
|
|
|
|
# - except in the case of invites created from a 3pid invite, which are exempt
|
|
|
|
# from this check, because the sender has to match that of the original 3pid
|
|
|
|
# invite, but the event may come from a different HS, for reasons that I don't
|
|
|
|
# entirely grok (why do the senders have to match? and if they do, why doesn't the
|
|
|
|
# joining server ask the inviting server to do the switcheroo with
|
|
|
|
# exchange_third_party_invite?).
|
|
|
|
#
|
|
|
|
# That's pretty awful, since redacting such an invite will render it invalid
|
|
|
|
# (because it will then look like a regular invite without a valid signature),
|
|
|
|
# and signatures are *supposed* to be valid whether or not an event has been
|
|
|
|
# redacted. But this isn't the worst of the ways that 3pid invites are broken.
|
|
|
|
#
|
|
|
|
# let's start by getting the domain for each pdu, and flattening the event back
|
|
|
|
# to JSON.
|
|
|
|
pdus_to_check = [
|
|
|
|
PduToCheckSig(
|
|
|
|
pdu=p,
|
|
|
|
redacted_pdu_json=prune_event(p).get_pdu_json(),
|
|
|
|
event_id_domain=get_domain_from_id(p.event_id),
|
|
|
|
sender_domain=get_domain_from_id(p.sender),
|
|
|
|
deferreds=[],
|
|
|
|
)
|
|
|
|
for p in pdus
|
|
|
|
]
|
|
|
|
|
|
|
|
# first make sure that the event is signed by the event_id's domain
|
|
|
|
deferreds = keyring.verify_json_objects_for_server([
|
|
|
|
(p.event_id_domain, p.redacted_pdu_json)
|
|
|
|
for p in pdus_to_check
|
|
|
|
])
|
|
|
|
|
|
|
|
for p, d in zip(pdus_to_check, deferreds):
|
|
|
|
p.deferreds.append(d)
|
|
|
|
|
|
|
|
# now let's look for events where the sender's domain is different to the
|
|
|
|
# event id's domain (normally only the case for joins/leaves), and add additional
|
|
|
|
# checks.
|
|
|
|
pdus_to_check_sender = [
|
|
|
|
p for p in pdus_to_check
|
|
|
|
if p.sender_domain != p.event_id_domain and not _is_invite_via_3pid(p.pdu)
|
|
|
|
]
|
|
|
|
|
|
|
|
more_deferreds = keyring.verify_json_objects_for_server([
|
|
|
|
(p.sender_domain, p.redacted_pdu_json)
|
|
|
|
for p in pdus_to_check_sender
|
|
|
|
])
|
|
|
|
|
|
|
|
for p, d in zip(pdus_to_check_sender, more_deferreds):
|
|
|
|
p.deferreds.append(d)
|
|
|
|
|
|
|
|
# replace lists of deferreds with single Deferreds
|
|
|
|
return [_flatten_deferred_list(p.deferreds) for p in pdus_to_check]
|
|
|
|
|
|
|
|
|
|
|
|
def _flatten_deferred_list(deferreds):
|
|
|
|
"""Given a list of one or more deferreds, either return the single deferred, or
|
|
|
|
combine into a DeferredList.
|
|
|
|
"""
|
|
|
|
if len(deferreds) > 1:
|
|
|
|
return DeferredList(deferreds, fireOnOneErrback=True, consumeErrors=True)
|
|
|
|
else:
|
|
|
|
assert len(deferreds) == 1
|
|
|
|
return deferreds[0]
|
|
|
|
|
|
|
|
|
|
|
|
def _is_invite_via_3pid(event):
|
|
|
|
return (
|
|
|
|
event.type == EventTypes.Member
|
|
|
|
and event.membership == Membership.INVITE
|
|
|
|
and "third_party_invite" in event.content
|
|
|
|
)
|
|
|
|
|
|
|
|
|
2017-12-30 11:40:19 -07:00
|
|
|
def event_from_pdu_json(pdu_json, outlier=False):
|
|
|
|
"""Construct a FrozenEvent from an event json received over federation
|
|
|
|
|
|
|
|
Args:
|
|
|
|
pdu_json (object): pdu as received over federation
|
|
|
|
outlier (bool): True to mark this event as an outlier
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
FrozenEvent
|
2017-12-30 11:40:19 -07:00
|
|
|
|
|
|
|
Raises:
|
2018-05-01 09:19:39 -06:00
|
|
|
SynapseError: if the pdu is missing required fields or is otherwise
|
|
|
|
not a valid matrix event
|
2017-12-30 11:40:19 -07:00
|
|
|
"""
|
2017-12-30 11:40:19 -07:00
|
|
|
# we could probably enforce a bunch of other fields here (room_id, sender,
|
|
|
|
# origin, etc etc)
|
2018-07-13 13:53:01 -06:00
|
|
|
assert_params_in_dict(pdu_json, ('event_id', 'type', 'depth'))
|
2018-05-01 09:19:39 -06:00
|
|
|
|
|
|
|
depth = pdu_json['depth']
|
|
|
|
if not isinstance(depth, six.integer_types):
|
|
|
|
raise SynapseError(400, "Depth %r not an intger" % (depth, ),
|
|
|
|
Codes.BAD_JSON)
|
|
|
|
|
|
|
|
if depth < 0:
|
|
|
|
raise SynapseError(400, "Depth too small", Codes.BAD_JSON)
|
|
|
|
elif depth > MAX_DEPTH:
|
|
|
|
raise SynapseError(400, "Depth too large", Codes.BAD_JSON)
|
|
|
|
|
2017-12-30 11:40:19 -07:00
|
|
|
event = FrozenEvent(
|
|
|
|
pdu_json
|
|
|
|
)
|
|
|
|
|
|
|
|
event.internal_metadata.outlier = outlier
|
|
|
|
|
|
|
|
return event
|