2016-11-04 09:35:25 -06:00
|
|
|
# -*- coding: utf-8 -*-
|
|
|
|
# Copyright 2014-2016 OpenMarket Ltd
|
|
|
|
#
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
|
|
|
|
2016-11-21 09:55:23 -07:00
|
|
|
"""A federation sender that forwards things to be sent across replication to
|
|
|
|
a worker process.
|
|
|
|
|
|
|
|
It assumes there is a single worker process feeding off of it.
|
|
|
|
|
|
|
|
Each row in the replication stream consists of a type and some json, where the
|
|
|
|
types indicate whether they are presence, or edus, etc.
|
|
|
|
|
|
|
|
Ephemeral or non-event data are queued up in-memory. When the worker requests
|
|
|
|
updates since a particular point, all in-memory data since before that point is
|
|
|
|
dropped. We also expire things in the queue after 5 minutes, to ensure that a
|
|
|
|
dead worker doesn't cause the queues to grow limitlessly.
|
|
|
|
|
|
|
|
Events are replicated via a separate events stream.
|
|
|
|
"""
|
|
|
|
|
2018-07-09 00:09:20 -06:00
|
|
|
import logging
|
|
|
|
from collections import namedtuple
|
2020-05-05 07:27:13 -06:00
|
|
|
from typing import Dict, List, Tuple, Type
|
2016-11-04 09:35:25 -06:00
|
|
|
|
2018-06-05 08:13:57 -06:00
|
|
|
from sortedcontainers import SortedDict
|
2016-11-16 10:34:44 -07:00
|
|
|
|
2019-10-29 09:08:22 -06:00
|
|
|
from twisted.internet import defer
|
|
|
|
|
2020-08-17 05:24:46 -06:00
|
|
|
from synapse.api.presence import UserPresenceState
|
2018-07-09 00:09:20 -06:00
|
|
|
from synapse.metrics import LaterGauge
|
|
|
|
from synapse.util.metrics import Measure
|
2017-04-07 04:51:28 -06:00
|
|
|
|
2018-07-09 00:09:20 -06:00
|
|
|
from .units import Edu
|
2018-04-15 08:51:07 -06:00
|
|
|
|
2017-04-07 04:51:28 -06:00
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
2016-11-16 10:34:44 -07:00
|
|
|
|
2020-09-04 04:54:56 -06:00
|
|
|
class FederationRemoteSendQueue:
|
2019-03-13 14:02:56 -06:00
|
|
|
"""A drop in replacement for FederationSender"""
|
2016-11-16 10:34:44 -07:00
|
|
|
|
2016-11-04 09:35:25 -06:00
|
|
|
def __init__(self, hs):
|
2016-11-17 08:46:44 -07:00
|
|
|
self.server_name = hs.hostname
|
2016-11-04 09:35:25 -06:00
|
|
|
self.clock = hs.get_clock()
|
2017-03-27 07:07:47 -06:00
|
|
|
self.notifier = hs.get_notifier()
|
2017-04-10 09:48:30 -06:00
|
|
|
self.is_mine_id = hs.is_mine_id
|
2016-11-04 09:35:25 -06:00
|
|
|
|
2020-07-10 11:26:36 -06:00
|
|
|
# We may have multiple federation sender instances, so we need to track
|
|
|
|
# their positions separately.
|
2020-07-29 16:22:13 -06:00
|
|
|
self._sender_instances = hs.config.worker.federation_shard_config.instances
|
2020-07-10 11:26:36 -06:00
|
|
|
self._sender_positions = {}
|
|
|
|
|
2020-05-05 07:27:13 -06:00
|
|
|
# Pending presence map user_id -> UserPresenceState
|
|
|
|
self.presence_map = {} # type: Dict[str, UserPresenceState]
|
|
|
|
|
|
|
|
# Stream position -> list[user_id]
|
|
|
|
self.presence_changed = SortedDict() # type: SortedDict[int, List[str]]
|
2019-03-26 06:45:22 -06:00
|
|
|
|
|
|
|
# Stores the destinations we need to explicitly send presence to about a
|
|
|
|
# given user.
|
|
|
|
# Stream position -> (user_id, destinations)
|
2020-05-05 07:27:13 -06:00
|
|
|
self.presence_destinations = (
|
|
|
|
SortedDict()
|
|
|
|
) # type: SortedDict[int, Tuple[str, List[str]]]
|
|
|
|
|
|
|
|
# (destination, key) -> EDU
|
|
|
|
self.keyed_edu = {} # type: Dict[Tuple[str, tuple], Edu]
|
2016-11-04 09:35:25 -06:00
|
|
|
|
2020-05-05 07:27:13 -06:00
|
|
|
# stream position -> (destination, key)
|
|
|
|
self.keyed_edu_changed = (
|
|
|
|
SortedDict()
|
|
|
|
) # type: SortedDict[int, Tuple[str, tuple]]
|
2016-11-04 09:35:25 -06:00
|
|
|
|
2020-05-05 07:27:13 -06:00
|
|
|
self.edus = SortedDict() # type: SortedDict[int, Edu]
|
2016-11-04 09:35:25 -06:00
|
|
|
|
2020-05-05 07:15:57 -06:00
|
|
|
# stream ID for the next entry into presence_changed/keyed_edu_changed/edus.
|
2016-11-04 09:35:25 -06:00
|
|
|
self.pos = 1
|
2020-05-05 07:15:57 -06:00
|
|
|
|
|
|
|
# map from stream ID to the time that stream entry was generated, so that we
|
|
|
|
# can clear out entries after a while
|
2020-05-05 07:27:13 -06:00
|
|
|
self.pos_time = SortedDict() # type: SortedDict[int, int]
|
2016-11-04 09:35:25 -06:00
|
|
|
|
2016-11-21 10:34:43 -07:00
|
|
|
# EVERYTHING IS SAD. In particular, python only makes new scopes when
|
|
|
|
# we make a new function, so we need to make a new function so the inner
|
2016-11-21 10:59:39 -07:00
|
|
|
# lambda binds to the queue rather than to the name of the queue which
|
|
|
|
# changes. ARGH.
|
2016-11-21 10:34:43 -07:00
|
|
|
def register(name, queue):
|
2019-06-20 03:32:02 -06:00
|
|
|
LaterGauge(
|
|
|
|
"synapse_federation_send_queue_%s_size" % (queue_name,),
|
|
|
|
"",
|
|
|
|
[],
|
|
|
|
lambda: len(queue),
|
|
|
|
)
|
2016-11-21 10:34:43 -07:00
|
|
|
|
|
|
|
for queue_name in [
|
2019-06-20 03:32:02 -06:00
|
|
|
"presence_map",
|
|
|
|
"presence_changed",
|
|
|
|
"keyed_edu",
|
|
|
|
"keyed_edu_changed",
|
|
|
|
"edus",
|
|
|
|
"pos_time",
|
|
|
|
"presence_destinations",
|
2016-11-21 10:34:43 -07:00
|
|
|
]:
|
|
|
|
register(queue_name, getattr(self, queue_name))
|
2016-11-17 08:46:44 -07:00
|
|
|
|
2016-11-04 09:35:25 -06:00
|
|
|
self.clock.looping_call(self._clear_queue, 30 * 1000)
|
|
|
|
|
|
|
|
def _next_pos(self):
|
|
|
|
pos = self.pos
|
|
|
|
self.pos += 1
|
|
|
|
self.pos_time[self.clock.time_msec()] = pos
|
|
|
|
return pos
|
|
|
|
|
|
|
|
def _clear_queue(self):
|
2016-11-21 09:55:23 -07:00
|
|
|
"""Clear the queues for anything older than N minutes"""
|
2016-11-04 09:35:25 -06:00
|
|
|
|
|
|
|
FIVE_MINUTES_AGO = 5 * 60 * 1000
|
|
|
|
now = self.clock.time_msec()
|
|
|
|
|
|
|
|
keys = self.pos_time.keys()
|
2018-06-05 17:17:52 -06:00
|
|
|
time = self.pos_time.bisect_left(now - FIVE_MINUTES_AGO)
|
2016-11-04 09:35:25 -06:00
|
|
|
if not keys[:time]:
|
|
|
|
return
|
|
|
|
|
|
|
|
position_to_delete = max(keys[:time])
|
|
|
|
for key in keys[:time]:
|
|
|
|
del self.pos_time[key]
|
|
|
|
|
|
|
|
self._clear_queue_before_pos(position_to_delete)
|
|
|
|
|
|
|
|
def _clear_queue_before_pos(self, position_to_delete):
|
2016-11-21 09:55:23 -07:00
|
|
|
"""Clear all the queues from before a given position"""
|
2016-11-21 10:34:43 -07:00
|
|
|
with Measure(self.clock, "send_queue._clear"):
|
|
|
|
# Delete things out of presence maps
|
|
|
|
keys = self.presence_changed.keys()
|
2018-06-05 17:17:52 -06:00
|
|
|
i = self.presence_changed.bisect_left(position_to_delete)
|
2016-11-21 10:34:43 -07:00
|
|
|
for key in keys[:i]:
|
|
|
|
del self.presence_changed[key]
|
|
|
|
|
2020-02-21 05:15:07 -07:00
|
|
|
user_ids = {
|
2019-06-20 03:32:02 -06:00
|
|
|
user_id for uids in self.presence_changed.values() for user_id in uids
|
2020-02-21 05:15:07 -07:00
|
|
|
}
|
2016-11-21 10:34:43 -07:00
|
|
|
|
2019-03-26 06:45:22 -06:00
|
|
|
keys = self.presence_destinations.keys()
|
|
|
|
i = self.presence_destinations.bisect_left(position_to_delete)
|
|
|
|
for key in keys[:i]:
|
|
|
|
del self.presence_destinations[key]
|
|
|
|
|
|
|
|
user_ids.update(
|
|
|
|
user_id for user_id, _ in self.presence_destinations.values()
|
|
|
|
)
|
|
|
|
|
2016-11-21 10:34:43 -07:00
|
|
|
to_del = [
|
|
|
|
user_id for user_id in self.presence_map if user_id not in user_ids
|
|
|
|
]
|
|
|
|
for user_id in to_del:
|
|
|
|
del self.presence_map[user_id]
|
|
|
|
|
|
|
|
# Delete things out of keyed edus
|
|
|
|
keys = self.keyed_edu_changed.keys()
|
2018-06-05 17:17:52 -06:00
|
|
|
i = self.keyed_edu_changed.bisect_left(position_to_delete)
|
2016-11-21 10:34:43 -07:00
|
|
|
for key in keys[:i]:
|
|
|
|
del self.keyed_edu_changed[key]
|
|
|
|
|
|
|
|
live_keys = set()
|
|
|
|
for edu_key in self.keyed_edu_changed.values():
|
|
|
|
live_keys.add(edu_key)
|
|
|
|
|
2020-05-05 07:27:13 -06:00
|
|
|
keys_to_del = [
|
|
|
|
edu_key for edu_key in self.keyed_edu if edu_key not in live_keys
|
|
|
|
]
|
|
|
|
for edu_key in keys_to_del:
|
2016-11-21 10:34:43 -07:00
|
|
|
del self.keyed_edu[edu_key]
|
|
|
|
|
|
|
|
# Delete things out of edu map
|
|
|
|
keys = self.edus.keys()
|
2018-06-05 17:17:52 -06:00
|
|
|
i = self.edus.bisect_left(position_to_delete)
|
2016-11-21 10:34:43 -07:00
|
|
|
for key in keys[:i]:
|
|
|
|
del self.edus[key]
|
|
|
|
|
2020-10-14 06:27:51 -06:00
|
|
|
def notify_new_events(self, max_token):
|
2019-03-13 14:02:56 -06:00
|
|
|
"""As per FederationSender"""
|
2016-11-21 09:55:23 -07:00
|
|
|
# We don't need to replicate this as it gets sent down a different
|
|
|
|
# stream.
|
2016-11-17 08:46:44 -07:00
|
|
|
pass
|
|
|
|
|
2019-03-04 05:57:44 -07:00
|
|
|
def build_and_send_edu(self, destination, edu_type, content, key=None):
|
2019-03-13 14:02:56 -06:00
|
|
|
"""As per FederationSender"""
|
2019-03-04 05:57:44 -07:00
|
|
|
if destination == self.server_name:
|
|
|
|
logger.info("Not sending EDU to ourselves")
|
|
|
|
return
|
|
|
|
|
2016-11-04 09:35:25 -06:00
|
|
|
pos = self._next_pos()
|
|
|
|
|
2016-11-16 10:34:44 -07:00
|
|
|
edu = Edu(
|
|
|
|
origin=self.server_name,
|
|
|
|
destination=destination,
|
|
|
|
edu_type=edu_type,
|
|
|
|
content=content,
|
|
|
|
)
|
|
|
|
|
2016-11-04 09:35:25 -06:00
|
|
|
if key:
|
2016-11-17 08:46:44 -07:00
|
|
|
assert isinstance(key, tuple)
|
2016-11-16 10:34:44 -07:00
|
|
|
self.keyed_edu[(destination, key)] = edu
|
|
|
|
self.keyed_edu_changed[pos] = (destination, key)
|
2016-11-04 09:35:25 -06:00
|
|
|
else:
|
|
|
|
self.edus[pos] = edu
|
|
|
|
|
2017-03-27 07:07:47 -06:00
|
|
|
self.notifier.on_new_replication_data()
|
|
|
|
|
2019-03-13 09:55:37 -06:00
|
|
|
def send_read_receipt(self, receipt):
|
2019-03-13 14:02:56 -06:00
|
|
|
"""As per FederationSender
|
2019-03-13 09:55:37 -06:00
|
|
|
|
|
|
|
Args:
|
|
|
|
receipt (synapse.types.ReadReceipt):
|
|
|
|
"""
|
|
|
|
# nothing to do here: the replication listener will handle it.
|
2019-10-29 09:08:22 -06:00
|
|
|
return defer.succeed(None)
|
2019-03-13 09:55:37 -06:00
|
|
|
|
2017-04-10 09:48:30 -06:00
|
|
|
def send_presence(self, states):
|
2019-03-13 14:02:56 -06:00
|
|
|
"""As per FederationSender
|
2017-04-11 08:34:55 -06:00
|
|
|
|
|
|
|
Args:
|
|
|
|
states (list(UserPresenceState))
|
|
|
|
"""
|
2016-11-04 09:35:25 -06:00
|
|
|
pos = self._next_pos()
|
|
|
|
|
2017-04-11 08:34:55 -06:00
|
|
|
# We only want to send presence for our own users, so lets always just
|
|
|
|
# filter here just in case.
|
2018-05-31 03:03:47 -06:00
|
|
|
local_states = list(filter(lambda s: self.is_mine_id(s.user_id), states))
|
2016-11-04 09:35:25 -06:00
|
|
|
|
2017-04-10 09:48:30 -06:00
|
|
|
self.presence_map.update({state.user_id: state for state in local_states})
|
|
|
|
self.presence_changed[pos] = [state.user_id for state in local_states]
|
2016-11-04 09:35:25 -06:00
|
|
|
|
2017-03-27 07:07:47 -06:00
|
|
|
self.notifier.on_new_replication_data()
|
|
|
|
|
2019-03-26 06:45:22 -06:00
|
|
|
def send_presence_to_destinations(self, states, destinations):
|
|
|
|
"""As per FederationSender
|
|
|
|
|
|
|
|
Args:
|
|
|
|
states (list[UserPresenceState])
|
|
|
|
destinations (list[str])
|
|
|
|
"""
|
|
|
|
for state in states:
|
|
|
|
pos = self._next_pos()
|
|
|
|
self.presence_map.update({state.user_id: state for state in states})
|
|
|
|
self.presence_destinations[pos] = (state.user_id, destinations)
|
|
|
|
|
|
|
|
self.notifier.on_new_replication_data()
|
|
|
|
|
2016-11-17 08:46:44 -07:00
|
|
|
def send_device_messages(self, destination):
|
2019-03-13 14:02:56 -06:00
|
|
|
"""As per FederationSender"""
|
2020-01-29 04:23:01 -07:00
|
|
|
# We don't need to replicate this as it gets sent down a different
|
|
|
|
# stream.
|
2016-11-04 09:35:25 -06:00
|
|
|
|
2016-11-16 10:34:44 -07:00
|
|
|
def get_current_token(self):
|
|
|
|
return self.pos - 1
|
|
|
|
|
2020-07-10 11:26:36 -06:00
|
|
|
def federation_ack(self, instance_name, token):
|
|
|
|
if self._sender_instances:
|
|
|
|
# If we have configured multiple federation sender instances we need
|
|
|
|
# to track their positions separately, and only clear the queue up
|
|
|
|
# to the token all instances have acked.
|
|
|
|
self._sender_positions[instance_name] = token
|
|
|
|
token = min(self._sender_positions.values())
|
|
|
|
|
2017-03-27 07:11:17 -06:00
|
|
|
self._clear_queue_before_pos(token)
|
|
|
|
|
2020-01-16 02:16:12 -07:00
|
|
|
async def get_replication_rows(
|
2020-05-05 07:15:57 -06:00
|
|
|
self, instance_name: str, from_token: int, to_token: int, target_row_count: int
|
|
|
|
) -> Tuple[List[Tuple[int, Tuple]], int, bool]:
|
2017-03-27 07:11:17 -06:00
|
|
|
"""Get rows to be sent over federation between the two tokens
|
|
|
|
|
2016-11-23 03:40:44 -07:00
|
|
|
Args:
|
2020-05-05 07:15:57 -06:00
|
|
|
instance_name: the name of the current process
|
|
|
|
from_token: the previous stream token: the starting point for fetching the
|
|
|
|
updates
|
|
|
|
to_token: the new stream token: the point to get updates up to
|
|
|
|
target_row_count: a target for the number of rows to be returned.
|
|
|
|
|
|
|
|
Returns: a triplet `(updates, new_last_token, limited)`, where:
|
|
|
|
* `updates` is a list of `(token, row)` entries.
|
|
|
|
* `new_last_token` is the new position in stream.
|
|
|
|
* `limited` is whether there are more updates to fetch.
|
2016-11-23 03:40:44 -07:00
|
|
|
"""
|
2020-05-05 07:15:57 -06:00
|
|
|
# TODO: Handle target_row_count.
|
2016-11-16 10:34:44 -07:00
|
|
|
|
|
|
|
# To handle restarts where we wrap around
|
2017-03-27 07:11:17 -06:00
|
|
|
if from_token > self.pos:
|
|
|
|
from_token = -1
|
2016-11-16 10:34:44 -07:00
|
|
|
|
2017-04-10 03:02:17 -06:00
|
|
|
# list of tuple(int, BaseFederationRow), where the first is the position
|
|
|
|
# of the federation stream.
|
2020-05-05 07:15:57 -06:00
|
|
|
rows = [] # type: List[Tuple[int, BaseFederationRow]]
|
2016-11-16 10:34:44 -07:00
|
|
|
|
2016-11-04 09:35:25 -06:00
|
|
|
# Fetch changed presence
|
2018-06-05 17:17:52 -06:00
|
|
|
i = self.presence_changed.bisect_right(from_token)
|
|
|
|
j = self.presence_changed.bisect_right(to_token) + 1
|
2017-04-10 09:48:30 -06:00
|
|
|
dest_user_ids = [
|
|
|
|
(pos, user_id)
|
2018-06-05 17:17:52 -06:00
|
|
|
for pos, user_id_list in self.presence_changed.items()[i:j]
|
|
|
|
for user_id in user_id_list
|
2017-04-10 09:48:30 -06:00
|
|
|
]
|
2016-11-16 10:34:44 -07:00
|
|
|
|
2017-04-10 09:48:30 -06:00
|
|
|
for (key, user_id) in dest_user_ids:
|
2019-06-20 03:32:02 -06:00
|
|
|
rows.append((key, PresenceRow(state=self.presence_map[user_id])))
|
2016-11-04 09:35:25 -06:00
|
|
|
|
2019-03-26 06:45:22 -06:00
|
|
|
# Fetch presence to send to destinations
|
|
|
|
i = self.presence_destinations.bisect_right(from_token)
|
|
|
|
j = self.presence_destinations.bisect_right(to_token) + 1
|
|
|
|
|
|
|
|
for pos, (user_id, dests) in self.presence_destinations.items()[i:j]:
|
2019-06-20 03:32:02 -06:00
|
|
|
rows.append(
|
|
|
|
(
|
|
|
|
pos,
|
|
|
|
PresenceDestinationsRow(
|
|
|
|
state=self.presence_map[user_id], destinations=list(dests)
|
|
|
|
),
|
|
|
|
)
|
|
|
|
)
|
2019-03-26 06:45:22 -06:00
|
|
|
|
2016-11-04 09:35:25 -06:00
|
|
|
# Fetch changes keyed edus
|
2018-06-05 17:17:52 -06:00
|
|
|
i = self.keyed_edu_changed.bisect_right(from_token)
|
|
|
|
j = self.keyed_edu_changed.bisect_right(to_token) + 1
|
2017-04-10 09:49:51 -06:00
|
|
|
# We purposefully clobber based on the key here, python dict comprehensions
|
|
|
|
# always use the last value, so this will correctly point to the last
|
|
|
|
# stream position.
|
2018-06-05 17:17:52 -06:00
|
|
|
keyed_edus = {v: k for k, v in self.keyed_edu_changed.items()[i:j]}
|
2016-11-04 09:35:25 -06:00
|
|
|
|
2020-06-15 05:03:36 -06:00
|
|
|
for ((destination, edu_key), pos) in keyed_edus.items():
|
2019-06-20 03:32:02 -06:00
|
|
|
rows.append(
|
|
|
|
(
|
|
|
|
pos,
|
|
|
|
KeyedEduRow(
|
|
|
|
key=edu_key, edu=self.keyed_edu[(destination, edu_key)]
|
|
|
|
),
|
|
|
|
)
|
|
|
|
)
|
2016-11-04 09:35:25 -06:00
|
|
|
|
|
|
|
# Fetch changed edus
|
2018-06-05 17:17:52 -06:00
|
|
|
i = self.edus.bisect_right(from_token)
|
|
|
|
j = self.edus.bisect_right(to_token) + 1
|
|
|
|
edus = self.edus.items()[i:j]
|
2016-11-04 09:35:25 -06:00
|
|
|
|
|
|
|
for (pos, edu) in edus:
|
2017-04-07 04:48:27 -06:00
|
|
|
rows.append((pos, EduRow(edu)))
|
2016-11-04 09:35:25 -06:00
|
|
|
|
|
|
|
# Sort rows based on pos
|
|
|
|
rows.sort()
|
|
|
|
|
2020-05-05 07:15:57 -06:00
|
|
|
return (
|
|
|
|
[(pos, (row.TypeId, row.to_data())) for pos, row in rows],
|
|
|
|
to_token,
|
|
|
|
False,
|
|
|
|
)
|
2017-04-07 04:48:27 -06:00
|
|
|
|
|
|
|
|
2020-09-04 04:54:56 -06:00
|
|
|
class BaseFederationRow:
|
2017-04-07 04:52:57 -06:00
|
|
|
"""Base class for rows to be sent in the federation stream.
|
|
|
|
|
|
|
|
Specifies how to identify, serialize and deserialize the different types.
|
|
|
|
"""
|
|
|
|
|
2020-07-09 07:52:58 -06:00
|
|
|
TypeId = "" # Unique string that ids the type. Must be overridden in sub classes.
|
2017-04-07 04:48:27 -06:00
|
|
|
|
|
|
|
@staticmethod
|
|
|
|
def from_data(data):
|
|
|
|
"""Parse the data from the federation stream into a row.
|
2017-04-07 04:52:57 -06:00
|
|
|
|
|
|
|
Args:
|
|
|
|
data: The value of ``data`` from FederationStreamRow.data, type
|
|
|
|
depends on the type of stream
|
2017-04-07 04:48:27 -06:00
|
|
|
"""
|
|
|
|
raise NotImplementedError()
|
|
|
|
|
|
|
|
def to_data(self):
|
2017-04-07 04:52:57 -06:00
|
|
|
"""Serialize this row to be sent over the federation stream.
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
The value to be sent in FederationStreamRow.data. The type depends
|
|
|
|
on the type of stream.
|
2017-04-07 04:48:27 -06:00
|
|
|
"""
|
|
|
|
raise NotImplementedError()
|
|
|
|
|
|
|
|
def add_to_buffer(self, buff):
|
|
|
|
"""Add this row to the appropriate field in the buffer ready for this
|
|
|
|
to be sent over federation.
|
|
|
|
|
|
|
|
We use a buffer so that we can batch up events that have come in at
|
|
|
|
the same time and send them all at once.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
buff (BufferedToSend)
|
|
|
|
"""
|
|
|
|
raise NotImplementedError()
|
|
|
|
|
|
|
|
|
2019-06-20 03:32:02 -06:00
|
|
|
class PresenceRow(
|
|
|
|
BaseFederationRow, namedtuple("PresenceRow", ("state",)) # UserPresenceState
|
|
|
|
):
|
2017-04-07 04:48:27 -06:00
|
|
|
TypeId = "p"
|
|
|
|
|
|
|
|
@staticmethod
|
|
|
|
def from_data(data):
|
2019-06-20 03:32:02 -06:00
|
|
|
return PresenceRow(state=UserPresenceState.from_dict(data))
|
2017-04-07 04:48:27 -06:00
|
|
|
|
|
|
|
def to_data(self):
|
2017-04-10 09:48:30 -06:00
|
|
|
return self.state.as_dict()
|
2017-04-07 04:48:27 -06:00
|
|
|
|
|
|
|
def add_to_buffer(self, buff):
|
2017-04-10 09:48:30 -06:00
|
|
|
buff.presence.append(self.state)
|
2017-04-07 04:48:27 -06:00
|
|
|
|
|
|
|
|
2019-06-20 03:32:02 -06:00
|
|
|
class PresenceDestinationsRow(
|
|
|
|
BaseFederationRow,
|
|
|
|
namedtuple(
|
|
|
|
"PresenceDestinationsRow",
|
|
|
|
("state", "destinations"), # UserPresenceState # list[str]
|
|
|
|
),
|
|
|
|
):
|
2019-03-26 06:45:22 -06:00
|
|
|
TypeId = "pd"
|
|
|
|
|
|
|
|
@staticmethod
|
|
|
|
def from_data(data):
|
|
|
|
return PresenceDestinationsRow(
|
2019-06-20 03:32:02 -06:00
|
|
|
state=UserPresenceState.from_dict(data["state"]), destinations=data["dests"]
|
2019-03-26 06:45:22 -06:00
|
|
|
)
|
|
|
|
|
|
|
|
def to_data(self):
|
2019-06-20 03:32:02 -06:00
|
|
|
return {"state": self.state.as_dict(), "dests": self.destinations}
|
2019-03-26 06:45:22 -06:00
|
|
|
|
|
|
|
def add_to_buffer(self, buff):
|
|
|
|
buff.presence_destinations.append((self.state, self.destinations))
|
|
|
|
|
|
|
|
|
2019-06-20 03:32:02 -06:00
|
|
|
class KeyedEduRow(
|
|
|
|
BaseFederationRow,
|
|
|
|
namedtuple(
|
|
|
|
"KeyedEduRow",
|
|
|
|
("key", "edu"), # tuple(str) - the edu key passed to send_edu # Edu
|
|
|
|
),
|
|
|
|
):
|
2017-04-12 03:36:29 -06:00
|
|
|
"""Streams EDUs that have an associated key that is ued to clobber. For example,
|
|
|
|
typing EDUs clobber based on room_id.
|
|
|
|
"""
|
|
|
|
|
2017-04-07 04:48:27 -06:00
|
|
|
TypeId = "k"
|
|
|
|
|
|
|
|
@staticmethod
|
|
|
|
def from_data(data):
|
2019-06-20 03:32:02 -06:00
|
|
|
return KeyedEduRow(key=tuple(data["key"]), edu=Edu(**data["edu"]))
|
2017-04-07 04:48:27 -06:00
|
|
|
|
|
|
|
def to_data(self):
|
2019-06-20 03:32:02 -06:00
|
|
|
return {"key": self.key, "edu": self.edu.get_internal_dict()}
|
2017-04-07 04:48:27 -06:00
|
|
|
|
|
|
|
def add_to_buffer(self, buff):
|
2019-06-20 03:32:02 -06:00
|
|
|
buff.keyed_edus.setdefault(self.edu.destination, {})[self.key] = self.edu
|
2017-04-07 04:48:27 -06:00
|
|
|
|
|
|
|
|
2019-06-20 03:32:02 -06:00
|
|
|
class EduRow(BaseFederationRow, namedtuple("EduRow", ("edu",))): # Edu
|
2017-04-12 03:36:29 -06:00
|
|
|
"""Streams EDUs that don't have keys. See KeyedEduRow
|
|
|
|
"""
|
2019-06-20 03:32:02 -06:00
|
|
|
|
2017-04-07 04:48:27 -06:00
|
|
|
TypeId = "e"
|
|
|
|
|
|
|
|
@staticmethod
|
|
|
|
def from_data(data):
|
|
|
|
return EduRow(Edu(**data))
|
|
|
|
|
|
|
|
def to_data(self):
|
|
|
|
return self.edu.get_internal_dict()
|
|
|
|
|
|
|
|
def add_to_buffer(self, buff):
|
|
|
|
buff.edus.setdefault(self.edu.destination, []).append(self.edu)
|
|
|
|
|
|
|
|
|
2020-05-05 07:27:13 -06:00
|
|
|
_rowtypes = (
|
|
|
|
PresenceRow,
|
|
|
|
PresenceDestinationsRow,
|
|
|
|
KeyedEduRow,
|
|
|
|
EduRow,
|
|
|
|
) # type: Tuple[Type[BaseFederationRow], ...]
|
|
|
|
|
|
|
|
TypeToRow = {Row.TypeId: Row for Row in _rowtypes}
|
2017-04-07 04:48:27 -06:00
|
|
|
|
|
|
|
|
2019-06-20 03:32:02 -06:00
|
|
|
ParsedFederationStreamData = namedtuple(
|
|
|
|
"ParsedFederationStreamData",
|
|
|
|
(
|
|
|
|
"presence", # list(UserPresenceState)
|
|
|
|
"presence_destinations", # list of tuples of UserPresenceState and destinations
|
|
|
|
"keyed_edus", # dict of destination -> { key -> Edu }
|
|
|
|
"edus", # dict of destination -> [Edu]
|
|
|
|
),
|
|
|
|
)
|
2017-04-07 04:48:27 -06:00
|
|
|
|
|
|
|
|
2017-04-10 05:44:17 -06:00
|
|
|
def process_rows_for_federation(transaction_queue, rows):
|
2017-04-10 03:02:17 -06:00
|
|
|
"""Parse a list of rows from the federation stream and put them in the
|
|
|
|
transaction queue ready for sending to the relevant homeservers.
|
2017-04-07 04:48:27 -06:00
|
|
|
|
|
|
|
Args:
|
2019-03-13 14:02:56 -06:00
|
|
|
transaction_queue (FederationSender)
|
2020-03-23 07:59:11 -06:00
|
|
|
rows (list(synapse.replication.tcp.streams.federation.FederationStream.FederationStreamRow))
|
2017-04-07 04:48:27 -06:00
|
|
|
"""
|
|
|
|
|
2017-04-10 03:02:17 -06:00
|
|
|
# The federation stream contains a bunch of different types of
|
2017-04-07 04:48:27 -06:00
|
|
|
# rows that need to be handled differently. We parse the rows, put
|
|
|
|
# them into the appropriate collection and then send them off.
|
|
|
|
|
2017-04-10 03:03:07 -06:00
|
|
|
buff = ParsedFederationStreamData(
|
2020-01-29 04:23:01 -07:00
|
|
|
presence=[], presence_destinations=[], keyed_edus={}, edus={},
|
2017-04-07 04:48:27 -06:00
|
|
|
)
|
|
|
|
|
|
|
|
# Parse the rows in the stream and add to the buffer
|
|
|
|
for row in rows:
|
2017-04-07 04:51:28 -06:00
|
|
|
if row.type not in TypeToRow:
|
|
|
|
logger.error("Unrecognized federation row type %r", row.type)
|
|
|
|
continue
|
|
|
|
|
2017-04-07 04:48:27 -06:00
|
|
|
RowType = TypeToRow[row.type]
|
|
|
|
parsed_row = RowType.from_data(row.data)
|
|
|
|
parsed_row.add_to_buffer(buff)
|
|
|
|
|
2017-04-10 09:48:30 -06:00
|
|
|
if buff.presence:
|
|
|
|
transaction_queue.send_presence(buff.presence)
|
2017-04-07 04:48:27 -06:00
|
|
|
|
2019-03-26 06:45:22 -06:00
|
|
|
for state, destinations in buff.presence_destinations:
|
|
|
|
transaction_queue.send_presence_to_destinations(
|
2019-06-20 03:32:02 -06:00
|
|
|
states=[state], destinations=destinations
|
2019-03-26 06:45:22 -06:00
|
|
|
)
|
|
|
|
|
2020-06-15 05:03:36 -06:00
|
|
|
for destination, edu_map in buff.keyed_edus.items():
|
2017-04-07 04:48:27 -06:00
|
|
|
for key, edu in edu_map.items():
|
2019-03-04 05:57:44 -07:00
|
|
|
transaction_queue.send_edu(edu, key)
|
2017-04-07 04:48:27 -06:00
|
|
|
|
2020-06-15 05:03:36 -06:00
|
|
|
for destination, edu_list in buff.edus.items():
|
2017-04-07 04:48:27 -06:00
|
|
|
for edu in edu_list:
|
2019-03-04 05:57:44 -07:00
|
|
|
transaction_queue.send_edu(edu, None)
|