2014-08-12 08:10:52 -06:00
|
|
|
# -*- coding: utf-8 -*-
|
2014-09-03 10:29:13 -06:00
|
|
|
# Copyright 2014 OpenMarket Ltd
|
2014-08-12 08:10:52 -06:00
|
|
|
#
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
2014-08-12 20:14:34 -06:00
|
|
|
|
2014-08-12 08:10:52 -06:00
|
|
|
"""This layer is responsible for replicating with remote home servers using
|
|
|
|
a given transport.
|
|
|
|
"""
|
|
|
|
|
|
|
|
from twisted.internet import defer
|
|
|
|
|
|
|
|
from .units import Transaction, Pdu, Edu
|
|
|
|
|
|
|
|
from .persistence import PduActions, TransactionActions
|
|
|
|
|
|
|
|
from synapse.util.logutils import log_function
|
|
|
|
|
|
|
|
import logging
|
|
|
|
|
|
|
|
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
|
|
|
|
|
|
|
class ReplicationLayer(object):
|
|
|
|
"""This layer is responsible for replicating with remote home servers over
|
|
|
|
the given transport. I.e., does the sending and receiving of PDUs to
|
|
|
|
remote home servers.
|
|
|
|
|
|
|
|
The layer communicates with the rest of the server via a registered
|
|
|
|
ReplicationHandler.
|
|
|
|
|
|
|
|
In more detail, the layer:
|
|
|
|
* Receives incoming data and processes it into transactions and pdus.
|
|
|
|
* Fetches any PDUs it thinks it might have missed.
|
|
|
|
* Keeps the current state for contexts up to date by applying the
|
|
|
|
suitable conflict resolution.
|
|
|
|
* Sends outgoing pdus wrapped in transactions.
|
|
|
|
* Fills out the references to previous pdus/transactions appropriately
|
|
|
|
for outgoing data.
|
|
|
|
"""
|
|
|
|
|
|
|
|
def __init__(self, hs, transport_layer):
|
|
|
|
self.server_name = hs.hostname
|
|
|
|
|
|
|
|
self.transport_layer = transport_layer
|
|
|
|
self.transport_layer.register_received_handler(self)
|
|
|
|
self.transport_layer.register_request_handler(self)
|
|
|
|
|
|
|
|
self.store = hs.get_datastore()
|
|
|
|
self.pdu_actions = PduActions(self.store)
|
|
|
|
self.transaction_actions = TransactionActions(self.store)
|
|
|
|
|
|
|
|
self._transaction_queue = _TransactionQueue(
|
|
|
|
hs, self.transaction_actions, transport_layer
|
|
|
|
)
|
|
|
|
|
|
|
|
self.handler = None
|
|
|
|
self.edu_handlers = {}
|
2014-08-13 09:55:53 -06:00
|
|
|
self.query_handlers = {}
|
2014-08-12 08:10:52 -06:00
|
|
|
|
|
|
|
self._order = 0
|
|
|
|
|
|
|
|
self._clock = hs.get_clock()
|
|
|
|
|
|
|
|
def set_handler(self, handler):
|
|
|
|
"""Sets the handler that the replication layer will use to communicate
|
|
|
|
receipt of new PDUs from other home servers. The required methods are
|
|
|
|
documented on :py:class:`.ReplicationHandler`.
|
|
|
|
"""
|
|
|
|
self.handler = handler
|
|
|
|
|
|
|
|
def register_edu_handler(self, edu_type, handler):
|
|
|
|
if edu_type in self.edu_handlers:
|
|
|
|
raise KeyError("Already have an EDU handler for %s" % (edu_type))
|
|
|
|
|
|
|
|
self.edu_handlers[edu_type] = handler
|
|
|
|
|
2014-08-13 09:55:53 -06:00
|
|
|
def register_query_handler(self, query_type, handler):
|
|
|
|
"""Sets the handler callable that will be used to handle an incoming
|
|
|
|
federation Query of the given type.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
query_type (str): Category name of the query, which should match
|
|
|
|
the string used by make_query.
|
|
|
|
handler (callable): Invoked to handle incoming queries of this type
|
|
|
|
|
|
|
|
handler is invoked as:
|
|
|
|
result = handler(args)
|
|
|
|
|
|
|
|
where 'args' is a dict mapping strings to strings of the query
|
|
|
|
arguments. It should return a Deferred that will eventually yield an
|
|
|
|
object to encode as JSON.
|
|
|
|
"""
|
|
|
|
if query_type in self.query_handlers:
|
|
|
|
raise KeyError("Already have a Query handler for %s" % (query_type))
|
|
|
|
|
|
|
|
self.query_handlers[query_type] = handler
|
|
|
|
|
2014-08-12 08:10:52 -06:00
|
|
|
@defer.inlineCallbacks
|
|
|
|
@log_function
|
|
|
|
def send_pdu(self, pdu):
|
|
|
|
"""Informs the replication layer about a new PDU generated within the
|
|
|
|
home server that should be transmitted to others.
|
|
|
|
|
|
|
|
This will fill out various attributes on the PDU object, e.g. the
|
|
|
|
`prev_pdus` key.
|
|
|
|
|
|
|
|
*Note:* The home server should always call `send_pdu` even if it knows
|
|
|
|
that it does not need to be replicated to other home servers. This is
|
|
|
|
in case e.g. someone else joins via a remote home server and then
|
2014-08-13 11:14:37 -06:00
|
|
|
backfills.
|
2014-08-12 08:10:52 -06:00
|
|
|
|
|
|
|
TODO: Figure out when we should actually resolve the deferred.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
pdu (Pdu): The new Pdu.
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
Deferred: Completes when we have successfully processed the PDU
|
|
|
|
and replicated it to any interested remote home servers.
|
|
|
|
"""
|
|
|
|
order = self._order
|
|
|
|
self._order += 1
|
|
|
|
|
|
|
|
logger.debug("[%s] Persisting PDU", pdu.pdu_id)
|
|
|
|
|
|
|
|
# Save *before* trying to send
|
2014-08-26 11:57:17 -06:00
|
|
|
yield self.store.persist_event(pdu=pdu)
|
2014-08-12 08:10:52 -06:00
|
|
|
|
|
|
|
logger.debug("[%s] Persisted PDU", pdu.pdu_id)
|
|
|
|
logger.debug("[%s] transaction_layer.enqueue_pdu... ", pdu.pdu_id)
|
|
|
|
|
|
|
|
# TODO, add errback, etc.
|
|
|
|
self._transaction_queue.enqueue_pdu(pdu, order)
|
|
|
|
|
|
|
|
logger.debug("[%s] transaction_layer.enqueue_pdu... done", pdu.pdu_id)
|
|
|
|
|
|
|
|
@log_function
|
|
|
|
def send_edu(self, destination, edu_type, content):
|
|
|
|
edu = Edu(
|
|
|
|
origin=self.server_name,
|
|
|
|
destination=destination,
|
|
|
|
edu_type=edu_type,
|
|
|
|
content=content,
|
|
|
|
)
|
|
|
|
|
|
|
|
# TODO, add errback, etc.
|
|
|
|
self._transaction_queue.enqueue_edu(edu)
|
2014-08-19 07:24:53 -06:00
|
|
|
return defer.succeed(None)
|
2014-08-12 08:10:52 -06:00
|
|
|
|
2014-08-13 09:55:53 -06:00
|
|
|
@log_function
|
2014-10-02 07:09:15 -06:00
|
|
|
def make_query(self, destination, query_type, args,
|
|
|
|
retry_on_dns_fail=True):
|
2014-08-13 09:55:53 -06:00
|
|
|
"""Sends a federation Query to a remote homeserver of the given type
|
|
|
|
and arguments.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
destination (str): Domain name of the remote homeserver
|
|
|
|
query_type (str): Category of the query type; should match the
|
|
|
|
handler name used in register_query_handler().
|
|
|
|
args (dict): Mapping of strings to strings containing the details
|
|
|
|
of the query request.
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
a Deferred which will eventually yield a JSON object from the
|
|
|
|
response
|
|
|
|
"""
|
2014-10-02 07:09:15 -06:00
|
|
|
return self.transport_layer.make_query(
|
|
|
|
destination, query_type, args, retry_on_dns_fail=retry_on_dns_fail
|
|
|
|
)
|
2014-08-13 09:55:53 -06:00
|
|
|
|
2014-08-12 08:10:52 -06:00
|
|
|
@defer.inlineCallbacks
|
|
|
|
@log_function
|
2014-08-13 11:14:37 -06:00
|
|
|
def backfill(self, dest, context, limit):
|
2014-08-12 08:10:52 -06:00
|
|
|
"""Requests some more historic PDUs for the given context from the
|
|
|
|
given destination server.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
dest (str): The remote home server to ask.
|
2014-08-13 11:14:37 -06:00
|
|
|
context (str): The context to backfill.
|
2014-08-12 08:10:52 -06:00
|
|
|
limit (int): The maximum number of PDUs to return.
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
Deferred: Results in the received PDUs.
|
|
|
|
"""
|
|
|
|
extremities = yield self.store.get_oldest_pdus_in_context(context)
|
|
|
|
|
2014-08-13 11:14:37 -06:00
|
|
|
logger.debug("backfill extrem=%s", extremities)
|
2014-08-12 08:10:52 -06:00
|
|
|
|
|
|
|
# If there are no extremeties then we've (probably) reached the start.
|
|
|
|
if not extremities:
|
|
|
|
return
|
|
|
|
|
2014-08-13 11:14:37 -06:00
|
|
|
transaction_data = yield self.transport_layer.backfill(
|
2014-08-12 08:10:52 -06:00
|
|
|
dest, context, extremities, limit)
|
|
|
|
|
2014-08-13 11:14:37 -06:00
|
|
|
logger.debug("backfill transaction_data=%s", repr(transaction_data))
|
2014-08-12 08:10:52 -06:00
|
|
|
|
|
|
|
transaction = Transaction(**transaction_data)
|
|
|
|
|
|
|
|
pdus = [Pdu(outlier=False, **p) for p in transaction.pdus]
|
|
|
|
for pdu in pdus:
|
2014-08-19 07:20:03 -06:00
|
|
|
yield self._handle_new_pdu(pdu, backfilled=True)
|
2014-08-12 08:10:52 -06:00
|
|
|
|
|
|
|
defer.returnValue(pdus)
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
@log_function
|
|
|
|
def get_pdu(self, destination, pdu_origin, pdu_id, outlier=False):
|
|
|
|
"""Requests the PDU with given origin and ID from the remote home
|
|
|
|
server.
|
|
|
|
|
|
|
|
This will persist the PDU locally upon receipt.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
destination (str): Which home server to query
|
|
|
|
pdu_origin (str): The home server that originally sent the pdu.
|
|
|
|
pdu_id (str)
|
|
|
|
outlier (bool): Indicates whether the PDU is an `outlier`, i.e. if
|
|
|
|
it's from an arbitary point in the context as opposed to part
|
|
|
|
of the current block of PDUs. Defaults to `False`
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
Deferred: Results in the requested PDU.
|
|
|
|
"""
|
|
|
|
|
|
|
|
transaction_data = yield self.transport_layer.get_pdu(
|
|
|
|
destination, pdu_origin, pdu_id)
|
|
|
|
|
|
|
|
transaction = Transaction(**transaction_data)
|
|
|
|
|
|
|
|
pdu_list = [Pdu(outlier=outlier, **p) for p in transaction.pdus]
|
|
|
|
|
|
|
|
pdu = None
|
|
|
|
if pdu_list:
|
|
|
|
pdu = pdu_list[0]
|
|
|
|
yield self._handle_new_pdu(pdu)
|
|
|
|
|
|
|
|
defer.returnValue(pdu)
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
@log_function
|
|
|
|
def get_state_for_context(self, destination, context):
|
|
|
|
"""Requests all of the `current` state PDUs for a given context from
|
|
|
|
a remote home server.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
destination (str): The remote homeserver to query for the state.
|
|
|
|
context (str): The context we're interested in.
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
Deferred: Results in a list of PDUs.
|
|
|
|
"""
|
|
|
|
|
|
|
|
transaction_data = yield self.transport_layer.get_context_state(
|
|
|
|
destination, context)
|
|
|
|
|
|
|
|
transaction = Transaction(**transaction_data)
|
|
|
|
|
|
|
|
pdus = [Pdu(outlier=True, **p) for p in transaction.pdus]
|
|
|
|
for pdu in pdus:
|
|
|
|
yield self._handle_new_pdu(pdu)
|
|
|
|
|
|
|
|
defer.returnValue(pdus)
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
@log_function
|
|
|
|
def on_context_pdus_request(self, context):
|
|
|
|
pdus = yield self.pdu_actions.get_all_pdus_from_context(
|
|
|
|
context
|
|
|
|
)
|
|
|
|
defer.returnValue((200, self._transaction_from_pdus(pdus).get_dict()))
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
@log_function
|
2014-08-13 11:14:37 -06:00
|
|
|
def on_backfill_request(self, context, versions, limit):
|
2014-08-12 08:10:52 -06:00
|
|
|
|
2014-08-13 11:14:37 -06:00
|
|
|
pdus = yield self.pdu_actions.backfill(context, versions, limit)
|
2014-08-12 08:10:52 -06:00
|
|
|
|
|
|
|
defer.returnValue((200, self._transaction_from_pdus(pdus).get_dict()))
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
@log_function
|
|
|
|
def on_incoming_transaction(self, transaction_data):
|
|
|
|
transaction = Transaction(**transaction_data)
|
|
|
|
|
2014-09-15 06:26:05 -06:00
|
|
|
for p in transaction.pdus:
|
2014-09-15 09:55:39 -06:00
|
|
|
if "age" in p:
|
|
|
|
p["age_ts"] = int(self._clock.time_msec()) - int(p["age"])
|
|
|
|
del p["age"]
|
2014-09-15 06:26:05 -06:00
|
|
|
|
|
|
|
pdu_list = [Pdu(**p) for p in transaction.pdus]
|
|
|
|
|
2014-08-12 08:10:52 -06:00
|
|
|
logger.debug("[%s] Got transaction", transaction.transaction_id)
|
|
|
|
|
|
|
|
response = yield self.transaction_actions.have_responded(transaction)
|
|
|
|
|
|
|
|
if response:
|
|
|
|
logger.debug("[%s] We've already responed to this request",
|
|
|
|
transaction.transaction_id)
|
|
|
|
defer.returnValue(response)
|
|
|
|
return
|
|
|
|
|
|
|
|
logger.debug("[%s] Transacition is new", transaction.transaction_id)
|
|
|
|
|
|
|
|
dl = []
|
|
|
|
for pdu in pdu_list:
|
|
|
|
dl.append(self._handle_new_pdu(pdu))
|
|
|
|
|
|
|
|
if hasattr(transaction, "edus"):
|
|
|
|
for edu in [Edu(**x) for x in transaction.edus]:
|
2014-10-14 09:44:27 -06:00
|
|
|
self.received_edu(transaction.origin, edu.edu_type, edu.content)
|
2014-08-12 08:10:52 -06:00
|
|
|
|
|
|
|
results = yield defer.DeferredList(dl)
|
|
|
|
|
|
|
|
ret = []
|
|
|
|
for r in results:
|
|
|
|
if r[0]:
|
|
|
|
ret.append({})
|
|
|
|
else:
|
|
|
|
logger.exception(r[1])
|
|
|
|
ret.append({"error": str(r[1])})
|
|
|
|
|
|
|
|
logger.debug("Returning: %s", str(ret))
|
|
|
|
|
|
|
|
yield self.transaction_actions.set_response(
|
|
|
|
transaction,
|
|
|
|
200, response
|
|
|
|
)
|
|
|
|
defer.returnValue((200, response))
|
|
|
|
|
|
|
|
def received_edu(self, origin, edu_type, content):
|
|
|
|
if edu_type in self.edu_handlers:
|
|
|
|
self.edu_handlers[edu_type](origin, content)
|
|
|
|
else:
|
|
|
|
logger.warn("Received EDU of type %s with no handler", edu_type)
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
@log_function
|
|
|
|
def on_context_state_request(self, context):
|
|
|
|
results = yield self.store.get_current_state_for_context(
|
|
|
|
context
|
|
|
|
)
|
|
|
|
|
|
|
|
logger.debug("Context returning %d results", len(results))
|
|
|
|
|
|
|
|
pdus = [Pdu.from_pdu_tuple(p) for p in results]
|
|
|
|
defer.returnValue((200, self._transaction_from_pdus(pdus).get_dict()))
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
@log_function
|
|
|
|
def on_pdu_request(self, pdu_origin, pdu_id):
|
|
|
|
pdu = yield self._get_persisted_pdu(pdu_id, pdu_origin)
|
|
|
|
|
|
|
|
if pdu:
|
|
|
|
defer.returnValue(
|
|
|
|
(200, self._transaction_from_pdus([pdu]).get_dict())
|
|
|
|
)
|
|
|
|
else:
|
|
|
|
defer.returnValue((404, ""))
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
@log_function
|
|
|
|
def on_pull_request(self, origin, versions):
|
|
|
|
transaction_id = max([int(v) for v in versions])
|
|
|
|
|
|
|
|
response = yield self.pdu_actions.after_transaction(
|
|
|
|
transaction_id,
|
|
|
|
origin,
|
|
|
|
self.server_name
|
|
|
|
)
|
|
|
|
|
|
|
|
if not response:
|
|
|
|
response = []
|
|
|
|
|
|
|
|
defer.returnValue(
|
|
|
|
(200, self._transaction_from_pdus(response).get_dict())
|
|
|
|
)
|
|
|
|
|
2014-08-13 09:55:53 -06:00
|
|
|
@defer.inlineCallbacks
|
|
|
|
def on_query_request(self, query_type, args):
|
|
|
|
if query_type in self.query_handlers:
|
|
|
|
response = yield self.query_handlers[query_type](args)
|
|
|
|
defer.returnValue((200, response))
|
|
|
|
else:
|
|
|
|
defer.returnValue((404, "No handler for Query type '%s'"
|
|
|
|
% (query_type)
|
|
|
|
))
|
|
|
|
|
2014-08-12 08:10:52 -06:00
|
|
|
@defer.inlineCallbacks
|
|
|
|
@log_function
|
|
|
|
def _get_persisted_pdu(self, pdu_id, pdu_origin):
|
|
|
|
""" Get a PDU from the database with given origin and id.
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
Deferred: Results in a `Pdu`.
|
|
|
|
"""
|
|
|
|
pdu_tuple = yield self.store.get_pdu(pdu_id, pdu_origin)
|
|
|
|
|
|
|
|
defer.returnValue(Pdu.from_pdu_tuple(pdu_tuple))
|
|
|
|
|
|
|
|
def _transaction_from_pdus(self, pdu_list):
|
|
|
|
"""Returns a new Transaction containing the given PDUs suitable for
|
|
|
|
transmission.
|
|
|
|
"""
|
2014-09-15 06:26:05 -06:00
|
|
|
pdus = [p.get_dict() for p in pdu_list]
|
|
|
|
for p in pdus:
|
|
|
|
if "age_ts" in pdus:
|
|
|
|
p["age"] = int(self.clock.time_msec()) - p["age_ts"]
|
|
|
|
|
2014-08-12 08:10:52 -06:00
|
|
|
return Transaction(
|
|
|
|
origin=self.server_name,
|
2014-09-15 06:26:05 -06:00
|
|
|
pdus=pdus,
|
2014-10-17 10:12:25 -06:00
|
|
|
origin_server_ts=int(self._clock.time_msec()),
|
2014-08-12 08:10:52 -06:00
|
|
|
destination=None,
|
|
|
|
)
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
@log_function
|
2014-08-19 07:20:03 -06:00
|
|
|
def _handle_new_pdu(self, pdu, backfilled=False):
|
2014-08-12 08:10:52 -06:00
|
|
|
# We reprocess pdus when we have seen them only as outliers
|
|
|
|
existing = yield self._get_persisted_pdu(pdu.pdu_id, pdu.origin)
|
|
|
|
|
|
|
|
if existing and (not existing.outlier or pdu.outlier):
|
|
|
|
logger.debug("Already seen pdu %s %s", pdu.pdu_id, pdu.origin)
|
|
|
|
defer.returnValue({})
|
|
|
|
return
|
|
|
|
|
|
|
|
# Get missing pdus if necessary.
|
|
|
|
is_new = yield self.pdu_actions.is_new(pdu)
|
|
|
|
if is_new and not pdu.outlier:
|
2014-08-13 11:14:37 -06:00
|
|
|
# We only backfill backwards to the min depth.
|
2014-08-12 08:10:52 -06:00
|
|
|
min_depth = yield self.store.get_min_depth_for_context(pdu.context)
|
|
|
|
|
|
|
|
if min_depth and pdu.depth > min_depth:
|
|
|
|
for pdu_id, origin in pdu.prev_pdus:
|
|
|
|
exists = yield self._get_persisted_pdu(pdu_id, origin)
|
|
|
|
|
|
|
|
if not exists:
|
|
|
|
logger.debug("Requesting pdu %s %s", pdu_id, origin)
|
|
|
|
|
|
|
|
try:
|
|
|
|
yield self.get_pdu(
|
|
|
|
pdu.origin,
|
|
|
|
pdu_id=pdu_id,
|
|
|
|
pdu_origin=origin
|
|
|
|
)
|
|
|
|
logger.debug("Processed pdu %s %s", pdu_id, origin)
|
|
|
|
except:
|
|
|
|
# TODO(erikj): Do some more intelligent retries.
|
|
|
|
logger.exception("Failed to get PDU")
|
|
|
|
|
|
|
|
# Persist the Pdu, but don't mark it as processed yet.
|
2014-08-26 11:57:17 -06:00
|
|
|
yield self.store.persist_event(pdu=pdu)
|
2014-08-12 08:10:52 -06:00
|
|
|
|
2014-08-19 07:20:03 -06:00
|
|
|
if not backfilled:
|
|
|
|
ret = yield self.handler.on_receive_pdu(pdu, backfilled=backfilled)
|
|
|
|
else:
|
|
|
|
ret = None
|
2014-08-12 08:10:52 -06:00
|
|
|
|
|
|
|
yield self.pdu_actions.mark_as_processed(pdu)
|
|
|
|
|
|
|
|
defer.returnValue(ret)
|
|
|
|
|
|
|
|
def __str__(self):
|
|
|
|
return "<ReplicationLayer(%s)>" % self.server_name
|
|
|
|
|
|
|
|
|
|
|
|
class ReplicationHandler(object):
|
|
|
|
"""This defines the methods that the :py:class:`.ReplicationLayer` will
|
|
|
|
use to communicate with the rest of the home server.
|
|
|
|
"""
|
|
|
|
def on_receive_pdu(self, pdu):
|
|
|
|
raise NotImplementedError("on_receive_pdu")
|
|
|
|
|
|
|
|
|
|
|
|
class _TransactionQueue(object):
|
|
|
|
"""This class makes sure we only have one transaction in flight at
|
|
|
|
a time for a given destination.
|
|
|
|
|
|
|
|
It batches pending PDUs into single transactions.
|
|
|
|
"""
|
|
|
|
|
|
|
|
def __init__(self, hs, transaction_actions, transport_layer):
|
|
|
|
self.server_name = hs.hostname
|
|
|
|
self.transaction_actions = transaction_actions
|
|
|
|
self.transport_layer = transport_layer
|
|
|
|
|
|
|
|
self._clock = hs.get_clock()
|
|
|
|
|
|
|
|
# Is a mapping from destinations -> deferreds. Used to keep track
|
|
|
|
# of which destinations have transactions in flight and when they are
|
|
|
|
# done
|
|
|
|
self.pending_transactions = {}
|
|
|
|
|
|
|
|
# Is a mapping from destination -> list of
|
|
|
|
# tuple(pending pdus, deferred, order)
|
|
|
|
self.pending_pdus_by_dest = {}
|
|
|
|
# destination -> list of tuple(edu, deferred)
|
|
|
|
self.pending_edus_by_dest = {}
|
|
|
|
|
|
|
|
# HACK to get unique tx id
|
|
|
|
self._next_txn_id = int(self._clock.time_msec())
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
@log_function
|
|
|
|
def enqueue_pdu(self, pdu, order):
|
|
|
|
# We loop through all destinations to see whether we already have
|
|
|
|
# a transaction in progress. If we do, stick it in the pending_pdus
|
|
|
|
# table and we'll get back to it later.
|
|
|
|
|
2014-08-24 07:35:13 -06:00
|
|
|
destinations = set([
|
2014-08-12 08:10:52 -06:00
|
|
|
d for d in pdu.destinations
|
|
|
|
if d != self.server_name
|
2014-08-24 07:35:13 -06:00
|
|
|
])
|
2014-08-12 08:10:52 -06:00
|
|
|
|
|
|
|
logger.debug("Sending to: %s", str(destinations))
|
|
|
|
|
|
|
|
if not destinations:
|
|
|
|
return
|
|
|
|
|
|
|
|
deferreds = []
|
|
|
|
|
|
|
|
for destination in destinations:
|
|
|
|
deferred = defer.Deferred()
|
|
|
|
self.pending_pdus_by_dest.setdefault(destination, []).append(
|
|
|
|
(pdu, deferred, order)
|
|
|
|
)
|
|
|
|
|
|
|
|
self._attempt_new_transaction(destination)
|
|
|
|
|
|
|
|
deferreds.append(deferred)
|
|
|
|
|
|
|
|
yield defer.DeferredList(deferreds)
|
|
|
|
|
|
|
|
# NO inlineCallbacks
|
|
|
|
def enqueue_edu(self, edu):
|
|
|
|
destination = edu.destination
|
|
|
|
|
|
|
|
deferred = defer.Deferred()
|
|
|
|
self.pending_edus_by_dest.setdefault(destination, []).append(
|
|
|
|
(edu, deferred)
|
|
|
|
)
|
|
|
|
|
|
|
|
def eb(failure):
|
2014-08-28 09:19:16 -06:00
|
|
|
if not deferred.called:
|
|
|
|
deferred.errback(failure)
|
2014-08-29 05:08:33 -06:00
|
|
|
else:
|
|
|
|
logger.exception("Failed to send edu", failure)
|
2014-08-12 08:10:52 -06:00
|
|
|
self._attempt_new_transaction(destination).addErrback(eb)
|
|
|
|
|
|
|
|
return deferred
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
@log_function
|
|
|
|
def _attempt_new_transaction(self, destination):
|
|
|
|
if destination in self.pending_transactions:
|
|
|
|
return
|
|
|
|
|
|
|
|
# list of (pending_pdu, deferred, order)
|
|
|
|
pending_pdus = self.pending_pdus_by_dest.pop(destination, [])
|
|
|
|
pending_edus = self.pending_edus_by_dest.pop(destination, [])
|
|
|
|
|
|
|
|
if not pending_pdus and not pending_edus:
|
|
|
|
return
|
|
|
|
|
|
|
|
logger.debug("TX [%s] Attempting new transaction", destination)
|
|
|
|
|
|
|
|
# Sort based on the order field
|
|
|
|
pending_pdus.sort(key=lambda t: t[2])
|
|
|
|
|
|
|
|
pdus = [x[0] for x in pending_pdus]
|
|
|
|
edus = [x[0] for x in pending_edus]
|
|
|
|
deferreds = [x[1] for x in pending_pdus + pending_edus]
|
|
|
|
|
|
|
|
try:
|
|
|
|
self.pending_transactions[destination] = 1
|
|
|
|
|
|
|
|
logger.debug("TX [%s] Persisting transaction...", destination)
|
|
|
|
|
|
|
|
transaction = Transaction.create_new(
|
2014-10-17 10:12:25 -06:00
|
|
|
origin_server_ts=self._clock.time_msec(),
|
2014-09-30 08:15:10 -06:00
|
|
|
transaction_id=str(self._next_txn_id),
|
2014-08-12 08:10:52 -06:00
|
|
|
origin=self.server_name,
|
|
|
|
destination=destination,
|
|
|
|
pdus=pdus,
|
|
|
|
edus=edus,
|
|
|
|
)
|
|
|
|
|
|
|
|
self._next_txn_id += 1
|
|
|
|
|
|
|
|
yield self.transaction_actions.prepare_to_send(transaction)
|
|
|
|
|
|
|
|
logger.debug("TX [%s] Persisted transaction", destination)
|
|
|
|
logger.debug("TX [%s] Sending transaction...", destination)
|
|
|
|
|
|
|
|
# Actually send the transaction
|
2014-09-15 07:54:25 -06:00
|
|
|
|
|
|
|
# FIXME (erikj): This is a bit of a hack to make the Pdu age
|
|
|
|
# keys work
|
2014-09-30 08:15:10 -06:00
|
|
|
def json_data_cb():
|
|
|
|
data = transaction.get_dict()
|
2014-09-15 07:54:25 -06:00
|
|
|
now = int(self._clock.time_msec())
|
2014-09-30 08:15:10 -06:00
|
|
|
if "pdus" in data:
|
|
|
|
for p in data["pdus"]:
|
2014-09-15 07:54:25 -06:00
|
|
|
if "age_ts" in p:
|
|
|
|
p["age"] = now - int(p["age_ts"])
|
2014-09-30 08:15:10 -06:00
|
|
|
return data
|
2014-09-15 07:54:25 -06:00
|
|
|
|
2014-08-12 08:10:52 -06:00
|
|
|
code, response = yield self.transport_layer.send_transaction(
|
2014-09-30 08:15:10 -06:00
|
|
|
transaction, json_data_cb
|
2014-08-12 08:10:52 -06:00
|
|
|
)
|
|
|
|
|
|
|
|
logger.debug("TX [%s] Sent transaction", destination)
|
|
|
|
logger.debug("TX [%s] Marking as delivered...", destination)
|
|
|
|
|
|
|
|
yield self.transaction_actions.delivered(
|
|
|
|
transaction, code, response
|
|
|
|
)
|
|
|
|
|
|
|
|
logger.debug("TX [%s] Marked as delivered", destination)
|
|
|
|
logger.debug("TX [%s] Yielding to callbacks...", destination)
|
|
|
|
|
|
|
|
for deferred in deferreds:
|
|
|
|
if code == 200:
|
|
|
|
deferred.callback(None)
|
|
|
|
else:
|
|
|
|
deferred.errback(RuntimeError("Got status %d" % code))
|
|
|
|
|
|
|
|
# Ensures we don't continue until all callbacks on that
|
|
|
|
# deferred have fired
|
|
|
|
yield deferred
|
|
|
|
|
|
|
|
logger.debug("TX [%s] Yielded to callbacks", destination)
|
|
|
|
|
|
|
|
except Exception as e:
|
|
|
|
logger.error("TX Problem in _attempt_transaction")
|
|
|
|
|
|
|
|
# We capture this here as there as nothing actually listens
|
|
|
|
# for this finishing functions deferred.
|
|
|
|
logger.exception(e)
|
|
|
|
|
|
|
|
for deferred in deferreds:
|
|
|
|
deferred.errback(e)
|
|
|
|
|
|
|
|
finally:
|
|
|
|
# We want to be *very* sure we delete this after we stop processing
|
|
|
|
self.pending_transactions.pop(destination, None)
|
|
|
|
|
|
|
|
# Check to see if there is anything else to send.
|
|
|
|
self._attempt_new_transaction(destination)
|