2016-08-25 10:35:37 -06:00
|
|
|
# -*- coding: utf-8 -*-
|
|
|
|
# Copyright 2016 OpenMarket Ltd
|
|
|
|
#
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
|
|
|
|
|
|
|
import logging
|
2018-06-28 07:49:57 -06:00
|
|
|
|
|
|
|
from canonicaljson import json
|
2016-08-25 10:35:37 -06:00
|
|
|
|
|
|
|
from twisted.internet import defer
|
|
|
|
|
2019-09-03 03:21:30 -06:00
|
|
|
from synapse.logging.opentracing import log_kv, set_tag, trace
|
2019-03-04 11:03:29 -07:00
|
|
|
from synapse.storage._base import SQLBaseStore
|
|
|
|
from synapse.storage.background_updates import BackgroundUpdateStore
|
2017-03-09 07:50:40 -07:00
|
|
|
from synapse.util.caches.expiringcache import ExpiringCache
|
|
|
|
|
2016-08-25 10:35:37 -06:00
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
|
|
|
|
2019-03-04 11:03:29 -07:00
|
|
|
class DeviceInboxWorkerStore(SQLBaseStore):
|
|
|
|
def get_to_device_stream_token(self):
|
|
|
|
return self._device_inbox_id_gen.get_current_token()
|
|
|
|
|
|
|
|
def get_new_messages_for_device(
|
|
|
|
self, user_id, device_id, last_stream_id, current_stream_id, limit=100
|
|
|
|
):
|
|
|
|
"""
|
|
|
|
Args:
|
|
|
|
user_id(str): The recipient user_id.
|
|
|
|
device_id(str): The recipient device_id.
|
|
|
|
current_stream_id(int): The current position of the to device
|
|
|
|
message stream.
|
|
|
|
Returns:
|
|
|
|
Deferred ([dict], int): List of messages for the device and where
|
|
|
|
in the stream the messages got to.
|
|
|
|
"""
|
|
|
|
has_changed = self._device_inbox_stream_cache.has_entity_changed(
|
|
|
|
user_id, last_stream_id
|
|
|
|
)
|
|
|
|
if not has_changed:
|
|
|
|
return defer.succeed(([], current_stream_id))
|
|
|
|
|
|
|
|
def get_new_messages_for_device_txn(txn):
|
|
|
|
sql = (
|
|
|
|
"SELECT stream_id, message_json FROM device_inbox"
|
|
|
|
" WHERE user_id = ? AND device_id = ?"
|
|
|
|
" AND ? < stream_id AND stream_id <= ?"
|
|
|
|
" ORDER BY stream_id ASC"
|
|
|
|
" LIMIT ?"
|
|
|
|
)
|
2019-04-03 03:07:29 -06:00
|
|
|
txn.execute(
|
|
|
|
sql, (user_id, device_id, last_stream_id, current_stream_id, limit)
|
|
|
|
)
|
2019-03-04 11:03:29 -07:00
|
|
|
messages = []
|
|
|
|
for row in txn:
|
|
|
|
stream_pos = row[0]
|
|
|
|
messages.append(json.loads(row[1]))
|
|
|
|
if len(messages) < limit:
|
|
|
|
stream_pos = current_stream_id
|
2019-08-30 09:28:26 -06:00
|
|
|
return messages, stream_pos
|
2019-03-04 11:03:29 -07:00
|
|
|
|
|
|
|
return self.runInteraction(
|
2019-04-03 03:07:29 -06:00
|
|
|
"get_new_messages_for_device", get_new_messages_for_device_txn
|
2019-03-04 11:03:29 -07:00
|
|
|
)
|
|
|
|
|
2019-09-03 03:21:30 -06:00
|
|
|
@trace
|
2019-03-04 11:03:29 -07:00
|
|
|
@defer.inlineCallbacks
|
|
|
|
def delete_messages_for_device(self, user_id, device_id, up_to_stream_id):
|
|
|
|
"""
|
|
|
|
Args:
|
|
|
|
user_id(str): The recipient user_id.
|
|
|
|
device_id(str): The recipient device_id.
|
|
|
|
up_to_stream_id(int): Where to delete messages up to.
|
|
|
|
Returns:
|
|
|
|
A deferred that resolves to the number of messages deleted.
|
|
|
|
"""
|
|
|
|
# If we have cached the last stream id we've deleted up to, we can
|
|
|
|
# check if there is likely to be anything that needs deleting
|
|
|
|
last_deleted_stream_id = self._last_device_delete_cache.get(
|
|
|
|
(user_id, device_id), None
|
|
|
|
)
|
2019-09-03 03:21:30 -06:00
|
|
|
|
|
|
|
set_tag("last_deleted_stream_id", last_deleted_stream_id)
|
|
|
|
|
2019-03-04 11:03:29 -07:00
|
|
|
if last_deleted_stream_id:
|
|
|
|
has_changed = self._device_inbox_stream_cache.has_entity_changed(
|
|
|
|
user_id, last_deleted_stream_id
|
|
|
|
)
|
|
|
|
if not has_changed:
|
2019-09-03 03:21:30 -06:00
|
|
|
log_kv({"message": "No changes in cache since last check"})
|
2019-07-23 07:00:55 -06:00
|
|
|
return 0
|
2019-03-04 11:03:29 -07:00
|
|
|
|
|
|
|
def delete_messages_for_device_txn(txn):
|
|
|
|
sql = (
|
|
|
|
"DELETE FROM device_inbox"
|
|
|
|
" WHERE user_id = ? AND device_id = ?"
|
|
|
|
" AND stream_id <= ?"
|
|
|
|
)
|
|
|
|
txn.execute(sql, (user_id, device_id, up_to_stream_id))
|
|
|
|
return txn.rowcount
|
|
|
|
|
|
|
|
count = yield self.runInteraction(
|
|
|
|
"delete_messages_for_device", delete_messages_for_device_txn
|
|
|
|
)
|
|
|
|
|
2019-09-03 03:21:30 -06:00
|
|
|
log_kv(
|
|
|
|
{"message": "deleted {} messages for device".format(count), "count": count}
|
|
|
|
)
|
|
|
|
|
2019-03-04 11:03:29 -07:00
|
|
|
# Update the cache, ensuring that we only ever increase the value
|
|
|
|
last_deleted_stream_id = self._last_device_delete_cache.get(
|
|
|
|
(user_id, device_id), 0
|
|
|
|
)
|
|
|
|
self._last_device_delete_cache[(user_id, device_id)] = max(
|
|
|
|
last_deleted_stream_id, up_to_stream_id
|
|
|
|
)
|
|
|
|
|
2019-07-23 07:00:55 -06:00
|
|
|
return count
|
2019-03-04 11:03:29 -07:00
|
|
|
|
2019-09-03 03:21:30 -06:00
|
|
|
@trace
|
2019-03-04 11:03:29 -07:00
|
|
|
def get_new_device_msgs_for_remote(
|
2019-05-09 04:01:41 -06:00
|
|
|
self, destination, last_stream_id, current_stream_id, limit
|
2019-03-04 11:03:29 -07:00
|
|
|
):
|
|
|
|
"""
|
|
|
|
Args:
|
|
|
|
destination(str): The name of the remote server.
|
|
|
|
last_stream_id(int|long): The last position of the device message stream
|
|
|
|
that the server sent up to.
|
|
|
|
current_stream_id(int|long): The current position of the device
|
|
|
|
message stream.
|
|
|
|
Returns:
|
|
|
|
Deferred ([dict], int|long): List of messages for the device and where
|
|
|
|
in the stream the messages got to.
|
|
|
|
"""
|
|
|
|
|
2019-09-03 03:21:30 -06:00
|
|
|
set_tag("destination", destination)
|
|
|
|
set_tag("last_stream_id", last_stream_id)
|
|
|
|
set_tag("current_stream_id", current_stream_id)
|
|
|
|
set_tag("limit", limit)
|
|
|
|
|
2019-03-04 11:03:29 -07:00
|
|
|
has_changed = self._device_federation_outbox_stream_cache.has_entity_changed(
|
|
|
|
destination, last_stream_id
|
|
|
|
)
|
|
|
|
if not has_changed or last_stream_id == current_stream_id:
|
2019-09-03 03:21:30 -06:00
|
|
|
log_kv({"message": "No new messages in stream"})
|
2019-03-04 11:03:29 -07:00
|
|
|
return defer.succeed(([], current_stream_id))
|
|
|
|
|
2019-06-10 09:21:42 -06:00
|
|
|
if limit <= 0:
|
|
|
|
# This can happen if we run out of room for EDUs in the transaction.
|
|
|
|
return defer.succeed(([], last_stream_id))
|
|
|
|
|
2019-09-03 03:21:30 -06:00
|
|
|
@trace
|
2019-03-04 11:03:29 -07:00
|
|
|
def get_new_messages_for_remote_destination_txn(txn):
|
|
|
|
sql = (
|
|
|
|
"SELECT stream_id, messages_json FROM device_federation_outbox"
|
|
|
|
" WHERE destination = ?"
|
|
|
|
" AND ? < stream_id AND stream_id <= ?"
|
|
|
|
" ORDER BY stream_id ASC"
|
|
|
|
" LIMIT ?"
|
|
|
|
)
|
2019-04-03 03:07:29 -06:00
|
|
|
txn.execute(sql, (destination, last_stream_id, current_stream_id, limit))
|
2019-03-04 11:03:29 -07:00
|
|
|
messages = []
|
|
|
|
for row in txn:
|
|
|
|
stream_pos = row[0]
|
|
|
|
messages.append(json.loads(row[1]))
|
|
|
|
if len(messages) < limit:
|
2019-09-03 03:21:30 -06:00
|
|
|
log_kv({"message": "Set stream position to current position"})
|
2019-03-04 11:03:29 -07:00
|
|
|
stream_pos = current_stream_id
|
2019-08-30 09:28:26 -06:00
|
|
|
return messages, stream_pos
|
2019-03-04 11:03:29 -07:00
|
|
|
|
|
|
|
return self.runInteraction(
|
|
|
|
"get_new_device_msgs_for_remote",
|
|
|
|
get_new_messages_for_remote_destination_txn,
|
|
|
|
)
|
|
|
|
|
2019-09-03 03:21:30 -06:00
|
|
|
@trace
|
2019-03-04 11:03:29 -07:00
|
|
|
def delete_device_msgs_for_remote(self, destination, up_to_stream_id):
|
|
|
|
"""Used to delete messages when the remote destination acknowledges
|
|
|
|
their receipt.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
destination(str): The destination server_name
|
|
|
|
up_to_stream_id(int): Where to delete messages up to.
|
|
|
|
Returns:
|
|
|
|
A deferred that resolves when the messages have been deleted.
|
|
|
|
"""
|
2019-04-03 03:07:29 -06:00
|
|
|
|
2019-03-04 11:03:29 -07:00
|
|
|
def delete_messages_for_remote_destination_txn(txn):
|
|
|
|
sql = (
|
|
|
|
"DELETE FROM device_federation_outbox"
|
|
|
|
" WHERE destination = ?"
|
|
|
|
" AND stream_id <= ?"
|
|
|
|
)
|
|
|
|
txn.execute(sql, (destination, up_to_stream_id))
|
|
|
|
|
|
|
|
return self.runInteraction(
|
2019-04-03 03:07:29 -06:00
|
|
|
"delete_device_msgs_for_remote", delete_messages_for_remote_destination_txn
|
2019-03-04 11:03:29 -07:00
|
|
|
)
|
|
|
|
|
|
|
|
|
|
|
|
class DeviceInboxStore(DeviceInboxWorkerStore, BackgroundUpdateStore):
|
2017-01-10 08:04:57 -07:00
|
|
|
DEVICE_INBOX_STREAM_ID = "device_inbox_stream_drop"
|
|
|
|
|
2017-11-09 11:51:27 -07:00
|
|
|
def __init__(self, db_conn, hs):
|
|
|
|
super(DeviceInboxStore, self).__init__(db_conn, hs)
|
2017-01-10 08:04:57 -07:00
|
|
|
|
|
|
|
self.register_background_index_update(
|
|
|
|
"device_inbox_stream_index",
|
|
|
|
index_name="device_inbox_stream_id_user_id",
|
|
|
|
table="device_inbox",
|
|
|
|
columns=["stream_id", "user_id"],
|
|
|
|
)
|
|
|
|
|
|
|
|
self.register_background_update_handler(
|
2019-04-03 03:07:29 -06:00
|
|
|
self.DEVICE_INBOX_STREAM_ID, self._background_drop_index_device_inbox
|
2017-01-10 08:04:57 -07:00
|
|
|
)
|
2016-08-25 10:35:37 -06:00
|
|
|
|
2017-03-09 07:50:40 -07:00
|
|
|
# Map of (user_id, device_id) to the last stream_id that has been
|
|
|
|
# deleted up to. This is so that we can no op deletions.
|
|
|
|
self._last_device_delete_cache = ExpiringCache(
|
|
|
|
cache_name="last_device_delete_cache",
|
|
|
|
clock=self._clock,
|
|
|
|
max_len=10000,
|
|
|
|
expiry_ms=30 * 60 * 1000,
|
|
|
|
)
|
|
|
|
|
2019-09-03 03:21:30 -06:00
|
|
|
@trace
|
2016-08-25 10:35:37 -06:00
|
|
|
@defer.inlineCallbacks
|
2019-04-03 03:07:29 -06:00
|
|
|
def add_messages_to_device_inbox(
|
|
|
|
self, local_messages_by_user_then_device, remote_messages_by_destination
|
|
|
|
):
|
2016-09-06 08:12:13 -06:00
|
|
|
"""Used to send messages from this server.
|
|
|
|
|
2016-08-25 10:35:37 -06:00
|
|
|
Args:
|
2016-09-06 08:12:13 -06:00
|
|
|
sender_user_id(str): The ID of the user sending these messages.
|
|
|
|
local_messages_by_user_and_device(dict):
|
2016-08-25 10:35:37 -06:00
|
|
|
Dictionary of user_id to device_id to message.
|
2016-09-06 08:12:13 -06:00
|
|
|
remote_messages_by_destination(dict):
|
|
|
|
Dictionary of destination server_name to the EDU JSON to send.
|
2016-08-25 10:35:37 -06:00
|
|
|
Returns:
|
2016-08-31 03:42:52 -06:00
|
|
|
A deferred stream_id that resolves when the messages have been
|
|
|
|
inserted.
|
2016-08-25 10:35:37 -06:00
|
|
|
"""
|
|
|
|
|
2016-09-07 05:03:37 -06:00
|
|
|
def add_messages_txn(txn, now_ms, stream_id):
|
|
|
|
# Add the local messages directly to the local inbox.
|
|
|
|
self._add_messages_to_local_device_inbox_txn(
|
|
|
|
txn, stream_id, local_messages_by_user_then_device
|
|
|
|
)
|
|
|
|
|
|
|
|
# Add the remote messages to the federation outbox.
|
|
|
|
# We'll send them to a remote server when we next send a
|
|
|
|
# federation transaction to that destination.
|
2016-09-06 08:12:13 -06:00
|
|
|
sql = (
|
|
|
|
"INSERT INTO device_federation_outbox"
|
|
|
|
" (destination, stream_id, queued_ts, messages_json)"
|
|
|
|
" VALUES (?,?,?,?)"
|
|
|
|
)
|
|
|
|
rows = []
|
|
|
|
for destination, edu in remote_messages_by_destination.items():
|
2018-06-28 07:49:57 -06:00
|
|
|
edu_json = json.dumps(edu)
|
2016-09-06 08:12:13 -06:00
|
|
|
rows.append((destination, stream_id, now_ms, edu_json))
|
|
|
|
txn.executemany(sql, rows)
|
|
|
|
|
2016-09-06 04:26:37 -06:00
|
|
|
with self._device_inbox_id_gen.get_next() as stream_id:
|
2016-09-06 11:16:20 -06:00
|
|
|
now_ms = self.clock.time_msec()
|
2016-09-06 04:26:37 -06:00
|
|
|
yield self.runInteraction(
|
2019-04-03 03:07:29 -06:00
|
|
|
"add_messages_to_device_inbox", add_messages_txn, now_ms, stream_id
|
2016-09-06 04:26:37 -06:00
|
|
|
)
|
2016-09-07 08:27:07 -06:00
|
|
|
for user_id in local_messages_by_user_then_device.keys():
|
2019-04-03 03:07:29 -06:00
|
|
|
self._device_inbox_stream_cache.entity_has_changed(user_id, stream_id)
|
2016-09-07 08:27:07 -06:00
|
|
|
for destination in remote_messages_by_destination.keys():
|
|
|
|
self._device_federation_outbox_stream_cache.entity_has_changed(
|
|
|
|
destination, stream_id
|
|
|
|
)
|
2016-09-06 04:26:37 -06:00
|
|
|
|
2019-07-23 07:00:55 -06:00
|
|
|
return self._device_inbox_id_gen.get_current_token()
|
2016-09-06 04:26:37 -06:00
|
|
|
|
2016-09-06 08:12:13 -06:00
|
|
|
@defer.inlineCallbacks
|
|
|
|
def add_messages_from_remote_to_device_inbox(
|
|
|
|
self, origin, message_id, local_messages_by_user_then_device
|
|
|
|
):
|
|
|
|
def add_messages_txn(txn, now_ms, stream_id):
|
2016-09-07 05:03:37 -06:00
|
|
|
# Check if we've already inserted a matching message_id for that
|
|
|
|
# origin. This can happen if the origin doesn't receive our
|
|
|
|
# acknowledgement from the first time we received the message.
|
2016-09-06 08:12:13 -06:00
|
|
|
already_inserted = self._simple_select_one_txn(
|
2019-04-03 03:07:29 -06:00
|
|
|
txn,
|
|
|
|
table="device_federation_inbox",
|
2016-09-06 08:12:13 -06:00
|
|
|
keyvalues={"origin": origin, "message_id": message_id},
|
|
|
|
retcols=("message_id",),
|
|
|
|
allow_none=True,
|
|
|
|
)
|
|
|
|
if already_inserted is not None:
|
|
|
|
return
|
|
|
|
|
2016-09-07 05:03:37 -06:00
|
|
|
# Add an entry for this message_id so that we know we've processed
|
|
|
|
# it.
|
2016-09-06 08:12:13 -06:00
|
|
|
self._simple_insert_txn(
|
2019-04-03 03:07:29 -06:00
|
|
|
txn,
|
|
|
|
table="device_federation_inbox",
|
2016-09-06 08:12:13 -06:00
|
|
|
values={
|
|
|
|
"origin": origin,
|
|
|
|
"message_id": message_id,
|
|
|
|
"received_ts": now_ms,
|
|
|
|
},
|
|
|
|
)
|
|
|
|
|
2016-09-07 05:03:37 -06:00
|
|
|
# Add the messages to the approriate local device inboxes so that
|
|
|
|
# they'll be sent to the devices when they next sync.
|
2016-09-06 08:12:13 -06:00
|
|
|
self._add_messages_to_local_device_inbox_txn(
|
|
|
|
txn, stream_id, local_messages_by_user_then_device
|
|
|
|
)
|
|
|
|
|
|
|
|
with self._device_inbox_id_gen.get_next() as stream_id:
|
2016-09-06 11:16:20 -06:00
|
|
|
now_ms = self.clock.time_msec()
|
2016-09-06 08:12:13 -06:00
|
|
|
yield self.runInteraction(
|
|
|
|
"add_messages_from_remote_to_device_inbox",
|
|
|
|
add_messages_txn,
|
|
|
|
now_ms,
|
|
|
|
stream_id,
|
|
|
|
)
|
2016-09-07 08:27:07 -06:00
|
|
|
for user_id in local_messages_by_user_then_device.keys():
|
2019-04-03 03:07:29 -06:00
|
|
|
self._device_inbox_stream_cache.entity_has_changed(user_id, stream_id)
|
2016-09-06 08:12:13 -06:00
|
|
|
|
2019-07-23 07:00:55 -06:00
|
|
|
return stream_id
|
2016-09-12 05:30:46 -06:00
|
|
|
|
2019-04-03 03:07:29 -06:00
|
|
|
def _add_messages_to_local_device_inbox_txn(
|
|
|
|
self, txn, stream_id, messages_by_user_then_device
|
|
|
|
):
|
|
|
|
sql = "UPDATE device_max_stream_id" " SET stream_id = ?" " WHERE stream_id < ?"
|
2016-09-09 04:48:23 -06:00
|
|
|
txn.execute(sql, (stream_id, stream_id))
|
|
|
|
|
2016-09-08 08:13:05 -06:00
|
|
|
local_by_user_then_device = {}
|
2016-09-06 04:26:37 -06:00
|
|
|
for user_id, messages_by_device in messages_by_user_then_device.items():
|
2016-09-08 08:13:05 -06:00
|
|
|
messages_json_for_user = {}
|
2018-08-30 08:19:58 -06:00
|
|
|
devices = list(messages_by_device.keys())
|
2016-09-08 08:13:05 -06:00
|
|
|
if len(devices) == 1 and devices[0] == "*":
|
|
|
|
# Handle wildcard device_ids.
|
2019-04-03 03:07:29 -06:00
|
|
|
sql = "SELECT device_id FROM devices" " WHERE user_id = ?"
|
2016-09-08 08:13:05 -06:00
|
|
|
txn.execute(sql, (user_id,))
|
2018-06-28 07:49:57 -06:00
|
|
|
message_json = json.dumps(messages_by_device["*"])
|
2017-03-23 11:53:49 -06:00
|
|
|
for row in txn:
|
2016-09-08 08:13:05 -06:00
|
|
|
# Add the message for all devices for this user on this
|
|
|
|
# server.
|
|
|
|
device = row[0]
|
|
|
|
messages_json_for_user[device] = message_json
|
|
|
|
else:
|
2016-09-09 04:48:23 -06:00
|
|
|
if not devices:
|
|
|
|
continue
|
2016-09-08 08:13:05 -06:00
|
|
|
sql = (
|
|
|
|
"SELECT device_id FROM devices"
|
|
|
|
" WHERE user_id = ? AND device_id IN ("
|
|
|
|
+ ",".join("?" * len(devices))
|
|
|
|
+ ")"
|
|
|
|
)
|
|
|
|
# TODO: Maybe this needs to be done in batches if there are
|
|
|
|
# too many local devices for a given user.
|
|
|
|
txn.execute(sql, [user_id] + devices)
|
2017-03-23 11:53:49 -06:00
|
|
|
for row in txn:
|
2016-09-08 08:13:05 -06:00
|
|
|
# Only insert into the local inbox if the device exists on
|
|
|
|
# this server
|
|
|
|
device = row[0]
|
2018-06-28 07:49:57 -06:00
|
|
|
message_json = json.dumps(messages_by_device[device])
|
2016-09-08 08:13:05 -06:00
|
|
|
messages_json_for_user[device] = message_json
|
|
|
|
|
2016-09-09 04:48:23 -06:00
|
|
|
if messages_json_for_user:
|
|
|
|
local_by_user_then_device[user_id] = messages_json_for_user
|
|
|
|
|
|
|
|
if not local_by_user_then_device:
|
|
|
|
return
|
2016-08-25 10:35:37 -06:00
|
|
|
|
2016-09-06 04:26:37 -06:00
|
|
|
sql = (
|
|
|
|
"INSERT INTO device_inbox"
|
|
|
|
" (user_id, device_id, stream_id, message_json)"
|
|
|
|
" VALUES (?,?,?,?)"
|
|
|
|
)
|
|
|
|
rows = []
|
2016-09-08 08:13:05 -06:00
|
|
|
for user_id, messages_by_device in local_by_user_then_device.items():
|
|
|
|
for device_id, message_json in messages_by_device.items():
|
|
|
|
rows.append((user_id, device_id, stream_id, message_json))
|
2016-08-25 10:35:37 -06:00
|
|
|
|
2016-09-06 04:26:37 -06:00
|
|
|
txn.executemany(sql, rows)
|
2016-08-31 03:42:52 -06:00
|
|
|
|
2016-08-31 03:38:58 -06:00
|
|
|
def get_all_new_device_messages(self, last_pos, current_pos, limit):
|
|
|
|
"""
|
|
|
|
Args:
|
|
|
|
last_pos(int):
|
|
|
|
current_pos(int):
|
|
|
|
limit(int):
|
|
|
|
Returns:
|
|
|
|
A deferred list of rows from the device inbox
|
|
|
|
"""
|
|
|
|
if last_pos == current_pos:
|
|
|
|
return defer.succeed([])
|
|
|
|
|
|
|
|
def get_all_new_device_messages_txn(txn):
|
2016-11-17 08:46:44 -07:00
|
|
|
# We limit like this as we might have multiple rows per stream_id, and
|
|
|
|
# we want to make sure we always get all entries for any stream_id
|
|
|
|
# we return.
|
|
|
|
upper_pos = min(current_pos, last_pos + limit)
|
2016-08-31 03:38:58 -06:00
|
|
|
sql = (
|
2017-04-04 09:21:21 -06:00
|
|
|
"SELECT max(stream_id), user_id"
|
2016-11-17 08:46:44 -07:00
|
|
|
" FROM device_inbox"
|
2016-08-31 03:38:58 -06:00
|
|
|
" WHERE ? < stream_id AND stream_id <= ?"
|
2017-04-04 09:21:21 -06:00
|
|
|
" GROUP BY user_id"
|
2016-08-31 03:38:58 -06:00
|
|
|
)
|
2016-11-17 08:46:44 -07:00
|
|
|
txn.execute(sql, (last_pos, upper_pos))
|
|
|
|
rows = txn.fetchall()
|
2016-08-31 03:38:58 -06:00
|
|
|
|
|
|
|
sql = (
|
2017-04-04 09:21:21 -06:00
|
|
|
"SELECT max(stream_id), destination"
|
2016-11-17 08:46:44 -07:00
|
|
|
" FROM device_federation_outbox"
|
2016-08-31 03:38:58 -06:00
|
|
|
" WHERE ? < stream_id AND stream_id <= ?"
|
2017-04-04 09:21:21 -06:00
|
|
|
" GROUP BY destination"
|
2016-08-31 03:38:58 -06:00
|
|
|
)
|
2016-11-17 08:46:44 -07:00
|
|
|
txn.execute(sql, (last_pos, upper_pos))
|
2017-03-23 11:53:49 -06:00
|
|
|
rows.extend(txn)
|
2016-11-17 08:46:44 -07:00
|
|
|
|
2017-04-04 09:21:21 -06:00
|
|
|
# Order by ascending stream ordering
|
|
|
|
rows.sort()
|
|
|
|
|
2016-11-17 08:46:44 -07:00
|
|
|
return rows
|
2016-08-31 03:38:58 -06:00
|
|
|
|
|
|
|
return self.runInteraction(
|
|
|
|
"get_all_new_device_messages", get_all_new_device_messages_txn
|
|
|
|
)
|
|
|
|
|
2017-01-10 08:04:57 -07:00
|
|
|
@defer.inlineCallbacks
|
|
|
|
def _background_drop_index_device_inbox(self, progress, batch_size):
|
|
|
|
def reindex_txn(conn):
|
2017-01-10 08:11:27 -07:00
|
|
|
txn = conn.cursor()
|
2019-04-03 03:07:29 -06:00
|
|
|
txn.execute("DROP INDEX IF EXISTS device_inbox_stream_id")
|
2017-01-10 08:15:25 -07:00
|
|
|
txn.close()
|
2017-01-10 08:04:57 -07:00
|
|
|
|
|
|
|
yield self.runWithConnection(reindex_txn)
|
|
|
|
|
|
|
|
yield self._end_background_update(self.DEVICE_INBOX_STREAM_ID)
|
|
|
|
|
2019-07-23 07:00:55 -06:00
|
|
|
return 1
|