2015-03-20 07:52:56 -06:00
|
|
|
# -*- coding: utf-8 -*-
|
2016-01-06 21:26:29 -07:00
|
|
|
# Copyright 2014-2016 OpenMarket Ltd
|
2019-05-30 04:22:59 -06:00
|
|
|
# Copyright 2018-2019 New Vector Ltd
|
|
|
|
# Copyright 2019 The Matrix.org Foundation C.I.C.
|
2015-03-20 07:52:56 -06:00
|
|
|
#
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
2018-04-17 11:30:53 -06:00
|
|
|
import itertools
|
2018-03-29 15:57:28 -06:00
|
|
|
import logging
|
2020-05-13 06:38:22 -06:00
|
|
|
from collections import OrderedDict, namedtuple
|
2020-09-23 09:42:14 -06:00
|
|
|
from typing import TYPE_CHECKING, Any, Dict, Iterable, List, Optional, Set, Tuple
|
2018-07-09 00:09:20 -06:00
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
import attr
|
2019-10-23 05:00:21 -06:00
|
|
|
from prometheus_client import Counter
|
2018-06-28 07:49:57 -06:00
|
|
|
|
2018-07-09 00:09:20 -06:00
|
|
|
import synapse.metrics
|
2020-07-07 07:20:40 -06:00
|
|
|
from synapse.api.constants import EventContentFields, EventTypes, RelationTypes
|
2020-01-29 04:01:32 -07:00
|
|
|
from synapse.api.room_versions import RoomVersions
|
2020-05-13 06:38:22 -06:00
|
|
|
from synapse.crypto.event_signing import compute_event_reference_hash
|
2018-07-09 00:09:20 -06:00
|
|
|
from synapse.events import EventBase # noqa: F401
|
|
|
|
from synapse.events.snapshot import EventContext # noqa: F401
|
2019-07-03 08:07:04 -06:00
|
|
|
from synapse.logging.utils import log_function
|
2020-07-16 09:32:19 -06:00
|
|
|
from synapse.storage._base import db_to_json, make_in_list_sql_clause
|
2020-08-05 14:38:57 -06:00
|
|
|
from synapse.storage.database import DatabasePool, LoggingTransaction
|
|
|
|
from synapse.storage.databases.main.search import SearchEntry
|
2020-09-14 03:16:41 -06:00
|
|
|
from synapse.storage.util.id_generators import MultiWriterIdGenerator
|
2020-05-13 06:38:22 -06:00
|
|
|
from synapse.types import StateMap, get_domain_from_id
|
2018-03-29 15:57:28 -06:00
|
|
|
from synapse.util.frozenutils import frozendict_json_encoder
|
2020-01-14 04:58:02 -07:00
|
|
|
from synapse.util.iterutils import batch_iter
|
2015-03-20 07:52:56 -06:00
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
if TYPE_CHECKING:
|
|
|
|
from synapse.server import HomeServer
|
2020-08-05 14:38:57 -06:00
|
|
|
from synapse.storage.databases.main import DataStore
|
2020-05-13 06:38:22 -06:00
|
|
|
|
|
|
|
|
2018-05-21 18:48:57 -06:00
|
|
|
logger = logging.getLogger(__name__)
|
2015-03-20 07:52:56 -06:00
|
|
|
|
2018-05-21 18:48:57 -06:00
|
|
|
persist_event_counter = Counter("synapse_storage_events_persisted_events", "")
|
2019-03-28 07:37:16 -06:00
|
|
|
event_counter = Counter(
|
|
|
|
"synapse_storage_events_persisted_events_sep",
|
|
|
|
"",
|
|
|
|
["type", "origin_type", "origin_entity"],
|
|
|
|
)
|
2018-03-27 06:13:38 -06:00
|
|
|
|
2016-06-06 04:58:09 -06:00
|
|
|
|
2016-06-03 10:55:32 -06:00
|
|
|
_EventCacheEntry = namedtuple("_EventCacheEntry", ("event", "redacted_event"))
|
|
|
|
|
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
@attr.s(slots=True)
|
|
|
|
class DeltaState:
|
|
|
|
"""Deltas to use to update the `current_state_events` table.
|
2019-05-28 11:52:41 -06:00
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
Attributes:
|
|
|
|
to_delete: List of type/state_keys to delete from current state
|
|
|
|
to_insert: Map of state to upsert into current state
|
|
|
|
no_longer_in_room: The server is not longer in the room, so the room
|
|
|
|
should e.g. be removed from `current_state_events` table.
|
|
|
|
"""
|
2019-06-13 06:40:52 -06:00
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
to_delete = attr.ib(type=List[Tuple[str, str]])
|
|
|
|
to_insert = attr.ib(type=StateMap[str])
|
|
|
|
no_longer_in_room = attr.ib(type=bool, default=False)
|
2019-06-13 06:40:52 -06:00
|
|
|
|
2019-07-03 02:31:27 -06:00
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
class PersistEventsStore:
|
|
|
|
"""Contains all the functions for writing events to the database.
|
2019-06-13 06:40:52 -06:00
|
|
|
|
2020-05-13 03:37:11 -06:00
|
|
|
Should only be instantiated on one process (when using a worker mode setup).
|
2020-05-13 06:38:22 -06:00
|
|
|
|
|
|
|
Note: This is not part of the `DataStore` mixin.
|
|
|
|
"""
|
2019-08-29 10:38:51 -06:00
|
|
|
|
2020-08-05 14:38:57 -06:00
|
|
|
def __init__(
|
|
|
|
self, hs: "HomeServer", db: DatabasePool, main_data_store: "DataStore"
|
|
|
|
):
|
2020-05-13 06:38:22 -06:00
|
|
|
self.hs = hs
|
2020-08-05 14:38:57 -06:00
|
|
|
self.db_pool = db
|
2020-05-13 06:38:22 -06:00
|
|
|
self.store = main_data_store
|
|
|
|
self.database_engine = db.engine
|
|
|
|
self._clock = hs.get_clock()
|
2020-09-14 03:16:41 -06:00
|
|
|
self._instance_name = hs.get_instance_name()
|
2019-08-29 10:38:51 -06:00
|
|
|
|
2019-12-03 12:19:45 -07:00
|
|
|
self._ephemeral_messages_enabled = hs.config.enable_ephemeral_messages
|
2020-01-15 07:59:33 -07:00
|
|
|
self.is_mine_id = hs.is_mine_id
|
2019-12-03 12:19:45 -07:00
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
# Ideally we'd move these ID gens here, unfortunately some other ID
|
|
|
|
# generators are chained off them so doing so is a bit of a PITA.
|
2020-09-14 03:16:41 -06:00
|
|
|
self._backfill_id_gen = (
|
|
|
|
self.store._backfill_id_gen
|
|
|
|
) # type: MultiWriterIdGenerator
|
|
|
|
self._stream_id_gen = self.store._stream_id_gen # type: MultiWriterIdGenerator
|
2019-06-13 06:40:52 -06:00
|
|
|
|
2020-05-22 09:11:35 -06:00
|
|
|
# This should only exist on instances that are configured to write
|
2020-05-13 06:38:22 -06:00
|
|
|
assert (
|
2020-09-14 03:16:41 -06:00
|
|
|
hs.get_instance_name() in hs.config.worker.writers.events
|
2020-05-22 09:11:35 -06:00
|
|
|
), "Can only instantiate EventsStore on master"
|
2019-06-13 06:40:52 -06:00
|
|
|
|
2020-08-14 08:05:19 -06:00
|
|
|
async def _persist_events_and_state_updates(
|
2019-10-23 05:00:21 -06:00
|
|
|
self,
|
2020-01-20 11:07:20 -07:00
|
|
|
events_and_contexts: List[Tuple[EventBase, EventContext]],
|
|
|
|
current_state_for_room: Dict[str, StateMap[str]],
|
|
|
|
state_delta_for_room: Dict[str, DeltaState],
|
|
|
|
new_forward_extremeties: Dict[str, List[str]],
|
|
|
|
backfilled: bool = False,
|
2020-08-14 08:05:19 -06:00
|
|
|
) -> None:
|
2019-10-30 07:33:38 -06:00
|
|
|
"""Persist a set of events alongside updates to the current state and
|
|
|
|
forward extremities tables.
|
2017-03-17 05:51:13 -06:00
|
|
|
|
|
|
|
Args:
|
2020-01-20 11:07:20 -07:00
|
|
|
events_and_contexts:
|
|
|
|
current_state_for_room: Map from room_id to the current state of
|
|
|
|
the room based on forward extremities
|
|
|
|
state_delta_for_room: Map from room_id to the delta to apply to
|
|
|
|
room state
|
|
|
|
new_forward_extremities: Map from room_id to list of event IDs
|
|
|
|
that are the new forward extremities of the room.
|
|
|
|
backfilled
|
2017-03-17 05:51:13 -06:00
|
|
|
|
|
|
|
Returns:
|
2020-08-14 08:05:19 -06:00
|
|
|
Resolves when the events have been persisted
|
2017-03-17 05:51:13 -06:00
|
|
|
"""
|
2015-06-25 10:18:19 -06:00
|
|
|
|
2019-10-23 05:00:21 -06:00
|
|
|
# We want to calculate the stream orderings as late as possible, as
|
|
|
|
# we only notify after all events with a lesser stream ordering have
|
|
|
|
# been persisted. I.e. if we spend 10s inside the with block then
|
|
|
|
# that will delay all subsequent events from being notified about.
|
|
|
|
# Hence why we do it down here rather than wrapping the entire
|
|
|
|
# function.
|
|
|
|
#
|
|
|
|
# Its safe to do this after calculating the state deltas etc as we
|
|
|
|
# only need to protect the *persistence* of the events. This is to
|
|
|
|
# ensure that queries of the form "fetch events since X" don't
|
|
|
|
# return events and stream positions after events that are still in
|
|
|
|
# flight, as otherwise subsequent requests "fetch event since Y"
|
|
|
|
# will not return those events.
|
|
|
|
#
|
|
|
|
# Note: Multiple instances of this function cannot be in flight at
|
|
|
|
# the same time for the same room.
|
|
|
|
if backfilled:
|
2020-09-23 09:11:18 -06:00
|
|
|
stream_ordering_manager = self._backfill_id_gen.get_next_mult(
|
2019-10-23 05:00:21 -06:00
|
|
|
len(events_and_contexts)
|
|
|
|
)
|
|
|
|
else:
|
2020-09-23 09:11:18 -06:00
|
|
|
stream_ordering_manager = self._stream_id_gen.get_next_mult(
|
2019-10-23 05:00:21 -06:00
|
|
|
len(events_and_contexts)
|
|
|
|
)
|
2019-08-06 06:27:22 -06:00
|
|
|
|
2020-09-23 09:11:18 -06:00
|
|
|
async with stream_ordering_manager as stream_orderings:
|
2019-10-23 05:00:21 -06:00
|
|
|
for (event, context), stream in zip(events_and_contexts, stream_orderings):
|
|
|
|
event.internal_metadata.stream_ordering = stream
|
2019-08-06 06:27:22 -06:00
|
|
|
|
2020-08-14 08:05:19 -06:00
|
|
|
await self.db_pool.runInteraction(
|
2019-10-23 05:00:21 -06:00
|
|
|
"persist_events",
|
|
|
|
self._persist_events_txn,
|
|
|
|
events_and_contexts=events_and_contexts,
|
|
|
|
backfilled=backfilled,
|
|
|
|
state_delta_for_room=state_delta_for_room,
|
|
|
|
new_forward_extremeties=new_forward_extremeties,
|
|
|
|
)
|
|
|
|
persist_event_counter.inc(len(events_and_contexts))
|
2019-08-06 06:27:22 -06:00
|
|
|
|
|
|
|
if not backfilled:
|
2019-10-23 05:00:21 -06:00
|
|
|
# backfilled events have negative stream orderings, so we don't
|
|
|
|
# want to set the event_persisted_position to that.
|
|
|
|
synapse.metrics.event_persisted_position.set(
|
|
|
|
events_and_contexts[-1][0].internal_metadata.stream_ordering
|
2016-08-30 09:54:40 -06:00
|
|
|
)
|
2017-01-20 07:28:53 -07:00
|
|
|
|
2019-10-23 05:00:21 -06:00
|
|
|
for event, context in events_and_contexts:
|
|
|
|
if context.app_service:
|
|
|
|
origin_type = "local"
|
|
|
|
origin_entity = context.app_service.id
|
|
|
|
elif self.hs.is_mine_id(event.sender):
|
|
|
|
origin_type = "local"
|
|
|
|
origin_entity = "*client*"
|
|
|
|
else:
|
|
|
|
origin_type = "remote"
|
|
|
|
origin_entity = get_domain_from_id(event.sender)
|
|
|
|
|
|
|
|
event_counter.labels(event.type, origin_type, origin_entity).inc()
|
|
|
|
|
2020-06-15 05:03:36 -06:00
|
|
|
for room_id, new_state in current_state_for_room.items():
|
2020-05-13 06:38:22 -06:00
|
|
|
self.store.get_current_state_ids.prefill((room_id,), new_state)
|
2019-10-23 05:00:21 -06:00
|
|
|
|
2020-06-15 05:03:36 -06:00
|
|
|
for room_id, latest_event_ids in new_forward_extremeties.items():
|
2020-05-13 06:38:22 -06:00
|
|
|
self.store.get_latest_event_ids_in_room.prefill(
|
2019-10-23 05:00:21 -06:00
|
|
|
(room_id,), list(latest_event_ids)
|
|
|
|
)
|
2017-01-20 07:28:53 -07:00
|
|
|
|
2020-08-14 08:05:19 -06:00
|
|
|
async def _get_events_which_are_prevs(self, event_ids: Iterable[str]) -> List[str]:
|
2018-10-02 16:33:29 -06:00
|
|
|
"""Filter the supplied list of event_ids to get those which are prev_events of
|
2018-10-03 03:19:41 -06:00
|
|
|
existing (non-outlier/rejected) events.
|
2018-10-02 16:33:29 -06:00
|
|
|
|
|
|
|
Args:
|
2020-08-14 08:05:19 -06:00
|
|
|
event_ids: event ids to filter
|
2018-10-02 16:33:29 -06:00
|
|
|
|
|
|
|
Returns:
|
2020-08-14 08:05:19 -06:00
|
|
|
Filtered event ids
|
2018-10-02 16:33:29 -06:00
|
|
|
"""
|
2020-09-11 05:22:55 -06:00
|
|
|
results = [] # type: List[str]
|
2018-10-02 16:33:29 -06:00
|
|
|
|
2019-05-29 04:56:24 -06:00
|
|
|
def _get_events_which_are_prevs_txn(txn, batch):
|
2018-10-02 16:33:29 -06:00
|
|
|
sql = """
|
2019-05-21 09:10:54 -06:00
|
|
|
SELECT prev_event_id, internal_metadata
|
2018-10-02 16:33:29 -06:00
|
|
|
FROM event_edges
|
|
|
|
INNER JOIN events USING (event_id)
|
|
|
|
LEFT JOIN rejections USING (event_id)
|
2019-05-21 09:10:54 -06:00
|
|
|
LEFT JOIN event_json USING (event_id)
|
2018-10-02 16:33:29 -06:00
|
|
|
WHERE
|
2019-10-02 12:07:07 -06:00
|
|
|
NOT events.outlier
|
2018-10-02 16:33:29 -06:00
|
|
|
AND rejections.event_id IS NULL
|
2019-10-02 12:07:07 -06:00
|
|
|
AND
|
|
|
|
"""
|
|
|
|
|
|
|
|
clause, args = make_in_list_sql_clause(
|
|
|
|
self.database_engine, "prev_event_id", batch
|
2018-10-02 16:33:29 -06:00
|
|
|
)
|
2017-01-20 07:28:53 -07:00
|
|
|
|
2019-10-02 12:07:07 -06:00
|
|
|
txn.execute(sql + clause, args)
|
2020-07-16 09:32:19 -06:00
|
|
|
results.extend(r[0] for r in txn if not db_to_json(r[1]).get("soft_failed"))
|
2018-10-02 16:33:29 -06:00
|
|
|
|
|
|
|
for chunk in batch_iter(event_ids, 100):
|
2020-08-14 08:05:19 -06:00
|
|
|
await self.db_pool.runInteraction(
|
2019-06-13 06:40:52 -06:00
|
|
|
"_get_events_which_are_prevs", _get_events_which_are_prevs_txn, chunk
|
2019-05-29 04:56:24 -06:00
|
|
|
)
|
2018-10-02 16:33:29 -06:00
|
|
|
|
2019-07-23 07:00:55 -06:00
|
|
|
return results
|
2017-01-20 07:28:53 -07:00
|
|
|
|
2020-08-14 08:05:19 -06:00
|
|
|
async def _get_prevs_before_rejected(self, event_ids: Iterable[str]) -> Set[str]:
|
2019-05-29 04:56:24 -06:00
|
|
|
"""Get soft-failed ancestors to remove from the extremities.
|
|
|
|
|
|
|
|
Given a set of events, find all those that have been soft-failed or
|
|
|
|
rejected. Returns those soft failed/rejected events and their prev
|
|
|
|
events (whether soft-failed/rejected or not), and recurses up the
|
|
|
|
prev-event graph until it finds no more soft-failed/rejected events.
|
|
|
|
|
|
|
|
This is used to find extremities that are ancestors of new events, but
|
|
|
|
are separated by soft failed events.
|
|
|
|
|
|
|
|
Args:
|
2020-08-14 08:05:19 -06:00
|
|
|
event_ids: Events to find prev events for. Note that these must have
|
|
|
|
already been persisted.
|
2019-05-29 04:56:24 -06:00
|
|
|
|
|
|
|
Returns:
|
2020-08-14 08:05:19 -06:00
|
|
|
The previous events.
|
2019-05-29 04:56:24 -06:00
|
|
|
"""
|
|
|
|
|
|
|
|
# The set of event_ids to return. This includes all soft-failed events
|
|
|
|
# and their prev events.
|
|
|
|
existing_prevs = set()
|
|
|
|
|
|
|
|
def _get_prevs_before_rejected_txn(txn, batch):
|
|
|
|
to_recursively_check = batch
|
|
|
|
|
|
|
|
while to_recursively_check:
|
|
|
|
sql = """
|
|
|
|
SELECT
|
|
|
|
event_id, prev_event_id, internal_metadata,
|
|
|
|
rejections.event_id IS NOT NULL
|
|
|
|
FROM event_edges
|
|
|
|
INNER JOIN events USING (event_id)
|
|
|
|
LEFT JOIN rejections USING (event_id)
|
|
|
|
LEFT JOIN event_json USING (event_id)
|
|
|
|
WHERE
|
2019-10-02 12:07:07 -06:00
|
|
|
NOT events.outlier
|
|
|
|
AND
|
|
|
|
"""
|
|
|
|
|
|
|
|
clause, args = make_in_list_sql_clause(
|
|
|
|
self.database_engine, "event_id", to_recursively_check
|
2019-05-29 04:56:24 -06:00
|
|
|
)
|
|
|
|
|
2019-10-02 12:07:07 -06:00
|
|
|
txn.execute(sql + clause, args)
|
2019-05-29 04:56:24 -06:00
|
|
|
to_recursively_check = []
|
|
|
|
|
|
|
|
for event_id, prev_event_id, metadata, rejected in txn:
|
|
|
|
if prev_event_id in existing_prevs:
|
|
|
|
continue
|
|
|
|
|
2020-07-16 09:32:19 -06:00
|
|
|
soft_failed = db_to_json(metadata).get("soft_failed")
|
2019-05-29 04:56:24 -06:00
|
|
|
if soft_failed or rejected:
|
|
|
|
to_recursively_check.append(prev_event_id)
|
|
|
|
existing_prevs.add(prev_event_id)
|
|
|
|
|
|
|
|
for chunk in batch_iter(event_ids, 100):
|
2020-08-14 08:05:19 -06:00
|
|
|
await self.db_pool.runInteraction(
|
2019-06-13 06:40:52 -06:00
|
|
|
"_get_prevs_before_rejected", _get_prevs_before_rejected_txn, chunk
|
2019-05-29 04:56:24 -06:00
|
|
|
)
|
|
|
|
|
2019-07-23 07:00:55 -06:00
|
|
|
return existing_prevs
|
2019-05-29 04:56:24 -06:00
|
|
|
|
2015-06-25 10:18:19 -06:00
|
|
|
@log_function
|
2019-03-28 07:37:16 -06:00
|
|
|
def _persist_events_txn(
|
|
|
|
self,
|
2020-01-20 11:07:20 -07:00
|
|
|
txn: LoggingTransaction,
|
|
|
|
events_and_contexts: List[Tuple[EventBase, EventContext]],
|
|
|
|
backfilled: bool,
|
|
|
|
state_delta_for_room: Dict[str, DeltaState] = {},
|
|
|
|
new_forward_extremeties: Dict[str, List[str]] = {},
|
2019-03-28 07:37:16 -06:00
|
|
|
):
|
2016-07-26 03:49:52 -06:00
|
|
|
"""Insert some number of room events into the necessary database tables.
|
|
|
|
|
|
|
|
Rejected events are only inserted into the events table, the events_json table,
|
|
|
|
and the rejections table. Things reading from those table will need to check
|
|
|
|
whether the event was rejected.
|
2016-08-04 08:02:15 -06:00
|
|
|
|
2017-03-17 05:51:13 -06:00
|
|
|
Args:
|
2020-01-20 11:07:20 -07:00
|
|
|
txn
|
|
|
|
events_and_contexts: events to persist
|
|
|
|
backfilled: True if the events were backfilled
|
|
|
|
delete_existing True to purge existing table rows for the events
|
|
|
|
from the database. This is useful when retrying due to
|
2017-03-17 05:51:13 -06:00
|
|
|
IntegrityError.
|
2020-01-20 11:07:20 -07:00
|
|
|
state_delta_for_room: The current-state delta for each room.
|
|
|
|
new_forward_extremetie: The new forward extremities for each room.
|
|
|
|
For each room, a list of the event ids which are the forward
|
|
|
|
extremities.
|
2017-03-17 05:51:13 -06:00
|
|
|
|
2016-07-26 03:49:52 -06:00
|
|
|
"""
|
2018-02-20 05:33:04 -07:00
|
|
|
all_events_and_contexts = events_and_contexts
|
|
|
|
|
2019-04-02 05:42:39 -06:00
|
|
|
min_stream_order = events_and_contexts[0][0].internal_metadata.stream_ordering
|
2017-01-20 07:28:53 -07:00
|
|
|
max_stream_order = events_and_contexts[-1][0].internal_metadata.stream_ordering
|
2017-05-30 07:41:42 -06:00
|
|
|
|
2020-10-05 07:43:14 -06:00
|
|
|
# stream orderings should have been assigned by now
|
|
|
|
assert min_stream_order
|
|
|
|
assert max_stream_order
|
|
|
|
|
2017-03-17 05:51:13 -06:00
|
|
|
self._update_forward_extremities_txn(
|
|
|
|
txn,
|
|
|
|
new_forward_extremities=new_forward_extremeties,
|
|
|
|
max_stream_order=max_stream_order,
|
|
|
|
)
|
|
|
|
|
|
|
|
# Ensure that we don't have the same event twice.
|
|
|
|
events_and_contexts = self._filter_events_and_contexts_for_duplicates(
|
2019-03-28 07:37:16 -06:00
|
|
|
events_and_contexts
|
2017-03-17 05:51:13 -06:00
|
|
|
)
|
|
|
|
|
|
|
|
self._update_room_depths_txn(
|
2019-03-28 07:37:16 -06:00
|
|
|
txn, events_and_contexts=events_and_contexts, backfilled=backfilled
|
2017-03-17 05:51:13 -06:00
|
|
|
)
|
|
|
|
|
|
|
|
# _update_outliers_txn filters out any events which have already been
|
|
|
|
# persisted, and returns the filtered list.
|
|
|
|
events_and_contexts = self._update_outliers_txn(
|
2019-03-28 07:37:16 -06:00
|
|
|
txn, events_and_contexts=events_and_contexts
|
2017-03-17 05:51:13 -06:00
|
|
|
)
|
|
|
|
|
|
|
|
# From this point onwards the events are only events that we haven't
|
|
|
|
# seen before.
|
|
|
|
|
2019-03-28 07:37:16 -06:00
|
|
|
self._store_event_txn(txn, events_and_contexts=events_and_contexts)
|
2017-03-17 05:51:13 -06:00
|
|
|
|
2020-10-13 05:07:56 -06:00
|
|
|
self._persist_transaction_ids_txn(txn, events_and_contexts)
|
|
|
|
|
2018-02-06 07:31:24 -07:00
|
|
|
# Insert into event_to_state_groups.
|
|
|
|
self._store_event_state_mappings_txn(txn, events_and_contexts)
|
2017-03-17 08:30:16 -06:00
|
|
|
|
2018-10-16 07:01:53 -06:00
|
|
|
# We want to store event_auth mappings for rejected events, as they're
|
|
|
|
# used in state res v2.
|
|
|
|
# This is only necessary if the rejected event appears in an accepted
|
|
|
|
# event's auth chain, but its easier for now just to store them (and
|
|
|
|
# it doesn't take much storage compared to storing the entire event
|
|
|
|
# anyway).
|
2020-08-05 14:38:57 -06:00
|
|
|
self.db_pool.simple_insert_many_txn(
|
2018-10-16 07:01:53 -06:00
|
|
|
txn,
|
|
|
|
table="event_auth",
|
|
|
|
values=[
|
|
|
|
{
|
|
|
|
"event_id": event.event_id,
|
|
|
|
"room_id": event.room_id,
|
|
|
|
"auth_id": auth_id,
|
|
|
|
}
|
|
|
|
for event, _ in events_and_contexts
|
2018-11-05 06:35:15 -07:00
|
|
|
for auth_id in event.auth_event_ids()
|
2018-10-16 07:01:53 -06:00
|
|
|
if event.is_state()
|
|
|
|
],
|
|
|
|
)
|
|
|
|
|
2017-03-17 05:51:13 -06:00
|
|
|
# _store_rejected_events_txn filters out any events which were
|
|
|
|
# rejected, and returns the filtered list.
|
|
|
|
events_and_contexts = self._store_rejected_events_txn(
|
2019-03-28 07:37:16 -06:00
|
|
|
txn, events_and_contexts=events_and_contexts
|
2017-03-17 05:51:13 -06:00
|
|
|
)
|
|
|
|
|
|
|
|
# From this point onwards the events are only ones that weren't
|
|
|
|
# rejected.
|
|
|
|
|
|
|
|
self._update_metadata_tables_txn(
|
|
|
|
txn,
|
|
|
|
events_and_contexts=events_and_contexts,
|
2018-02-20 05:33:04 -07:00
|
|
|
all_events_and_contexts=all_events_and_contexts,
|
2017-03-17 05:51:13 -06:00
|
|
|
backfilled=backfilled,
|
|
|
|
)
|
|
|
|
|
2019-07-17 08:33:37 -06:00
|
|
|
# We call this last as it assumes we've inserted the events into
|
|
|
|
# room_memberships, where applicable.
|
|
|
|
self._update_current_state_txn(txn, state_delta_for_room, min_stream_order)
|
|
|
|
|
2020-10-13 05:07:56 -06:00
|
|
|
def _persist_transaction_ids_txn(
|
|
|
|
self,
|
|
|
|
txn: LoggingTransaction,
|
|
|
|
events_and_contexts: List[Tuple[EventBase, EventContext]],
|
|
|
|
):
|
|
|
|
"""Persist the mapping from transaction IDs to event IDs (if defined).
|
|
|
|
"""
|
|
|
|
|
|
|
|
to_insert = []
|
|
|
|
for event, _ in events_and_contexts:
|
|
|
|
token_id = getattr(event.internal_metadata, "token_id", None)
|
|
|
|
txn_id = getattr(event.internal_metadata, "txn_id", None)
|
|
|
|
if token_id and txn_id:
|
|
|
|
to_insert.append(
|
|
|
|
{
|
|
|
|
"event_id": event.event_id,
|
|
|
|
"room_id": event.room_id,
|
|
|
|
"user_id": event.sender,
|
|
|
|
"token_id": token_id,
|
|
|
|
"txn_id": txn_id,
|
|
|
|
"inserted_ts": self._clock.time_msec(),
|
|
|
|
}
|
|
|
|
)
|
|
|
|
|
|
|
|
if to_insert:
|
|
|
|
self.db_pool.simple_insert_many_txn(
|
|
|
|
txn, table="event_txn_id", values=to_insert,
|
|
|
|
)
|
|
|
|
|
2020-01-20 11:07:20 -07:00
|
|
|
def _update_current_state_txn(
|
|
|
|
self,
|
|
|
|
txn: LoggingTransaction,
|
|
|
|
state_delta_by_room: Dict[str, DeltaState],
|
|
|
|
stream_id: int,
|
|
|
|
):
|
2020-06-15 05:03:36 -06:00
|
|
|
for room_id, delta_state in state_delta_by_room.items():
|
2020-01-20 11:07:20 -07:00
|
|
|
to_delete = delta_state.to_delete
|
|
|
|
to_insert = delta_state.to_insert
|
2017-01-20 08:40:04 -07:00
|
|
|
|
2020-01-29 04:01:32 -07:00
|
|
|
if delta_state.no_longer_in_room:
|
|
|
|
# Server is no longer in the room so we delete the room from
|
|
|
|
# current_state_events, being careful we've already updated the
|
|
|
|
# rooms.room_version column (which gets populated in a
|
|
|
|
# background task).
|
|
|
|
self._upsert_room_version_txn(txn, room_id)
|
|
|
|
|
|
|
|
# Before deleting we populate the current_state_delta_stream
|
|
|
|
# so that async background tasks get told what happened.
|
|
|
|
sql = """
|
|
|
|
INSERT INTO current_state_delta_stream
|
2020-10-09 06:10:33 -06:00
|
|
|
(stream_id, instance_name, room_id, type, state_key, event_id, prev_event_id)
|
|
|
|
SELECT ?, ?, room_id, type, state_key, null, event_id
|
2020-01-29 04:01:32 -07:00
|
|
|
FROM current_state_events
|
|
|
|
WHERE room_id = ?
|
|
|
|
"""
|
2020-10-09 06:10:33 -06:00
|
|
|
txn.execute(sql, (stream_id, self._instance_name, room_id))
|
2020-01-29 04:01:32 -07:00
|
|
|
|
2020-08-05 14:38:57 -06:00
|
|
|
self.db_pool.simple_delete_txn(
|
2020-01-29 04:01:32 -07:00
|
|
|
txn, table="current_state_events", keyvalues={"room_id": room_id},
|
2017-01-20 08:40:04 -07:00
|
|
|
)
|
2020-01-29 04:01:32 -07:00
|
|
|
else:
|
|
|
|
# We're still in the room, so we update the current state as normal.
|
|
|
|
|
|
|
|
# First we add entries to the current_state_delta_stream. We
|
|
|
|
# do this before updating the current_state_events table so
|
|
|
|
# that we can use it to calculate the `prev_event_id`. (This
|
|
|
|
# allows us to not have to pull out the existing state
|
|
|
|
# unnecessarily).
|
|
|
|
#
|
|
|
|
# The stream_id for the update is chosen to be the minimum of the stream_ids
|
|
|
|
# for the batch of the events that we are persisting; that means we do not
|
|
|
|
# end up in a situation where workers see events before the
|
|
|
|
# current_state_delta updates.
|
|
|
|
#
|
|
|
|
sql = """
|
|
|
|
INSERT INTO current_state_delta_stream
|
2020-10-09 06:10:33 -06:00
|
|
|
(stream_id, instance_name, room_id, type, state_key, event_id, prev_event_id)
|
|
|
|
SELECT ?, ?, ?, ?, ?, ?, (
|
2020-01-29 04:01:32 -07:00
|
|
|
SELECT event_id FROM current_state_events
|
|
|
|
WHERE room_id = ? AND type = ? AND state_key = ?
|
2019-03-28 07:37:16 -06:00
|
|
|
)
|
2020-01-29 04:01:32 -07:00
|
|
|
"""
|
|
|
|
txn.executemany(
|
|
|
|
sql,
|
2019-03-28 07:37:16 -06:00
|
|
|
(
|
2020-01-29 04:01:32 -07:00
|
|
|
(
|
|
|
|
stream_id,
|
2020-10-09 06:10:33 -06:00
|
|
|
self._instance_name,
|
2020-01-29 04:01:32 -07:00
|
|
|
room_id,
|
|
|
|
etype,
|
|
|
|
state_key,
|
|
|
|
to_insert.get((etype, state_key)),
|
|
|
|
room_id,
|
|
|
|
etype,
|
|
|
|
state_key,
|
|
|
|
)
|
|
|
|
for etype, state_key in itertools.chain(to_delete, to_insert)
|
|
|
|
),
|
|
|
|
)
|
|
|
|
# Now we actually update the current_state_events table
|
2019-01-30 03:53:17 -07:00
|
|
|
|
2020-01-29 04:01:32 -07:00
|
|
|
txn.executemany(
|
|
|
|
"DELETE FROM current_state_events"
|
|
|
|
" WHERE room_id = ? AND type = ? AND state_key = ?",
|
|
|
|
(
|
|
|
|
(room_id, etype, state_key)
|
|
|
|
for etype, state_key in itertools.chain(to_delete, to_insert)
|
|
|
|
),
|
|
|
|
)
|
2019-01-30 03:53:17 -07:00
|
|
|
|
2020-01-29 04:01:32 -07:00
|
|
|
# We include the membership in the current state table, hence we do
|
|
|
|
# a lookup when we insert. This assumes that all events have already
|
|
|
|
# been inserted into room_memberships.
|
|
|
|
txn.executemany(
|
|
|
|
"""INSERT INTO current_state_events
|
|
|
|
(room_id, type, state_key, event_id, membership)
|
|
|
|
VALUES (?, ?, ?, ?, (SELECT membership FROM room_memberships WHERE event_id = ?))
|
|
|
|
""",
|
|
|
|
[
|
|
|
|
(room_id, key[0], key[1], ev_id, ev_id)
|
2020-06-15 05:03:36 -06:00
|
|
|
for key, ev_id in to_insert.items()
|
2020-01-29 04:01:32 -07:00
|
|
|
],
|
|
|
|
)
|
2017-01-20 08:40:04 -07:00
|
|
|
|
2020-01-29 04:01:32 -07:00
|
|
|
# We now update `local_current_membership`. We do this regardless
|
|
|
|
# of whether we're still in the room or not to handle the case where
|
|
|
|
# e.g. we just got banned (where we need to record that fact here).
|
2017-01-20 08:40:04 -07:00
|
|
|
|
2020-01-15 07:59:33 -07:00
|
|
|
# Note: Do we really want to delete rows here (that we do not
|
|
|
|
# subsequently reinsert below)? While technically correct it means
|
|
|
|
# we have no record of the fact the user *was* a member of the
|
|
|
|
# room but got, say, state reset out of it.
|
|
|
|
if to_delete or to_insert:
|
|
|
|
txn.executemany(
|
|
|
|
"DELETE FROM local_current_membership"
|
|
|
|
" WHERE room_id = ? AND user_id = ?",
|
|
|
|
(
|
|
|
|
(room_id, state_key)
|
|
|
|
for etype, state_key in itertools.chain(to_delete, to_insert)
|
|
|
|
if etype == EventTypes.Member and self.is_mine_id(state_key)
|
|
|
|
),
|
|
|
|
)
|
|
|
|
|
|
|
|
if to_insert:
|
|
|
|
txn.executemany(
|
|
|
|
"""INSERT INTO local_current_membership
|
|
|
|
(room_id, user_id, event_id, membership)
|
|
|
|
VALUES (?, ?, ?, (SELECT membership FROM room_memberships WHERE event_id = ?))
|
|
|
|
""",
|
|
|
|
[
|
|
|
|
(room_id, key[1], ev_id, ev_id)
|
|
|
|
for key, ev_id in to_insert.items()
|
|
|
|
if key[0] == EventTypes.Member and self.is_mine_id(key[1])
|
|
|
|
],
|
|
|
|
)
|
|
|
|
|
2019-01-30 03:53:17 -07:00
|
|
|
txn.call_after(
|
2020-05-13 06:38:22 -06:00
|
|
|
self.store._curr_state_delta_stream_cache.entity_has_changed,
|
2019-03-28 07:37:16 -06:00
|
|
|
room_id,
|
2019-04-02 05:42:39 -06:00
|
|
|
stream_id,
|
2019-01-30 03:53:17 -07:00
|
|
|
)
|
2017-06-13 02:56:18 -06:00
|
|
|
|
2019-01-30 03:53:17 -07:00
|
|
|
# Invalidate the various caches
|
|
|
|
|
|
|
|
# Figure out the changes of membership to invalidate the
|
|
|
|
# `get_rooms_for_user` cache.
|
|
|
|
# We find out which membership events we may have deleted
|
|
|
|
# and which we have added, then we invlidate the caches for all
|
|
|
|
# those users.
|
2020-02-21 05:15:07 -07:00
|
|
|
members_changed = {
|
2019-01-30 03:53:17 -07:00
|
|
|
state_key
|
|
|
|
for ev_type, state_key in itertools.chain(to_delete, to_insert)
|
|
|
|
if ev_type == EventTypes.Member
|
2020-02-21 05:15:07 -07:00
|
|
|
}
|
2019-01-30 03:53:17 -07:00
|
|
|
|
2019-04-02 05:42:39 -06:00
|
|
|
for member in members_changed:
|
|
|
|
txn.call_after(
|
2020-05-13 06:38:22 -06:00
|
|
|
self.store.get_rooms_for_user_with_stream_ordering.invalidate,
|
|
|
|
(member,),
|
2019-04-02 05:42:39 -06:00
|
|
|
)
|
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
self.store._invalidate_state_caches_and_stream(
|
|
|
|
txn, room_id, members_changed
|
|
|
|
)
|
2019-01-30 03:53:17 -07:00
|
|
|
|
2020-01-29 04:01:32 -07:00
|
|
|
def _upsert_room_version_txn(self, txn: LoggingTransaction, room_id: str):
|
|
|
|
"""Update the room version in the database based off current state
|
|
|
|
events.
|
|
|
|
|
|
|
|
This is used when we're about to delete current state and we want to
|
|
|
|
ensure that the `rooms.room_version` column is up to date.
|
|
|
|
"""
|
|
|
|
|
|
|
|
sql = """
|
|
|
|
SELECT json FROM event_json
|
|
|
|
INNER JOIN current_state_events USING (room_id, event_id)
|
|
|
|
WHERE room_id = ? AND type = ? AND state_key = ?
|
|
|
|
"""
|
|
|
|
txn.execute(sql, (room_id, EventTypes.Create, ""))
|
|
|
|
row = txn.fetchone()
|
|
|
|
if row:
|
2020-07-16 09:32:19 -06:00
|
|
|
event_json = db_to_json(row[0])
|
2020-01-29 04:01:32 -07:00
|
|
|
content = event_json.get("content", {})
|
|
|
|
creator = content.get("creator")
|
|
|
|
room_version_id = content.get("room_version", RoomVersions.V1.identifier)
|
|
|
|
|
2020-08-05 14:38:57 -06:00
|
|
|
self.db_pool.simple_upsert_txn(
|
2020-01-29 04:01:32 -07:00
|
|
|
txn,
|
|
|
|
table="rooms",
|
|
|
|
keyvalues={"room_id": room_id},
|
|
|
|
values={"room_version": room_version_id},
|
|
|
|
insertion_values={"is_public": False, "creator": creator},
|
|
|
|
)
|
|
|
|
|
2019-03-28 07:37:16 -06:00
|
|
|
def _update_forward_extremities_txn(
|
|
|
|
self, txn, new_forward_extremities, max_stream_order
|
|
|
|
):
|
2020-06-15 05:03:36 -06:00
|
|
|
for room_id, new_extrem in new_forward_extremities.items():
|
2020-08-05 14:38:57 -06:00
|
|
|
self.db_pool.simple_delete_txn(
|
2019-03-28 07:37:16 -06:00
|
|
|
txn, table="event_forward_extremities", keyvalues={"room_id": room_id}
|
2017-01-20 07:28:53 -07:00
|
|
|
)
|
2020-05-13 06:38:22 -06:00
|
|
|
txn.call_after(
|
|
|
|
self.store.get_latest_event_ids_in_room.invalidate, (room_id,)
|
|
|
|
)
|
2017-01-20 07:28:53 -07:00
|
|
|
|
2020-08-05 14:38:57 -06:00
|
|
|
self.db_pool.simple_insert_many_txn(
|
2017-01-20 07:28:53 -07:00
|
|
|
txn,
|
|
|
|
table="event_forward_extremities",
|
|
|
|
values=[
|
2019-03-28 07:37:16 -06:00
|
|
|
{"event_id": ev_id, "room_id": room_id}
|
2020-06-15 05:03:36 -06:00
|
|
|
for room_id, new_extrem in new_forward_extremities.items()
|
2017-01-20 07:28:53 -07:00
|
|
|
for ev_id in new_extrem
|
|
|
|
],
|
|
|
|
)
|
|
|
|
# We now insert into stream_ordering_to_exterm a mapping from room_id,
|
|
|
|
# new stream_ordering to new forward extremeties in the room.
|
|
|
|
# This allows us to later efficiently look up the forward extremeties
|
|
|
|
# for a room before a given stream_ordering
|
2020-08-05 14:38:57 -06:00
|
|
|
self.db_pool.simple_insert_many_txn(
|
2017-01-20 07:28:53 -07:00
|
|
|
txn,
|
|
|
|
table="stream_ordering_to_exterm",
|
|
|
|
values=[
|
|
|
|
{
|
|
|
|
"room_id": room_id,
|
|
|
|
"event_id": event_id,
|
|
|
|
"stream_ordering": max_stream_order,
|
|
|
|
}
|
2020-06-15 05:03:36 -06:00
|
|
|
for room_id, new_extrem in new_forward_extremities.items()
|
2017-01-20 07:28:53 -07:00
|
|
|
for event_id in new_extrem
|
2019-03-28 07:37:16 -06:00
|
|
|
],
|
2017-01-20 07:28:53 -07:00
|
|
|
)
|
|
|
|
|
2017-03-17 05:51:13 -06:00
|
|
|
@classmethod
|
2020-09-11 05:22:55 -06:00
|
|
|
def _filter_events_and_contexts_for_duplicates(
|
|
|
|
cls, events_and_contexts: List[Tuple[EventBase, EventContext]]
|
|
|
|
) -> List[Tuple[EventBase, EventContext]]:
|
2017-03-17 05:51:13 -06:00
|
|
|
"""Ensure that we don't have the same event twice.
|
|
|
|
|
|
|
|
Pick the earliest non-outlier if there is one, else the earliest one.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
events_and_contexts (list[(EventBase, EventContext)]):
|
|
|
|
Returns:
|
|
|
|
list[(EventBase, EventContext)]: filtered list
|
|
|
|
"""
|
2020-09-11 05:22:55 -06:00
|
|
|
new_events_and_contexts = (
|
|
|
|
OrderedDict()
|
|
|
|
) # type: OrderedDict[str, Tuple[EventBase, EventContext]]
|
2016-08-03 04:23:39 -06:00
|
|
|
for event, context in events_and_contexts:
|
|
|
|
prev_event_context = new_events_and_contexts.get(event.event_id)
|
|
|
|
if prev_event_context:
|
|
|
|
if not event.internal_metadata.is_outlier():
|
|
|
|
if prev_event_context[0].internal_metadata.is_outlier():
|
|
|
|
# To ensure correct ordering we pop, as OrderedDict is
|
|
|
|
# ordered by first insertion.
|
|
|
|
new_events_and_contexts.pop(event.event_id, None)
|
|
|
|
new_events_and_contexts[event.event_id] = (event, context)
|
|
|
|
else:
|
|
|
|
new_events_and_contexts[event.event_id] = (event, context)
|
2018-05-31 03:03:47 -06:00
|
|
|
return list(new_events_and_contexts.values())
|
2016-08-03 04:23:39 -06:00
|
|
|
|
2020-09-11 05:22:55 -06:00
|
|
|
def _update_room_depths_txn(
|
|
|
|
self,
|
|
|
|
txn,
|
|
|
|
events_and_contexts: List[Tuple[EventBase, EventContext]],
|
|
|
|
backfilled: bool,
|
|
|
|
):
|
2017-03-17 05:51:13 -06:00
|
|
|
"""Update min_depth for each room
|
2016-08-03 04:23:39 -06:00
|
|
|
|
2017-03-17 05:51:13 -06:00
|
|
|
Args:
|
|
|
|
txn (twisted.enterprise.adbapi.Connection): db connection
|
|
|
|
events_and_contexts (list[(EventBase, EventContext)]): events
|
|
|
|
we are persisting
|
|
|
|
backfilled (bool): True if the events were backfilled
|
|
|
|
"""
|
2020-09-11 05:22:55 -06:00
|
|
|
depth_updates = {} # type: Dict[str, int]
|
2016-02-09 09:19:15 -07:00
|
|
|
for event, context in events_and_contexts:
|
|
|
|
# Remove the any existing cache entries for the event_ids
|
2020-05-13 06:38:22 -06:00
|
|
|
txn.call_after(self.store._invalidate_get_event_cache, event.event_id)
|
2016-01-28 08:02:37 -07:00
|
|
|
if not backfilled:
|
|
|
|
txn.call_after(
|
2020-05-13 06:38:22 -06:00
|
|
|
self.store._events_stream_cache.entity_has_changed,
|
2019-03-28 07:37:16 -06:00
|
|
|
event.room_id,
|
|
|
|
event.internal_metadata.stream_ordering,
|
2016-01-28 08:02:37 -07:00
|
|
|
)
|
|
|
|
|
2016-07-25 11:44:30 -06:00
|
|
|
if not event.internal_metadata.is_outlier() and not context.rejected:
|
2016-02-09 09:19:15 -07:00
|
|
|
depth_updates[event.room_id] = max(
|
|
|
|
event.depth, depth_updates.get(event.room_id, event.depth)
|
|
|
|
)
|
|
|
|
|
2020-06-15 05:03:36 -06:00
|
|
|
for room_id, depth in depth_updates.items():
|
2015-06-25 10:18:19 -06:00
|
|
|
self._update_min_depth_for_room_txn(txn, room_id, depth)
|
|
|
|
|
2017-03-17 05:51:13 -06:00
|
|
|
def _update_outliers_txn(self, txn, events_and_contexts):
|
|
|
|
"""Update any outliers with new event info.
|
|
|
|
|
|
|
|
This turns outliers into ex-outliers (unless the new event was
|
|
|
|
rejected).
|
|
|
|
|
|
|
|
Args:
|
|
|
|
txn (twisted.enterprise.adbapi.Connection): db connection
|
|
|
|
events_and_contexts (list[(EventBase, EventContext)]): events
|
|
|
|
we are persisting
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
list[(EventBase, EventContext)] new list, without events which
|
|
|
|
are already in the events table.
|
|
|
|
"""
|
2015-06-25 10:18:19 -06:00
|
|
|
txn.execute(
|
2019-03-28 07:37:16 -06:00
|
|
|
"SELECT event_id, outlier FROM events WHERE event_id in (%s)"
|
|
|
|
% (",".join(["?"] * len(events_and_contexts)),),
|
|
|
|
[event.event_id for event, _ in events_and_contexts],
|
2015-03-20 07:52:56 -06:00
|
|
|
)
|
2016-07-25 11:44:30 -06:00
|
|
|
|
2019-03-28 07:37:16 -06:00
|
|
|
have_persisted = {event_id: outlier for event_id, outlier in txn}
|
2015-06-25 10:18:19 -06:00
|
|
|
|
|
|
|
to_remove = set()
|
|
|
|
for event, context in events_and_contexts:
|
|
|
|
if event.event_id not in have_persisted:
|
|
|
|
continue
|
|
|
|
|
|
|
|
to_remove.add(event)
|
|
|
|
|
2017-03-17 05:51:13 -06:00
|
|
|
if context.rejected:
|
|
|
|
# If the event is rejected then we don't care if the event
|
|
|
|
# was an outlier or not.
|
|
|
|
continue
|
|
|
|
|
2015-06-25 10:18:19 -06:00
|
|
|
outlier_persisted = have_persisted[event.event_id]
|
|
|
|
if not event.internal_metadata.is_outlier() and outlier_persisted:
|
2016-07-26 03:49:52 -06:00
|
|
|
# We received a copy of an event that we had already stored as
|
|
|
|
# an outlier in the database. We now have some state at that
|
|
|
|
# so we need to update the state_groups table with that state.
|
|
|
|
|
2018-02-06 07:31:24 -07:00
|
|
|
# insert into event_to_state_groups.
|
2016-09-02 03:41:38 -06:00
|
|
|
try:
|
2018-02-06 07:31:24 -07:00
|
|
|
self._store_event_state_mappings_txn(txn, ((event, context),))
|
2016-09-02 03:41:38 -06:00
|
|
|
except Exception:
|
|
|
|
logger.exception("")
|
|
|
|
raise
|
2015-03-20 07:52:56 -06:00
|
|
|
|
2020-10-01 06:09:18 -06:00
|
|
|
metadata_json = frozendict_json_encoder.encode(
|
|
|
|
event.internal_metadata.get_dict()
|
|
|
|
)
|
2015-05-12 07:14:58 -06:00
|
|
|
|
2019-11-21 05:00:14 -07:00
|
|
|
sql = "UPDATE event_json SET internal_metadata = ? WHERE event_id = ?"
|
2019-03-28 07:37:16 -06:00
|
|
|
txn.execute(sql, (metadata_json, event.event_id))
|
2015-03-20 07:52:56 -06:00
|
|
|
|
2016-07-26 03:49:52 -06:00
|
|
|
# Add an entry to the ex_outlier_stream table to replicate the
|
|
|
|
# change in outlier status to our workers.
|
2016-03-30 10:19:56 -06:00
|
|
|
stream_order = event.internal_metadata.stream_ordering
|
2016-08-31 03:09:46 -06:00
|
|
|
state_group_id = context.state_group
|
2020-08-05 14:38:57 -06:00
|
|
|
self.db_pool.simple_insert_txn(
|
2016-03-30 10:19:56 -06:00
|
|
|
txn,
|
|
|
|
table="ex_outlier_stream",
|
|
|
|
values={
|
|
|
|
"event_stream_ordering": stream_order,
|
|
|
|
"event_id": event.event_id,
|
|
|
|
"state_group": state_group_id,
|
2020-10-09 06:10:33 -06:00
|
|
|
"instance_name": self._instance_name,
|
2019-03-28 07:37:16 -06:00
|
|
|
},
|
2016-03-30 10:19:56 -06:00
|
|
|
)
|
|
|
|
|
2019-11-21 05:00:14 -07:00
|
|
|
sql = "UPDATE events SET outlier = ? WHERE event_id = ?"
|
2019-03-28 07:37:16 -06:00
|
|
|
txn.execute(sql, (False, event.event_id))
|
2016-07-25 11:44:30 -06:00
|
|
|
|
2016-07-26 03:49:52 -06:00
|
|
|
# Update the event_backward_extremities table now that this
|
|
|
|
# event isn't an outlier any more.
|
2017-01-20 07:40:31 -07:00
|
|
|
self._update_backward_extremeties(txn, [event])
|
2015-09-15 09:34:42 -06:00
|
|
|
|
2019-03-28 07:37:16 -06:00
|
|
|
return [ec for ec in events_and_contexts if ec[0] not in to_remove]
|
2015-03-24 10:20:26 -06:00
|
|
|
|
2017-03-17 05:51:13 -06:00
|
|
|
def _store_event_txn(self, txn, events_and_contexts):
|
|
|
|
"""Insert new events into the event and event_json tables
|
|
|
|
|
|
|
|
Args:
|
|
|
|
txn (twisted.enterprise.adbapi.Connection): db connection
|
|
|
|
events_and_contexts (list[(EventBase, EventContext)]): events
|
|
|
|
we are persisting
|
|
|
|
"""
|
|
|
|
|
|
|
|
if not events_and_contexts:
|
|
|
|
# nothing to do here
|
|
|
|
return
|
|
|
|
|
|
|
|
def event_dict(event):
|
2017-03-24 04:57:02 -06:00
|
|
|
d = event.get_dict()
|
|
|
|
d.pop("redacted", None)
|
|
|
|
d.pop("redacted_because", None)
|
|
|
|
return d
|
2016-08-04 08:02:15 -06:00
|
|
|
|
2020-08-05 14:38:57 -06:00
|
|
|
self.db_pool.simple_insert_many_txn(
|
2015-06-25 10:18:19 -06:00
|
|
|
txn,
|
|
|
|
table="event_json",
|
|
|
|
values=[
|
|
|
|
{
|
|
|
|
"event_id": event.event_id,
|
|
|
|
"room_id": event.room_id,
|
2020-10-01 06:09:18 -06:00
|
|
|
"internal_metadata": frozendict_json_encoder.encode(
|
2015-08-24 09:17:38 -06:00
|
|
|
event.internal_metadata.get_dict()
|
2018-08-01 08:54:06 -06:00
|
|
|
),
|
2020-10-01 06:09:18 -06:00
|
|
|
"json": frozendict_json_encoder.encode(event_dict(event)),
|
2019-01-23 04:11:52 -07:00
|
|
|
"format_version": event.format_version,
|
2015-06-25 10:18:19 -06:00
|
|
|
}
|
|
|
|
for event, _ in events_and_contexts
|
|
|
|
],
|
2015-04-15 03:24:24 -06:00
|
|
|
)
|
2015-03-20 07:52:56 -06:00
|
|
|
|
2020-08-05 14:38:57 -06:00
|
|
|
self.db_pool.simple_insert_many_txn(
|
2015-06-25 10:18:19 -06:00
|
|
|
txn,
|
|
|
|
table="events",
|
|
|
|
values=[
|
|
|
|
{
|
2020-09-14 03:16:41 -06:00
|
|
|
"instance_name": self._instance_name,
|
2015-06-25 10:18:19 -06:00
|
|
|
"stream_ordering": event.internal_metadata.stream_ordering,
|
|
|
|
"topological_ordering": event.depth,
|
|
|
|
"depth": event.depth,
|
|
|
|
"event_id": event.event_id,
|
|
|
|
"room_id": event.room_id,
|
|
|
|
"type": event.type,
|
|
|
|
"processed": True,
|
|
|
|
"outlier": event.internal_metadata.is_outlier(),
|
2015-11-30 10:45:31 -07:00
|
|
|
"origin_server_ts": int(event.origin_server_ts),
|
2016-04-19 07:24:36 -06:00
|
|
|
"received_ts": self._clock.time_msec(),
|
2016-07-14 08:15:22 -06:00
|
|
|
"sender": event.sender,
|
|
|
|
"contains_url": (
|
2020-06-16 06:51:47 -06:00
|
|
|
"url" in event.content and isinstance(event.content["url"], str)
|
2016-07-14 08:15:22 -06:00
|
|
|
),
|
2015-06-25 10:18:19 -06:00
|
|
|
}
|
2020-08-06 10:15:35 -06:00
|
|
|
for event, _ in events_and_contexts
|
2015-06-25 10:18:19 -06:00
|
|
|
],
|
2015-03-20 07:52:56 -06:00
|
|
|
)
|
|
|
|
|
2019-10-01 04:05:48 -06:00
|
|
|
for event, _ in events_and_contexts:
|
|
|
|
if not event.internal_metadata.is_redacted():
|
|
|
|
# If we're persisting an unredacted event we go and ensure
|
|
|
|
# that we mark any redactions that reference this event as
|
|
|
|
# requiring censoring.
|
2020-08-05 14:38:57 -06:00
|
|
|
self.db_pool.simple_update_txn(
|
2019-10-01 04:05:48 -06:00
|
|
|
txn,
|
|
|
|
table="redactions",
|
2019-10-01 06:23:34 -06:00
|
|
|
keyvalues={"redacts": event.event_id},
|
|
|
|
updatevalues={"have_censored": False},
|
2019-10-01 04:05:48 -06:00
|
|
|
)
|
|
|
|
|
2017-03-17 05:51:13 -06:00
|
|
|
def _store_rejected_events_txn(self, txn, events_and_contexts):
|
|
|
|
"""Add rows to the 'rejections' table for received events which were
|
|
|
|
rejected
|
|
|
|
|
|
|
|
Args:
|
|
|
|
txn (twisted.enterprise.adbapi.Connection): db connection
|
|
|
|
events_and_contexts (list[(EventBase, EventContext)]): events
|
|
|
|
we are persisting
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
list[(EventBase, EventContext)] new list, without the rejected
|
|
|
|
events.
|
|
|
|
"""
|
2016-07-26 03:49:52 -06:00
|
|
|
# Remove the rejected events from the list now that we've added them
|
|
|
|
# to the events table and the events_json table.
|
2016-07-25 11:44:30 -06:00
|
|
|
to_remove = set()
|
2016-07-25 09:12:16 -06:00
|
|
|
for event, context in events_and_contexts:
|
|
|
|
if context.rejected:
|
2016-07-26 03:49:52 -06:00
|
|
|
# Insert the event_id into the rejections table
|
2019-03-28 07:37:16 -06:00
|
|
|
self._store_rejections_txn(txn, event.event_id, context.rejected)
|
2016-07-26 03:49:52 -06:00
|
|
|
to_remove.add(event)
|
2016-07-25 11:44:30 -06:00
|
|
|
|
2019-03-28 07:37:16 -06:00
|
|
|
return [ec for ec in events_and_contexts if ec[0] not in to_remove]
|
2016-07-25 11:44:30 -06:00
|
|
|
|
2019-03-28 07:37:16 -06:00
|
|
|
def _update_metadata_tables_txn(
|
|
|
|
self, txn, events_and_contexts, all_events_and_contexts, backfilled
|
|
|
|
):
|
2017-03-17 05:51:13 -06:00
|
|
|
"""Update all the miscellaneous tables for new events
|
|
|
|
|
|
|
|
Args:
|
|
|
|
txn (twisted.enterprise.adbapi.Connection): db connection
|
|
|
|
events_and_contexts (list[(EventBase, EventContext)]): events
|
|
|
|
we are persisting
|
2018-02-20 05:33:04 -07:00
|
|
|
all_events_and_contexts (list[(EventBase, EventContext)]): all
|
|
|
|
events that we were going to persist. This includes events
|
|
|
|
we've already persisted, etc, that wouldn't appear in
|
|
|
|
events_and_context.
|
2017-03-17 05:51:13 -06:00
|
|
|
backfilled (bool): True if the events were backfilled
|
|
|
|
"""
|
|
|
|
|
2018-02-20 05:29:50 -07:00
|
|
|
# Insert all the push actions into the event_push_actions table.
|
|
|
|
self._set_push_actions_for_event_and_users_txn(
|
|
|
|
txn,
|
|
|
|
events_and_contexts=events_and_contexts,
|
2018-02-20 05:33:04 -07:00
|
|
|
all_events_and_contexts=all_events_and_contexts,
|
2018-02-20 05:29:50 -07:00
|
|
|
)
|
|
|
|
|
2016-07-25 11:44:30 -06:00
|
|
|
if not events_and_contexts:
|
2017-03-17 05:51:13 -06:00
|
|
|
# nothing to do here
|
2016-07-25 11:44:30 -06:00
|
|
|
return
|
|
|
|
|
|
|
|
for event, context in events_and_contexts:
|
2016-07-26 04:05:39 -06:00
|
|
|
if event.type == EventTypes.Redaction and event.redacts is not None:
|
|
|
|
# Remove the entries in the event_push_actions table for the
|
|
|
|
# redacted event.
|
|
|
|
self._remove_push_actions_for_event_id_txn(
|
|
|
|
txn, event.room_id, event.redacts
|
|
|
|
)
|
|
|
|
|
2019-05-20 03:13:05 -06:00
|
|
|
# Remove from relations table.
|
|
|
|
self._handle_redaction(txn, event.redacts)
|
|
|
|
|
2016-07-26 03:49:52 -06:00
|
|
|
# Update the event_forward_extremities, event_backward_extremities and
|
|
|
|
# event_edges tables.
|
2016-07-25 11:44:30 -06:00
|
|
|
self._handle_mult_prev_events(
|
2019-03-28 07:37:16 -06:00
|
|
|
txn, events=[event for event, _ in events_and_contexts]
|
2016-07-25 11:44:30 -06:00
|
|
|
)
|
|
|
|
|
|
|
|
for event, _ in events_and_contexts:
|
|
|
|
if event.type == EventTypes.Name:
|
2019-08-21 06:16:28 -06:00
|
|
|
# Insert into the event_search table.
|
2016-07-25 11:44:30 -06:00
|
|
|
self._store_room_name_txn(txn, event)
|
|
|
|
elif event.type == EventTypes.Topic:
|
2019-08-21 06:16:28 -06:00
|
|
|
# Insert into the event_search table.
|
2016-07-25 11:44:30 -06:00
|
|
|
self._store_room_topic_txn(txn, event)
|
|
|
|
elif event.type == EventTypes.Message:
|
2016-07-26 03:49:52 -06:00
|
|
|
# Insert into the event_search table.
|
2016-07-25 11:44:30 -06:00
|
|
|
self._store_room_message_txn(txn, event)
|
2020-01-23 08:19:03 -07:00
|
|
|
elif event.type == EventTypes.Redaction and event.redacts is not None:
|
2016-07-26 03:49:52 -06:00
|
|
|
# Insert into the redactions table.
|
2016-07-25 11:44:30 -06:00
|
|
|
self._store_redaction(txn, event)
|
2019-11-04 10:09:22 -07:00
|
|
|
elif event.type == EventTypes.Retention:
|
|
|
|
# Update the room_retention table.
|
|
|
|
self._store_retention_policy_for_room_txn(txn, event)
|
2016-07-25 11:44:30 -06:00
|
|
|
|
2019-05-14 09:59:21 -06:00
|
|
|
self._handle_event_relations(txn, event)
|
|
|
|
|
2019-10-29 12:35:49 -06:00
|
|
|
# Store the labels for this event.
|
2019-11-01 10:22:44 -06:00
|
|
|
labels = event.content.get(EventContentFields.LABELS)
|
2019-10-29 12:35:49 -06:00
|
|
|
if labels:
|
2019-11-01 05:47:28 -06:00
|
|
|
self.insert_labels_for_event_txn(
|
|
|
|
txn, event.event_id, labels, event.room_id, event.depth
|
|
|
|
)
|
2019-10-29 12:35:49 -06:00
|
|
|
|
2019-12-03 12:19:45 -07:00
|
|
|
if self._ephemeral_messages_enabled:
|
|
|
|
# If there's an expiry timestamp on the event, store it.
|
|
|
|
expiry_ts = event.content.get(EventContentFields.SELF_DESTRUCT_AFTER)
|
|
|
|
if isinstance(expiry_ts, int) and not event.is_state():
|
|
|
|
self._insert_event_expiry_txn(txn, event.event_id, expiry_ts)
|
|
|
|
|
2016-07-26 03:49:52 -06:00
|
|
|
# Insert into the room_memberships table.
|
2016-07-25 11:44:30 -06:00
|
|
|
self._store_room_members_txn(
|
|
|
|
txn,
|
|
|
|
[
|
|
|
|
event
|
|
|
|
for event, _ in events_and_contexts
|
|
|
|
if event.type == EventTypes.Member
|
|
|
|
],
|
|
|
|
backfilled=backfilled,
|
|
|
|
)
|
|
|
|
|
2016-07-26 03:49:52 -06:00
|
|
|
# Insert event_reference_hashes table.
|
2015-06-25 10:18:19 -06:00
|
|
|
self._store_event_reference_hashes_txn(
|
|
|
|
txn, [event for event, _ in events_and_contexts]
|
|
|
|
)
|
|
|
|
|
2016-03-31 08:00:42 -06:00
|
|
|
state_events_and_contexts = [
|
|
|
|
ec for ec in events_and_contexts if ec[0].is_state()
|
|
|
|
]
|
2015-04-27 06:22:30 -06:00
|
|
|
|
2015-06-25 10:18:19 -06:00
|
|
|
state_values = []
|
|
|
|
for event, context in state_events_and_contexts:
|
2015-03-20 07:52:56 -06:00
|
|
|
vals = {
|
|
|
|
"event_id": event.event_id,
|
|
|
|
"room_id": event.room_id,
|
|
|
|
"type": event.type,
|
|
|
|
"state_key": event.state_key,
|
|
|
|
}
|
|
|
|
|
|
|
|
# TODO: How does this work with backfilling?
|
|
|
|
if hasattr(event, "replaces_state"):
|
|
|
|
vals["prev_state"] = event.replaces_state
|
|
|
|
|
2015-06-25 10:18:19 -06:00
|
|
|
state_values.append(vals)
|
2015-03-20 07:52:56 -06:00
|
|
|
|
2020-08-05 14:38:57 -06:00
|
|
|
self.db_pool.simple_insert_many_txn(
|
|
|
|
txn, table="state_events", values=state_values
|
|
|
|
)
|
2015-03-20 07:52:56 -06:00
|
|
|
|
2016-07-26 04:05:39 -06:00
|
|
|
# Prefill the event cache
|
2016-06-06 04:08:12 -06:00
|
|
|
self._add_to_cache(txn, events_and_contexts)
|
|
|
|
|
|
|
|
def _add_to_cache(self, txn, events_and_contexts):
|
|
|
|
to_prefill = []
|
|
|
|
|
|
|
|
rows = []
|
|
|
|
N = 200
|
|
|
|
for i in range(0, len(events_and_contexts), N):
|
2019-03-28 07:37:16 -06:00
|
|
|
ev_map = {e[0].event_id: e[0] for e in events_and_contexts[i : i + N]}
|
2016-06-06 04:08:12 -06:00
|
|
|
if not ev_map:
|
|
|
|
break
|
|
|
|
|
|
|
|
sql = (
|
|
|
|
"SELECT "
|
|
|
|
" e.event_id as event_id, "
|
|
|
|
" r.redacts as redacts,"
|
|
|
|
" rej.event_id as rejects "
|
|
|
|
" FROM events as e"
|
|
|
|
" LEFT JOIN rejections as rej USING (event_id)"
|
|
|
|
" LEFT JOIN redactions as r ON e.event_id = r.redacts"
|
2019-10-02 12:07:07 -06:00
|
|
|
" WHERE "
|
|
|
|
)
|
|
|
|
|
|
|
|
clause, args = make_in_list_sql_clause(
|
|
|
|
self.database_engine, "e.event_id", list(ev_map)
|
|
|
|
)
|
2016-06-06 04:08:12 -06:00
|
|
|
|
2019-10-02 12:07:07 -06:00
|
|
|
txn.execute(sql + clause, args)
|
2020-08-05 14:38:57 -06:00
|
|
|
rows = self.db_pool.cursor_to_dict(txn)
|
2016-06-06 04:08:12 -06:00
|
|
|
for row in rows:
|
|
|
|
event = ev_map[row["event_id"]]
|
|
|
|
if not row["rejects"] and not row["redacts"]:
|
2019-03-28 07:37:16 -06:00
|
|
|
to_prefill.append(
|
|
|
|
_EventCacheEntry(event=event, redacted_event=None)
|
|
|
|
)
|
2016-06-06 04:08:12 -06:00
|
|
|
|
|
|
|
def prefill():
|
|
|
|
for cache_entry in to_prefill:
|
2020-10-19 05:20:29 -06:00
|
|
|
self.store._get_event_cache.set((cache_entry[0].event_id,), cache_entry)
|
2019-03-28 07:37:16 -06:00
|
|
|
|
2016-06-06 04:08:12 -06:00
|
|
|
txn.call_after(prefill)
|
|
|
|
|
2015-03-20 07:52:56 -06:00
|
|
|
def _store_redaction(self, txn, event):
|
|
|
|
# invalidate the cache for the redacted event
|
2020-05-13 06:38:22 -06:00
|
|
|
txn.call_after(self.store._invalidate_get_event_cache, event.redacts)
|
2019-10-01 06:23:34 -06:00
|
|
|
|
2020-08-05 14:38:57 -06:00
|
|
|
self.db_pool.simple_insert_txn(
|
2019-10-01 06:23:34 -06:00
|
|
|
txn,
|
|
|
|
table="redactions",
|
|
|
|
values={
|
|
|
|
"event_id": event.event_id,
|
|
|
|
"redacts": event.redacts,
|
|
|
|
"received_ts": self._clock.time_msec(),
|
|
|
|
},
|
2015-03-20 07:52:56 -06:00
|
|
|
)
|
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
def insert_labels_for_event_txn(
|
|
|
|
self, txn, event_id, labels, room_id, topological_ordering
|
|
|
|
):
|
|
|
|
"""Store the mapping between an event's ID and its labels, with one row per
|
|
|
|
(event_id, label) tuple.
|
2019-08-29 10:38:51 -06:00
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
Args:
|
|
|
|
txn (LoggingTransaction): The transaction to execute.
|
|
|
|
event_id (str): The event's ID.
|
|
|
|
labels (list[str]): A list of text labels.
|
|
|
|
room_id (str): The ID of the room the event was sent to.
|
|
|
|
topological_ordering (int): The position of the event in the room's topology.
|
2019-08-29 10:38:51 -06:00
|
|
|
"""
|
2020-08-05 14:38:57 -06:00
|
|
|
return self.db_pool.simple_insert_many_txn(
|
2020-05-13 06:38:22 -06:00
|
|
|
txn=txn,
|
|
|
|
table="event_labels",
|
|
|
|
values=[
|
|
|
|
{
|
|
|
|
"event_id": event_id,
|
|
|
|
"label": label,
|
|
|
|
"room_id": room_id,
|
|
|
|
"topological_ordering": topological_ordering,
|
|
|
|
}
|
|
|
|
for label in labels
|
|
|
|
],
|
|
|
|
)
|
2019-08-29 10:38:51 -06:00
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
def _insert_event_expiry_txn(self, txn, event_id, expiry_ts):
|
|
|
|
"""Save the expiry timestamp associated with a given event ID.
|
2019-08-29 10:38:51 -06:00
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
Args:
|
|
|
|
txn (LoggingTransaction): The database transaction to use.
|
|
|
|
event_id (str): The event ID the expiry timestamp is associated with.
|
|
|
|
expiry_ts (int): The timestamp at which to expire (delete) the event.
|
2019-08-29 10:38:51 -06:00
|
|
|
"""
|
2020-08-05 14:38:57 -06:00
|
|
|
return self.db_pool.simple_insert_txn(
|
2020-05-13 06:38:22 -06:00
|
|
|
txn=txn,
|
|
|
|
table="event_expiry",
|
|
|
|
values={"event_id": event_id, "expiry_ts": expiry_ts},
|
2019-12-04 06:52:46 -07:00
|
|
|
)
|
2019-08-29 10:38:51 -06:00
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
def _store_event_reference_hashes_txn(self, txn, events):
|
|
|
|
"""Store a hash for a PDU
|
|
|
|
Args:
|
|
|
|
txn (cursor):
|
|
|
|
events (list): list of Events.
|
|
|
|
"""
|
2019-08-29 10:38:51 -06:00
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
vals = []
|
|
|
|
for event in events:
|
|
|
|
ref_alg, ref_hash_bytes = compute_event_reference_hash(event)
|
|
|
|
vals.append(
|
|
|
|
{
|
|
|
|
"event_id": event.event_id,
|
|
|
|
"algorithm": ref_alg,
|
|
|
|
"hash": memoryview(ref_hash_bytes),
|
|
|
|
}
|
2019-08-29 10:38:51 -06:00
|
|
|
)
|
|
|
|
|
2020-08-05 14:38:57 -06:00
|
|
|
self.db_pool.simple_insert_many_txn(
|
|
|
|
txn, table="event_reference_hashes", values=vals
|
|
|
|
)
|
2019-08-29 10:38:51 -06:00
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
def _store_room_members_txn(self, txn, events, backfilled):
|
|
|
|
"""Store a room member in the database.
|
2019-12-03 12:19:45 -07:00
|
|
|
"""
|
2020-09-23 09:42:14 -06:00
|
|
|
|
|
|
|
def str_or_none(val: Any) -> Optional[str]:
|
|
|
|
return val if isinstance(val, str) else None
|
|
|
|
|
2020-08-05 14:38:57 -06:00
|
|
|
self.db_pool.simple_insert_many_txn(
|
2019-12-03 12:19:45 -07:00
|
|
|
txn,
|
2020-05-13 06:38:22 -06:00
|
|
|
table="room_memberships",
|
|
|
|
values=[
|
|
|
|
{
|
|
|
|
"event_id": event.event_id,
|
|
|
|
"user_id": event.state_key,
|
|
|
|
"sender": event.user_id,
|
|
|
|
"room_id": event.room_id,
|
|
|
|
"membership": event.membership,
|
2020-09-23 09:42:14 -06:00
|
|
|
"display_name": str_or_none(event.content.get("displayname")),
|
|
|
|
"avatar_url": str_or_none(event.content.get("avatar_url")),
|
2020-05-13 06:38:22 -06:00
|
|
|
}
|
|
|
|
for event in events
|
|
|
|
],
|
2019-12-03 12:19:45 -07:00
|
|
|
)
|
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
for event in events:
|
|
|
|
txn.call_after(
|
|
|
|
self.store._membership_stream_cache.entity_has_changed,
|
|
|
|
event.state_key,
|
|
|
|
event.internal_metadata.stream_ordering,
|
2016-03-01 07:49:41 -07:00
|
|
|
)
|
2020-05-13 06:38:22 -06:00
|
|
|
txn.call_after(
|
|
|
|
self.store.get_invited_rooms_for_local_user.invalidate,
|
|
|
|
(event.state_key,),
|
2016-03-01 07:49:41 -07:00
|
|
|
)
|
2016-03-30 10:19:56 -06:00
|
|
|
|
2020-07-07 07:20:40 -06:00
|
|
|
# We update the local_current_membership table only if the event is
|
|
|
|
# "current", i.e., its something that has just happened.
|
|
|
|
#
|
|
|
|
# This will usually get updated by the `current_state_events` handling,
|
|
|
|
# unless its an outlier, and an outlier is only "current" if it's an "out of
|
|
|
|
# band membership", like a remote invite or a rejection of a remote invite.
|
|
|
|
if (
|
|
|
|
self.is_mine_id(event.state_key)
|
|
|
|
and not backfilled
|
|
|
|
and event.internal_metadata.is_outlier()
|
|
|
|
and event.internal_metadata.is_out_of_band_membership()
|
|
|
|
):
|
2020-08-05 14:38:57 -06:00
|
|
|
self.db_pool.simple_upsert_txn(
|
2020-07-07 07:20:40 -06:00
|
|
|
txn,
|
|
|
|
table="local_current_membership",
|
|
|
|
keyvalues={"room_id": event.room_id, "user_id": event.state_key},
|
|
|
|
values={
|
|
|
|
"event_id": event.event_id,
|
|
|
|
"membership": event.membership,
|
|
|
|
},
|
|
|
|
)
|
2016-07-05 03:28:51 -06:00
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
def _handle_event_relations(self, txn, event):
|
|
|
|
"""Handles inserting relation data during peristence of events
|
2019-10-30 09:12:49 -06:00
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
Args:
|
|
|
|
txn
|
|
|
|
event (EventBase)
|
2016-07-04 09:02:50 -06:00
|
|
|
"""
|
2020-05-13 06:38:22 -06:00
|
|
|
relation = event.content.get("m.relates_to")
|
|
|
|
if not relation:
|
|
|
|
# No relations
|
|
|
|
return
|
2016-07-04 09:02:50 -06:00
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
rel_type = relation.get("rel_type")
|
|
|
|
if rel_type not in (
|
|
|
|
RelationTypes.ANNOTATION,
|
|
|
|
RelationTypes.REFERENCE,
|
|
|
|
RelationTypes.REPLACE,
|
|
|
|
):
|
|
|
|
# Unknown relation type
|
|
|
|
return
|
2016-07-05 03:28:51 -06:00
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
parent_id = relation.get("event_id")
|
|
|
|
if not parent_id:
|
|
|
|
# Invalid relation
|
|
|
|
return
|
2018-02-14 09:41:12 -07:00
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
aggregation_key = relation.get("key")
|
2018-02-14 09:41:12 -07:00
|
|
|
|
2020-08-05 14:38:57 -06:00
|
|
|
self.db_pool.simple_insert_txn(
|
2020-05-13 06:38:22 -06:00
|
|
|
txn,
|
|
|
|
table="event_relations",
|
|
|
|
values={
|
|
|
|
"event_id": event.event_id,
|
|
|
|
"relates_to_id": parent_id,
|
|
|
|
"relation_type": rel_type,
|
|
|
|
"aggregation_key": aggregation_key,
|
|
|
|
},
|
2018-02-14 09:41:12 -07:00
|
|
|
)
|
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
txn.call_after(self.store.get_relations_for_event.invalidate_many, (parent_id,))
|
|
|
|
txn.call_after(
|
|
|
|
self.store.get_aggregation_groups_for_event.invalidate_many, (parent_id,)
|
2016-07-04 09:02:50 -06:00
|
|
|
)
|
2018-02-14 04:02:22 -07:00
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
if rel_type == RelationTypes.REPLACE:
|
|
|
|
txn.call_after(self.store.get_applicable_edit.invalidate, (parent_id,))
|
2018-02-14 04:02:22 -07:00
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
def _handle_redaction(self, txn, redacted_event_id):
|
|
|
|
"""Handles receiving a redaction and checking whether we need to remove
|
|
|
|
any redacted relations from the database.
|
2018-09-13 08:05:52 -06:00
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
Args:
|
|
|
|
txn
|
|
|
|
redacted_event_id (str): The event that was redacted.
|
|
|
|
"""
|
2018-09-13 08:05:52 -06:00
|
|
|
|
2020-08-05 14:38:57 -06:00
|
|
|
self.db_pool.simple_delete_txn(
|
2020-05-13 06:38:22 -06:00
|
|
|
txn, table="event_relations", keyvalues={"event_id": redacted_event_id}
|
2018-09-13 08:05:52 -06:00
|
|
|
)
|
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
def _store_room_topic_txn(self, txn, event):
|
|
|
|
if hasattr(event, "content") and "topic" in event.content:
|
|
|
|
self.store_event_search_txn(
|
|
|
|
txn, event, "content.topic", event.content["topic"]
|
|
|
|
)
|
2016-09-02 03:41:38 -06:00
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
def _store_room_name_txn(self, txn, event):
|
|
|
|
if hasattr(event, "content") and "name" in event.content:
|
|
|
|
self.store_event_search_txn(
|
|
|
|
txn, event, "content.name", event.content["name"]
|
|
|
|
)
|
2017-05-10 10:46:41 -06:00
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
def _store_room_message_txn(self, txn, event):
|
|
|
|
if hasattr(event, "content") and "body" in event.content:
|
|
|
|
self.store_event_search_txn(
|
|
|
|
txn, event, "content.body", event.content["body"]
|
|
|
|
)
|
2016-07-04 09:02:50 -06:00
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
def _store_retention_policy_for_room_txn(self, txn, event):
|
2020-10-14 05:00:52 -06:00
|
|
|
if not event.is_state():
|
|
|
|
logger.debug("Ignoring non-state m.room.retention event")
|
|
|
|
return
|
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
if hasattr(event, "content") and (
|
|
|
|
"min_lifetime" in event.content or "max_lifetime" in event.content
|
|
|
|
):
|
|
|
|
if (
|
|
|
|
"min_lifetime" in event.content
|
2020-06-16 06:51:47 -06:00
|
|
|
and not isinstance(event.content.get("min_lifetime"), int)
|
2020-05-13 06:38:22 -06:00
|
|
|
) or (
|
|
|
|
"max_lifetime" in event.content
|
2020-06-16 06:51:47 -06:00
|
|
|
and not isinstance(event.content.get("max_lifetime"), int)
|
2020-05-13 06:38:22 -06:00
|
|
|
):
|
|
|
|
# Ignore the event if one of the value isn't an integer.
|
|
|
|
return
|
2017-05-10 10:46:41 -06:00
|
|
|
|
2020-08-05 14:38:57 -06:00
|
|
|
self.db_pool.simple_insert_txn(
|
2020-05-13 06:38:22 -06:00
|
|
|
txn=txn,
|
|
|
|
table="room_retention",
|
|
|
|
values={
|
|
|
|
"room_id": event.room_id,
|
|
|
|
"event_id": event.event_id,
|
|
|
|
"min_lifetime": event.content.get("min_lifetime"),
|
|
|
|
"max_lifetime": event.content.get("max_lifetime"),
|
|
|
|
},
|
|
|
|
)
|
2016-07-15 07:23:15 -06:00
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
self.store._invalidate_cache_and_stream(
|
|
|
|
txn, self.store.get_retention_policy_for_room, (event.room_id,)
|
|
|
|
)
|
2016-07-15 07:23:15 -06:00
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
def store_event_search_txn(self, txn, event, key, value):
|
|
|
|
"""Add event to the search table
|
2017-05-10 10:46:41 -06:00
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
Args:
|
|
|
|
txn (cursor):
|
|
|
|
event (EventBase):
|
|
|
|
key (str):
|
|
|
|
value (str):
|
2019-03-28 07:37:16 -06:00
|
|
|
"""
|
2020-05-13 06:38:22 -06:00
|
|
|
self.store.store_search_entries_txn(
|
|
|
|
txn,
|
|
|
|
(
|
|
|
|
SearchEntry(
|
|
|
|
key=key,
|
|
|
|
value=value,
|
|
|
|
event_id=event.event_id,
|
|
|
|
room_id=event.room_id,
|
|
|
|
stream_ordering=event.internal_metadata.stream_ordering,
|
|
|
|
origin_server_ts=event.origin_server_ts,
|
|
|
|
),
|
|
|
|
),
|
2019-03-28 07:37:16 -06:00
|
|
|
)
|
2016-09-05 07:49:08 -06:00
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
def _set_push_actions_for_event_and_users_txn(
|
|
|
|
self, txn, events_and_contexts, all_events_and_contexts
|
|
|
|
):
|
|
|
|
"""Handles moving push actions from staging table to main
|
|
|
|
event_push_actions table for all events in `events_and_contexts`.
|
2018-10-04 08:18:52 -06:00
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
Also ensures that all events in `all_events_and_contexts` are removed
|
|
|
|
from the push action staging area.
|
2016-07-04 09:02:50 -06:00
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
Args:
|
|
|
|
events_and_contexts (list[(EventBase, EventContext)]): events
|
|
|
|
we are persisting
|
|
|
|
all_events_and_contexts (list[(EventBase, EventContext)]): all
|
|
|
|
events that we were going to persist. This includes events
|
|
|
|
we've already persisted, etc, that wouldn't appear in
|
|
|
|
events_and_context.
|
|
|
|
"""
|
2017-05-10 10:46:41 -06:00
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
sql = """
|
|
|
|
INSERT INTO event_push_actions (
|
|
|
|
room_id, event_id, user_id, actions, stream_ordering,
|
2020-09-02 10:19:37 -06:00
|
|
|
topological_ordering, notif, highlight, unread
|
2016-07-04 09:02:50 -06:00
|
|
|
)
|
2020-09-02 10:19:37 -06:00
|
|
|
SELECT ?, event_id, user_id, actions, ?, ?, notif, highlight, unread
|
2020-05-13 06:38:22 -06:00
|
|
|
FROM event_push_actions_staging
|
|
|
|
WHERE event_id = ?
|
|
|
|
"""
|
2016-07-04 09:02:50 -06:00
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
if events_and_contexts:
|
|
|
|
txn.executemany(
|
|
|
|
sql,
|
|
|
|
(
|
|
|
|
(
|
|
|
|
event.room_id,
|
|
|
|
event.internal_metadata.stream_ordering,
|
|
|
|
event.depth,
|
|
|
|
event.event_id,
|
|
|
|
)
|
|
|
|
for event, _ in events_and_contexts
|
|
|
|
),
|
|
|
|
)
|
2018-02-14 08:44:51 -07:00
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
for event, _ in events_and_contexts:
|
2020-08-05 14:38:57 -06:00
|
|
|
user_ids = self.db_pool.simple_select_onecol_txn(
|
2020-05-13 06:38:22 -06:00
|
|
|
txn,
|
|
|
|
table="event_push_actions_staging",
|
|
|
|
keyvalues={"event_id": event.event_id},
|
|
|
|
retcol="user_id",
|
2018-02-14 08:44:51 -07:00
|
|
|
)
|
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
for uid in user_ids:
|
|
|
|
txn.call_after(
|
|
|
|
self.store.get_unread_event_push_actions_by_room_for_user.invalidate_many,
|
|
|
|
(event.room_id, uid),
|
|
|
|
)
|
|
|
|
|
|
|
|
# Now we delete the staging area for *all* events that were being
|
|
|
|
# persisted.
|
|
|
|
txn.executemany(
|
|
|
|
"DELETE FROM event_push_actions_staging WHERE event_id = ?",
|
|
|
|
((event.event_id,) for event, _ in all_events_and_contexts),
|
2018-02-14 04:02:22 -07:00
|
|
|
)
|
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
def _remove_push_actions_for_event_id_txn(self, txn, room_id, event_id):
|
|
|
|
# Sad that we have to blow away the cache for the whole room here
|
|
|
|
txn.call_after(
|
|
|
|
self.store.get_unread_event_push_actions_by_room_for_user.invalidate_many,
|
2019-03-28 07:37:16 -06:00
|
|
|
(room_id,),
|
|
|
|
)
|
2018-02-09 05:13:34 -07:00
|
|
|
txn.execute(
|
2020-05-13 06:38:22 -06:00
|
|
|
"DELETE FROM event_push_actions WHERE room_id = ? AND event_id = ?",
|
|
|
|
(room_id, event_id),
|
2018-02-09 05:13:34 -07:00
|
|
|
)
|
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
def _store_rejections_txn(self, txn, event_id, reason):
|
2020-08-05 14:38:57 -06:00
|
|
|
self.db_pool.simple_insert_txn(
|
2020-05-13 06:38:22 -06:00
|
|
|
txn,
|
|
|
|
table="rejections",
|
|
|
|
values={
|
|
|
|
"event_id": event_id,
|
|
|
|
"reason": reason,
|
|
|
|
"last_check": self._clock.time_msec(),
|
|
|
|
},
|
|
|
|
)
|
2019-08-22 03:42:59 -06:00
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
def _store_event_state_mappings_txn(
|
|
|
|
self, txn, events_and_contexts: Iterable[Tuple[EventBase, EventContext]]
|
|
|
|
):
|
|
|
|
state_groups = {}
|
|
|
|
for event, context in events_and_contexts:
|
|
|
|
if event.internal_metadata.is_outlier():
|
|
|
|
continue
|
2019-11-04 06:36:57 -07:00
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
# if the event was rejected, just give it the same state as its
|
|
|
|
# predecessor.
|
|
|
|
if context.rejected:
|
|
|
|
state_groups[event.event_id] = context.state_group_before_event
|
|
|
|
continue
|
2019-08-22 03:42:59 -06:00
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
state_groups[event.event_id] = context.state_group
|
2019-08-22 03:42:59 -06:00
|
|
|
|
2020-08-05 14:38:57 -06:00
|
|
|
self.db_pool.simple_insert_many_txn(
|
2020-05-13 06:38:22 -06:00
|
|
|
txn,
|
|
|
|
table="event_to_state_groups",
|
|
|
|
values=[
|
|
|
|
{"state_group": state_group_id, "event_id": event_id}
|
2020-06-15 05:03:36 -06:00
|
|
|
for event_id, state_group_id in state_groups.items()
|
2020-05-13 06:38:22 -06:00
|
|
|
],
|
2019-08-22 03:42:59 -06:00
|
|
|
)
|
|
|
|
|
2020-06-15 05:03:36 -06:00
|
|
|
for event_id, state_group_id in state_groups.items():
|
2020-05-13 06:38:22 -06:00
|
|
|
txn.call_after(
|
|
|
|
self.store._get_state_group_for_event.prefill,
|
|
|
|
(event_id,),
|
|
|
|
state_group_id,
|
2019-08-22 03:42:59 -06:00
|
|
|
)
|
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
def _update_min_depth_for_room_txn(self, txn, room_id, depth):
|
|
|
|
min_depth = self.store._get_min_depth_interaction(txn, room_id)
|
2019-08-22 03:42:59 -06:00
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
if min_depth is not None and depth >= min_depth:
|
|
|
|
return
|
2017-04-11 10:34:09 -06:00
|
|
|
|
2020-08-05 14:38:57 -06:00
|
|
|
self.db_pool.simple_upsert_txn(
|
2020-05-13 06:38:22 -06:00
|
|
|
txn,
|
|
|
|
table="room_depth",
|
|
|
|
keyvalues={"room_id": room_id},
|
|
|
|
values={"min_depth": depth},
|
2017-04-11 10:34:09 -06:00
|
|
|
)
|
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
def _handle_mult_prev_events(self, txn, events):
|
2019-11-01 04:30:51 -06:00
|
|
|
"""
|
2020-05-13 06:38:22 -06:00
|
|
|
For the given event, update the event edges table and forward and
|
|
|
|
backward extremities tables.
|
|
|
|
"""
|
2020-08-05 14:38:57 -06:00
|
|
|
self.db_pool.simple_insert_many_txn(
|
2020-05-13 06:38:22 -06:00
|
|
|
txn,
|
|
|
|
table="event_edges",
|
2019-11-01 05:47:28 -06:00
|
|
|
values=[
|
|
|
|
{
|
2020-05-13 06:38:22 -06:00
|
|
|
"event_id": ev.event_id,
|
|
|
|
"prev_event_id": e_id,
|
|
|
|
"room_id": ev.room_id,
|
|
|
|
"is_state": False,
|
2019-11-01 05:47:28 -06:00
|
|
|
}
|
2020-05-13 06:38:22 -06:00
|
|
|
for ev in events
|
|
|
|
for e_id in ev.prev_event_ids()
|
2019-11-01 05:47:28 -06:00
|
|
|
],
|
2019-10-29 12:35:49 -06:00
|
|
|
)
|
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
self._update_backward_extremeties(txn, events)
|
2019-12-03 12:19:45 -07:00
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
def _update_backward_extremeties(self, txn, events):
|
|
|
|
"""Updates the event_backward_extremities tables based on the new/updated
|
|
|
|
events being persisted.
|
2019-12-03 12:19:45 -07:00
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
This is called for new events *and* for events that were outliers, but
|
|
|
|
are now being persisted as non-outliers.
|
2019-12-03 12:19:45 -07:00
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
Forward extremities are handled when we first start persisting the events.
|
2019-12-03 12:19:45 -07:00
|
|
|
"""
|
2020-09-11 05:22:55 -06:00
|
|
|
events_by_room = {} # type: Dict[str, List[EventBase]]
|
2020-05-13 06:38:22 -06:00
|
|
|
for ev in events:
|
|
|
|
events_by_room.setdefault(ev.room_id, []).append(ev)
|
|
|
|
|
|
|
|
query = (
|
|
|
|
"INSERT INTO event_backward_extremities (event_id, room_id)"
|
|
|
|
" SELECT ?, ? WHERE NOT EXISTS ("
|
|
|
|
" SELECT 1 FROM event_backward_extremities"
|
|
|
|
" WHERE event_id = ? AND room_id = ?"
|
|
|
|
" )"
|
|
|
|
" AND NOT EXISTS ("
|
|
|
|
" SELECT 1 FROM events WHERE event_id = ? AND room_id = ? "
|
|
|
|
" AND outlier = ?"
|
|
|
|
" )"
|
2019-12-03 12:19:45 -07:00
|
|
|
)
|
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
txn.executemany(
|
|
|
|
query,
|
|
|
|
[
|
|
|
|
(e_id, ev.room_id, e_id, ev.room_id, e_id, ev.room_id, False)
|
|
|
|
for ev in events
|
|
|
|
for e_id in ev.prev_event_ids()
|
|
|
|
if not ev.internal_metadata.is_outlier()
|
|
|
|
],
|
2019-12-03 12:19:45 -07:00
|
|
|
)
|
|
|
|
|
2020-05-13 06:38:22 -06:00
|
|
|
query = (
|
|
|
|
"DELETE FROM event_backward_extremities"
|
|
|
|
" WHERE event_id = ? AND room_id = ?"
|
|
|
|
)
|
|
|
|
txn.executemany(
|
|
|
|
query,
|
|
|
|
[
|
|
|
|
(ev.event_id, ev.room_id)
|
|
|
|
for ev in events
|
|
|
|
if not ev.internal_metadata.is_outlier()
|
|
|
|
],
|
|
|
|
)
|