2018-08-07 08:22:40 -06:00
|
|
|
# -*- coding: utf-8 -*-
|
|
|
|
# Copyright 2018 New Vector Ltd
|
|
|
|
#
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
|
|
|
|
|
|
|
import hashlib
|
|
|
|
import logging
|
2020-07-24 08:59:51 -06:00
|
|
|
from typing import Awaitable, Callable, Dict, List, Optional
|
2018-08-07 08:22:40 -06:00
|
|
|
|
|
|
|
from synapse import event_auth
|
2019-04-01 03:24:38 -06:00
|
|
|
from synapse.api.constants import EventTypes
|
2018-08-07 08:22:40 -06:00
|
|
|
from synapse.api.errors import AuthError
|
2019-04-01 03:24:38 -06:00
|
|
|
from synapse.api.room_versions import RoomVersions
|
2019-12-13 05:55:32 -07:00
|
|
|
from synapse.events import EventBase
|
2020-01-16 06:31:22 -07:00
|
|
|
from synapse.types import StateMap
|
2018-08-07 08:22:40 -06:00
|
|
|
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
|
|
|
|
|
|
|
POWER_KEY = (EventTypes.PowerLevels, "")
|
|
|
|
|
|
|
|
|
2020-07-24 08:59:51 -06:00
|
|
|
async def resolve_events_with_store(
|
2019-12-13 05:55:32 -07:00
|
|
|
room_id: str,
|
2020-01-16 06:31:22 -07:00
|
|
|
state_sets: List[StateMap[str]],
|
2019-12-13 05:55:32 -07:00
|
|
|
event_map: Optional[Dict[str, EventBase]],
|
2020-07-24 08:59:51 -06:00
|
|
|
state_map_factory: Callable[[List[str]], Awaitable],
|
2019-12-13 05:55:32 -07:00
|
|
|
):
|
2018-08-07 08:22:40 -06:00
|
|
|
"""
|
|
|
|
Args:
|
2019-12-13 05:55:32 -07:00
|
|
|
room_id: the room we are working in
|
|
|
|
|
|
|
|
state_sets: List of dicts of (type, state_key) -> event_id,
|
2018-08-07 08:22:40 -06:00
|
|
|
which are the different state groups to resolve.
|
|
|
|
|
2019-12-13 05:55:32 -07:00
|
|
|
event_map:
|
2018-08-07 08:22:40 -06:00
|
|
|
a dict from event_id to event, for any events that we happen to
|
|
|
|
have in flight (eg, those currently being persisted). This will be
|
|
|
|
used as a starting point fof finding the state we need; any missing
|
|
|
|
events will be requested via state_map_factory.
|
|
|
|
|
|
|
|
If None, all events will be fetched via state_map_factory.
|
|
|
|
|
2019-12-13 05:55:32 -07:00
|
|
|
state_map_factory: will be called
|
2018-08-07 08:22:40 -06:00
|
|
|
with a list of event_ids that are needed, and should return with
|
2020-07-24 08:59:51 -06:00
|
|
|
an Awaitable that resolves to a dict of event_id to event.
|
2018-08-07 08:22:40 -06:00
|
|
|
|
2019-12-13 05:55:32 -07:00
|
|
|
Returns:
|
2018-08-07 08:22:40 -06:00
|
|
|
Deferred[dict[(str, str), str]]:
|
|
|
|
a map from (type, state_key) to event_id.
|
|
|
|
"""
|
|
|
|
if len(state_sets) == 1:
|
2019-07-23 07:00:55 -06:00
|
|
|
return state_sets[0]
|
2018-08-07 08:22:40 -06:00
|
|
|
|
2019-06-20 03:32:02 -06:00
|
|
|
unconflicted_state, conflicted_state = _seperate(state_sets)
|
2018-08-07 08:22:40 -06:00
|
|
|
|
2020-02-21 05:15:07 -07:00
|
|
|
needed_events = {
|
2020-06-15 05:03:36 -06:00
|
|
|
event_id for event_ids in conflicted_state.values() for event_id in event_ids
|
2020-02-21 05:15:07 -07:00
|
|
|
}
|
2018-09-26 00:56:06 -06:00
|
|
|
needed_event_count = len(needed_events)
|
2018-08-07 08:22:40 -06:00
|
|
|
if event_map is not None:
|
2020-06-15 05:03:36 -06:00
|
|
|
needed_events -= set(event_map.keys())
|
2018-08-07 08:22:40 -06:00
|
|
|
|
2018-09-26 00:56:06 -06:00
|
|
|
logger.info(
|
2019-06-20 03:32:02 -06:00
|
|
|
"Asking for %d/%d conflicted events", len(needed_events), needed_event_count
|
2018-09-26 00:56:06 -06:00
|
|
|
)
|
2018-08-07 08:22:40 -06:00
|
|
|
|
|
|
|
# dict[str, FrozenEvent]: a map from state event id to event. Only includes
|
|
|
|
# the state events which are in conflict (and those in event_map)
|
2020-07-24 08:59:51 -06:00
|
|
|
state_map = await state_map_factory(needed_events)
|
2018-08-07 08:22:40 -06:00
|
|
|
if event_map is not None:
|
|
|
|
state_map.update(event_map)
|
|
|
|
|
2019-12-13 05:55:32 -07:00
|
|
|
# everything in the state map should be in the right room
|
|
|
|
for event in state_map.values():
|
|
|
|
if event.room_id != room_id:
|
|
|
|
raise Exception(
|
|
|
|
"Attempting to state-resolve for room %s with event %s which is in %s"
|
|
|
|
% (room_id, event.event_id, event.room_id,)
|
|
|
|
)
|
|
|
|
|
2018-08-07 08:22:40 -06:00
|
|
|
# get the ids of the auth events which allow us to authenticate the
|
|
|
|
# conflicted state, picking only from the unconflicting state.
|
|
|
|
#
|
|
|
|
# dict[(str, str), str]: a map from state key to event id
|
|
|
|
auth_events = _create_auth_events_from_maps(
|
|
|
|
unconflicted_state, conflicted_state, state_map
|
|
|
|
)
|
|
|
|
|
2020-06-15 05:03:36 -06:00
|
|
|
new_needed_events = set(auth_events.values())
|
2018-09-26 00:56:06 -06:00
|
|
|
new_needed_event_count = len(new_needed_events)
|
2018-08-07 08:22:40 -06:00
|
|
|
new_needed_events -= needed_events
|
|
|
|
if event_map is not None:
|
2020-06-15 05:03:36 -06:00
|
|
|
new_needed_events -= set(event_map.keys())
|
2018-08-07 08:22:40 -06:00
|
|
|
|
2018-09-26 00:56:06 -06:00
|
|
|
logger.info(
|
2019-06-20 03:32:02 -06:00
|
|
|
"Asking for %d/%d auth events", len(new_needed_events), new_needed_event_count
|
2018-09-26 00:56:06 -06:00
|
|
|
)
|
2018-08-07 08:22:40 -06:00
|
|
|
|
2020-07-24 08:59:51 -06:00
|
|
|
state_map_new = await state_map_factory(new_needed_events)
|
2019-12-13 05:55:32 -07:00
|
|
|
for event in state_map_new.values():
|
|
|
|
if event.room_id != room_id:
|
|
|
|
raise Exception(
|
|
|
|
"Attempting to state-resolve for room %s with event %s which is in %s"
|
|
|
|
% (room_id, event.event_id, event.room_id,)
|
|
|
|
)
|
|
|
|
|
2018-08-07 08:22:40 -06:00
|
|
|
state_map.update(state_map_new)
|
|
|
|
|
2019-07-23 07:00:55 -06:00
|
|
|
return _resolve_with_state(
|
|
|
|
unconflicted_state, conflicted_state, auth_events, state_map
|
2019-06-20 03:32:02 -06:00
|
|
|
)
|
2018-08-07 08:22:40 -06:00
|
|
|
|
|
|
|
|
|
|
|
def _seperate(state_sets):
|
|
|
|
"""Takes the state_sets and figures out which keys are conflicted and
|
|
|
|
which aren't. i.e., which have multiple different event_ids associated
|
|
|
|
with them in different state sets.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
state_sets(iterable[dict[(str, str), str]]):
|
|
|
|
List of dicts of (type, state_key) -> event_id, which are the
|
|
|
|
different state groups to resolve.
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
(dict[(str, str), str], dict[(str, str), set[str]]):
|
|
|
|
A tuple of (unconflicted_state, conflicted_state), where:
|
|
|
|
|
|
|
|
unconflicted_state is a dict mapping (type, state_key)->event_id
|
|
|
|
for unconflicted state keys.
|
|
|
|
|
|
|
|
conflicted_state is a dict mapping (type, state_key) to a set of
|
|
|
|
event ids for conflicted state keys.
|
|
|
|
"""
|
|
|
|
state_set_iterator = iter(state_sets)
|
|
|
|
unconflicted_state = dict(next(state_set_iterator))
|
|
|
|
conflicted_state = {}
|
|
|
|
|
|
|
|
for state_set in state_set_iterator:
|
2020-06-15 05:03:36 -06:00
|
|
|
for key, value in state_set.items():
|
2018-08-07 08:22:40 -06:00
|
|
|
# Check if there is an unconflicted entry for the state key.
|
|
|
|
unconflicted_value = unconflicted_state.get(key)
|
|
|
|
if unconflicted_value is None:
|
|
|
|
# There isn't an unconflicted entry so check if there is a
|
|
|
|
# conflicted entry.
|
|
|
|
ls = conflicted_state.get(key)
|
|
|
|
if ls is None:
|
|
|
|
# There wasn't a conflicted entry so haven't seen this key before.
|
|
|
|
# Therefore it isn't conflicted yet.
|
|
|
|
unconflicted_state[key] = value
|
|
|
|
else:
|
|
|
|
# This key is already conflicted, add our value to the conflict set.
|
|
|
|
ls.add(value)
|
|
|
|
elif unconflicted_value != value:
|
|
|
|
# If the unconflicted value is not the same as our value then we
|
|
|
|
# have a new conflict. So move the key from the unconflicted_state
|
|
|
|
# to the conflicted state.
|
|
|
|
conflicted_state[key] = {value, unconflicted_value}
|
|
|
|
unconflicted_state.pop(key, None)
|
|
|
|
|
|
|
|
return unconflicted_state, conflicted_state
|
|
|
|
|
|
|
|
|
|
|
|
def _create_auth_events_from_maps(unconflicted_state, conflicted_state, state_map):
|
|
|
|
auth_events = {}
|
2020-06-15 05:03:36 -06:00
|
|
|
for event_ids in conflicted_state.values():
|
2018-08-07 08:22:40 -06:00
|
|
|
for event_id in event_ids:
|
|
|
|
if event_id in state_map:
|
|
|
|
keys = event_auth.auth_types_for_event(state_map[event_id])
|
|
|
|
for key in keys:
|
|
|
|
if key not in auth_events:
|
|
|
|
event_id = unconflicted_state.get(key, None)
|
|
|
|
if event_id:
|
|
|
|
auth_events[key] = event_id
|
|
|
|
return auth_events
|
|
|
|
|
|
|
|
|
2019-06-20 03:32:02 -06:00
|
|
|
def _resolve_with_state(
|
|
|
|
unconflicted_state_ids, conflicted_state_ids, auth_event_ids, state_map
|
|
|
|
):
|
2018-08-07 08:22:40 -06:00
|
|
|
conflicted_state = {}
|
2020-06-15 05:03:36 -06:00
|
|
|
for key, event_ids in conflicted_state_ids.items():
|
2018-08-07 08:22:40 -06:00
|
|
|
events = [state_map[ev_id] for ev_id in event_ids if ev_id in state_map]
|
|
|
|
if len(events) > 1:
|
|
|
|
conflicted_state[key] = events
|
|
|
|
elif len(events) == 1:
|
|
|
|
unconflicted_state_ids[key] = events[0].event_id
|
|
|
|
|
|
|
|
auth_events = {
|
|
|
|
key: state_map[ev_id]
|
2020-06-15 05:03:36 -06:00
|
|
|
for key, ev_id in auth_event_ids.items()
|
2018-08-07 08:22:40 -06:00
|
|
|
if ev_id in state_map
|
|
|
|
}
|
|
|
|
|
|
|
|
try:
|
2019-06-20 03:32:02 -06:00
|
|
|
resolved_state = _resolve_state_events(conflicted_state, auth_events)
|
2018-08-07 08:22:40 -06:00
|
|
|
except Exception:
|
|
|
|
logger.exception("Failed to resolve state")
|
|
|
|
raise
|
|
|
|
|
|
|
|
new_state = unconflicted_state_ids
|
2020-06-15 05:03:36 -06:00
|
|
|
for key, event in resolved_state.items():
|
2018-08-07 08:22:40 -06:00
|
|
|
new_state[key] = event.event_id
|
|
|
|
|
|
|
|
return new_state
|
|
|
|
|
|
|
|
|
|
|
|
def _resolve_state_events(conflicted_state, auth_events):
|
|
|
|
""" This is where we actually decide which of the conflicted state to
|
|
|
|
use.
|
|
|
|
|
|
|
|
We resolve conflicts in the following order:
|
|
|
|
1. power levels
|
|
|
|
2. join rules
|
|
|
|
3. memberships
|
|
|
|
4. other events.
|
|
|
|
"""
|
|
|
|
resolved_state = {}
|
|
|
|
if POWER_KEY in conflicted_state:
|
|
|
|
events = conflicted_state[POWER_KEY]
|
|
|
|
logger.debug("Resolving conflicted power levels %r", events)
|
2019-06-20 03:32:02 -06:00
|
|
|
resolved_state[POWER_KEY] = _resolve_auth_events(events, auth_events)
|
2018-08-07 08:22:40 -06:00
|
|
|
|
|
|
|
auth_events.update(resolved_state)
|
|
|
|
|
2020-06-15 05:03:36 -06:00
|
|
|
for key, events in conflicted_state.items():
|
2018-08-07 08:22:40 -06:00
|
|
|
if key[0] == EventTypes.JoinRules:
|
|
|
|
logger.debug("Resolving conflicted join rules %r", events)
|
2019-06-20 03:32:02 -06:00
|
|
|
resolved_state[key] = _resolve_auth_events(events, auth_events)
|
2018-08-07 08:22:40 -06:00
|
|
|
|
|
|
|
auth_events.update(resolved_state)
|
|
|
|
|
2020-06-15 05:03:36 -06:00
|
|
|
for key, events in conflicted_state.items():
|
2018-08-07 08:22:40 -06:00
|
|
|
if key[0] == EventTypes.Member:
|
|
|
|
logger.debug("Resolving conflicted member lists %r", events)
|
2019-06-20 03:32:02 -06:00
|
|
|
resolved_state[key] = _resolve_auth_events(events, auth_events)
|
2018-08-07 08:22:40 -06:00
|
|
|
|
|
|
|
auth_events.update(resolved_state)
|
|
|
|
|
2020-06-15 05:03:36 -06:00
|
|
|
for key, events in conflicted_state.items():
|
2018-08-07 08:22:40 -06:00
|
|
|
if key not in resolved_state:
|
|
|
|
logger.debug("Resolving conflicted state %r:%r", key, events)
|
2019-06-20 03:32:02 -06:00
|
|
|
resolved_state[key] = _resolve_normal_events(events, auth_events)
|
2018-08-07 08:22:40 -06:00
|
|
|
|
|
|
|
return resolved_state
|
|
|
|
|
|
|
|
|
|
|
|
def _resolve_auth_events(events, auth_events):
|
2020-02-21 05:15:07 -07:00
|
|
|
reverse = list(reversed(_ordered_events(events)))
|
2018-08-07 08:22:40 -06:00
|
|
|
|
2020-02-21 05:15:07 -07:00
|
|
|
auth_keys = {
|
2019-06-20 03:32:02 -06:00
|
|
|
key for event in events for key in event_auth.auth_types_for_event(event)
|
2020-02-21 05:15:07 -07:00
|
|
|
}
|
2018-08-07 08:22:40 -06:00
|
|
|
|
|
|
|
new_auth_events = {}
|
|
|
|
for key in auth_keys:
|
|
|
|
auth_event = auth_events.get(key, None)
|
|
|
|
if auth_event:
|
|
|
|
new_auth_events[key] = auth_event
|
|
|
|
|
|
|
|
auth_events = new_auth_events
|
|
|
|
|
|
|
|
prev_event = reverse[0]
|
|
|
|
for event in reverse[1:]:
|
|
|
|
auth_events[(prev_event.type, prev_event.state_key)] = prev_event
|
|
|
|
try:
|
|
|
|
# The signatures have already been checked at this point
|
2019-01-25 11:31:41 -07:00
|
|
|
event_auth.check(
|
2020-01-28 07:18:29 -07:00
|
|
|
RoomVersions.V1,
|
2019-04-01 03:24:38 -06:00
|
|
|
event,
|
|
|
|
auth_events,
|
2019-01-25 11:31:41 -07:00
|
|
|
do_sig_check=False,
|
|
|
|
do_size_check=False,
|
|
|
|
)
|
2018-08-07 08:22:40 -06:00
|
|
|
prev_event = event
|
|
|
|
except AuthError:
|
|
|
|
return prev_event
|
|
|
|
|
|
|
|
return event
|
|
|
|
|
|
|
|
|
|
|
|
def _resolve_normal_events(events, auth_events):
|
|
|
|
for event in _ordered_events(events):
|
|
|
|
try:
|
|
|
|
# The signatures have already been checked at this point
|
2019-01-25 11:31:41 -07:00
|
|
|
event_auth.check(
|
2020-01-28 07:18:29 -07:00
|
|
|
RoomVersions.V1,
|
2019-04-01 03:24:38 -06:00
|
|
|
event,
|
|
|
|
auth_events,
|
2019-01-25 11:31:41 -07:00
|
|
|
do_sig_check=False,
|
|
|
|
do_size_check=False,
|
|
|
|
)
|
2018-08-07 08:22:40 -06:00
|
|
|
return event
|
|
|
|
except AuthError:
|
|
|
|
pass
|
|
|
|
|
|
|
|
# Use the last event (the one with the least depth) if they all fail
|
|
|
|
# the auth check.
|
|
|
|
return event
|
|
|
|
|
|
|
|
|
|
|
|
def _ordered_events(events):
|
|
|
|
def key_func(e):
|
2018-12-03 03:47:48 -07:00
|
|
|
# we have to use utf-8 rather than ascii here because it turns out we allow
|
|
|
|
# people to send us events with non-ascii event IDs :/
|
2019-06-20 03:32:02 -06:00
|
|
|
return -int(e.depth), hashlib.sha1(e.event_id.encode("utf-8")).hexdigest()
|
2018-08-07 08:22:40 -06:00
|
|
|
|
|
|
|
return sorted(events, key=key_func)
|