2021-08-13 13:37:24 -06:00
|
|
|
# Copyright 2016 OpenMarket Ltd
|
|
|
|
#
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
|
|
|
|
|
|
|
import logging
|
|
|
|
import re
|
2021-09-15 02:30:58 -06:00
|
|
|
from http import HTTPStatus
|
2021-10-08 17:35:00 -06:00
|
|
|
from typing import TYPE_CHECKING, Awaitable, Tuple
|
2021-09-01 09:59:32 -06:00
|
|
|
|
|
|
|
from twisted.web.server import Request
|
2021-08-13 13:37:24 -06:00
|
|
|
|
2021-10-08 17:35:00 -06:00
|
|
|
from synapse.api.constants import EventContentFields
|
2021-08-13 13:37:24 -06:00
|
|
|
from synapse.api.errors import AuthError, Codes, SynapseError
|
2021-09-01 09:59:32 -06:00
|
|
|
from synapse.http.server import HttpServer
|
2021-08-13 13:37:24 -06:00
|
|
|
from synapse.http.servlet import (
|
|
|
|
RestServlet,
|
|
|
|
assert_params_in_dict,
|
|
|
|
parse_json_object_from_request,
|
|
|
|
parse_string,
|
|
|
|
parse_strings_from_args,
|
|
|
|
)
|
2021-09-01 09:59:32 -06:00
|
|
|
from synapse.http.site import SynapseRequest
|
2021-08-13 13:37:24 -06:00
|
|
|
from synapse.rest.client.transactions import HttpTransactionCache
|
2021-10-08 17:35:00 -06:00
|
|
|
from synapse.types import JsonDict
|
2021-08-13 13:37:24 -06:00
|
|
|
|
2021-09-01 09:59:32 -06:00
|
|
|
if TYPE_CHECKING:
|
|
|
|
from synapse.server import HomeServer
|
|
|
|
|
2021-08-13 13:37:24 -06:00
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
|
|
|
|
|
|
|
class RoomBatchSendEventRestServlet(RestServlet):
|
|
|
|
"""
|
2021-09-21 14:06:28 -06:00
|
|
|
API endpoint which can insert a batch of events historically back in time
|
2021-08-13 13:37:24 -06:00
|
|
|
next to the given `prev_event`.
|
|
|
|
|
2021-09-21 14:06:28 -06:00
|
|
|
`batch_id` comes from `next_batch_id `in the response of the batch send
|
|
|
|
endpoint and is derived from the "insertion" events added to each batch.
|
2021-08-13 13:37:24 -06:00
|
|
|
It's not required for the first batch send.
|
|
|
|
|
|
|
|
`state_events_at_start` is used to define the historical state events
|
|
|
|
needed to auth the events like join events. These events will float
|
|
|
|
outside of the normal DAG as outlier's and won't be visible in the chat
|
2021-09-21 14:06:28 -06:00
|
|
|
history which also allows us to insert multiple batches without having a bunch
|
|
|
|
of `@mxid joined the room` noise between each batch.
|
2021-08-13 13:37:24 -06:00
|
|
|
|
2021-09-21 14:06:28 -06:00
|
|
|
`events` is chronological list of events you want to insert.
|
|
|
|
There is a reverse-chronological constraint on batches so once you insert
|
2021-08-13 13:37:24 -06:00
|
|
|
some messages, you can only insert older ones after that.
|
2021-09-21 14:06:28 -06:00
|
|
|
tldr; Insert batches from your most recent history -> oldest history.
|
2021-08-13 13:37:24 -06:00
|
|
|
|
2021-09-21 14:06:28 -06:00
|
|
|
POST /_matrix/client/unstable/org.matrix.msc2716/rooms/<roomID>/batch_send?prev_event_id=<eventID>&batch_id=<batchID>
|
2021-08-13 13:37:24 -06:00
|
|
|
{
|
|
|
|
"events": [ ... ],
|
|
|
|
"state_events_at_start": [ ... ]
|
|
|
|
}
|
|
|
|
"""
|
|
|
|
|
|
|
|
PATTERNS = (
|
|
|
|
re.compile(
|
|
|
|
"^/_matrix/client/unstable/org.matrix.msc2716"
|
|
|
|
"/rooms/(?P<room_id>[^/]*)/batch_send$"
|
|
|
|
),
|
|
|
|
)
|
|
|
|
|
2021-09-01 09:59:32 -06:00
|
|
|
def __init__(self, hs: "HomeServer"):
|
2021-08-13 13:37:24 -06:00
|
|
|
super().__init__()
|
2022-02-23 04:04:02 -07:00
|
|
|
self.store = hs.get_datastores().main
|
2021-08-13 13:37:24 -06:00
|
|
|
self.event_creation_handler = hs.get_event_creation_handler()
|
|
|
|
self.auth = hs.get_auth()
|
2021-10-08 17:35:00 -06:00
|
|
|
self.room_batch_handler = hs.get_room_batch_handler()
|
2021-08-13 13:37:24 -06:00
|
|
|
self.txns = HttpTransactionCache(hs)
|
|
|
|
|
2021-09-01 09:59:32 -06:00
|
|
|
async def on_POST(
|
|
|
|
self, request: SynapseRequest, room_id: str
|
|
|
|
) -> Tuple[int, JsonDict]:
|
2021-08-13 13:37:24 -06:00
|
|
|
requester = await self.auth.get_user_by_req(request, allow_guest=False)
|
|
|
|
|
|
|
|
if not requester.app_service:
|
|
|
|
raise AuthError(
|
2021-09-15 02:30:58 -06:00
|
|
|
HTTPStatus.FORBIDDEN,
|
2021-08-13 13:37:24 -06:00
|
|
|
"Only application services can use the /batchsend endpoint",
|
|
|
|
)
|
|
|
|
|
|
|
|
body = parse_json_object_from_request(request)
|
|
|
|
assert_params_in_dict(body, ["state_events_at_start", "events"])
|
|
|
|
|
2021-09-01 09:59:32 -06:00
|
|
|
assert request.args is not None
|
2021-09-21 07:10:01 -06:00
|
|
|
prev_event_ids_from_query = parse_strings_from_args(
|
|
|
|
request.args, "prev_event_id"
|
|
|
|
)
|
2021-09-21 14:06:28 -06:00
|
|
|
batch_id_from_query = parse_string(request, "batch_id")
|
2021-08-13 13:37:24 -06:00
|
|
|
|
2021-09-21 07:10:01 -06:00
|
|
|
if prev_event_ids_from_query is None:
|
2021-08-13 13:37:24 -06:00
|
|
|
raise SynapseError(
|
2021-09-15 02:30:58 -06:00
|
|
|
HTTPStatus.BAD_REQUEST,
|
2021-08-13 13:37:24 -06:00
|
|
|
"prev_event query parameter is required when inserting historical messages back in time",
|
|
|
|
errcode=Codes.MISSING_PARAM,
|
|
|
|
)
|
|
|
|
|
2022-10-27 03:52:23 -06:00
|
|
|
if await self.store.is_partial_state_room(room_id):
|
|
|
|
raise SynapseError(
|
|
|
|
HTTPStatus.BAD_REQUEST,
|
|
|
|
"Cannot insert history batches until we have fully joined the room",
|
|
|
|
errcode=Codes.UNABLE_DUE_TO_PARTIAL_STATE,
|
|
|
|
)
|
|
|
|
|
2021-10-08 17:35:00 -06:00
|
|
|
# Verify the batch_id_from_query corresponds to an actual insertion event
|
|
|
|
# and have the batch connected.
|
|
|
|
if batch_id_from_query:
|
|
|
|
corresponding_insertion_event_id = (
|
2021-11-08 20:21:10 -07:00
|
|
|
await self.store.get_insertion_event_id_by_batch_id(
|
2021-10-08 17:35:00 -06:00
|
|
|
room_id, batch_id_from_query
|
|
|
|
)
|
|
|
|
)
|
|
|
|
if corresponding_insertion_event_id is None:
|
|
|
|
raise SynapseError(
|
|
|
|
HTTPStatus.BAD_REQUEST,
|
|
|
|
"No insertion event corresponds to the given ?batch_id",
|
|
|
|
errcode=Codes.INVALID_PARAM,
|
|
|
|
)
|
|
|
|
|
2022-03-29 07:56:25 -06:00
|
|
|
# Make sure that the prev_event_ids exist and aren't outliers - ie, they are
|
|
|
|
# regular parts of the room DAG where we know the state.
|
|
|
|
non_outlier_prev_events = await self.store.have_events_in_timeline(
|
|
|
|
prev_event_ids_from_query
|
|
|
|
)
|
|
|
|
for prev_event_id in prev_event_ids_from_query:
|
|
|
|
if prev_event_id not in non_outlier_prev_events:
|
|
|
|
raise SynapseError(
|
|
|
|
HTTPStatus.BAD_REQUEST,
|
|
|
|
"prev_event %s does not exist, or is an outlier" % (prev_event_id,),
|
|
|
|
errcode=Codes.INVALID_PARAM,
|
|
|
|
)
|
|
|
|
|
2021-09-21 07:10:01 -06:00
|
|
|
# For the event we are inserting next to (`prev_event_ids_from_query`),
|
2022-03-25 08:21:06 -06:00
|
|
|
# find the most recent state events that allowed that message to be
|
|
|
|
# sent. We will use that as a base to auth our historical messages
|
|
|
|
# against.
|
|
|
|
state_event_ids = await self.room_batch_handler.get_most_recent_full_state_ids_from_event_id_list(
|
2021-10-08 17:35:00 -06:00
|
|
|
prev_event_ids_from_query
|
2021-08-13 13:37:24 -06:00
|
|
|
)
|
|
|
|
|
2021-11-03 02:13:51 -06:00
|
|
|
state_event_ids_at_start = []
|
2021-10-08 17:35:00 -06:00
|
|
|
# Create and persist all of the state events that float off on their own
|
|
|
|
# before the batch. These will most likely be all of the invite/member
|
|
|
|
# state events used to auth the upcoming historical messages.
|
2021-11-03 02:13:51 -06:00
|
|
|
if body["state_events_at_start"]:
|
|
|
|
state_event_ids_at_start = (
|
|
|
|
await self.room_batch_handler.persist_state_events_at_start(
|
|
|
|
state_events_at_start=body["state_events_at_start"],
|
|
|
|
room_id=room_id,
|
2022-03-25 08:21:06 -06:00
|
|
|
initial_state_event_ids=state_event_ids,
|
2021-11-03 02:13:51 -06:00
|
|
|
app_service_requester=requester,
|
|
|
|
)
|
2021-08-13 13:37:24 -06:00
|
|
|
)
|
2021-11-03 02:13:51 -06:00
|
|
|
# Update our ongoing auth event ID list with all of the new state we
|
|
|
|
# just created
|
2022-03-25 08:21:06 -06:00
|
|
|
state_event_ids.extend(state_event_ids_at_start)
|
2021-08-13 13:37:24 -06:00
|
|
|
|
2021-10-08 17:35:00 -06:00
|
|
|
inherited_depth = await self.room_batch_handler.inherit_depth_from_prev_ids(
|
2021-09-21 07:10:01 -06:00
|
|
|
prev_event_ids_from_query
|
2021-08-13 13:37:24 -06:00
|
|
|
)
|
|
|
|
|
2021-10-08 17:35:00 -06:00
|
|
|
events_to_create = body["events"]
|
|
|
|
|
2021-09-21 14:06:28 -06:00
|
|
|
# Figure out which batch to connect to. If they passed in
|
|
|
|
# batch_id_from_query let's use it. The batch ID passed in comes
|
|
|
|
# from the batch_id in the "insertion" event from the previous batch.
|
|
|
|
last_event_in_batch = events_to_create[-1]
|
2021-08-13 13:37:24 -06:00
|
|
|
base_insertion_event = None
|
2021-09-21 14:06:28 -06:00
|
|
|
if batch_id_from_query:
|
2021-10-08 17:35:00 -06:00
|
|
|
batch_id_to_connect_to = batch_id_from_query
|
2021-08-13 13:37:24 -06:00
|
|
|
# Otherwise, create an insertion event to act as a starting point.
|
|
|
|
#
|
|
|
|
# We don't always have an insertion event to start hanging more history
|
|
|
|
# off of (ideally there would be one in the main DAG, but that's not the
|
|
|
|
# case if we're wanting to add history to e.g. existing rooms without
|
|
|
|
# an insertion event), in which case we just create a new insertion event
|
|
|
|
# that can then get pointed to by a "marker" event later.
|
|
|
|
else:
|
2021-10-08 17:35:00 -06:00
|
|
|
base_insertion_event_dict = (
|
|
|
|
self.room_batch_handler.create_insertion_event_dict(
|
|
|
|
sender=requester.user.to_string(),
|
|
|
|
room_id=room_id,
|
|
|
|
origin_server_ts=last_event_in_batch["origin_server_ts"],
|
|
|
|
)
|
2021-08-13 13:37:24 -06:00
|
|
|
)
|
2021-10-13 16:44:00 -06:00
|
|
|
base_insertion_event_dict["prev_events"] = prev_event_ids_from_query.copy()
|
2021-08-13 13:37:24 -06:00
|
|
|
|
|
|
|
(
|
|
|
|
base_insertion_event,
|
|
|
|
_,
|
|
|
|
) = await self.event_creation_handler.create_and_send_nonmember_event(
|
2021-10-08 17:35:00 -06:00
|
|
|
await self.room_batch_handler.create_requester_for_user_id_from_app_service(
|
2021-08-13 13:37:24 -06:00
|
|
|
base_insertion_event_dict["sender"],
|
|
|
|
requester.app_service,
|
|
|
|
),
|
|
|
|
base_insertion_event_dict,
|
|
|
|
prev_event_ids=base_insertion_event_dict.get("prev_events"),
|
2022-03-25 08:21:06 -06:00
|
|
|
# Also set the explicit state here because we want to resolve
|
|
|
|
# any `state_events_at_start` here too. It's not strictly
|
|
|
|
# necessary to accomplish anything but if someone asks for the
|
|
|
|
# state at this point, we probably want to show them the
|
|
|
|
# historical state that was part of this batch.
|
|
|
|
state_event_ids=state_event_ids,
|
2021-08-13 13:37:24 -06:00
|
|
|
historical=True,
|
|
|
|
depth=inherited_depth,
|
|
|
|
)
|
|
|
|
|
2021-11-02 07:55:52 -06:00
|
|
|
batch_id_to_connect_to = base_insertion_event.content[
|
2021-09-21 14:06:28 -06:00
|
|
|
EventContentFields.MSC2716_NEXT_BATCH_ID
|
2021-08-13 13:37:24 -06:00
|
|
|
]
|
|
|
|
|
2021-10-08 17:35:00 -06:00
|
|
|
# Create and persist all of the historical events as well as insertion
|
|
|
|
# and batch meta events to make the batch navigable in the DAG.
|
|
|
|
event_ids, next_batch_id = await self.room_batch_handler.handle_batch_of_events(
|
|
|
|
events_to_create=events_to_create,
|
2021-08-13 13:37:24 -06:00
|
|
|
room_id=room_id,
|
2021-10-08 17:35:00 -06:00
|
|
|
batch_id_to_connect_to=batch_id_to_connect_to,
|
|
|
|
inherited_depth=inherited_depth,
|
2022-03-25 08:21:06 -06:00
|
|
|
initial_state_event_ids=state_event_ids,
|
2021-10-08 17:35:00 -06:00
|
|
|
app_service_requester=requester,
|
2021-08-13 13:37:24 -06:00
|
|
|
)
|
|
|
|
|
2021-09-15 02:30:58 -06:00
|
|
|
insertion_event_id = event_ids[0]
|
2021-09-21 14:06:28 -06:00
|
|
|
batch_event_id = event_ids[-1]
|
2021-09-15 02:30:58 -06:00
|
|
|
historical_event_ids = event_ids[1:-1]
|
2021-08-13 13:37:24 -06:00
|
|
|
|
2021-09-15 02:30:58 -06:00
|
|
|
response_dict = {
|
|
|
|
"state_event_ids": state_event_ids_at_start,
|
|
|
|
"event_ids": historical_event_ids,
|
2021-10-08 17:35:00 -06:00
|
|
|
"next_batch_id": next_batch_id,
|
2021-09-15 02:30:58 -06:00
|
|
|
"insertion_event_id": insertion_event_id,
|
2021-09-21 14:06:28 -06:00
|
|
|
"batch_event_id": batch_event_id,
|
2021-08-13 13:37:24 -06:00
|
|
|
}
|
2021-09-15 02:30:58 -06:00
|
|
|
if base_insertion_event is not None:
|
|
|
|
response_dict["base_insertion_event_id"] = base_insertion_event.event_id
|
|
|
|
|
|
|
|
return HTTPStatus.OK, response_dict
|
2021-08-13 13:37:24 -06:00
|
|
|
|
2021-09-01 09:59:32 -06:00
|
|
|
def on_GET(self, request: Request, room_id: str) -> Tuple[int, str]:
|
2021-09-15 02:30:58 -06:00
|
|
|
return HTTPStatus.NOT_IMPLEMENTED, "Not implemented"
|
2021-08-13 13:37:24 -06:00
|
|
|
|
2021-09-01 09:59:32 -06:00
|
|
|
def on_PUT(
|
|
|
|
self, request: SynapseRequest, room_id: str
|
|
|
|
) -> Awaitable[Tuple[int, JsonDict]]:
|
2021-08-13 13:37:24 -06:00
|
|
|
return self.txns.fetch_or_execute_request(
|
|
|
|
request, self.on_POST, request, room_id
|
|
|
|
)
|
|
|
|
|
|
|
|
|
2021-09-01 09:59:32 -06:00
|
|
|
def register_servlets(hs: "HomeServer", http_server: HttpServer) -> None:
|
2021-08-13 13:37:24 -06:00
|
|
|
msc2716_enabled = hs.config.experimental.msc2716_enabled
|
|
|
|
|
|
|
|
if msc2716_enabled:
|
|
|
|
RoomBatchSendEventRestServlet(hs).register(http_server)
|