2019-05-14 09:59:21 -06:00
|
|
|
# -*- coding: utf-8 -*-
|
|
|
|
# Copyright 2019 New Vector Ltd
|
|
|
|
#
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
|
|
|
|
|
|
|
"""This class implements the proposed relation APIs from MSC 1849.
|
|
|
|
|
|
|
|
Since the MSC has not been approved all APIs here are unstable and may change at
|
|
|
|
any time to reflect changes in the MSC.
|
|
|
|
"""
|
|
|
|
|
|
|
|
import logging
|
|
|
|
|
2019-05-14 09:59:21 -06:00
|
|
|
from synapse.api.constants import EventTypes, RelationTypes
|
2019-05-14 09:59:21 -06:00
|
|
|
from synapse.api.errors import SynapseError
|
|
|
|
from synapse.http.servlet import (
|
|
|
|
RestServlet,
|
2019-05-14 09:59:21 -06:00
|
|
|
parse_integer,
|
2019-05-14 09:59:21 -06:00
|
|
|
parse_json_object_from_request,
|
|
|
|
parse_string,
|
|
|
|
)
|
2019-05-15 10:28:33 -06:00
|
|
|
from synapse.rest.client.transactions import HttpTransactionCache
|
2019-06-20 03:32:02 -06:00
|
|
|
from synapse.storage.relations import (
|
|
|
|
AggregationPaginationToken,
|
2019-07-18 07:41:42 -06:00
|
|
|
PaginationChunk,
|
2019-06-20 03:32:02 -06:00
|
|
|
RelationPaginationToken,
|
|
|
|
)
|
2019-05-14 09:59:21 -06:00
|
|
|
|
2019-06-03 05:28:59 -06:00
|
|
|
from ._base import client_patterns
|
2019-05-14 09:59:21 -06:00
|
|
|
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
|
|
|
|
|
|
|
class RelationSendServlet(RestServlet):
|
|
|
|
"""Helper API for sending events that have relation data.
|
|
|
|
|
|
|
|
Example API shape to send a 👍 reaction to a room:
|
|
|
|
|
|
|
|
POST /rooms/!foo/send_relation/$bar/m.annotation/m.reaction?key=%F0%9F%91%8D
|
|
|
|
{}
|
|
|
|
|
|
|
|
{
|
|
|
|
"event_id": "$foobar"
|
|
|
|
}
|
|
|
|
"""
|
|
|
|
|
|
|
|
PATTERN = (
|
|
|
|
"/rooms/(?P<room_id>[^/]*)/send_relation"
|
|
|
|
"/(?P<parent_id>[^/]*)/(?P<relation_type>[^/]*)/(?P<event_type>[^/]*)"
|
|
|
|
)
|
|
|
|
|
|
|
|
def __init__(self, hs):
|
|
|
|
super(RelationSendServlet, self).__init__()
|
|
|
|
self.auth = hs.get_auth()
|
|
|
|
self.event_creation_handler = hs.get_event_creation_handler()
|
2019-05-15 10:28:33 -06:00
|
|
|
self.txns = HttpTransactionCache(hs)
|
2019-05-14 09:59:21 -06:00
|
|
|
|
|
|
|
def register(self, http_server):
|
|
|
|
http_server.register_paths(
|
|
|
|
"POST",
|
2019-06-03 05:28:59 -06:00
|
|
|
client_patterns(self.PATTERN + "$", releases=()),
|
2019-05-14 09:59:21 -06:00
|
|
|
self.on_PUT_or_POST,
|
2019-07-24 06:07:35 -06:00
|
|
|
self.__class__.__name__,
|
2019-05-14 09:59:21 -06:00
|
|
|
)
|
|
|
|
http_server.register_paths(
|
|
|
|
"PUT",
|
2019-06-03 05:28:59 -06:00
|
|
|
client_patterns(self.PATTERN + "/(?P<txn_id>[^/]*)$", releases=()),
|
2019-05-15 10:28:33 -06:00
|
|
|
self.on_PUT,
|
2019-07-24 06:07:35 -06:00
|
|
|
self.__class__.__name__,
|
2019-05-15 10:28:33 -06:00
|
|
|
)
|
|
|
|
|
|
|
|
def on_PUT(self, request, *args, **kwargs):
|
|
|
|
return self.txns.fetch_or_execute_request(
|
|
|
|
request, self.on_PUT_or_POST, request, *args, **kwargs
|
2019-05-14 09:59:21 -06:00
|
|
|
)
|
|
|
|
|
2019-12-05 09:46:37 -07:00
|
|
|
async def on_PUT_or_POST(
|
2019-05-14 09:59:21 -06:00
|
|
|
self, request, room_id, parent_id, relation_type, event_type, txn_id=None
|
|
|
|
):
|
2019-12-05 09:46:37 -07:00
|
|
|
requester = await self.auth.get_user_by_req(request, allow_guest=True)
|
2019-05-14 09:59:21 -06:00
|
|
|
|
|
|
|
if event_type == EventTypes.Member:
|
|
|
|
# Add relations to a membership is meaningless, so we just deny it
|
|
|
|
# at the CS API rather than trying to handle it correctly.
|
|
|
|
raise SynapseError(400, "Cannot send member events with relations")
|
|
|
|
|
|
|
|
content = parse_json_object_from_request(request)
|
|
|
|
|
|
|
|
aggregation_key = parse_string(request, "key", encoding="utf-8")
|
|
|
|
|
|
|
|
content["m.relates_to"] = {
|
|
|
|
"event_id": parent_id,
|
|
|
|
"key": aggregation_key,
|
|
|
|
"rel_type": relation_type,
|
|
|
|
}
|
|
|
|
|
|
|
|
event_dict = {
|
|
|
|
"type": event_type,
|
|
|
|
"content": content,
|
|
|
|
"room_id": room_id,
|
|
|
|
"sender": requester.user.to_string(),
|
|
|
|
}
|
|
|
|
|
2020-05-22 07:21:54 -06:00
|
|
|
event, _ = await self.event_creation_handler.create_and_send_nonmember_event(
|
2019-05-14 09:59:21 -06:00
|
|
|
requester, event_dict=event_dict, txn_id=txn_id
|
|
|
|
)
|
|
|
|
|
2019-08-30 09:28:26 -06:00
|
|
|
return 200, {"event_id": event.event_id}
|
2019-05-14 09:59:21 -06:00
|
|
|
|
|
|
|
|
2019-05-14 09:59:21 -06:00
|
|
|
class RelationPaginationServlet(RestServlet):
|
|
|
|
"""API to paginate relations on an event by topological ordering, optionally
|
|
|
|
filtered by relation type and event type.
|
|
|
|
"""
|
|
|
|
|
2019-06-03 05:28:59 -06:00
|
|
|
PATTERNS = client_patterns(
|
2019-05-14 09:59:21 -06:00
|
|
|
"/rooms/(?P<room_id>[^/]*)/relations/(?P<parent_id>[^/]*)"
|
|
|
|
"(/(?P<relation_type>[^/]*)(/(?P<event_type>[^/]*))?)?$",
|
|
|
|
releases=(),
|
|
|
|
)
|
|
|
|
|
|
|
|
def __init__(self, hs):
|
|
|
|
super(RelationPaginationServlet, self).__init__()
|
|
|
|
self.auth = hs.get_auth()
|
|
|
|
self.store = hs.get_datastore()
|
|
|
|
self.clock = hs.get_clock()
|
|
|
|
self._event_serializer = hs.get_event_client_serializer()
|
2019-05-16 07:19:06 -06:00
|
|
|
self.event_handler = hs.get_event_handler()
|
2019-05-14 09:59:21 -06:00
|
|
|
|
2019-12-05 09:46:37 -07:00
|
|
|
async def on_GET(
|
|
|
|
self, request, room_id, parent_id, relation_type=None, event_type=None
|
|
|
|
):
|
|
|
|
requester = await self.auth.get_user_by_req(request, allow_guest=True)
|
2019-05-14 09:59:21 -06:00
|
|
|
|
2020-02-18 16:14:57 -07:00
|
|
|
await self.auth.check_user_in_room_or_world_readable(
|
|
|
|
room_id, requester.user.to_string(), allow_departed_users=True
|
2019-05-14 09:59:21 -06:00
|
|
|
)
|
|
|
|
|
2019-07-09 06:43:08 -06:00
|
|
|
# This gets the original event and checks that a) the event exists and
|
|
|
|
# b) the user is allowed to view it.
|
2019-12-05 09:46:37 -07:00
|
|
|
event = await self.event_handler.get_event(requester.user, room_id, parent_id)
|
2019-05-16 07:19:06 -06:00
|
|
|
|
2019-05-14 09:59:21 -06:00
|
|
|
limit = parse_integer(request, "limit", default=5)
|
2019-05-14 09:59:21 -06:00
|
|
|
from_token = parse_string(request, "from")
|
|
|
|
to_token = parse_string(request, "to")
|
2019-05-14 09:59:21 -06:00
|
|
|
|
2019-07-18 07:41:42 -06:00
|
|
|
if event.internal_metadata.is_redacted():
|
|
|
|
# If the event is redacted, return an empty list of relations
|
|
|
|
pagination_chunk = PaginationChunk(chunk=[])
|
|
|
|
else:
|
|
|
|
# Return the relations
|
|
|
|
if from_token:
|
|
|
|
from_token = RelationPaginationToken.from_string(from_token)
|
|
|
|
|
|
|
|
if to_token:
|
|
|
|
to_token = RelationPaginationToken.from_string(to_token)
|
|
|
|
|
2019-12-05 09:46:37 -07:00
|
|
|
pagination_chunk = await self.store.get_relations_for_event(
|
2019-07-18 07:41:42 -06:00
|
|
|
event_id=parent_id,
|
|
|
|
relation_type=relation_type,
|
|
|
|
event_type=event_type,
|
|
|
|
limit=limit,
|
|
|
|
from_token=from_token,
|
|
|
|
to_token=to_token,
|
|
|
|
)
|
2019-05-14 09:59:21 -06:00
|
|
|
|
2019-12-05 09:46:37 -07:00
|
|
|
events = await self.store.get_events_as_list(
|
2019-07-18 07:41:42 -06:00
|
|
|
[c["event_id"] for c in pagination_chunk.chunk]
|
2019-05-14 09:59:21 -06:00
|
|
|
)
|
|
|
|
|
|
|
|
now = self.clock.time_msec()
|
2019-07-10 07:43:11 -06:00
|
|
|
# We set bundle_aggregations to False when retrieving the original
|
|
|
|
# event because we want the content before relations were applied to
|
|
|
|
# it.
|
2019-12-05 09:46:37 -07:00
|
|
|
original_event = await self._event_serializer.serialize_event(
|
2019-07-10 07:43:11 -06:00
|
|
|
event, now, bundle_aggregations=False
|
|
|
|
)
|
|
|
|
# Similarly, we don't allow relations to be applied to relations, so we
|
|
|
|
# return the original relations without any aggregations on top of them
|
|
|
|
# here.
|
2019-12-05 09:46:37 -07:00
|
|
|
events = await self._event_serializer.serialize_events(
|
2019-07-10 07:43:11 -06:00
|
|
|
events, now, bundle_aggregations=False
|
|
|
|
)
|
2019-05-14 09:59:21 -06:00
|
|
|
|
2019-07-18 07:41:42 -06:00
|
|
|
return_value = pagination_chunk.to_dict()
|
2019-05-14 09:59:21 -06:00
|
|
|
return_value["chunk"] = events
|
2019-07-09 06:43:08 -06:00
|
|
|
return_value["original_event"] = original_event
|
2019-05-14 09:59:21 -06:00
|
|
|
|
2019-08-30 09:28:26 -06:00
|
|
|
return 200, return_value
|
2019-05-14 09:59:21 -06:00
|
|
|
|
|
|
|
|
2019-05-14 09:59:21 -06:00
|
|
|
class RelationAggregationPaginationServlet(RestServlet):
|
|
|
|
"""API to paginate aggregation groups of relations, e.g. paginate the
|
|
|
|
types and counts of the reactions on the events.
|
|
|
|
|
|
|
|
Example request and response:
|
|
|
|
|
|
|
|
GET /rooms/{room_id}/aggregations/{parent_id}
|
|
|
|
|
|
|
|
{
|
|
|
|
chunk: [
|
|
|
|
{
|
|
|
|
"type": "m.reaction",
|
|
|
|
"key": "👍",
|
|
|
|
"count": 3
|
|
|
|
}
|
|
|
|
]
|
|
|
|
}
|
|
|
|
"""
|
|
|
|
|
2019-06-03 05:28:59 -06:00
|
|
|
PATTERNS = client_patterns(
|
2019-05-14 09:59:21 -06:00
|
|
|
"/rooms/(?P<room_id>[^/]*)/aggregations/(?P<parent_id>[^/]*)"
|
|
|
|
"(/(?P<relation_type>[^/]*)(/(?P<event_type>[^/]*))?)?$",
|
|
|
|
releases=(),
|
|
|
|
)
|
|
|
|
|
|
|
|
def __init__(self, hs):
|
|
|
|
super(RelationAggregationPaginationServlet, self).__init__()
|
|
|
|
self.auth = hs.get_auth()
|
|
|
|
self.store = hs.get_datastore()
|
2019-05-16 07:19:06 -06:00
|
|
|
self.event_handler = hs.get_event_handler()
|
2019-05-14 09:59:21 -06:00
|
|
|
|
2019-12-05 09:46:37 -07:00
|
|
|
async def on_GET(
|
|
|
|
self, request, room_id, parent_id, relation_type=None, event_type=None
|
|
|
|
):
|
|
|
|
requester = await self.auth.get_user_by_req(request, allow_guest=True)
|
2019-05-14 09:59:21 -06:00
|
|
|
|
2020-02-18 16:14:57 -07:00
|
|
|
await self.auth.check_user_in_room_or_world_readable(
|
|
|
|
room_id, requester.user.to_string(), allow_departed_users=True,
|
2019-05-14 09:59:21 -06:00
|
|
|
)
|
|
|
|
|
2019-05-16 07:19:06 -06:00
|
|
|
# This checks that a) the event exists and b) the user is allowed to
|
|
|
|
# view it.
|
2019-12-05 09:46:37 -07:00
|
|
|
event = await self.event_handler.get_event(requester.user, room_id, parent_id)
|
2019-05-16 07:19:06 -06:00
|
|
|
|
2019-05-14 09:59:21 -06:00
|
|
|
if relation_type not in (RelationTypes.ANNOTATION, None):
|
|
|
|
raise SynapseError(400, "Relation type must be 'annotation'")
|
|
|
|
|
|
|
|
limit = parse_integer(request, "limit", default=5)
|
|
|
|
from_token = parse_string(request, "from")
|
|
|
|
to_token = parse_string(request, "to")
|
|
|
|
|
2019-07-18 07:41:42 -06:00
|
|
|
if event.internal_metadata.is_redacted():
|
|
|
|
# If the event is redacted, return an empty list of relations
|
|
|
|
pagination_chunk = PaginationChunk(chunk=[])
|
|
|
|
else:
|
|
|
|
# Return the relations
|
|
|
|
if from_token:
|
|
|
|
from_token = AggregationPaginationToken.from_string(from_token)
|
|
|
|
|
|
|
|
if to_token:
|
|
|
|
to_token = AggregationPaginationToken.from_string(to_token)
|
|
|
|
|
2019-12-05 09:46:37 -07:00
|
|
|
pagination_chunk = await self.store.get_aggregation_groups_for_event(
|
2019-07-18 07:41:42 -06:00
|
|
|
event_id=parent_id,
|
|
|
|
event_type=event_type,
|
|
|
|
limit=limit,
|
|
|
|
from_token=from_token,
|
|
|
|
to_token=to_token,
|
|
|
|
)
|
|
|
|
|
2019-08-30 09:28:26 -06:00
|
|
|
return 200, pagination_chunk.to_dict()
|
2019-05-14 09:59:21 -06:00
|
|
|
|
|
|
|
|
|
|
|
class RelationAggregationGroupPaginationServlet(RestServlet):
|
|
|
|
"""API to paginate within an aggregation group of relations, e.g. paginate
|
|
|
|
all the 👍 reactions on an event.
|
|
|
|
|
|
|
|
Example request and response:
|
|
|
|
|
|
|
|
GET /rooms/{room_id}/aggregations/{parent_id}/m.annotation/m.reaction/👍
|
|
|
|
|
|
|
|
{
|
|
|
|
chunk: [
|
|
|
|
{
|
|
|
|
"type": "m.reaction",
|
|
|
|
"content": {
|
|
|
|
"m.relates_to": {
|
|
|
|
"rel_type": "m.annotation",
|
|
|
|
"key": "👍"
|
|
|
|
}
|
|
|
|
}
|
|
|
|
},
|
|
|
|
...
|
|
|
|
]
|
|
|
|
}
|
|
|
|
"""
|
|
|
|
|
2019-06-03 05:28:59 -06:00
|
|
|
PATTERNS = client_patterns(
|
2019-05-14 09:59:21 -06:00
|
|
|
"/rooms/(?P<room_id>[^/]*)/aggregations/(?P<parent_id>[^/]*)"
|
|
|
|
"/(?P<relation_type>[^/]*)/(?P<event_type>[^/]*)/(?P<key>[^/]*)$",
|
|
|
|
releases=(),
|
|
|
|
)
|
|
|
|
|
|
|
|
def __init__(self, hs):
|
|
|
|
super(RelationAggregationGroupPaginationServlet, self).__init__()
|
|
|
|
self.auth = hs.get_auth()
|
|
|
|
self.store = hs.get_datastore()
|
|
|
|
self.clock = hs.get_clock()
|
|
|
|
self._event_serializer = hs.get_event_client_serializer()
|
2019-05-16 07:19:06 -06:00
|
|
|
self.event_handler = hs.get_event_handler()
|
2019-05-14 09:59:21 -06:00
|
|
|
|
2019-12-05 09:46:37 -07:00
|
|
|
async def on_GET(self, request, room_id, parent_id, relation_type, event_type, key):
|
|
|
|
requester = await self.auth.get_user_by_req(request, allow_guest=True)
|
2019-05-14 09:59:21 -06:00
|
|
|
|
2020-02-18 16:14:57 -07:00
|
|
|
await self.auth.check_user_in_room_or_world_readable(
|
|
|
|
room_id, requester.user.to_string(), allow_departed_users=True,
|
2019-05-14 09:59:21 -06:00
|
|
|
)
|
|
|
|
|
2019-05-16 07:19:06 -06:00
|
|
|
# This checks that a) the event exists and b) the user is allowed to
|
|
|
|
# view it.
|
2019-12-05 09:46:37 -07:00
|
|
|
await self.event_handler.get_event(requester.user, room_id, parent_id)
|
2019-05-16 07:19:06 -06:00
|
|
|
|
2019-05-14 09:59:21 -06:00
|
|
|
if relation_type != RelationTypes.ANNOTATION:
|
|
|
|
raise SynapseError(400, "Relation type must be 'annotation'")
|
|
|
|
|
|
|
|
limit = parse_integer(request, "limit", default=5)
|
|
|
|
from_token = parse_string(request, "from")
|
|
|
|
to_token = parse_string(request, "to")
|
|
|
|
|
2019-05-16 07:24:58 -06:00
|
|
|
if from_token:
|
|
|
|
from_token = RelationPaginationToken.from_string(from_token)
|
|
|
|
|
|
|
|
if to_token:
|
|
|
|
to_token = RelationPaginationToken.from_string(to_token)
|
|
|
|
|
2019-12-05 09:46:37 -07:00
|
|
|
result = await self.store.get_relations_for_event(
|
2019-05-14 09:59:21 -06:00
|
|
|
event_id=parent_id,
|
|
|
|
relation_type=relation_type,
|
|
|
|
event_type=event_type,
|
|
|
|
aggregation_key=key,
|
|
|
|
limit=limit,
|
|
|
|
from_token=from_token,
|
|
|
|
to_token=to_token,
|
|
|
|
)
|
|
|
|
|
2019-12-05 09:46:37 -07:00
|
|
|
events = await self.store.get_events_as_list(
|
2019-05-14 09:59:21 -06:00
|
|
|
[c["event_id"] for c in result.chunk]
|
|
|
|
)
|
|
|
|
|
|
|
|
now = self.clock.time_msec()
|
2019-12-05 09:46:37 -07:00
|
|
|
events = await self._event_serializer.serialize_events(events, now)
|
2019-05-14 09:59:21 -06:00
|
|
|
|
|
|
|
return_value = result.to_dict()
|
|
|
|
return_value["chunk"] = events
|
|
|
|
|
2019-08-30 09:28:26 -06:00
|
|
|
return 200, return_value
|
2019-05-14 09:59:21 -06:00
|
|
|
|
|
|
|
|
2019-05-14 09:59:21 -06:00
|
|
|
def register_servlets(hs, http_server):
|
|
|
|
RelationSendServlet(hs).register(http_server)
|
2019-05-14 09:59:21 -06:00
|
|
|
RelationPaginationServlet(hs).register(http_server)
|
2019-05-14 09:59:21 -06:00
|
|
|
RelationAggregationPaginationServlet(hs).register(http_server)
|
|
|
|
RelationAggregationGroupPaginationServlet(hs).register(http_server)
|