2015-01-23 11:31:29 -07:00
|
|
|
# -*- coding: utf-8 -*-
|
2015-01-26 11:53:31 -07:00
|
|
|
# Copyright 2015 OpenMarket Ltd
|
2015-01-23 11:31:29 -07:00
|
|
|
#
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
|
|
|
|
|
|
|
from twisted.internet import defer
|
|
|
|
|
2015-04-21 09:35:53 -06:00
|
|
|
from synapse.http.servlet import (
|
|
|
|
RestServlet, parse_string, parse_integer, parse_boolean
|
|
|
|
)
|
2015-01-26 11:53:31 -07:00
|
|
|
from synapse.handlers.sync import SyncConfig
|
|
|
|
from synapse.types import StreamToken
|
2015-01-28 19:45:33 -07:00
|
|
|
from synapse.events.utils import (
|
|
|
|
serialize_event, format_event_for_client_v2_without_event_id,
|
|
|
|
)
|
2015-01-29 11:11:28 -07:00
|
|
|
from synapse.api.filtering import Filter
|
2015-01-23 11:31:29 -07:00
|
|
|
from ._base import client_v2_pattern
|
|
|
|
|
|
|
|
import logging
|
|
|
|
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
|
|
|
|
|
|
|
class SyncRestServlet(RestServlet):
|
|
|
|
"""
|
|
|
|
|
|
|
|
GET parameters::
|
|
|
|
timeout(int): How long to wait for new events in milliseconds.
|
|
|
|
limit(int): Maxiumum number of events per room to return.
|
|
|
|
gap(bool): Create gaps the message history if limit is exceeded to
|
|
|
|
ensure that the client has the most recent messages. Defaults to
|
|
|
|
"true".
|
|
|
|
sort(str,str): tuple of sort key (e.g. "timeline") and direction
|
|
|
|
(e.g. "asc", "desc"). Defaults to "timeline,asc".
|
|
|
|
since(batch_token): Batch token when asking for incremental deltas.
|
|
|
|
set_presence(str): What state the device presence should be set to.
|
|
|
|
default is "online".
|
|
|
|
backfill(bool): Should the HS request message history from other
|
|
|
|
servers. This may take a long time making it unsuitable for clients
|
|
|
|
expecting a prompt response. Defaults to "true".
|
|
|
|
filter(filter_id): A filter to apply to the events returned.
|
|
|
|
filter_*: Filter override parameters.
|
|
|
|
|
|
|
|
Response JSON::
|
|
|
|
{
|
|
|
|
"next_batch": // batch token for the next /sync
|
|
|
|
"private_user_data": // private events for this user.
|
|
|
|
"public_user_data": // public events for all users including the
|
|
|
|
// public events for this user.
|
|
|
|
"rooms": [{ // List of rooms with updates.
|
|
|
|
"room_id": // Id of the room being updated
|
|
|
|
"limited": // Was the per-room event limit exceeded?
|
|
|
|
"published": // Is the room published by our HS?
|
|
|
|
"event_map": // Map of EventID -> event JSON.
|
|
|
|
"events": { // The recent events in the room if gap is "true"
|
|
|
|
// otherwise the next events in the room.
|
|
|
|
"batch": [] // list of EventIDs in the "event_map".
|
|
|
|
"prev_batch": // back token for getting previous events.
|
|
|
|
}
|
|
|
|
"state": [] // list of EventIDs updating the current state to
|
|
|
|
// be what it should be at the end of the batch.
|
2015-01-29 09:41:21 -07:00
|
|
|
"ephemeral": []
|
2015-01-23 11:31:29 -07:00
|
|
|
}]
|
|
|
|
}
|
|
|
|
"""
|
|
|
|
|
|
|
|
PATTERN = client_v2_pattern("/sync$")
|
|
|
|
ALLOWED_SORT = set(["timeline,asc", "timeline,desc"])
|
|
|
|
ALLOWED_PRESENCE = set(["online", "offline", "idle"])
|
|
|
|
|
|
|
|
def __init__(self, hs):
|
|
|
|
super(SyncRestServlet, self).__init__()
|
|
|
|
self.auth = hs.get_auth()
|
2015-01-26 11:53:31 -07:00
|
|
|
self.sync_handler = hs.get_handlers().sync_handler
|
|
|
|
self.clock = hs.get_clock()
|
2015-01-29 11:11:28 -07:00
|
|
|
self.filtering = hs.get_filtering()
|
2015-01-23 11:31:29 -07:00
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def on_GET(self, request):
|
2015-01-28 10:32:41 -07:00
|
|
|
user, client = yield self.auth.get_user_by_req(request)
|
2015-01-23 11:31:29 -07:00
|
|
|
|
2015-04-21 09:35:53 -06:00
|
|
|
timeout = parse_integer(request, "timeout", default=0)
|
|
|
|
limit = parse_integer(request, "limit", required=True)
|
|
|
|
gap = parse_boolean(request, "gap", default=True)
|
|
|
|
sort = parse_string(
|
2015-01-23 11:31:29 -07:00
|
|
|
request, "sort", default="timeline,asc",
|
|
|
|
allowed_values=self.ALLOWED_SORT
|
|
|
|
)
|
2015-04-21 09:35:53 -06:00
|
|
|
since = parse_string(request, "since")
|
|
|
|
set_presence = parse_string(
|
2015-01-23 11:31:29 -07:00
|
|
|
request, "set_presence", default="online",
|
|
|
|
allowed_values=self.ALLOWED_PRESENCE
|
|
|
|
)
|
2015-04-21 09:35:53 -06:00
|
|
|
backfill = parse_boolean(request, "backfill", default=False)
|
|
|
|
filter_id = parse_string(request, "filter", default=None)
|
2015-01-23 11:31:29 -07:00
|
|
|
|
|
|
|
logger.info(
|
|
|
|
"/sync: user=%r, timeout=%r, limit=%r, gap=%r, sort=%r, since=%r,"
|
|
|
|
" set_presence=%r, backfill=%r, filter_id=%r" % (
|
|
|
|
user, timeout, limit, gap, sort, since, set_presence,
|
|
|
|
backfill, filter_id
|
|
|
|
)
|
|
|
|
)
|
|
|
|
|
|
|
|
# TODO(mjark): Load filter and apply overrides.
|
2015-01-29 11:11:28 -07:00
|
|
|
try:
|
|
|
|
filter = yield self.filtering.get_user_filter(
|
|
|
|
user.localpart, filter_id
|
|
|
|
)
|
|
|
|
except:
|
2015-01-30 04:32:35 -07:00
|
|
|
filter = Filter({})
|
2015-01-23 11:31:29 -07:00
|
|
|
# filter = filter.apply_overrides(http_request)
|
2015-02-10 10:58:36 -07:00
|
|
|
# if filter.matches(event):
|
2015-01-23 11:31:29 -07:00
|
|
|
# # stuff
|
|
|
|
|
2015-01-26 11:53:31 -07:00
|
|
|
sync_config = SyncConfig(
|
|
|
|
user=user,
|
2015-01-30 04:50:15 -07:00
|
|
|
client_info=client,
|
2015-01-26 11:53:31 -07:00
|
|
|
gap=gap,
|
|
|
|
limit=limit,
|
|
|
|
sort=sort,
|
|
|
|
backfill=backfill,
|
2015-01-29 11:11:28 -07:00
|
|
|
filter=filter,
|
2015-01-26 11:53:31 -07:00
|
|
|
)
|
|
|
|
|
|
|
|
if since is not None:
|
|
|
|
since_token = StreamToken.from_string(since)
|
2015-01-23 11:31:29 -07:00
|
|
|
else:
|
2015-01-26 11:53:31 -07:00
|
|
|
since_token = None
|
2015-01-23 11:31:29 -07:00
|
|
|
|
2015-01-26 11:53:31 -07:00
|
|
|
sync_result = yield self.sync_handler.wait_for_sync_for_user(
|
|
|
|
sync_config, since_token=since_token, timeout=timeout
|
|
|
|
)
|
2015-01-23 11:31:29 -07:00
|
|
|
|
2015-01-26 11:53:31 -07:00
|
|
|
time_now = self.clock.time_msec()
|
|
|
|
|
|
|
|
response_content = {
|
2015-01-28 20:33:51 -07:00
|
|
|
"public_user_data": self.encode_user_data(
|
2015-01-26 11:53:31 -07:00
|
|
|
sync_result.public_user_data, filter, time_now
|
|
|
|
),
|
2015-01-28 20:33:51 -07:00
|
|
|
"private_user_data": self.encode_user_data(
|
2015-01-26 11:53:31 -07:00
|
|
|
sync_result.private_user_data, filter, time_now
|
|
|
|
),
|
2015-01-28 19:45:33 -07:00
|
|
|
"rooms": self.encode_rooms(
|
|
|
|
sync_result.rooms, filter, time_now, client.token_id
|
|
|
|
),
|
2015-01-26 11:53:31 -07:00
|
|
|
"next_batch": sync_result.next_batch.to_string(),
|
|
|
|
}
|
2015-01-23 11:31:29 -07:00
|
|
|
|
|
|
|
defer.returnValue((200, response_content))
|
|
|
|
|
2015-01-28 20:33:51 -07:00
|
|
|
def encode_user_data(self, events, filter, time_now):
|
|
|
|
return events
|
2015-01-26 11:53:31 -07:00
|
|
|
|
2015-01-28 19:45:33 -07:00
|
|
|
def encode_rooms(self, rooms, filter, time_now, token_id):
|
|
|
|
return [
|
|
|
|
self.encode_room(room, filter, time_now, token_id)
|
|
|
|
for room in rooms
|
|
|
|
]
|
2015-01-26 11:53:31 -07:00
|
|
|
|
|
|
|
@staticmethod
|
2015-01-28 19:45:33 -07:00
|
|
|
def encode_room(room, filter, time_now, token_id):
|
2015-01-26 11:53:31 -07:00
|
|
|
event_map = {}
|
2015-01-29 11:11:28 -07:00
|
|
|
state_events = filter.filter_room_state(room.state)
|
|
|
|
recent_events = filter.filter_room_events(room.events)
|
2015-01-26 11:53:31 -07:00
|
|
|
state_event_ids = []
|
|
|
|
recent_event_ids = []
|
2015-01-29 11:11:28 -07:00
|
|
|
for event in state_events:
|
2015-01-26 11:53:31 -07:00
|
|
|
# TODO(mjark): Respect formatting requirements in the filter.
|
|
|
|
event_map[event.event_id] = serialize_event(
|
2015-01-28 19:45:33 -07:00
|
|
|
event, time_now, token_id=token_id,
|
|
|
|
event_format=format_event_for_client_v2_without_event_id,
|
2015-01-26 11:53:31 -07:00
|
|
|
)
|
|
|
|
state_event_ids.append(event.event_id)
|
|
|
|
|
2015-01-29 11:11:28 -07:00
|
|
|
for event in recent_events:
|
2015-01-26 11:53:31 -07:00
|
|
|
# TODO(mjark): Respect formatting requirements in the filter.
|
|
|
|
event_map[event.event_id] = serialize_event(
|
2015-01-28 19:45:33 -07:00
|
|
|
event, time_now, token_id=token_id,
|
|
|
|
event_format=format_event_for_client_v2_without_event_id,
|
2015-01-26 11:53:31 -07:00
|
|
|
)
|
|
|
|
recent_event_ids.append(event.event_id)
|
2015-01-28 20:33:51 -07:00
|
|
|
result = {
|
2015-01-26 11:53:31 -07:00
|
|
|
"room_id": room.room_id,
|
|
|
|
"event_map": event_map,
|
|
|
|
"events": {
|
|
|
|
"batch": recent_event_ids,
|
|
|
|
"prev_batch": room.prev_batch.to_string(),
|
|
|
|
},
|
|
|
|
"state": state_event_ids,
|
|
|
|
"limited": room.limited,
|
|
|
|
"published": room.published,
|
2015-01-29 09:41:21 -07:00
|
|
|
"ephemeral": room.ephemeral,
|
2015-01-26 11:53:31 -07:00
|
|
|
}
|
2015-01-28 20:33:51 -07:00
|
|
|
return result
|
2015-01-26 11:53:31 -07:00
|
|
|
|
2015-01-23 11:31:29 -07:00
|
|
|
|
|
|
|
def register_servlets(hs, http_server):
|
|
|
|
SyncRestServlet(hs).register(http_server)
|