Merge branch 'develop' into application-services
This commit is contained in:
commit
53557fc532
46
README.rst
46
README.rst
|
@ -96,6 +96,11 @@ Installing prerequisites on Ubuntu or Debian::
|
||||||
$ sudo apt-get install build-essential python2.7-dev libffi-dev \
|
$ sudo apt-get install build-essential python2.7-dev libffi-dev \
|
||||||
python-pip python-setuptools sqlite3 \
|
python-pip python-setuptools sqlite3 \
|
||||||
libssl-dev python-virtualenv libjpeg-dev
|
libssl-dev python-virtualenv libjpeg-dev
|
||||||
|
|
||||||
|
Installing prerequisites on ArchLinux::
|
||||||
|
|
||||||
|
$ sudo pacman -S base-devel python2 python-pip \
|
||||||
|
python-setuptools python-virtualenv sqlite3
|
||||||
|
|
||||||
Installing prerequisites on Mac OS X::
|
Installing prerequisites on Mac OS X::
|
||||||
|
|
||||||
|
@ -148,6 +153,39 @@ failing, e.g.::
|
||||||
On OSX, if you encounter clang: error: unknown argument: '-mno-fused-madd' you
|
On OSX, if you encounter clang: error: unknown argument: '-mno-fused-madd' you
|
||||||
will need to export CFLAGS=-Qunused-arguments.
|
will need to export CFLAGS=-Qunused-arguments.
|
||||||
|
|
||||||
|
ArchLinux
|
||||||
|
---------
|
||||||
|
|
||||||
|
Installation on ArchLinux may encounter a few hiccups as Arch defaults to
|
||||||
|
python 3, but synapse currently assumes python 2.7 by default.
|
||||||
|
|
||||||
|
pip may be outdated (6.0.7-1 and needs to be upgraded to 6.0.8-1 )::
|
||||||
|
|
||||||
|
$ sudo pip2.7 install --upgrade pip
|
||||||
|
|
||||||
|
You also may need to explicitly specify python 2.7 again during the install
|
||||||
|
request::
|
||||||
|
|
||||||
|
$ pip2.7 install --process-dependency-links \
|
||||||
|
https://github.com/matrix-org/synapse/tarball/master
|
||||||
|
|
||||||
|
If you encounter an error with lib bcrypt causing an Wrong ELF Class:
|
||||||
|
ELFCLASS32 (x64 Systems), you may need to reinstall py-bcrypt to correctly
|
||||||
|
compile it under the right architecture. (This should not be needed if
|
||||||
|
installing under virtualenv)::
|
||||||
|
|
||||||
|
$ sudo pip2.7 uninstall py-bcrypt
|
||||||
|
$ sudo pip2.7 install py-bcrypt
|
||||||
|
|
||||||
|
During setup of homeserver you need to call python2.7 directly again::
|
||||||
|
|
||||||
|
$ python2.7 -m synapse.app.homeserver \
|
||||||
|
--server-name machine.my.domain.name \
|
||||||
|
--config-path homeserver.yaml \
|
||||||
|
--generate-config
|
||||||
|
|
||||||
|
...substituting your host and domain name as appropriate.
|
||||||
|
|
||||||
Windows Install
|
Windows Install
|
||||||
---------------
|
---------------
|
||||||
Synapse can be installed on Cygwin. It requires the following Cygwin packages:
|
Synapse can be installed on Cygwin. It requires the following Cygwin packages:
|
||||||
|
@ -207,6 +245,14 @@ fix try re-installing from PyPI or directly from
|
||||||
$ # Install from github
|
$ # Install from github
|
||||||
$ pip install --user https://github.com/pyca/pynacl/tarball/master
|
$ pip install --user https://github.com/pyca/pynacl/tarball/master
|
||||||
|
|
||||||
|
ArchLinux
|
||||||
|
---------
|
||||||
|
|
||||||
|
If running `$ synctl start` fails wit 'returned non-zero exit status 1', you will need to explicitly call Python2.7 - either running as::
|
||||||
|
|
||||||
|
$ python2.7 -m synapse.app.homeserver --daemonize -c homeserver.yaml --pid-file homeserver.pid
|
||||||
|
|
||||||
|
...or by editing synctl with the correct python executable.
|
||||||
|
|
||||||
Homeserver Development
|
Homeserver Development
|
||||||
======================
|
======================
|
||||||
|
|
|
@ -0,0 +1,39 @@
|
||||||
|
#!/usr/bin/env perl
|
||||||
|
|
||||||
|
use strict;
|
||||||
|
use warnings;
|
||||||
|
|
||||||
|
use DBI;
|
||||||
|
use DBD::SQLite;
|
||||||
|
use JSON;
|
||||||
|
use Getopt::Long;
|
||||||
|
|
||||||
|
my $db; # = "homeserver.db";
|
||||||
|
my $server = "http://localhost:8008";
|
||||||
|
my $size = 320;
|
||||||
|
|
||||||
|
GetOptions("db|d=s", \$db,
|
||||||
|
"server|s=s", \$server,
|
||||||
|
"width|w=i", \$size) or usage();
|
||||||
|
|
||||||
|
usage() unless $db;
|
||||||
|
|
||||||
|
my $dbh = DBI->connect("dbi:SQLite:dbname=$db","","") || die $DBI::errstr;
|
||||||
|
|
||||||
|
my $res = $dbh->selectall_arrayref("select token, name from access_tokens, users where access_tokens.user_id = users.id group by user_id") || die $DBI::errstr;
|
||||||
|
|
||||||
|
foreach (@$res) {
|
||||||
|
my ($token, $mxid) = ($_->[0], $_->[1]);
|
||||||
|
my ($user_id) = ($mxid =~ m/@(.*):/);
|
||||||
|
my ($url) = $dbh->selectrow_array("select avatar_url from profiles where user_id=?", undef, $user_id);
|
||||||
|
if (!$url || $url =~ /#auto$/) {
|
||||||
|
`curl -s -o tmp.png "$server/_matrix/media/v1/identicon?name=${mxid}&width=$size&height=$size"`;
|
||||||
|
my $json = `curl -s -X POST -H "Content-Type: image/png" -T "tmp.png" $server/_matrix/media/v1/upload?access_token=$token`;
|
||||||
|
my $content_uri = from_json($json)->{content_uri};
|
||||||
|
`curl -X PUT -H "Content-Type: application/json" --data '{ "avatar_url": "${content_uri}#auto"}' $server/_matrix/client/api/v1/profile/${mxid}/avatar_url?access_token=$token`;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
sub usage {
|
||||||
|
die "usage: ./make-identicons.pl\n\t-d database [e.g. homeserver.db]\n\t-s homeserver (default: http://localhost:8008)\n\t-w identicon size in pixels (default 320)";
|
||||||
|
}
|
|
@ -68,7 +68,7 @@ class SynapseHomeServer(HomeServer):
|
||||||
return ClientV2AlphaRestResource(self)
|
return ClientV2AlphaRestResource(self)
|
||||||
|
|
||||||
def build_resource_for_federation(self):
|
def build_resource_for_federation(self):
|
||||||
return JsonResource()
|
return JsonResource(self)
|
||||||
|
|
||||||
def build_resource_for_app_services(self):
|
def build_resource_for_app_services(self):
|
||||||
return AppServiceRestResource(self)
|
return AppServiceRestResource(self)
|
||||||
|
@ -279,6 +279,8 @@ def setup():
|
||||||
|
|
||||||
hs.get_pusherpool().start()
|
hs.get_pusherpool().start()
|
||||||
|
|
||||||
|
hs.get_datastore().start_profiling()
|
||||||
|
|
||||||
if config.daemonize:
|
if config.daemonize:
|
||||||
print config.pid_file
|
print config.pid_file
|
||||||
daemon = Daemonize(
|
daemon = Daemonize(
|
||||||
|
|
|
@ -130,7 +130,9 @@ class FederationHandler(BaseHandler):
|
||||||
if auth_chain:
|
if auth_chain:
|
||||||
event_ids |= {e.event_id for e in auth_chain}
|
event_ids |= {e.event_id for e in auth_chain}
|
||||||
|
|
||||||
seen_ids = (yield self.store.have_events(event_ids)).keys()
|
seen_ids = set(
|
||||||
|
(yield self.store.have_events(event_ids)).keys()
|
||||||
|
)
|
||||||
|
|
||||||
if state and auth_chain is not None:
|
if state and auth_chain is not None:
|
||||||
# If we have any state or auth_chain given to us by the replication
|
# If we have any state or auth_chain given to us by the replication
|
||||||
|
@ -332,8 +334,11 @@ class FederationHandler(BaseHandler):
|
||||||
|
|
||||||
# Try the host we successfully got a response to /make_join/
|
# Try the host we successfully got a response to /make_join/
|
||||||
# request first.
|
# request first.
|
||||||
target_hosts.remove(origin)
|
try:
|
||||||
target_hosts.insert(0, origin)
|
target_hosts.remove(origin)
|
||||||
|
target_hosts.insert(0, origin)
|
||||||
|
except ValueError:
|
||||||
|
pass
|
||||||
|
|
||||||
ret = yield self.replication_layer.send_join(
|
ret = yield self.replication_layer.send_join(
|
||||||
target_hosts,
|
target_hosts,
|
||||||
|
@ -521,7 +526,7 @@ class FederationHandler(BaseHandler):
|
||||||
"Failed to get destination from event %s", s.event_id
|
"Failed to get destination from event %s", s.event_id
|
||||||
)
|
)
|
||||||
|
|
||||||
destinations.remove(origin)
|
destinations.discard(origin)
|
||||||
|
|
||||||
logger.debug(
|
logger.debug(
|
||||||
"on_send_join_request: Sending event: %s, signatures: %s",
|
"on_send_join_request: Sending event: %s, signatures: %s",
|
||||||
|
@ -786,12 +791,12 @@ class FederationHandler(BaseHandler):
|
||||||
@log_function
|
@log_function
|
||||||
def do_auth(self, origin, event, context, auth_events):
|
def do_auth(self, origin, event, context, auth_events):
|
||||||
# Check if we have all the auth events.
|
# Check if we have all the auth events.
|
||||||
res = yield self.store.have_events(
|
have_events = yield self.store.have_events(
|
||||||
[e_id for e_id, _ in event.auth_events]
|
[e_id for e_id, _ in event.auth_events]
|
||||||
)
|
)
|
||||||
|
|
||||||
event_auth_events = set(e_id for e_id, _ in event.auth_events)
|
event_auth_events = set(e_id for e_id, _ in event.auth_events)
|
||||||
seen_events = set(res.keys())
|
seen_events = set(have_events.keys())
|
||||||
|
|
||||||
missing_auth = event_auth_events - seen_events
|
missing_auth = event_auth_events - seen_events
|
||||||
|
|
||||||
|
@ -834,6 +839,11 @@ class FederationHandler(BaseHandler):
|
||||||
auth_events[(e.type, e.state_key)] = e
|
auth_events[(e.type, e.state_key)] = e
|
||||||
except AuthError:
|
except AuthError:
|
||||||
pass
|
pass
|
||||||
|
|
||||||
|
have_events = yield self.store.have_events(
|
||||||
|
[e_id for e_id, _ in event.auth_events]
|
||||||
|
)
|
||||||
|
seen_events = set(have_events.keys())
|
||||||
except:
|
except:
|
||||||
# FIXME:
|
# FIXME:
|
||||||
logger.exception("Failed to get auth chain")
|
logger.exception("Failed to get auth chain")
|
||||||
|
@ -847,64 +857,75 @@ class FederationHandler(BaseHandler):
|
||||||
# Do auth conflict res.
|
# Do auth conflict res.
|
||||||
logger.debug("Different auth: %s", different_auth)
|
logger.debug("Different auth: %s", different_auth)
|
||||||
|
|
||||||
# 1. Get what we think is the auth chain.
|
# Only do auth resolution if we have something new to say.
|
||||||
auth_ids = self.auth.compute_auth_events(
|
# We can't rove an auth failure.
|
||||||
event, context.current_state
|
do_resolution = False
|
||||||
)
|
for e_id in different_auth:
|
||||||
local_auth_chain = yield self.store.get_auth_chain(auth_ids)
|
if e_id in have_events:
|
||||||
|
if have_events[e_id] != RejectedReason.AUTH_ERROR:
|
||||||
|
do_resolution = True
|
||||||
|
break
|
||||||
|
|
||||||
try:
|
if do_resolution:
|
||||||
# 2. Get remote difference.
|
# 1. Get what we think is the auth chain.
|
||||||
result = yield self.replication_layer.query_auth(
|
auth_ids = self.auth.compute_auth_events(
|
||||||
origin,
|
event, context.current_state
|
||||||
event.room_id,
|
|
||||||
event.event_id,
|
|
||||||
local_auth_chain,
|
|
||||||
)
|
)
|
||||||
|
local_auth_chain = yield self.store.get_auth_chain(auth_ids)
|
||||||
|
|
||||||
seen_remotes = yield self.store.have_events(
|
try:
|
||||||
[e.event_id for e in result["auth_chain"]]
|
# 2. Get remote difference.
|
||||||
)
|
result = yield self.replication_layer.query_auth(
|
||||||
|
origin,
|
||||||
|
event.room_id,
|
||||||
|
event.event_id,
|
||||||
|
local_auth_chain,
|
||||||
|
)
|
||||||
|
|
||||||
# 3. Process any remote auth chain events we haven't seen.
|
seen_remotes = yield self.store.have_events(
|
||||||
for ev in result["auth_chain"]:
|
[e.event_id for e in result["auth_chain"]]
|
||||||
if ev.event_id in seen_remotes.keys():
|
)
|
||||||
continue
|
|
||||||
|
|
||||||
if ev.event_id == event.event_id:
|
# 3. Process any remote auth chain events we haven't seen.
|
||||||
continue
|
for ev in result["auth_chain"]:
|
||||||
|
if ev.event_id in seen_remotes.keys():
|
||||||
|
continue
|
||||||
|
|
||||||
try:
|
if ev.event_id == event.event_id:
|
||||||
auth_ids = [e_id for e_id, _ in ev.auth_events]
|
continue
|
||||||
auth = {
|
|
||||||
(e.type, e.state_key): e for e in result["auth_chain"]
|
|
||||||
if e.event_id in auth_ids
|
|
||||||
}
|
|
||||||
ev.internal_metadata.outlier = True
|
|
||||||
|
|
||||||
logger.debug(
|
try:
|
||||||
"do_auth %s different_auth: %s",
|
auth_ids = [e_id for e_id, _ in ev.auth_events]
|
||||||
event.event_id, e.event_id
|
auth = {
|
||||||
)
|
(e.type, e.state_key): e
|
||||||
|
for e in result["auth_chain"]
|
||||||
|
if e.event_id in auth_ids
|
||||||
|
}
|
||||||
|
ev.internal_metadata.outlier = True
|
||||||
|
|
||||||
yield self._handle_new_event(
|
logger.debug(
|
||||||
origin, ev, auth_events=auth
|
"do_auth %s different_auth: %s",
|
||||||
)
|
event.event_id, e.event_id
|
||||||
|
)
|
||||||
|
|
||||||
if ev.event_id in event_auth_events:
|
yield self._handle_new_event(
|
||||||
auth_events[(ev.type, ev.state_key)] = ev
|
origin, ev, auth_events=auth
|
||||||
except AuthError:
|
)
|
||||||
pass
|
|
||||||
|
|
||||||
except:
|
if ev.event_id in event_auth_events:
|
||||||
# FIXME:
|
auth_events[(ev.type, ev.state_key)] = ev
|
||||||
logger.exception("Failed to query auth chain")
|
except AuthError:
|
||||||
|
pass
|
||||||
|
|
||||||
# 4. Look at rejects and their proofs.
|
except:
|
||||||
# TODO.
|
# FIXME:
|
||||||
|
logger.exception("Failed to query auth chain")
|
||||||
|
|
||||||
context.current_state.update(auth_events)
|
# 4. Look at rejects and their proofs.
|
||||||
context.state_group = None
|
# TODO.
|
||||||
|
|
||||||
|
context.current_state.update(auth_events)
|
||||||
|
context.state_group = None
|
||||||
|
|
||||||
try:
|
try:
|
||||||
self.auth.check(event, auth_events=auth_events)
|
self.auth.check(event, auth_events=auth_events)
|
||||||
|
@ -1013,16 +1034,19 @@ class FederationHandler(BaseHandler):
|
||||||
for e in missing_remotes:
|
for e in missing_remotes:
|
||||||
for e_id, _ in e.auth_events:
|
for e_id, _ in e.auth_events:
|
||||||
if e_id in missing_remote_ids:
|
if e_id in missing_remote_ids:
|
||||||
base_remote_rejected.remove(e)
|
try:
|
||||||
|
base_remote_rejected.remove(e)
|
||||||
|
except ValueError:
|
||||||
|
pass
|
||||||
|
|
||||||
reason_map = {}
|
reason_map = {}
|
||||||
|
|
||||||
for e in base_remote_rejected:
|
for e in base_remote_rejected:
|
||||||
reason = yield self.store.get_rejection_reason(e.event_id)
|
reason = yield self.store.get_rejection_reason(e.event_id)
|
||||||
if reason is None:
|
if reason is None:
|
||||||
# FIXME: ERRR?!
|
# TODO: e is not in the current state, so we should
|
||||||
logger.warn("Could not find reason for %s", e.event_id)
|
# construct some proof of that.
|
||||||
raise RuntimeError("Could not find reason for %s" % e.event_id)
|
continue
|
||||||
|
|
||||||
reason_map[e.event_id] = reason
|
reason_map[e.event_id] = reason
|
||||||
|
|
||||||
|
|
|
@ -104,6 +104,23 @@ class RegistrationHandler(BaseHandler):
|
||||||
raise RegistrationError(
|
raise RegistrationError(
|
||||||
500, "Cannot generate user ID.")
|
500, "Cannot generate user ID.")
|
||||||
|
|
||||||
|
# create a default avatar for the user
|
||||||
|
# XXX: ideally clients would explicitly specify one, but given they don't
|
||||||
|
# and we want consistent and pretty identicons for random users, we'll
|
||||||
|
# do it here.
|
||||||
|
try:
|
||||||
|
auth_user = UserID.from_string(user_id)
|
||||||
|
identicon_resource = self.hs.get_resource_for_media_repository().getChildWithDefault("identicon", None)
|
||||||
|
upload_resource = self.hs.get_resource_for_media_repository().getChildWithDefault("upload", None)
|
||||||
|
identicon_bytes = identicon_resource.generate_identicon(user_id, 320, 320)
|
||||||
|
content_uri = yield upload_resource.create_content(
|
||||||
|
"image/png", None, identicon_bytes, len(identicon_bytes), auth_user
|
||||||
|
)
|
||||||
|
profile_handler = self.hs.get_handlers().profile_handler
|
||||||
|
profile_handler.set_avatar_url(auth_user, auth_user, ("%s#auto" % content_uri))
|
||||||
|
except NotImplementedError:
|
||||||
|
pass # make tests pass without messing around creating default avatars
|
||||||
|
|
||||||
defer.returnValue((user_id, token))
|
defer.returnValue((user_id, token))
|
||||||
|
|
||||||
@defer.inlineCallbacks
|
@defer.inlineCallbacks
|
||||||
|
|
|
@ -298,15 +298,17 @@ class SyncHandler(BaseHandler):
|
||||||
load_limit = max(sync_config.limit * filtering_factor, 100)
|
load_limit = max(sync_config.limit * filtering_factor, 100)
|
||||||
max_repeat = 3 # Only try a few times per room, otherwise
|
max_repeat = 3 # Only try a few times per room, otherwise
|
||||||
room_key = now_token.room_key
|
room_key = now_token.room_key
|
||||||
|
end_key = room_key
|
||||||
|
|
||||||
while limited and len(recents) < sync_config.limit and max_repeat:
|
while limited and len(recents) < sync_config.limit and max_repeat:
|
||||||
events, keys = yield self.store.get_recent_events_for_room(
|
events, keys = yield self.store.get_recent_events_for_room(
|
||||||
room_id,
|
room_id,
|
||||||
limit=load_limit + 1,
|
limit=load_limit + 1,
|
||||||
from_token=since_token.room_key if since_token else None,
|
from_token=since_token.room_key if since_token else None,
|
||||||
end_token=room_key,
|
end_token=end_key,
|
||||||
)
|
)
|
||||||
(room_key, _) = keys
|
(room_key, _) = keys
|
||||||
|
end_key = "s" + room_key.split('-')[-1]
|
||||||
loaded_recents = sync_config.filter.filter_room_events(events)
|
loaded_recents = sync_config.filter.filter_room_events(events)
|
||||||
loaded_recents.extend(recents)
|
loaded_recents.extend(recents)
|
||||||
recents = loaded_recents
|
recents = loaded_recents
|
||||||
|
|
|
@ -169,8 +169,9 @@ class MatrixFederationHttpClient(object):
|
||||||
else:
|
else:
|
||||||
# :'(
|
# :'(
|
||||||
# Update transactions table?
|
# Update transactions table?
|
||||||
|
body = yield readBody(response)
|
||||||
raise HttpResponseException(
|
raise HttpResponseException(
|
||||||
response.code, response.phrase, response
|
response.code, response.phrase, body
|
||||||
)
|
)
|
||||||
|
|
||||||
defer.returnValue(response)
|
defer.returnValue(response)
|
||||||
|
|
|
@ -69,9 +69,10 @@ class JsonResource(HttpServer, resource.Resource):
|
||||||
|
|
||||||
_PathEntry = collections.namedtuple("_PathEntry", ["pattern", "callback"])
|
_PathEntry = collections.namedtuple("_PathEntry", ["pattern", "callback"])
|
||||||
|
|
||||||
def __init__(self):
|
def __init__(self, hs):
|
||||||
resource.Resource.__init__(self)
|
resource.Resource.__init__(self)
|
||||||
|
|
||||||
|
self.clock = hs.get_clock()
|
||||||
self.path_regexs = {}
|
self.path_regexs = {}
|
||||||
|
|
||||||
def register_path(self, method, path_pattern, callback):
|
def register_path(self, method, path_pattern, callback):
|
||||||
|
@ -111,6 +112,7 @@ class JsonResource(HttpServer, resource.Resource):
|
||||||
This checks if anyone has registered a callback for that method and
|
This checks if anyone has registered a callback for that method and
|
||||||
path.
|
path.
|
||||||
"""
|
"""
|
||||||
|
code = None
|
||||||
try:
|
try:
|
||||||
# Just say yes to OPTIONS.
|
# Just say yes to OPTIONS.
|
||||||
if request.method == "OPTIONS":
|
if request.method == "OPTIONS":
|
||||||
|
@ -130,6 +132,13 @@ class JsonResource(HttpServer, resource.Resource):
|
||||||
urllib.unquote(u).decode("UTF-8") for u in m.groups()
|
urllib.unquote(u).decode("UTF-8") for u in m.groups()
|
||||||
]
|
]
|
||||||
|
|
||||||
|
logger.info(
|
||||||
|
"Received request: %s %s",
|
||||||
|
request.method, request.path
|
||||||
|
)
|
||||||
|
|
||||||
|
start = self.clock.time_msec()
|
||||||
|
|
||||||
code, response = yield path_entry.callback(
|
code, response = yield path_entry.callback(
|
||||||
request,
|
request,
|
||||||
*args
|
*args
|
||||||
|
@ -145,9 +154,11 @@ class JsonResource(HttpServer, resource.Resource):
|
||||||
logger.info("%s SynapseError: %s - %s", request, e.code, e.msg)
|
logger.info("%s SynapseError: %s - %s", request, e.code, e.msg)
|
||||||
else:
|
else:
|
||||||
logger.exception(e)
|
logger.exception(e)
|
||||||
|
|
||||||
|
code = e.code
|
||||||
self._send_response(
|
self._send_response(
|
||||||
request,
|
request,
|
||||||
e.code,
|
code,
|
||||||
cs_exception(e),
|
cs_exception(e),
|
||||||
response_code_message=e.response_code_message
|
response_code_message=e.response_code_message
|
||||||
)
|
)
|
||||||
|
@ -158,6 +169,14 @@ class JsonResource(HttpServer, resource.Resource):
|
||||||
500,
|
500,
|
||||||
{"error": "Internal server error"}
|
{"error": "Internal server error"}
|
||||||
)
|
)
|
||||||
|
finally:
|
||||||
|
code = str(code) if code else "-"
|
||||||
|
|
||||||
|
end = self.clock.time_msec()
|
||||||
|
logger.info(
|
||||||
|
"Processed request: %dms %s %s %s",
|
||||||
|
end-start, code, request.method, request.path
|
||||||
|
)
|
||||||
|
|
||||||
def _send_response(self, request, code, response_json_object,
|
def _send_response(self, request, code, response_json_object,
|
||||||
response_code_message=None):
|
response_code_message=None):
|
||||||
|
|
|
@ -22,7 +22,6 @@ import synapse.util.async
|
||||||
import baserules
|
import baserules
|
||||||
|
|
||||||
import logging
|
import logging
|
||||||
import fnmatch
|
|
||||||
import json
|
import json
|
||||||
import re
|
import re
|
||||||
|
|
||||||
|
@ -130,26 +129,35 @@ class Pusher(object):
|
||||||
|
|
||||||
defer.returnValue(Pusher.DEFAULT_ACTIONS)
|
defer.returnValue(Pusher.DEFAULT_ACTIONS)
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def _glob_to_regexp(glob):
|
||||||
|
r = re.escape(glob)
|
||||||
|
r = re.sub(r'\\\*', r'.*?', r)
|
||||||
|
r = re.sub(r'\\\?', r'.', r)
|
||||||
|
|
||||||
|
# handle [abc], [a-z] and [!a-z] style ranges.
|
||||||
|
r = re.sub(r'\\\[(\\\!|)(.*)\\\]',
|
||||||
|
lambda x: ('[%s%s]' % (x.group(1) and '^' or '',
|
||||||
|
re.sub(r'\\\-', '-', x.group(2)))), r)
|
||||||
|
return r
|
||||||
|
|
||||||
def _event_fulfills_condition(self, ev, condition, display_name, room_member_count):
|
def _event_fulfills_condition(self, ev, condition, display_name, room_member_count):
|
||||||
if condition['kind'] == 'event_match':
|
if condition['kind'] == 'event_match':
|
||||||
if 'pattern' not in condition:
|
if 'pattern' not in condition:
|
||||||
logger.warn("event_match condition with no pattern")
|
logger.warn("event_match condition with no pattern")
|
||||||
return False
|
return False
|
||||||
pat = condition['pattern']
|
# XXX: optimisation: cache our pattern regexps
|
||||||
|
r = r'\b%s\b' % self._glob_to_regexp(condition['pattern'])
|
||||||
if pat.strip("*?[]") == pat:
|
|
||||||
# no special glob characters so we assume the user means
|
|
||||||
# 'contains this string' rather than 'is this string'
|
|
||||||
pat = "*%s*" % (pat,)
|
|
||||||
|
|
||||||
val = _value_for_dotted_key(condition['key'], ev)
|
val = _value_for_dotted_key(condition['key'], ev)
|
||||||
if val is None:
|
if val is None:
|
||||||
return False
|
return False
|
||||||
return fnmatch.fnmatch(val.upper(), pat.upper())
|
return re.match(r, val, flags=re.IGNORECASE) != None
|
||||||
|
|
||||||
elif condition['kind'] == 'device':
|
elif condition['kind'] == 'device':
|
||||||
if 'profile_tag' not in condition:
|
if 'profile_tag' not in condition:
|
||||||
return True
|
return True
|
||||||
return condition['profile_tag'] == self.profile_tag
|
return condition['profile_tag'] == self.profile_tag
|
||||||
|
|
||||||
elif condition['kind'] == 'contains_display_name':
|
elif condition['kind'] == 'contains_display_name':
|
||||||
# This is special because display names can be different
|
# This is special because display names can be different
|
||||||
# between rooms and so you can't really hard code it in a rule.
|
# between rooms and so you can't really hard code it in a rule.
|
||||||
|
@ -159,9 +167,9 @@ class Pusher(object):
|
||||||
return False
|
return False
|
||||||
if not display_name:
|
if not display_name:
|
||||||
return False
|
return False
|
||||||
return fnmatch.fnmatch(
|
return re.match("\b%s\b" % re.escape(display_name),
|
||||||
ev['content']['body'].upper(), "*%s*" % (display_name.upper(),)
|
ev['content']['body'], flags=re.IGNORECASE) != None
|
||||||
)
|
|
||||||
elif condition['kind'] == 'room_member_count':
|
elif condition['kind'] == 'room_member_count':
|
||||||
if 'is' not in condition:
|
if 'is' not in condition:
|
||||||
return False
|
return False
|
||||||
|
|
|
@ -4,24 +4,24 @@ def list_with_base_rules(rawrules, user_name):
|
||||||
ruleslist = []
|
ruleslist = []
|
||||||
|
|
||||||
# shove the server default rules for each kind onto the end of each
|
# shove the server default rules for each kind onto the end of each
|
||||||
current_prio_class = 1
|
current_prio_class = PRIORITY_CLASS_INVERSE_MAP.keys()[-1]
|
||||||
for r in rawrules:
|
for r in rawrules:
|
||||||
if r['priority_class'] > current_prio_class:
|
if r['priority_class'] < current_prio_class:
|
||||||
while current_prio_class < r['priority_class']:
|
while r['priority_class'] < current_prio_class:
|
||||||
ruleslist.extend(make_base_rules(
|
ruleslist.extend(make_base_rules(
|
||||||
user_name,
|
user_name,
|
||||||
PRIORITY_CLASS_INVERSE_MAP[current_prio_class])
|
PRIORITY_CLASS_INVERSE_MAP[current_prio_class])
|
||||||
)
|
)
|
||||||
current_prio_class += 1
|
current_prio_class -= 1
|
||||||
|
|
||||||
ruleslist.append(r)
|
ruleslist.append(r)
|
||||||
|
|
||||||
while current_prio_class <= PRIORITY_CLASS_INVERSE_MAP.keys()[-1]:
|
while current_prio_class > 0:
|
||||||
ruleslist.extend(make_base_rules(
|
ruleslist.extend(make_base_rules(
|
||||||
user_name,
|
user_name,
|
||||||
PRIORITY_CLASS_INVERSE_MAP[current_prio_class])
|
PRIORITY_CLASS_INVERSE_MAP[current_prio_class])
|
||||||
)
|
)
|
||||||
current_prio_class += 1
|
current_prio_class -= 1
|
||||||
|
|
||||||
return ruleslist
|
return ruleslist
|
||||||
|
|
||||||
|
|
|
@ -25,7 +25,7 @@ class ClientV1RestResource(JsonResource):
|
||||||
"""A resource for version 1 of the matrix client API."""
|
"""A resource for version 1 of the matrix client API."""
|
||||||
|
|
||||||
def __init__(self, hs):
|
def __init__(self, hs):
|
||||||
JsonResource.__init__(self)
|
JsonResource.__init__(self, hs)
|
||||||
self.register_servlets(self, hs)
|
self.register_servlets(self, hs)
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
|
|
|
@ -25,7 +25,7 @@ class ClientV2AlphaRestResource(JsonResource):
|
||||||
"""A resource for version 2 alpha of the matrix client API."""
|
"""A resource for version 2 alpha of the matrix client API."""
|
||||||
|
|
||||||
def __init__(self, hs):
|
def __init__(self, hs):
|
||||||
JsonResource.__init__(self)
|
JsonResource.__init__(self, hs)
|
||||||
self.register_servlets(self, hs)
|
self.register_servlets(self, hs)
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
|
|
|
@ -82,7 +82,7 @@ class BaseMediaResource(Resource):
|
||||||
raise SynapseError(
|
raise SynapseError(
|
||||||
404,
|
404,
|
||||||
"Invalid media id token %r" % (request.postpath,),
|
"Invalid media id token %r" % (request.postpath,),
|
||||||
Codes.UNKKOWN,
|
Codes.UNKNOWN,
|
||||||
)
|
)
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
|
|
|
@ -34,6 +34,7 @@ class MediaRepositoryResource(Resource):
|
||||||
|
|
||||||
=> POST /_matrix/media/v1/upload HTTP/1.1
|
=> POST /_matrix/media/v1/upload HTTP/1.1
|
||||||
Content-Type: <media-type>
|
Content-Type: <media-type>
|
||||||
|
Content-Length: <content-length>
|
||||||
|
|
||||||
<media>
|
<media>
|
||||||
|
|
||||||
|
|
|
@ -38,6 +38,35 @@ class UploadResource(BaseMediaResource):
|
||||||
def render_OPTIONS(self, request):
|
def render_OPTIONS(self, request):
|
||||||
respond_with_json(request, 200, {}, send_cors=True)
|
respond_with_json(request, 200, {}, send_cors=True)
|
||||||
return NOT_DONE_YET
|
return NOT_DONE_YET
|
||||||
|
|
||||||
|
@defer.inlineCallbacks
|
||||||
|
def create_content(self, media_type, upload_name, content, content_length, auth_user):
|
||||||
|
media_id = random_string(24)
|
||||||
|
|
||||||
|
fname = self.filepaths.local_media_filepath(media_id)
|
||||||
|
self._makedirs(fname)
|
||||||
|
|
||||||
|
# This shouldn't block for very long because the content will have
|
||||||
|
# already been uploaded at this point.
|
||||||
|
with open(fname, "wb") as f:
|
||||||
|
f.write(content)
|
||||||
|
|
||||||
|
yield self.store.store_local_media(
|
||||||
|
media_id=media_id,
|
||||||
|
media_type=media_type,
|
||||||
|
time_now_ms=self.clock.time_msec(),
|
||||||
|
upload_name=upload_name,
|
||||||
|
media_length=content_length,
|
||||||
|
user_id=auth_user,
|
||||||
|
)
|
||||||
|
media_info = {
|
||||||
|
"media_type": media_type,
|
||||||
|
"media_length": content_length,
|
||||||
|
}
|
||||||
|
|
||||||
|
yield self._generate_local_thumbnails(media_id, media_info)
|
||||||
|
|
||||||
|
defer.returnValue("mxc://%s/%s" % (self.server_name, media_id))
|
||||||
|
|
||||||
@defer.inlineCallbacks
|
@defer.inlineCallbacks
|
||||||
def _async_render_POST(self, request):
|
def _async_render_POST(self, request):
|
||||||
|
@ -70,32 +99,10 @@ class UploadResource(BaseMediaResource):
|
||||||
# disposition = headers.getRawHeaders("Content-Disposition")[0]
|
# disposition = headers.getRawHeaders("Content-Disposition")[0]
|
||||||
# TODO(markjh): parse content-dispostion
|
# TODO(markjh): parse content-dispostion
|
||||||
|
|
||||||
media_id = random_string(24)
|
content_uri = yield self.create_content(
|
||||||
|
media_type, None, request.content.read(),
|
||||||
fname = self.filepaths.local_media_filepath(media_id)
|
content_length, auth_user
|
||||||
self._makedirs(fname)
|
|
||||||
|
|
||||||
# This shouldn't block for very long because the content will have
|
|
||||||
# already been uploaded at this point.
|
|
||||||
with open(fname, "wb") as f:
|
|
||||||
f.write(request.content.read())
|
|
||||||
|
|
||||||
yield self.store.store_local_media(
|
|
||||||
media_id=media_id,
|
|
||||||
media_type=media_type,
|
|
||||||
time_now_ms=self.clock.time_msec(),
|
|
||||||
upload_name=None,
|
|
||||||
media_length=content_length,
|
|
||||||
user_id=auth_user,
|
|
||||||
)
|
)
|
||||||
media_info = {
|
|
||||||
"media_type": media_type,
|
|
||||||
"media_length": content_length,
|
|
||||||
}
|
|
||||||
|
|
||||||
yield self._generate_local_thumbnails(media_id, media_info)
|
|
||||||
|
|
||||||
content_uri = "mxc://%s/%s" % (self.server_name, media_id)
|
|
||||||
|
|
||||||
respond_with_json(
|
respond_with_json(
|
||||||
request, 200, {"content_uri": content_uri}, send_cors=True
|
request, 200, {"content_uri": content_uri}, send_cors=True
|
||||||
|
|
|
@ -85,6 +85,28 @@ class SQLBaseStore(object):
|
||||||
self._db_pool = hs.get_db_pool()
|
self._db_pool = hs.get_db_pool()
|
||||||
self._clock = hs.get_clock()
|
self._clock = hs.get_clock()
|
||||||
|
|
||||||
|
self._previous_txn_total_time = 0
|
||||||
|
self._current_txn_total_time = 0
|
||||||
|
self._previous_loop_ts = 0
|
||||||
|
|
||||||
|
def start_profiling(self):
|
||||||
|
self._previous_loop_ts = self._clock.time_msec()
|
||||||
|
|
||||||
|
def loop():
|
||||||
|
curr = self._current_txn_total_time
|
||||||
|
prev = self._previous_txn_total_time
|
||||||
|
self._previous_txn_total_time = curr
|
||||||
|
|
||||||
|
time_now = self._clock.time_msec()
|
||||||
|
time_then = self._previous_loop_ts
|
||||||
|
self._previous_loop_ts = time_now
|
||||||
|
|
||||||
|
ratio = (curr - prev)/(time_now - time_then)
|
||||||
|
|
||||||
|
logger.info("Total database time: %.3f%%", ratio * 100)
|
||||||
|
|
||||||
|
self._clock.looping_call(loop, 10000)
|
||||||
|
|
||||||
@defer.inlineCallbacks
|
@defer.inlineCallbacks
|
||||||
def runInteraction(self, desc, func, *args, **kwargs):
|
def runInteraction(self, desc, func, *args, **kwargs):
|
||||||
"""Wraps the .runInteraction() method on the underlying db_pool."""
|
"""Wraps the .runInteraction() method on the underlying db_pool."""
|
||||||
|
@ -114,6 +136,9 @@ class SQLBaseStore(object):
|
||||||
"[TXN END] {%s} %f",
|
"[TXN END] {%s} %f",
|
||||||
name, end - start
|
name, end - start
|
||||||
)
|
)
|
||||||
|
|
||||||
|
self._current_txn_total_time += end - start
|
||||||
|
|
||||||
with PreserveLoggingContext():
|
with PreserveLoggingContext():
|
||||||
result = yield self._db_pool.runInteraction(
|
result = yield self._db_pool.runInteraction(
|
||||||
inner_func, *args, **kwargs
|
inner_func, *args, **kwargs
|
||||||
|
|
|
@ -52,3 +52,14 @@ CREATE TABLE IF NOT EXISTS push_rules (
|
||||||
);
|
);
|
||||||
|
|
||||||
CREATE INDEX IF NOT EXISTS push_rules_user_name on push_rules (user_name);
|
CREATE INDEX IF NOT EXISTS push_rules_user_name on push_rules (user_name);
|
||||||
|
|
||||||
|
CREATE TABLE IF NOT EXISTS user_filters(
|
||||||
|
user_id TEXT,
|
||||||
|
filter_id INTEGER,
|
||||||
|
filter_json TEXT,
|
||||||
|
FOREIGN KEY(user_id) REFERENCES users(id)
|
||||||
|
);
|
||||||
|
|
||||||
|
CREATE INDEX IF NOT EXISTS user_filters_by_user_id_filter_id ON user_filters(
|
||||||
|
user_id, filter_id
|
||||||
|
);
|
||||||
|
|
|
@ -1,24 +0,0 @@
|
||||||
/* Copyright 2015 OpenMarket Ltd
|
|
||||||
*
|
|
||||||
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
||||||
* you may not use this file except in compliance with the License.
|
|
||||||
* You may obtain a copy of the License at
|
|
||||||
*
|
|
||||||
* http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
*
|
|
||||||
* Unless required by applicable law or agreed to in writing, software
|
|
||||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
* See the License for the specific language governing permissions and
|
|
||||||
* limitations under the License.
|
|
||||||
*/
|
|
||||||
CREATE TABLE IF NOT EXISTS user_filters(
|
|
||||||
user_id TEXT,
|
|
||||||
filter_id INTEGER,
|
|
||||||
filter_json TEXT,
|
|
||||||
FOREIGN KEY(user_id) REFERENCES users(id)
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE INDEX IF NOT EXISTS user_filters_by_user_id_filter_id ON user_filters(
|
|
||||||
user_id, filter_id
|
|
||||||
);
|
|
|
@ -15,7 +15,7 @@
|
||||||
|
|
||||||
from synapse.util.logcontext import LoggingContext
|
from synapse.util.logcontext import LoggingContext
|
||||||
|
|
||||||
from twisted.internet import reactor
|
from twisted.internet import reactor, task
|
||||||
|
|
||||||
import time
|
import time
|
||||||
|
|
||||||
|
@ -35,6 +35,14 @@ class Clock(object):
|
||||||
"""Returns the current system time in miliseconds since epoch."""
|
"""Returns the current system time in miliseconds since epoch."""
|
||||||
return self.time() * 1000
|
return self.time() * 1000
|
||||||
|
|
||||||
|
def looping_call(self, f, msec):
|
||||||
|
l = task.LoopingCall(f)
|
||||||
|
l.start(msec/1000.0, now=False)
|
||||||
|
return l
|
||||||
|
|
||||||
|
def stop_looping_call(self, loop):
|
||||||
|
loop.stop()
|
||||||
|
|
||||||
def call_later(self, delay, callback):
|
def call_later(self, delay, callback):
|
||||||
current_context = LoggingContext.current_context()
|
current_context = LoggingContext.current_context()
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue