2016-01-06 21:26:29 -07:00
|
|
|
# Copyright 2015, 2016 OpenMarket Ltd
|
2022-03-30 07:39:27 -06:00
|
|
|
# Copyright 2022 The Matrix.org Foundation C.I.C.
|
2015-02-03 07:44:16 -07:00
|
|
|
#
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
2018-07-09 00:09:20 -06:00
|
|
|
import logging
|
2021-12-14 10:02:46 -07:00
|
|
|
import urllib.parse
|
2023-02-06 05:49:06 -07:00
|
|
|
from typing import (
|
|
|
|
TYPE_CHECKING,
|
|
|
|
Any,
|
|
|
|
Dict,
|
|
|
|
Iterable,
|
|
|
|
List,
|
|
|
|
Mapping,
|
|
|
|
Optional,
|
|
|
|
Sequence,
|
|
|
|
Tuple,
|
|
|
|
)
|
2018-07-09 00:09:20 -06:00
|
|
|
|
|
|
|
from prometheus_client import Counter
|
2022-05-06 06:35:20 -06:00
|
|
|
from typing_extensions import TypeGuard
|
2018-07-09 00:09:20 -06:00
|
|
|
|
2021-06-09 12:39:51 -06:00
|
|
|
from synapse.api.constants import EventTypes, Membership, ThirdPartyEntityKind
|
2015-02-04 10:32:44 -07:00
|
|
|
from synapse.api.errors import CodeMessageException
|
2022-02-24 10:55:45 -07:00
|
|
|
from synapse.appservice import (
|
|
|
|
ApplicationService,
|
2022-11-28 09:17:29 -07:00
|
|
|
TransactionOneTimeKeysCount,
|
2022-02-24 10:55:45 -07:00
|
|
|
TransactionUnusedFallbackKeys,
|
|
|
|
)
|
2020-10-15 10:33:28 -06:00
|
|
|
from synapse.events import EventBase
|
2022-03-03 08:43:06 -07:00
|
|
|
from synapse.events.utils import SerializeEventConfig, serialize_event
|
2018-07-09 00:09:20 -06:00
|
|
|
from synapse.http.client import SimpleHttpClient
|
2022-03-30 07:39:27 -06:00
|
|
|
from synapse.types import DeviceListUpdates, JsonDict, ThirdPartyInstanceID
|
2018-07-09 00:09:20 -06:00
|
|
|
from synapse.util.caches.response_cache import ResponseCache
|
2018-06-05 06:17:55 -06:00
|
|
|
|
2020-09-01 09:03:49 -06:00
|
|
|
if TYPE_CHECKING:
|
2021-10-22 11:15:41 -06:00
|
|
|
from synapse.server import HomeServer
|
2020-09-01 09:03:49 -06:00
|
|
|
|
2015-02-04 09:44:53 -07:00
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
2018-06-05 06:17:55 -06:00
|
|
|
sent_transactions_counter = Counter(
|
|
|
|
"synapse_appservice_api_sent_transactions",
|
|
|
|
"Number of /transactions/ requests sent",
|
|
|
|
["service"],
|
|
|
|
)
|
|
|
|
|
|
|
|
failed_transactions_counter = Counter(
|
|
|
|
"synapse_appservice_api_failed_transactions",
|
|
|
|
"Number of /transactions/ requests that failed to send",
|
|
|
|
["service"],
|
|
|
|
)
|
|
|
|
|
|
|
|
sent_events_counter = Counter(
|
|
|
|
"synapse_appservice_api_sent_events", "Number of events sent to the AS", ["service"]
|
|
|
|
)
|
2015-02-04 09:44:53 -07:00
|
|
|
|
2022-07-13 08:18:20 -06:00
|
|
|
sent_ephemeral_counter = Counter(
|
|
|
|
"synapse_appservice_api_sent_ephemeral",
|
|
|
|
"Number of ephemeral events sent to the AS",
|
|
|
|
["service"],
|
|
|
|
)
|
|
|
|
|
|
|
|
sent_todevice_counter = Counter(
|
|
|
|
"synapse_appservice_api_sent_todevice",
|
|
|
|
"Number of todevice messages sent to the AS",
|
|
|
|
["service"],
|
|
|
|
)
|
|
|
|
|
2016-08-25 08:56:27 -06:00
|
|
|
HOUR_IN_MS = 60 * 60 * 1000
|
|
|
|
|
|
|
|
|
2016-08-25 11:06:29 -06:00
|
|
|
APP_SERVICE_PREFIX = "/_matrix/app/unstable"
|
|
|
|
|
|
|
|
|
2021-12-14 10:02:46 -07:00
|
|
|
def _is_valid_3pe_metadata(info: JsonDict) -> bool:
|
2016-09-09 08:09:46 -06:00
|
|
|
if "instances" not in info:
|
|
|
|
return False
|
|
|
|
if not isinstance(info["instances"], list):
|
|
|
|
return False
|
|
|
|
return True
|
|
|
|
|
|
|
|
|
2022-05-06 06:35:20 -06:00
|
|
|
def _is_valid_3pe_result(r: object, field: str) -> TypeGuard[JsonDict]:
|
2016-08-18 10:33:56 -06:00
|
|
|
if not isinstance(r, dict):
|
|
|
|
return False
|
|
|
|
|
|
|
|
for k in (field, "protocol"):
|
|
|
|
if k not in r:
|
|
|
|
return False
|
|
|
|
if not isinstance(r[k], str):
|
|
|
|
return False
|
|
|
|
|
|
|
|
if "fields" not in r:
|
|
|
|
return False
|
|
|
|
fields = r["fields"]
|
|
|
|
if not isinstance(fields, dict):
|
|
|
|
return False
|
|
|
|
|
|
|
|
return True
|
|
|
|
|
|
|
|
|
2015-02-04 09:44:53 -07:00
|
|
|
class ApplicationServiceApi(SimpleHttpClient):
|
2015-02-04 04:19:18 -07:00
|
|
|
"""This class manages HS -> AS communications, including querying and
|
|
|
|
pushing.
|
|
|
|
"""
|
|
|
|
|
2021-10-22 11:15:41 -06:00
|
|
|
def __init__(self, hs: "HomeServer"):
|
2020-09-18 07:56:44 -06:00
|
|
|
super().__init__(hs)
|
2015-02-05 06:42:35 -07:00
|
|
|
self.clock = hs.get_clock()
|
2015-02-04 04:19:18 -07:00
|
|
|
|
2021-07-15 04:02:43 -06:00
|
|
|
self.protocol_meta_cache: ResponseCache[Tuple[str, str]] = ResponseCache(
|
2021-03-08 12:00:07 -07:00
|
|
|
hs.get_clock(), "as_protocol_meta", timeout_ms=HOUR_IN_MS
|
2021-07-15 04:02:43 -06:00
|
|
|
)
|
2016-08-25 08:56:27 -06:00
|
|
|
|
2021-12-14 10:02:46 -07:00
|
|
|
async def query_user(self, service: "ApplicationService", user_id: str) -> bool:
|
2016-08-30 10:16:00 -06:00
|
|
|
if service.url is None:
|
2019-07-23 07:00:55 -06:00
|
|
|
return False
|
2021-12-14 10:02:46 -07:00
|
|
|
|
|
|
|
# This is required by the configuration.
|
|
|
|
assert service.hs_token is not None
|
|
|
|
|
2018-09-05 08:10:47 -06:00
|
|
|
uri = service.url + ("/users/%s" % urllib.parse.quote(user_id))
|
2015-02-04 09:44:53 -07:00
|
|
|
try:
|
2022-10-04 05:06:41 -06:00
|
|
|
response = await self.get_json(
|
|
|
|
uri,
|
|
|
|
{"access_token": service.hs_token},
|
2022-10-26 07:00:01 -06:00
|
|
|
headers={"Authorization": [f"Bearer {service.hs_token}"]},
|
2022-10-04 05:06:41 -06:00
|
|
|
)
|
2015-02-09 08:01:28 -07:00
|
|
|
if response is not None: # just an empty json object
|
2019-07-23 07:00:55 -06:00
|
|
|
return True
|
2015-02-04 10:32:44 -07:00
|
|
|
except CodeMessageException as e:
|
|
|
|
if e.code == 404:
|
2019-07-23 07:00:55 -06:00
|
|
|
return False
|
2015-02-04 10:32:44 -07:00
|
|
|
logger.warning("query_user to %s received %s", uri, e.code)
|
2015-02-05 06:19:46 -07:00
|
|
|
except Exception as ex:
|
|
|
|
logger.warning("query_user to %s threw exception %s", uri, ex)
|
2019-07-23 07:00:55 -06:00
|
|
|
return False
|
2015-02-04 04:19:18 -07:00
|
|
|
|
2021-12-14 10:02:46 -07:00
|
|
|
async def query_alias(self, service: "ApplicationService", alias: str) -> bool:
|
2016-08-30 10:16:00 -06:00
|
|
|
if service.url is None:
|
2019-07-23 07:00:55 -06:00
|
|
|
return False
|
2021-12-14 10:02:46 -07:00
|
|
|
|
|
|
|
# This is required by the configuration.
|
|
|
|
assert service.hs_token is not None
|
|
|
|
|
2018-09-05 08:10:47 -06:00
|
|
|
uri = service.url + ("/rooms/%s" % urllib.parse.quote(alias))
|
2015-02-04 09:44:53 -07:00
|
|
|
try:
|
2022-10-04 05:06:41 -06:00
|
|
|
response = await self.get_json(
|
|
|
|
uri,
|
|
|
|
{"access_token": service.hs_token},
|
2022-10-26 07:00:01 -06:00
|
|
|
headers={"Authorization": [f"Bearer {service.hs_token}"]},
|
2022-10-04 05:06:41 -06:00
|
|
|
)
|
2015-02-09 08:01:28 -07:00
|
|
|
if response is not None: # just an empty json object
|
2019-07-23 07:00:55 -06:00
|
|
|
return True
|
2015-02-04 10:32:44 -07:00
|
|
|
except CodeMessageException as e:
|
2015-02-09 08:01:28 -07:00
|
|
|
logger.warning("query_alias to %s received %s", uri, e.code)
|
2015-02-04 10:32:44 -07:00
|
|
|
if e.code == 404:
|
2019-07-23 07:00:55 -06:00
|
|
|
return False
|
2015-02-05 06:19:46 -07:00
|
|
|
except Exception as ex:
|
|
|
|
logger.warning("query_alias to %s threw exception %s", uri, ex)
|
2019-07-23 07:00:55 -06:00
|
|
|
return False
|
2015-02-05 06:19:46 -07:00
|
|
|
|
2021-12-14 10:02:46 -07:00
|
|
|
async def query_3pe(
|
|
|
|
self,
|
|
|
|
service: "ApplicationService",
|
|
|
|
kind: str,
|
|
|
|
protocol: str,
|
|
|
|
fields: Dict[bytes, List[bytes]],
|
|
|
|
) -> List[JsonDict]:
|
2016-08-18 10:19:55 -06:00
|
|
|
if kind == ThirdPartyEntityKind.USER:
|
2016-08-18 10:33:56 -06:00
|
|
|
required_field = "userid"
|
2016-08-18 10:19:55 -06:00
|
|
|
elif kind == ThirdPartyEntityKind.LOCATION:
|
2016-08-18 10:33:56 -06:00
|
|
|
required_field = "alias"
|
2016-08-18 10:19:55 -06:00
|
|
|
else:
|
|
|
|
raise ValueError("Unrecognised 'kind' argument %r to query_3pe()", kind)
|
2016-08-30 10:16:00 -06:00
|
|
|
if service.url is None:
|
2019-07-23 07:00:55 -06:00
|
|
|
return []
|
2016-08-17 06:15:06 -06:00
|
|
|
|
2022-05-24 08:39:54 -06:00
|
|
|
# This is required by the configuration.
|
|
|
|
assert service.hs_token is not None
|
|
|
|
|
2016-08-25 11:06:29 -06:00
|
|
|
uri = "%s%s/thirdparty/%s/%s" % (
|
|
|
|
service.url,
|
|
|
|
APP_SERVICE_PREFIX,
|
2016-08-25 11:35:38 -06:00
|
|
|
kind,
|
2018-09-05 08:10:47 -06:00
|
|
|
urllib.parse.quote(protocol),
|
2016-08-25 11:06:29 -06:00
|
|
|
)
|
2016-08-18 09:09:50 -06:00
|
|
|
try:
|
2022-05-24 08:39:54 -06:00
|
|
|
args: Mapping[Any, Any] = {
|
|
|
|
**fields,
|
|
|
|
b"access_token": service.hs_token,
|
|
|
|
}
|
2022-10-04 05:06:41 -06:00
|
|
|
response = await self.get_json(
|
2022-10-26 07:00:01 -06:00
|
|
|
uri,
|
|
|
|
args=args,
|
|
|
|
headers={"Authorization": [f"Bearer {service.hs_token}"]},
|
2022-10-04 05:06:41 -06:00
|
|
|
)
|
2016-08-18 10:33:56 -06:00
|
|
|
if not isinstance(response, list):
|
|
|
|
logger.warning(
|
|
|
|
"query_3pe to %s returned an invalid response %r", uri, response
|
|
|
|
)
|
2019-07-23 07:00:55 -06:00
|
|
|
return []
|
2016-08-18 10:33:56 -06:00
|
|
|
|
|
|
|
ret = []
|
|
|
|
for r in response:
|
|
|
|
if _is_valid_3pe_result(r, field=required_field):
|
|
|
|
ret.append(r)
|
|
|
|
else:
|
|
|
|
logger.warning(
|
|
|
|
"query_3pe to %s returned an invalid result %r", uri, r
|
|
|
|
)
|
|
|
|
|
2019-07-23 07:00:55 -06:00
|
|
|
return ret
|
2016-08-18 09:09:50 -06:00
|
|
|
except Exception as ex:
|
2016-08-18 10:19:55 -06:00
|
|
|
logger.warning("query_3pe to %s threw exception %s", uri, ex)
|
2019-07-23 07:00:55 -06:00
|
|
|
return []
|
2016-08-18 09:09:50 -06:00
|
|
|
|
2020-09-01 09:03:49 -06:00
|
|
|
async def get_3pe_protocol(
|
|
|
|
self, service: "ApplicationService", protocol: str
|
|
|
|
) -> Optional[JsonDict]:
|
2016-08-30 10:16:00 -06:00
|
|
|
if service.url is None:
|
2019-07-23 07:00:55 -06:00
|
|
|
return {}
|
2016-08-30 09:30:12 -06:00
|
|
|
|
2020-09-01 09:03:49 -06:00
|
|
|
async def _get() -> Optional[JsonDict]:
|
2022-05-24 08:39:54 -06:00
|
|
|
# This is required by the configuration.
|
|
|
|
assert service.hs_token is not None
|
2016-08-25 11:06:29 -06:00
|
|
|
uri = "%s%s/thirdparty/protocol/%s" % (
|
|
|
|
service.url,
|
|
|
|
APP_SERVICE_PREFIX,
|
2018-09-05 08:10:47 -06:00
|
|
|
urllib.parse.quote(protocol),
|
2016-08-25 11:06:29 -06:00
|
|
|
)
|
2016-08-25 08:56:27 -06:00
|
|
|
try:
|
2022-10-04 05:06:41 -06:00
|
|
|
info = await self.get_json(
|
|
|
|
uri,
|
|
|
|
{"access_token": service.hs_token},
|
2022-10-26 07:00:01 -06:00
|
|
|
headers={"Authorization": [f"Bearer {service.hs_token}"]},
|
2022-10-04 05:06:41 -06:00
|
|
|
)
|
2016-09-09 08:07:04 -06:00
|
|
|
|
2016-09-09 08:09:46 -06:00
|
|
|
if not _is_valid_3pe_metadata(info):
|
|
|
|
logger.warning(
|
2019-11-21 05:00:14 -07:00
|
|
|
"query_3pe_protocol to %s did not return a valid result", uri
|
2016-09-09 08:09:46 -06:00
|
|
|
)
|
2019-07-23 07:00:55 -06:00
|
|
|
return None
|
2016-09-09 08:07:04 -06:00
|
|
|
|
2016-12-06 03:43:48 -07:00
|
|
|
for instance in info.get("instances", []):
|
|
|
|
network_id = instance.get("network_id", None)
|
|
|
|
if network_id is not None:
|
2016-12-12 07:46:13 -07:00
|
|
|
instance["instance_id"] = ThirdPartyInstanceID(
|
2016-12-06 03:43:48 -07:00
|
|
|
service.id, network_id
|
|
|
|
).to_string()
|
|
|
|
|
2019-07-23 07:00:55 -06:00
|
|
|
return info
|
2016-08-25 08:56:27 -06:00
|
|
|
except Exception as ex:
|
|
|
|
logger.warning("query_3pe_protocol to %s threw exception %s", uri, ex)
|
2019-07-23 07:00:55 -06:00
|
|
|
return None
|
2016-08-25 08:56:27 -06:00
|
|
|
|
|
|
|
key = (service.id, protocol)
|
2020-09-01 09:03:49 -06:00
|
|
|
return await self.protocol_meta_cache.wrap(key, _get)
|
2016-08-25 08:10:06 -06:00
|
|
|
|
2020-10-15 10:33:28 -06:00
|
|
|
async def push_bulk(
|
|
|
|
self,
|
|
|
|
service: "ApplicationService",
|
2023-02-06 05:49:06 -07:00
|
|
|
events: Sequence[EventBase],
|
2020-10-15 10:33:28 -06:00
|
|
|
ephemeral: List[JsonDict],
|
2022-02-01 07:13:38 -07:00
|
|
|
to_device_messages: List[JsonDict],
|
2022-11-28 09:17:29 -07:00
|
|
|
one_time_keys_count: TransactionOneTimeKeysCount,
|
2022-02-24 10:55:45 -07:00
|
|
|
unused_fallback_keys: TransactionUnusedFallbackKeys,
|
2022-03-30 07:39:27 -06:00
|
|
|
device_list_summary: DeviceListUpdates,
|
2020-10-15 10:33:28 -06:00
|
|
|
txn_id: Optional[int] = None,
|
2021-12-14 10:02:46 -07:00
|
|
|
) -> bool:
|
2022-02-01 07:13:38 -07:00
|
|
|
"""
|
|
|
|
Push data to an application service.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
service: The application service to send to.
|
|
|
|
events: The persistent events to send.
|
|
|
|
ephemeral: The ephemeral events to send.
|
|
|
|
to_device_messages: The to-device messages to send.
|
|
|
|
txn_id: An unique ID to assign to this transaction. Application services should
|
|
|
|
deduplicate transactions received with identitical IDs.
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
True if the task succeeded, False if it failed.
|
|
|
|
"""
|
2016-08-30 10:16:00 -06:00
|
|
|
if service.url is None:
|
2019-07-23 07:00:55 -06:00
|
|
|
return True
|
2016-08-30 09:21:16 -06:00
|
|
|
|
2021-12-14 10:02:46 -07:00
|
|
|
# This is required by the configuration.
|
|
|
|
assert service.hs_token is not None
|
|
|
|
|
|
|
|
serialized_events = self._serialize(service, events)
|
2015-02-05 06:42:35 -07:00
|
|
|
|
2015-03-05 08:40:07 -07:00
|
|
|
if txn_id is None:
|
|
|
|
logger.warning(
|
|
|
|
"push_bulk: Missing txn ID sending events to %s", service.url
|
|
|
|
)
|
2020-10-15 10:33:28 -06:00
|
|
|
txn_id = 0
|
|
|
|
|
|
|
|
uri = service.url + ("/transactions/%s" % urllib.parse.quote(str(txn_id)))
|
|
|
|
|
|
|
|
# Never send ephemeral events to appservices that do not support it
|
2022-02-24 10:55:45 -07:00
|
|
|
body: JsonDict = {"events": serialized_events}
|
2020-10-15 10:33:28 -06:00
|
|
|
if service.supports_ephemeral:
|
2022-02-01 07:13:38 -07:00
|
|
|
body.update(
|
|
|
|
{
|
|
|
|
# TODO: Update to stable prefixes once MSC2409 completes FCP merge.
|
|
|
|
"de.sorunome.msc2409.ephemeral": ephemeral,
|
|
|
|
"de.sorunome.msc2409.to_device": to_device_messages,
|
|
|
|
}
|
|
|
|
)
|
2015-03-05 08:40:07 -07:00
|
|
|
|
2022-03-30 07:39:27 -06:00
|
|
|
# TODO: Update to stable prefixes once MSC3202 completes FCP merge
|
2022-02-24 10:55:45 -07:00
|
|
|
if service.msc3202_transaction_extensions:
|
2022-11-28 09:17:29 -07:00
|
|
|
if one_time_keys_count:
|
2022-02-24 10:55:45 -07:00
|
|
|
body[
|
|
|
|
"org.matrix.msc3202.device_one_time_key_counts"
|
2022-11-28 09:17:29 -07:00
|
|
|
] = one_time_keys_count
|
|
|
|
body[
|
|
|
|
"org.matrix.msc3202.device_one_time_keys_count"
|
|
|
|
] = one_time_keys_count
|
2022-02-24 10:55:45 -07:00
|
|
|
if unused_fallback_keys:
|
|
|
|
body[
|
2022-04-22 09:03:46 -06:00
|
|
|
"org.matrix.msc3202.device_unused_fallback_key_types"
|
2022-02-24 10:55:45 -07:00
|
|
|
] = unused_fallback_keys
|
2022-03-30 07:39:27 -06:00
|
|
|
if device_list_summary:
|
|
|
|
body["org.matrix.msc3202.device_lists"] = {
|
|
|
|
"changed": list(device_list_summary.changed),
|
|
|
|
"left": list(device_list_summary.left),
|
|
|
|
}
|
2022-02-24 10:55:45 -07:00
|
|
|
|
2015-02-05 02:43:22 -07:00
|
|
|
try:
|
2020-07-30 05:27:39 -06:00
|
|
|
await self.put_json(
|
2020-10-15 10:33:28 -06:00
|
|
|
uri=uri,
|
|
|
|
json_body=body,
|
|
|
|
args={"access_token": service.hs_token},
|
2022-10-26 07:00:01 -06:00
|
|
|
headers={"Authorization": [f"Bearer {service.hs_token}"]},
|
2015-02-05 02:43:22 -07:00
|
|
|
)
|
2021-11-18 13:16:08 -07:00
|
|
|
if logger.isEnabledFor(logging.DEBUG):
|
|
|
|
logger.debug(
|
|
|
|
"push_bulk to %s succeeded! events=%s",
|
|
|
|
uri,
|
|
|
|
[event.get("event_id") for event in events],
|
|
|
|
)
|
2018-06-05 10:30:45 -06:00
|
|
|
sent_transactions_counter.labels(service.id).inc()
|
2021-12-14 10:02:46 -07:00
|
|
|
sent_events_counter.labels(service.id).inc(len(serialized_events))
|
2022-07-13 08:18:20 -06:00
|
|
|
sent_ephemeral_counter.labels(service.id).inc(len(ephemeral))
|
|
|
|
sent_todevice_counter.labels(service.id).inc(len(to_device_messages))
|
2019-07-23 07:00:55 -06:00
|
|
|
return True
|
2015-02-05 02:43:22 -07:00
|
|
|
except CodeMessageException as e:
|
2021-11-18 13:16:08 -07:00
|
|
|
logger.warning(
|
|
|
|
"push_bulk to %s received code=%s msg=%s",
|
|
|
|
uri,
|
|
|
|
e.code,
|
|
|
|
e.msg,
|
|
|
|
exc_info=logger.isEnabledFor(logging.DEBUG),
|
|
|
|
)
|
2015-02-05 06:19:46 -07:00
|
|
|
except Exception as ex:
|
2021-11-18 13:16:08 -07:00
|
|
|
logger.warning(
|
|
|
|
"push_bulk to %s threw exception(%s) %s args=%s",
|
|
|
|
uri,
|
|
|
|
type(ex).__name__,
|
|
|
|
ex,
|
|
|
|
ex.args,
|
|
|
|
exc_info=logger.isEnabledFor(logging.DEBUG),
|
|
|
|
)
|
2018-06-05 10:30:45 -06:00
|
|
|
failed_transactions_counter.labels(service.id).inc()
|
2019-07-23 07:00:55 -06:00
|
|
|
return False
|
2015-02-04 04:19:18 -07:00
|
|
|
|
2021-12-14 10:02:46 -07:00
|
|
|
def _serialize(
|
|
|
|
self, service: "ApplicationService", events: Iterable[EventBase]
|
|
|
|
) -> List[JsonDict]:
|
2015-02-05 06:42:35 -07:00
|
|
|
time_now = self.clock.time_msec()
|
2020-07-10 11:44:56 -06:00
|
|
|
return [
|
|
|
|
serialize_event(
|
|
|
|
e,
|
|
|
|
time_now,
|
2022-03-03 08:43:06 -07:00
|
|
|
config=SerializeEventConfig(
|
|
|
|
as_client_event=True,
|
|
|
|
# If this is an invite or a knock membership event, and we're interested
|
|
|
|
# in this user, then include any stripped state alongside the event.
|
|
|
|
include_stripped_room_state=(
|
|
|
|
e.type == EventTypes.Member
|
|
|
|
and (
|
|
|
|
e.membership == Membership.INVITE
|
|
|
|
or e.membership == Membership.KNOCK
|
|
|
|
)
|
|
|
|
and service.is_interested_in_user(e.state_key)
|
|
|
|
),
|
2020-07-10 11:44:56 -06:00
|
|
|
),
|
|
|
|
)
|
|
|
|
for e in events
|
|
|
|
]
|