2015-02-24 11:03:39 -07:00
|
|
|
# -*- coding: utf-8 -*-
|
2016-01-06 21:26:29 -07:00
|
|
|
# Copyright 2015, 2016 OpenMarket Ltd
|
2015-02-24 11:03:39 -07:00
|
|
|
#
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
|
|
|
|
2015-08-13 04:38:59 -06:00
|
|
|
import functools
|
2016-05-09 03:13:25 -06:00
|
|
|
import gc
|
2020-09-29 15:26:28 -06:00
|
|
|
import itertools
|
2018-07-09 00:09:20 -06:00
|
|
|
import logging
|
2018-05-23 12:03:56 -06:00
|
|
|
import os
|
2018-01-07 17:53:32 -07:00
|
|
|
import platform
|
2018-09-14 07:39:59 -06:00
|
|
|
import threading
|
2018-07-09 00:09:20 -06:00
|
|
|
import time
|
2020-03-19 04:00:24 -06:00
|
|
|
from typing import Callable, Dict, Iterable, Optional, Tuple, Union
|
2015-08-13 04:38:59 -06:00
|
|
|
|
2018-07-09 00:09:20 -06:00
|
|
|
import attr
|
|
|
|
from prometheus_client import Counter, Gauge, Histogram
|
2020-05-22 04:08:41 -06:00
|
|
|
from prometheus_client.core import (
|
|
|
|
REGISTRY,
|
|
|
|
CounterMetricFamily,
|
2020-09-29 15:26:28 -06:00
|
|
|
GaugeHistogramMetricFamily,
|
2020-05-22 04:08:41 -06:00
|
|
|
GaugeMetricFamily,
|
|
|
|
)
|
2015-03-04 10:13:09 -07:00
|
|
|
|
2018-05-21 18:47:37 -06:00
|
|
|
from twisted.internet import reactor
|
2015-02-24 11:03:39 -07:00
|
|
|
|
2019-09-09 08:14:58 -06:00
|
|
|
import synapse
|
2019-07-18 07:57:15 -06:00
|
|
|
from synapse.metrics._exposition import (
|
|
|
|
MetricsResource,
|
|
|
|
generate_latest,
|
|
|
|
start_http_server,
|
|
|
|
)
|
2019-09-09 08:14:58 -06:00
|
|
|
from synapse.util.versionstring import get_version_string
|
2019-07-18 07:57:15 -06:00
|
|
|
|
2015-03-04 10:13:09 -07:00
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
2019-07-18 07:57:15 -06:00
|
|
|
METRICS_PREFIX = "/_synapse/metrics"
|
|
|
|
|
2018-05-22 16:32:57 -06:00
|
|
|
running_on_pypy = platform.python_implementation() == "PyPy"
|
2020-09-29 15:26:28 -06:00
|
|
|
all_gauges = {} # type: Dict[str, Union[LaterGauge, InFlightGauge]]
|
2015-02-24 11:03:39 -07:00
|
|
|
|
2018-05-28 19:22:27 -06:00
|
|
|
HAVE_PROC_SELF_STAT = os.path.exists("/proc/self/stat")
|
2018-05-22 15:28:23 -06:00
|
|
|
|
2018-05-28 19:32:15 -06:00
|
|
|
|
2020-09-04 04:54:56 -06:00
|
|
|
class RegistryProxy:
|
2018-05-31 03:04:50 -06:00
|
|
|
@staticmethod
|
|
|
|
def collect():
|
2018-05-22 15:28:23 -06:00
|
|
|
for metric in REGISTRY.collect():
|
|
|
|
if not metric.name.startswith("__"):
|
|
|
|
yield metric
|
|
|
|
|
|
|
|
|
2020-09-14 10:50:06 -06:00
|
|
|
@attr.s(slots=True, hash=True)
|
2020-09-04 04:54:56 -06:00
|
|
|
class LaterGauge:
|
2015-02-24 11:03:39 -07:00
|
|
|
|
2020-03-19 04:00:24 -06:00
|
|
|
name = attr.ib(type=str)
|
|
|
|
desc = attr.ib(type=str)
|
|
|
|
labels = attr.ib(hash=False, type=Optional[Iterable[str]])
|
|
|
|
# callback: should either return a value (if there are no labels for this metric),
|
|
|
|
# or dict mapping from a label tuple to a value
|
|
|
|
caller = attr.ib(type=Callable[[], Union[Dict[Tuple[str, ...], float], float]])
|
2018-04-11 04:07:33 -06:00
|
|
|
|
2018-05-21 18:47:37 -06:00
|
|
|
def collect(self):
|
2015-03-04 08:47:23 -07:00
|
|
|
|
2018-05-22 15:28:23 -06:00
|
|
|
g = GaugeMetricFamily(self.name, self.desc, labels=self.labels)
|
2015-03-04 08:47:23 -07:00
|
|
|
|
2018-05-21 18:47:37 -06:00
|
|
|
try:
|
|
|
|
calls = self.caller()
|
2018-06-04 08:59:14 -06:00
|
|
|
except Exception:
|
2019-06-13 06:40:52 -06:00
|
|
|
logger.exception("Exception running callback for LaterGauge(%s)", self.name)
|
2018-05-21 18:47:37 -06:00
|
|
|
yield g
|
2018-06-04 08:59:14 -06:00
|
|
|
return
|
2015-02-24 11:03:39 -07:00
|
|
|
|
2018-05-21 18:47:37 -06:00
|
|
|
if isinstance(calls, dict):
|
2020-06-15 05:03:36 -06:00
|
|
|
for k, v in calls.items():
|
2018-05-21 18:47:37 -06:00
|
|
|
g.add_metric(k, v)
|
|
|
|
else:
|
|
|
|
g.add_metric([], calls)
|
2016-07-20 08:47:28 -06:00
|
|
|
|
2018-05-21 18:47:37 -06:00
|
|
|
yield g
|
2016-07-20 08:47:28 -06:00
|
|
|
|
2018-05-22 09:50:26 -06:00
|
|
|
def __attrs_post_init__(self):
|
|
|
|
self._register()
|
2016-07-20 08:47:28 -06:00
|
|
|
|
2018-05-22 09:50:26 -06:00
|
|
|
def _register(self):
|
2018-05-21 18:47:37 -06:00
|
|
|
if self.name in all_gauges.keys():
|
2018-05-22 09:50:26 -06:00
|
|
|
logger.warning("%s already registered, reregistering" % (self.name,))
|
2018-09-14 07:39:59 -06:00
|
|
|
REGISTRY.unregister(all_gauges.pop(self.name))
|
|
|
|
|
|
|
|
REGISTRY.register(self)
|
|
|
|
all_gauges[self.name] = self
|
|
|
|
|
|
|
|
|
2020-09-04 04:54:56 -06:00
|
|
|
class InFlightGauge:
|
2018-09-14 07:39:59 -06:00
|
|
|
"""Tracks number of things (e.g. requests, Measure blocks, etc) in flight
|
|
|
|
at any given time.
|
|
|
|
|
|
|
|
Each InFlightGauge will create a metric called `<name>_total` that counts
|
|
|
|
the number of in flight blocks, as well as a metrics for each item in the
|
|
|
|
given `sub_metrics` as `<name>_<sub_metric>` which will get updated by the
|
|
|
|
callbacks.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
name (str)
|
|
|
|
desc (str)
|
|
|
|
labels (list[str])
|
|
|
|
sub_metrics (list[str]): A list of sub metrics that the callbacks
|
|
|
|
will update.
|
|
|
|
"""
|
|
|
|
|
|
|
|
def __init__(self, name, desc, labels, sub_metrics):
|
|
|
|
self.name = name
|
|
|
|
self.desc = desc
|
|
|
|
self.labels = labels
|
|
|
|
self.sub_metrics = sub_metrics
|
|
|
|
|
|
|
|
# Create a class which have the sub_metrics values as attributes, which
|
|
|
|
# default to 0 on initialization. Used to pass to registered callbacks.
|
|
|
|
self._metrics_class = attr.make_class(
|
2019-06-13 06:40:52 -06:00
|
|
|
"_MetricsEntry", attrs={x: attr.ib(0) for x in sub_metrics}, slots=True
|
2018-09-14 07:39:59 -06:00
|
|
|
)
|
|
|
|
|
|
|
|
# Counts number of in flight blocks for a given set of label values
|
2019-10-02 06:29:01 -06:00
|
|
|
self._registrations = {} # type: Dict
|
2018-09-14 07:39:59 -06:00
|
|
|
|
|
|
|
# Protects access to _registrations
|
|
|
|
self._lock = threading.Lock()
|
|
|
|
|
|
|
|
self._register_with_collector()
|
|
|
|
|
|
|
|
def register(self, key, callback):
|
|
|
|
"""Registers that we've entered a new block with labels `key`.
|
|
|
|
|
|
|
|
`callback` gets called each time the metrics are collected. The same
|
|
|
|
value must also be given to `unregister`.
|
|
|
|
|
|
|
|
`callback` gets called with an object that has an attribute per
|
|
|
|
sub_metric, which should be updated with the necessary values. Note that
|
|
|
|
the metrics object is shared between all callbacks registered with the
|
|
|
|
same key.
|
|
|
|
|
|
|
|
Note that `callback` may be called on a separate thread.
|
|
|
|
"""
|
|
|
|
with self._lock:
|
|
|
|
self._registrations.setdefault(key, set()).add(callback)
|
|
|
|
|
|
|
|
def unregister(self, key, callback):
|
2021-02-16 15:32:34 -07:00
|
|
|
"""Registers that we've exited a block with labels `key`."""
|
2018-09-14 07:39:59 -06:00
|
|
|
|
|
|
|
with self._lock:
|
|
|
|
self._registrations.setdefault(key, set()).discard(callback)
|
|
|
|
|
|
|
|
def collect(self):
|
|
|
|
"""Called by prometheus client when it reads metrics.
|
|
|
|
|
|
|
|
Note: may be called by a separate thread.
|
|
|
|
"""
|
2019-06-13 06:40:52 -06:00
|
|
|
in_flight = GaugeMetricFamily(
|
|
|
|
self.name + "_total", self.desc, labels=self.labels
|
|
|
|
)
|
2018-09-14 07:39:59 -06:00
|
|
|
|
|
|
|
metrics_by_key = {}
|
|
|
|
|
|
|
|
# We copy so that we don't mutate the list while iterating
|
|
|
|
with self._lock:
|
|
|
|
keys = list(self._registrations)
|
|
|
|
|
|
|
|
for key in keys:
|
|
|
|
with self._lock:
|
|
|
|
callbacks = set(self._registrations[key])
|
|
|
|
|
|
|
|
in_flight.add_metric(key, len(callbacks))
|
|
|
|
|
|
|
|
metrics = self._metrics_class()
|
|
|
|
metrics_by_key[key] = metrics
|
|
|
|
for callback in callbacks:
|
|
|
|
callback(metrics)
|
|
|
|
|
|
|
|
yield in_flight
|
|
|
|
|
|
|
|
for name in self.sub_metrics:
|
2019-06-13 06:40:52 -06:00
|
|
|
gauge = GaugeMetricFamily(
|
|
|
|
"_".join([self.name, name]), "", labels=self.labels
|
|
|
|
)
|
2020-06-15 05:03:36 -06:00
|
|
|
for key, metrics in metrics_by_key.items():
|
2018-09-14 07:39:59 -06:00
|
|
|
gauge.add_metric(key, getattr(metrics, name))
|
|
|
|
yield gauge
|
|
|
|
|
|
|
|
def _register_with_collector(self):
|
|
|
|
if self.name in all_gauges.keys():
|
|
|
|
logger.warning("%s already registered, reregistering" % (self.name,))
|
2018-05-21 18:47:37 -06:00
|
|
|
REGISTRY.unregister(all_gauges.pop(self.name))
|
2015-03-06 09:18:21 -07:00
|
|
|
|
2018-05-21 18:47:37 -06:00
|
|
|
REGISTRY.register(self)
|
|
|
|
all_gauges[self.name] = self
|
2015-02-24 11:03:39 -07:00
|
|
|
|
|
|
|
|
2020-09-29 15:26:28 -06:00
|
|
|
class GaugeBucketCollector:
|
|
|
|
"""Like a Histogram, but the buckets are Gauges which are updated atomically.
|
2019-06-13 06:40:52 -06:00
|
|
|
|
2020-09-29 15:26:28 -06:00
|
|
|
The data is updated by calling `update_data` with an iterable of measurements.
|
2019-06-13 06:40:52 -06:00
|
|
|
|
2020-09-29 15:26:28 -06:00
|
|
|
We assume that the data is updated less frequently than it is reported to
|
|
|
|
Prometheus, and optimise for that case.
|
2019-06-13 06:40:52 -06:00
|
|
|
"""
|
|
|
|
|
2020-09-29 15:26:28 -06:00
|
|
|
__slots__ = ("_name", "_documentation", "_bucket_bounds", "_metric")
|
2019-06-13 06:40:52 -06:00
|
|
|
|
2020-09-29 15:26:28 -06:00
|
|
|
def __init__(
|
|
|
|
self,
|
|
|
|
name: str,
|
|
|
|
documentation: str,
|
|
|
|
buckets: Iterable[float],
|
|
|
|
registry=REGISTRY,
|
|
|
|
):
|
|
|
|
"""
|
|
|
|
Args:
|
|
|
|
name: base name of metric to be exported to Prometheus. (a _bucket suffix
|
|
|
|
will be added.)
|
|
|
|
documentation: help text for the metric
|
|
|
|
buckets: The top bounds of the buckets to report
|
|
|
|
registry: metric registry to register with
|
|
|
|
"""
|
|
|
|
self._name = name
|
|
|
|
self._documentation = documentation
|
2019-06-13 06:40:52 -06:00
|
|
|
|
2020-09-29 15:26:28 -06:00
|
|
|
# the tops of the buckets
|
|
|
|
self._bucket_bounds = [float(b) for b in buckets]
|
|
|
|
if self._bucket_bounds != sorted(self._bucket_bounds):
|
|
|
|
raise ValueError("Buckets not in sorted order")
|
2019-06-13 06:40:52 -06:00
|
|
|
|
2020-09-29 15:26:28 -06:00
|
|
|
if self._bucket_bounds[-1] != float("inf"):
|
|
|
|
self._bucket_bounds.append(float("inf"))
|
2019-06-13 06:40:52 -06:00
|
|
|
|
2020-09-29 15:26:28 -06:00
|
|
|
self._metric = self._values_to_metric([])
|
|
|
|
registry.register(self)
|
2019-06-13 06:40:52 -06:00
|
|
|
|
2020-09-29 15:26:28 -06:00
|
|
|
def collect(self):
|
|
|
|
yield self._metric
|
2019-06-13 06:40:52 -06:00
|
|
|
|
2020-09-29 15:26:28 -06:00
|
|
|
def update_data(self, values: Iterable[float]):
|
|
|
|
"""Update the data to be reported by the metric
|
2019-06-13 06:40:52 -06:00
|
|
|
|
2020-09-29 15:26:28 -06:00
|
|
|
The existing data is cleared, and each measurement in the input is assigned
|
|
|
|
to the relevant bucket.
|
|
|
|
"""
|
|
|
|
self._metric = self._values_to_metric(values)
|
|
|
|
|
|
|
|
def _values_to_metric(self, values: Iterable[float]) -> GaugeHistogramMetricFamily:
|
|
|
|
total = 0.0
|
|
|
|
bucket_values = [0 for _ in self._bucket_bounds]
|
|
|
|
|
|
|
|
for v in values:
|
|
|
|
# assign each value to a bucket
|
|
|
|
for i, bound in enumerate(self._bucket_bounds):
|
|
|
|
if v <= bound:
|
|
|
|
bucket_values[i] += 1
|
|
|
|
break
|
|
|
|
|
|
|
|
# ... and increment the sum
|
|
|
|
total += v
|
|
|
|
|
|
|
|
# now, aggregate the bucket values so that they count the number of entries in
|
|
|
|
# that bucket or below.
|
|
|
|
accumulated_values = itertools.accumulate(bucket_values)
|
|
|
|
|
|
|
|
return GaugeHistogramMetricFamily(
|
|
|
|
self._name,
|
|
|
|
self._documentation,
|
|
|
|
buckets=list(
|
|
|
|
zip((str(b) for b in self._bucket_bounds), accumulated_values)
|
|
|
|
),
|
|
|
|
gsum_value=total,
|
2019-06-13 06:40:52 -06:00
|
|
|
)
|
|
|
|
|
|
|
|
|
2018-05-23 12:03:56 -06:00
|
|
|
#
|
|
|
|
# Detailed CPU metrics
|
|
|
|
#
|
|
|
|
|
|
|
|
|
2020-09-04 04:54:56 -06:00
|
|
|
class CPUMetrics:
|
2018-05-23 12:03:56 -06:00
|
|
|
def __init__(self):
|
|
|
|
ticks_per_sec = 100
|
|
|
|
try:
|
|
|
|
# Try and get the system config
|
2019-06-20 03:32:02 -06:00
|
|
|
ticks_per_sec = os.sysconf("SC_CLK_TCK")
|
2018-05-23 12:03:56 -06:00
|
|
|
except (ValueError, TypeError, AttributeError):
|
|
|
|
pass
|
|
|
|
|
|
|
|
self.ticks_per_sec = ticks_per_sec
|
|
|
|
|
|
|
|
def collect(self):
|
2018-05-28 19:22:27 -06:00
|
|
|
if not HAVE_PROC_SELF_STAT:
|
|
|
|
return
|
2018-05-23 12:03:56 -06:00
|
|
|
|
|
|
|
with open("/proc/self/stat") as s:
|
|
|
|
line = s.read()
|
|
|
|
raw_stats = line.split(") ", 1)[1].split(" ")
|
|
|
|
|
|
|
|
user = GaugeMetricFamily("process_cpu_user_seconds_total", "")
|
|
|
|
user.add_metric([], float(raw_stats[11]) / self.ticks_per_sec)
|
|
|
|
yield user
|
|
|
|
|
|
|
|
sys = GaugeMetricFamily("process_cpu_system_seconds_total", "")
|
|
|
|
sys.add_metric([], float(raw_stats[12]) / self.ticks_per_sec)
|
|
|
|
yield sys
|
|
|
|
|
2018-05-23 12:08:59 -06:00
|
|
|
|
2018-05-23 12:03:56 -06:00
|
|
|
REGISTRY.register(CPUMetrics())
|
|
|
|
|
2018-05-21 18:47:37 -06:00
|
|
|
#
|
|
|
|
# Python GC metrics
|
|
|
|
#
|
2015-02-24 11:03:39 -07:00
|
|
|
|
2018-05-21 18:47:37 -06:00
|
|
|
gc_unreachable = Gauge("python_gc_unreachable_total", "Unreachable GC objects", ["gen"])
|
2018-05-22 16:32:57 -06:00
|
|
|
gc_time = Histogram(
|
|
|
|
"python_gc_time",
|
2018-05-28 03:10:27 -06:00
|
|
|
"Time taken to GC (sec)",
|
2018-05-22 16:32:57 -06:00
|
|
|
["gen"],
|
2019-06-13 06:40:52 -06:00
|
|
|
buckets=[
|
|
|
|
0.0025,
|
|
|
|
0.005,
|
|
|
|
0.01,
|
|
|
|
0.025,
|
|
|
|
0.05,
|
|
|
|
0.10,
|
|
|
|
0.25,
|
|
|
|
0.50,
|
|
|
|
1.00,
|
|
|
|
2.50,
|
|
|
|
5.00,
|
|
|
|
7.50,
|
|
|
|
15.00,
|
|
|
|
30.00,
|
|
|
|
45.00,
|
|
|
|
60.00,
|
|
|
|
],
|
2018-05-22 16:32:57 -06:00
|
|
|
)
|
|
|
|
|
2015-03-05 09:15:21 -07:00
|
|
|
|
2020-09-04 04:54:56 -06:00
|
|
|
class GCCounts:
|
2018-05-21 18:47:37 -06:00
|
|
|
def collect(self):
|
2018-06-21 03:02:42 -06:00
|
|
|
cm = GaugeMetricFamily("python_gc_counts", "GC object counts", labels=["gen"])
|
2018-05-21 18:47:37 -06:00
|
|
|
for n, m in enumerate(gc.get_count()):
|
|
|
|
cm.add_metric([str(n)], m)
|
2015-03-06 12:08:47 -07:00
|
|
|
|
2018-05-21 18:47:37 -06:00
|
|
|
yield cm
|
2015-03-05 09:15:21 -07:00
|
|
|
|
2018-05-22 16:32:57 -06:00
|
|
|
|
2018-06-28 07:49:57 -06:00
|
|
|
if not running_on_pypy:
|
|
|
|
REGISTRY.register(GCCounts())
|
2015-03-05 09:15:21 -07:00
|
|
|
|
2020-05-22 04:08:41 -06:00
|
|
|
|
|
|
|
#
|
|
|
|
# PyPy GC / memory metrics
|
|
|
|
#
|
|
|
|
|
|
|
|
|
2020-09-04 04:54:56 -06:00
|
|
|
class PyPyGCStats:
|
2020-05-22 04:08:41 -06:00
|
|
|
def collect(self):
|
|
|
|
|
|
|
|
# @stats is a pretty-printer object with __str__() returning a nice table,
|
|
|
|
# plus some fields that contain data from that table.
|
|
|
|
# unfortunately, fields are pretty-printed themselves (i. e. '4.5MB').
|
|
|
|
stats = gc.get_stats(memory_pressure=False) # type: ignore
|
|
|
|
# @s contains same fields as @stats, but as actual integers.
|
|
|
|
s = stats._s # type: ignore
|
|
|
|
|
|
|
|
# also note that field naming is completely braindead
|
|
|
|
# and only vaguely correlates with the pretty-printed table.
|
|
|
|
# >>>> gc.get_stats(False)
|
|
|
|
# Total memory consumed:
|
|
|
|
# GC used: 8.7MB (peak: 39.0MB) # s.total_gc_memory, s.peak_memory
|
|
|
|
# in arenas: 3.0MB # s.total_arena_memory
|
|
|
|
# rawmalloced: 1.7MB # s.total_rawmalloced_memory
|
|
|
|
# nursery: 4.0MB # s.nursery_size
|
|
|
|
# raw assembler used: 31.0kB # s.jit_backend_used
|
|
|
|
# -----------------------------
|
|
|
|
# Total: 8.8MB # stats.memory_used_sum
|
|
|
|
#
|
|
|
|
# Total memory allocated:
|
|
|
|
# GC allocated: 38.7MB (peak: 41.1MB) # s.total_allocated_memory, s.peak_allocated_memory
|
|
|
|
# in arenas: 30.9MB # s.peak_arena_memory
|
|
|
|
# rawmalloced: 4.1MB # s.peak_rawmalloced_memory
|
|
|
|
# nursery: 4.0MB # s.nursery_size
|
|
|
|
# raw assembler allocated: 1.0MB # s.jit_backend_allocated
|
|
|
|
# -----------------------------
|
|
|
|
# Total: 39.7MB # stats.memory_allocated_sum
|
|
|
|
#
|
|
|
|
# Total time spent in GC: 0.073 # s.total_gc_time
|
|
|
|
|
|
|
|
pypy_gc_time = CounterMetricFamily(
|
2021-02-16 15:32:34 -07:00
|
|
|
"pypy_gc_time_seconds_total",
|
|
|
|
"Total time spent in PyPy GC",
|
|
|
|
labels=[],
|
2020-05-22 04:08:41 -06:00
|
|
|
)
|
|
|
|
pypy_gc_time.add_metric([], s.total_gc_time / 1000)
|
|
|
|
yield pypy_gc_time
|
|
|
|
|
|
|
|
pypy_mem = GaugeMetricFamily(
|
|
|
|
"pypy_memory_bytes",
|
|
|
|
"Memory tracked by PyPy allocator",
|
|
|
|
labels=["state", "class", "kind"],
|
|
|
|
)
|
|
|
|
# memory used by JIT assembler
|
|
|
|
pypy_mem.add_metric(["used", "", "jit"], s.jit_backend_used)
|
|
|
|
pypy_mem.add_metric(["allocated", "", "jit"], s.jit_backend_allocated)
|
|
|
|
# memory used by GCed objects
|
|
|
|
pypy_mem.add_metric(["used", "", "arenas"], s.total_arena_memory)
|
|
|
|
pypy_mem.add_metric(["allocated", "", "arenas"], s.peak_arena_memory)
|
|
|
|
pypy_mem.add_metric(["used", "", "rawmalloced"], s.total_rawmalloced_memory)
|
|
|
|
pypy_mem.add_metric(["allocated", "", "rawmalloced"], s.peak_rawmalloced_memory)
|
|
|
|
pypy_mem.add_metric(["used", "", "nursery"], s.nursery_size)
|
|
|
|
pypy_mem.add_metric(["allocated", "", "nursery"], s.nursery_size)
|
|
|
|
# totals
|
|
|
|
pypy_mem.add_metric(["used", "totals", "gc"], s.total_gc_memory)
|
|
|
|
pypy_mem.add_metric(["allocated", "totals", "gc"], s.total_allocated_memory)
|
|
|
|
pypy_mem.add_metric(["used", "totals", "gc_peak"], s.peak_memory)
|
|
|
|
pypy_mem.add_metric(["allocated", "totals", "gc_peak"], s.peak_allocated_memory)
|
|
|
|
yield pypy_mem
|
|
|
|
|
|
|
|
|
|
|
|
if running_on_pypy:
|
|
|
|
REGISTRY.register(PyPyGCStats())
|
|
|
|
|
|
|
|
|
2018-05-21 18:47:37 -06:00
|
|
|
#
|
|
|
|
# Twisted reactor metrics
|
|
|
|
#
|
2016-11-03 11:03:52 -06:00
|
|
|
|
2018-05-22 16:32:57 -06:00
|
|
|
tick_time = Histogram(
|
|
|
|
"python_twisted_reactor_tick_time",
|
2018-05-28 03:10:27 -06:00
|
|
|
"Tick time of the Twisted reactor (sec)",
|
2018-05-28 03:16:09 -06:00
|
|
|
buckets=[0.001, 0.002, 0.005, 0.01, 0.025, 0.05, 0.1, 0.2, 0.5, 1, 2, 5],
|
2018-05-22 16:32:57 -06:00
|
|
|
)
|
|
|
|
pending_calls_metric = Histogram(
|
|
|
|
"python_twisted_reactor_pending_calls",
|
|
|
|
"Pending calls",
|
|
|
|
buckets=[1, 2, 5, 10, 25, 50, 100, 250, 500, 1000],
|
|
|
|
)
|
2015-08-13 04:38:59 -06:00
|
|
|
|
2018-05-21 18:47:37 -06:00
|
|
|
#
|
|
|
|
# Federation Metrics
|
|
|
|
#
|
2015-08-13 04:38:59 -06:00
|
|
|
|
2018-05-21 18:47:37 -06:00
|
|
|
sent_transactions_counter = Counter("synapse_federation_client_sent_transactions", "")
|
2016-10-27 11:09:34 -06:00
|
|
|
|
2018-05-21 18:47:37 -06:00
|
|
|
events_processed_counter = Counter("synapse_federation_client_events_processed", "")
|
2018-04-11 04:07:51 -06:00
|
|
|
|
2018-08-07 12:09:48 -06:00
|
|
|
event_processing_loop_counter = Counter(
|
2019-06-13 06:40:52 -06:00
|
|
|
"synapse_event_processing_loop_count", "Event processing loop iterations", ["name"]
|
2018-08-07 12:09:48 -06:00
|
|
|
)
|
|
|
|
|
|
|
|
event_processing_loop_room_count = Counter(
|
|
|
|
"synapse_event_processing_loop_room_count",
|
|
|
|
"Rooms seen per event processing loop iteration",
|
|
|
|
["name"],
|
|
|
|
)
|
|
|
|
|
|
|
|
|
2018-04-11 04:07:51 -06:00
|
|
|
# Used to track where various components have processed in the event stream,
|
|
|
|
# e.g. federation sending, appservice sending, etc.
|
2018-05-21 18:47:37 -06:00
|
|
|
event_processing_positions = Gauge("synapse_event_processing_positions", "", ["name"])
|
2018-04-11 04:07:51 -06:00
|
|
|
|
|
|
|
# Used to track the current max events stream position
|
2018-05-21 18:47:37 -06:00
|
|
|
event_persisted_position = Gauge("synapse_event_persisted_position", "")
|
2018-04-11 04:07:51 -06:00
|
|
|
|
2018-04-11 04:52:19 -06:00
|
|
|
# Used to track the received_ts of the last event processed by various
|
|
|
|
# components
|
2018-05-21 18:47:37 -06:00
|
|
|
event_processing_last_ts = Gauge("synapse_event_processing_last_ts", "", ["name"])
|
2018-04-11 04:52:19 -06:00
|
|
|
|
|
|
|
# Used to track the lag processing events. This is the time difference
|
|
|
|
# between the last processed event's received_ts and the time it was
|
|
|
|
# finished being processed.
|
2018-05-21 18:47:37 -06:00
|
|
|
event_processing_lag = Gauge("synapse_event_processing_lag", "", ["name"])
|
2015-08-13 04:38:59 -06:00
|
|
|
|
2020-06-30 09:58:06 -06:00
|
|
|
event_processing_lag_by_event = Histogram(
|
|
|
|
"synapse_event_processing_lag_by_event",
|
|
|
|
"Time between an event being persisted and it being queued up to be sent to the relevant remote servers",
|
|
|
|
["name"],
|
|
|
|
)
|
|
|
|
|
2019-09-09 08:14:58 -06:00
|
|
|
# Build info of the running server.
|
|
|
|
build_info = Gauge(
|
|
|
|
"synapse_build_info", "Build information", ["pythonversion", "version", "osversion"]
|
|
|
|
)
|
|
|
|
build_info.labels(
|
|
|
|
" ".join([platform.python_implementation(), platform.python_version()]),
|
|
|
|
get_version_string(synapse),
|
|
|
|
" ".join([platform.system(), platform.release()]),
|
|
|
|
).set(1)
|
|
|
|
|
2018-06-14 04:26:59 -06:00
|
|
|
last_ticked = time.time()
|
|
|
|
|
2020-11-13 05:03:51 -07:00
|
|
|
# 3PID send info
|
|
|
|
threepid_send_requests = Histogram(
|
|
|
|
"synapse_threepid_send_requests_with_tries",
|
|
|
|
documentation="Number of requests for a 3pid token by try count. Note if"
|
|
|
|
" there is a request with try count of 4, then there would have been one"
|
|
|
|
" each for 1, 2 and 3",
|
|
|
|
buckets=(1, 2, 3, 4, 5, 10),
|
|
|
|
labelnames=("type", "reason"),
|
|
|
|
)
|
|
|
|
|
2018-06-14 04:26:59 -06:00
|
|
|
|
2020-09-04 04:54:56 -06:00
|
|
|
class ReactorLastSeenMetric:
|
2018-06-14 04:26:59 -06:00
|
|
|
def collect(self):
|
|
|
|
cm = GaugeMetricFamily(
|
|
|
|
"python_twisted_reactor_last_seen",
|
|
|
|
"Seconds since the Twisted reactor was last seen",
|
|
|
|
)
|
|
|
|
cm.add_metric([], time.time() - last_ticked)
|
|
|
|
yield cm
|
|
|
|
|
|
|
|
|
|
|
|
REGISTRY.register(ReactorLastSeenMetric())
|
|
|
|
|
2018-05-22 16:32:57 -06:00
|
|
|
|
2021-03-03 13:47:38 -07:00
|
|
|
def runUntilCurrentTimer(reactor, func):
|
2015-08-13 04:38:59 -06:00
|
|
|
@functools.wraps(func)
|
|
|
|
def f(*args, **kwargs):
|
2015-08-14 08:42:52 -06:00
|
|
|
now = reactor.seconds()
|
|
|
|
num_pending = 0
|
|
|
|
|
|
|
|
# _newTimedCalls is one long list of *all* pending calls. Below loop
|
|
|
|
# is based off of impl of reactor.runUntilCurrent
|
2015-08-18 04:47:00 -06:00
|
|
|
for delayed_call in reactor._newTimedCalls:
|
|
|
|
if delayed_call.time > now:
|
2015-08-14 08:42:52 -06:00
|
|
|
break
|
|
|
|
|
2015-08-18 04:47:00 -06:00
|
|
|
if delayed_call.delayed_time > 0:
|
2015-08-14 08:42:52 -06:00
|
|
|
continue
|
|
|
|
|
|
|
|
num_pending += 1
|
|
|
|
|
|
|
|
num_pending += len(reactor.threadCallQueue)
|
2018-05-28 03:10:27 -06:00
|
|
|
start = time.time()
|
2015-08-13 04:38:59 -06:00
|
|
|
ret = func(*args, **kwargs)
|
2018-05-28 03:10:27 -06:00
|
|
|
end = time.time()
|
2018-01-19 16:51:04 -07:00
|
|
|
|
|
|
|
# record the amount of wallclock time spent running pending calls.
|
|
|
|
# This is a proxy for the actual amount of time between reactor polls,
|
|
|
|
# since about 25% of time is actually spent running things triggered by
|
|
|
|
# I/O events, but that is harder to capture without rewriting half the
|
|
|
|
# reactor.
|
2018-05-21 18:47:37 -06:00
|
|
|
tick_time.observe(end - start)
|
|
|
|
pending_calls_metric.observe(num_pending)
|
2016-05-09 03:13:25 -06:00
|
|
|
|
2018-06-14 04:26:59 -06:00
|
|
|
# Update the time we last ticked, for the metric to test whether
|
|
|
|
# Synapse's reactor has frozen
|
|
|
|
global last_ticked
|
|
|
|
last_ticked = end
|
|
|
|
|
2018-01-07 17:53:32 -07:00
|
|
|
if running_on_pypy:
|
|
|
|
return ret
|
|
|
|
|
2016-05-13 09:31:08 -06:00
|
|
|
# Check if we need to do a manual GC (since its been disabled), and do
|
|
|
|
# one if necessary.
|
2016-05-09 03:13:25 -06:00
|
|
|
threshold = gc.get_threshold()
|
|
|
|
counts = gc.get_count()
|
2016-06-07 06:40:22 -06:00
|
|
|
for i in (2, 1, 0):
|
2016-05-09 03:13:25 -06:00
|
|
|
if threshold[i] < counts[i]:
|
2019-06-28 05:45:33 -06:00
|
|
|
if i == 0:
|
|
|
|
logger.debug("Collecting gc %d", i)
|
|
|
|
else:
|
|
|
|
logger.info("Collecting gc %d", i)
|
2016-05-16 02:32:29 -06:00
|
|
|
|
2018-05-28 03:10:27 -06:00
|
|
|
start = time.time()
|
2016-06-07 06:40:22 -06:00
|
|
|
unreachable = gc.collect(i)
|
2018-05-28 03:10:27 -06:00
|
|
|
end = time.time()
|
2016-05-16 02:32:29 -06:00
|
|
|
|
2018-05-21 18:47:37 -06:00
|
|
|
gc_time.labels(i).observe(end - start)
|
|
|
|
gc_unreachable.labels(i).set(unreachable)
|
2016-05-09 03:13:25 -06:00
|
|
|
|
2015-08-13 04:38:59 -06:00
|
|
|
return ret
|
|
|
|
|
|
|
|
return f
|
|
|
|
|
|
|
|
|
2015-08-18 04:51:08 -06:00
|
|
|
try:
|
|
|
|
# Ensure the reactor has all the attributes we expect
|
2021-03-03 13:47:38 -07:00
|
|
|
reactor.seconds # type: ignore
|
|
|
|
reactor.runUntilCurrent # type: ignore
|
|
|
|
reactor._newTimedCalls # type: ignore
|
|
|
|
reactor.threadCallQueue # type: ignore
|
2015-08-18 04:51:08 -06:00
|
|
|
|
2015-08-13 04:38:59 -06:00
|
|
|
# runUntilCurrent is called when we have pending calls. It is called once
|
|
|
|
# per iteratation after fd polling.
|
2021-03-03 13:47:38 -07:00
|
|
|
reactor.runUntilCurrent = runUntilCurrentTimer(reactor, reactor.runUntilCurrent) # type: ignore
|
2016-05-13 09:31:08 -06:00
|
|
|
|
|
|
|
# We manually run the GC each reactor tick so that we can get some metrics
|
|
|
|
# about time spent doing GC,
|
2018-01-07 17:53:32 -07:00
|
|
|
if not running_on_pypy:
|
|
|
|
gc.disable()
|
2015-08-18 04:51:08 -06:00
|
|
|
except AttributeError:
|
|
|
|
pass
|
2019-07-18 07:57:15 -06:00
|
|
|
|
|
|
|
__all__ = [
|
|
|
|
"MetricsResource",
|
|
|
|
"generate_latest",
|
|
|
|
"start_http_server",
|
|
|
|
"LaterGauge",
|
|
|
|
"InFlightGauge",
|
|
|
|
"BucketCollector",
|
|
|
|
]
|