Fix unread counts on large servers (#13140)
This commit is contained in:
parent
a3a05c812d
commit
dbce28b2f1
|
@ -0,0 +1 @@
|
||||||
|
Fix unread counts for users on large servers. Introduced in v1.62.0rc1.
|
|
@ -854,18 +854,20 @@ class EventPushActionsWorkerStore(ReceiptsWorkerStore, EventsWorkerStore, SQLBas
|
||||||
|
|
||||||
limit = 100
|
limit = 100
|
||||||
|
|
||||||
min_stream_id = self.db_pool.simple_select_one_onecol_txn(
|
min_receipts_stream_id = self.db_pool.simple_select_one_onecol_txn(
|
||||||
txn,
|
txn,
|
||||||
table="event_push_summary_last_receipt_stream_id",
|
table="event_push_summary_last_receipt_stream_id",
|
||||||
keyvalues={},
|
keyvalues={},
|
||||||
retcol="stream_id",
|
retcol="stream_id",
|
||||||
)
|
)
|
||||||
|
|
||||||
|
max_receipts_stream_id = self._receipts_id_gen.get_current_token()
|
||||||
|
|
||||||
sql = """
|
sql = """
|
||||||
SELECT r.stream_id, r.room_id, r.user_id, e.stream_ordering
|
SELECT r.stream_id, r.room_id, r.user_id, e.stream_ordering
|
||||||
FROM receipts_linearized AS r
|
FROM receipts_linearized AS r
|
||||||
INNER JOIN events AS e USING (event_id)
|
INNER JOIN events AS e USING (event_id)
|
||||||
WHERE r.stream_id > ? AND user_id LIKE ?
|
WHERE ? < r.stream_id AND r.stream_id <= ? AND user_id LIKE ?
|
||||||
ORDER BY r.stream_id ASC
|
ORDER BY r.stream_id ASC
|
||||||
LIMIT ?
|
LIMIT ?
|
||||||
"""
|
"""
|
||||||
|
@ -877,13 +879,21 @@ class EventPushActionsWorkerStore(ReceiptsWorkerStore, EventsWorkerStore, SQLBas
|
||||||
txn.execute(
|
txn.execute(
|
||||||
sql,
|
sql,
|
||||||
(
|
(
|
||||||
min_stream_id,
|
min_receipts_stream_id,
|
||||||
|
max_receipts_stream_id,
|
||||||
user_filter,
|
user_filter,
|
||||||
limit,
|
limit,
|
||||||
),
|
),
|
||||||
)
|
)
|
||||||
rows = txn.fetchall()
|
rows = txn.fetchall()
|
||||||
|
|
||||||
|
old_rotate_stream_ordering = self.db_pool.simple_select_one_onecol_txn(
|
||||||
|
txn,
|
||||||
|
table="event_push_summary_stream_ordering",
|
||||||
|
keyvalues={},
|
||||||
|
retcol="stream_ordering",
|
||||||
|
)
|
||||||
|
|
||||||
# For each new read receipt we delete push actions from before it and
|
# For each new read receipt we delete push actions from before it and
|
||||||
# recalculate the summary.
|
# recalculate the summary.
|
||||||
for _, room_id, user_id, stream_ordering in rows:
|
for _, room_id, user_id, stream_ordering in rows:
|
||||||
|
@ -902,13 +912,6 @@ class EventPushActionsWorkerStore(ReceiptsWorkerStore, EventsWorkerStore, SQLBas
|
||||||
(room_id, user_id, stream_ordering),
|
(room_id, user_id, stream_ordering),
|
||||||
)
|
)
|
||||||
|
|
||||||
old_rotate_stream_ordering = self.db_pool.simple_select_one_onecol_txn(
|
|
||||||
txn,
|
|
||||||
table="event_push_summary_stream_ordering",
|
|
||||||
keyvalues={},
|
|
||||||
retcol="stream_ordering",
|
|
||||||
)
|
|
||||||
|
|
||||||
notif_count, unread_count = self._get_notif_unread_count_for_user_room(
|
notif_count, unread_count = self._get_notif_unread_count_for_user_room(
|
||||||
txn, room_id, user_id, stream_ordering, old_rotate_stream_ordering
|
txn, room_id, user_id, stream_ordering, old_rotate_stream_ordering
|
||||||
)
|
)
|
||||||
|
@ -927,18 +930,19 @@ class EventPushActionsWorkerStore(ReceiptsWorkerStore, EventsWorkerStore, SQLBas
|
||||||
|
|
||||||
# We always update `event_push_summary_last_receipt_stream_id` to
|
# We always update `event_push_summary_last_receipt_stream_id` to
|
||||||
# ensure that we don't rescan the same receipts for remote users.
|
# ensure that we don't rescan the same receipts for remote users.
|
||||||
#
|
|
||||||
# This requires repeatable read to be safe, as we need the
|
upper_limit = max_receipts_stream_id
|
||||||
# `MAX(stream_id)` to not include any new rows that have been committed
|
if len(rows) >= limit:
|
||||||
# since the start of the transaction (since those rows won't have been
|
# If we pulled out a limited number of rows we only update the
|
||||||
# returned by the query above). Alternatively we could query the max
|
# position to the last receipt we processed, so we continue
|
||||||
# stream ID at the start of the transaction and bound everything by
|
# processing the rest next iteration.
|
||||||
# that.
|
upper_limit = rows[-1][0]
|
||||||
txn.execute(
|
|
||||||
"""
|
self.db_pool.simple_update_txn(
|
||||||
UPDATE event_push_summary_last_receipt_stream_id
|
txn,
|
||||||
SET stream_id = (SELECT COALESCE(MAX(stream_id), 0) FROM receipts_linearized)
|
table="event_push_summary_last_receipt_stream_id",
|
||||||
"""
|
keyvalues={},
|
||||||
|
updatevalues={"stream_id": upper_limit},
|
||||||
)
|
)
|
||||||
|
|
||||||
return len(rows) < limit
|
return len(rows) < limit
|
||||||
|
|
|
@ -134,15 +134,12 @@ class EventPushActionsStoreTestCase(HomeserverTestCase):
|
||||||
last_read_stream_ordering[0] = stream
|
last_read_stream_ordering[0] = stream
|
||||||
|
|
||||||
self.get_success(
|
self.get_success(
|
||||||
self.store.db_pool.runInteraction(
|
self.store.insert_receipt(
|
||||||
"",
|
|
||||||
self.store._insert_linearized_receipt_txn,
|
|
||||||
room_id,
|
room_id,
|
||||||
"m.read",
|
"m.read",
|
||||||
user_id,
|
user_id=user_id,
|
||||||
f"$test{stream}:example.com",
|
event_ids=[f"$test{stream}:example.com"],
|
||||||
{},
|
data={},
|
||||||
stream,
|
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -166,6 +163,7 @@ class EventPushActionsStoreTestCase(HomeserverTestCase):
|
||||||
|
|
||||||
_inject_actions(6, PlAIN_NOTIF)
|
_inject_actions(6, PlAIN_NOTIF)
|
||||||
_rotate(7)
|
_rotate(7)
|
||||||
|
_assert_counts(1, 0)
|
||||||
|
|
||||||
self.get_success(
|
self.get_success(
|
||||||
self.store.db_pool.simple_delete(
|
self.store.db_pool.simple_delete(
|
||||||
|
|
Loading…
Reference in New Issue