Merge pull request #1129 from matrix-org/erikj/fix_pubroom_pag
Fix and clean up publicRooms pagination
This commit is contained in:
commit
eb24aecf8c
|
@ -115,32 +115,109 @@ class RoomListHandler(BaseHandler):
|
||||||
sorted_entries = sorted(rooms_to_order_value.items(), key=lambda e: e[1])
|
sorted_entries = sorted(rooms_to_order_value.items(), key=lambda e: e[1])
|
||||||
sorted_rooms = [room_id for room_id, _ in sorted_entries]
|
sorted_rooms = [room_id for room_id, _ in sorted_entries]
|
||||||
|
|
||||||
|
# `sorted_rooms` should now be a list of all public room ids that is
|
||||||
|
# stable across pagination. Therefore, we can use indices into this
|
||||||
|
# list as our pagination tokens.
|
||||||
|
|
||||||
|
# Filter out rooms that we don't want to return
|
||||||
|
rooms_to_scan = [
|
||||||
|
r for r in sorted_rooms
|
||||||
|
if r not in newly_unpublished and rooms_to_num_joined[room_id] > 0
|
||||||
|
]
|
||||||
|
|
||||||
if since_token:
|
if since_token:
|
||||||
|
# Filter out rooms we've already returned previously
|
||||||
|
# `since_token.current_limit` is the index of the last room we
|
||||||
|
# sent down, so we exclude it and everything before/after it.
|
||||||
if since_token.direction_is_forward:
|
if since_token.direction_is_forward:
|
||||||
sorted_rooms = sorted_rooms[since_token.current_limit:]
|
rooms_to_scan = rooms_to_scan[since_token.current_limit + 1:]
|
||||||
else:
|
else:
|
||||||
sorted_rooms = sorted_rooms[:since_token.current_limit]
|
rooms_to_scan = rooms_to_scan[:since_token.current_limit]
|
||||||
sorted_rooms.reverse()
|
rooms_to_scan.reverse()
|
||||||
|
|
||||||
rooms_to_scan = sorted_rooms
|
# If there's no search filter just limit the range since we'll
|
||||||
|
# return the vast majority of things.
|
||||||
if limit and not search_filter:
|
if limit and not search_filter:
|
||||||
rooms_to_scan = sorted_rooms[:limit + 1]
|
rooms_to_scan = rooms_to_scan[:limit + 1]
|
||||||
|
|
||||||
|
# Actually generate the entries. _generate_room_entry will append to
|
||||||
|
# chunk but will stop if len(chunk) > limit
|
||||||
chunk = []
|
chunk = []
|
||||||
|
yield concurrently_execute(
|
||||||
|
lambda r: self._generate_room_entry(
|
||||||
|
r, rooms_to_num_joined[r],
|
||||||
|
chunk, limit, search_filter
|
||||||
|
),
|
||||||
|
rooms_to_scan, 10
|
||||||
|
)
|
||||||
|
|
||||||
|
chunk.sort(key=lambda e: (-e["num_joined_members"], e["room_id"]))
|
||||||
|
|
||||||
|
# Work out the new limit of the batch for pagination, or None if we
|
||||||
|
# know there are no more results that would be returned.
|
||||||
|
# i.e., [since_token.current_limit..new_limit] is the batch of rooms
|
||||||
|
# we've returned (or the reverse if we paginated backwards)
|
||||||
|
# We tried to pull out limit + 1 rooms above, so if we have <= limit
|
||||||
|
# then we know there are no more results to return
|
||||||
|
new_limit = None
|
||||||
|
if chunk and (not limit or len(chunk) > limit):
|
||||||
|
|
||||||
|
if not since_token or since_token.direction_is_forward:
|
||||||
|
if limit:
|
||||||
|
chunk = chunk[:limit]
|
||||||
|
last_room_id = chunk[-1]["room_id"]
|
||||||
|
else:
|
||||||
|
if limit:
|
||||||
|
chunk = chunk[-limit:]
|
||||||
|
last_room_id = chunk[0]["room_id"]
|
||||||
|
|
||||||
|
new_limit = sorted_rooms.index(last_room_id)
|
||||||
|
|
||||||
|
results = {
|
||||||
|
"chunk": chunk,
|
||||||
|
}
|
||||||
|
|
||||||
|
if since_token:
|
||||||
|
results["new_rooms"] = bool(newly_visible)
|
||||||
|
|
||||||
|
if not since_token or since_token.direction_is_forward:
|
||||||
|
if new_limit is not None:
|
||||||
|
results["next_batch"] = RoomListNextBatch(
|
||||||
|
stream_ordering=stream_token,
|
||||||
|
public_room_stream_id=public_room_stream_id,
|
||||||
|
current_limit=new_limit,
|
||||||
|
direction_is_forward=True,
|
||||||
|
).to_token()
|
||||||
|
|
||||||
|
if since_token:
|
||||||
|
results["prev_batch"] = since_token.copy_and_replace(
|
||||||
|
direction_is_forward=False,
|
||||||
|
current_limit=since_token.current_limit + 1,
|
||||||
|
).to_token()
|
||||||
|
else:
|
||||||
|
if new_limit is not None:
|
||||||
|
results["prev_batch"] = RoomListNextBatch(
|
||||||
|
stream_ordering=stream_token,
|
||||||
|
public_room_stream_id=public_room_stream_id,
|
||||||
|
current_limit=new_limit,
|
||||||
|
direction_is_forward=False,
|
||||||
|
).to_token()
|
||||||
|
|
||||||
|
if since_token:
|
||||||
|
results["next_batch"] = since_token.copy_and_replace(
|
||||||
|
direction_is_forward=True,
|
||||||
|
current_limit=since_token.current_limit - 1,
|
||||||
|
).to_token()
|
||||||
|
|
||||||
|
defer.returnValue(results)
|
||||||
|
|
||||||
@defer.inlineCallbacks
|
@defer.inlineCallbacks
|
||||||
def handle_room(room_id):
|
def _generate_room_entry(self, room_id, num_joined_users, chunk, limit,
|
||||||
|
search_filter):
|
||||||
if limit and len(chunk) > limit + 1:
|
if limit and len(chunk) > limit + 1:
|
||||||
# We've already got enough, so lets just drop it.
|
# We've already got enough, so lets just drop it.
|
||||||
return
|
return
|
||||||
|
|
||||||
num_joined_users = rooms_to_num_joined[room_id]
|
|
||||||
if num_joined_users == 0:
|
|
||||||
return
|
|
||||||
|
|
||||||
if room_id in newly_unpublished:
|
|
||||||
return
|
|
||||||
|
|
||||||
result = {
|
result = {
|
||||||
"room_id": room_id,
|
"room_id": room_id,
|
||||||
"num_joined_members": num_joined_users,
|
"num_joined_members": num_joined_users,
|
||||||
|
@ -216,70 +293,6 @@ class RoomListHandler(BaseHandler):
|
||||||
if _matches_room_entry(result, search_filter):
|
if _matches_room_entry(result, search_filter):
|
||||||
chunk.append(result)
|
chunk.append(result)
|
||||||
|
|
||||||
yield concurrently_execute(handle_room, rooms_to_scan, 10)
|
|
||||||
|
|
||||||
chunk.sort(key=lambda e: (-e["num_joined_members"], e["room_id"]))
|
|
||||||
|
|
||||||
# Work out the new limit of the batch for pagination, or None if we
|
|
||||||
# know there are no more results that would be returned.
|
|
||||||
new_limit = None
|
|
||||||
if chunk and (not limit or len(chunk) > limit):
|
|
||||||
if limit:
|
|
||||||
chunk = chunk[:limit]
|
|
||||||
|
|
||||||
addition = 1
|
|
||||||
if since_token:
|
|
||||||
addition += since_token.current_limit
|
|
||||||
|
|
||||||
if not since_token or since_token.direction_is_forward:
|
|
||||||
last_room_id = chunk[-1]["room_id"]
|
|
||||||
else:
|
|
||||||
last_room_id = chunk[0]["room_id"]
|
|
||||||
addition *= -1
|
|
||||||
|
|
||||||
try:
|
|
||||||
new_limit = sorted_rooms.index(last_room_id) + addition
|
|
||||||
if new_limit >= len(sorted_rooms):
|
|
||||||
new_limit = None
|
|
||||||
except ValueError:
|
|
||||||
pass
|
|
||||||
|
|
||||||
results = {
|
|
||||||
"chunk": chunk,
|
|
||||||
}
|
|
||||||
|
|
||||||
if since_token:
|
|
||||||
results["new_rooms"] = bool(newly_visible)
|
|
||||||
|
|
||||||
if not since_token or since_token.direction_is_forward:
|
|
||||||
if new_limit:
|
|
||||||
results["next_batch"] = RoomListNextBatch(
|
|
||||||
stream_ordering=stream_token,
|
|
||||||
public_room_stream_id=public_room_stream_id,
|
|
||||||
current_limit=new_limit,
|
|
||||||
direction_is_forward=True,
|
|
||||||
).to_token()
|
|
||||||
|
|
||||||
if since_token:
|
|
||||||
results["prev_batch"] = since_token.copy_and_replace(
|
|
||||||
direction_is_forward=False,
|
|
||||||
).to_token()
|
|
||||||
else:
|
|
||||||
if new_limit:
|
|
||||||
results["prev_batch"] = RoomListNextBatch(
|
|
||||||
stream_ordering=stream_token,
|
|
||||||
public_room_stream_id=public_room_stream_id,
|
|
||||||
current_limit=new_limit,
|
|
||||||
direction_is_forward=False,
|
|
||||||
).to_token()
|
|
||||||
|
|
||||||
if since_token:
|
|
||||||
results["next_batch"] = since_token.copy_and_replace(
|
|
||||||
direction_is_forward=True,
|
|
||||||
).to_token()
|
|
||||||
|
|
||||||
defer.returnValue(results)
|
|
||||||
|
|
||||||
@defer.inlineCallbacks
|
@defer.inlineCallbacks
|
||||||
def get_remote_public_room_list(self, server_name, limit=None, since_token=None,
|
def get_remote_public_room_list(self, server_name, limit=None, since_token=None,
|
||||||
search_filter=None):
|
search_filter=None):
|
||||||
|
|
Loading…
Reference in New Issue