Refactor to rewrite the SQL instead

This commit is contained in:
Erik Johnston 2019-01-22 16:31:05 +00:00
parent 1c9704f8ab
commit 7f503f83b9
1 changed files with 16 additions and 13 deletions

View File

@ -140,16 +140,8 @@ class ClientIpStore(background_updates.BackgroundUpdateStore):
"user_ips_dups_get_last_seen", get_last_seen
)
if end_last_seen is None:
# If we get a None then we're reaching the end and just need to
# delete the last batch.
last = True
# We fake not having an upper bound by using a future date, by
# just multiplying the current time by two....
last_seen = int(self.clock.time_msec()) * 2
else:
last = False
# If it returns None, then we're processing the last batch
last = end_last_seen is None
def remove(txn):
# This works by looking at all entries in the given time span, and
@ -160,6 +152,16 @@ class ClientIpStore(background_updates.BackgroundUpdateStore):
# all other duplicates.
# It is efficient due to the existence of (user_id, access_token,
# ip) and (last_seen) indices.
# Define the search space, which requires handling the last batch in
# a different way
if last:
clause = "? <= last_seen"
args = (begin_last_seen,)
else:
clause = "? <= last_seen AND last_seen < ?"
args = (begin_last_seen, end_last_seen)
txn.execute(
"""
SELECT user_id, access_token, ip,
@ -167,13 +169,14 @@ class ClientIpStore(background_updates.BackgroundUpdateStore):
FROM (
SELECT user_id, access_token, ip
FROM user_ips
WHERE ? <= last_seen AND last_seen < ?
WHERE {}
ORDER BY last_seen
) c
INNER JOIN user_ips USING (user_id, access_token, ip)
GROUP BY user_id, access_token, ip
HAVING count(*) > 1""",
(begin_last_seen, end_last_seen)
HAVING count(*) > 1
""".format(clause),
args
)
res = txn.fetchall()