use farey function

This commit is contained in:
Erik Johnston 2018-06-01 14:54:19 +01:00
parent ca11acf388
commit 1205362f1d
2 changed files with 89 additions and 72 deletions

View File

@ -21,6 +21,7 @@ from gmpy2 import mpq as Fraction
from fractions import Fraction as FractionPy from fractions import Fraction as FractionPy
from synapse.storage._base import SQLBaseStore from synapse.storage._base import SQLBaseStore
from synapse.storage.engines import PostgresEngine
from synapse.util.katriel_bodlaender import OrderedListStore from synapse.util.katriel_bodlaender import OrderedListStore
import synapse.metrics import synapse.metrics
@ -91,12 +92,13 @@ class ChunkDBOrderedListStore(OrderedListStore):
this. this.
""" """
def __init__(self, def __init__(self,
txn, room_id, clock, txn, room_id, clock, database_engine,
rebalance_max_denominator=100, rebalance_max_denominator=100,
max_denominator=100000): max_denominator=100000):
self.txn = txn self.txn = txn
self.room_id = room_id self.room_id = room_id
self.clock = clock self.clock = clock
self.database_engine = database_engine
self.rebalance_md = rebalance_max_denominator self.rebalance_md = rebalance_max_denominator
self.max_denominator = max_denominator self.max_denominator = max_denominator
@ -390,69 +392,43 @@ class ChunkDBOrderedListStore(OrderedListStore):
logger.info("Rebalancing room %s, chunk %s", self.room_id, node_id) logger.info("Rebalancing room %s, chunk %s", self.room_id, node_id)
old_order = self._get_order(node_id) old_order = self._get_order(node_id)
new_order = FractionPy(
int(old_order.numerator),
int(old_order.denominator),
).limit_denominator(
self.rebalance_md,
)
new_order = Fraction(new_order.numerator, new_order.denominator)
if new_order < old_order:
new_order += Fraction(1, new_order.denominator)
count_nodes = [node_id] a, b, c, d = find_farey_terms(old_order, self.rebalance_md)
next_id = node_id n = max(b, d)
while True:
next_id = self.get_next(next_id)
if not next_id: with_sql = """
max_order = None WITH RECURSIVE chunks (chunk_id, next, n, a, b, c, d) AS (
break SELECT chunk_id, next_chunk_id, ?, ?, ?, ?, ?
FROM chunk_linearized WHERE chunk_id = ?
count_nodes.append(next_id) UNION ALL
SELECT n.chunk_id, n.next_chunk_id, n, c, d, ((n + b) / d) * c - a, ((n + b) / d) * d - b
max_order = self._get_order(next_id) FROM chunks AS c
INNER JOIN chunk_linearized AS l ON l.chunk_id = c.chunk_id
if len(count_nodes) < self.rebalance_md * (max_order - new_order): INNER JOIN chunk_linearized AS n ON n.chunk_id = l.next_chunk_id
break WHERE c * 1.0 / d > n.numerator * 1.0 / n.denominator
if len(count_nodes) == 1:
orders = [new_order]
if max_order:
orders = stern_brocot_range(len(count_nodes), new_order, max_order)
orders.sort(reverse=True)
else:
orders = [
Fraction(int(math.ceil(new_order)) + i, 1)
for i in xrange(0, len(count_nodes))
]
orders.reverse()
assert len(count_nodes) == len(orders)
next_id = node_id
prev_order = old_order
while orders:
order = orders.pop()
if max_order:
assert old_order <= new_order <= max_order
else:
assert old_order <= new_order
assert prev_order < order
SQLBaseStore._simple_update_txn(
self.txn,
table="chunk_linearized",
keyvalues={"chunk_id": next_id},
updatevalues={
"numerator": int(order.numerator),
"denominator": int(order.denominator),
},
) )
"""
next_id = self.get_next(next_id) if isinstance(self.database_engine, PostgresEngine):
sql = with_sql + """
UPDATE chunk_linearized AS l
SET numerator = a, denominator = b
FROM chunks AS c
WHERE c.chunk_id = l.chunk_id
"""
else:
sql = with_sql + """
UPDATE chunk_linearized
SET (numerator, denominator) = (
SELECT a, b FROM chunks
WHERE chunks.chunk_id = chunk_linearized.chunk_id
)
WHERE chunk_id in (SELECT chunk_id FROM chunks)
"""
self.txn.execute(sql, (
n, a, b, c, d, node_id
))
rebalance_counter.inc() rebalance_counter.inc()
@ -512,7 +488,7 @@ def stern_brocot_single(min_frac, max_frac):
def stern_brocot_range_depth(min_frac, max_denom): def stern_brocot_range_depth(min_frac, max_denom):
assert 0 < min_frac assert 0 < min_frac
states = stern_brocot_single(min_frac) states = stern_brocot_singless(min_frac)
while len(states): while len(states):
a, b, c, d = states.pop() a, b, c, d = states.pop()
@ -534,22 +510,51 @@ def stern_brocot_range_depth(min_frac, max_denom):
def stern_brocot_single(min_frac): def stern_brocot_state(min_frac, target_d):
assert 0 <= min_frac assert 0 <= min_frac
states = [] states = []
a, b, c, d = 0, 1, 1, 0 a, b, c, d = 0, 1, 1, 0
states.append((a, b, c, d))
while True: while True:
f = Fraction(a + c, b + d) f = Fraction(a + c, b + d)
if b + d >= target_d:
return a + c, b + d, c, d
if f < min_frac: if f < min_frac:
a, b, c, d = a + c, b + d, c, d a, b, c, d = a + c, b + d, c, d
elif f == min_frac: elif f == min_frac:
return states return a + c, b + d, c, d
else: else:
a, b, c, d = a, b, a + c, b + d a, b, c, d = a, b, a + c, b + d
states.append((a, b, c, d))
def find_farey_terms(min_frac, max_denom):
states = deque([(0, 1, 1, 0)])
while True:
a, b, c, d = states.popleft()
left = a / float(b)
mid = (a + c) / float(b + d)
right = c / float(d) if d > 0 else None
if min_frac < left:
if b >= max_denom or d >= max_denom:
return a, b, c, d
if b + d >= max_denom:
return a + c, b + d, c, d
states.append((a, b, a + c, b + d))
elif min_frac < mid:
if b + d >= max_denom:
return a + c, b + d, c, d
states.append((a, b, a + c, b + d))
states.append((a + c, b + d, c, d))
elif right is None or min_frac < right:
states.append((a + c, b + d, c, d))
else:
states.append((a + c, b + d, c, d))

View File

@ -44,7 +44,9 @@ class ChunkLinearizerStoreTestCase(tests.unittest.TestCase):
def test_txn(txn): def test_txn(txn):
table = ChunkDBOrderedListStore( table = ChunkDBOrderedListStore(
txn, room_id, self.clock, 5, 100, txn, room_id, self.clock,
self.store.database_engine,
5, 100,
) )
table.add_node("A") table.add_node("A")
@ -71,7 +73,9 @@ class ChunkLinearizerStoreTestCase(tests.unittest.TestCase):
def test_txn(txn): def test_txn(txn):
table = ChunkDBOrderedListStore( table = ChunkDBOrderedListStore(
txn, room_id, self.clock, 5, 100, txn, room_id, self.clock,
self.store.database_engine,
5, 100,
) )
nodes = [(i, "node_%d" % (i,)) for i in xrange(1, 1000)] nodes = [(i, "node_%d" % (i,)) for i in xrange(1, 1000)]
@ -116,7 +120,9 @@ class ChunkLinearizerStoreTestCase(tests.unittest.TestCase):
def test_txn(txn): def test_txn(txn):
table = ChunkDBOrderedListStore( table = ChunkDBOrderedListStore(
txn, room_id, self.clock, 5, 1000, txn, room_id, self.clock,
self.store.database_engine,
5, 1000,
) )
table.add_node("a") table.add_node("a")
@ -152,7 +158,9 @@ class ChunkLinearizerStoreTestCase(tests.unittest.TestCase):
def test_txn(txn): def test_txn(txn):
table = ChunkDBOrderedListStore( table = ChunkDBOrderedListStore(
txn, room_id, self.clock, 5, 100, txn, room_id, self.clock,
self.store.database_engine,
5, 100,
) )
table.add_node("a") table.add_node("a")
@ -193,7 +201,9 @@ class ChunkLinearizerStoreTestCase(tests.unittest.TestCase):
def test_txn(txn): def test_txn(txn):
table = ChunkDBOrderedListStore( table = ChunkDBOrderedListStore(
txn, room_id, self.clock, 5, 100, txn, room_id, self.clock,
self.store.database_engine,
5, 100,
) )
table.add_node("A") table.add_node("A")
@ -216,7 +226,9 @@ class ChunkLinearizerStoreTestCase(tests.unittest.TestCase):
def test_txn(txn): def test_txn(txn):
table = ChunkDBOrderedListStore( table = ChunkDBOrderedListStore(
txn, room_id, self.clock, 5, 100, txn, room_id, self.clock,
self.store.database_engine,
5, 100,
) )
table.add_node("A") table.add_node("A")