You've already forked pg_probackup
mirror of
https://github.com/postgrespro/pg_probackup.git
synced 2025-07-12 06:40:11 +02:00
tests: for pgpro-2573
This commit is contained in:
@ -837,7 +837,7 @@ class ProbackupTest(object):
|
|||||||
def delete_expired(
|
def delete_expired(
|
||||||
self, backup_dir, instance, options=[], old_binary=False):
|
self, backup_dir, instance, options=[], old_binary=False):
|
||||||
cmd_list = [
|
cmd_list = [
|
||||||
'delete', '--expired', '--wal',
|
'delete',
|
||||||
'-B', backup_dir,
|
'-B', backup_dir,
|
||||||
'--instance={0}'.format(instance)
|
'--instance={0}'.format(instance)
|
||||||
]
|
]
|
||||||
|
@ -40,7 +40,8 @@ class RetentionTest(ProbackupTest, unittest.TestCase):
|
|||||||
self.assertEqual(len(self.show_pb(backup_dir, 'node')), 4)
|
self.assertEqual(len(self.show_pb(backup_dir, 'node')), 4)
|
||||||
|
|
||||||
# Purge backups
|
# Purge backups
|
||||||
log = self.delete_expired(backup_dir, 'node')
|
log = self.delete_expired(
|
||||||
|
backup_dir, 'node', options=['--expired', '--wal'])
|
||||||
self.assertEqual(len(self.show_pb(backup_dir, 'node')), 2)
|
self.assertEqual(len(self.show_pb(backup_dir, 'node')), 2)
|
||||||
|
|
||||||
# Check that WAL segments were deleted
|
# Check that WAL segments were deleted
|
||||||
@ -115,7 +116,7 @@ class RetentionTest(ProbackupTest, unittest.TestCase):
|
|||||||
self.assertEqual(len(self.show_pb(backup_dir, 'node')), 4)
|
self.assertEqual(len(self.show_pb(backup_dir, 'node')), 4)
|
||||||
|
|
||||||
# Purge backups
|
# Purge backups
|
||||||
self.delete_expired(backup_dir, 'node')
|
self.delete_expired(backup_dir, 'node', options=['--expired'])
|
||||||
self.assertEqual(len(self.show_pb(backup_dir, 'node')), 2)
|
self.assertEqual(len(self.show_pb(backup_dir, 'node')), 2)
|
||||||
|
|
||||||
# Clean after yourself
|
# Clean after yourself
|
||||||
@ -158,7 +159,7 @@ class RetentionTest(ProbackupTest, unittest.TestCase):
|
|||||||
|
|
||||||
# Purge backups
|
# Purge backups
|
||||||
self.delete_expired(
|
self.delete_expired(
|
||||||
backup_dir, 'node', options=['--retention-window=1'])
|
backup_dir, 'node', options=['--retention-window=1', '--expired'])
|
||||||
|
|
||||||
self.assertEqual(len(self.show_pb(backup_dir, 'node')), 0)
|
self.assertEqual(len(self.show_pb(backup_dir, 'node')), 0)
|
||||||
|
|
||||||
@ -207,7 +208,8 @@ class RetentionTest(ProbackupTest, unittest.TestCase):
|
|||||||
|
|
||||||
# Purge backups
|
# Purge backups
|
||||||
self.delete_expired(
|
self.delete_expired(
|
||||||
backup_dir, 'node', options=['--retention-window=1'])
|
backup_dir, 'node',
|
||||||
|
options=['--retention-window=1', '--expired', '--wal'])
|
||||||
|
|
||||||
self.assertEqual(len(self.show_pb(backup_dir, 'node')), 0)
|
self.assertEqual(len(self.show_pb(backup_dir, 'node')), 0)
|
||||||
|
|
||||||
@ -221,7 +223,8 @@ class RetentionTest(ProbackupTest, unittest.TestCase):
|
|||||||
self.assertTrue(n_wals > 0)
|
self.assertTrue(n_wals > 0)
|
||||||
|
|
||||||
self.delete_expired(
|
self.delete_expired(
|
||||||
backup_dir, 'node', options=['--retention-window=1'])
|
backup_dir, 'node',
|
||||||
|
options=['--retention-window=1', '--expired', '--wal'])
|
||||||
|
|
||||||
# count again
|
# count again
|
||||||
n_wals = len(os.listdir(wals_dir))
|
n_wals = len(os.listdir(wals_dir))
|
||||||
@ -231,7 +234,7 @@ class RetentionTest(ProbackupTest, unittest.TestCase):
|
|||||||
self.del_test_dir(module_name, fname)
|
self.del_test_dir(module_name, fname)
|
||||||
|
|
||||||
# @unittest.skip("skip")
|
# @unittest.skip("skip")
|
||||||
def test_retention_interleaved_incremental_chains(self):
|
def test_window_expire_interleaved_incremental_chains(self):
|
||||||
"""complicated case of interleaved backup chains"""
|
"""complicated case of interleaved backup chains"""
|
||||||
fname = self.id().split('.')[3]
|
fname = self.id().split('.')[3]
|
||||||
node = self.make_simple_node(
|
node = self.make_simple_node(
|
||||||
@ -335,7 +338,8 @@ class RetentionTest(ProbackupTest, unittest.TestCase):
|
|||||||
|
|
||||||
|
|
||||||
self.delete_expired(
|
self.delete_expired(
|
||||||
backup_dir, 'node', options=['--retention-window=1'])
|
backup_dir, 'node',
|
||||||
|
options=['--retention-window=1', '--expired'])
|
||||||
|
|
||||||
self.assertEqual(len(self.show_pb(backup_dir, 'node')), 6)
|
self.assertEqual(len(self.show_pb(backup_dir, 'node')), 6)
|
||||||
|
|
||||||
@ -344,3 +348,900 @@ class RetentionTest(ProbackupTest, unittest.TestCase):
|
|||||||
|
|
||||||
# Clean after yourself
|
# Clean after yourself
|
||||||
self.del_test_dir(module_name, fname)
|
self.del_test_dir(module_name, fname)
|
||||||
|
|
||||||
|
# @unittest.skip("skip")
|
||||||
|
def test_redundancy_expire_interleaved_incremental_chains(self):
|
||||||
|
"""complicated case of interleaved backup chains"""
|
||||||
|
fname = self.id().split('.')[3]
|
||||||
|
node = self.make_simple_node(
|
||||||
|
base_dir=os.path.join(module_name, fname, 'node'),
|
||||||
|
initdb_params=['--data-checksums'])
|
||||||
|
|
||||||
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
||||||
|
self.init_pb(backup_dir)
|
||||||
|
self.add_instance(backup_dir, 'node', node)
|
||||||
|
self.set_archiving(backup_dir, 'node', node)
|
||||||
|
node.slow_start()
|
||||||
|
|
||||||
|
|
||||||
|
# Take FULL BACKUPs
|
||||||
|
backup_id_a = self.backup_node(backup_dir, 'node', node)
|
||||||
|
backup_id_b = self.backup_node(backup_dir, 'node', node)
|
||||||
|
|
||||||
|
# Change FULL B backup status to ERROR
|
||||||
|
self.change_backup_status(backup_dir, 'node', backup_id_b, 'ERROR')
|
||||||
|
|
||||||
|
# FULLb ERROR
|
||||||
|
# FULLa OK
|
||||||
|
# Take PAGEa1 backup
|
||||||
|
page_id_a1 = self.backup_node(
|
||||||
|
backup_dir, 'node', node, backup_type='page')
|
||||||
|
|
||||||
|
# PAGEa1 OK
|
||||||
|
# FULLb ERROR
|
||||||
|
# FULLa OK
|
||||||
|
# Change FULL B backup status to OK
|
||||||
|
self.change_backup_status(backup_dir, 'node', backup_id_b, 'OK')
|
||||||
|
|
||||||
|
# Change PAGEa1 backup status to ERROR
|
||||||
|
self.change_backup_status(backup_dir, 'node', page_id_a1, 'ERROR')
|
||||||
|
|
||||||
|
# PAGEa1 ERROR
|
||||||
|
# FULLb OK
|
||||||
|
# FULLa OK
|
||||||
|
page_id_b1 = self.backup_node(
|
||||||
|
backup_dir, 'node', node, backup_type='page')
|
||||||
|
|
||||||
|
# PAGEb1 OK
|
||||||
|
# PAGEa1 ERROR
|
||||||
|
# FULLb OK
|
||||||
|
# FULLa OK
|
||||||
|
# Now we start to play with first generation of PAGE backups
|
||||||
|
# Change PAGEb1 status to ERROR
|
||||||
|
self.change_backup_status(backup_dir, 'node', page_id_b1, 'ERROR')
|
||||||
|
|
||||||
|
# Change PAGEa1 status to OK
|
||||||
|
self.change_backup_status(backup_dir, 'node', page_id_a1, 'OK')
|
||||||
|
|
||||||
|
# PAGEb1 ERROR
|
||||||
|
# PAGEa1 OK
|
||||||
|
# FULLb OK
|
||||||
|
# FULLa OK
|
||||||
|
page_id_a2 = self.backup_node(
|
||||||
|
backup_dir, 'node', node, backup_type='page')
|
||||||
|
|
||||||
|
# PAGEa2 OK
|
||||||
|
# PAGEb1 ERROR
|
||||||
|
# PAGEa1 OK
|
||||||
|
# FULLb OK
|
||||||
|
# FULLa OK
|
||||||
|
# Change PAGEa2 status to ERROR
|
||||||
|
self.change_backup_status(backup_dir, 'node', page_id_a2, 'ERROR')
|
||||||
|
|
||||||
|
# Change PAGEb1 status to OK
|
||||||
|
self.change_backup_status(backup_dir, 'node', page_id_b1, 'OK')
|
||||||
|
|
||||||
|
# PAGEa2 ERROR
|
||||||
|
# PAGEb1 OK
|
||||||
|
# PAGEa1 OK
|
||||||
|
# FULLb OK
|
||||||
|
# FULLa OK
|
||||||
|
page_id_b2 = self.backup_node(
|
||||||
|
backup_dir, 'node', node, backup_type='page')
|
||||||
|
|
||||||
|
# Change PAGEa2 status to OK
|
||||||
|
self.change_backup_status(backup_dir, 'node', page_id_a2, 'OK')
|
||||||
|
|
||||||
|
# PAGEb2 OK
|
||||||
|
# PAGEa2 OK
|
||||||
|
# PAGEb1 OK
|
||||||
|
# PAGEa1 OK
|
||||||
|
# FULLb OK
|
||||||
|
# FULLa OK
|
||||||
|
|
||||||
|
self.delete_expired(
|
||||||
|
backup_dir, 'node',
|
||||||
|
options=['--retention-redundancy=1', '--expired'])
|
||||||
|
|
||||||
|
self.assertEqual(len(self.show_pb(backup_dir, 'node')), 3)
|
||||||
|
|
||||||
|
print(self.show_pb(
|
||||||
|
backup_dir, 'node', as_json=False, as_text=True))
|
||||||
|
|
||||||
|
# Clean after yourself
|
||||||
|
self.del_test_dir(module_name, fname)
|
||||||
|
|
||||||
|
# @unittest.skip("skip")
|
||||||
|
def test_window_merge_interleaved_incremental_chains(self):
|
||||||
|
"""complicated case of interleaved backup chains"""
|
||||||
|
fname = self.id().split('.')[3]
|
||||||
|
node = self.make_simple_node(
|
||||||
|
base_dir=os.path.join(module_name, fname, 'node'),
|
||||||
|
initdb_params=['--data-checksums'])
|
||||||
|
|
||||||
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
||||||
|
self.init_pb(backup_dir)
|
||||||
|
self.add_instance(backup_dir, 'node', node)
|
||||||
|
self.set_archiving(backup_dir, 'node', node)
|
||||||
|
node.slow_start()
|
||||||
|
|
||||||
|
# Take FULL BACKUPs
|
||||||
|
backup_id_a = self.backup_node(backup_dir, 'node', node)
|
||||||
|
backup_id_b = self.backup_node(backup_dir, 'node', node)
|
||||||
|
|
||||||
|
# Change FULL B backup status to ERROR
|
||||||
|
self.change_backup_status(backup_dir, 'node', backup_id_b, 'ERROR')
|
||||||
|
|
||||||
|
# FULLb ERROR
|
||||||
|
# FULLa OK
|
||||||
|
# Take PAGEa1 backup
|
||||||
|
page_id_a1 = self.backup_node(
|
||||||
|
backup_dir, 'node', node, backup_type='page')
|
||||||
|
|
||||||
|
# PAGEa1 OK
|
||||||
|
# FULLb ERROR
|
||||||
|
# FULLa OK
|
||||||
|
# Change FULL B backup status to OK
|
||||||
|
self.change_backup_status(backup_dir, 'node', backup_id_b, 'OK')
|
||||||
|
|
||||||
|
# Change PAGEa1 backup status to ERROR
|
||||||
|
self.change_backup_status(backup_dir, 'node', page_id_a1, 'ERROR')
|
||||||
|
|
||||||
|
# PAGEa1 ERROR
|
||||||
|
# FULLb OK
|
||||||
|
# FULLa OK
|
||||||
|
page_id_b1 = self.backup_node(
|
||||||
|
backup_dir, 'node', node, backup_type='page')
|
||||||
|
|
||||||
|
# PAGEb1 OK
|
||||||
|
# PAGEa1 ERROR
|
||||||
|
# FULLb OK
|
||||||
|
# FULLa OK
|
||||||
|
# Now we start to play with first generation of PAGE backups
|
||||||
|
# Change PAGEb1 status to ERROR
|
||||||
|
self.change_backup_status(backup_dir, 'node', page_id_b1, 'ERROR')
|
||||||
|
|
||||||
|
# Change PAGEa1 status to OK
|
||||||
|
self.change_backup_status(backup_dir, 'node', page_id_a1, 'OK')
|
||||||
|
|
||||||
|
# PAGEb1 ERROR
|
||||||
|
# PAGEa1 OK
|
||||||
|
# FULLb OK
|
||||||
|
# FULLa OK
|
||||||
|
page_id_a2 = self.backup_node(
|
||||||
|
backup_dir, 'node', node, backup_type='page')
|
||||||
|
|
||||||
|
# PAGEa2 OK
|
||||||
|
# PAGEb1 ERROR
|
||||||
|
# PAGEa1 OK
|
||||||
|
# FULLb OK
|
||||||
|
# FULLa OK
|
||||||
|
# Change PAGEa2 status to ERROR
|
||||||
|
self.change_backup_status(backup_dir, 'node', page_id_a2, 'ERROR')
|
||||||
|
|
||||||
|
# Change PAGEb1 status to OK
|
||||||
|
self.change_backup_status(backup_dir, 'node', page_id_b1, 'OK')
|
||||||
|
|
||||||
|
# PAGEa2 ERROR
|
||||||
|
# PAGEb1 OK
|
||||||
|
# PAGEa1 OK
|
||||||
|
# FULLb OK
|
||||||
|
# FULLa OK
|
||||||
|
page_id_b2 = self.backup_node(
|
||||||
|
backup_dir, 'node', node, backup_type='page')
|
||||||
|
|
||||||
|
# Change PAGEa2 status to OK
|
||||||
|
self.change_backup_status(backup_dir, 'node', page_id_a2, 'OK')
|
||||||
|
|
||||||
|
# PAGEb2 OK
|
||||||
|
# PAGEa2 OK
|
||||||
|
# PAGEb1 OK
|
||||||
|
# PAGEa1 OK
|
||||||
|
# FULLb OK
|
||||||
|
# FULLa OK
|
||||||
|
|
||||||
|
# Purge backups
|
||||||
|
backups = os.path.join(backup_dir, 'backups', 'node')
|
||||||
|
for backup in os.listdir(backups):
|
||||||
|
if backup in [page_id_a2, page_id_b2, 'pg_probackup.conf']:
|
||||||
|
continue
|
||||||
|
|
||||||
|
with open(
|
||||||
|
os.path.join(
|
||||||
|
backups, backup, "backup.control"), "a") as conf:
|
||||||
|
conf.write("recovery_time='{:%Y-%m-%d %H:%M:%S}'\n".format(
|
||||||
|
datetime.now() - timedelta(days=3)))
|
||||||
|
|
||||||
|
output = self.delete_expired(
|
||||||
|
backup_dir, 'node',
|
||||||
|
options=['--retention-window=1', '--expired', '--merge-expired'])
|
||||||
|
|
||||||
|
self.assertIn(
|
||||||
|
"Merge incremental chain between FULL backup {0} and backup {1}".format(
|
||||||
|
backup_id_a, page_id_a2),
|
||||||
|
output)
|
||||||
|
|
||||||
|
self.assertIn(
|
||||||
|
"Merging backup {0} with backup {1}".format(
|
||||||
|
page_id_a1, backup_id_a), output)
|
||||||
|
|
||||||
|
self.assertIn(
|
||||||
|
"Rename {0} to {1}".format(
|
||||||
|
backup_id_a, page_id_a1), output)
|
||||||
|
|
||||||
|
self.assertIn(
|
||||||
|
"Merging backup {0} with backup {1}".format(
|
||||||
|
page_id_a2, page_id_a1), output)
|
||||||
|
|
||||||
|
self.assertIn(
|
||||||
|
"Rename {0} to {1}".format(
|
||||||
|
page_id_a1, page_id_a2), output)
|
||||||
|
|
||||||
|
self.assertIn(
|
||||||
|
"Merge incremental chain between FULL backup {0} and backup {1}".format(
|
||||||
|
backup_id_b, page_id_b2),
|
||||||
|
output)
|
||||||
|
|
||||||
|
self.assertIn(
|
||||||
|
"Merging backup {0} with backup {1}".format(
|
||||||
|
page_id_b1, backup_id_b), output)
|
||||||
|
|
||||||
|
self.assertIn(
|
||||||
|
"Rename {0} to {1}".format(
|
||||||
|
backup_id_b, page_id_b1), output)
|
||||||
|
|
||||||
|
self.assertIn(
|
||||||
|
"Merging backup {0} with backup {1}".format(
|
||||||
|
page_id_b2, page_id_b1), output)
|
||||||
|
|
||||||
|
self.assertIn(
|
||||||
|
"Rename {0} to {1}".format(
|
||||||
|
page_id_b1, page_id_b2), output)
|
||||||
|
|
||||||
|
self.assertEqual(len(self.show_pb(backup_dir, 'node')), 2)
|
||||||
|
|
||||||
|
# Clean after yourself
|
||||||
|
self.del_test_dir(module_name, fname)
|
||||||
|
|
||||||
|
# @unittest.skip("skip")
|
||||||
|
def test_window_merge_interleaved_incremental_chains_1(self):
|
||||||
|
"""
|
||||||
|
PAGEb3
|
||||||
|
PAGEb2
|
||||||
|
PAGEb1
|
||||||
|
PAGEa1
|
||||||
|
FULLb
|
||||||
|
FULLa
|
||||||
|
"""
|
||||||
|
fname = self.id().split('.')[3]
|
||||||
|
node = self.make_simple_node(
|
||||||
|
base_dir=os.path.join(module_name, fname, 'node'),
|
||||||
|
initdb_params=['--data-checksums'])
|
||||||
|
|
||||||
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
||||||
|
self.init_pb(backup_dir)
|
||||||
|
self.add_instance(backup_dir, 'node', node)
|
||||||
|
self.set_archiving(backup_dir, 'node', node)
|
||||||
|
node.slow_start()
|
||||||
|
|
||||||
|
node.pgbench_init(scale=3)
|
||||||
|
|
||||||
|
# Take FULL BACKUPs
|
||||||
|
backup_id_a = self.backup_node(backup_dir, 'node', node)
|
||||||
|
pgbench = node.pgbench(options=['-t', '10', '-c', '2'])
|
||||||
|
pgbench.wait()
|
||||||
|
|
||||||
|
backup_id_b = self.backup_node(backup_dir, 'node', node)
|
||||||
|
pgbench = node.pgbench(options=['-t', '10', '-c', '2'])
|
||||||
|
pgbench.wait()
|
||||||
|
|
||||||
|
# Change FULL B backup status to ERROR
|
||||||
|
self.change_backup_status(backup_dir, 'node', backup_id_b, 'ERROR')
|
||||||
|
|
||||||
|
page_id_a1 = self.backup_node(
|
||||||
|
backup_dir, 'node', node, backup_type='page')
|
||||||
|
|
||||||
|
pgdata_a1 = self.pgdata_content(node.data_dir)
|
||||||
|
|
||||||
|
pgbench = node.pgbench(options=['-t', '10', '-c', '2'])
|
||||||
|
pgbench.wait()
|
||||||
|
|
||||||
|
# PAGEa1 OK
|
||||||
|
# FULLb ERROR
|
||||||
|
# FULLa OK
|
||||||
|
# Change FULL B backup status to OK
|
||||||
|
self.change_backup_status(backup_dir, 'node', backup_id_b, 'OK')
|
||||||
|
|
||||||
|
# Change PAGEa1 backup status to ERROR
|
||||||
|
self.change_backup_status(backup_dir, 'node', page_id_a1, 'ERROR')
|
||||||
|
|
||||||
|
# PAGEa1 ERROR
|
||||||
|
# FULLb OK
|
||||||
|
# FULLa OK
|
||||||
|
page_id_b1 = self.backup_node(
|
||||||
|
backup_dir, 'node', node, backup_type='page')
|
||||||
|
|
||||||
|
pgbench = node.pgbench(options=['-t', '10', '-c', '2'])
|
||||||
|
pgbench.wait()
|
||||||
|
|
||||||
|
page_id_b2 = self.backup_node(
|
||||||
|
backup_dir, 'node', node, backup_type='page')
|
||||||
|
|
||||||
|
pgbench = node.pgbench(options=['-t', '10', '-c', '2'])
|
||||||
|
pgbench.wait()
|
||||||
|
|
||||||
|
page_id_b3 = self.backup_node(
|
||||||
|
backup_dir, 'node', node, backup_type='page')
|
||||||
|
|
||||||
|
pgdata_b3 = self.pgdata_content(node.data_dir)
|
||||||
|
|
||||||
|
pgbench = node.pgbench(options=['-t', '10', '-c', '2'])
|
||||||
|
pgbench.wait()
|
||||||
|
|
||||||
|
# PAGEb3 OK
|
||||||
|
# PAGEb2 OK
|
||||||
|
# PAGEb1 OK
|
||||||
|
# PAGEa1 ERROR
|
||||||
|
# FULLb OK
|
||||||
|
# FULLa OK
|
||||||
|
|
||||||
|
# Change PAGEa1 backup status to ERROR
|
||||||
|
self.change_backup_status(backup_dir, 'node', page_id_a1, 'OK')
|
||||||
|
|
||||||
|
# PAGEb3 OK
|
||||||
|
# PAGEb2 OK
|
||||||
|
# PAGEb1 OK
|
||||||
|
# PAGEa1 OK
|
||||||
|
# FULLb OK
|
||||||
|
# FULLa OK
|
||||||
|
|
||||||
|
# Purge backups
|
||||||
|
backups = os.path.join(backup_dir, 'backups', 'node')
|
||||||
|
for backup in os.listdir(backups):
|
||||||
|
if backup in [page_id_a1, page_id_b3, 'pg_probackup.conf']:
|
||||||
|
continue
|
||||||
|
|
||||||
|
with open(
|
||||||
|
os.path.join(
|
||||||
|
backups, backup, "backup.control"), "a") as conf:
|
||||||
|
conf.write("recovery_time='{:%Y-%m-%d %H:%M:%S}'\n".format(
|
||||||
|
datetime.now() - timedelta(days=3)))
|
||||||
|
|
||||||
|
output = self.delete_expired(
|
||||||
|
backup_dir, 'node',
|
||||||
|
options=['--retention-window=1', '--expired', '--merge-expired'])
|
||||||
|
|
||||||
|
|
||||||
|
self.assertEqual(len(self.show_pb(backup_dir, 'node')), 2)
|
||||||
|
|
||||||
|
self.assertEqual(
|
||||||
|
self.show_pb(backup_dir, 'node')[1]['id'],
|
||||||
|
page_id_b3)
|
||||||
|
|
||||||
|
self.assertEqual(
|
||||||
|
self.show_pb(backup_dir, 'node')[0]['id'],
|
||||||
|
page_id_a1)
|
||||||
|
|
||||||
|
self.assertEqual(
|
||||||
|
self.show_pb(backup_dir, 'node')[1]['backup-mode'],
|
||||||
|
'FULL')
|
||||||
|
|
||||||
|
self.assertEqual(
|
||||||
|
self.show_pb(backup_dir, 'node')[0]['backup-mode'],
|
||||||
|
'FULL')
|
||||||
|
|
||||||
|
node.cleanup()
|
||||||
|
|
||||||
|
# Data correctness of PAGEa3
|
||||||
|
self.restore_node(backup_dir, 'node', node, backup_id=page_id_a1)
|
||||||
|
pgdata_restored_a1 = self.pgdata_content(node.data_dir)
|
||||||
|
self.compare_pgdata(pgdata_a1, pgdata_restored_a1)
|
||||||
|
|
||||||
|
node.cleanup()
|
||||||
|
|
||||||
|
# Data correctness of PAGEb3
|
||||||
|
self.restore_node(backup_dir, 'node', node, backup_id=page_id_b3)
|
||||||
|
pgdata_restored_b3 = self.pgdata_content(node.data_dir)
|
||||||
|
self.compare_pgdata(pgdata_b3, pgdata_restored_b3)
|
||||||
|
|
||||||
|
# Clean after yourself
|
||||||
|
self.del_test_dir(module_name, fname)
|
||||||
|
|
||||||
|
# @unittest.skip("skip")
|
||||||
|
def test_window_merge_multiple_descendants(self):
|
||||||
|
"""
|
||||||
|
PAGEb3
|
||||||
|
| PAGEa3
|
||||||
|
-----------------------------retention window
|
||||||
|
PAGEb2 /
|
||||||
|
| PAGEa2 / should be deleted
|
||||||
|
PAGEb1 \ /
|
||||||
|
| PAGEa1
|
||||||
|
FULLb |
|
||||||
|
FULLa
|
||||||
|
"""
|
||||||
|
fname = self.id().split('.')[3]
|
||||||
|
node = self.make_simple_node(
|
||||||
|
base_dir=os.path.join(module_name, fname, 'node'),
|
||||||
|
initdb_params=['--data-checksums'])
|
||||||
|
|
||||||
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
||||||
|
self.init_pb(backup_dir)
|
||||||
|
self.add_instance(backup_dir, 'node', node)
|
||||||
|
self.set_archiving(backup_dir, 'node', node)
|
||||||
|
node.slow_start()
|
||||||
|
|
||||||
|
node.pgbench_init(scale=3)
|
||||||
|
|
||||||
|
# Take FULL BACKUPs
|
||||||
|
backup_id_a = self.backup_node(backup_dir, 'node', node)
|
||||||
|
pgbench = node.pgbench(options=['-T', '10', '-c', '2'])
|
||||||
|
pgbench.wait()
|
||||||
|
|
||||||
|
backup_id_b = self.backup_node(backup_dir, 'node', node)
|
||||||
|
pgbench = node.pgbench(options=['-T', '10', '-c', '2'])
|
||||||
|
pgbench.wait()
|
||||||
|
|
||||||
|
# Change FULLb backup status to ERROR
|
||||||
|
self.change_backup_status(backup_dir, 'node', backup_id_b, 'ERROR')
|
||||||
|
|
||||||
|
page_id_a1 = self.backup_node(
|
||||||
|
backup_dir, 'node', node, backup_type='page')
|
||||||
|
|
||||||
|
pgbench = node.pgbench(options=['-T', '10', '-c', '2'])
|
||||||
|
pgbench.wait()
|
||||||
|
|
||||||
|
# Change FULLb backup status to OK
|
||||||
|
self.change_backup_status(backup_dir, 'node', backup_id_b, 'OK')
|
||||||
|
|
||||||
|
# Change PAGEa1 backup status to ERROR
|
||||||
|
self.change_backup_status(backup_dir, 'node', page_id_a1, 'ERROR')
|
||||||
|
|
||||||
|
# PAGEa1 ERROR
|
||||||
|
# FULLb OK
|
||||||
|
# FULLa OK
|
||||||
|
|
||||||
|
page_id_b1 = self.backup_node(
|
||||||
|
backup_dir, 'node', node, backup_type='page')
|
||||||
|
|
||||||
|
# PAGEb1 OK
|
||||||
|
# PAGEa1 ERROR
|
||||||
|
# FULLb OK
|
||||||
|
# FULLa OK
|
||||||
|
|
||||||
|
pgbench = node.pgbench(options=['-T', '10', '-c', '2'])
|
||||||
|
pgbench.wait()
|
||||||
|
|
||||||
|
# Change PAGEa1 backup status to OK
|
||||||
|
self.change_backup_status(backup_dir, 'node', page_id_a1, 'OK')
|
||||||
|
|
||||||
|
# Change PAGEb1 backup status to ERROR
|
||||||
|
self.change_backup_status(backup_dir, 'node', page_id_b1, 'ERROR')
|
||||||
|
|
||||||
|
# PAGEb1 ERROR
|
||||||
|
# PAGEa1 OK
|
||||||
|
# FULLb OK
|
||||||
|
# FULLa OK
|
||||||
|
|
||||||
|
page_id_a2 = self.backup_node(
|
||||||
|
backup_dir, 'node', node, backup_type='page')
|
||||||
|
|
||||||
|
pgbench = node.pgbench(options=['-T', '10', '-c', '2'])
|
||||||
|
pgbench.wait()
|
||||||
|
|
||||||
|
# PAGEa2 OK
|
||||||
|
# PAGEb1 ERROR
|
||||||
|
# PAGEa1 OK
|
||||||
|
# FULLb OK
|
||||||
|
# FULLa OK
|
||||||
|
|
||||||
|
# Change PAGEb1 backup status to OK
|
||||||
|
self.change_backup_status(backup_dir, 'node', page_id_b1, 'OK')
|
||||||
|
|
||||||
|
# Change PAGEa2 backup status to ERROR
|
||||||
|
self.change_backup_status(backup_dir, 'node', page_id_a2, 'ERROR')
|
||||||
|
|
||||||
|
# PAGEa2 ERROR
|
||||||
|
# PAGEb1 OK
|
||||||
|
# PAGEa1 OK
|
||||||
|
# FULLb OK
|
||||||
|
# FULLa OK
|
||||||
|
|
||||||
|
page_id_b2 = self.backup_node(
|
||||||
|
backup_dir, 'node', node, backup_type='page')
|
||||||
|
|
||||||
|
pgbench = node.pgbench(options=['-T', '10', '-c', '2'])
|
||||||
|
pgbench.wait()
|
||||||
|
|
||||||
|
# PAGEb2 OK
|
||||||
|
# PAGEa2 ERROR
|
||||||
|
# PAGEb1 OK
|
||||||
|
# PAGEa1 OK
|
||||||
|
# FULLb OK
|
||||||
|
# FULLa OK
|
||||||
|
|
||||||
|
# Change PAGEb2 and PAGEb1 status to ERROR
|
||||||
|
self.change_backup_status(backup_dir, 'node', page_id_b2, 'ERROR')
|
||||||
|
self.change_backup_status(backup_dir, 'node', page_id_b1, 'ERROR')
|
||||||
|
|
||||||
|
# PAGEb2 ERROR
|
||||||
|
# PAGEa2 ERROR
|
||||||
|
# PAGEb1 ERROR
|
||||||
|
# PAGEa1 OK
|
||||||
|
# FULLb OK
|
||||||
|
# FULLa OK
|
||||||
|
|
||||||
|
page_id_a3 = self.backup_node(
|
||||||
|
backup_dir, 'node', node, backup_type='page')
|
||||||
|
pgbench = node.pgbench(options=['-T', '10', '-c', '2'])
|
||||||
|
pgbench.wait()
|
||||||
|
|
||||||
|
# PAGEa3 OK
|
||||||
|
# PAGEb2 ERROR
|
||||||
|
# PAGEa2 ERROR
|
||||||
|
# PAGEb1 ERROR
|
||||||
|
# PAGEa1 OK
|
||||||
|
# FULLb OK
|
||||||
|
# FULLa OK
|
||||||
|
|
||||||
|
# Change PAGEa3 status to ERROR
|
||||||
|
self.change_backup_status(backup_dir, 'node', page_id_a3, 'ERROR')
|
||||||
|
|
||||||
|
# Change PAGEb2 status to OK
|
||||||
|
self.change_backup_status(backup_dir, 'node', page_id_b2, 'OK')
|
||||||
|
|
||||||
|
page_id_b3 = self.backup_node(
|
||||||
|
backup_dir, 'node', node, backup_type='page')
|
||||||
|
|
||||||
|
# PAGEb3 OK
|
||||||
|
# PAGEa3 ERROR
|
||||||
|
# PAGEb2 OK
|
||||||
|
# PAGEa2 ERROR
|
||||||
|
# PAGEb1 ERROR
|
||||||
|
# PAGEa1 OK
|
||||||
|
# FULLb OK
|
||||||
|
# FULLa OK
|
||||||
|
|
||||||
|
# Change PAGEa3, PAGEa2 and PAGEb1 status to OK
|
||||||
|
self.change_backup_status(backup_dir, 'node', page_id_a3, 'OK')
|
||||||
|
self.change_backup_status(backup_dir, 'node', page_id_a2, 'OK')
|
||||||
|
self.change_backup_status(backup_dir, 'node', page_id_b1, 'OK')
|
||||||
|
|
||||||
|
# PAGEb3 OK
|
||||||
|
# PAGEa3 OK
|
||||||
|
# PAGEb2 OK
|
||||||
|
# PAGEa2 OK
|
||||||
|
# PAGEb1 OK
|
||||||
|
# PAGEa1 OK
|
||||||
|
# FULLb OK
|
||||||
|
# FULLa OK
|
||||||
|
|
||||||
|
# Check that page_id_a3 and page_id_a2 are both direct descendants of page_id_a1
|
||||||
|
self.assertEqual(
|
||||||
|
self.show_pb(backup_dir, 'node', backup_id=page_id_a3)['parent-backup-id'],
|
||||||
|
page_id_a1)
|
||||||
|
|
||||||
|
self.assertEqual(
|
||||||
|
self.show_pb(backup_dir, 'node', backup_id=page_id_a2)['parent-backup-id'],
|
||||||
|
page_id_a1)
|
||||||
|
|
||||||
|
print("Backups {0} and {1} are children of {2}".format(
|
||||||
|
page_id_a3, page_id_a2, page_id_a1))
|
||||||
|
|
||||||
|
# Purge backups
|
||||||
|
backups = os.path.join(backup_dir, 'backups', 'node')
|
||||||
|
for backup in os.listdir(backups):
|
||||||
|
if backup in [page_id_a3, page_id_b3, 'pg_probackup.conf']:
|
||||||
|
continue
|
||||||
|
|
||||||
|
with open(
|
||||||
|
os.path.join(
|
||||||
|
backups, backup, "backup.control"), "a") as conf:
|
||||||
|
conf.write("recovery_time='{:%Y-%m-%d %H:%M:%S}'\n".format(
|
||||||
|
datetime.now() - timedelta(days=3)))
|
||||||
|
|
||||||
|
output = self.delete_expired(
|
||||||
|
backup_dir, 'node',
|
||||||
|
options=[
|
||||||
|
'--retention-window=1', '--expired',
|
||||||
|
'--merge-expired', '--log-level-console=log'])
|
||||||
|
|
||||||
|
print(output)
|
||||||
|
|
||||||
|
self.assertEqual(len(self.show_pb(backup_dir, 'node')), 3)
|
||||||
|
|
||||||
|
# Merging chain A
|
||||||
|
self.assertIn(
|
||||||
|
"Merge incremental chain between FULL backup {0} and backup {1}".format(
|
||||||
|
backup_id_a, page_id_a3),
|
||||||
|
output)
|
||||||
|
|
||||||
|
self.assertIn(
|
||||||
|
"Merging backup {0} with backup {1}".format(
|
||||||
|
page_id_a1, backup_id_a), output)
|
||||||
|
|
||||||
|
self.assertIn(
|
||||||
|
"INFO: Rename {0} to {1}".format(
|
||||||
|
backup_id_a, page_id_a1), output)
|
||||||
|
|
||||||
|
self.assertIn(
|
||||||
|
"WARNING: Backup {0} has multiple valid descendants. "
|
||||||
|
"Automatic merge is not possible.".format(
|
||||||
|
page_id_a1), output)
|
||||||
|
|
||||||
|
# Merge chain B
|
||||||
|
self.assertIn(
|
||||||
|
"Merge incremental chain between FULL backup {0} and backup {1}".format(
|
||||||
|
backup_id_b, page_id_b3),
|
||||||
|
output)
|
||||||
|
|
||||||
|
self.assertIn(
|
||||||
|
"Merging backup {0} with backup {1}".format(
|
||||||
|
page_id_b1, backup_id_b), output)
|
||||||
|
|
||||||
|
self.assertIn(
|
||||||
|
"INFO: Rename {0} to {1}".format(
|
||||||
|
backup_id_b, page_id_b1), output)
|
||||||
|
|
||||||
|
self.assertIn(
|
||||||
|
"Merging backup {0} with backup {1}".format(
|
||||||
|
page_id_b2, page_id_b1), output)
|
||||||
|
|
||||||
|
self.assertIn(
|
||||||
|
"INFO: Rename {0} to {1}".format(
|
||||||
|
page_id_b1, page_id_b2), output)
|
||||||
|
|
||||||
|
self.assertIn(
|
||||||
|
"Merging backup {0} with backup {1}".format(
|
||||||
|
page_id_b3, page_id_b2), output)
|
||||||
|
|
||||||
|
self.assertIn(
|
||||||
|
"INFO: Rename {0} to {1}".format(
|
||||||
|
page_id_b2, page_id_b3), output)
|
||||||
|
|
||||||
|
# this backup deleted because it is not guarded by retention
|
||||||
|
self.assertIn(
|
||||||
|
"INFO: Delete: {0}".format(
|
||||||
|
page_id_a1), output)
|
||||||
|
|
||||||
|
|
||||||
|
self.assertEqual(
|
||||||
|
self.show_pb(backup_dir, 'node')[2]['id'],
|
||||||
|
page_id_b3)
|
||||||
|
|
||||||
|
self.assertEqual(
|
||||||
|
self.show_pb(backup_dir, 'node')[1]['id'],
|
||||||
|
page_id_a3)
|
||||||
|
|
||||||
|
self.assertEqual(
|
||||||
|
self.show_pb(backup_dir, 'node')[0]['id'],
|
||||||
|
page_id_a1)
|
||||||
|
|
||||||
|
self.assertEqual(
|
||||||
|
self.show_pb(backup_dir, 'node')[2]['backup-mode'],
|
||||||
|
'FULL')
|
||||||
|
|
||||||
|
self.assertEqual(
|
||||||
|
self.show_pb(backup_dir, 'node')[1]['backup-mode'],
|
||||||
|
'PAGE')
|
||||||
|
|
||||||
|
self.assertEqual(
|
||||||
|
self.show_pb(backup_dir, 'node')[0]['backup-mode'],
|
||||||
|
'FULL')
|
||||||
|
|
||||||
|
# Clean after yourself
|
||||||
|
self.del_test_dir(module_name, fname)
|
||||||
|
|
||||||
|
# @unittest.skip("skip")
|
||||||
|
def test_window_chains(self):
|
||||||
|
"""
|
||||||
|
PAGE
|
||||||
|
-------window
|
||||||
|
PAGE
|
||||||
|
PAGE
|
||||||
|
FULL
|
||||||
|
PAGE
|
||||||
|
PAGE
|
||||||
|
FULL
|
||||||
|
"""
|
||||||
|
fname = self.id().split('.')[3]
|
||||||
|
node = self.make_simple_node(
|
||||||
|
base_dir=os.path.join(module_name, fname, 'node'),
|
||||||
|
initdb_params=['--data-checksums'])
|
||||||
|
|
||||||
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
||||||
|
self.init_pb(backup_dir)
|
||||||
|
self.add_instance(backup_dir, 'node', node)
|
||||||
|
self.set_archiving(backup_dir, 'node', node)
|
||||||
|
node.slow_start()
|
||||||
|
|
||||||
|
node.pgbench_init(scale=3)
|
||||||
|
|
||||||
|
# Chain A
|
||||||
|
backup_id_a = self.backup_node(backup_dir, 'node', node)
|
||||||
|
page_id_a1 = self.backup_node(
|
||||||
|
backup_dir, 'node', node, backup_type='page')
|
||||||
|
|
||||||
|
page_id_a2 = self.backup_node(
|
||||||
|
backup_dir, 'node', node, backup_type='page')
|
||||||
|
|
||||||
|
# Chain B
|
||||||
|
backup_id_b = self.backup_node(backup_dir, 'node', node)
|
||||||
|
|
||||||
|
pgbench = node.pgbench(options=['-T', '10', '-c', '2'])
|
||||||
|
pgbench.wait()
|
||||||
|
|
||||||
|
page_id_b1 = self.backup_node(
|
||||||
|
backup_dir, 'node', node, backup_type='delta')
|
||||||
|
|
||||||
|
pgbench = node.pgbench(options=['-T', '10', '-c', '2'])
|
||||||
|
pgbench.wait()
|
||||||
|
|
||||||
|
page_id_b2 = self.backup_node(
|
||||||
|
backup_dir, 'node', node, backup_type='page')
|
||||||
|
|
||||||
|
pgbench = node.pgbench(options=['-T', '10', '-c', '2'])
|
||||||
|
pgbench.wait()
|
||||||
|
|
||||||
|
page_id_b3 = self.backup_node(
|
||||||
|
backup_dir, 'node', node, backup_type='delta')
|
||||||
|
|
||||||
|
pgdata = self.pgdata_content(node.data_dir)
|
||||||
|
|
||||||
|
# Purge backups
|
||||||
|
backups = os.path.join(backup_dir, 'backups', 'node')
|
||||||
|
for backup in os.listdir(backups):
|
||||||
|
if backup in [page_id_b3, 'pg_probackup.conf']:
|
||||||
|
continue
|
||||||
|
|
||||||
|
with open(
|
||||||
|
os.path.join(
|
||||||
|
backups, backup, "backup.control"), "a") as conf:
|
||||||
|
conf.write("recovery_time='{:%Y-%m-%d %H:%M:%S}'\n".format(
|
||||||
|
datetime.now() - timedelta(days=3)))
|
||||||
|
|
||||||
|
output = self.delete_expired(
|
||||||
|
backup_dir, 'node',
|
||||||
|
options=[
|
||||||
|
'--retention-window=1', '--expired',
|
||||||
|
'--merge-expired', '--log-level-console=log'])
|
||||||
|
|
||||||
|
self.assertEqual(len(self.show_pb(backup_dir, 'node')), 1)
|
||||||
|
|
||||||
|
node.cleanup()
|
||||||
|
|
||||||
|
self.restore_node(backup_dir, 'node', node)
|
||||||
|
|
||||||
|
pgdata_restored = self.pgdata_content(node.data_dir)
|
||||||
|
self.compare_pgdata(pgdata, pgdata_restored)
|
||||||
|
|
||||||
|
# Clean after yourself
|
||||||
|
self.del_test_dir(module_name, fname)
|
||||||
|
|
||||||
|
|
||||||
|
# @unittest.skip("skip")
|
||||||
|
def test_window_chains_1(self):
|
||||||
|
"""
|
||||||
|
PAGE
|
||||||
|
-------window
|
||||||
|
PAGE
|
||||||
|
PAGE
|
||||||
|
FULL
|
||||||
|
PAGE
|
||||||
|
PAGE
|
||||||
|
FULL
|
||||||
|
"""
|
||||||
|
fname = self.id().split('.')[3]
|
||||||
|
node = self.make_simple_node(
|
||||||
|
base_dir=os.path.join(module_name, fname, 'node'),
|
||||||
|
initdb_params=['--data-checksums'])
|
||||||
|
|
||||||
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
||||||
|
self.init_pb(backup_dir)
|
||||||
|
self.add_instance(backup_dir, 'node', node)
|
||||||
|
self.set_archiving(backup_dir, 'node', node)
|
||||||
|
node.slow_start()
|
||||||
|
|
||||||
|
node.pgbench_init(scale=3)
|
||||||
|
|
||||||
|
# Chain A
|
||||||
|
backup_id_a = self.backup_node(backup_dir, 'node', node)
|
||||||
|
page_id_a1 = self.backup_node(
|
||||||
|
backup_dir, 'node', node, backup_type='page')
|
||||||
|
|
||||||
|
page_id_a2 = self.backup_node(
|
||||||
|
backup_dir, 'node', node, backup_type='page')
|
||||||
|
|
||||||
|
# Chain B
|
||||||
|
backup_id_b = self.backup_node(backup_dir, 'node', node)
|
||||||
|
|
||||||
|
page_id_b1 = self.backup_node(
|
||||||
|
backup_dir, 'node', node, backup_type='delta')
|
||||||
|
|
||||||
|
page_id_b2 = self.backup_node(
|
||||||
|
backup_dir, 'node', node, backup_type='page')
|
||||||
|
|
||||||
|
page_id_b3 = self.backup_node(
|
||||||
|
backup_dir, 'node', node, backup_type='delta')
|
||||||
|
|
||||||
|
pgdata = self.pgdata_content(node.data_dir)
|
||||||
|
|
||||||
|
# Purge backups
|
||||||
|
backups = os.path.join(backup_dir, 'backups', 'node')
|
||||||
|
for backup in os.listdir(backups):
|
||||||
|
if backup in [page_id_b3, 'pg_probackup.conf']:
|
||||||
|
continue
|
||||||
|
|
||||||
|
with open(
|
||||||
|
os.path.join(
|
||||||
|
backups, backup, "backup.control"), "a") as conf:
|
||||||
|
conf.write("recovery_time='{:%Y-%m-%d %H:%M:%S}'\n".format(
|
||||||
|
datetime.now() - timedelta(days=3)))
|
||||||
|
|
||||||
|
output = self.delete_expired(
|
||||||
|
backup_dir, 'node',
|
||||||
|
options=[
|
||||||
|
'--retention-window=1',
|
||||||
|
'--merge-expired', '--log-level-console=log'])
|
||||||
|
|
||||||
|
self.assertEqual(len(self.show_pb(backup_dir, 'node')), 4)
|
||||||
|
|
||||||
|
self.assertIn(
|
||||||
|
"There are no backups to delete by retention policy",
|
||||||
|
output)
|
||||||
|
|
||||||
|
self.assertIn(
|
||||||
|
"Retention merging finished",
|
||||||
|
output)
|
||||||
|
|
||||||
|
output = self.delete_expired(
|
||||||
|
backup_dir, 'node',
|
||||||
|
options=[
|
||||||
|
'--retention-window=1',
|
||||||
|
'--expired', '--log-level-console=log'])
|
||||||
|
|
||||||
|
self.assertEqual(len(self.show_pb(backup_dir, 'node')), 1)
|
||||||
|
|
||||||
|
self.assertIn(
|
||||||
|
"There are no backups to merge by retention policy",
|
||||||
|
output)
|
||||||
|
|
||||||
|
self.assertIn(
|
||||||
|
"Purging finished",
|
||||||
|
output)
|
||||||
|
|
||||||
|
# Clean after yourself
|
||||||
|
self.del_test_dir(module_name, fname)
|
||||||
|
|
||||||
|
# @unittest.skip("skip")
|
||||||
|
def test_window_error_backups(self):
|
||||||
|
"""
|
||||||
|
PAGE ERROR
|
||||||
|
-------window
|
||||||
|
PAGE ERROR
|
||||||
|
PAGE ERROR
|
||||||
|
PAGE ERROR
|
||||||
|
FULL ERROR
|
||||||
|
FULL
|
||||||
|
-------redundancy
|
||||||
|
"""
|
||||||
|
fname = self.id().split('.')[3]
|
||||||
|
node = self.make_simple_node(
|
||||||
|
base_dir=os.path.join(module_name, fname, 'node'),
|
||||||
|
initdb_params=['--data-checksums'])
|
||||||
|
|
||||||
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
||||||
|
self.init_pb(backup_dir)
|
||||||
|
self.add_instance(backup_dir, 'node', node)
|
||||||
|
self.set_archiving(backup_dir, 'node', node)
|
||||||
|
node.slow_start()
|
||||||
|
|
||||||
|
node.pgbench_init(scale=3)
|
||||||
|
|
||||||
|
# Take FULL BACKUPs
|
||||||
|
backup_id_a1 = self.backup_node(backup_dir, 'node', node)
|
||||||
|
page_id_a2 = self.backup_node(
|
||||||
|
backup_dir, 'node', node, backup_type='page')
|
||||||
|
|
||||||
|
# Change FULLb backup status to ERROR
|
||||||
|
#self.change_backup_status(backup_dir, 'node', backup_id_b, 'ERROR')
|
||||||
|
Reference in New Issue
Block a user