mirror of
https://github.com/postgrespro/pg_probackup.git
synced 2025-03-17 21:18:00 +02:00
Merge branch 'master' into issue_63
This commit is contained in:
commit
cdc42241b3
195
tests/merge.py
195
tests/merge.py
@ -1666,6 +1666,141 @@ class MergeTest(ProbackupTest, unittest.TestCase):
|
||||
|
||||
self.del_test_dir(module_name, fname)
|
||||
|
||||
def test_failed_merge_after_delete(self):
|
||||
"""
|
||||
"""
|
||||
fname = self.id().split('.')[3]
|
||||
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
||||
node = self.make_simple_node(
|
||||
base_dir=os.path.join(module_name, fname, 'node'),
|
||||
set_replication=True,
|
||||
initdb_params=['--data-checksums'],
|
||||
pg_options={'autovacuum': 'off'})
|
||||
|
||||
self.init_pb(backup_dir)
|
||||
self.add_instance(backup_dir, 'node', node)
|
||||
self.set_archiving(backup_dir, 'node', node)
|
||||
node.slow_start()
|
||||
|
||||
# add database
|
||||
node.safe_psql(
|
||||
'postgres',
|
||||
'CREATE DATABASE testdb')
|
||||
|
||||
dboid = node.safe_psql(
|
||||
"postgres",
|
||||
"select oid from pg_database where datname = 'testdb'").rstrip()
|
||||
|
||||
# take FULL backup
|
||||
full_id = self.backup_node(
|
||||
backup_dir, 'node', node, options=['--stream'])
|
||||
|
||||
# drop database
|
||||
node.safe_psql(
|
||||
'postgres',
|
||||
'DROP DATABASE testdb')
|
||||
|
||||
# take PAGE backup
|
||||
page_id = self.backup_node(
|
||||
backup_dir, 'node', node, backup_type='page')
|
||||
|
||||
gdb = self.merge_backup(
|
||||
backup_dir, 'node', page_id,
|
||||
gdb=True, options=['--log-level-console=verbose'])
|
||||
|
||||
gdb.set_breakpoint('delete_backup_files')
|
||||
gdb.run_until_break()
|
||||
|
||||
gdb.set_breakpoint('parray_bsearch')
|
||||
gdb.continue_execution_until_break()
|
||||
|
||||
gdb.set_breakpoint('pgFileDelete')
|
||||
gdb.continue_execution_until_break(20)
|
||||
|
||||
gdb._execute('signal SIGKILL')
|
||||
|
||||
# backup half-merged
|
||||
self.assertEqual(
|
||||
'OK', self.show_pb(backup_dir, 'node')[0]['status'])
|
||||
|
||||
self.assertEqual(
|
||||
full_id, self.show_pb(backup_dir, 'node')[0]['id'])
|
||||
|
||||
db_path = os.path.join(
|
||||
backup_dir, 'backups', 'node',
|
||||
full_id, 'database', 'base', dboid)
|
||||
|
||||
self.assertFalse(
|
||||
os.path.isdir(db_path),
|
||||
'Directory {0} should not exist'.format(
|
||||
db_path, full_id))
|
||||
|
||||
self.del_test_dir(module_name, fname)
|
||||
|
||||
def test_failed_merge_after_delete_1(self):
|
||||
"""
|
||||
"""
|
||||
fname = self.id().split('.')[3]
|
||||
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
||||
node = self.make_simple_node(
|
||||
base_dir=os.path.join(module_name, fname, 'node'),
|
||||
set_replication=True,
|
||||
initdb_params=['--data-checksums'],
|
||||
pg_options={'autovacuum': 'off'})
|
||||
|
||||
self.init_pb(backup_dir)
|
||||
self.add_instance(backup_dir, 'node', node)
|
||||
self.set_archiving(backup_dir, 'node', node)
|
||||
node.slow_start()
|
||||
|
||||
# add database
|
||||
node.pgbench_init(scale=1)
|
||||
|
||||
# take FULL backup
|
||||
full_id = self.backup_node(
|
||||
backup_dir, 'node', node, options=['--stream'])
|
||||
|
||||
pgdata = self.pgdata_content(node.data_dir)
|
||||
|
||||
# drop database
|
||||
pgbench = node.pgbench(options=['-T', '10', '-c', '2', '--no-vacuum'])
|
||||
pgbench.wait()
|
||||
|
||||
# take PAGE backup
|
||||
page_id = self.backup_node(
|
||||
backup_dir, 'node', node, backup_type='page')
|
||||
|
||||
gdb = self.merge_backup(
|
||||
backup_dir, 'node', page_id,
|
||||
gdb=True, options=['--log-level-console=verbose'])
|
||||
|
||||
gdb.set_breakpoint('delete_backup_files')
|
||||
gdb.run_until_break()
|
||||
|
||||
gdb.set_breakpoint('parray_bsearch')
|
||||
gdb.continue_execution_until_break()
|
||||
|
||||
gdb.set_breakpoint('pgFileDelete')
|
||||
gdb.continue_execution_until_break(30)
|
||||
|
||||
gdb._execute('signal SIGKILL')
|
||||
|
||||
# backup half-merged
|
||||
self.assertEqual(
|
||||
'OK', self.show_pb(backup_dir, 'node')[0]['status'])
|
||||
|
||||
self.assertEqual(
|
||||
full_id, self.show_pb(backup_dir, 'node')[0]['id'])
|
||||
|
||||
# restore
|
||||
node.cleanup()
|
||||
self.restore_node(backup_dir, 'node', node)
|
||||
|
||||
pgdata_restored = self.pgdata_content(node.data_dir)
|
||||
self.compare_pgdata(pgdata, pgdata_restored)
|
||||
|
||||
self.del_test_dir(module_name, fname)
|
||||
|
||||
# @unittest.skip("skip")
|
||||
def test_merge_backup_from_future(self):
|
||||
"""
|
||||
@ -1906,6 +2041,66 @@ class MergeTest(ProbackupTest, unittest.TestCase):
|
||||
# Clean after yourself
|
||||
self.del_test_dir(module_name, fname)
|
||||
|
||||
# @unittest.skip("skip")
|
||||
def test_smart_merge(self):
|
||||
"""
|
||||
make node, create database, take full backup, drop database,
|
||||
take PAGE backup and merge it into FULL,
|
||||
make sure that files from dropped database are not
|
||||
copied during restore
|
||||
https://github.com/postgrespro/pg_probackup/issues/63
|
||||
"""
|
||||
fname = self.id().split('.')[3]
|
||||
node = self.make_simple_node(
|
||||
base_dir=os.path.join(module_name, fname, 'node'),
|
||||
set_replication=True,
|
||||
initdb_params=['--data-checksums'])
|
||||
|
||||
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
||||
self.init_pb(backup_dir)
|
||||
self.add_instance(backup_dir, 'node', node)
|
||||
self.set_archiving(backup_dir, 'node', node)
|
||||
node.slow_start()
|
||||
|
||||
# create database
|
||||
node.safe_psql(
|
||||
"postgres",
|
||||
"CREATE DATABASE testdb")
|
||||
|
||||
# take FULL backup
|
||||
full_id = self.backup_node(backup_dir, 'node', node)
|
||||
|
||||
# drop database
|
||||
node.safe_psql(
|
||||
"postgres",
|
||||
"DROP DATABASE testdb")
|
||||
|
||||
# take PAGE backup
|
||||
page_id = self.backup_node(
|
||||
backup_dir, 'node', node, backup_type='page')
|
||||
|
||||
# get delta between FULL and PAGE filelists
|
||||
filelist_full = self.get_backup_filelist(
|
||||
backup_dir, 'node', full_id)
|
||||
|
||||
filelist_page = self.get_backup_filelist(
|
||||
backup_dir, 'node', page_id)
|
||||
|
||||
filelist_diff = self.get_backup_filelist_diff(
|
||||
filelist_full, filelist_page)
|
||||
|
||||
# merge PAGE backup
|
||||
self.merge_backup(
|
||||
backup_dir, 'node', page_id,
|
||||
options=['--log-level-file=VERBOSE'])
|
||||
|
||||
logfile = os.path.join(backup_dir, 'log', 'pg_probackup.log')
|
||||
with open(logfile, 'r') as f:
|
||||
logfile_content = f.read()
|
||||
|
||||
# Clean after yourself
|
||||
self.del_test_dir(module_name, fname)
|
||||
|
||||
|
||||
# 1. always use parent link when merging (intermediates may be from different chain)
|
||||
# 2. page backup we are merging with may disappear after failed merge,
|
||||
|
Loading…
x
Reference in New Issue
Block a user