1
0
mirror of https://github.com/postgrespro/pg_probackup.git synced 2024-11-24 08:52:38 +02:00
pg_probackup/tests/restore_test.py

1244 lines
44 KiB
Python
Raw Normal View History

import os
2017-06-27 11:43:45 +02:00
import unittest
2017-06-27 07:42:52 +02:00
from .helpers.ptrack_helpers import ProbackupTest, ProbackupException
2016-12-13 19:45:50 +02:00
import subprocess
from datetime import datetime
import sys
import time
2016-12-13 19:45:50 +02:00
2017-07-12 16:28:28 +02:00
module_name = 'restore'
2016-12-13 19:45:50 +02:00
2017-07-12 16:28:28 +02:00
class RestoreTest(ProbackupTest, unittest.TestCase):
2017-05-03 13:14:48 +02:00
2017-06-20 12:57:23 +02:00
# @unittest.skip("skip")
# @unittest.expectedFailure
2017-05-03 13:14:48 +02:00
def test_restore_full_to_latest(self):
"""recovery to latest from full backup"""
fname = self.id().split('.')[3]
node = self.make_simple_node(
base_dir="{0}/{1}/node".format(module_name, fname),
2017-05-03 13:14:48 +02:00
initdb_params=['--data-checksums'],
pg_options={'wal_level': 'replica'}
)
2017-11-03 14:49:31 +02:00
2017-07-12 16:28:28 +02:00
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
2017-06-20 12:57:23 +02:00
self.init_pb(backup_dir)
self.add_instance(backup_dir, 'node', node)
self.set_archiving(backup_dir, 'node', node)
2017-05-03 13:14:48 +02:00
node.start()
2017-06-20 12:57:23 +02:00
2017-05-03 13:14:48 +02:00
node.pgbench_init(scale=2)
pgbench = node.pgbench(
stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
2017-05-03 13:14:48 +02:00
pgbench.wait()
pgbench.stdout.close()
before = node.execute("postgres", "SELECT * FROM pgbench_branches")
2017-06-20 12:57:23 +02:00
backup_id = self.backup_node(backup_dir, 'node', node)
2017-05-03 13:14:48 +02:00
2017-09-28 09:32:06 +02:00
node.stop()
2017-05-03 13:14:48 +02:00
node.cleanup()
# 1 - Test recovery from latest
self.assertIn(
"INFO: Restore of backup {0} completed.".format(backup_id),
self.restore_node(
backup_dir, 'node', node,
options=["-j", "4", "--recovery-target-action=promote"]),
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
repr(self.output), self.cmd))
2017-05-03 13:14:48 +02:00
# 2 - Test that recovery.conf was created
2017-06-20 12:57:23 +02:00
recovery_conf = os.path.join(node.data_dir, "recovery.conf")
self.assertEqual(os.path.isfile(recovery_conf), True)
2017-05-03 13:14:48 +02:00
node.slow_start()
2017-05-03 13:14:48 +02:00
after = node.execute("postgres", "SELECT * FROM pgbench_branches")
self.assertEqual(before, after)
2017-06-27 07:42:52 +02:00
# Clean after yourself
2017-07-12 16:28:28 +02:00
self.del_test_dir(module_name, fname)
2017-05-03 13:14:48 +02:00
2017-06-20 12:57:23 +02:00
# @unittest.skip("skip")
2017-05-03 13:14:48 +02:00
def test_restore_full_page_to_latest(self):
"""recovery to latest from full + page backups"""
fname = self.id().split('.')[3]
node = self.make_simple_node(
base_dir="{0}/{1}/node".format(module_name, fname),
2017-05-03 13:14:48 +02:00
initdb_params=['--data-checksums'],
pg_options={'wal_level': 'replica'}
)
2017-07-12 16:28:28 +02:00
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
2017-06-20 12:57:23 +02:00
self.init_pb(backup_dir)
self.add_instance(backup_dir, 'node', node)
self.set_archiving(backup_dir, 'node', node)
2017-05-03 13:14:48 +02:00
node.start()
2017-06-20 12:57:23 +02:00
2017-05-03 13:14:48 +02:00
node.pgbench_init(scale=2)
2017-06-20 12:57:23 +02:00
self.backup_node(backup_dir, 'node', node)
2017-05-03 13:14:48 +02:00
pgbench = node.pgbench(
stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
2017-05-03 13:14:48 +02:00
pgbench.wait()
pgbench.stdout.close()
backup_id = self.backup_node(
backup_dir, 'node', node, backup_type="page")
2017-05-03 13:14:48 +02:00
before = node.execute("postgres", "SELECT * FROM pgbench_branches")
2017-09-28 09:32:06 +02:00
node.stop()
2017-05-03 13:14:48 +02:00
node.cleanup()
self.assertIn(
"INFO: Restore of backup {0} completed.".format(backup_id),
self.restore_node(
backup_dir, 'node', node,
options=["-j", "4", "--recovery-target-action=promote"]),
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
repr(self.output), self.cmd))
2017-05-03 13:14:48 +02:00
node.slow_start()
2017-05-03 13:14:48 +02:00
after = node.execute("postgres", "SELECT * FROM pgbench_branches")
self.assertEqual(before, after)
2017-06-27 07:42:52 +02:00
# Clean after yourself
2017-07-12 16:28:28 +02:00
self.del_test_dir(module_name, fname)
2017-05-03 13:14:48 +02:00
2017-09-28 09:32:06 +02:00
# @unittest.skip("skip")
2017-06-27 07:42:52 +02:00
def test_restore_to_specific_timeline(self):
2017-05-03 13:14:48 +02:00
"""recovery to target timeline"""
fname = self.id().split('.')[3]
node = self.make_simple_node(
base_dir="{0}/{1}/node".format(module_name, fname),
2017-05-03 13:14:48 +02:00
initdb_params=['--data-checksums'],
pg_options={'wal_level': 'replica'}
)
2017-07-12 16:28:28 +02:00
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
2017-06-20 12:57:23 +02:00
self.init_pb(backup_dir)
self.add_instance(backup_dir, 'node', node)
self.set_archiving(backup_dir, 'node', node)
2017-05-03 13:14:48 +02:00
node.start()
2017-06-20 12:57:23 +02:00
2017-05-03 13:14:48 +02:00
node.pgbench_init(scale=2)
before = node.execute("postgres", "SELECT * FROM pgbench_branches")
2017-06-20 12:57:23 +02:00
backup_id = self.backup_node(backup_dir, 'node', node)
2017-05-03 13:14:48 +02:00
target_tli = int(
node.get_control_data()["Latest checkpoint's TimeLineID"])
2017-09-28 09:32:06 +02:00
node.stop()
2017-05-03 13:14:48 +02:00
node.cleanup()
self.assertIn(
"INFO: Restore of backup {0} completed.".format(backup_id),
self.restore_node(
backup_dir, 'node', node,
options=["-j", "4", "--recovery-target-action=promote"]),
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
repr(self.output), self.cmd))
2017-05-03 13:14:48 +02:00
node.slow_start()
pgbench = node.pgbench(
stdout=subprocess.PIPE, stderr=subprocess.STDOUT,
2017-10-09 14:32:48 +02:00
options=['-T', '10', '-c', '2', '--no-vacuum'])
2017-05-03 13:14:48 +02:00
pgbench.wait()
pgbench.stdout.close()
2017-06-27 07:42:52 +02:00
self.backup_node(backup_dir, 'node', node)
2017-05-03 13:14:48 +02:00
2017-09-28 09:32:06 +02:00
node.stop()
2017-05-03 13:14:48 +02:00
node.cleanup()
2017-06-20 12:57:23 +02:00
# Correct Backup must be choosen for restore
self.assertIn(
"INFO: Restore of backup {0} completed.".format(backup_id),
self.restore_node(
backup_dir, 'node', node,
options=[
"-j", "4", "--timeline={0}".format(target_tli),
"--recovery-target-action=promote"]
),
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
repr(self.output), self.cmd))
2017-05-03 13:14:48 +02:00
recovery_target_timeline = self.get_recovery_conf(
node)["recovery_target_timeline"]
2017-05-03 13:14:48 +02:00
self.assertEqual(int(recovery_target_timeline), target_tli)
node.slow_start()
2017-05-03 13:14:48 +02:00
after = node.execute("postgres", "SELECT * FROM pgbench_branches")
self.assertEqual(before, after)
2017-06-27 07:42:52 +02:00
# Clean after yourself
2017-07-12 16:28:28 +02:00
self.del_test_dir(module_name, fname)
2017-05-03 13:14:48 +02:00
2017-06-20 12:57:23 +02:00
# @unittest.skip("skip")
2017-05-03 13:14:48 +02:00
def test_restore_to_time(self):
2017-06-20 12:57:23 +02:00
"""recovery to target time"""
2017-05-03 13:14:48 +02:00
fname = self.id().split('.')[3]
node = self.make_simple_node(
base_dir="{0}/{1}/node".format(module_name, fname),
2017-05-03 13:14:48 +02:00
initdb_params=['--data-checksums'],
pg_options={'wal_level': 'replica'}
)
2017-07-12 16:28:28 +02:00
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
2017-06-20 12:57:23 +02:00
self.init_pb(backup_dir)
self.add_instance(backup_dir, 'node', node)
self.set_archiving(backup_dir, 'node', node)
2017-10-09 14:32:48 +02:00
node.append_conf("postgresql.auto.conf", "TimeZone = Europe/Moscow")
2017-05-03 13:14:48 +02:00
node.start()
2017-06-20 12:57:23 +02:00
node.pgbench_init(scale=2)
2017-05-03 13:14:48 +02:00
before = node.execute("postgres", "SELECT * FROM pgbench_branches")
2017-06-20 12:57:23 +02:00
backup_id = self.backup_node(backup_dir, 'node', node)
2017-05-03 13:14:48 +02:00
target_time = datetime.now().strftime("%Y-%m-%d %H:%M:%S")
pgbench = node.pgbench(
stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
2017-05-03 13:14:48 +02:00
pgbench.wait()
pgbench.stdout.close()
2017-06-07 16:52:07 +02:00
node.stop()
2017-05-03 13:14:48 +02:00
node.cleanup()
self.assertIn(
"INFO: Restore of backup {0} completed.".format(backup_id),
self.restore_node(
backup_dir, 'node', node,
options=[
"-j", "4", '--time={0}'.format(target_time),
"--recovery-target-action=promote"
]
),
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
repr(self.output), self.cmd))
2017-05-03 13:14:48 +02:00
node.slow_start()
2017-05-03 13:14:48 +02:00
after = node.execute("postgres", "SELECT * FROM pgbench_branches")
self.assertEqual(before, after)
2017-06-27 07:42:52 +02:00
# Clean after yourself
self.del_test_dir(module_name, fname)
2017-05-03 13:14:48 +02:00
2017-06-20 12:57:23 +02:00
# @unittest.skip("skip")
2017-10-09 14:32:48 +02:00
def test_restore_to_xid_inclusive(self):
2017-05-03 13:14:48 +02:00
"""recovery to target xid"""
fname = self.id().split('.')[3]
node = self.make_simple_node(
base_dir="{0}/{1}/node".format(module_name, fname),
2017-05-03 13:14:48 +02:00
initdb_params=['--data-checksums'],
pg_options={'wal_level': 'replica'}
)
2017-07-12 16:28:28 +02:00
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
2017-06-20 12:57:23 +02:00
self.init_pb(backup_dir)
self.add_instance(backup_dir, 'node', node)
self.set_archiving(backup_dir, 'node', node)
2017-05-03 13:14:48 +02:00
node.start()
2017-06-20 12:57:23 +02:00
2017-05-03 13:14:48 +02:00
node.pgbench_init(scale=2)
with node.connect("postgres") as con:
con.execute("CREATE TABLE tbl0005 (a text)")
con.commit()
2017-06-20 12:57:23 +02:00
backup_id = self.backup_node(backup_dir, 'node', node)
2017-05-03 13:14:48 +02:00
pgbench = node.pgbench(
stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
2017-05-03 13:14:48 +02:00
pgbench.wait()
pgbench.stdout.close()
2017-10-09 14:32:48 +02:00
before = node.safe_psql("postgres", "SELECT * FROM pgbench_branches")
2017-05-03 13:14:48 +02:00
with node.connect("postgres") as con:
res = con.execute("INSERT INTO tbl0005 VALUES ('inserted') RETURNING (xmin)")
con.commit()
target_xid = res[0][0]
pgbench = node.pgbench(
stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
2017-05-03 13:14:48 +02:00
pgbench.wait()
pgbench.stdout.close()
2017-09-28 09:32:06 +02:00
node.stop()
2017-05-03 13:14:48 +02:00
node.cleanup()
self.assertIn(
"INFO: Restore of backup {0} completed.".format(backup_id),
self.restore_node(
backup_dir, 'node', node,
options=[
"-j", "4", '--xid={0}'.format(target_xid),
"--recovery-target-action=promote"]
),
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
repr(self.output), self.cmd))
2017-05-03 13:14:48 +02:00
node.slow_start()
2017-10-09 14:32:48 +02:00
after = node.safe_psql("postgres", "SELECT * FROM pgbench_branches")
self.assertEqual(before, after)
self.assertEqual(
len(node.execute("postgres", "SELECT * FROM tbl0005")), 1)
2017-10-09 14:32:48 +02:00
# Clean after yourself
self.del_test_dir(module_name, fname)
# @unittest.skip("skip")
def test_restore_to_xid_not_inclusive(self):
"""recovery with target inclusive false"""
fname = self.id().split('.')[3]
node = self.make_simple_node(
base_dir="{0}/{1}/node".format(module_name, fname),
2017-10-09 14:32:48 +02:00
initdb_params=['--data-checksums'],
pg_options={
'wal_level': 'replica',
'ptrack_enable': 'on',
'max_wal_senders': '2'}
2017-10-09 14:32:48 +02:00
)
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
self.init_pb(backup_dir)
self.add_instance(backup_dir, 'node', node)
self.set_archiving(backup_dir, 'node', node)
node.start()
node.pgbench_init(scale=2)
with node.connect("postgres") as con:
con.execute("CREATE TABLE tbl0005 (a text)")
con.commit()
backup_id = self.backup_node(backup_dir, 'node', node)
pgbench = node.pgbench(
stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
2017-10-09 14:32:48 +02:00
pgbench.wait()
pgbench.stdout.close()
before = node.execute("postgres", "SELECT * FROM pgbench_branches")
with node.connect("postgres") as con:
result = con.execute("INSERT INTO tbl0005 VALUES ('inserted') RETURNING (xmin)")
con.commit()
target_xid = result[0][0]
pgbench = node.pgbench(
stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
2017-10-09 14:32:48 +02:00
pgbench.wait()
pgbench.stdout.close()
node.stop()
node.cleanup()
self.assertIn(
"INFO: Restore of backup {0} completed.".format(backup_id),
self.restore_node(
backup_dir, 'node', node,
options=[
"-j", "4",
'--xid={0}'.format(target_xid),
"--inclusive=false",
"--recovery-target-action=promote"]),
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
repr(self.output), self.cmd))
2017-10-09 14:32:48 +02:00
node.slow_start()
2017-05-03 13:14:48 +02:00
after = node.execute("postgres", "SELECT * FROM pgbench_branches")
self.assertEqual(before, after)
self.assertEqual(
len(node.execute("postgres", "SELECT * FROM tbl0005")), 0)
2017-05-03 13:14:48 +02:00
2017-06-27 07:42:52 +02:00
# Clean after yourself
2017-07-12 16:28:28 +02:00
self.del_test_dir(module_name, fname)
2017-05-03 13:14:48 +02:00
2018-06-20 15:00:44 +02:00
# @unittest.skip("skip")
def test_restore_to_lsn_inclusive(self):
"""recovery to target lsn"""
fname = self.id().split('.')[3]
node = self.make_simple_node(
base_dir="{0}/{1}/node".format(module_name, fname),
initdb_params=['--data-checksums'],
pg_options={'wal_level': 'replica'}
)
if self.get_version(node) < self.version_to_num('10.0'):
self.del_test_dir(module_name, fname)
return
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
self.init_pb(backup_dir)
self.add_instance(backup_dir, 'node', node)
self.set_archiving(backup_dir, 'node', node)
node.start()
node.pgbench_init(scale=2)
with node.connect("postgres") as con:
con.execute("CREATE TABLE tbl0005 (a int)")
con.commit()
backup_id = self.backup_node(backup_dir, 'node', node)
pgbench = node.pgbench(
stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
pgbench.wait()
pgbench.stdout.close()
before = node.safe_psql("postgres", "SELECT * FROM pgbench_branches")
with node.connect("postgres") as con:
con.execute("INSERT INTO tbl0005 VALUES (1)")
con.commit()
res = con.execute("SELECT pg_current_wal_lsn()")
con.commit()
con.execute("INSERT INTO tbl0005 VALUES (2)")
con.commit()
xlogid, xrecoff = res[0][0].split('/')
xrecoff = hex(int(xrecoff, 16) + 1)[2:]
target_lsn = "{0}/{1}".format(xlogid, xrecoff)
pgbench = node.pgbench(
stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
pgbench.wait()
pgbench.stdout.close()
node.stop()
node.cleanup()
self.assertIn(
"INFO: Restore of backup {0} completed.".format(backup_id),
self.restore_node(
backup_dir, 'node', node,
options=[
"-j", "4", '--lsn={0}'.format(target_lsn),
"--recovery-target-action=promote"]
),
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
repr(self.output), self.cmd))
node.slow_start()
after = node.safe_psql("postgres", "SELECT * FROM pgbench_branches")
self.assertEqual(before, after)
self.assertEqual(
len(node.execute("postgres", "SELECT * FROM tbl0005")), 2)
# Clean after yourself
self.del_test_dir(module_name, fname)
# @unittest.skip("skip")
def test_restore_to_lsn_not_inclusive(self):
"""recovery to target lsn"""
fname = self.id().split('.')[3]
node = self.make_simple_node(
base_dir="{0}/{1}/node".format(module_name, fname),
initdb_params=['--data-checksums'],
pg_options={'wal_level': 'replica'}
)
if self.get_version(node) < self.version_to_num('10.0'):
self.del_test_dir(module_name, fname)
return
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
self.init_pb(backup_dir)
self.add_instance(backup_dir, 'node', node)
self.set_archiving(backup_dir, 'node', node)
node.start()
node.pgbench_init(scale=2)
with node.connect("postgres") as con:
con.execute("CREATE TABLE tbl0005 (a int)")
con.commit()
backup_id = self.backup_node(backup_dir, 'node', node)
pgbench = node.pgbench(
stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
pgbench.wait()
pgbench.stdout.close()
before = node.safe_psql("postgres", "SELECT * FROM pgbench_branches")
with node.connect("postgres") as con:
con.execute("INSERT INTO tbl0005 VALUES (1)")
con.commit()
res = con.execute("SELECT pg_current_wal_lsn()")
con.commit()
con.execute("INSERT INTO tbl0005 VALUES (2)")
con.commit()
xlogid, xrecoff = res[0][0].split('/')
xrecoff = hex(int(xrecoff, 16) + 1)[2:]
target_lsn = "{0}/{1}".format(xlogid, xrecoff)
pgbench = node.pgbench(
stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
pgbench.wait()
pgbench.stdout.close()
node.stop()
node.cleanup()
self.assertIn(
"INFO: Restore of backup {0} completed.".format(backup_id),
self.restore_node(
backup_dir, 'node', node,
options=[
"--inclusive=false",
"-j", "4", '--lsn={0}'.format(target_lsn),
"--recovery-target-action=promote"]
),
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
repr(self.output), self.cmd))
node.slow_start()
after = node.safe_psql("postgres", "SELECT * FROM pgbench_branches")
self.assertEqual(before, after)
self.assertEqual(
len(node.execute("postgres", "SELECT * FROM tbl0005")), 1)
# Clean after yourself
self.del_test_dir(module_name, fname)
2017-06-20 12:57:23 +02:00
# @unittest.skip("skip")
2017-06-07 16:52:07 +02:00
def test_restore_full_ptrack_archive(self):
2017-06-20 12:57:23 +02:00
"""recovery to latest from archive full+ptrack backups"""
2017-05-03 13:14:48 +02:00
fname = self.id().split('.')[3]
node = self.make_simple_node(
base_dir="{0}/{1}/node".format(module_name, fname),
2017-05-03 13:14:48 +02:00
initdb_params=['--data-checksums'],
pg_options={'wal_level': 'replica', 'ptrack_enable': 'on'}
)
2017-07-12 16:28:28 +02:00
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
2017-06-20 12:57:23 +02:00
self.init_pb(backup_dir)
self.add_instance(backup_dir, 'node', node)
self.set_archiving(backup_dir, 'node', node)
2017-05-03 13:14:48 +02:00
node.start()
2017-06-20 12:57:23 +02:00
node.pgbench_init(scale=2)
2017-05-03 13:14:48 +02:00
2017-06-20 12:57:23 +02:00
self.backup_node(backup_dir, 'node', node)
2017-05-03 13:14:48 +02:00
pgbench = node.pgbench(
stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
2017-05-03 13:14:48 +02:00
pgbench.wait()
pgbench.stdout.close()
backup_id = self.backup_node(
backup_dir, 'node', node, backup_type="ptrack")
2017-05-03 13:14:48 +02:00
before = node.execute("postgres", "SELECT * FROM pgbench_branches")
2017-09-28 09:32:06 +02:00
node.stop()
2017-05-03 13:14:48 +02:00
node.cleanup()
self.assertIn(
"INFO: Restore of backup {0} completed.".format(backup_id),
self.restore_node(
backup_dir, 'node', node,
options=[
"-j", "4", "--recovery-target-action=promote"]),
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
repr(self.output), self.cmd))
2017-05-03 13:14:48 +02:00
node.slow_start()
2017-05-03 13:14:48 +02:00
after = node.execute("postgres", "SELECT * FROM pgbench_branches")
self.assertEqual(before, after)
2017-06-27 07:42:52 +02:00
# Clean after yourself
2017-07-12 16:28:28 +02:00
self.del_test_dir(module_name, fname)
2017-05-03 13:14:48 +02:00
2017-06-20 12:57:23 +02:00
# @unittest.skip("skip")
def test_restore_ptrack(self):
"""recovery to latest from archive full+ptrack+ptrack backups"""
2017-05-03 13:14:48 +02:00
fname = self.id().split('.')[3]
node = self.make_simple_node(
base_dir="{0}/{1}/node".format(module_name, fname),
2017-05-03 13:14:48 +02:00
initdb_params=['--data-checksums'],
pg_options={'wal_level': 'replica', 'ptrack_enable': 'on'}
)
2017-07-12 16:28:28 +02:00
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
2017-06-20 12:57:23 +02:00
self.init_pb(backup_dir)
self.add_instance(backup_dir, 'node', node)
self.set_archiving(backup_dir, 'node', node)
2017-05-03 13:14:48 +02:00
node.start()
2017-06-20 12:57:23 +02:00
node.pgbench_init(scale=2)
2017-05-03 13:14:48 +02:00
2017-06-20 12:57:23 +02:00
self.backup_node(backup_dir, 'node', node)
2017-05-03 13:14:48 +02:00
pgbench = node.pgbench(
stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
2017-05-03 13:14:48 +02:00
pgbench.wait()
pgbench.stdout.close()
2017-06-20 12:57:23 +02:00
self.backup_node(backup_dir, 'node', node, backup_type="ptrack")
2017-05-03 13:14:48 +02:00
pgbench = node.pgbench(
stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
2017-05-03 13:14:48 +02:00
pgbench.wait()
pgbench.stdout.close()
backup_id = self.backup_node(
backup_dir, 'node', node, backup_type="ptrack")
2017-05-03 13:14:48 +02:00
before = node.execute("postgres", "SELECT * FROM pgbench_branches")
2017-09-28 09:32:06 +02:00
node.stop()
2017-05-03 13:14:48 +02:00
node.cleanup()
self.assertIn(
"INFO: Restore of backup {0} completed.".format(backup_id),
self.restore_node(
backup_dir, 'node', node,
options=[
"-j", "4", "--recovery-target-action=promote"]),
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
repr(self.output), self.cmd))
2017-05-03 13:14:48 +02:00
node.slow_start()
2017-05-03 13:14:48 +02:00
after = node.execute("postgres", "SELECT * FROM pgbench_branches")
self.assertEqual(before, after)
2017-06-27 07:42:52 +02:00
# Clean after yourself
2017-07-12 16:28:28 +02:00
self.del_test_dir(module_name, fname)
2017-05-03 13:14:48 +02:00
2017-06-20 12:57:23 +02:00
# @unittest.skip("skip")
2017-05-03 13:14:48 +02:00
def test_restore_full_ptrack_stream(self):
"""recovery in stream mode to latest from full + ptrack backups"""
fname = self.id().split('.')[3]
node = self.make_simple_node(
base_dir="{0}/{1}/node".format(module_name, fname),
2017-05-03 13:14:48 +02:00
set_replication=True,
initdb_params=['--data-checksums'],
pg_options={
'wal_level': 'replica',
'ptrack_enable': 'on',
'max_wal_senders': '2'}
2017-05-03 13:14:48 +02:00
)
2017-07-12 16:28:28 +02:00
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
2017-06-20 12:57:23 +02:00
self.init_pb(backup_dir)
self.add_instance(backup_dir, 'node', node)
self.set_archiving(backup_dir, 'node', node)
2017-05-03 13:14:48 +02:00
node.start()
2017-06-20 12:57:23 +02:00
2017-05-03 13:14:48 +02:00
node.pgbench_init(scale=2)
2017-06-20 12:57:23 +02:00
self.backup_node(backup_dir, 'node', node, options=["--stream"])
2017-05-03 13:14:48 +02:00
pgbench = node.pgbench(
stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
2017-05-03 13:14:48 +02:00
pgbench.wait()
pgbench.stdout.close()
backup_id = self.backup_node(
backup_dir, 'node', node,
backup_type="ptrack", options=["--stream"])
2017-05-03 13:14:48 +02:00
before = node.execute("postgres", "SELECT * FROM pgbench_branches")
2017-06-07 16:52:07 +02:00
node.stop()
2017-05-03 13:14:48 +02:00
node.cleanup()
self.assertIn(
"INFO: Restore of backup {0} completed.".format(backup_id),
self.restore_node(
backup_dir, 'node', node,
options=["-j", "4", "--recovery-target-action=promote"]),
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
repr(self.output), self.cmd))
2017-05-03 13:14:48 +02:00
node.slow_start()
2017-05-03 13:14:48 +02:00
after = node.execute("postgres", "SELECT * FROM pgbench_branches")
self.assertEqual(before, after)
2017-06-27 07:42:52 +02:00
# Clean after yourself
2017-07-12 16:28:28 +02:00
self.del_test_dir(module_name, fname)
2017-05-03 13:14:48 +02:00
2017-06-20 12:57:23 +02:00
# @unittest.skip("skip")
2017-05-03 13:14:48 +02:00
def test_restore_full_ptrack_under_load(self):
"""
recovery to latest from full + ptrack backups
with loads when ptrack backup do
"""
2017-05-03 13:14:48 +02:00
fname = self.id().split('.')[3]
node = self.make_simple_node(
base_dir="{0}/{1}/node".format(module_name, fname),
2017-05-05 15:21:49 +02:00
set_replication=True,
2017-05-03 13:14:48 +02:00
initdb_params=['--data-checksums'],
pg_options={
'wal_level': 'replica',
'ptrack_enable': 'on',
'max_wal_senders': '2'}
2017-05-03 13:14:48 +02:00
)
2017-07-12 16:28:28 +02:00
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
2017-06-20 12:57:23 +02:00
self.init_pb(backup_dir)
self.add_instance(backup_dir, 'node', node)
self.set_archiving(backup_dir, 'node', node)
2017-05-03 13:14:48 +02:00
node.start()
2017-06-20 12:57:23 +02:00
2017-05-03 13:14:48 +02:00
node.pgbench_init(scale=2)
2017-06-20 12:57:23 +02:00
self.backup_node(backup_dir, 'node', node)
2017-05-03 13:14:48 +02:00
pgbench = node.pgbench(
stdout=subprocess.PIPE,
stderr=subprocess.STDOUT,
options=["-c", "4", "-T", "8"]
)
backup_id = self.backup_node(
backup_dir, 'node', node,
backup_type="ptrack", options=["--stream"])
2017-05-03 13:14:48 +02:00
pgbench.wait()
pgbench.stdout.close()
bbalance = node.execute(
"postgres", "SELECT sum(bbalance) FROM pgbench_branches")
delta = node.execute(
"postgres", "SELECT sum(delta) FROM pgbench_history")
2017-05-03 13:14:48 +02:00
self.assertEqual(bbalance, delta)
2017-09-28 09:32:06 +02:00
node.stop()
2017-05-03 13:14:48 +02:00
node.cleanup()
self.assertIn(
"INFO: Restore of backup {0} completed.".format(backup_id),
self.restore_node(
backup_dir, 'node', node,
options=["-j", "4", "--recovery-target-action=promote"]),
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
repr(self.output), self.cmd))
2017-05-03 13:14:48 +02:00
node.slow_start()
bbalance = node.execute(
"postgres", "SELECT sum(bbalance) FROM pgbench_branches")
delta = node.execute(
"postgres", "SELECT sum(delta) FROM pgbench_history")
2017-05-03 13:14:48 +02:00
self.assertEqual(bbalance, delta)
2017-06-27 07:42:52 +02:00
# Clean after yourself
2017-07-12 16:28:28 +02:00
self.del_test_dir(module_name, fname)
2017-05-03 13:14:48 +02:00
2017-06-20 12:57:23 +02:00
# @unittest.skip("skip")
2017-05-03 13:14:48 +02:00
def test_restore_full_under_load_ptrack(self):
"""
recovery to latest from full + page backups
with loads when full backup do
"""
2017-05-03 13:14:48 +02:00
fname = self.id().split('.')[3]
node = self.make_simple_node(
base_dir="{0}/{1}/node".format(module_name, fname),
2017-05-05 15:21:49 +02:00
set_replication=True,
2017-05-03 13:14:48 +02:00
initdb_params=['--data-checksums'],
pg_options={
'wal_level': 'replica',
'ptrack_enable': 'on',
'max_wal_senders': '2'}
2017-05-03 13:14:48 +02:00
)
2017-07-12 16:28:28 +02:00
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
2017-06-20 12:57:23 +02:00
self.init_pb(backup_dir)
self.add_instance(backup_dir, 'node', node)
self.set_archiving(backup_dir, 'node', node)
2017-05-03 13:14:48 +02:00
node.start()
2017-06-20 12:57:23 +02:00
# wal_segment_size = self.guc_wal_segment_size(node)
node.pgbench_init(scale=2)
2017-05-03 13:14:48 +02:00
pgbench = node.pgbench(
stdout=subprocess.PIPE,
stderr=subprocess.STDOUT,
options=["-c", "4", "-T", "8"]
)
2017-06-20 12:57:23 +02:00
self.backup_node(backup_dir, 'node', node)
2017-05-03 13:14:48 +02:00
pgbench.wait()
pgbench.stdout.close()
backup_id = self.backup_node(
backup_dir, 'node', node,
backup_type="ptrack", options=["--stream"])
2017-05-03 13:14:48 +02:00
bbalance = node.execute(
"postgres", "SELECT sum(bbalance) FROM pgbench_branches")
delta = node.execute(
"postgres", "SELECT sum(delta) FROM pgbench_history")
2017-05-03 13:14:48 +02:00
self.assertEqual(bbalance, delta)
2017-09-28 09:32:06 +02:00
node.stop()
2017-05-03 13:14:48 +02:00
node.cleanup()
# self.wrong_wal_clean(node, wal_segment_size)
2017-05-03 13:14:48 +02:00
self.assertIn(
"INFO: Restore of backup {0} completed.".format(backup_id),
self.restore_node(
backup_dir, 'node', node,
options=["-j", "4", "--recovery-target-action=promote"]),
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
repr(self.output), self.cmd))
node.slow_start()
bbalance = node.execute(
"postgres", "SELECT sum(bbalance) FROM pgbench_branches")
delta = node.execute(
"postgres", "SELECT sum(delta) FROM pgbench_history")
2017-05-03 13:14:48 +02:00
self.assertEqual(bbalance, delta)
2017-06-27 07:42:52 +02:00
# Clean after yourself
2017-07-12 16:28:28 +02:00
self.del_test_dir(module_name, fname)
2017-05-03 13:14:48 +02:00
2017-06-20 12:57:23 +02:00
# @unittest.skip("skip")
2017-05-03 13:14:48 +02:00
def test_restore_with_tablespace_mapping_1(self):
"""recovery using tablespace-mapping option"""
fname = self.id().split('.')[3]
node = self.make_simple_node(
base_dir="{0}/{1}/node".format(module_name, fname),
2017-05-03 13:14:48 +02:00
initdb_params=['--data-checksums'],
pg_options={
'wal_level': 'replica',
'ptrack_enable': 'on',
'max_wal_senders': '2'}
2017-05-03 13:14:48 +02:00
)
2017-07-12 16:28:28 +02:00
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
2017-06-20 12:57:23 +02:00
self.init_pb(backup_dir)
self.add_instance(backup_dir, 'node', node)
self.set_archiving(backup_dir, 'node', node)
2017-05-03 13:14:48 +02:00
node.start()
# Create tablespace
2017-06-20 12:57:23 +02:00
tblspc_path = os.path.join(node.base_dir, "tblspc")
2017-05-03 13:14:48 +02:00
os.makedirs(tblspc_path)
with node.connect("postgres") as con:
con.connection.autocommit = True
con.execute("CREATE TABLESPACE tblspc LOCATION '%s'" % tblspc_path)
con.connection.autocommit = False
con.execute("CREATE TABLE test (id int) TABLESPACE tblspc")
con.execute("INSERT INTO test VALUES (1)")
con.commit()
2017-06-20 12:57:23 +02:00
backup_id = self.backup_node(backup_dir, 'node', node)
2018-06-02 19:35:37 +02:00
self.assertEqual(self.show_pb(backup_dir, 'node')[0]['status'], "OK")
2017-05-03 13:14:48 +02:00
# 1 - Try to restore to existing directory
node.stop()
try:
2017-06-20 12:57:23 +02:00
self.restore_node(backup_dir, 'node', node)
# we should die here because exception is what we expect to happen
self.assertEqual(
1, 0,
"Expecting Error because restore destionation is not empty.\n "
"Output: {0} \n CMD: {1}".format(
repr(self.output), self.cmd))
2017-06-27 07:42:52 +02:00
except ProbackupException as e:
self.assertEqual(
e.message,
'ERROR: restore destination is not empty: "{0}"\n'.format(
node.data_dir),
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
repr(e.message), self.cmd))
2017-05-03 13:14:48 +02:00
# 2 - Try to restore to existing tablespace directory
2017-06-20 12:57:23 +02:00
node.cleanup()
2017-05-03 13:14:48 +02:00
try:
2017-06-20 12:57:23 +02:00
self.restore_node(backup_dir, 'node', node)
# we should die here because exception is what we expect to happen
self.assertEqual(
1, 0,
"Expecting Error because restore tablespace destination is "
"not empty.\n Output: {0} \n CMD: {1}".format(
repr(self.output), self.cmd))
2017-06-27 07:42:52 +02:00
except ProbackupException as e:
self.assertEqual(
e.message,
'ERROR: restore tablespace destination '
'is not empty: "{0}"\n'.format(tblspc_path),
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
repr(e.message), self.cmd))
2017-05-03 13:14:48 +02:00
# 3 - Restore using tablespace-mapping
2017-06-20 12:57:23 +02:00
tblspc_path_new = os.path.join(node.base_dir, "tblspc_new")
self.assertIn(
"INFO: Restore of backup {0} completed.".format(backup_id),
self.restore_node(
backup_dir, 'node', node,
options=[
"-T", "%s=%s" % (tblspc_path, tblspc_path_new),
"--recovery-target-action=promote"]
),
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
repr(self.output), self.cmd))
2017-05-03 13:14:48 +02:00
node.slow_start()
2017-06-27 07:42:52 +02:00
result = node.execute("postgres", "SELECT id FROM test")
self.assertEqual(result[0][0], 1)
2017-05-03 13:14:48 +02:00
# 4 - Restore using tablespace-mapping using page backup
2017-06-20 12:57:23 +02:00
self.backup_node(backup_dir, 'node', node)
2017-05-03 13:14:48 +02:00
with node.connect("postgres") as con:
con.execute("INSERT INTO test VALUES (2)")
con.commit()
backup_id = self.backup_node(
backup_dir, 'node', node, backup_type="page")
2017-05-03 13:14:48 +02:00
2017-06-20 12:57:23 +02:00
show_pb = self.show_pb(backup_dir, 'node')
2018-06-02 19:35:37 +02:00
self.assertEqual(show_pb[1]['status'], "OK")
self.assertEqual(show_pb[2]['status'], "OK")
2017-05-03 13:14:48 +02:00
node.stop()
node.cleanup()
2017-06-20 12:57:23 +02:00
tblspc_path_page = os.path.join(node.base_dir, "tblspc_page")
self.assertIn(
"INFO: Restore of backup {0} completed.".format(backup_id),
self.restore_node(
backup_dir, 'node', node,
options=[
"-T", "%s=%s" % (tblspc_path_new, tblspc_path_page),
"--recovery-target-action=promote"]),
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
repr(self.output), self.cmd))
2017-05-03 13:14:48 +02:00
node.slow_start()
2017-06-27 07:42:52 +02:00
result = node.execute("postgres", "SELECT id FROM test OFFSET 1")
self.assertEqual(result[0][0], 2)
2017-05-03 13:14:48 +02:00
2017-06-27 07:42:52 +02:00
# Clean after yourself
2017-07-12 16:28:28 +02:00
self.del_test_dir(module_name, fname)
2017-05-03 13:14:48 +02:00
2017-06-20 12:57:23 +02:00
# @unittest.skip("skip")
2017-05-03 13:14:48 +02:00
def test_restore_with_tablespace_mapping_2(self):
"""recovery using tablespace-mapping option and page backup"""
fname = self.id().split('.')[3]
node = self.make_simple_node(
base_dir="{0}/{1}/node".format(module_name, fname),
2017-05-03 13:14:48 +02:00
initdb_params=['--data-checksums'],
pg_options={'wal_level': 'replica'}
)
2017-07-12 16:28:28 +02:00
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
2017-06-20 12:57:23 +02:00
self.init_pb(backup_dir)
self.add_instance(backup_dir, 'node', node)
self.set_archiving(backup_dir, 'node', node)
2017-05-03 13:14:48 +02:00
node.start()
# Full backup
2017-06-20 12:57:23 +02:00
self.backup_node(backup_dir, 'node', node)
2018-06-02 19:35:37 +02:00
self.assertEqual(self.show_pb(backup_dir, 'node')[0]['status'], "OK")
2017-05-03 13:14:48 +02:00
# Create tablespace
2017-06-20 12:57:23 +02:00
tblspc_path = os.path.join(node.base_dir, "tblspc")
2017-05-03 13:14:48 +02:00
os.makedirs(tblspc_path)
with node.connect("postgres") as con:
con.connection.autocommit = True
con.execute("CREATE TABLESPACE tblspc LOCATION '%s'" % tblspc_path)
con.connection.autocommit = False
con.execute(
"CREATE TABLE tbl AS SELECT * "
"FROM generate_series(0,3) AS integer")
2017-05-03 13:14:48 +02:00
con.commit()
# First page backup
2017-06-20 12:57:23 +02:00
self.backup_node(backup_dir, 'node', node, backup_type="page")
2018-06-02 19:35:37 +02:00
self.assertEqual(self.show_pb(backup_dir, 'node')[1]['status'], "OK")
self.assertEqual(
self.show_pb(backup_dir, 'node')[1]['backup-mode'], "PAGE")
2017-05-03 13:14:48 +02:00
# Create tablespace table
with node.connect("postgres") as con:
con.connection.autocommit = True
con.execute("CHECKPOINT")
con.connection.autocommit = False
con.execute("CREATE TABLE tbl1 (a int) TABLESPACE tblspc")
con.execute(
"INSERT INTO tbl1 SELECT * "
"FROM generate_series(0,3) AS integer")
2017-05-03 13:14:48 +02:00
con.commit()
# Second page backup
backup_id = self.backup_node(
backup_dir, 'node', node, backup_type="page")
2018-06-02 19:35:37 +02:00
self.assertEqual(self.show_pb(backup_dir, 'node')[2]['status'], "OK")
self.assertEqual(
self.show_pb(backup_dir, 'node')[2]['backup-mode'], "PAGE")
2017-05-03 13:14:48 +02:00
node.stop()
node.cleanup()
2017-06-20 12:57:23 +02:00
tblspc_path_new = os.path.join(node.base_dir, "tblspc_new")
self.assertIn(
"INFO: Restore of backup {0} completed.".format(backup_id),
self.restore_node(
backup_dir, 'node', node,
options=[
"-T", "%s=%s" % (tblspc_path, tblspc_path_new),
"--recovery-target-action=promote"]),
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
repr(self.output), self.cmd))
node.slow_start()
2017-05-03 13:14:48 +02:00
count = node.execute("postgres", "SELECT count(*) FROM tbl")
self.assertEqual(count[0][0], 4)
count = node.execute("postgres", "SELECT count(*) FROM tbl1")
self.assertEqual(count[0][0], 4)
2017-05-18 12:01:30 +02:00
2017-06-27 07:42:52 +02:00
# Clean after yourself
2017-07-12 16:28:28 +02:00
self.del_test_dir(module_name, fname)
2017-06-27 07:42:52 +02:00
2017-09-28 09:32:06 +02:00
# @unittest.skip("skip")
2017-05-18 12:01:30 +02:00
def test_archive_node_backup_stream_restore_to_recovery_time(self):
"""
make node with archiving, make stream backup,
make PITR to Recovery Time
"""
2017-05-18 12:01:30 +02:00
fname = self.id().split('.')[3]
node = self.make_simple_node(
base_dir="{0}/{1}/node".format(module_name, fname),
2017-05-18 12:01:30 +02:00
set_replication=True,
initdb_params=['--data-checksums'],
pg_options={'wal_level': 'replica', 'max_wal_senders': '2'}
)
2017-07-12 16:28:28 +02:00
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
2017-06-20 12:57:23 +02:00
self.init_pb(backup_dir)
self.add_instance(backup_dir, 'node', node)
2017-06-27 07:42:52 +02:00
self.set_archiving(backup_dir, 'node', node)
2017-05-18 12:01:30 +02:00
node.start()
backup_id = self.backup_node(
backup_dir, 'node', node, options=["--stream"])
2017-07-12 16:28:28 +02:00
node.safe_psql("postgres", "create table t_heap(a int)")
node.safe_psql("postgres", "select pg_switch_xlog()")
2017-06-07 16:52:07 +02:00
node.stop()
2017-05-18 12:01:30 +02:00
node.cleanup()
recovery_time = self.show_pb(
backup_dir, 'node', backup_id)['recovery-time']
2017-06-07 16:52:07 +02:00
self.assertIn(
"INFO: Restore of backup {0} completed.".format(backup_id),
self.restore_node(
backup_dir, 'node', node,
options=[
"-j", "4", '--time={0}'.format(recovery_time),
"--recovery-target-action=promote"
]
),
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
repr(self.output), self.cmd))
node.slow_start()
2017-06-07 16:52:07 +02:00
2017-06-27 07:42:52 +02:00
result = node.psql("postgres", 'select * from t_heap')
2017-07-12 16:28:28 +02:00
self.assertTrue('does not exist' in result[2].decode("utf-8"))
2017-06-27 07:42:52 +02:00
# Clean after yourself
2017-07-12 16:28:28 +02:00
self.del_test_dir(module_name, fname)
2017-06-27 07:42:52 +02:00
2017-09-28 09:32:06 +02:00
# @unittest.skip("skip")
2017-10-09 14:32:48 +02:00
# @unittest.expectedFailure
2017-06-27 07:42:52 +02:00
def test_archive_node_backup_stream_restore_to_recovery_time(self):
"""
make node with archiving, make stream backup,
make PITR to Recovery Time
"""
2017-06-27 07:42:52 +02:00
fname = self.id().split('.')[3]
node = self.make_simple_node(
base_dir="{0}/{1}/node".format(module_name, fname),
2017-06-27 07:42:52 +02:00
set_replication=True,
initdb_params=['--data-checksums'],
pg_options={'wal_level': 'replica', 'max_wal_senders': '2'}
)
2017-07-12 16:28:28 +02:00
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
2017-06-27 07:42:52 +02:00
self.init_pb(backup_dir)
self.add_instance(backup_dir, 'node', node)
self.set_archiving(backup_dir, 'node', node)
node.start()
backup_id = self.backup_node(
backup_dir, 'node', node, options=["--stream"])
2017-07-12 16:28:28 +02:00
node.safe_psql("postgres", "create table t_heap(a int)")
2017-05-18 12:01:30 +02:00
node.stop()
2017-06-27 07:42:52 +02:00
node.cleanup()
recovery_time = self.show_pb(
backup_dir, 'node', backup_id)['recovery-time']
2017-06-27 07:42:52 +02:00
self.assertIn(
"INFO: Restore of backup {0} completed.".format(backup_id),
self.restore_node(
backup_dir, 'node', node,
options=[
"-j", "4", '--time={0}'.format(recovery_time),
"--recovery-target-action=promote"
]
),
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
repr(self.output), self.cmd))
2017-06-27 07:42:52 +02:00
node.slow_start()
2017-06-27 07:42:52 +02:00
result = node.psql("postgres", 'select * from t_heap')
2017-07-12 16:28:28 +02:00
self.assertTrue('does not exist' in result[2].decode("utf-8"))
2017-06-27 07:42:52 +02:00
# Clean after yourself
2017-07-12 16:28:28 +02:00
self.del_test_dir(module_name, fname)
2017-05-18 12:01:30 +02:00
2017-06-20 12:57:23 +02:00
# @unittest.skip("skip")
2017-10-09 14:32:48 +02:00
# @unittest.expectedFailure
2017-06-07 16:52:07 +02:00
def test_archive_node_backup_stream_pitr(self):
"""
make node with archiving, make stream backup,
create table t_heap, make pitr to Recovery Time,
check that t_heap do not exists
"""
2017-06-07 16:52:07 +02:00
fname = self.id().split('.')[3]
node = self.make_simple_node(
base_dir="{0}/{1}/node".format(module_name, fname),
2017-06-07 16:52:07 +02:00
set_replication=True,
initdb_params=['--data-checksums'],
pg_options={'wal_level': 'replica', 'max_wal_senders': '2'}
)
2017-07-12 16:28:28 +02:00
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
2017-06-20 12:57:23 +02:00
self.init_pb(backup_dir)
self.add_instance(backup_dir, 'node', node)
self.set_archiving(backup_dir, 'node', node)
2017-06-07 16:52:07 +02:00
node.start()
backup_id = self.backup_node(
backup_dir, 'node', node, options=["--stream"])
2017-07-12 16:28:28 +02:00
node.safe_psql("postgres", "create table t_heap(a int)")
2017-06-07 16:52:07 +02:00
node.cleanup()
recovery_time = self.show_pb(
backup_dir, 'node', backup_id)['recovery-time']
2017-06-07 16:52:07 +02:00
self.assertIn(
"INFO: Restore of backup {0} completed.".format(backup_id),
self.restore_node(
backup_dir, 'node', node,
options=[
"-j", "4", '--time={0}'.format(recovery_time),
"--recovery-target-action=promote"
]
),
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
repr(self.output), self.cmd))
2017-06-07 16:52:07 +02:00
node.slow_start()
2017-06-07 16:52:07 +02:00
2017-06-27 07:42:52 +02:00
result = node.psql("postgres", 'select * from t_heap')
self.assertEqual(True, 'does not exist' in result[2].decode("utf-8"))
# Clean after yourself
2017-07-12 16:28:28 +02:00
self.del_test_dir(module_name, fname)
2017-06-07 16:52:07 +02:00
2017-06-20 12:57:23 +02:00
# @unittest.skip("skip")
2017-10-09 14:32:48 +02:00
# @unittest.expectedFailure
2017-06-20 12:57:23 +02:00
def test_archive_node_backup_archive_pitr_2(self):
"""
make node with archiving, make archive backup,
create table t_heap, make pitr to Recovery Time,
check that t_heap do not exists
"""
2017-05-18 12:01:30 +02:00
fname = self.id().split('.')[3]
node = self.make_simple_node(
base_dir="{0}/{1}/node".format(module_name, fname),
2017-05-18 12:01:30 +02:00
initdb_params=['--data-checksums'],
2017-06-20 12:57:23 +02:00
pg_options={'wal_level': 'replica'}
2017-05-18 12:01:30 +02:00
)
2017-07-12 16:28:28 +02:00
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
2017-06-20 12:57:23 +02:00
self.init_pb(backup_dir)
self.add_instance(backup_dir, 'node', node)
self.set_archiving(backup_dir, 'node', node)
2017-05-18 12:01:30 +02:00
node.start()
2017-06-20 12:57:23 +02:00
backup_id = self.backup_node(backup_dir, 'node', node)
2018-03-03 10:42:10 +02:00
if self.paranoia:
pgdata = self.pgdata_content(node.data_dir)
2017-07-12 16:28:28 +02:00
node.safe_psql("postgres", "create table t_heap(a int)")
2017-09-28 09:32:06 +02:00
node.stop()
2017-05-18 12:01:30 +02:00
node.cleanup()
recovery_time = self.show_pb(
backup_dir, 'node', backup_id)['recovery-time']
2017-06-07 16:52:07 +02:00
self.assertIn(
"INFO: Restore of backup {0} completed.".format(backup_id),
self.restore_node(
backup_dir, 'node', node,
options=[
"-j", "4", '--time={0}'.format(recovery_time),
"--recovery-target-action=promote"
]
),
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
repr(self.output), self.cmd))
2018-03-03 10:42:10 +02:00
if self.paranoia:
pgdata_restored = self.pgdata_content(node.data_dir)
self.compare_pgdata(pgdata, pgdata_restored)
node.slow_start()
2017-05-03 13:14:48 +02:00
2017-06-27 07:42:52 +02:00
result = node.psql("postgres", 'select * from t_heap')
self.assertTrue('does not exist' in result[2].decode("utf-8"))
# Clean after yourself
2017-07-12 16:28:28 +02:00
self.del_test_dir(module_name, fname)
# @unittest.skip("skip")
# @unittest.expectedFailure
def test_archive_restore_to_restore_point(self):
"""
make node with archiving, make archive backup,
create table t_heap, make pitr to Recovery Time,
check that t_heap do not exists
"""
fname = self.id().split('.')[3]
node = self.make_simple_node(
base_dir="{0}/{1}/node".format(module_name, fname),
initdb_params=['--data-checksums'],
pg_options={'wal_level': 'replica'}
)
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
self.init_pb(backup_dir)
self.add_instance(backup_dir, 'node', node)
self.set_archiving(backup_dir, 'node', node)
node.start()
self.backup_node(backup_dir, 'node', node)
node.safe_psql(
"postgres",
"create table t_heap as select generate_series(0,10000)")
result = node.safe_psql(
"postgres",
"select * from t_heap")
node.safe_psql(
"postgres", "select pg_create_restore_point('savepoint')")
node.safe_psql(
"postgres",
"create table t_heap_1 as select generate_series(0,10000)")
node.cleanup()
self.restore_node(
backup_dir, 'node', node,
options=[
"--recovery-target-name=savepoint",
"--recovery-target-action=promote"])
node.slow_start()
result_new = node.safe_psql("postgres", "select * from t_heap")
res = node.psql("postgres", "select * from t_heap_1")
self.assertEqual(
res[0], 1,
"Table t_heap_1 should not exist in restored instance")
self.assertEqual(result, result_new)
# Clean after yourself
self.del_test_dir(module_name, fname)