1
0
mirror of https://github.com/postgrespro/pg_probackup.git synced 2024-12-01 09:51:43 +02:00
pg_probackup/tests/show.py
2021-08-03 22:24:33 +03:00

553 lines
19 KiB
Python

import os
import unittest
from .helpers.ptrack_helpers import ProbackupTest, ProbackupException
module_name = 'show'
class ShowTest(ProbackupTest, unittest.TestCase):
# @unittest.skip("skip")
# @unittest.expectedFailure
def test_show_1(self):
"""Status DONE and OK"""
fname = self.id().split('.')[3]
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
node = self.make_simple_node(
base_dir=os.path.join(module_name, fname, 'node'),
initdb_params=['--data-checksums'])
self.init_pb(backup_dir)
self.add_instance(backup_dir, 'node', node)
self.set_archiving(backup_dir, 'node', node)
node.slow_start()
self.assertEqual(
self.backup_node(
backup_dir, 'node', node,
options=["--log-level-console=off"]),
None
)
self.assertIn("OK", self.show_pb(backup_dir, 'node', as_text=True))
# Clean after yourself
self.del_test_dir(module_name, fname)
# @unittest.skip("skip")
# @unittest.expectedFailure
def test_show_json(self):
"""Status DONE and OK"""
fname = self.id().split('.')[3]
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
node = self.make_simple_node(
base_dir=os.path.join(module_name, fname, 'node'),
initdb_params=['--data-checksums'])
self.init_pb(backup_dir)
self.add_instance(backup_dir, 'node', node)
self.set_archiving(backup_dir, 'node', node)
node.slow_start()
self.assertEqual(
self.backup_node(
backup_dir, 'node', node,
options=["--log-level-console=off"]),
None
)
self.backup_node(backup_dir, 'node', node)
self.assertIn("OK", self.show_pb(backup_dir, 'node', as_text=True))
# Clean after yourself
self.del_test_dir(module_name, fname)
# @unittest.skip("skip")
def test_corrupt_2(self):
"""Status CORRUPT"""
fname = self.id().split('.')[3]
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
node = self.make_simple_node(
base_dir=os.path.join(module_name, fname, 'node'),
initdb_params=['--data-checksums'])
self.init_pb(backup_dir)
self.add_instance(backup_dir, 'node', node)
self.set_archiving(backup_dir, 'node', node)
node.slow_start()
backup_id = self.backup_node(backup_dir, 'node', node)
# delete file which belong to backup
file = os.path.join(
backup_dir, "backups", "node",
backup_id, "database", "postgresql.conf")
os.remove(file)
try:
self.validate_pb(backup_dir, 'node', backup_id)
# we should die here because exception is what we expect to happen
self.assertEqual(
1, 0,
"Expecting Error because backup corrupted."
" Output: {0} \n CMD: {1}".format(
repr(self.output), self.cmd
)
)
except ProbackupException as e:
self.assertIn(
'data files are corrupted',
e.message,
'\n Unexpected Error Message: {0}\n'
' CMD: {1}'.format(repr(e.message), self.cmd)
)
self.assertIn("CORRUPT", self.show_pb(backup_dir, as_text=True))
# Clean after yourself
self.del_test_dir(module_name, fname)
# @unittest.skip("skip")
def test_no_control_file(self):
"""backup.control doesn't exist"""
fname = self.id().split('.')[3]
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
node = self.make_simple_node(
base_dir=os.path.join(module_name, fname, 'node'),
initdb_params=['--data-checksums'])
self.init_pb(backup_dir)
self.add_instance(backup_dir, 'node', node)
self.set_archiving(backup_dir, 'node', node)
node.slow_start()
backup_id = self.backup_node(backup_dir, 'node', node)
# delete backup.control file
file = os.path.join(
backup_dir, "backups", "node",
backup_id, "backup.control")
os.remove(file)
output = self.show_pb(backup_dir, 'node', as_text=True, as_json=False)
self.assertIn(
'Control file',
output)
self.assertIn(
'doesn\'t exist',
output)
# Clean after yourself
self.del_test_dir(module_name, fname)
# @unittest.skip("skip")
def test_empty_control_file(self):
"""backup.control is empty"""
fname = self.id().split('.')[3]
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
node = self.make_simple_node(
base_dir=os.path.join(module_name, fname, 'node'),
initdb_params=['--data-checksums'])
self.init_pb(backup_dir)
self.add_instance(backup_dir, 'node', node)
self.set_archiving(backup_dir, 'node', node)
node.slow_start()
backup_id = self.backup_node(backup_dir, 'node', node)
# truncate backup.control file
file = os.path.join(
backup_dir, "backups", "node",
backup_id, "backup.control")
fd = open(file, 'w')
fd.close()
output = self.show_pb(backup_dir, 'node', as_text=True, as_json=False)
self.assertIn(
'Control file',
output)
self.assertIn(
'is empty',
output)
# Clean after yourself
self.del_test_dir(module_name, fname)
# @unittest.skip("skip")
# @unittest.expectedFailure
def test_corrupt_control_file(self):
"""backup.control contains invalid option"""
fname = self.id().split('.')[3]
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
node = self.make_simple_node(
base_dir=os.path.join(module_name, fname, 'node'),
initdb_params=['--data-checksums'])
self.init_pb(backup_dir)
self.add_instance(backup_dir, 'node', node)
self.set_archiving(backup_dir, 'node', node)
node.slow_start()
backup_id = self.backup_node(backup_dir, 'node', node)
# corrupt backup.control file
file = os.path.join(
backup_dir, "backups", "node",
backup_id, "backup.control")
fd = open(file, 'a')
fd.write("statuss = OK")
fd.close()
self.assertIn(
'WARNING: Invalid option "statuss" in file',
self.show_pb(backup_dir, 'node', as_json=False, as_text=True))
# Clean after yourself
self.del_test_dir(module_name, fname)
# @unittest.skip("skip")
# @unittest.expectedFailure
def test_corrupt_correctness(self):
"""backup.control contains invalid option"""
if not self.remote:
self.skipTest("You must enable PGPROBACKUP_SSH_REMOTE"
" for run this test")
fname = self.id().split('.')[3]
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
node = self.make_simple_node(
base_dir=os.path.join(module_name, fname, 'node'),
initdb_params=['--data-checksums'])
self.init_pb(backup_dir)
self.add_instance(backup_dir, 'node', node)
self.set_archiving(backup_dir, 'node', node)
node.slow_start()
node.pgbench_init(scale=1)
# FULL
backup_local_id = self.backup_node(
backup_dir, 'node', node, no_remote=True)
output_local = self.show_pb(
backup_dir, 'node', as_json=False, backup_id=backup_local_id)
backup_remote_id = self.backup_node(backup_dir, 'node', node)
output_remote = self.show_pb(
backup_dir, 'node', as_json=False, backup_id=backup_remote_id)
# check correctness
self.assertEqual(
output_local['data-bytes'],
output_remote['data-bytes'])
self.assertEqual(
output_local['uncompressed-bytes'],
output_remote['uncompressed-bytes'])
# DELTA
backup_local_id = self.backup_node(
backup_dir, 'node', node,
backup_type='delta', no_remote=True)
output_local = self.show_pb(
backup_dir, 'node', as_json=False, backup_id=backup_local_id)
self.delete_pb(backup_dir, 'node', backup_local_id)
backup_remote_id = self.backup_node(
backup_dir, 'node', node, backup_type='delta')
output_remote = self.show_pb(
backup_dir, 'node', as_json=False, backup_id=backup_remote_id)
self.delete_pb(backup_dir, 'node', backup_remote_id)
# check correctness
self.assertEqual(
output_local['data-bytes'],
output_remote['data-bytes'])
self.assertEqual(
output_local['uncompressed-bytes'],
output_remote['uncompressed-bytes'])
# PAGE
backup_local_id = self.backup_node(
backup_dir, 'node', node,
backup_type='page', no_remote=True)
output_local = self.show_pb(
backup_dir, 'node', as_json=False, backup_id=backup_local_id)
self.delete_pb(backup_dir, 'node', backup_local_id)
backup_remote_id = self.backup_node(
backup_dir, 'node', node, backup_type='page')
output_remote = self.show_pb(
backup_dir, 'node', as_json=False, backup_id=backup_remote_id)
self.delete_pb(backup_dir, 'node', backup_remote_id)
# check correctness
self.assertEqual(
output_local['data-bytes'],
output_remote['data-bytes'])
self.assertEqual(
output_local['uncompressed-bytes'],
output_remote['uncompressed-bytes'])
# Clean after yourself
self.del_test_dir(module_name, fname)
# @unittest.skip("skip")
# @unittest.expectedFailure
def test_corrupt_correctness_1(self):
"""backup.control contains invalid option"""
if not self.remote:
self.skipTest("You must enable PGPROBACKUP_SSH_REMOTE"
" for run this test")
fname = self.id().split('.')[3]
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
node = self.make_simple_node(
base_dir=os.path.join(module_name, fname, 'node'),
initdb_params=['--data-checksums'])
self.init_pb(backup_dir)
self.add_instance(backup_dir, 'node', node)
self.set_archiving(backup_dir, 'node', node)
node.slow_start()
node.pgbench_init(scale=1)
# FULL
backup_local_id = self.backup_node(
backup_dir, 'node', node, no_remote=True)
output_local = self.show_pb(
backup_dir, 'node', as_json=False, backup_id=backup_local_id)
backup_remote_id = self.backup_node(backup_dir, 'node', node)
output_remote = self.show_pb(
backup_dir, 'node', as_json=False, backup_id=backup_remote_id)
# check correctness
self.assertEqual(
output_local['data-bytes'],
output_remote['data-bytes'])
self.assertEqual(
output_local['uncompressed-bytes'],
output_remote['uncompressed-bytes'])
# change data
pgbench = node.pgbench(options=['-T', '10', '--no-vacuum'])
pgbench.wait()
# DELTA
backup_local_id = self.backup_node(
backup_dir, 'node', node,
backup_type='delta', no_remote=True)
output_local = self.show_pb(
backup_dir, 'node', as_json=False, backup_id=backup_local_id)
self.delete_pb(backup_dir, 'node', backup_local_id)
backup_remote_id = self.backup_node(
backup_dir, 'node', node, backup_type='delta')
output_remote = self.show_pb(
backup_dir, 'node', as_json=False, backup_id=backup_remote_id)
self.delete_pb(backup_dir, 'node', backup_remote_id)
# check correctness
self.assertEqual(
output_local['data-bytes'],
output_remote['data-bytes'])
self.assertEqual(
output_local['uncompressed-bytes'],
output_remote['uncompressed-bytes'])
# PAGE
backup_local_id = self.backup_node(
backup_dir, 'node', node,
backup_type='page', no_remote=True)
output_local = self.show_pb(
backup_dir, 'node', as_json=False, backup_id=backup_local_id)
self.delete_pb(backup_dir, 'node', backup_local_id)
backup_remote_id = self.backup_node(
backup_dir, 'node', node, backup_type='page')
output_remote = self.show_pb(
backup_dir, 'node', as_json=False, backup_id=backup_remote_id)
self.delete_pb(backup_dir, 'node', backup_remote_id)
# check correctness
self.assertEqual(
output_local['data-bytes'],
output_remote['data-bytes'])
self.assertEqual(
output_local['uncompressed-bytes'],
output_remote['uncompressed-bytes'])
# Clean after yourself
self.del_test_dir(module_name, fname)
# @unittest.skip("skip")
# @unittest.expectedFailure
def test_corrupt_correctness_2(self):
"""backup.control contains invalid option"""
if not self.remote:
self.skipTest("You must enable PGPROBACKUP_SSH_REMOTE"
" for run this test")
fname = self.id().split('.')[3]
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
node = self.make_simple_node(
base_dir=os.path.join(module_name, fname, 'node'),
initdb_params=['--data-checksums'])
self.init_pb(backup_dir)
self.add_instance(backup_dir, 'node', node)
self.set_archiving(backup_dir, 'node', node)
node.slow_start()
node.pgbench_init(scale=1)
# FULL
backup_local_id = self.backup_node(
backup_dir, 'node', node,
options=['--compress'], no_remote=True)
output_local = self.show_pb(
backup_dir, 'node', as_json=False, backup_id=backup_local_id)
if self.remote:
backup_remote_id = self.backup_node(
backup_dir, 'node', node, options=['--compress'])
else:
backup_remote_id = self.backup_node(
backup_dir, 'node', node,
options=['--remote-proto=ssh', '--remote-host=localhost', '--compress'])
output_remote = self.show_pb(
backup_dir, 'node', as_json=False, backup_id=backup_remote_id)
# check correctness
self.assertEqual(
output_local['data-bytes'],
output_remote['data-bytes'])
self.assertEqual(
output_local['uncompressed-bytes'],
output_remote['uncompressed-bytes'])
# change data
pgbench = node.pgbench(options=['-T', '10', '--no-vacuum'])
pgbench.wait()
# DELTA
backup_local_id = self.backup_node(
backup_dir, 'node', node,
backup_type='delta', options=['--compress'], no_remote=True)
output_local = self.show_pb(
backup_dir, 'node', as_json=False, backup_id=backup_local_id)
self.delete_pb(backup_dir, 'node', backup_local_id)
if self.remote:
backup_remote_id = self.backup_node(
backup_dir, 'node', node, backup_type='delta', options=['--compress'])
else:
backup_remote_id = self.backup_node(
backup_dir, 'node', node, backup_type='delta',
options=['--remote-proto=ssh', '--remote-host=localhost', '--compress'])
output_remote = self.show_pb(
backup_dir, 'node', as_json=False, backup_id=backup_remote_id)
self.delete_pb(backup_dir, 'node', backup_remote_id)
# check correctness
self.assertEqual(
output_local['data-bytes'],
output_remote['data-bytes'])
self.assertEqual(
output_local['uncompressed-bytes'],
output_remote['uncompressed-bytes'])
# PAGE
backup_local_id = self.backup_node(
backup_dir, 'node', node,
backup_type='page', options=['--compress'], no_remote=True)
output_local = self.show_pb(
backup_dir, 'node', as_json=False, backup_id=backup_local_id)
self.delete_pb(backup_dir, 'node', backup_local_id)
if self.remote:
backup_remote_id = self.backup_node(
backup_dir, 'node', node, backup_type='page', options=['--compress'])
else:
backup_remote_id = self.backup_node(
backup_dir, 'node', node, backup_type='page',
options=['--remote-proto=ssh', '--remote-host=localhost', '--compress'])
output_remote = self.show_pb(
backup_dir, 'node', as_json=False, backup_id=backup_remote_id)
self.delete_pb(backup_dir, 'node', backup_remote_id)
# check correctness
self.assertEqual(
output_local['data-bytes'],
output_remote['data-bytes'])
self.assertEqual(
output_local['uncompressed-bytes'],
output_remote['uncompressed-bytes'])
# Clean after yourself
self.del_test_dir(module_name, fname)
# @unittest.skip("skip")
# @unittest.expectedFailure
def test_color_with_no_terminal(self):
"""backup.control contains invalid option"""
fname = self.id().split('.')[3]
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
node = self.make_simple_node(
base_dir=os.path.join(module_name, fname, 'node'),
initdb_params=['--data-checksums'],
pg_options={'autovacuum': 'off'})
self.init_pb(backup_dir)
self.add_instance(backup_dir, 'node', node)
node.slow_start()
node.pgbench_init(scale=1)
# FULL
try:
self.backup_node(
backup_dir, 'node', node, options=['--archive-timeout=1s'])
# we should die here because exception is what we expect to happen
self.assertEqual(
1, 0,
"Expecting Error because archiving is disabled\n "
"Output: {0} \n CMD: {1}".format(
repr(self.output), self.cmd))
except ProbackupException as e:
self.assertNotIn(
'[0m', e.message,
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
repr(e.message), self.cmd))
# Clean after yourself
self.del_test_dir(module_name, fname)