mirror of
https://github.com/postgrespro/pg_probackup.git
synced 2024-11-24 08:52:38 +02:00
4251 lines
173 KiB
Python
4251 lines
173 KiB
Python
import os
|
|
import unittest
|
|
from .helpers.ptrack_helpers import ProbackupTest, ProbackupException
|
|
from datetime import datetime, timedelta
|
|
from pathlib import Path
|
|
import subprocess
|
|
from sys import exit
|
|
import time
|
|
import hashlib
|
|
|
|
|
|
module_name = 'validate'
|
|
|
|
|
|
class ValidateTest(ProbackupTest, unittest.TestCase):
|
|
|
|
# @unittest.skip("skip")
|
|
# @unittest.expectedFailure
|
|
def test_basic_validate_nullified_heap_page_backup(self):
|
|
"""
|
|
make node with nullified heap block
|
|
"""
|
|
fname = self.id().split('.')[3]
|
|
node = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node'),
|
|
initdb_params=['--data-checksums'])
|
|
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
self.set_archiving(backup_dir, 'node', node)
|
|
node.slow_start()
|
|
|
|
node.pgbench_init(scale=3)
|
|
|
|
file_path = node.safe_psql(
|
|
"postgres",
|
|
"select pg_relation_filepath('pgbench_accounts')").decode('utf-8').rstrip()
|
|
|
|
node.safe_psql(
|
|
"postgres",
|
|
"CHECKPOINT")
|
|
|
|
# Nullify some block in PostgreSQL
|
|
file = os.path.join(node.data_dir, file_path)
|
|
with open(file, 'r+b') as f:
|
|
f.seek(8192)
|
|
f.write(b"\x00"*8192)
|
|
f.flush()
|
|
f.close
|
|
|
|
self.backup_node(
|
|
backup_dir, 'node', node, options=['--log-level-file=verbose'])
|
|
|
|
pgdata = self.pgdata_content(node.data_dir)
|
|
|
|
if not self.remote:
|
|
log_file_path = os.path.join(backup_dir, "log", "pg_probackup.log")
|
|
with open(log_file_path) as f:
|
|
log_content = f.read()
|
|
self.assertIn(
|
|
'File: "{0}" blknum 1, empty page'.format(Path(file).as_posix()),
|
|
log_content,
|
|
'Failed to detect nullified block')
|
|
|
|
self.validate_pb(backup_dir, options=["-j", "4"])
|
|
node.cleanup()
|
|
|
|
self.restore_node(backup_dir, 'node', node)
|
|
|
|
pgdata_restored = self.pgdata_content(node.data_dir)
|
|
self.compare_pgdata(pgdata, pgdata_restored)
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
# @unittest.skip("skip")
|
|
# @unittest.expectedFailure
|
|
def test_validate_wal_unreal_values(self):
|
|
"""
|
|
make node with archiving, make archive backup
|
|
validate to both real and unreal values
|
|
"""
|
|
fname = self.id().split('.')[3]
|
|
node = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node'),
|
|
initdb_params=['--data-checksums'])
|
|
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
self.set_archiving(backup_dir, 'node', node)
|
|
node.slow_start()
|
|
|
|
node.pgbench_init(scale=3)
|
|
with node.connect("postgres") as con:
|
|
con.execute("CREATE TABLE tbl0005 (a text)")
|
|
con.commit()
|
|
|
|
backup_id = self.backup_node(backup_dir, 'node', node)
|
|
|
|
node.pgbench_init(scale=3)
|
|
|
|
target_time = self.show_pb(
|
|
backup_dir, 'node', backup_id)['recovery-time']
|
|
after_backup_time = datetime.now().replace(second=0, microsecond=0)
|
|
|
|
# Validate to real time
|
|
self.assertIn(
|
|
"INFO: Backup validation completed successfully",
|
|
self.validate_pb(
|
|
backup_dir, 'node',
|
|
options=["--time={0}".format(target_time), "-j", "4"]),
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(self.output), self.cmd))
|
|
|
|
# Validate to unreal time
|
|
unreal_time_1 = after_backup_time - timedelta(days=2)
|
|
try:
|
|
self.validate_pb(
|
|
backup_dir, 'node', options=["--time={0}".format(
|
|
unreal_time_1), "-j", "4"])
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because of validation to unreal time.\n "
|
|
"Output: {0} \n CMD: {1}".format(
|
|
repr(self.output), self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertIn(
|
|
'ERROR: Backup satisfying target options is not found',
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
# Validate to unreal time #2
|
|
unreal_time_2 = after_backup_time + timedelta(days=2)
|
|
try:
|
|
self.validate_pb(
|
|
backup_dir, 'node',
|
|
options=["--time={0}".format(unreal_time_2), "-j", "4"])
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because of validation to unreal time.\n "
|
|
"Output: {0} \n CMD: {1}".format(
|
|
repr(self.output), self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertTrue(
|
|
'ERROR: Not enough WAL records to time' in e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
# Validate to real xid
|
|
target_xid = None
|
|
with node.connect("postgres") as con:
|
|
res = con.execute(
|
|
"INSERT INTO tbl0005 VALUES ('inserted') RETURNING (xmin)")
|
|
con.commit()
|
|
target_xid = res[0][0]
|
|
self.switch_wal_segment(node)
|
|
time.sleep(5)
|
|
|
|
self.assertIn(
|
|
"INFO: Backup validation completed successfully",
|
|
self.validate_pb(
|
|
backup_dir, 'node', options=["--xid={0}".format(target_xid),
|
|
"-j", "4"]),
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(self.output), self.cmd))
|
|
|
|
# Validate to unreal xid
|
|
unreal_xid = int(target_xid) + 1000
|
|
try:
|
|
self.validate_pb(
|
|
backup_dir, 'node', options=["--xid={0}".format(unreal_xid),
|
|
"-j", "4"])
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because of validation to unreal xid.\n "
|
|
"Output: {0} \n CMD: {1}".format(
|
|
repr(self.output), self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertTrue(
|
|
'ERROR: Not enough WAL records to xid' in e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
# Validate with backup ID
|
|
output = self.validate_pb(backup_dir, 'node', backup_id,
|
|
options=["-j", "4"])
|
|
self.assertIn(
|
|
"INFO: Validating backup {0}".format(backup_id),
|
|
output,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(self.output), self.cmd))
|
|
self.assertIn(
|
|
"INFO: Backup {0} data files are valid".format(backup_id),
|
|
output,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(self.output), self.cmd))
|
|
self.assertIn(
|
|
"INFO: Backup {0} WAL segments are valid".format(backup_id),
|
|
output,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(self.output), self.cmd))
|
|
self.assertIn(
|
|
"INFO: Backup {0} is valid".format(backup_id),
|
|
output,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(self.output), self.cmd))
|
|
self.assertIn(
|
|
"INFO: Validate of backup {0} completed".format(backup_id),
|
|
output,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(self.output), self.cmd))
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
# @unittest.skip("skip")
|
|
def test_basic_validate_corrupted_intermediate_backup(self):
|
|
"""
|
|
make archive node, take FULL, PAGE1, PAGE2 backups,
|
|
corrupt file in PAGE1 backup,
|
|
run validate on PAGE1, expect PAGE1 to gain status CORRUPT
|
|
and PAGE2 gain status ORPHAN
|
|
"""
|
|
fname = self.id().split('.')[3]
|
|
node = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node'),
|
|
initdb_params=['--data-checksums'])
|
|
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
self.set_archiving(backup_dir, 'node', node)
|
|
node.slow_start()
|
|
|
|
# FULL
|
|
backup_id_1 = self.backup_node(backup_dir, 'node', node)
|
|
|
|
node.safe_psql(
|
|
"postgres",
|
|
"create table t_heap as select i as id, md5(i::text) as text, "
|
|
"md5(repeat(i::text,10))::tsvector as tsvector "
|
|
"from generate_series(0,10000) i")
|
|
file_path = node.safe_psql(
|
|
"postgres",
|
|
"select pg_relation_filepath('t_heap')").decode('utf-8').rstrip()
|
|
# PAGE1
|
|
backup_id_2 = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
|
|
node.safe_psql(
|
|
"postgres",
|
|
"insert into t_heap select i as id, md5(i::text) as text, "
|
|
"md5(repeat(i::text,10))::tsvector as tsvector "
|
|
"from generate_series(10000,20000) i")
|
|
# PAGE2
|
|
backup_id_3 = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
|
|
# Corrupt some file
|
|
file = os.path.join(
|
|
backup_dir, 'backups', 'node',
|
|
backup_id_2, 'database', file_path)
|
|
with open(file, "r+b", 0) as f:
|
|
f.seek(42)
|
|
f.write(b"blah")
|
|
f.flush()
|
|
f.close
|
|
|
|
# Simple validate
|
|
try:
|
|
self.validate_pb(
|
|
backup_dir, 'node', backup_id=backup_id_2, options=["-j", "4"])
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because of data files corruption.\n "
|
|
"Output: {0} \n CMD: {1}".format(
|
|
repr(self.output), self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertTrue(
|
|
'INFO: Validating parents for backup {0}'.format(
|
|
backup_id_2) in e.message and
|
|
'ERROR: Backup {0} is corrupt'.format(
|
|
backup_id_2) in e.message and
|
|
'WARNING: Backup {0} data files are corrupted'.format(
|
|
backup_id_2) in e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertEqual(
|
|
'CORRUPT',
|
|
self.show_pb(backup_dir, 'node', backup_id_2)['status'],
|
|
'Backup STATUS should be "CORRUPT"')
|
|
self.assertEqual(
|
|
'ORPHAN',
|
|
self.show_pb(backup_dir, 'node', backup_id_3)['status'],
|
|
'Backup STATUS should be "ORPHAN"')
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
# @unittest.skip("skip")
|
|
def test_validate_corrupted_intermediate_backups(self):
|
|
"""
|
|
make archive node, take FULL, PAGE1, PAGE2 backups,
|
|
corrupt file in FULL and PAGE1 backups, run validate on PAGE1,
|
|
expect FULL and PAGE1 to gain status CORRUPT and
|
|
PAGE2 gain status ORPHAN
|
|
"""
|
|
fname = self.id().split('.')[3]
|
|
node = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node'),
|
|
initdb_params=['--data-checksums'])
|
|
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
self.set_archiving(backup_dir, 'node', node)
|
|
node.slow_start()
|
|
|
|
node.safe_psql(
|
|
"postgres",
|
|
"create table t_heap as select i as id, md5(i::text) as text, "
|
|
"md5(repeat(i::text,10))::tsvector as tsvector "
|
|
"from generate_series(0,10000) i")
|
|
file_path_t_heap = node.safe_psql(
|
|
"postgres",
|
|
"select pg_relation_filepath('t_heap')").decode('utf-8').rstrip()
|
|
# FULL
|
|
backup_id_1 = self.backup_node(backup_dir, 'node', node)
|
|
|
|
node.safe_psql(
|
|
"postgres",
|
|
"create table t_heap_1 as select i as id, md5(i::text) as text, "
|
|
"md5(repeat(i::text,10))::tsvector as tsvector "
|
|
"from generate_series(0,10000) i")
|
|
file_path_t_heap_1 = node.safe_psql(
|
|
"postgres",
|
|
"select pg_relation_filepath('t_heap_1')").decode('utf-8').rstrip()
|
|
# PAGE1
|
|
backup_id_2 = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
|
|
node.safe_psql(
|
|
"postgres",
|
|
"insert into t_heap select i as id, md5(i::text) as text, "
|
|
"md5(repeat(i::text,10))::tsvector as tsvector "
|
|
"from generate_series(20000,30000) i")
|
|
# PAGE2
|
|
backup_id_3 = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
|
|
# Corrupt some file in FULL backup
|
|
file_full = os.path.join(
|
|
backup_dir, 'backups', 'node',
|
|
backup_id_1, 'database', file_path_t_heap)
|
|
with open(file_full, "rb+", 0) as f:
|
|
f.seek(84)
|
|
f.write(b"blah")
|
|
f.flush()
|
|
f.close
|
|
|
|
# Corrupt some file in PAGE1 backup
|
|
file_page1 = os.path.join(
|
|
backup_dir, 'backups', 'node',
|
|
backup_id_2, 'database', file_path_t_heap_1)
|
|
with open(file_page1, "rb+", 0) as f:
|
|
f.seek(42)
|
|
f.write(b"blah")
|
|
f.flush()
|
|
f.close
|
|
|
|
# Validate PAGE1
|
|
try:
|
|
self.validate_pb(
|
|
backup_dir, 'node', backup_id=backup_id_2, options=["-j", "4"])
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because of data files corruption.\n "
|
|
"Output: {0} \n CMD: {1}".format(
|
|
repr(self.output), self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertTrue(
|
|
'INFO: Validating parents for backup {0}'.format(
|
|
backup_id_2) in e.message,
|
|
'\n Unexpected Error Message: {0}\n '
|
|
'CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertTrue(
|
|
'INFO: Validating backup {0}'.format(
|
|
backup_id_1) in e.message and
|
|
'WARNING: Invalid CRC of backup file' in e.message and
|
|
'WARNING: Backup {0} data files are corrupted'.format(
|
|
backup_id_1) in e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertTrue(
|
|
'WARNING: Backup {0} is orphaned because his parent'.format(
|
|
backup_id_2) in e.message and
|
|
'WARNING: Backup {0} is orphaned because his parent'.format(
|
|
backup_id_3) in e.message and
|
|
'ERROR: Backup {0} is orphan.'.format(
|
|
backup_id_2) in e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertEqual(
|
|
'CORRUPT',
|
|
self.show_pb(backup_dir, 'node', backup_id_1)['status'],
|
|
'Backup STATUS should be "CORRUPT"')
|
|
self.assertEqual(
|
|
'ORPHAN',
|
|
self.show_pb(backup_dir, 'node', backup_id_2)['status'],
|
|
'Backup STATUS should be "ORPHAN"')
|
|
self.assertEqual(
|
|
'ORPHAN',
|
|
self.show_pb(backup_dir, 'node', backup_id_3)['status'],
|
|
'Backup STATUS should be "ORPHAN"')
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
# @unittest.skip("skip")
|
|
def test_validate_specific_error_intermediate_backups(self):
|
|
"""
|
|
make archive node, take FULL, PAGE1, PAGE2 backups,
|
|
change backup status of FULL and PAGE1 to ERROR,
|
|
run validate on PAGE1
|
|
purpose of this test is to be sure that not only
|
|
CORRUPT backup descendants can be orphanized
|
|
"""
|
|
fname = self.id().split('.')[3]
|
|
node = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node'),
|
|
initdb_params=['--data-checksums'])
|
|
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
self.set_archiving(backup_dir, 'node', node)
|
|
node.slow_start()
|
|
|
|
# FULL
|
|
backup_id_1 = self.backup_node(backup_dir, 'node', node)
|
|
|
|
# PAGE1
|
|
backup_id_2 = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
|
|
# PAGE2
|
|
backup_id_3 = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
|
|
# Change FULL backup status to ERROR
|
|
control_path = os.path.join(
|
|
backup_dir, 'backups', 'node', backup_id_1, 'backup.control')
|
|
|
|
with open(control_path, 'r') as f:
|
|
actual_control = f.read()
|
|
|
|
new_control_file = ''
|
|
for line in actual_control.splitlines():
|
|
new_control_file += line.replace(
|
|
'status = OK', 'status = ERROR')
|
|
new_control_file += '\n'
|
|
|
|
with open(control_path, 'wt') as f:
|
|
f.write(new_control_file)
|
|
f.flush()
|
|
f.close()
|
|
|
|
# Validate PAGE1
|
|
try:
|
|
self.validate_pb(
|
|
backup_dir, 'node', backup_id=backup_id_2, options=["-j", "4"])
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because backup has status ERROR.\n "
|
|
"Output: {0} \n CMD: {1}".format(
|
|
repr(self.output), self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertTrue(
|
|
'WARNING: Backup {0} is orphaned because '
|
|
'his parent {1} has status: ERROR'.format(
|
|
backup_id_2, backup_id_1) in e.message and
|
|
'INFO: Validating parents for backup {0}'.format(
|
|
backup_id_2) in e.message and
|
|
'WARNING: Backup {0} has status ERROR. Skip validation.'.format(
|
|
backup_id_1) and
|
|
'ERROR: Backup {0} is orphan.'.format(backup_id_2) in e.message,
|
|
'\n Unexpected Error Message: {0}\n '
|
|
'CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertEqual(
|
|
'ERROR',
|
|
self.show_pb(backup_dir, 'node', backup_id_1)['status'],
|
|
'Backup STATUS should be "ERROR"')
|
|
self.assertEqual(
|
|
'ORPHAN',
|
|
self.show_pb(backup_dir, 'node', backup_id_2)['status'],
|
|
'Backup STATUS should be "ORPHAN"')
|
|
self.assertEqual(
|
|
'ORPHAN',
|
|
self.show_pb(backup_dir, 'node', backup_id_3)['status'],
|
|
'Backup STATUS should be "ORPHAN"')
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
# @unittest.skip("skip")
|
|
def test_validate_error_intermediate_backups(self):
|
|
"""
|
|
make archive node, take FULL, PAGE1, PAGE2 backups,
|
|
change backup status of FULL and PAGE1 to ERROR,
|
|
run validate on instance
|
|
purpose of this test is to be sure that not only
|
|
CORRUPT backup descendants can be orphanized
|
|
"""
|
|
fname = self.id().split('.')[3]
|
|
node = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node'),
|
|
initdb_params=['--data-checksums'])
|
|
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
self.set_archiving(backup_dir, 'node', node)
|
|
node.slow_start()
|
|
|
|
# FULL
|
|
backup_id_1 = self.backup_node(backup_dir, 'node', node)
|
|
|
|
# PAGE1
|
|
backup_id_2 = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
|
|
# PAGE2
|
|
backup_id_3 = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
|
|
# Change FULL backup status to ERROR
|
|
control_path = os.path.join(
|
|
backup_dir, 'backups', 'node', backup_id_1, 'backup.control')
|
|
|
|
with open(control_path, 'r') as f:
|
|
actual_control = f.read()
|
|
|
|
new_control_file = ''
|
|
for line in actual_control.splitlines():
|
|
new_control_file += line.replace(
|
|
'status = OK', 'status = ERROR')
|
|
new_control_file += '\n'
|
|
|
|
with open(control_path, 'wt') as f:
|
|
f.write(new_control_file)
|
|
f.flush()
|
|
f.close()
|
|
|
|
# Validate instance
|
|
try:
|
|
self.validate_pb(backup_dir, options=["-j", "4"])
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because backup has status ERROR.\n "
|
|
"Output: {0} \n CMD: {1}".format(
|
|
repr(self.output), self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertTrue(
|
|
"WARNING: Backup {0} is orphaned because "
|
|
"his parent {1} has status: ERROR".format(
|
|
backup_id_2, backup_id_1) in e.message and
|
|
'WARNING: Backup {0} has status ERROR. Skip validation'.format(
|
|
backup_id_1) in e.message and
|
|
"WARNING: Some backups are not valid" in e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertEqual(
|
|
'ERROR',
|
|
self.show_pb(backup_dir, 'node', backup_id_1)['status'],
|
|
'Backup STATUS should be "ERROR"')
|
|
self.assertEqual(
|
|
'ORPHAN',
|
|
self.show_pb(backup_dir, 'node', backup_id_2)['status'],
|
|
'Backup STATUS should be "ORPHAN"')
|
|
self.assertEqual(
|
|
'ORPHAN',
|
|
self.show_pb(backup_dir, 'node', backup_id_3)['status'],
|
|
'Backup STATUS should be "ORPHAN"')
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
# @unittest.skip("skip")
|
|
def test_validate_corrupted_intermediate_backups_1(self):
|
|
"""
|
|
make archive node, FULL1, PAGE1, PAGE2, PAGE3, PAGE4, PAGE5, FULL2,
|
|
corrupt file in PAGE1 and PAGE4, run validate on PAGE3,
|
|
expect PAGE1 to gain status CORRUPT, PAGE2, PAGE3, PAGE4 and PAGE5
|
|
to gain status ORPHAN
|
|
"""
|
|
fname = self.id().split('.')[3]
|
|
node = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node'),
|
|
initdb_params=['--data-checksums'])
|
|
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
self.set_archiving(backup_dir, 'node', node)
|
|
node.slow_start()
|
|
|
|
# FULL1
|
|
backup_id_1 = self.backup_node(backup_dir, 'node', node)
|
|
|
|
# PAGE1
|
|
node.safe_psql(
|
|
"postgres",
|
|
"create table t_heap as select i as id, md5(i::text) as text, "
|
|
"md5(repeat(i::text,10))::tsvector as tsvector "
|
|
"from generate_series(0,10000) i")
|
|
backup_id_2 = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
|
|
# PAGE2
|
|
node.safe_psql(
|
|
"postgres",
|
|
"insert into t_heap select i as id, md5(i::text) as text, "
|
|
"md5(repeat(i::text,10))::tsvector as tsvector "
|
|
"from generate_series(0,10000) i")
|
|
file_page_2 = node.safe_psql(
|
|
"postgres",
|
|
"select pg_relation_filepath('t_heap')").decode('utf-8').rstrip()
|
|
backup_id_3 = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
|
|
# PAGE3
|
|
node.safe_psql(
|
|
"postgres",
|
|
"insert into t_heap select i as id, md5(i::text) as text, "
|
|
"md5(repeat(i::text,10))::tsvector as tsvector "
|
|
"from generate_series(10000,20000) i")
|
|
backup_id_4 = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
|
|
# PAGE4
|
|
node.safe_psql(
|
|
"postgres",
|
|
"insert into t_heap select i as id, md5(i::text) as text, "
|
|
"md5(repeat(i::text,10))::tsvector as tsvector "
|
|
"from generate_series(20000,30000) i")
|
|
backup_id_5 = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
|
|
# PAGE5
|
|
node.safe_psql(
|
|
"postgres",
|
|
"create table t_heap1 as select i as id, md5(i::text) as text, "
|
|
"md5(repeat(i::text,10))::tsvector as tsvector "
|
|
"from generate_series(0,10000) i")
|
|
file_page_5 = node.safe_psql(
|
|
"postgres",
|
|
"select pg_relation_filepath('t_heap1')").decode('utf-8').rstrip()
|
|
backup_id_6 = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
|
|
# PAGE6
|
|
node.safe_psql(
|
|
"postgres",
|
|
"insert into t_heap select i as id, md5(i::text) as text, "
|
|
"md5(repeat(i::text,10))::tsvector as tsvector "
|
|
"from generate_series(30000,40000) i")
|
|
backup_id_7 = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
|
|
# FULL2
|
|
backup_id_8 = self.backup_node(backup_dir, 'node', node)
|
|
|
|
# Corrupt some file in PAGE2 and PAGE5 backups
|
|
file_page1 = os.path.join(
|
|
backup_dir, 'backups', 'node', backup_id_3, 'database', file_page_2)
|
|
with open(file_page1, "rb+", 0) as f:
|
|
f.seek(84)
|
|
f.write(b"blah")
|
|
f.flush()
|
|
f.close
|
|
|
|
file_page4 = os.path.join(
|
|
backup_dir, 'backups', 'node', backup_id_6, 'database', file_page_5)
|
|
with open(file_page4, "rb+", 0) as f:
|
|
f.seek(42)
|
|
f.write(b"blah")
|
|
f.flush()
|
|
f.close
|
|
|
|
# Validate PAGE3
|
|
try:
|
|
self.validate_pb(
|
|
backup_dir, 'node',
|
|
backup_id=backup_id_4, options=["-j", "4"])
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because of data files corruption.\n"
|
|
" Output: {0} \n CMD: {1}".format(
|
|
repr(self.output), self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertTrue(
|
|
'INFO: Validating parents for backup {0}'.format(
|
|
backup_id_4) in e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertTrue(
|
|
'INFO: Validating backup {0}'.format(
|
|
backup_id_1) in e.message and
|
|
'INFO: Backup {0} data files are valid'.format(
|
|
backup_id_1) in e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertTrue(
|
|
'INFO: Validating backup {0}'.format(
|
|
backup_id_2) in e.message and
|
|
'INFO: Backup {0} data files are valid'.format(
|
|
backup_id_2) in e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertTrue(
|
|
'INFO: Validating backup {0}'.format(
|
|
backup_id_3) in e.message and
|
|
'WARNING: Invalid CRC of backup file' in e.message and
|
|
'WARNING: Backup {0} data files are corrupted'.format(
|
|
backup_id_3) in e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertTrue(
|
|
'WARNING: Backup {0} is orphaned because '
|
|
'his parent {1} has status: CORRUPT'.format(
|
|
backup_id_4, backup_id_3) in e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertTrue(
|
|
'WARNING: Backup {0} is orphaned because '
|
|
'his parent {1} has status: CORRUPT'.format(
|
|
backup_id_5, backup_id_3) in e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertTrue(
|
|
'WARNING: Backup {0} is orphaned because '
|
|
'his parent {1} has status: CORRUPT'.format(
|
|
backup_id_6, backup_id_3) in e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertTrue(
|
|
'WARNING: Backup {0} is orphaned because '
|
|
'his parent {1} has status: CORRUPT'.format(
|
|
backup_id_7, backup_id_3) in e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertTrue(
|
|
'ERROR: Backup {0} is orphan'.format(backup_id_4) in e.message,
|
|
'\n Unexpected Error Message: {0}\n '
|
|
'CMD: {1}'.format(repr(e.message), self.cmd))
|
|
|
|
self.assertEqual(
|
|
'OK', self.show_pb(backup_dir, 'node', backup_id_1)['status'],
|
|
'Backup STATUS should be "OK"')
|
|
self.assertEqual(
|
|
'OK', self.show_pb(backup_dir, 'node', backup_id_2)['status'],
|
|
'Backup STATUS should be "OK"')
|
|
self.assertEqual(
|
|
'CORRUPT', self.show_pb(backup_dir, 'node', backup_id_3)['status'],
|
|
'Backup STATUS should be "CORRUPT"')
|
|
self.assertEqual(
|
|
'ORPHAN', self.show_pb(backup_dir, 'node', backup_id_4)['status'],
|
|
'Backup STATUS should be "ORPHAN"')
|
|
self.assertEqual(
|
|
'ORPHAN', self.show_pb(backup_dir, 'node', backup_id_5)['status'],
|
|
'Backup STATUS should be "ORPHAN"')
|
|
self.assertEqual(
|
|
'ORPHAN', self.show_pb(backup_dir, 'node', backup_id_6)['status'],
|
|
'Backup STATUS should be "ORPHAN"')
|
|
self.assertEqual(
|
|
'ORPHAN', self.show_pb(backup_dir, 'node', backup_id_7)['status'],
|
|
'Backup STATUS should be "ORPHAN"')
|
|
self.assertEqual(
|
|
'OK', self.show_pb(backup_dir, 'node', backup_id_8)['status'],
|
|
'Backup STATUS should be "OK"')
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
# @unittest.skip("skip")
|
|
def test_validate_specific_target_corrupted_intermediate_backups(self):
|
|
"""
|
|
make archive node, take FULL1, PAGE1, PAGE2, PAGE3, PAGE4, PAGE5, FULL2
|
|
corrupt file in PAGE1 and PAGE4, run validate on PAGE3 to specific xid,
|
|
expect PAGE1 to gain status CORRUPT, PAGE2, PAGE3, PAGE4 and PAGE5 to
|
|
gain status ORPHAN
|
|
"""
|
|
fname = self.id().split('.')[3]
|
|
node = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node'),
|
|
initdb_params=['--data-checksums'])
|
|
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
self.set_archiving(backup_dir, 'node', node)
|
|
node.slow_start()
|
|
|
|
# FULL1
|
|
backup_id_1 = self.backup_node(backup_dir, 'node', node)
|
|
|
|
# PAGE1
|
|
node.safe_psql(
|
|
"postgres",
|
|
"create table t_heap as select i as id, md5(i::text) as text, "
|
|
"md5(repeat(i::text,10))::tsvector as tsvector "
|
|
"from generate_series(0,10000) i")
|
|
backup_id_2 = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
|
|
# PAGE2
|
|
node.safe_psql(
|
|
"postgres",
|
|
"insert into t_heap select i as id, md5(i::text) as text, "
|
|
"md5(repeat(i::text,10))::tsvector as tsvector "
|
|
"from generate_series(0,10000) i")
|
|
file_page_2 = node.safe_psql(
|
|
"postgres",
|
|
"select pg_relation_filepath('t_heap')").decode('utf-8').rstrip()
|
|
backup_id_3 = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
|
|
# PAGE3
|
|
node.safe_psql(
|
|
"postgres",
|
|
"insert into t_heap select i as id, md5(i::text) as text, "
|
|
"md5(repeat(i::text,10))::tsvector as tsvector "
|
|
"from generate_series(10000,20000) i")
|
|
backup_id_4 = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
|
|
# PAGE4
|
|
node.safe_psql(
|
|
"postgres",
|
|
"insert into t_heap select i as id, md5(i::text) as text, "
|
|
"md5(repeat(i::text,10))::tsvector as tsvector "
|
|
"from generate_series(20000,30000) i")
|
|
|
|
target_xid = node.safe_psql(
|
|
"postgres",
|
|
"insert into t_heap select i as id, md5(i::text) as text, "
|
|
"md5(repeat(i::text,10))::tsvector as tsvector "
|
|
"from generate_series(30001, 30001) i RETURNING (xmin)").decode('utf-8').rstrip()
|
|
|
|
backup_id_5 = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
|
|
# PAGE5
|
|
node.safe_psql(
|
|
"postgres",
|
|
"create table t_heap1 as select i as id, md5(i::text) as text, "
|
|
"md5(repeat(i::text,10))::tsvector as tsvector "
|
|
"from generate_series(0,10000) i")
|
|
file_page_5 = node.safe_psql(
|
|
"postgres",
|
|
"select pg_relation_filepath('t_heap1')").decode('utf-8').rstrip()
|
|
backup_id_6 = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
|
|
# PAGE6
|
|
node.safe_psql(
|
|
"postgres",
|
|
"insert into t_heap select i as id, md5(i::text) as text, "
|
|
"md5(repeat(i::text,10))::tsvector as tsvector "
|
|
"from generate_series(30000,40000) i")
|
|
backup_id_7 = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
|
|
# FULL2
|
|
backup_id_8 = self.backup_node(backup_dir, 'node', node)
|
|
|
|
# Corrupt some file in PAGE2 and PAGE5 backups
|
|
file_page1 = os.path.join(
|
|
backup_dir, 'backups', 'node',
|
|
backup_id_3, 'database', file_page_2)
|
|
with open(file_page1, "rb+", 0) as f:
|
|
f.seek(84)
|
|
f.write(b"blah")
|
|
f.flush()
|
|
f.close
|
|
|
|
file_page4 = os.path.join(
|
|
backup_dir, 'backups', 'node',
|
|
backup_id_6, 'database', file_page_5)
|
|
with open(file_page4, "rb+", 0) as f:
|
|
f.seek(42)
|
|
f.write(b"blah")
|
|
f.flush()
|
|
f.close
|
|
|
|
# Validate PAGE3
|
|
try:
|
|
self.validate_pb(
|
|
backup_dir, 'node',
|
|
options=[
|
|
'-i', backup_id_4, '--xid={0}'.format(target_xid), "-j", "4"])
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because of data files corruption.\n "
|
|
"Output: {0} \n CMD: {1}".format(
|
|
repr(self.output), self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertTrue(
|
|
'INFO: Validating parents for backup {0}'.format(
|
|
backup_id_4) in e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertTrue(
|
|
'INFO: Validating backup {0}'.format(
|
|
backup_id_1) in e.message and
|
|
'INFO: Backup {0} data files are valid'.format(
|
|
backup_id_1) in e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertTrue(
|
|
'INFO: Validating backup {0}'.format(
|
|
backup_id_2) in e.message and
|
|
'INFO: Backup {0} data files are valid'.format(
|
|
backup_id_2) in e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertTrue(
|
|
'INFO: Validating backup {0}'.format(
|
|
backup_id_3) in e.message and
|
|
'WARNING: Invalid CRC of backup file' in e.message and
|
|
'WARNING: Backup {0} data files are corrupted'.format(
|
|
backup_id_3) in e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertTrue(
|
|
'WARNING: Backup {0} is orphaned because his '
|
|
'parent {1} has status: CORRUPT'.format(
|
|
backup_id_4, backup_id_3) in e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertTrue(
|
|
'WARNING: Backup {0} is orphaned because his '
|
|
'parent {1} has status: CORRUPT'.format(
|
|
backup_id_5, backup_id_3) in e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertTrue(
|
|
'WARNING: Backup {0} is orphaned because his '
|
|
'parent {1} has status: CORRUPT'.format(
|
|
backup_id_6, backup_id_3) in e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertTrue(
|
|
'WARNING: Backup {0} is orphaned because his '
|
|
'parent {1} has status: CORRUPT'.format(
|
|
backup_id_7, backup_id_3) in e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertTrue(
|
|
'ERROR: Backup {0} is orphan'.format(
|
|
backup_id_4) in e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertEqual('OK', self.show_pb(backup_dir, 'node', backup_id_1)['status'], 'Backup STATUS should be "OK"')
|
|
self.assertEqual('OK', self.show_pb(backup_dir, 'node', backup_id_2)['status'], 'Backup STATUS should be "OK"')
|
|
self.assertEqual('CORRUPT', self.show_pb(backup_dir, 'node', backup_id_3)['status'], 'Backup STATUS should be "CORRUPT"')
|
|
self.assertEqual('ORPHAN', self.show_pb(backup_dir, 'node', backup_id_4)['status'], 'Backup STATUS should be "ORPHAN"')
|
|
self.assertEqual('ORPHAN', self.show_pb(backup_dir, 'node', backup_id_5)['status'], 'Backup STATUS should be "ORPHAN"')
|
|
self.assertEqual('ORPHAN', self.show_pb(backup_dir, 'node', backup_id_6)['status'], 'Backup STATUS should be "ORPHAN"')
|
|
self.assertEqual('ORPHAN', self.show_pb(backup_dir, 'node', backup_id_7)['status'], 'Backup STATUS should be "ORPHAN"')
|
|
self.assertEqual('OK', self.show_pb(backup_dir, 'node', backup_id_8)['status'], 'Backup STATUS should be "OK"')
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
# @unittest.skip("skip")
|
|
def test_validate_instance_with_several_corrupt_backups(self):
|
|
"""
|
|
make archive node, take FULL1, PAGE1_1, FULL2, PAGE2_1 backups, FULL3
|
|
corrupt file in FULL and FULL2 and run validate on instance,
|
|
expect FULL1 to gain status CORRUPT, PAGE1_1 to gain status ORPHAN
|
|
FULL2 to gain status CORRUPT, PAGE2_1 to gain status ORPHAN
|
|
"""
|
|
fname = self.id().split('.')[3]
|
|
node = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node'),
|
|
initdb_params=['--data-checksums'])
|
|
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
self.set_archiving(backup_dir, 'node', node)
|
|
node.slow_start()
|
|
|
|
node.safe_psql(
|
|
"postgres",
|
|
"create table t_heap as select generate_series(0,1) i")
|
|
# FULL1
|
|
backup_id_1 = self.backup_node(
|
|
backup_dir, 'node', node, options=['--no-validate'])
|
|
|
|
# FULL2
|
|
backup_id_2 = self.backup_node(backup_dir, 'node', node)
|
|
rel_path = node.safe_psql(
|
|
"postgres",
|
|
"select pg_relation_filepath('t_heap')").decode('utf-8').rstrip()
|
|
|
|
node.safe_psql(
|
|
"postgres",
|
|
"insert into t_heap values(2)")
|
|
|
|
backup_id_3 = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
|
|
# FULL3
|
|
backup_id_4 = self.backup_node(backup_dir, 'node', node)
|
|
|
|
node.safe_psql(
|
|
"postgres",
|
|
"insert into t_heap values(3)")
|
|
|
|
backup_id_5 = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
|
|
# FULL4
|
|
backup_id_6 = self.backup_node(
|
|
backup_dir, 'node', node, options=['--no-validate'])
|
|
|
|
# Corrupt some files in FULL2 and FULL3 backup
|
|
os.remove(os.path.join(
|
|
backup_dir, 'backups', 'node', backup_id_2,
|
|
'database', rel_path))
|
|
os.remove(os.path.join(
|
|
backup_dir, 'backups', 'node', backup_id_4,
|
|
'database', rel_path))
|
|
|
|
# Validate Instance
|
|
try:
|
|
self.validate_pb(backup_dir, 'node', options=["-j", "4", "--log-level-file=LOG"])
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because of data files corruption.\n "
|
|
"Output: {0} \n CMD: {1}".format(
|
|
repr(self.output), self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertTrue(
|
|
"INFO: Validate backups of the instance 'node'" in e.message,
|
|
"\n Unexpected Error Message: {0}\n "
|
|
"CMD: {1}".format(repr(e.message), self.cmd))
|
|
self.assertTrue(
|
|
'WARNING: Some backups are not valid' in e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertEqual(
|
|
'OK', self.show_pb(backup_dir, 'node', backup_id_1)['status'],
|
|
'Backup STATUS should be "OK"')
|
|
self.assertEqual(
|
|
'CORRUPT', self.show_pb(backup_dir, 'node', backup_id_2)['status'],
|
|
'Backup STATUS should be "CORRUPT"')
|
|
self.assertEqual(
|
|
'ORPHAN', self.show_pb(backup_dir, 'node', backup_id_3)['status'],
|
|
'Backup STATUS should be "ORPHAN"')
|
|
self.assertEqual(
|
|
'CORRUPT', self.show_pb(backup_dir, 'node', backup_id_4)['status'],
|
|
'Backup STATUS should be "CORRUPT"')
|
|
self.assertEqual(
|
|
'ORPHAN', self.show_pb(backup_dir, 'node', backup_id_5)['status'],
|
|
'Backup STATUS should be "ORPHAN"')
|
|
self.assertEqual(
|
|
'OK', self.show_pb(backup_dir, 'node', backup_id_6)['status'],
|
|
'Backup STATUS should be "OK"')
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
# @unittest.skip("skip")
|
|
def test_validate_instance_with_several_corrupt_backups_interrupt(self):
|
|
"""
|
|
check that interrupt during validation is handled correctly
|
|
"""
|
|
self._check_gdb_flag_or_skip_test()
|
|
|
|
fname = self.id().split('.')[3]
|
|
node = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node'),
|
|
initdb_params=['--data-checksums'])
|
|
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
self.set_archiving(backup_dir, 'node', node)
|
|
node.slow_start()
|
|
|
|
node.safe_psql(
|
|
"postgres",
|
|
"create table t_heap as select generate_series(0,1) i")
|
|
# FULL1
|
|
backup_id_1 = self.backup_node(
|
|
backup_dir, 'node', node, options=['--no-validate'])
|
|
|
|
# FULL2
|
|
backup_id_2 = self.backup_node(backup_dir, 'node', node)
|
|
rel_path = node.safe_psql(
|
|
"postgres",
|
|
"select pg_relation_filepath('t_heap')").decode('utf-8').rstrip()
|
|
|
|
node.safe_psql(
|
|
"postgres",
|
|
"insert into t_heap values(2)")
|
|
|
|
backup_id_3 = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
|
|
# FULL3
|
|
backup_id_4 = self.backup_node(backup_dir, 'node', node)
|
|
|
|
node.safe_psql(
|
|
"postgres",
|
|
"insert into t_heap values(3)")
|
|
|
|
backup_id_5 = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
|
|
# FULL4
|
|
backup_id_6 = self.backup_node(
|
|
backup_dir, 'node', node, options=['--no-validate'])
|
|
|
|
# Corrupt some files in FULL2 and FULL3 backup
|
|
os.remove(os.path.join(
|
|
backup_dir, 'backups', 'node', backup_id_1,
|
|
'database', rel_path))
|
|
os.remove(os.path.join(
|
|
backup_dir, 'backups', 'node', backup_id_3,
|
|
'database', rel_path))
|
|
|
|
# Validate Instance
|
|
gdb = self.validate_pb(
|
|
backup_dir, 'node', options=["-j", "4", "--log-level-file=LOG"], gdb=True)
|
|
|
|
gdb.set_breakpoint('validate_file_pages')
|
|
gdb.run_until_break()
|
|
gdb.continue_execution_until_break()
|
|
gdb.remove_all_breakpoints()
|
|
gdb._execute('signal SIGINT')
|
|
gdb.continue_execution_until_error()
|
|
|
|
self.assertEqual(
|
|
'DONE', self.show_pb(backup_dir, 'node', backup_id_1)['status'],
|
|
'Backup STATUS should be "OK"')
|
|
self.assertEqual(
|
|
'OK', self.show_pb(backup_dir, 'node', backup_id_2)['status'],
|
|
'Backup STATUS should be "OK"')
|
|
self.assertEqual(
|
|
'OK', self.show_pb(backup_dir, 'node', backup_id_3)['status'],
|
|
'Backup STATUS should be "CORRUPT"')
|
|
self.assertEqual(
|
|
'OK', self.show_pb(backup_dir, 'node', backup_id_4)['status'],
|
|
'Backup STATUS should be "ORPHAN"')
|
|
self.assertEqual(
|
|
'OK', self.show_pb(backup_dir, 'node', backup_id_5)['status'],
|
|
'Backup STATUS should be "OK"')
|
|
self.assertEqual(
|
|
'DONE', self.show_pb(backup_dir, 'node', backup_id_6)['status'],
|
|
'Backup STATUS should be "OK"')
|
|
|
|
log_file = os.path.join(backup_dir, 'log', 'pg_probackup.log')
|
|
with open(log_file, 'r') as f:
|
|
log_content = f.read()
|
|
self.assertNotIn(
|
|
'Interrupted while locking backup', log_content)
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
# @unittest.skip("skip")
|
|
def test_validate_instance_with_corrupted_page(self):
|
|
"""
|
|
make archive node, take FULL, PAGE1, PAGE2, FULL2, PAGE3 backups,
|
|
corrupt file in PAGE1 backup and run validate on instance,
|
|
expect PAGE1 to gain status CORRUPT, PAGE2 to gain status ORPHAN
|
|
"""
|
|
fname = self.id().split('.')[3]
|
|
node = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node'),
|
|
initdb_params=['--data-checksums'])
|
|
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
self.set_archiving(backup_dir, 'node', node)
|
|
node.slow_start()
|
|
|
|
node.safe_psql(
|
|
"postgres",
|
|
"create table t_heap as select i as id, md5(i::text) as text, "
|
|
"md5(repeat(i::text,10))::tsvector as tsvector "
|
|
"from generate_series(0,10000) i")
|
|
# FULL1
|
|
backup_id_1 = self.backup_node(backup_dir, 'node', node)
|
|
|
|
node.safe_psql(
|
|
"postgres",
|
|
"create table t_heap1 as select i as id, md5(i::text) as text, "
|
|
"md5(repeat(i::text,10))::tsvector as tsvector "
|
|
"from generate_series(0,10000) i")
|
|
file_path_t_heap1 = node.safe_psql(
|
|
"postgres",
|
|
"select pg_relation_filepath('t_heap1')").decode('utf-8').rstrip()
|
|
# PAGE1
|
|
backup_id_2 = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
|
|
node.safe_psql(
|
|
"postgres",
|
|
"insert into t_heap select i as id, md5(i::text) as text, "
|
|
"md5(repeat(i::text,10))::tsvector as tsvector "
|
|
"from generate_series(20000,30000) i")
|
|
# PAGE2
|
|
backup_id_3 = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
# FULL1
|
|
backup_id_4 = self.backup_node(
|
|
backup_dir, 'node', node)
|
|
# PAGE3
|
|
backup_id_5 = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
|
|
# Corrupt some file in FULL backup
|
|
file_full = os.path.join(
|
|
backup_dir, 'backups', 'node', backup_id_2,
|
|
'database', file_path_t_heap1)
|
|
with open(file_full, "rb+", 0) as f:
|
|
f.seek(84)
|
|
f.write(b"blah")
|
|
f.flush()
|
|
f.close
|
|
|
|
# Validate Instance
|
|
try:
|
|
self.validate_pb(backup_dir, 'node', options=["-j", "4"])
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because of data files corruption.\n "
|
|
"Output: {0} \n CMD: {1}".format(
|
|
repr(self.output), self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertTrue(
|
|
"INFO: Validate backups of the instance 'node'" in e.message,
|
|
"\n Unexpected Error Message: {0}\n "
|
|
"CMD: {1}".format(repr(e.message), self.cmd))
|
|
self.assertTrue(
|
|
'INFO: Validating backup {0}'.format(
|
|
backup_id_5) in e.message and
|
|
'INFO: Backup {0} data files are valid'.format(
|
|
backup_id_5) in e.message and
|
|
'INFO: Backup {0} WAL segments are valid'.format(
|
|
backup_id_5) in e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertTrue(
|
|
'INFO: Validating backup {0}'.format(
|
|
backup_id_4) in e.message and
|
|
'INFO: Backup {0} data files are valid'.format(
|
|
backup_id_4) in e.message and
|
|
'INFO: Backup {0} WAL segments are valid'.format(
|
|
backup_id_4) in e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertTrue(
|
|
'INFO: Validating backup {0}'.format(
|
|
backup_id_3) in e.message and
|
|
'INFO: Backup {0} data files are valid'.format(
|
|
backup_id_3) in e.message and
|
|
'INFO: Backup {0} WAL segments are valid'.format(
|
|
backup_id_3) in e.message and
|
|
'WARNING: Backup {0} is orphaned because '
|
|
'his parent {1} has status: CORRUPT'.format(
|
|
backup_id_3, backup_id_2) in e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertTrue(
|
|
'INFO: Validating backup {0}'.format(
|
|
backup_id_2) in e.message and
|
|
'WARNING: Invalid CRC of backup file' in e.message and
|
|
'WARNING: Backup {0} data files are corrupted'.format(
|
|
backup_id_2) in e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertTrue(
|
|
'INFO: Validating backup {0}'.format(
|
|
backup_id_1) in e.message and
|
|
'INFO: Backup {0} data files are valid'.format(
|
|
backup_id_1) in e.message and
|
|
'INFO: Backup {0} WAL segments are valid'.format(
|
|
backup_id_1) in e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertTrue(
|
|
'WARNING: Some backups are not valid' in e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertEqual(
|
|
'OK', self.show_pb(backup_dir, 'node', backup_id_1)['status'],
|
|
'Backup STATUS should be "OK"')
|
|
self.assertEqual(
|
|
'CORRUPT', self.show_pb(backup_dir, 'node', backup_id_2)['status'],
|
|
'Backup STATUS should be "CORRUPT"')
|
|
self.assertEqual(
|
|
'ORPHAN', self.show_pb(backup_dir, 'node', backup_id_3)['status'],
|
|
'Backup STATUS should be "ORPHAN"')
|
|
self.assertEqual(
|
|
'OK', self.show_pb(backup_dir, 'node', backup_id_4)['status'],
|
|
'Backup STATUS should be "OK"')
|
|
self.assertEqual(
|
|
'OK', self.show_pb(backup_dir, 'node', backup_id_5)['status'],
|
|
'Backup STATUS should be "OK"')
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
# @unittest.skip("skip")
|
|
def test_validate_instance_with_corrupted_full_and_try_restore(self):
|
|
"""make archive node, take FULL, PAGE1, PAGE2, FULL2, PAGE3 backups,
|
|
corrupt file in FULL backup and run validate on instance,
|
|
expect FULL to gain status CORRUPT, PAGE1 and PAGE2 to gain status ORPHAN,
|
|
try to restore backup with --no-validation option"""
|
|
fname = self.id().split('.')[3]
|
|
node = self.make_simple_node(base_dir=os.path.join(module_name, fname, 'node'),
|
|
initdb_params=['--data-checksums'])
|
|
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
self.set_archiving(backup_dir, 'node', node)
|
|
node.slow_start()
|
|
|
|
node.safe_psql(
|
|
"postgres",
|
|
"create table t_heap as select i as id, md5(i::text) as text, "
|
|
"md5(repeat(i::text,10))::tsvector as tsvector "
|
|
"from generate_series(0,10000) i")
|
|
file_path_t_heap = node.safe_psql(
|
|
"postgres",
|
|
"select pg_relation_filepath('t_heap')").decode('utf-8').rstrip()
|
|
# FULL1
|
|
backup_id_1 = self.backup_node(backup_dir, 'node', node)
|
|
|
|
node.safe_psql(
|
|
"postgres",
|
|
"insert into t_heap select i as id, md5(i::text) as text, "
|
|
"md5(repeat(i::text,10))::tsvector as tsvector "
|
|
"from generate_series(0,10000) i")
|
|
# PAGE1
|
|
backup_id_2 = self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
|
|
# PAGE2
|
|
node.safe_psql(
|
|
"postgres",
|
|
"insert into t_heap select i as id, md5(i::text) as text, "
|
|
"md5(repeat(i::text,10))::tsvector as tsvector "
|
|
"from generate_series(20000,30000) i")
|
|
backup_id_3 = self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
|
|
# FULL1
|
|
backup_id_4 = self.backup_node(backup_dir, 'node', node)
|
|
|
|
# PAGE3
|
|
node.safe_psql(
|
|
"postgres",
|
|
"insert into t_heap select i as id, "
|
|
"md5(i::text) as text, md5(repeat(i::text,10))::tsvector as tsvector "
|
|
"from generate_series(30000,40000) i")
|
|
backup_id_5 = self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
|
|
# Corrupt some file in FULL backup
|
|
file_full = os.path.join(
|
|
backup_dir, 'backups', 'node',
|
|
backup_id_1, 'database', file_path_t_heap)
|
|
with open(file_full, "rb+", 0) as f:
|
|
f.seek(84)
|
|
f.write(b"blah")
|
|
f.flush()
|
|
f.close
|
|
|
|
# Validate Instance
|
|
try:
|
|
self.validate_pb(backup_dir, 'node', options=["-j", "4"])
|
|
self.assertEqual(1, 0, "Expecting Error because of data files corruption.\n Output: {0} \n CMD: {1}".format(
|
|
repr(self.output), self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertTrue(
|
|
'INFO: Validating backup {0}'.format(backup_id_1) in e.message
|
|
and "INFO: Validate backups of the instance 'node'" in e.message
|
|
and 'WARNING: Invalid CRC of backup file' in e.message
|
|
and 'WARNING: Backup {0} data files are corrupted'.format(backup_id_1) in e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(repr(e.message), self.cmd))
|
|
|
|
self.assertEqual('CORRUPT', self.show_pb(backup_dir, 'node', backup_id_1)['status'], 'Backup STATUS should be "CORRUPT"')
|
|
self.assertEqual('ORPHAN', self.show_pb(backup_dir, 'node', backup_id_2)['status'], 'Backup STATUS should be "ORPHAN"')
|
|
self.assertEqual('ORPHAN', self.show_pb(backup_dir, 'node', backup_id_3)['status'], 'Backup STATUS should be "ORPHAN"')
|
|
self.assertEqual('OK', self.show_pb(backup_dir, 'node', backup_id_4)['status'], 'Backup STATUS should be "OK"')
|
|
self.assertEqual('OK', self.show_pb(backup_dir, 'node', backup_id_5)['status'], 'Backup STATUS should be "OK"')
|
|
|
|
node.cleanup()
|
|
restore_out = self.restore_node(
|
|
backup_dir, 'node', node,
|
|
options=["--no-validate"])
|
|
self.assertIn(
|
|
"INFO: Restore of backup {0} completed.".format(backup_id_5),
|
|
restore_out,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(self.output), self.cmd))
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
# @unittest.skip("skip")
|
|
def test_validate_instance_with_corrupted_full(self):
|
|
"""make archive node, take FULL, PAGE1, PAGE2, FULL2, PAGE3 backups,
|
|
corrupt file in FULL backup and run validate on instance,
|
|
expect FULL to gain status CORRUPT, PAGE1 and PAGE2 to gain status ORPHAN"""
|
|
fname = self.id().split('.')[3]
|
|
node = self.make_simple_node(base_dir=os.path.join(module_name, fname, 'node'),
|
|
initdb_params=['--data-checksums'])
|
|
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
self.set_archiving(backup_dir, 'node', node)
|
|
node.slow_start()
|
|
|
|
node.safe_psql(
|
|
"postgres",
|
|
"create table t_heap as select i as id, "
|
|
"md5(i::text) as text, md5(repeat(i::text,10))::tsvector as tsvector "
|
|
"from generate_series(0,10000) i")
|
|
|
|
file_path_t_heap = node.safe_psql(
|
|
"postgres",
|
|
"select pg_relation_filepath('t_heap')").decode('utf-8').rstrip()
|
|
# FULL1
|
|
backup_id_1 = self.backup_node(backup_dir, 'node', node)
|
|
|
|
node.safe_psql(
|
|
"postgres",
|
|
"insert into t_heap select i as id, md5(i::text) as text, "
|
|
"md5(repeat(i::text,10))::tsvector as tsvector "
|
|
"from generate_series(0,10000) i")
|
|
|
|
# PAGE1
|
|
backup_id_2 = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
|
|
# PAGE2
|
|
node.safe_psql(
|
|
"postgres",
|
|
"insert into t_heap select i as id, md5(i::text) as text, "
|
|
"md5(repeat(i::text,10))::tsvector as tsvector "
|
|
"from generate_series(20000,30000) i")
|
|
|
|
backup_id_3 = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
|
|
# FULL1
|
|
backup_id_4 = self.backup_node(
|
|
backup_dir, 'node', node)
|
|
|
|
# PAGE3
|
|
node.safe_psql(
|
|
"postgres",
|
|
"insert into t_heap select i as id, "
|
|
"md5(i::text) as text, md5(repeat(i::text,10))::tsvector as tsvector "
|
|
"from generate_series(30000,40000) i")
|
|
backup_id_5 = self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
|
|
# Corrupt some file in FULL backup
|
|
file_full = os.path.join(
|
|
backup_dir, 'backups', 'node',
|
|
backup_id_1, 'database', file_path_t_heap)
|
|
with open(file_full, "rb+", 0) as f:
|
|
f.seek(84)
|
|
f.write(b"blah")
|
|
f.flush()
|
|
f.close
|
|
|
|
# Validate Instance
|
|
try:
|
|
self.validate_pb(backup_dir, 'node', options=["-j", "4"])
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because of data files corruption.\n "
|
|
"Output: {0} \n CMD: {1}".format(
|
|
repr(self.output), self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertTrue(
|
|
'INFO: Validating backup {0}'.format(backup_id_1) in e.message
|
|
and "INFO: Validate backups of the instance 'node'" in e.message
|
|
and 'WARNING: Invalid CRC of backup file' in e.message
|
|
and 'WARNING: Backup {0} data files are corrupted'.format(backup_id_1) in e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(repr(e.message), self.cmd))
|
|
|
|
self.assertEqual('CORRUPT', self.show_pb(backup_dir, 'node', backup_id_1)['status'], 'Backup STATUS should be "CORRUPT"')
|
|
self.assertEqual('ORPHAN', self.show_pb(backup_dir, 'node', backup_id_2)['status'], 'Backup STATUS should be "ORPHAN"')
|
|
self.assertEqual('ORPHAN', self.show_pb(backup_dir, 'node', backup_id_3)['status'], 'Backup STATUS should be "ORPHAN"')
|
|
self.assertEqual('OK', self.show_pb(backup_dir, 'node', backup_id_4)['status'], 'Backup STATUS should be "OK"')
|
|
self.assertEqual('OK', self.show_pb(backup_dir, 'node', backup_id_5)['status'], 'Backup STATUS should be "OK"')
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
# @unittest.skip("skip")
|
|
def test_validate_corrupt_wal_1(self):
|
|
"""make archive node, take FULL1, PAGE1,PAGE2,FULL2,PAGE3,PAGE4 backups, corrupt all wal files, run validate, expect errors"""
|
|
fname = self.id().split('.')[3]
|
|
node = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node'),
|
|
initdb_params=['--data-checksums'])
|
|
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
self.set_archiving(backup_dir, 'node', node)
|
|
node.slow_start()
|
|
|
|
backup_id_1 = self.backup_node(backup_dir, 'node', node)
|
|
|
|
with node.connect("postgres") as con:
|
|
con.execute("CREATE TABLE tbl0005 (a text)")
|
|
con.commit()
|
|
|
|
backup_id_2 = self.backup_node(backup_dir, 'node', node)
|
|
|
|
# Corrupt WAL
|
|
wals_dir = os.path.join(backup_dir, 'wal', 'node')
|
|
wals = [f for f in os.listdir(wals_dir) if os.path.isfile(os.path.join(wals_dir, f)) and not f.endswith('.backup')]
|
|
wals.sort()
|
|
for wal in wals:
|
|
with open(os.path.join(wals_dir, wal), "rb+", 0) as f:
|
|
f.seek(42)
|
|
f.write(b"blablablaadssaaaaaaaaaaaaaaa")
|
|
f.flush()
|
|
f.close
|
|
|
|
# Simple validate
|
|
try:
|
|
self.validate_pb(backup_dir, 'node', options=["-j", "4"])
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because of wal segments corruption.\n"
|
|
" Output: {0} \n CMD: {1}".format(
|
|
repr(self.output), self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertTrue(
|
|
'WARNING: Backup' in e.message and
|
|
'WAL segments are corrupted' in e.message and
|
|
"WARNING: There are not enough WAL "
|
|
"records to consistenly restore backup" in e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertEqual(
|
|
'CORRUPT',
|
|
self.show_pb(backup_dir, 'node', backup_id_1)['status'],
|
|
'Backup STATUS should be "CORRUPT"')
|
|
self.assertEqual(
|
|
'CORRUPT',
|
|
self.show_pb(backup_dir, 'node', backup_id_2)['status'],
|
|
'Backup STATUS should be "CORRUPT"')
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
# @unittest.skip("skip")
|
|
def test_validate_corrupt_wal_2(self):
|
|
"""make archive node, make full backup, corrupt all wal files, run validate to real xid, expect errors"""
|
|
fname = self.id().split('.')[3]
|
|
node = self.make_simple_node(base_dir=os.path.join(module_name, fname, 'node'),
|
|
initdb_params=['--data-checksums'])
|
|
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
self.set_archiving(backup_dir, 'node', node)
|
|
node.slow_start()
|
|
|
|
with node.connect("postgres") as con:
|
|
con.execute("CREATE TABLE tbl0005 (a text)")
|
|
con.commit()
|
|
|
|
backup_id = self.backup_node(backup_dir, 'node', node)
|
|
target_xid = None
|
|
with node.connect("postgres") as con:
|
|
res = con.execute(
|
|
"INSERT INTO tbl0005 VALUES ('inserted') RETURNING (xmin)")
|
|
con.commit()
|
|
target_xid = res[0][0]
|
|
|
|
# Corrupt WAL
|
|
wals_dir = os.path.join(backup_dir, 'wal', 'node')
|
|
wals = [f for f in os.listdir(wals_dir) if os.path.isfile(os.path.join(wals_dir, f)) and not f.endswith('.backup')]
|
|
wals.sort()
|
|
for wal in wals:
|
|
with open(os.path.join(wals_dir, wal), "rb+", 0) as f:
|
|
f.seek(128)
|
|
f.write(b"blablablaadssaaaaaaaaaaaaaaa")
|
|
f.flush()
|
|
f.close
|
|
|
|
# Validate to xid
|
|
try:
|
|
self.validate_pb(
|
|
backup_dir,
|
|
'node',
|
|
backup_id,
|
|
options=[
|
|
"--xid={0}".format(target_xid), "-j", "4"])
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because of wal segments corruption.\n"
|
|
" Output: {0} \n CMD: {1}".format(
|
|
repr(self.output), self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertTrue(
|
|
'WARNING: Backup' in e.message and
|
|
'WAL segments are corrupted' in e.message and
|
|
"WARNING: There are not enough WAL "
|
|
"records to consistenly restore backup" in e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertEqual(
|
|
'CORRUPT',
|
|
self.show_pb(backup_dir, 'node', backup_id)['status'],
|
|
'Backup STATUS should be "CORRUPT"')
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
# @unittest.skip("skip")
|
|
def test_validate_wal_lost_segment_1(self):
|
|
"""make archive node, make archive full backup,
|
|
delete from archive wal segment which belong to previous backup
|
|
run validate, expecting error because of missing wal segment
|
|
make sure that backup status is 'CORRUPT'
|
|
"""
|
|
fname = self.id().split('.')[3]
|
|
node = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node'),
|
|
initdb_params=['--data-checksums'])
|
|
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
self.set_archiving(backup_dir, 'node', node)
|
|
node.slow_start()
|
|
|
|
node.pgbench_init(scale=3)
|
|
|
|
backup_id = self.backup_node(backup_dir, 'node', node)
|
|
|
|
# Delete wal segment
|
|
wals_dir = os.path.join(backup_dir, 'wal', 'node')
|
|
wals = [f for f in os.listdir(wals_dir) if os.path.isfile(os.path.join(wals_dir, f)) and not f.endswith('.backup')]
|
|
wals.sort()
|
|
file = os.path.join(backup_dir, 'wal', 'node', wals[-1])
|
|
os.remove(file)
|
|
|
|
# cut out '.gz'
|
|
if self.archive_compress:
|
|
file = file[:-3]
|
|
|
|
try:
|
|
self.validate_pb(backup_dir, 'node', options=["-j", "4"])
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because of wal segment disappearance.\n"
|
|
" Output: {0} \n CMD: {1}".format(
|
|
repr(self.output), self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertTrue(
|
|
"is absent" in e.message and
|
|
"WARNING: There are not enough WAL records to consistenly "
|
|
"restore backup {0}".format(backup_id) in e.message and
|
|
"WARNING: Backup {0} WAL segments are corrupted".format(
|
|
backup_id) in e.message and
|
|
"WARNING: Some backups are not valid" in e.message,
|
|
"\n Unexpected Error Message: {0}\n CMD: {1}".format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertEqual(
|
|
'CORRUPT',
|
|
self.show_pb(backup_dir, 'node', backup_id)['status'],
|
|
'Backup {0} should have STATUS "CORRUPT"')
|
|
|
|
# Run validate again
|
|
try:
|
|
self.validate_pb(backup_dir, 'node', backup_id, options=["-j", "4"])
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because of backup corruption.\n"
|
|
" Output: {0} \n CMD: {1}".format(
|
|
repr(self.output), self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertIn(
|
|
'INFO: Revalidating backup {0}'.format(backup_id), e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertIn(
|
|
'ERROR: Backup {0} is corrupt.'.format(backup_id), e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
# @unittest.skip("skip")
|
|
def test_validate_corrupt_wal_between_backups(self):
|
|
"""
|
|
make archive node, make full backup, corrupt all wal files,
|
|
run validate to real xid, expect errors
|
|
"""
|
|
fname = self.id().split('.')[3]
|
|
node = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node'),
|
|
initdb_params=['--data-checksums'])
|
|
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
self.set_archiving(backup_dir, 'node', node)
|
|
node.slow_start()
|
|
|
|
backup_id = self.backup_node(backup_dir, 'node', node)
|
|
|
|
# make some wals
|
|
node.pgbench_init(scale=3)
|
|
|
|
with node.connect("postgres") as con:
|
|
con.execute("CREATE TABLE tbl0005 (a text)")
|
|
con.commit()
|
|
|
|
with node.connect("postgres") as con:
|
|
res = con.execute(
|
|
"INSERT INTO tbl0005 VALUES ('inserted') RETURNING (xmin)")
|
|
con.commit()
|
|
target_xid = res[0][0]
|
|
|
|
if self.get_version(node) < self.version_to_num('10.0'):
|
|
walfile = node.safe_psql(
|
|
'postgres',
|
|
'select pg_xlogfile_name(pg_current_xlog_location())').decode('utf-8').rstrip()
|
|
else:
|
|
walfile = node.safe_psql(
|
|
'postgres',
|
|
'select pg_walfile_name(pg_current_wal_lsn())').decode('utf-8').rstrip()
|
|
|
|
if self.archive_compress:
|
|
walfile = walfile + '.gz'
|
|
self.switch_wal_segment(node)
|
|
|
|
# generate some wals
|
|
node.pgbench_init(scale=3)
|
|
|
|
self.backup_node(backup_dir, 'node', node)
|
|
|
|
# Corrupt WAL
|
|
wals_dir = os.path.join(backup_dir, 'wal', 'node')
|
|
with open(os.path.join(wals_dir, walfile), "rb+", 0) as f:
|
|
f.seek(9000)
|
|
f.write(b"b")
|
|
f.flush()
|
|
f.close
|
|
|
|
# Validate to xid
|
|
try:
|
|
self.validate_pb(
|
|
backup_dir,
|
|
'node',
|
|
backup_id,
|
|
options=[
|
|
"--xid={0}".format(target_xid), "-j", "4"])
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because of wal segments corruption.\n"
|
|
" Output: {0} \n CMD: {1}".format(
|
|
repr(self.output), self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertTrue(
|
|
'ERROR: Not enough WAL records to xid' in e.message and
|
|
'WARNING: Recovery can be done up to time' in e.message and
|
|
"ERROR: Not enough WAL records to xid {0}\n".format(
|
|
target_xid),
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertEqual(
|
|
'OK',
|
|
self.show_pb(backup_dir, 'node')[0]['status'],
|
|
'Backup STATUS should be "OK"')
|
|
|
|
self.assertEqual(
|
|
'OK',
|
|
self.show_pb(backup_dir, 'node')[1]['status'],
|
|
'Backup STATUS should be "OK"')
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
# @unittest.skip("skip")
|
|
def test_pgpro702_688(self):
|
|
"""
|
|
make node without archiving, make stream backup,
|
|
get Recovery Time, validate to Recovery Time
|
|
"""
|
|
fname = self.id().split('.')[3]
|
|
node = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node'),
|
|
set_replication=True,
|
|
initdb_params=['--data-checksums'])
|
|
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
node.slow_start()
|
|
|
|
backup_id = self.backup_node(
|
|
backup_dir, 'node', node, options=["--stream"])
|
|
recovery_time = self.show_pb(
|
|
backup_dir, 'node', backup_id=backup_id)['recovery-time']
|
|
|
|
try:
|
|
self.validate_pb(
|
|
backup_dir, 'node',
|
|
options=["--time={0}".format(recovery_time), "-j", "4"])
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because of wal segment disappearance.\n "
|
|
"Output: {0} \n CMD: {1}".format(
|
|
self.output, self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertIn(
|
|
'WAL archive is empty. You cannot restore backup to a '
|
|
'recovery target without WAL archive', e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
# @unittest.skip("skip")
|
|
def test_pgpro688(self):
|
|
"""
|
|
make node with archiving, make backup, get Recovery Time,
|
|
validate to Recovery Time. Waiting PGPRO-688. RESOLVED
|
|
"""
|
|
fname = self.id().split('.')[3]
|
|
node = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node'),
|
|
set_replication=True,
|
|
initdb_params=['--data-checksums'])
|
|
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
self.set_archiving(backup_dir, 'node', node)
|
|
node.slow_start()
|
|
|
|
backup_id = self.backup_node(backup_dir, 'node', node)
|
|
recovery_time = self.show_pb(
|
|
backup_dir, 'node', backup_id)['recovery-time']
|
|
|
|
self.validate_pb(
|
|
backup_dir, 'node', options=["--time={0}".format(recovery_time),
|
|
"-j", "4"])
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
# @unittest.skip("skip")
|
|
# @unittest.expectedFailure
|
|
def test_pgpro561(self):
|
|
"""
|
|
make node with archiving, make stream backup,
|
|
restore it to node1, check that archiving is not successful on node1
|
|
"""
|
|
fname = self.id().split('.')[3]
|
|
node1 = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node1'),
|
|
set_replication=True,
|
|
initdb_params=['--data-checksums'])
|
|
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node1', node1)
|
|
self.set_archiving(backup_dir, 'node1', node1)
|
|
node1.slow_start()
|
|
|
|
backup_id = self.backup_node(
|
|
backup_dir, 'node1', node1, options=["--stream"])
|
|
|
|
node2 = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node2'))
|
|
node2.cleanup()
|
|
|
|
node1.psql(
|
|
"postgres",
|
|
"create table t_heap as select i as id, md5(i::text) as text, "
|
|
"md5(repeat(i::text,10))::tsvector as tsvector "
|
|
"from generate_series(0,256) i")
|
|
|
|
self.backup_node(
|
|
backup_dir, 'node1', node1,
|
|
backup_type='page', options=["--stream"])
|
|
self.restore_node(backup_dir, 'node1', data_dir=node2.data_dir)
|
|
|
|
self.set_auto_conf(
|
|
node2, {'port': node2.port, 'archive_mode': 'off'})
|
|
|
|
node2.slow_start()
|
|
|
|
self.set_auto_conf(
|
|
node2, {'archive_mode': 'on'})
|
|
|
|
node2.stop()
|
|
node2.slow_start()
|
|
|
|
timeline_node1 = node1.get_control_data()["Latest checkpoint's TimeLineID"]
|
|
timeline_node2 = node2.get_control_data()["Latest checkpoint's TimeLineID"]
|
|
self.assertEqual(
|
|
timeline_node1, timeline_node2,
|
|
"Timelines on Master and Node1 should be equal. "
|
|
"This is unexpected")
|
|
|
|
archive_command_node1 = node1.safe_psql(
|
|
"postgres", "show archive_command")
|
|
archive_command_node2 = node2.safe_psql(
|
|
"postgres", "show archive_command")
|
|
self.assertEqual(
|
|
archive_command_node1, archive_command_node2,
|
|
"Archive command on Master and Node should be equal. "
|
|
"This is unexpected")
|
|
|
|
# result = node2.safe_psql("postgres", "select last_failed_wal from pg_stat_get_archiver() where last_failed_wal is not NULL")
|
|
## self.assertEqual(res, six.b(""), 'Restored Node1 failed to archive segment {0} due to having the same archive command as Master'.format(res.rstrip()))
|
|
# if result == "":
|
|
# self.assertEqual(1, 0, 'Error is expected due to Master and Node1 having the common archive and archive_command')
|
|
|
|
node1.psql(
|
|
"postgres",
|
|
"create table t_heap_1 as select i as id, md5(i::text) as text, "
|
|
"md5(repeat(i::text,10))::tsvector as tsvector "
|
|
"from generate_series(0,10) i")
|
|
|
|
self.switch_wal_segment(node1)
|
|
|
|
# wals_dir = os.path.join(backup_dir, 'wal', 'node1')
|
|
# wals = [f for f in os.listdir(wals_dir) if os.path.isfile(os.path.join(
|
|
# wals_dir, f)) and not f.endswith('.backup') and not f.endswith('.part')]
|
|
# wals = map(str, wals)
|
|
# print(wals)
|
|
|
|
self.switch_wal_segment(node2)
|
|
|
|
# wals_dir = os.path.join(backup_dir, 'wal', 'node1')
|
|
# wals = [f for f in os.listdir(wals_dir) if os.path.isfile(os.path.join(
|
|
# wals_dir, f)) and not f.endswith('.backup') and not f.endswith('.part')]
|
|
# wals = map(str, wals)
|
|
# print(wals)
|
|
|
|
time.sleep(5)
|
|
|
|
log_file = os.path.join(node2.logs_dir, 'postgresql.log')
|
|
with open(log_file, 'r') as f:
|
|
log_content = f.read()
|
|
self.assertTrue(
|
|
'LOG: archive command failed with exit code 1' in log_content and
|
|
'DETAIL: The failed archive command was:' in log_content and
|
|
'WAL file already exists in archive with different checksum' in log_content,
|
|
'Expecting error messages about failed archive_command'
|
|
)
|
|
self.assertFalse(
|
|
'pg_probackup archive-push completed successfully' in log_content)
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
# @unittest.skip("skip")
|
|
def test_validate_corrupted_full(self):
|
|
"""
|
|
make node with archiving, take full backup, and three page backups,
|
|
take another full backup and three page backups
|
|
corrupt second full backup, run validate, check that
|
|
second full backup became CORRUPT and his page backups are ORPHANs
|
|
remove corruption and run valudate again, check that
|
|
second full backup and his page backups are OK
|
|
"""
|
|
fname = self.id().split('.')[3]
|
|
node = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node'),
|
|
set_replication=True,
|
|
initdb_params=['--data-checksums'],
|
|
pg_options={
|
|
'checkpoint_timeout': '30'})
|
|
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
self.set_archiving(backup_dir, 'node', node)
|
|
node.slow_start()
|
|
|
|
self.backup_node(backup_dir, 'node', node)
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
|
|
backup_id = self.backup_node(backup_dir, 'node', node)
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
|
|
node.safe_psql(
|
|
"postgres",
|
|
"alter system set archive_command = 'false'")
|
|
node.reload()
|
|
try:
|
|
self.backup_node(
|
|
backup_dir, 'node', node,
|
|
backup_type='page', options=['--archive-timeout=1s'])
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because of data file dissapearance.\n "
|
|
"Output: {0} \n CMD: {1}".format(
|
|
self.output, self.cmd))
|
|
except ProbackupException as e:
|
|
pass
|
|
|
|
self.assertTrue(
|
|
self.show_pb(backup_dir, 'node')[6]['status'] == 'ERROR')
|
|
self.set_archiving(backup_dir, 'node', node)
|
|
node.reload()
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
|
|
file = os.path.join(
|
|
backup_dir, 'backups', 'node',
|
|
backup_id, 'database', 'postgresql.auto.conf')
|
|
|
|
file_new = os.path.join(backup_dir, 'postgresql.auto.conf')
|
|
os.rename(file, file_new)
|
|
|
|
try:
|
|
self.validate_pb(backup_dir, options=["-j", "4"])
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because of data file dissapearance.\n "
|
|
"Output: {0} \n CMD: {1}".format(
|
|
self.output, self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertIn(
|
|
'Validating backup {0}'.format(backup_id), e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertIn(
|
|
'WARNING: Backup {0} data files are corrupted'.format(
|
|
backup_id), e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertIn(
|
|
'WARNING: Some backups are not valid'.format(
|
|
backup_id), e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[0]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[1]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[2]['status'] == 'OK')
|
|
self.assertTrue(
|
|
self.show_pb(backup_dir, 'node')[3]['status'] == 'CORRUPT')
|
|
self.assertTrue(
|
|
self.show_pb(backup_dir, 'node')[4]['status'] == 'ORPHAN')
|
|
self.assertTrue(
|
|
self.show_pb(backup_dir, 'node')[5]['status'] == 'ORPHAN')
|
|
self.assertTrue(
|
|
self.show_pb(backup_dir, 'node')[6]['status'] == 'ERROR')
|
|
self.assertTrue(
|
|
self.show_pb(backup_dir, 'node')[7]['status'] == 'ORPHAN')
|
|
|
|
os.rename(file_new, file)
|
|
try:
|
|
self.validate_pb(backup_dir, options=["-j", "4"])
|
|
except ProbackupException as e:
|
|
self.assertIn(
|
|
'WARNING: Some backups are not valid'.format(
|
|
backup_id), e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[0]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[1]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[2]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[3]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[4]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[5]['status'] == 'OK')
|
|
self.assertTrue(
|
|
self.show_pb(backup_dir, 'node')[6]['status'] == 'ERROR')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[7]['status'] == 'OK')
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
# @unittest.skip("skip")
|
|
def test_validate_corrupted_full_1(self):
|
|
"""
|
|
make node with archiving, take full backup, and three page backups,
|
|
take another full backup and four page backups
|
|
corrupt second full backup, run validate, check that
|
|
second full backup became CORRUPT and his page backups are ORPHANs
|
|
remove corruption from full backup and corrupt his second page backup
|
|
run valudate again, check that
|
|
second full backup and his firts page backups are OK,
|
|
second page should be CORRUPT
|
|
third page should be ORPHAN
|
|
"""
|
|
fname = self.id().split('.')[3]
|
|
node = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node'),
|
|
set_replication=True,
|
|
initdb_params=['--data-checksums'])
|
|
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
self.set_archiving(backup_dir, 'node', node)
|
|
node.slow_start()
|
|
|
|
self.backup_node(backup_dir, 'node', node)
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
|
|
backup_id = self.backup_node(backup_dir, 'node', node)
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
backup_id_page = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
|
|
file = os.path.join(
|
|
backup_dir, 'backups', 'node',
|
|
backup_id, 'database', 'postgresql.auto.conf')
|
|
|
|
file_new = os.path.join(backup_dir, 'postgresql.auto.conf')
|
|
os.rename(file, file_new)
|
|
|
|
try:
|
|
self.validate_pb(backup_dir, options=["-j", "4"])
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because of data file dissapearance.\n "
|
|
"Output: {0} \n CMD: {1}".format(
|
|
self.output, self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertIn(
|
|
'Validating backup {0}'.format(backup_id), e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertIn(
|
|
'WARNING: Backup {0} data files are corrupted'.format(
|
|
backup_id), e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertIn(
|
|
'WARNING: Some backups are not valid'.format(
|
|
backup_id), e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[6]['status'] == 'ORPHAN')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[5]['status'] == 'ORPHAN')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[4]['status'] == 'ORPHAN')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[3]['status'] == 'CORRUPT')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[2]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[1]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[0]['status'] == 'OK')
|
|
|
|
os.rename(file_new, file)
|
|
|
|
file = os.path.join(
|
|
backup_dir, 'backups', 'node',
|
|
backup_id_page, 'database', 'backup_label')
|
|
|
|
file_new = os.path.join(backup_dir, 'backup_label')
|
|
os.rename(file, file_new)
|
|
|
|
try:
|
|
self.validate_pb(backup_dir, options=["-j", "4"])
|
|
except ProbackupException as e:
|
|
self.assertIn(
|
|
'WARNING: Some backups are not valid'.format(
|
|
backup_id), e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[0]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[1]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[2]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[3]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[4]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[5]['status'] == 'CORRUPT')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[6]['status'] == 'ORPHAN')
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
# @unittest.skip("skip")
|
|
def test_validate_corrupted_full_2(self):
|
|
"""
|
|
PAGE2_2b
|
|
PAGE2_2a
|
|
PAGE2_4
|
|
PAGE2_4 <- validate
|
|
PAGE2_3
|
|
PAGE2_2 <- CORRUPT
|
|
PAGE2_1
|
|
FULL2
|
|
PAGE1_1
|
|
FULL1
|
|
corrupt second page backup, run validate on PAGE2_3, check that
|
|
PAGE2_2 became CORRUPT and his descendants are ORPHANs,
|
|
take two more PAGE backups, which now trace their origin
|
|
to PAGE2_1 - latest OK backup,
|
|
run validate on PAGE2_3, check that PAGE2_2a and PAGE2_2b are OK,
|
|
|
|
remove corruption from PAGE2_2 and run validate on PAGE2_4
|
|
"""
|
|
|
|
fname = self.id().split('.')[3]
|
|
node = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node'),
|
|
set_replication=True,
|
|
initdb_params=['--data-checksums'])
|
|
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
self.set_archiving(backup_dir, 'node', node)
|
|
node.slow_start()
|
|
|
|
self.backup_node(backup_dir, 'node', node)
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
|
|
self.backup_node(backup_dir, 'node', node)
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
corrupt_id = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
validate_id = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
|
|
file = os.path.join(
|
|
backup_dir, 'backups', 'node',
|
|
corrupt_id, 'database', 'backup_label')
|
|
|
|
file_new = os.path.join(backup_dir, 'backup_label')
|
|
os.rename(file, file_new)
|
|
|
|
try:
|
|
self.validate_pb(backup_dir, 'node', validate_id,
|
|
options=["-j", "4"])
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because of data file dissapearance.\n "
|
|
"Output: {0} \n CMD: {1}".format(
|
|
self.output, self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertIn(
|
|
'INFO: Validating parents for backup {0}'.format(validate_id),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertIn(
|
|
'INFO: Validating backup {0}'.format(
|
|
self.show_pb(backup_dir, 'node')[2]['id']), e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertIn(
|
|
'INFO: Validating backup {0}'.format(
|
|
self.show_pb(backup_dir, 'node')[3]['id']), e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertIn(
|
|
'INFO: Validating backup {0}'.format(
|
|
corrupt_id), e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertIn(
|
|
'WARNING: Backup {0} data files are corrupted'.format(
|
|
corrupt_id), e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[7]['status'] == 'ORPHAN')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[6]['status'] == 'ORPHAN')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[5]['status'] == 'ORPHAN')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[4]['status'] == 'CORRUPT')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[3]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[2]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[1]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[0]['status'] == 'OK')
|
|
|
|
# THIS IS GOLD!!!!
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
|
|
try:
|
|
self.validate_pb(backup_dir, 'node', options=["-j", "4"])
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because of data file dissapearance.\n "
|
|
"Output: {0} \n CMD: {1}".format(
|
|
self.output, self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertIn(
|
|
'Backup {0} data files are valid'.format(
|
|
self.show_pb(backup_dir, 'node')[9]['id']),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertIn(
|
|
'Backup {0} data files are valid'.format(
|
|
self.show_pb(backup_dir, 'node')[8]['id']),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertIn(
|
|
'WARNING: Backup {0} has parent {1} with status: CORRUPT'.format(
|
|
self.show_pb(backup_dir, 'node')[7]['id'], corrupt_id),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertIn(
|
|
'WARNING: Backup {0} has parent {1} with status: CORRUPT'.format(
|
|
self.show_pb(backup_dir, 'node')[6]['id'], corrupt_id),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertIn(
|
|
'WARNING: Backup {0} has parent {1} with status: CORRUPT'.format(
|
|
self.show_pb(backup_dir, 'node')[5]['id'], corrupt_id),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertIn(
|
|
'INFO: Revalidating backup {0}'.format(
|
|
corrupt_id),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertIn(
|
|
'WARNING: Some backups are not valid', e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[9]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[8]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[7]['status'] == 'ORPHAN')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[6]['status'] == 'ORPHAN')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[5]['status'] == 'ORPHAN')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[4]['status'] == 'CORRUPT')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[3]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[2]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[1]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[0]['status'] == 'OK')
|
|
|
|
# revalidate again
|
|
|
|
try:
|
|
self.validate_pb(backup_dir, 'node', validate_id,
|
|
options=["-j", "4"])
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because of data file dissapearance.\n "
|
|
"Output: {0} \n CMD: {1}".format(
|
|
self.output, self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertIn(
|
|
'WARNING: Backup {0} has status: ORPHAN'.format(validate_id),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertIn(
|
|
'Backup {0} has parent {1} with status: CORRUPT'.format(
|
|
self.show_pb(backup_dir, 'node')[7]['id'], corrupt_id),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertIn(
|
|
'Backup {0} has parent {1} with status: CORRUPT'.format(
|
|
self.show_pb(backup_dir, 'node')[6]['id'], corrupt_id),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertIn(
|
|
'Backup {0} has parent {1} with status: CORRUPT'.format(
|
|
self.show_pb(backup_dir, 'node')[5]['id'], corrupt_id),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertIn(
|
|
'INFO: Validating parents for backup {0}'.format(
|
|
validate_id), e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertIn(
|
|
'INFO: Validating backup {0}'.format(
|
|
self.show_pb(backup_dir, 'node')[2]['id']), e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertIn(
|
|
'INFO: Validating backup {0}'.format(
|
|
self.show_pb(backup_dir, 'node')[3]['id']), e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertIn(
|
|
'INFO: Revalidating backup {0}'.format(
|
|
corrupt_id), e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertIn(
|
|
'WARNING: Backup {0} data files are corrupted'.format(
|
|
corrupt_id), e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertIn(
|
|
'ERROR: Backup {0} is orphan.'.format(
|
|
validate_id), e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
# Fix CORRUPT
|
|
os.rename(file_new, file)
|
|
|
|
output = self.validate_pb(backup_dir, 'node', validate_id,
|
|
options=["-j", "4"])
|
|
|
|
self.assertIn(
|
|
'WARNING: Backup {0} has status: ORPHAN'.format(validate_id),
|
|
output,
|
|
'\n Unexpected Output Message: {0}\n'.format(
|
|
repr(output)))
|
|
|
|
self.assertIn(
|
|
'Backup {0} has parent {1} with status: CORRUPT'.format(
|
|
self.show_pb(backup_dir, 'node')[7]['id'], corrupt_id),
|
|
output,
|
|
'\n Unexpected Output Message: {0}\n'.format(
|
|
repr(output)))
|
|
|
|
self.assertIn(
|
|
'Backup {0} has parent {1} with status: CORRUPT'.format(
|
|
self.show_pb(backup_dir, 'node')[6]['id'], corrupt_id),
|
|
output,
|
|
'\n Unexpected Output Message: {0}\n'.format(
|
|
repr(output)))
|
|
|
|
self.assertIn(
|
|
'Backup {0} has parent {1} with status: CORRUPT'.format(
|
|
self.show_pb(backup_dir, 'node')[5]['id'], corrupt_id),
|
|
output,
|
|
'\n Unexpected Output Message: {0}\n'.format(
|
|
repr(output)))
|
|
|
|
self.assertIn(
|
|
'INFO: Validating parents for backup {0}'.format(
|
|
validate_id), output,
|
|
'\n Unexpected Output Message: {0}\n'.format(
|
|
repr(output)))
|
|
|
|
self.assertIn(
|
|
'INFO: Validating backup {0}'.format(
|
|
self.show_pb(backup_dir, 'node')[2]['id']), output,
|
|
'\n Unexpected Output Message: {0}\n'.format(
|
|
repr(output)))
|
|
|
|
self.assertIn(
|
|
'INFO: Validating backup {0}'.format(
|
|
self.show_pb(backup_dir, 'node')[3]['id']), output,
|
|
'\n Unexpected Output Message: {0}\n'.format(
|
|
repr(output)))
|
|
|
|
self.assertIn(
|
|
'INFO: Revalidating backup {0}'.format(
|
|
corrupt_id), output,
|
|
'\n Unexpected Output Message: {0}\n'.format(
|
|
repr(output)))
|
|
|
|
self.assertIn(
|
|
'Backup {0} data files are valid'.format(
|
|
corrupt_id), output,
|
|
'\n Unexpected Output Message: {0}\n'.format(
|
|
repr(output)))
|
|
|
|
self.assertIn(
|
|
'INFO: Revalidating backup {0}'.format(
|
|
self.show_pb(backup_dir, 'node')[5]['id']), output,
|
|
'\n Unexpected Output Message: {0}\n'.format(
|
|
repr(output)))
|
|
|
|
self.assertIn(
|
|
'Backup {0} data files are valid'.format(
|
|
self.show_pb(backup_dir, 'node')[5]['id']), output,
|
|
'\n Unexpected Output Message: {0}\n'.format(
|
|
repr(output)))
|
|
|
|
self.assertIn(
|
|
'INFO: Revalidating backup {0}'.format(
|
|
validate_id), output,
|
|
'\n Unexpected Output Message: {0}\n'.format(
|
|
repr(output)))
|
|
|
|
self.assertIn(
|
|
'Backup {0} data files are valid'.format(
|
|
validate_id), output,
|
|
'\n Unexpected Output Message: {0}\n'.format(
|
|
repr(output)))
|
|
|
|
self.assertIn(
|
|
'INFO: Backup {0} WAL segments are valid'.format(
|
|
validate_id), output,
|
|
'\n Unexpected Output Message: {0}\n'.format(
|
|
repr(output)))
|
|
|
|
self.assertIn(
|
|
'INFO: Backup {0} is valid.'.format(
|
|
validate_id), output,
|
|
'\n Unexpected Output Message: {0}\n'.format(
|
|
repr(output)))
|
|
|
|
self.assertIn(
|
|
'INFO: Validate of backup {0} completed.'.format(
|
|
validate_id), output,
|
|
'\n Unexpected Output Message: {0}\n'.format(
|
|
repr(output)))
|
|
|
|
# Now we have two perfectly valid backup chains based on FULL2
|
|
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[9]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[8]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[7]['status'] == 'ORPHAN')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[6]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[5]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[4]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[3]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[2]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[1]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[0]['status'] == 'OK')
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
# @unittest.skip("skip")
|
|
def test_validate_corrupted_full_missing(self):
|
|
"""
|
|
make node with archiving, take full backup, and three page backups,
|
|
take another full backup and four page backups
|
|
corrupt second full backup, run validate, check that
|
|
second full backup became CORRUPT and his page backups are ORPHANs
|
|
remove corruption from full backup and remove his second page backup
|
|
run valudate again, check that
|
|
second full backup and his firts page backups are OK,
|
|
third page should be ORPHAN
|
|
"""
|
|
fname = self.id().split('.')[3]
|
|
node = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node'),
|
|
set_replication=True,
|
|
initdb_params=['--data-checksums'])
|
|
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
self.set_archiving(backup_dir, 'node', node)
|
|
node.slow_start()
|
|
|
|
self.backup_node(backup_dir, 'node', node)
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
|
|
backup_id = self.backup_node(backup_dir, 'node', node)
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
backup_id_page = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
|
|
file = os.path.join(
|
|
backup_dir, 'backups', 'node',
|
|
backup_id, 'database', 'postgresql.auto.conf')
|
|
|
|
file_new = os.path.join(backup_dir, 'postgresql.auto.conf')
|
|
os.rename(file, file_new)
|
|
|
|
try:
|
|
self.validate_pb(backup_dir, options=["-j", "4"])
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because of data file dissapearance.\n "
|
|
"Output: {0} \n CMD: {1}".format(
|
|
self.output, self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertIn(
|
|
'Validating backup {0}'.format(backup_id), e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertIn(
|
|
'WARNING: Backup {0} data files are corrupted'.format(
|
|
backup_id), e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertIn(
|
|
'WARNING: Backup {0} is orphaned because his parent {1} has status: CORRUPT'.format(
|
|
self.show_pb(backup_dir, 'node')[5]['id'], backup_id), e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[8]['status'] == 'ORPHAN')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[7]['status'] == 'ORPHAN')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[6]['status'] == 'ORPHAN')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[5]['status'] == 'ORPHAN')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[4]['status'] == 'CORRUPT')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[3]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[2]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[1]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[0]['status'] == 'OK')
|
|
|
|
# Full backup is fixed
|
|
os.rename(file_new, file)
|
|
|
|
# break PAGE
|
|
old_directory = os.path.join(
|
|
backup_dir, 'backups', 'node', backup_id_page)
|
|
new_directory = os.path.join(backup_dir, backup_id_page)
|
|
os.rename(old_directory, new_directory)
|
|
|
|
try:
|
|
self.validate_pb(backup_dir, options=["-j", "4"])
|
|
except ProbackupException as e:
|
|
self.assertIn(
|
|
'WARNING: Some backups are not valid', e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertIn(
|
|
'WARNING: Backup {0} has missing parent {1}'.format(
|
|
self.show_pb(backup_dir, 'node')[7]['id'],
|
|
backup_id_page),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertIn(
|
|
'WARNING: Backup {0} has missing parent {1}'.format(
|
|
self.show_pb(backup_dir, 'node')[6]['id'],
|
|
backup_id_page),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertIn(
|
|
'WARNING: Backup {0} has parent {1} with status: CORRUPT'.format(
|
|
self.show_pb(backup_dir, 'node')[5]['id'], backup_id),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[7]['status'] == 'ORPHAN')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[6]['status'] == 'ORPHAN')
|
|
# missing backup is here
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[5]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[4]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[3]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[2]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[1]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[0]['status'] == 'OK')
|
|
|
|
# validate should be idempotent - user running validate
|
|
# second time must be provided with ID of missing backup
|
|
|
|
try:
|
|
self.validate_pb(backup_dir, options=["-j", "4"])
|
|
except ProbackupException as e:
|
|
self.assertIn(
|
|
'WARNING: Some backups are not valid', e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertIn(
|
|
'WARNING: Backup {0} has missing parent {1}'.format(
|
|
self.show_pb(backup_dir, 'node')[7]['id'],
|
|
backup_id_page), e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertIn(
|
|
'WARNING: Backup {0} has missing parent {1}'.format(
|
|
self.show_pb(backup_dir, 'node')[6]['id'],
|
|
backup_id_page), e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[7]['status'] == 'ORPHAN')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[6]['status'] == 'ORPHAN')
|
|
# missing backup is here
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[5]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[4]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[3]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[2]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[1]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[0]['status'] == 'OK')
|
|
|
|
# fix missing PAGE backup
|
|
os.rename(new_directory, old_directory)
|
|
# exit(1)
|
|
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[8]['status'] == 'ORPHAN')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[7]['status'] == 'ORPHAN')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[6]['status'] == 'ORPHAN')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[5]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[4]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[3]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[2]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[1]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[0]['status'] == 'OK')
|
|
|
|
output = self.validate_pb(backup_dir, options=["-j", "4"])
|
|
|
|
self.assertIn(
|
|
'INFO: All backups are valid',
|
|
output,
|
|
'\n Unexpected Error Message: {0}\n'.format(
|
|
repr(output)))
|
|
|
|
self.assertIn(
|
|
'WARNING: Backup {0} has parent {1} with status: ORPHAN'.format(
|
|
self.show_pb(backup_dir, 'node')[8]['id'],
|
|
self.show_pb(backup_dir, 'node')[6]['id']),
|
|
output,
|
|
'\n Unexpected Error Message: {0}\n'.format(
|
|
repr(output)))
|
|
|
|
self.assertIn(
|
|
'WARNING: Backup {0} has parent {1} with status: ORPHAN'.format(
|
|
self.show_pb(backup_dir, 'node')[7]['id'],
|
|
self.show_pb(backup_dir, 'node')[6]['id']),
|
|
output,
|
|
'\n Unexpected Error Message: {0}\n'.format(
|
|
repr(output)))
|
|
|
|
self.assertIn(
|
|
'Revalidating backup {0}'.format(
|
|
self.show_pb(backup_dir, 'node')[6]['id']),
|
|
output,
|
|
'\n Unexpected Error Message: {0}\n'.format(
|
|
repr(output)))
|
|
|
|
self.assertIn(
|
|
'Revalidating backup {0}'.format(
|
|
self.show_pb(backup_dir, 'node')[7]['id']),
|
|
output,
|
|
'\n Unexpected Error Message: {0}\n'.format(
|
|
repr(output)))
|
|
|
|
self.assertIn(
|
|
'Revalidating backup {0}'.format(
|
|
self.show_pb(backup_dir, 'node')[8]['id']),
|
|
output,
|
|
'\n Unexpected Error Message: {0}\n'.format(
|
|
repr(output)))
|
|
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[8]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[7]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[6]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[5]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[4]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[3]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[2]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[1]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[0]['status'] == 'OK')
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
def test_file_size_corruption_no_validate(self):
|
|
|
|
fname = self.id().split('.')[3]
|
|
node = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node'),
|
|
# initdb_params=['--data-checksums'],
|
|
)
|
|
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
self.set_archiving(backup_dir, 'node', node)
|
|
|
|
node.slow_start()
|
|
|
|
node.safe_psql(
|
|
"postgres",
|
|
"create table t_heap as select 1 as id, md5(i::text) as text, "
|
|
"md5(repeat(i::text,10))::tsvector as tsvector "
|
|
"from generate_series(0,1000) i")
|
|
node.safe_psql(
|
|
"postgres",
|
|
"CHECKPOINT;")
|
|
|
|
heap_path = node.safe_psql(
|
|
"postgres",
|
|
"select pg_relation_filepath('t_heap')").decode('utf-8').rstrip()
|
|
heap_size = node.safe_psql(
|
|
"postgres",
|
|
"select pg_relation_size('t_heap')")
|
|
|
|
backup_id = self.backup_node(
|
|
backup_dir, 'node', node, backup_type="full",
|
|
options=["-j", "4"], asynchronous=False, gdb=False)
|
|
|
|
node.stop()
|
|
node.cleanup()
|
|
|
|
# Let`s do file corruption
|
|
with open(
|
|
os.path.join(
|
|
backup_dir, "backups", 'node', backup_id,
|
|
"database", heap_path), "rb+", 0) as f:
|
|
f.truncate(int(heap_size) - 4096)
|
|
f.flush()
|
|
f.close
|
|
|
|
node.cleanup()
|
|
|
|
try:
|
|
self.restore_node(
|
|
backup_dir, 'node', node,
|
|
options=["--no-validate"])
|
|
except ProbackupException as e:
|
|
self.assertTrue(
|
|
"ERROR: Backup files restoring failed" in e.message,
|
|
repr(e.message))
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
# @unittest.skip("skip")
|
|
def test_validate_specific_backup_with_missing_backup(self):
|
|
"""
|
|
PAGE3_2
|
|
PAGE3_1
|
|
FULL3
|
|
PAGE2_5
|
|
PAGE2_4 <- validate
|
|
PAGE2_3
|
|
PAGE2_2 <- missing
|
|
PAGE2_1
|
|
FULL2
|
|
PAGE1_2
|
|
PAGE1_1
|
|
FULL1
|
|
"""
|
|
fname = self.id().split('.')[3]
|
|
node = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node'),
|
|
set_replication=True,
|
|
initdb_params=['--data-checksums'])
|
|
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
self.set_archiving(backup_dir, 'node', node)
|
|
node.slow_start()
|
|
|
|
# CHAIN1
|
|
self.backup_node(backup_dir, 'node', node)
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
|
|
# CHAIN2
|
|
self.backup_node(backup_dir, 'node', node)
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
missing_id = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
validate_id = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
|
|
# CHAIN3
|
|
self.backup_node(backup_dir, 'node', node)
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
|
|
old_directory = os.path.join(backup_dir, 'backups', 'node', missing_id)
|
|
new_directory = os.path.join(backup_dir, missing_id)
|
|
|
|
os.rename(old_directory, new_directory)
|
|
|
|
try:
|
|
self.validate_pb(backup_dir, 'node', validate_id,
|
|
options=["-j", "4"])
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because of backup dissapearance.\n "
|
|
"Output: {0} \n CMD: {1}".format(
|
|
self.output, self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertIn(
|
|
'WARNING: Backup {0} is orphaned because his parent {1} is missing'.format(
|
|
self.show_pb(backup_dir, 'node')[7]['id'], missing_id),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertIn(
|
|
'WARNING: Backup {0} is orphaned because his parent {1} is missing'.format(
|
|
self.show_pb(backup_dir, 'node')[6]['id'], missing_id),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertIn(
|
|
'WARNING: Backup {0} is orphaned because his parent {1} is missing'.format(
|
|
self.show_pb(backup_dir, 'node')[5]['id'], missing_id),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[10]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[9]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[8]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[7]['status'] == 'ORPHAN')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[6]['status'] == 'ORPHAN')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[5]['status'] == 'ORPHAN')
|
|
# missing backup
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[4]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[3]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[2]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[1]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[0]['status'] == 'OK')
|
|
|
|
try:
|
|
self.validate_pb(backup_dir, 'node', validate_id,
|
|
options=["-j", "4"])
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because of backup dissapearance.\n "
|
|
"Output: {0} \n CMD: {1}".format(
|
|
self.output, self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertIn(
|
|
'WARNING: Backup {0} has missing parent {1}'.format(
|
|
self.show_pb(backup_dir, 'node')[7]['id'], missing_id),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertIn(
|
|
'WARNING: Backup {0} has missing parent {1}'.format(
|
|
self.show_pb(backup_dir, 'node')[6]['id'], missing_id),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertIn(
|
|
'WARNING: Backup {0} has missing parent {1}'.format(
|
|
self.show_pb(backup_dir, 'node')[5]['id'], missing_id),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
os.rename(new_directory, old_directory)
|
|
|
|
# Revalidate backup chain
|
|
self.validate_pb(backup_dir, 'node', validate_id, options=["-j", "4"])
|
|
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[11]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[10]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[9]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[8]['status'] == 'ORPHAN')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[7]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[6]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[5]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[4]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[3]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[2]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[1]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[0]['status'] == 'OK')
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
# @unittest.skip("skip")
|
|
def test_validate_specific_backup_with_missing_backup_1(self):
|
|
"""
|
|
PAGE3_2
|
|
PAGE3_1
|
|
FULL3
|
|
PAGE2_5
|
|
PAGE2_4 <- validate
|
|
PAGE2_3
|
|
PAGE2_2 <- missing
|
|
PAGE2_1
|
|
FULL2 <- missing
|
|
PAGE1_2
|
|
PAGE1_1
|
|
FULL1
|
|
"""
|
|
fname = self.id().split('.')[3]
|
|
node = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node'),
|
|
set_replication=True,
|
|
initdb_params=['--data-checksums'])
|
|
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
self.set_archiving(backup_dir, 'node', node)
|
|
node.slow_start()
|
|
|
|
# CHAIN1
|
|
self.backup_node(backup_dir, 'node', node)
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
|
|
# CHAIN2
|
|
missing_full_id = self.backup_node(backup_dir, 'node', node)
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
missing_page_id = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
validate_id = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
|
|
# CHAIN3
|
|
self.backup_node(backup_dir, 'node', node)
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
|
|
page_old_directory = os.path.join(
|
|
backup_dir, 'backups', 'node', missing_page_id)
|
|
page_new_directory = os.path.join(backup_dir, missing_page_id)
|
|
os.rename(page_old_directory, page_new_directory)
|
|
|
|
full_old_directory = os.path.join(
|
|
backup_dir, 'backups', 'node', missing_full_id)
|
|
full_new_directory = os.path.join(backup_dir, missing_full_id)
|
|
os.rename(full_old_directory, full_new_directory)
|
|
|
|
try:
|
|
self.validate_pb(backup_dir, 'node', validate_id,
|
|
options=["-j", "4"])
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because of backup dissapearance.\n "
|
|
"Output: {0} \n CMD: {1}".format(
|
|
self.output, self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertIn(
|
|
'WARNING: Backup {0} is orphaned because his parent {1} is missing'.format(
|
|
self.show_pb(backup_dir, 'node')[6]['id'], missing_page_id),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertIn(
|
|
'WARNING: Backup {0} is orphaned because his parent {1} is missing'.format(
|
|
self.show_pb(backup_dir, 'node')[5]['id'], missing_page_id),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertIn(
|
|
'WARNING: Backup {0} is orphaned because his parent {1} is missing'.format(
|
|
self.show_pb(backup_dir, 'node')[4]['id'], missing_page_id),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[9]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[8]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[7]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[6]['status'] == 'ORPHAN')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[5]['status'] == 'ORPHAN')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[4]['status'] == 'ORPHAN')
|
|
# PAGE2_1
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[3]['status'] == 'OK') # <- SHit
|
|
# FULL2
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[2]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[1]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[0]['status'] == 'OK')
|
|
|
|
os.rename(page_new_directory, page_old_directory)
|
|
os.rename(full_new_directory, full_old_directory)
|
|
|
|
# Revalidate backup chain
|
|
self.validate_pb(backup_dir, 'node', validate_id, options=["-j", "4"])
|
|
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[11]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[10]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[9]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[8]['status'] == 'ORPHAN') # <- Fail
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[7]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[6]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[5]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[4]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[3]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[2]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[1]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[0]['status'] == 'OK')
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
# @unittest.skip("skip")
|
|
def test_validate_with_missing_backup_1(self):
|
|
"""
|
|
PAGE3_2
|
|
PAGE3_1
|
|
FULL3
|
|
PAGE2_5
|
|
PAGE2_4 <- validate
|
|
PAGE2_3
|
|
PAGE2_2 <- missing
|
|
PAGE2_1
|
|
FULL2 <- missing
|
|
PAGE1_2
|
|
PAGE1_1
|
|
FULL1
|
|
"""
|
|
fname = self.id().split('.')[3]
|
|
node = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node'),
|
|
set_replication=True,
|
|
initdb_params=['--data-checksums'])
|
|
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
self.set_archiving(backup_dir, 'node', node)
|
|
node.slow_start()
|
|
|
|
# CHAIN1
|
|
self.backup_node(backup_dir, 'node', node)
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
|
|
# CHAIN2
|
|
missing_full_id = self.backup_node(backup_dir, 'node', node)
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
missing_page_id = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
validate_id = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
|
|
# CHAIN3
|
|
self.backup_node(backup_dir, 'node', node)
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
|
|
# Break PAGE
|
|
page_old_directory = os.path.join(
|
|
backup_dir, 'backups', 'node', missing_page_id)
|
|
page_new_directory = os.path.join(backup_dir, missing_page_id)
|
|
os.rename(page_old_directory, page_new_directory)
|
|
|
|
# Break FULL
|
|
full_old_directory = os.path.join(
|
|
backup_dir, 'backups', 'node', missing_full_id)
|
|
full_new_directory = os.path.join(backup_dir, missing_full_id)
|
|
os.rename(full_old_directory, full_new_directory)
|
|
|
|
try:
|
|
self.validate_pb(backup_dir, 'node', validate_id,
|
|
options=["-j", "4"])
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because of backup dissapearance.\n "
|
|
"Output: {0} \n CMD: {1}".format(
|
|
self.output, self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertIn(
|
|
'WARNING: Backup {0} is orphaned because his parent {1} is missing'.format(
|
|
self.show_pb(backup_dir, 'node')[6]['id'], missing_page_id),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertIn(
|
|
'WARNING: Backup {0} is orphaned because his parent {1} is missing'.format(
|
|
self.show_pb(backup_dir, 'node')[5]['id'], missing_page_id),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertIn(
|
|
'WARNING: Backup {0} is orphaned because his parent {1} is missing'.format(
|
|
self.show_pb(backup_dir, 'node')[4]['id'], missing_page_id),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[9]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[8]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[7]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[6]['status'] == 'ORPHAN')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[5]['status'] == 'ORPHAN')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[4]['status'] == 'ORPHAN')
|
|
# PAGE2_2 is missing
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[3]['status'] == 'OK')
|
|
# FULL1 - is missing
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[2]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[1]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[0]['status'] == 'OK')
|
|
|
|
os.rename(page_new_directory, page_old_directory)
|
|
|
|
# Revalidate backup chain
|
|
try:
|
|
self.validate_pb(backup_dir, 'node', validate_id,
|
|
options=["-j", "4"])
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because of backup dissapearance.\n "
|
|
"Output: {0} \n CMD: {1}".format(
|
|
self.output, self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertIn(
|
|
'WARNING: Backup {0} has status: ORPHAN'.format(
|
|
validate_id),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertIn(
|
|
'WARNING: Backup {0} has missing parent {1}'.format(
|
|
self.show_pb(backup_dir, 'node')[7]['id'],
|
|
missing_full_id),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertIn(
|
|
'WARNING: Backup {0} has missing parent {1}'.format(
|
|
self.show_pb(backup_dir, 'node')[6]['id'],
|
|
missing_full_id),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertIn(
|
|
'WARNING: Backup {0} has missing parent {1}'.format(
|
|
self.show_pb(backup_dir, 'node')[5]['id'],
|
|
missing_full_id),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertIn(
|
|
'WARNING: Backup {0} is orphaned because his parent {1} is missing'.format(
|
|
self.show_pb(backup_dir, 'node')[4]['id'],
|
|
missing_full_id),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertIn(
|
|
'WARNING: Backup {0} is orphaned because his parent {1} is missing'.format(
|
|
self.show_pb(backup_dir, 'node')[3]['id'],
|
|
missing_full_id),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[10]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[9]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[8]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[7]['status'] == 'ORPHAN')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[6]['status'] == 'ORPHAN')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[5]['status'] == 'ORPHAN')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[4]['status'] == 'ORPHAN')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[3]['status'] == 'ORPHAN')
|
|
# FULL1 - is missing
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[2]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[1]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[0]['status'] == 'OK')
|
|
|
|
os.rename(full_new_directory, full_old_directory)
|
|
|
|
# Revalidate chain
|
|
self.validate_pb(backup_dir, 'node', validate_id, options=["-j", "4"])
|
|
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[11]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[10]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[9]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[8]['status'] == 'ORPHAN')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[7]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[6]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[5]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[4]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[3]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[2]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[1]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[0]['status'] == 'OK')
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
# @unittest.skip("skip")
|
|
def test_validate_with_missing_backup_2(self):
|
|
"""
|
|
PAGE3_2
|
|
PAGE3_1
|
|
FULL3
|
|
PAGE2_5
|
|
PAGE2_4
|
|
PAGE2_3
|
|
PAGE2_2 <- missing
|
|
PAGE2_1
|
|
FULL2 <- missing
|
|
PAGE1_2
|
|
PAGE1_1
|
|
FULL1
|
|
"""
|
|
fname = self.id().split('.')[3]
|
|
node = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node'),
|
|
set_replication=True,
|
|
initdb_params=['--data-checksums'])
|
|
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
self.set_archiving(backup_dir, 'node', node)
|
|
node.slow_start()
|
|
|
|
# CHAIN1
|
|
self.backup_node(backup_dir, 'node', node)
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
|
|
# CHAIN2
|
|
missing_full_id = self.backup_node(backup_dir, 'node', node)
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
missing_page_id = self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
self.backup_node(
|
|
backup_dir, 'node', node, backup_type='page')
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
|
|
# CHAIN3
|
|
self.backup_node(backup_dir, 'node', node)
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page')
|
|
|
|
page_old_directory = os.path.join(backup_dir, 'backups', 'node', missing_page_id)
|
|
page_new_directory = os.path.join(backup_dir, missing_page_id)
|
|
os.rename(page_old_directory, page_new_directory)
|
|
|
|
full_old_directory = os.path.join(backup_dir, 'backups', 'node', missing_full_id)
|
|
full_new_directory = os.path.join(backup_dir, missing_full_id)
|
|
os.rename(full_old_directory, full_new_directory)
|
|
|
|
try:
|
|
self.validate_pb(backup_dir, 'node', options=["-j", "4"])
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because of backup dissapearance.\n "
|
|
"Output: {0} \n CMD: {1}".format(
|
|
self.output, self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertIn(
|
|
'WARNING: Backup {0} is orphaned because his parent {1} is missing'.format(
|
|
self.show_pb(backup_dir, 'node')[6]['id'], missing_page_id),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertIn(
|
|
'WARNING: Backup {0} is orphaned because his parent {1} is missing'.format(
|
|
self.show_pb(backup_dir, 'node')[5]['id'], missing_page_id),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertIn(
|
|
'WARNING: Backup {0} is orphaned because his parent {1} is missing'.format(
|
|
self.show_pb(backup_dir, 'node')[4]['id'], missing_page_id),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertIn(
|
|
'WARNING: Backup {0} is orphaned because his parent {1} is missing'.format(
|
|
self.show_pb(backup_dir, 'node')[3]['id'], missing_full_id),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[9]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[8]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[7]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[6]['status'] == 'ORPHAN')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[5]['status'] == 'ORPHAN')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[4]['status'] == 'ORPHAN')
|
|
# PAGE2_2 is missing
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[3]['status'] == 'ORPHAN')
|
|
# FULL1 - is missing
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[2]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[1]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[0]['status'] == 'OK')
|
|
|
|
os.rename(page_new_directory, page_old_directory)
|
|
|
|
# Revalidate backup chain
|
|
try:
|
|
self.validate_pb(backup_dir, 'node', options=["-j", "4"])
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because of backup dissapearance.\n "
|
|
"Output: {0} \n CMD: {1}".format(
|
|
self.output, self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertIn(
|
|
'WARNING: Backup {0} has missing parent {1}'.format(
|
|
self.show_pb(backup_dir, 'node')[7]['id'], missing_full_id),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertIn(
|
|
'WARNING: Backup {0} has missing parent {1}'.format(
|
|
self.show_pb(backup_dir, 'node')[6]['id'], missing_full_id),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertIn(
|
|
'WARNING: Backup {0} has missing parent {1}'.format(
|
|
self.show_pb(backup_dir, 'node')[5]['id'], missing_full_id),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertIn(
|
|
'WARNING: Backup {0} is orphaned because his parent {1} is missing'.format(
|
|
self.show_pb(backup_dir, 'node')[4]['id'], missing_full_id),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
self.assertIn(
|
|
'WARNING: Backup {0} has missing parent {1}'.format(
|
|
self.show_pb(backup_dir, 'node')[3]['id'], missing_full_id),
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[10]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[9]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[8]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[7]['status'] == 'ORPHAN')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[6]['status'] == 'ORPHAN')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[5]['status'] == 'ORPHAN')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[4]['status'] == 'ORPHAN')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[3]['status'] == 'ORPHAN')
|
|
# FULL1 - is missing
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[2]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[1]['status'] == 'OK')
|
|
self.assertTrue(self.show_pb(backup_dir, 'node')[0]['status'] == 'OK')
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
# @unittest.skip("skip")
|
|
def test_corrupt_pg_control_via_resetxlog(self):
|
|
""" PGPRO-2096 """
|
|
fname = self.id().split('.')[3]
|
|
node = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node'),
|
|
set_replication=True,
|
|
initdb_params=['--data-checksums'])
|
|
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
self.set_archiving(backup_dir, 'node', node)
|
|
node.slow_start()
|
|
|
|
backup_id = self.backup_node(backup_dir, 'node', node)
|
|
|
|
if self.get_version(node) < 100000:
|
|
pg_resetxlog_path = self.get_bin_path('pg_resetxlog')
|
|
wal_dir = 'pg_xlog'
|
|
else:
|
|
pg_resetxlog_path = self.get_bin_path('pg_resetwal')
|
|
wal_dir = 'pg_wal'
|
|
|
|
os.mkdir(
|
|
os.path.join(
|
|
backup_dir, 'backups', 'node', backup_id, 'database', wal_dir, 'archive_status'))
|
|
|
|
pg_control_path = os.path.join(
|
|
backup_dir, 'backups', 'node',
|
|
backup_id, 'database', 'global', 'pg_control')
|
|
|
|
md5_before = hashlib.md5(
|
|
open(pg_control_path, 'rb').read()).hexdigest()
|
|
|
|
self.run_binary(
|
|
[
|
|
pg_resetxlog_path,
|
|
'-D',
|
|
os.path.join(backup_dir, 'backups', 'node', backup_id, 'database'),
|
|
'-o 42',
|
|
'-f'
|
|
],
|
|
asynchronous=False)
|
|
|
|
md5_after = hashlib.md5(
|
|
open(pg_control_path, 'rb').read()).hexdigest()
|
|
|
|
if self.verbose:
|
|
print('\n MD5 BEFORE resetxlog: {0}\n MD5 AFTER resetxlog: {1}'.format(
|
|
md5_before, md5_after))
|
|
|
|
# Validate backup
|
|
try:
|
|
self.validate_pb(backup_dir, 'node', options=["-j", "4"])
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because of pg_control change.\n "
|
|
"Output: {0} \n CMD: {1}".format(
|
|
self.output, self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertIn(
|
|
'data files are corrupted',
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
# @unittest.skip("skip")
|
|
def test_validation_after_backup(self):
|
|
""""""
|
|
self._check_gdb_flag_or_skip_test()
|
|
|
|
fname = self.id().split('.')[3]
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
node = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node'),
|
|
set_replication=True,
|
|
initdb_params=['--data-checksums'])
|
|
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
node.slow_start()
|
|
|
|
# FULL backup
|
|
gdb = self.backup_node(
|
|
backup_dir, 'node', node, gdb=True, options=['--stream'])
|
|
|
|
gdb.set_breakpoint('pgBackupValidate')
|
|
gdb.run_until_break()
|
|
|
|
backup_id = self.show_pb(backup_dir, 'node')[0]['id']
|
|
|
|
file = os.path.join(
|
|
backup_dir, "backups", "node", backup_id,
|
|
"database", "postgresql.conf")
|
|
os.remove(file)
|
|
|
|
gdb.continue_execution_until_exit()
|
|
|
|
self.assertEqual(
|
|
'CORRUPT',
|
|
self.show_pb(backup_dir, 'node', backup_id)['status'],
|
|
'Backup STATUS should be "ERROR"')
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
# @unittest.expectedFailure
|
|
# @unittest.skip("skip")
|
|
def test_validate_corrupt_tablespace_map(self):
|
|
"""
|
|
Check that corruption in tablespace_map is detected
|
|
"""
|
|
fname = self.id().split('.')[3]
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
node = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node'),
|
|
set_replication=True,
|
|
initdb_params=['--data-checksums'])
|
|
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
node.slow_start()
|
|
|
|
self.create_tblspace_in_node(node, 'external_dir')
|
|
|
|
node.safe_psql(
|
|
'postgres',
|
|
'CREATE TABLE t_heap(a int) TABLESPACE "external_dir"')
|
|
|
|
# FULL backup
|
|
backup_id = self.backup_node(
|
|
backup_dir, 'node', node, options=['--stream'])
|
|
|
|
tablespace_map = os.path.join(
|
|
backup_dir, 'backups', 'node',
|
|
backup_id, 'database', 'tablespace_map')
|
|
|
|
# Corrupt tablespace_map file in FULL backup
|
|
with open(tablespace_map, "rb+", 0) as f:
|
|
f.seek(84)
|
|
f.write(b"blah")
|
|
f.flush()
|
|
f.close
|
|
|
|
try:
|
|
self.validate_pb(backup_dir, 'node', backup_id=backup_id)
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because tablespace_map is corrupted.\n "
|
|
"Output: {0} \n CMD: {1}".format(
|
|
self.output, self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertIn(
|
|
'WARNING: Invalid CRC of backup file',
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
#TODO fix the test
|
|
@unittest.expectedFailure
|
|
# @unittest.skip("skip")
|
|
def test_validate_target_lsn(self):
|
|
"""
|
|
Check validation to specific LSN
|
|
"""
|
|
fname = self.id().split('.')[3]
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
node = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node'),
|
|
set_replication=True,
|
|
initdb_params=['--data-checksums'])
|
|
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
self.set_archiving(backup_dir, 'node', node)
|
|
node.slow_start()
|
|
|
|
# FULL backup
|
|
self.backup_node(backup_dir, 'node', node)
|
|
|
|
node.safe_psql(
|
|
"postgres",
|
|
"create table t_heap as select 1 as id, md5(i::text) as text, "
|
|
"md5(repeat(i::text,10))::tsvector as tsvector "
|
|
"from generate_series(0,10000) i")
|
|
|
|
node_restored = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node_restored'))
|
|
node_restored.cleanup()
|
|
|
|
self.restore_node(backup_dir, 'node', node_restored)
|
|
|
|
self.set_auto_conf(
|
|
node_restored, {'port': node_restored.port})
|
|
|
|
node_restored.slow_start()
|
|
|
|
self.switch_wal_segment(node)
|
|
|
|
backup_id = self.backup_node(
|
|
backup_dir, 'node', node_restored,
|
|
data_dir=node_restored.data_dir)
|
|
|
|
target_lsn = self.show_pb(backup_dir, 'node')[1]['stop-lsn']
|
|
|
|
self.delete_pb(backup_dir, 'node', backup_id)
|
|
|
|
self.validate_pb(
|
|
backup_dir, 'node',
|
|
options=[
|
|
'--recovery-target-timeline=2',
|
|
'--recovery-target-lsn={0}'.format(target_lsn)])
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
@unittest.skip("skip")
|
|
def test_partial_validate_empty_and_mangled_database_map(self):
|
|
"""
|
|
"""
|
|
fname = self.id().split('.')[3]
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
node = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node'),
|
|
set_replication=True,
|
|
initdb_params=['--data-checksums'])
|
|
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
|
|
node.slow_start()
|
|
|
|
# create databases
|
|
for i in range(1, 10, 1):
|
|
node.safe_psql(
|
|
'postgres',
|
|
'CREATE database db{0}'.format(i))
|
|
|
|
# FULL backup with database_map
|
|
backup_id = self.backup_node(
|
|
backup_dir, 'node', node, options=['--stream'])
|
|
pgdata = self.pgdata_content(node.data_dir)
|
|
|
|
# truncate database_map
|
|
path = os.path.join(
|
|
backup_dir, 'backups', 'node',
|
|
backup_id, 'database', 'database_map')
|
|
with open(path, "w") as f:
|
|
f.close()
|
|
|
|
try:
|
|
self.validate_pb(
|
|
backup_dir, 'node',
|
|
options=["--db-include=db1"])
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because database_map is empty.\n "
|
|
"Output: {0} \n CMD: {1}".format(
|
|
self.output, self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertIn(
|
|
"WARNING: Backup {0} data files are corrupted".format(
|
|
backup_id), e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
# mangle database_map
|
|
with open(path, "w") as f:
|
|
f.write("42")
|
|
f.close()
|
|
|
|
try:
|
|
self.validate_pb(
|
|
backup_dir, 'node',
|
|
options=["--db-include=db1"])
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because database_map is empty.\n "
|
|
"Output: {0} \n CMD: {1}".format(
|
|
self.output, self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertIn(
|
|
"WARNING: Backup {0} data files are corrupted".format(
|
|
backup_id), e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
@unittest.skip("skip")
|
|
def test_partial_validate_exclude(self):
|
|
""""""
|
|
fname = self.id().split('.')[3]
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
node = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node'),
|
|
initdb_params=['--data-checksums'])
|
|
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
self.set_archiving(backup_dir, 'node', node)
|
|
node.slow_start()
|
|
|
|
for i in range(1, 10, 1):
|
|
node.safe_psql(
|
|
'postgres',
|
|
'CREATE database db{0}'.format(i))
|
|
|
|
# FULL backup
|
|
backup_id = self.backup_node(backup_dir, 'node', node)
|
|
|
|
try:
|
|
self.validate_pb(
|
|
backup_dir, 'node',
|
|
options=[
|
|
"--db-include=db1",
|
|
"--db-exclude=db2"])
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because of 'db-exclude' and 'db-include'.\n "
|
|
"Output: {0} \n CMD: {1}".format(
|
|
self.output, self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertIn(
|
|
"ERROR: You cannot specify '--db-include' "
|
|
"and '--db-exclude' together", e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
try:
|
|
self.validate_pb(
|
|
backup_dir, 'node',
|
|
options=[
|
|
"--db-exclude=db1",
|
|
"--db-exclude=db5",
|
|
"--log-level-console=verbose"])
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because of missing backup ID.\n "
|
|
"Output: {0} \n CMD: {1}".format(
|
|
self.output, self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertIn(
|
|
"ERROR: You must specify parameter (-i, --backup-id) for partial validation",
|
|
e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
output = self.validate_pb(
|
|
backup_dir, 'node', backup_id,
|
|
options=[
|
|
"--db-exclude=db1",
|
|
"--db-exclude=db5",
|
|
"--log-level-console=verbose"])
|
|
|
|
self.assertIn(
|
|
"VERBOSE: Skip file validation due to partial restore", output)
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
@unittest.skip("skip")
|
|
def test_partial_validate_include(self):
|
|
"""
|
|
"""
|
|
fname = self.id().split('.')[3]
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
node = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node'),
|
|
initdb_params=['--data-checksums'])
|
|
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
self.set_archiving(backup_dir, 'node', node)
|
|
node.slow_start()
|
|
|
|
for i in range(1, 10, 1):
|
|
node.safe_psql(
|
|
'postgres',
|
|
'CREATE database db{0}'.format(i))
|
|
|
|
# FULL backup
|
|
backup_id = self.backup_node(backup_dir, 'node', node)
|
|
|
|
try:
|
|
self.validate_pb(
|
|
backup_dir, 'node',
|
|
options=[
|
|
"--db-include=db1",
|
|
"--db-exclude=db2"])
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because of 'db-exclude' and 'db-include'.\n "
|
|
"Output: {0} \n CMD: {1}".format(
|
|
self.output, self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertIn(
|
|
"ERROR: You cannot specify '--db-include' "
|
|
"and '--db-exclude' together", e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
output = self.validate_pb(
|
|
backup_dir, 'node', backup_id,
|
|
options=[
|
|
"--db-include=db1",
|
|
"--db-include=db5",
|
|
"--db-include=postgres",
|
|
"--log-level-console=verbose"])
|
|
|
|
self.assertIn(
|
|
"VERBOSE: Skip file validation due to partial restore", output)
|
|
|
|
output = self.validate_pb(
|
|
backup_dir, 'node', backup_id,
|
|
options=["--log-level-console=verbose"])
|
|
|
|
self.assertNotIn(
|
|
"VERBOSE: Skip file validation due to partial restore", output)
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
# @unittest.skip("skip")
|
|
def test_not_validate_diffenent_pg_version(self):
|
|
"""Do not validate backup, if binary is compiled with different PG version"""
|
|
fname = self.id().split('.')[3]
|
|
node = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node'),
|
|
initdb_params=['--data-checksums'])
|
|
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
self.set_archiving(backup_dir, 'node', node)
|
|
node.slow_start()
|
|
|
|
backup_id = self.backup_node(backup_dir, 'node', node)
|
|
|
|
control_file = os.path.join(
|
|
backup_dir, "backups", "node", backup_id,
|
|
"backup.control")
|
|
|
|
pg_version = node.major_version
|
|
|
|
if pg_version.is_integer():
|
|
pg_version = int(pg_version)
|
|
|
|
fake_new_pg_version = pg_version + 1
|
|
|
|
with open(control_file, 'r') as f:
|
|
data = f.read();
|
|
|
|
data = data.replace(
|
|
"server-version = {0}".format(str(pg_version)),
|
|
"server-version = {0}".format(str(fake_new_pg_version)))
|
|
|
|
with open(control_file, 'w') as f:
|
|
f.write(data);
|
|
|
|
try:
|
|
self.validate_pb(backup_dir)
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because validation is forbidden if server version of backup "
|
|
"is different from the server version of pg_probackup.\n Output: {0} \n CMD: {1}".format(
|
|
repr(self.output), self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertIn(
|
|
"ERROR: Backup {0} has server version".format(backup_id),
|
|
e.message,
|
|
"\n Unexpected Error Message: {0}\n CMD: {1}".format(
|
|
repr(e.message), self.cmd))
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
# @unittest.expectedFailure
|
|
# @unittest.skip("skip")
|
|
def test_validate_corrupt_page_header_map(self):
|
|
"""
|
|
Check that corruption in page_header_map is detected
|
|
"""
|
|
fname = self.id().split('.')[3]
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
node = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node'),
|
|
set_replication=True,
|
|
initdb_params=['--data-checksums'])
|
|
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
node.slow_start()
|
|
|
|
ok_1 = self.backup_node(backup_dir, 'node', node, options=['--stream'])
|
|
|
|
# FULL backup
|
|
backup_id = self.backup_node(
|
|
backup_dir, 'node', node, options=['--stream'])
|
|
|
|
ok_2 = self.backup_node(backup_dir, 'node', node, options=['--stream'])
|
|
|
|
page_header_map = os.path.join(
|
|
backup_dir, 'backups', 'node', backup_id, 'page_header_map')
|
|
|
|
# Corrupt tablespace_map file in FULL backup
|
|
with open(page_header_map, "rb+", 0) as f:
|
|
f.seek(42)
|
|
f.write(b"blah")
|
|
f.flush()
|
|
f.close
|
|
|
|
try:
|
|
self.validate_pb(backup_dir, 'node', backup_id=backup_id)
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because page_header is corrupted.\n "
|
|
"Output: {0} \n CMD: {1}".format(
|
|
self.output, self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertRegex(
|
|
e.message,
|
|
r'WARNING: An error occured during metadata decompression for file "[\w/]+": (data|buffer) error',
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertIn("Backup {0} is corrupt".format(backup_id), e.message)
|
|
|
|
try:
|
|
self.validate_pb(backup_dir)
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because page_header is corrupted.\n "
|
|
"Output: {0} \n CMD: {1}".format(
|
|
self.output, self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertTrue(
|
|
'WARNING: An error occured during metadata decompression' in e.message and
|
|
'data error' in e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
self.assertIn("INFO: Backup {0} data files are valid".format(ok_1), e.message)
|
|
self.assertIn("WARNING: Backup {0} data files are corrupted".format(backup_id), e.message)
|
|
self.assertIn("INFO: Backup {0} data files are valid".format(ok_2), e.message)
|
|
|
|
self.assertIn("WARNING: Some backups are not valid", e.message)
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
# @unittest.expectedFailure
|
|
# @unittest.skip("skip")
|
|
def test_validate_truncated_page_header_map(self):
|
|
"""
|
|
Check that corruption in page_header_map is detected
|
|
"""
|
|
fname = self.id().split('.')[3]
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
node = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node'),
|
|
set_replication=True,
|
|
initdb_params=['--data-checksums'])
|
|
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
node.slow_start()
|
|
|
|
ok_1 = self.backup_node(backup_dir, 'node', node, options=['--stream'])
|
|
|
|
# FULL backup
|
|
backup_id = self.backup_node(
|
|
backup_dir, 'node', node, options=['--stream'])
|
|
|
|
ok_2 = self.backup_node(backup_dir, 'node', node, options=['--stream'])
|
|
|
|
page_header_map = os.path.join(
|
|
backup_dir, 'backups', 'node', backup_id, 'page_header_map')
|
|
|
|
# truncate page_header_map file
|
|
with open(page_header_map, "rb+", 0) as f:
|
|
f.truncate(121)
|
|
f.flush()
|
|
f.close
|
|
|
|
try:
|
|
self.validate_pb(backup_dir, 'node', backup_id=backup_id)
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because page_header is corrupted.\n "
|
|
"Output: {0} \n CMD: {1}".format(
|
|
self.output, self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertIn(
|
|
'ERROR: Backup {0} is corrupt'.format(backup_id), e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
try:
|
|
self.validate_pb(backup_dir)
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because page_header is corrupted.\n "
|
|
"Output: {0} \n CMD: {1}".format(
|
|
self.output, self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertIn("INFO: Backup {0} data files are valid".format(ok_1), e.message)
|
|
self.assertIn("WARNING: Backup {0} data files are corrupted".format(backup_id), e.message)
|
|
self.assertIn("INFO: Backup {0} data files are valid".format(ok_2), e.message)
|
|
self.assertIn("WARNING: Some backups are not valid", e.message)
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
# @unittest.expectedFailure
|
|
# @unittest.skip("skip")
|
|
def test_validate_missing_page_header_map(self):
|
|
"""
|
|
Check that corruption in page_header_map is detected
|
|
"""
|
|
fname = self.id().split('.')[3]
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
node = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node'),
|
|
set_replication=True,
|
|
initdb_params=['--data-checksums'])
|
|
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
node.slow_start()
|
|
|
|
ok_1 = self.backup_node(backup_dir, 'node', node, options=['--stream'])
|
|
|
|
# FULL backup
|
|
backup_id = self.backup_node(
|
|
backup_dir, 'node', node, options=['--stream'])
|
|
|
|
ok_2 = self.backup_node(backup_dir, 'node', node, options=['--stream'])
|
|
|
|
page_header_map = os.path.join(
|
|
backup_dir, 'backups', 'node', backup_id, 'page_header_map')
|
|
|
|
# unlink page_header_map file
|
|
os.remove(page_header_map)
|
|
|
|
try:
|
|
self.validate_pb(backup_dir, 'node', backup_id=backup_id)
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because page_header is corrupted.\n "
|
|
"Output: {0} \n CMD: {1}".format(
|
|
self.output, self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertIn(
|
|
'ERROR: Backup {0} is corrupt'.format(backup_id), e.message,
|
|
'\n Unexpected Error Message: {0}\n CMD: {1}'.format(
|
|
repr(e.message), self.cmd))
|
|
|
|
try:
|
|
self.validate_pb(backup_dir)
|
|
self.assertEqual(
|
|
1, 0,
|
|
"Expecting Error because page_header is corrupted.\n "
|
|
"Output: {0} \n CMD: {1}".format(
|
|
self.output, self.cmd))
|
|
except ProbackupException as e:
|
|
self.assertIn("INFO: Backup {0} data files are valid".format(ok_1), e.message)
|
|
self.assertIn("WARNING: Backup {0} data files are corrupted".format(backup_id), e.message)
|
|
self.assertIn("INFO: Backup {0} data files are valid".format(ok_2), e.message)
|
|
self.assertIn("WARNING: Some backups are not valid", e.message)
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
# @unittest.expectedFailure
|
|
# @unittest.skip("skip")
|
|
def test_no_validate_tablespace_map(self):
|
|
"""
|
|
Check that --no-validate is propagated to tablespace_map
|
|
"""
|
|
fname = self.id().split('.')[3]
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
node = self.make_simple_node(
|
|
base_dir=os.path.join(module_name, fname, 'node'),
|
|
set_replication=True,
|
|
initdb_params=['--data-checksums'])
|
|
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
node.slow_start()
|
|
|
|
self.create_tblspace_in_node(node, 'external_dir')
|
|
|
|
node.safe_psql(
|
|
'postgres',
|
|
'CREATE TABLE t_heap(a int) TABLESPACE "external_dir"')
|
|
|
|
tblspace_new = self.get_tblspace_path(node, 'external_dir_new')
|
|
|
|
oid = node.safe_psql(
|
|
'postgres',
|
|
"select oid from pg_tablespace where spcname = 'external_dir'").decode('utf-8').rstrip()
|
|
|
|
# FULL backup
|
|
backup_id = self.backup_node(
|
|
backup_dir, 'node', node, options=['--stream'])
|
|
|
|
pgdata = self.pgdata_content(node.data_dir)
|
|
|
|
tablespace_map = os.path.join(
|
|
backup_dir, 'backups', 'node',
|
|
backup_id, 'database', 'tablespace_map')
|
|
|
|
# overwrite tablespace_map file
|
|
with open(tablespace_map, "w") as f:
|
|
f.write("{0} {1}".format(oid, tblspace_new))
|
|
f.close
|
|
|
|
node.cleanup()
|
|
|
|
self.restore_node(backup_dir, 'node', node, options=['--no-validate'])
|
|
|
|
pgdata_restored = self.pgdata_content(node.data_dir)
|
|
self.compare_pgdata(pgdata, pgdata_restored)
|
|
|
|
# check that tablespace restore as symlink
|
|
tablespace_link = os.path.join(node.data_dir, 'pg_tblspc', oid)
|
|
self.assertTrue(
|
|
os.path.islink(tablespace_link),
|
|
"'%s' is not a symlink" % tablespace_link)
|
|
|
|
self.assertEqual(
|
|
os.readlink(tablespace_link),
|
|
tblspace_new,
|
|
"Symlink '{0}' do not points to '{1}'".format(tablespace_link, tblspace_new))
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
# validate empty backup list
|
|
# page from future during validate
|
|
# page from future during backup
|
|
|
|
# corrupt block, so file become unaligned:
|
|
# 712 Assert(header.compressed_size <= BLCKSZ);
|
|
# 713
|
|
# 714 read_len = fread(compressed_page.data, 1,
|
|
# 715 MAXALIGN(header.compressed_size), in);
|
|
# 716 if (read_len != MAXALIGN(header.compressed_size))
|
|
# -> 717 elog(ERROR, "cannot read block %u of \"%s\" read %lu of %d",
|
|
# 718 blknum, file->path, read_len, header.compressed_size);
|