mirror of
https://github.com/postgrespro/pg_probackup.git
synced 2024-12-11 11:41:33 +02:00
180 lines
9.3 KiB
Python
180 lines
9.3 KiB
Python
import os
|
|
import unittest
|
|
from .helpers.ptrack_helpers import ProbackupTest, idx_ptrack
|
|
|
|
|
|
module_name = 'ptrack_clean'
|
|
|
|
|
|
class SimpleTest(ProbackupTest, unittest.TestCase):
|
|
|
|
# @unittest.skip("skip")
|
|
# @unittest.expectedFailure
|
|
def test_ptrack_clean(self):
|
|
fname = self.id().split('.')[3]
|
|
node = self.make_simple_node(base_dir="{0}/{1}/node".format(module_name, fname),
|
|
set_replication=True,
|
|
initdb_params=['--data-checksums'],
|
|
pg_options={'ptrack_enable': 'on', 'wal_level': 'replica', 'max_wal_senders': '2'})
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'node', node)
|
|
self.set_archiving(backup_dir, 'node', node)
|
|
node.start()
|
|
|
|
self.create_tblspace_in_node(node, 'somedata')
|
|
|
|
# Create table and indexes
|
|
node.safe_psql(
|
|
"postgres",
|
|
"create sequence t_seq; create table t_heap tablespace somedata as select i as id, nextval('t_seq') as t_seq, md5(i::text) as text, md5(repeat(i::text,10))::tsvector as tsvector from generate_series(0,256) i")
|
|
for i in idx_ptrack:
|
|
if idx_ptrack[i]['type'] != 'heap' and idx_ptrack[i]['type'] != 'seq':
|
|
node.safe_psql("postgres", "create index {0} on {1} using {2}({3}) tablespace somedata".format(
|
|
i, idx_ptrack[i]['relation'], idx_ptrack[i]['type'], idx_ptrack[i]['column']))
|
|
|
|
# Make full backup to clean every ptrack
|
|
self.backup_node(backup_dir, 'node', node, options=['-j100', '--stream'])
|
|
|
|
for i in idx_ptrack:
|
|
# get fork size and calculate it in pages
|
|
idx_ptrack[i]['size'] = self.get_fork_size(node, i)
|
|
# get path to heap and index files
|
|
idx_ptrack[i]['path'] = self.get_fork_path(node, i)
|
|
# get ptrack for every idx
|
|
idx_ptrack[i]['ptrack'] = self.get_ptrack_bits_per_page_for_fork(
|
|
node, idx_ptrack[i]['path'], [idx_ptrack[i]['size']])
|
|
self.check_ptrack_clean(idx_ptrack[i], idx_ptrack[i]['size'])
|
|
|
|
# Update everything, vacuum it and make PTRACK BACKUP
|
|
node.safe_psql('postgres', "update t_heap set t_seq = nextval('t_seq'), text = md5(text), tsvector = md5(repeat(tsvector::text, 10))::tsvector;")
|
|
node.safe_psql('postgres', 'vacuum t_heap')
|
|
|
|
backup_id = self.backup_node(backup_dir, 'node', node, backup_type='ptrack', options=['-j100'])
|
|
node.safe_psql('postgres', 'checkpoint')
|
|
|
|
for i in idx_ptrack:
|
|
# get new size of heap and indexes and calculate it in pages
|
|
idx_ptrack[i]['size'] = self.get_fork_size(node, i)
|
|
# update path to heap and index files in case they`ve changed
|
|
idx_ptrack[i]['path'] = self.get_fork_path(node, i)
|
|
# # get ptrack for every idx
|
|
idx_ptrack[i]['ptrack'] = self.get_ptrack_bits_per_page_for_fork(
|
|
node, idx_ptrack[i]['path'], [idx_ptrack[i]['size']])
|
|
# check that ptrack bits are cleaned
|
|
self.check_ptrack_clean(idx_ptrack[i], idx_ptrack[i]['size'])
|
|
|
|
# Update everything, vacuum it and make PAGE BACKUP
|
|
node.safe_psql('postgres', "update t_heap set t_seq = nextval('t_seq'), text = md5(text), tsvector = md5(repeat(tsvector::text, 10))::tsvector;")
|
|
node.safe_psql('postgres', 'vacuum t_heap')
|
|
|
|
# Make page backup to clean every ptrack
|
|
self.backup_node(backup_dir, 'node', node, backup_type='page', options=['-j100'])
|
|
node.safe_psql('postgres', 'checkpoint')
|
|
|
|
for i in idx_ptrack:
|
|
# get new size of heap and indexes and calculate it in pages
|
|
idx_ptrack[i]['size'] = self.get_fork_size(node, i)
|
|
# update path to heap and index files in case they`ve changed
|
|
idx_ptrack[i]['path'] = self.get_fork_path(node, i)
|
|
# # get ptrack for every idx
|
|
idx_ptrack[i]['ptrack'] = self.get_ptrack_bits_per_page_for_fork(
|
|
node, idx_ptrack[i]['path'], [idx_ptrack[i]['size']])
|
|
# check that ptrack bits are cleaned
|
|
self.check_ptrack_clean(idx_ptrack[i], idx_ptrack[i]['size'])
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|
|
|
|
# @unittest.skip("skip")
|
|
# @unittest.expectedFailure
|
|
def test_ptrack_clean_replica(self):
|
|
fname = self.id().split('.')[3]
|
|
master = self.make_simple_node(base_dir="{0}/{1}/master".format(module_name, fname),
|
|
set_replication=True,
|
|
initdb_params=['--data-checksums'],
|
|
pg_options={'ptrack_enable': 'on', 'wal_level': 'replica', 'max_wal_senders': '2'})
|
|
backup_dir = os.path.join(self.tmp_path, module_name, fname, 'backup')
|
|
self.init_pb(backup_dir)
|
|
self.add_instance(backup_dir, 'master', master)
|
|
master.start()
|
|
|
|
self.create_tblspace_in_node(master, 'somedata')
|
|
self.backup_node(backup_dir, 'master', master, options=['--stream'])
|
|
|
|
replica = self.make_simple_node(base_dir="{0}/{1}/replica".format(module_name, fname))
|
|
replica.cleanup()
|
|
|
|
self.restore_node(backup_dir, 'master', replica, options=['-T', '{0}={1}'.format(
|
|
self.get_tblspace_path(master, 'somedata'), self.get_tblspace_path(replica, 'somedata_new'))])
|
|
|
|
self.add_instance(backup_dir, 'replica', replica)
|
|
self.set_replica(master, replica)
|
|
self.set_archiving(backup_dir, 'replica', replica, replica=True)
|
|
replica.start()
|
|
|
|
# Create table and indexes
|
|
master.safe_psql(
|
|
"postgres",
|
|
"create sequence t_seq; create table t_heap tablespace somedata as select i as id, nextval('t_seq') as t_seq, md5(i::text) as text, md5(repeat(i::text,10))::tsvector as tsvector from generate_series(0,256) i")
|
|
for i in idx_ptrack:
|
|
if idx_ptrack[i]['type'] != 'heap' and idx_ptrack[i]['type'] != 'seq':
|
|
master.safe_psql("postgres", "create index {0} on {1} using {2}({3}) tablespace somedata".format(
|
|
i, idx_ptrack[i]['relation'], idx_ptrack[i]['type'], idx_ptrack[i]['column']))
|
|
|
|
# Make full backup to clean every ptrack
|
|
self.backup_node(backup_dir, 'replica', replica, options=['-j100', '--stream',
|
|
'--master-host=localhost', '--master-db=postgres', '--master-port={0}'.format(master.port)])
|
|
|
|
for i in idx_ptrack:
|
|
# get fork size and calculate it in pages
|
|
idx_ptrack[i]['size'] = self.get_fork_size(replica, i)
|
|
# get path to heap and index files
|
|
idx_ptrack[i]['path'] = self.get_fork_path(replica, i)
|
|
# get ptrack for every idx
|
|
idx_ptrack[i]['ptrack'] = self.get_ptrack_bits_per_page_for_fork(
|
|
replica, idx_ptrack[i]['path'], [idx_ptrack[i]['size']])
|
|
self.check_ptrack_clean(idx_ptrack[i], idx_ptrack[i]['size'])
|
|
|
|
# Update everything, vacuum it and make PTRACK BACKUP
|
|
master.safe_psql('postgres', "update t_heap set t_seq = nextval('t_seq'), text = md5(text), tsvector = md5(repeat(tsvector::text, 10))::tsvector;")
|
|
master.safe_psql('postgres', 'vacuum t_heap')
|
|
|
|
backup_id = self.backup_node(backup_dir, 'replica', replica, backup_type='ptrack', options=['-j100', '--stream',
|
|
'--master-host=localhost', '--master-db=postgres', '--master-port={0}'.format(master.port)])
|
|
|
|
master.safe_psql('postgres', 'checkpoint')
|
|
for i in idx_ptrack:
|
|
# get new size of heap and indexes and calculate it in pages
|
|
idx_ptrack[i]['size'] = self.get_fork_size(replica, i)
|
|
# update path to heap and index files in case they`ve changed
|
|
idx_ptrack[i]['path'] = self.get_fork_path(replica, i)
|
|
# # get ptrack for every idx
|
|
idx_ptrack[i]['ptrack'] = self.get_ptrack_bits_per_page_for_fork(
|
|
replica, idx_ptrack[i]['path'], [idx_ptrack[i]['size']])
|
|
# check that ptrack bits are cleaned
|
|
self.check_ptrack_clean(idx_ptrack[i], idx_ptrack[i]['size'])
|
|
|
|
# WAITING FOR PGPRO-820
|
|
## Update everything, vacuum it and make PAGE BACKUP
|
|
# master.safe_psql('postgres', "update t_heap set t_seq = nextval('t_seq'), text = md5(text), tsvector = md5(repeat(tsvector::text, 10))::tsvector;")
|
|
# master.safe_psql('postgres', 'vacuum t_heap')
|
|
# master.safe_psql('postgres', 'checkpoint')
|
|
|
|
## Make page backup to clean every ptrack
|
|
# self.backup_node(backup_dir, 'replica', replica, backup_type='page', options=['-j100',
|
|
# '--master-host=localhost', '--master-db=postgres', '--master-port={0}'.format(master.port)])
|
|
# for i in idx_ptrack:
|
|
# # get new size of heap and indexes and calculate it in pages
|
|
# idx_ptrack[i]['size'] = self.get_fork_size(replica, i)
|
|
# # update path to heap and index files in case they`ve changed
|
|
# idx_ptrack[i]['path'] = self.get_fork_path(replica, i)
|
|
# # # get ptrack for every idx
|
|
# idx_ptrack[i]['ptrack'] = self.get_ptrack_bits_per_page_for_fork(
|
|
# replica, idx_ptrack[i]['path'], idx_ptrack[i]['size'], idx_ptrack[i]['size'])
|
|
# # check that ptrack bits are cleaned
|
|
# self.check_ptrack_clean(idx_ptrack[i], idx_ptrack[i]['size'])
|
|
|
|
# Clean after yourself
|
|
self.del_test_dir(module_name, fname)
|