This user guide is intended to be followed sequentially from beginning to end — each section depends on the last. For example the Backup section relies on setup that is performed in the Quick Start section. Once you have up and running then skipping around is possible but it is recommended to follow the user guide in order the first time through.
Although the examples are targeted at Ubuntu and 9.4 they will also work fine on Debian and it should be fairly easy to apply this guide to any Unix distribution and version. The only OS-specific commands are those to create, start, stop, and drop clusters. The commands will be the same on any Unix system though the locations to install Perl libraries and executables may vary. Configuring archiving is different on versions <= 8.4 and configuration information can be found in the documentation.
A somewhat novel approach is taken to documentation in this user guide. Each command is run on a virtual machine when the documentation is built from the XML source. This means you can have a high confidence that the commands work correctly in the order presented. Output is captured and displayed below the command when appropriate. If the output is not included it is because it was deemed not relevant or was considered a distraction from the narrative.
All commands are intended to be run as an unprivileged user that has sudo privileges for both the root and postgres users. It's also possible to run the commands directly as their respective users without modification and in that case the sudo commands can be stripped off.
Concepts
The following concepts are defined as they are relevant to , , and this user guide.
Backup
A backup is a consistent copy of a database cluster that can be restored to recover from a hardware failure, to perform Point-In-Time Recovery, or to bring up a new standby.
Full Backup: copies the entire contents of the database cluster to the backup server. The first backup of the database cluster is always a Full Backup. is always able to restore a full backup directly. The full backup does not depend on any files outside of the full backup for consistency.
Differential Backup: copies only those database cluster files that have changed since the last full backup. restores a differential backup by copying all of the files in the chosen differential backup and the appropriate unchanged files from the previous full backup. The advantage of a differential backup is that it requires less disk space than a full backup, however, the differential backup and the full backup must both be valid to restore the differential backup.
Incremental Backup: copies only those database cluster files that have changed since the last backup (which can be another incremental backup, a differential backup, or a full backup). As an incremental backup only includes those files changed since the prior backup, they are generally much smaller than full or differential backups. As with the differential backup, the incremental backup depends on other backups to be valid to restore the incremental backup. Since the incremental backup includes only those files since the last backup, all prior incremental backups back to the prior differential, the prior differential backup, and the prior full backup must all be valid to perform a restore of the incremental backup. If no differential backup exists then all prior incremental backups back to the prior full backup, which must exist, and the full backup itself must be valid to restore the incremental backup.
Restore
A restore is the act of copying a backup to a system where it will be started as a live database cluster. A restore requires the backup files and one or more WAL segments in order to work correctly.
Write Ahead Log (WAL)
WAL is the mechanism that uses to ensure that no committed changes are lost. Transactions are written sequentially to the WAL and a transaction is considered to be committed when those writes are flushed to disk. Afterwards, a background process writes the changes into the main database cluster files (also known as the heap). In the event of a crash, the WAL is replayed to make the database consistent.
WAL is conceptually infinite but in practice is broken up into individual 16MB files called segments. WAL segments follow the naming convention 0000000100000A1E000000FE where the first 8 hexadecimal digits represent the timeline and the next 16 digits are the logical sequence number (LSN).
A valid backup will always include at least one WAL segment even if explicit writes were made to the database between backups.
No Debian/Ubuntu packages are currently available for but it is easy to download the source and install manually.
No RHEL/CentOS packages are currently available for but it is easy to download the source and install manually.
Download version {[version]} of wget -q -O -
https://github.com/pgmasters/backrest/archive/release/{[version]}.tar.gz |
tar zx -C ~Install cp -r ~/backrest-release-{[version]}/lib/BackRest
{[perl-lib-path]}find {[perl-lib-path]}/BackRest -type f -exec chmod 644 {} +find {[perl-lib-path]}/BackRest -type d -exec chmod 755 {} +cp ~/backrest-release-{[version]}/bin/{[project-exe]} {[perl-bin-path]}/{[project-exe]}chmod 755 {[perl-bin-path]}/{[project-exe]}
should now be properly installed but it is best to check. If any dependencies were missed then you will get an error when running from the command line.
Make sure the installation worked{[project-exe]}Quick Start
The Quick Start section will cover basic configuration of and and introduce the backup, restore, and info commands.
Setup Demo Cluster
Creating the demo cluster is optional but commands in the user guide may not work in other environments without significant modification. The cluster won't be started immediately because there is still some configuration to do.
Create the demo cluster{[db-cluster-create]}
By default will only accept local connections. The examples in this guide will require connections from other servers so listen_addresses is configured to listen on all interfaces. This may not be appropriate for secure installations.
Set listen_addresses'*'
For demonstration purposes the log_line_prefix setting will be minimally configured. This keeps the log output as brief as possible to better illustrate important information.
Set log_line_prefix''
By default CentOS/RHEL includes the day of the week in the log filename. This makes automating the user guide a bit more complicated so the log_filename is set to a constant.
Set log_filename'postgresql.log'Configure Cluster Stanza
Demo describes the purpose of this cluster accurately so that will also make a good stanza name.
needs to know where the base data directory for the cluster is located. The path can be requested from directly but in a recovery scenario the process will not be available. During backups the value supplied to will be compared against the path that is running on and they must be equal or the backup will return an error. Make sure that db-path is exactly equal to data_directory in postgresql.conf.
By default Ubuntu stores clusters in /var/lib/postgresql/[version]/[cluster] so it is easy to determine the correct path for the data directory.
Configure the cluster data directory{[db-path]}
configuration files follow the Windows INI convention. Sections are denoted by text in brackets and key/value pairs are contained in each section. Lines beginning with # are ignored and can be used as comments.
Create the Repository
For this demonstration the repository will be stored on the same host as the server. This is the simplest configuration and is useful in cases where traditional backup software is employed to backup the database host.
Create the repositorymkdir {[backrest-repo-path]}chmod 750 {[backrest-repo-path]}chown postgres:postgres {[backrest-repo-path]}
The repository path must be configured so knows where to find it.
Configure the repository path{[backrest-repo-path]}Configure Archiving
Backing up a running cluster requires WAL archiving to be enabled.
The wal_level setting must be set to archive at a minimum but hot_standby and logical also work fine for backups. Setting wal_level to hot_standy and increasing max_wal_senders is a good idea even if you do not currently run a hot standby as this will allow them to be added later without restarting the master cluster.
The cluster must be restarted after making these changes and before performing a backup.
Restart the {[postgres-cluster-demo]} cluster{[db-cluster-restart]}{[db-cluster-wait]}Perform a Backup
To perform a backup of the cluster run with the backup command.
By default will attempt to perform an incremental backup. However, an incremental backup must be based on a full backup and since no full backup existed ran a full backup instead.
The type option can be used to specify a full or differential backup.
Differential backup of the {[postgres-cluster-demo]} cluster{[project-exe]} {[dash]}-stanza={[postgres-cluster-demo]} {[dash]}-type=diff
--log-level-console=info backupdiff backup size
This time there was no warning because a full backup already existed. While incremental backups can be based on a full or differential backup, differential backups must be based on a full backup. A full backup can be performed by running the backup command with {[dash]}-type=full.
Use the info command to get information about backups.
Get info for the {[postgres-cluster-demo]} cluster{[project-exe]} info(oldest|latest) backup label
The oldest and newest backups are shown in the info output. The oldest backup will always be a full backup (indicated by an F at the end of the label) but the newest backup can be full, differential (ends with D), or incremental (ends with I).
More information about the backup command can be found in the Backup section.
Restore a Backup
Backups can protect you from a number of disaster scenarios, the most common of which are hardware failure and data corruption. The easiest way to simulate data corruption is to remove an important cluster file.
Stop the {[postgres-cluster-demo]} cluster and delete the pg_control file{[db-cluster-stop]}rm {[db-path]}/global/pg_control
Starting the cluster without this important file will result in an error.
Attempt to start the corrupted {[postgres-cluster-demo]} cluster{[db-cluster-start]}could not find the database system{[db-cluster-wait]}rm -f {[postgres-log-pgstartup-demo]}{[db-cluster-start]}cat {[postgres-log-pgstartup-demo]}errorcould not find the database system
To restore a backup of the cluster run with the restore command. The cluster needs to be stopped (in this case it is already stopped) and all files must be removed from the data directory.
Remove old files from {[postgres-cluster-demo]} clusterfind {[db-path]} -mindepth 1 -deleteRestore the {[postgres-cluster-demo]} cluster and start {[project-exe]} {[dash]}-stanza={[postgres-cluster-demo]} restore{[db-cluster-start]}{[db-cluster-wait]}
This time the cluster started successfully since the restore replaced the missing pg_control file.
More information about the restore command can be found in the Restore section.
Backup
The Backup section introduces additional backup command features.
Fast Start Option
By default will wait for the next regularly scheduled checkpoint before starting a backup. Depending on the checkpoint_timeout and checkpoint_segments settings in it may be quite some time before a checkpoint completes and the backup can begin.
Incremental backup of the {[postgres-cluster-demo]} cluster with the regularly scheduled checkpoint{[project-exe]} {[dash]}-stanza={[postgres-cluster-demo]} {[dash]}-type=incr
--log-level-console=info backupbackup begins after the next regular checkpoint completes
By setting start-fast on the command-line or in {[backrest-config-demo]} an immediate checkpoint is requested and the backup will start more quickly. This is convenient for testing and for ad-hoc backups. For instance, if a backup is being taken at the beginning of a release window it makes no sense to wait for a checkpoint. Since regularly scheduled backups generally only happen once per day it is unlikely that enabling the start-fast in {[backrest-config-demo]} will negatively affect performance, however for high-volume transactional systems you may want to pass {[dash]}-start-fast on the command-line instead.
Enable the start-fast optionyIncremental backup of the {[postgres-cluster-demo]} cluster with an immediate checkpoint{[project-exe]} {[dash]}-stanza={[postgres-cluster-demo]} {[dash]}-type=incr
--log-level-console=info backupbackup begins after the requested immediate checkpoint completesAutomatic Stop Option
Sometimes will exit unexpectedly and the backup in progress on the cluster will not be properly stopped. exits as quickly as possible when an error occurs so that the cause can be reported accurately and is not masked by another problem that might happen during a more extensive cleanup.
Here an error in intentionally caused by removing repository permissions.
Revoke write privileges in the repository and attempt a backupchmod 550 {[backrest-repo-path]}/temp{[project-exe]} {[dash]}-stanza={[postgres-cluster-demo]} {[dash]}-type=incr
--log-level-console=info backupERROR:
Even when the permissions are fixed will still be unable to perform a backup because the cluster is stuck in backup mode.
Restore write privileges in the repository and attempt a backupchmod 750 {[backrest-repo-path]}/temp{[project-exe]} {[dash]}-stanza={[postgres-cluster-demo]} {[dash]}-type=incr
--log-level-console=info backupERROR:
Enabling the stop-auto option allows to stop the current backup if it detects that no other backup process is running.
Enable the stop-auto optiony
Now will stop the old backup and start a new one so the process completes successfully.
Perform an incremental backup{[project-exe]} {[dash]}-stanza={[postgres-cluster-demo]} {[dash]}-type=incr
--log-level-console=info backupcluster is already in backup mode|backup begins after the requested immediate checkpoint completes
Although useful this feature may not be appropriate when another third-party backup solution is being used to take online backups as will not recognize that the other software is running and may terminate a backup started by that software. However, it would be unusual to run more than one third-party backup solution at the same time so this is not likely to be a problem.
Note that pg_dump and pg_base_backup do not take online backups so are not affected. It is safe to run them in conjunction with .
Retention
Generally it is best to retain as many backups as possible to provide a greater window for Point-in-Time Recovery, but practical concerns such as disk space must also be considered. Retention options remove older backups once they are no longer needed.
Full Backup Retention
Set full-retention to the number of full backups required. New backups must be completed before expiration will occur — that means if retention-full=2 then there will be three full backups stored before the oldest one is expired.
Configure retention-full2
Backup retention-full=2 but currently there is only one full backup so the next full backup to run will not expire any full backups.
Perform a full backup{[project-exe]} {[dash]}-stanza={[postgres-cluster-demo]} --type=full
--log-level-console=info backuparchive retention on backup {[backup-full-first]}|remove archive{[cmd-backup-last]}
Archive is expired because WAL segments were generated before the oldest backup. These are not useful for recovery — only WAL segments generated after a backup can be used to recover that backup.
Perform a full backup{[project-exe]} {[dash]}-stanza={[postgres-cluster-demo]} --type=full
--log-level-console=info backupexpire full backup set\: {[backup-full-first]}|archive retention on backup {[backup-full-second]}|remove archive
The {[backup-full-first]} full backup is expired and archive retention is based on the {[backup-full-second]} which is now the oldest full backup.
Differential Backup Retention
Set retention-diff to the number of differential backups required. Differentials only rely on the prior full backup so it is possible to create a rolling set of differentials for the last day or more. This allows quick restores to recent points-in-time but reduces overall space consumption.
Configure retention-diff1
Backup retention-diff=2 so two differentials will need to be performed before one is expired. An incremental backup is added to demonstrate incremental expiration. Incremental backups cannot be expired independently — they are always expired with their related full or differential backup.
The Restore section introduces additional restore command features.
Delta Option
Restore a Backup in Quick Start required the database cluster directory to be cleaned before the restore could be performed. The delta allows to automatically determine which files in the database cluster directory can be preserved and which ones need to be restored from the backup. This is accomplished by calculating a SHA-1 cryptographic hash for each file in the database cluster directory. If the SHA-1 hash does not match the hash stored in the backup then that file will be restored. This operation is very efficient when combined with the thread-max option. Since the process is shut down during the restore, a larger number of threads can be used than might be desirable during a backup when the process is running.
Stop the {[postgres-cluster-demo]} cluster, perform delta restore{[db-cluster-stop]}{[project-exe]} {[dash]}-stanza={[postgres-cluster-demo]} {[dash]}-delta
--log-level-console=info restoredemo\/PG_VERSION - exists and matches backup|check\/clean db path|restore global\/pg_controlRestart {[db-cluster-start]}{[db-cluster-wait]}Point-in-Time Recovery
Restore a Backup in Quick Start performed default recovery, which is to play all the way to the end of the WAL stream. In the case of a hardware failure this is probably the most appropriate action but for data corruption scenarios (whether machine or human in origin) there is a better alternative called Point-in-Time Recovery (PITR).
Point-in-Time Recovery (PITR) allows the WAL to be played from the last backup to a specified time, transaction id, or recovery point. For common recovery scenarios time-based recovery is arguably the most useful. A typical recovery scenario is to restore a table that was accidentally dropped or data that was accidentally deleted. Recovering a dropped table is more dramatic so that's the example given here but deleted data would be recovered in exactly the same way.
Backup the {[postgres-cluster-demo]} cluster and create a table with very important data{[project-exe]} {[dash]}-stanza={[postgres-cluster-demo]} --type=diff backup
psql -c "begin;
create table important_table (message text);
insert into important_table values ('{[test-table-data]}');
commit;
select * from important_table";
{[test-table-data]}
It is important to represent the time as reckoned by and to include timezone offsets. This reduces the possibility of unintended timezone conversions and an unexpected recovery result.
Get the time from
psql -Atc "select current_timestamp"
Now that the time has been recorded the table is dropped. In practice finding the exact time that the table was dropped is a lot harder than in this example. It may not be possible to find the exact time, but some forensic work should be able to get you close.
Drop the important tablepsql -c "begin;
drop table important_table;
commit;
select * from important_table;"does not exist
Now the restore can be performed with time-based recovery to bring back the missing table.
Stop , restore the {[postgres-cluster-demo]} cluster to {[time-recovery-timestamp]}, and display recovery.conf{[db-cluster-stop]}{[project-exe]} {[dash]}-stanza={[postgres-cluster-demo]} {[dash]}-delta
{[dash]}-type=time "{[dash]}-target={[time-recovery-timestamp]}" restorerm {[postgres-log-demo]}cat {[postgres-recovery-demo]}recovery_target_time
The recovery.conf file has been automatically generated by so can be started immediately. Once has finished recovery the table will exist again and can be queried.
Start and check that the important table exists{[db-cluster-start]}{[db-cluster-wait]}psql -c "select * from important_table"{[test-table-data]}
The log also contains valuable information. It will indicate the time and transaction where the recovery stopped and also give the time of the last transaction to be applied.
This example was rigged to give the correct result. If a backup after the required time is chosen then will not be able to recover the lost table. can only play forward, not backward. To demonstrate this the important table must be dropped (again).
Drop the important table (again)psql -c "begin;
drop table important_table;
commit;
select * from important_table;"does not exist
Now take a new backup and attempt recovery from the new backup.
Perform a backup then attempt recovery from that backup{[cmd-backup-last]}{[project-exe]} {[dash]}-stanza={[postgres-cluster-demo]} {[dash]}-type=incr backup{[db-cluster-stop]}{[project-exe]} {[dash]}-stanza={[postgres-cluster-demo]} {[dash]}-delta
{[dash]}-type=time "{[dash]}-target={[time-recovery-timestamp]}" restorerm {[postgres-log-demo]}{[db-cluster-start]}{[db-cluster-wait]}psql -c "select * from important_table"does not exist
Looking at the log output it's not obvious that recovery failed to restore the table. The key is to look for the presence of the recovery stopping before... and last completed transaction... log messages. If they are not present then the recovery to the specified point-in-time was not successful.
Examine the log output to discover the recovery was not successfulcat {[postgres-log-demo]}starting point-in-time recovery
Using an earlier backup will allow to play forward to the correct time again. The default behavior for restore is to use the most recent backup but an earlier backup can be specified with the {[dash]}-set option.
Stop , restore from the previous backup, and start {[db-cluster-stop]}
{[project-exe]} {[dash]}-stanza={[postgres-cluster-demo]} {[dash]}-delta
{[dash]}-type=time "{[dash]}-target={[time-recovery-timestamp]}"
{[dash]}-set={[backup-last]} restore
rm {[postgres-log-demo]}{[db-cluster-start]}{[db-cluster-wait]}psql -c "select * from important_table"{[test-table-data]}
Now the the log output will contain the expected recovery stopping before... and last completed transaction... messages showing that the recovery was successful.
Examine the log output for log messages indicating successcat {[postgres-log-demo]}recovery stopping before|last completed transaction|starting point-in-time recoveryDedicated Backup Host
The configuration described in Quickstart is suitable for simple installations but for enterprise configurations it is more typical to have a dedicated backup host. This separates the backups and WAL archive from the database server so database host failures have less impact. It is still a good idea to employ traditional backup software to backup the backup host.
Installation and Configurationcp -r /backrest/lib/BackRest {[perl-lib-path]}find {[perl-lib-path]}/BackRest -type f -exec chmod 644 {} +find {[perl-lib-path]}/BackRest -type d -exec chmod 755 {} +cp /backrest/bin/{[project-exe]} {[perl-bin-path]}/{[project-exe]}chmod 755 {[perl-bin-path]}/{[project-exe]}Configure the repository path{[backrest-repo-path]}
For this example a new host named backup has been created to store the cluster backups. Follow the instructions in Installation to install and Create the Repository to create the repository. The backup host must also be configured with the database host/user and database path.
Configure db-host/db-user and db-path{[db-path]}{[host-db-master]}y
The database host must be configured with the backup host/user.
Commands are run the same as on a single host configuration except that the backup and expire command are run from the backup host and all other commands are run from the database host.
Perform a Backup
To perform a backup of the cluster run with the backup command on the backup host.
Backup the {[postgres-cluster-demo]} cluster{[project-exe]} {[dash]}-stanza={[postgres-cluster-demo]} backup
Since a new repository was created on the backup host the warning about the incremental backup changing to a full backup was emitted.
Restore a Backup
To perform a restore of the cluster run with the restore command on the database host.
Stop the {[postgres-cluster-demo]} cluster, restore, and restart {[db-cluster-stop]}{[project-exe]} {[dash]}-stanza={[postgres-cluster-demo]} {[dash]}-delta restore{[db-cluster-start]}{[db-cluster-wait]}
A new backup must be performed to due to the timeline switch.
Backup the {[postgres-cluster-demo]} cluster{[project-exe]} {[dash]}-stanza={[postgres-cluster-demo]} backupStarting and Stopping
Sometimes it is useful to prevent from running on a system. For example, when failing over from a master to a standby it's best to prevent from running on the old master in case gets restarted or can't be completely killed. This will also prevent from running on cron.
Stop the services{[project-exe]} stop
New processes will no longer run.
Attempt a backup{[project-exe]} {[dash]}-stanza={[postgres-cluster-demo]} backupstop file exists for all stanzas
Specify the --force option to terminate any process that are currently running. If is already stopped then stopping again will generate a warning.
Stop the services again{[project-exe]} stop
Start processes again with the start command.
Start the services{[project-exe]} start
It is also possible to stop for a single stanza.
Stop services for the demo stanza{[project-exe]} {[dash]}-stanza={[postgres-cluster-demo]} stop{[project-exe]} {[dash]}-stanza={[postgres-cluster-demo]} backupstop file exists for stanza demo
The stanza must also be specified when starting the processes for a single stanza.
Start the services for the demo stanza{[project-exe]} {[dash]}-stanza={[postgres-cluster-demo]} startReplication
Replication allows multiple copies of a cluster (called standbys) to be created from a single master. The standbys are useful for balancing reads and to provide redundancy in case the master host fails.
Hot Standby
A hot standby performs replication using the WAL archive and allows read-only queries.
A new host named db-standby will be created to run the standby. Follow the instructions in Installation to install , Setup Demo Cluster to setup the demo cluster, and Create the Repository to create the repository on the db-standby host.
configuration is very similar to db-master except that the standby_mode setting will be enabled to keep the cluster in recovery mode when the end of the WAL stream has been reached.
Configure on the standby{[db-path]}{[backrest-repo-path]}{[host-backup]}on
Now the standby can be created with the restore command.
Restore the {[postgres-cluster-demo]} standby cluster{[project-exe]} {[dash]}-stanza={[postgres-cluster-demo]} {[dash]}-delta restorecat {[postgres-recovery-demo]}
Note that the standby_mode setting has been written into the recovery.conf file. Configuring recovery settings in means that the recovery.conf file does not need to be stored elsewhere since it will be properly recreated with each restore. The --type=preserve option can be used with the restore to leave the existing recovery.conf file in place if that behavior is preferred.
The hot_standby setting must be enabled before starting to allow read-only connections on db-standby. Otherwise, connection attempts will be refused.
The log gives valuable information about the recovery. Note especially that the cluster has entered standby mode and is ready to accept read-only connections.
Examine the log output for log messages indicating successcat {[postgres-log-demo]}entering standby mode|database system is ready to accept read only connections
An easy way to test that replication is properly configured is to create a table on db-master.
Create a new table on the master
psql -c "
begin;
create table replicated_table (message text);
insert into replicated_table values ('{[test-table-data]}');
commit;
select * from replicated_table";
{[test-table-data]}
And then query the same table on db-standby.
Query new table on the standbypsql -c "select * from replicated_table;"does not exist
So, what went wrong? Since is pulling WAL segments from the archive to perform replication, changes won't be seen on the standby until the WAL segment that contains those changes is pushed from db-master.
This can be done manually by calling pg_switch_xlog() which pushes the current WAL segment to the archive (a new WAL segment is created to contain further changes).
Call pg_switch_xlog()
psql -c "select *, current_timestamp from pg_switch_xlog()";
Now after a short delay the table will appear on db-standby.
Now the new table exists on the standby (may require a few retries)psql -c "
select *, current_timestamp from replicated_table"{[test-table-data]}Streaming Replication
Instead of relying solely on the WAL archive, streaming replication makes a direct connection to the master and applies changes as soon as they are made on the master. This results in much less lag between the master and standby.
Streaming replication requires a user with the replication privilege.
Create replication user
psql -c "
create user replicator password 'jw8s0F4' replication";
The pg_hba.conf file must be updated to allow the standby to connection as the replication user. Be sure to replace the IP address below with the actual IP address of your db-master. A reload will be required after modifying the pg_hba.conf file.
Create pg_hba.conf entry for replication user
sh -c 'echo
"host replication replicator {[host-db-standby-ip]}/32 md5"
>> {[postgres-hba-demo]}'
{[db-cluster-reload]}
The standby needs to know how to contact the master so the primary_conninfo setting will be configured in .
Set primary_conninfohost={[host-db-master-ip]} port=5432 user=replicator
It is possible to configure a password in the primary_conninfo setting but using a .pgpass file is more flexible and secure.
Configure the replication password in the .pgpass file.
sh -c 'echo
"{[host-db-master-ip]}:*:replication:replicator:jw8s0F4"
>> {[postgres-pgpass]}'
chmod 600 {[postgres-pgpass]}
Now the standby can be created with the restore command.
Stop and restore the {[postgres-cluster-demo]} standby cluster{[db-cluster-stop]}{[project-exe]} {[dash]}-stanza={[postgres-cluster-demo]} {[dash]}-delta restorecat {[postgres-recovery-demo]}
By default CentOS/RHEL stores the postgresql.conf file in the data directory. That means the change made to postgresql.conf was overwritten by the last restore and the hot_standby setting must be enabled again. Other solutions to this problem are to store the postgresql.conf file elsewhere or to enable the hot_standby setting on the db-master host where it will be ignored.
The log will confirm that streaming replication has started.
Examine the log output for log messages indicating successcat {[postgres-log-demo]}started streaming WAL from primary
Now when a table is created on db-master it will appear on db-standby quickly and without the need to call pg_switch_xlog().
Create a new table on the master
psql -c "
begin;
create table stream_table (message text);
insert into stream_table values ('{[test-table-data]}');
commit;
select *, current_timestamp from stream_table";
{[test-table-data]}Query table on the standbypsql -c "
select *, current_timestamp from stream_table"{[test-table-data]}