@Adrian,

> Seems to me the settings for nice and ionice above would, on a busy
> machine, slow down the transfer. Has there always been a notable time
> difference in the transfer or has it gotten worse over time?
>
> Yep... I also thought about that. Specially because the master is
constantly getting 100% of IO (we use SATA disks still)...

I'm thinking about removing that `ionice` command... I don't need to
restart Postgres eh?? Just reload the confs?


@John R Pierce,

normally, you would ship the archived wal files to a file server via
> cp-over-nfs or scp, and have the slaves access them as needed via the
> recovery.conf

What if the NFS server goes down? Networking goes down? We have had that
kind of problem in the past, that's why I'm shipping the wal_files to each
slave, separately. Also, to have an extra copy of them.


@Venkata Balaji N,


Not sure why the script is so complex. Do you see any messages in the
> postgresql log file on master ? and on slave ? which indicates the reason
> for delayed shipping of WAL archives. Did you notice any network level
> issues ?

Yes the script is complex.. I've hidden almost all of it for privacy
purpose.. sorry....

I don't see any messages on the log files... not on the master and not on
the slaves as well. I just see the message of the wal_files
being successfully shipped to the slaves.

Also, no networking level issues.. because I got four slaves with streaming
replication and all of them are working fine... also, my backup server has
never failed... so no networking issues.


Thanks,

Patrick

Reply via email to