Sat Mar 26 17:02:58 PDT 2011
- Previous message: [Slony1-general] Slony replication problem - logswitch failure
- Next message: [Slony1-general] Slony replication problem - logswitch failure
- Messages sorted by: [ date ] [ thread ] [ subject ] [ author ]
It wasn't mission critical changes lost. Postgres log was full of messages saying it couldn't switch the log because it was already in progress. Uptime was showing load averages of 60. Checking sl_log_1 it only had 4 entries. nuking it and re-initing the log switch reduced the load average to between 4 and 12. That's with 15 slaves to one master! ------Original Message------ From: Jan Wieck To: tim.lloyd at censornet.com Cc: slony1-general at lists.slony.info Subject: Re: [Slony1-general] Slony replication problem - logswitch failure Sent: Mar 26, 2011 23:52 On 3/26/2011 6:54 PM, Tim Lloyd wrote: > Only way to stop Postgres chewing all the CPU What did you do after this to verify that the unqualified delete from sl_log_1 did not make you lose any changes? Jan > > > ------Original Message------ > From: Jan Wieck > To: Tim Lloyd > Cc: slony1-general at lists.slony.info > Subject: Re: [Slony1-general] Slony replication problem - logswitch failure > Sent: Mar 26, 2011 22:53 > > On 3/26/2011 6:43 AM, Tim Lloyd wrote: >> Hi Venkat >> >> I found a way to get Slony out of this state without rebuilding the db. >> >> 1) Connect to your database, e.g. psql<dbname> as user postgres >> >> 2) delete from _schema.sl_log_1; > > Looks a little bit dangerous to me. YMMV. > > > Jan > -- Anyone who trades liberty for security deserves neither liberty nor security. -- Benjamin Franklin Sent from my BlackBerry® wireless device
- Previous message: [Slony1-general] Slony replication problem - logswitch failure
- Next message: [Slony1-general] Slony replication problem - logswitch failure
- Messages sorted by: [ date ] [ thread ] [ subject ] [ author ]
More information about the Slony1-general mailing list