>
>
> Since you are on RHEL 6 I would use ext4 throughout.
>
Great, I'll look into it. However my research suggested a journalled
filesystem is unnecessary for xlogs and I assume ext4 is journalled?


> You say you have I/O problems when "stats jobs" run.  Can you describe
> those jobs
> and what they are doing?
>

They summarise all the activity in the logs since the last run, essentially
counting logs. Since there are around 30 columns in each log and I want to
summarise in a number of ways - think multiple but separate groups bys with
a regexp to filter out some of the logs which are not interesting -
inevitably there are several passes over the table. Of course regexps are
also very slow in postgres. I found various indices did not help at all,
query planner thinks sequential scan is the way to go.


>
> If you have a lot of sequential scans on a 10GB table, that will suck
> majorly
> no matter how you tune it.
>

Ok.. but they I don't care how long they take, only that they don't affect
new writes to the table. So will splitting xlogs off into a different
partition/lun etc help?


>
> Two more things that you can try out:
> - Change the I/O scheduler to "deadline" by booting with
> "elevator=deadline".
>
I'll check it out.


> - Split the 400GB LUN into several smaller LUNs and use tablespaces.
>

This could be worth doing on the active logging table if splitting off
xlogs won't help. I archive (copy) these records off to the a new table
every month and then use inheritance to access the entire logset..

log_master
..inherits
  log_active
  log_201205
  log_201204
  ..etc

This is how I got the active table down to 10GB :)


>
> I don't say that that is guaranteed to help, but I have made good
> experiences
> with it.
>
> Yours,
> Laurenz Albe
>

thanks,

Ben

Reply via email to