Hi guys,

I'm interested in a solution that will allow our customers to run reports - 
which may involve complicated queries - on data which is as up-to-date as 
possible.

One thing I don't want to do is to let the reporting system connect to the 
production database. I want the indexes in production to be limited to what 
production needs, and not add indexes that are required for reports, for 
instance. And basically, I don't want a customer to run a complicated report 
and degrade the performance of my production system.

A replication solution is not very good, either, because of course I can't 
define indexes differently, I don't want *all* transactions in all tables to be 
sent, and also, because I may want to cross reference data from different 
systems. So ideally, I want to have a reporting database, where specific tables 
(or maybe even just specific columns) from various databases are collected, and 
have a reporting tool connect to this database. But I want to push the data 
into into that database as close to real time as possible.

The most important data I am currently considering are two tables which have an 
average of 7,600 transactions per hour (standard deviation 10,000, maximum in 
May is 62,000 transactions per hour). There may be similar pairs of tables 
collected from more than one database.

I assume this is not an uncommon scenario. What solutions would you recommend?


Herouth
-- 
Sent via pgsql-general mailing list (pgsql-general@postgresql.org)
To make changes to your subscription:
http://www.postgresql.org/mailpref/pgsql-general

Reply via email to