Re: [PERFORM] optimizing LIKE '%2345' queries
Am Sonntag, 2. Juli 2006 23:50 schrieb Gene: > Is there any way to create a reverse index on string columns so that > queries of the form: > > where column like '%2345'; > > can use an index and perform as fast as searching with like '2345%'? > > Is the only way to create a reverse function and create an index using > the reverse function and modify queries to use: > > where reverse(column) like reverse('%2345') ? > > thanks create a trigger that computes this at insert/update time, index this fix, and rewrite the query this way: where inverted_column like '5432%'; ---(end of broadcast)--- TIP 5: don't forget to increase your free space map settings
[PERFORM] Is postgresql ca do the job for software deployed in ASP ou SaaS mode?
Hi all, I've been working on my personal project for 3.5 years now. I developed an ERP system in web/java. Now the people I will work with suggest to offers it in Saas mode. Which means my customer will connect to my website and found they ERP software and data there. It's not the deployment I planned initially so if you can just validate some technicals points to be sure it's not crazy using Postgresl here and not a big $$$ db to do the job. Typically I will have 1db per client and around 150 tables per db. So since I hope I didn`t work all those year for nothing .. I expect to have bunch of clients witch means the same amount of db since I have 1 db/client. Can I hope having several hundred of db on 1 db server? Like 250 dbs = 250 client = 360 000 tables !!! So is there a limit for the number of db in the db server ?(this spec is not on the website) What about the performance? Can I expect to have the same performance? Since I put everything on the web I do needs an High Availability infrastructure. I looked into SlonyI and Mammoth to replicate the db but since SlonyI use triggers can I expect it to do the job? Is Mammoth is the only available solution? Last question and not the least I'm reading this performance list for several years now and know suggestion about hardware to run postgresl is discussed. Since I wrote software there severals points about hardware that I don`t understand. Do you have any suggestion of platform to run into my Saas configuration? I do need the WISE one! I'm pretty sure that if I was a big company I would be able throw bunch of but it's not my case. I'm pretty sure it exists out there some piece of Hardware that would do the job perfectly with a fair price. So far I did understand that Postgresql loves Opteron and I have looked into the dl145 series of HP. I did understand that Dell Hardware it`s not reliable. But it's still not clear what should be my requirement for memory, disk, nb cpu, cpu power, etc. I'm pretty sure it`s better to have more slower CPUs that having the latest Opteron available on the market, or more slower servers that having the fastest one... am I right? But agains what it`s the optimal choice? Thanks you to share your knowledge on those point. I do consider using Postgresql is the Smart choice in my project since the beginning but before putting all the money (That I don`t have ..:-)) to buy some hardware I just want to be sure I'm not crazy! Thanks for your help I really appreciate it!! Best Regards /David ---(end of broadcast)--- TIP 4: Have you searched our list archives? http://archives.postgresql.org
Re: [PERFORM] Is postgresql ca do the job for software deployed in ASP ou SaaS mode?
> Typically I will have 1db per client and around 150 tables per db. So > since I hope I didn`t work all those year for nothing .. I expect to > have bunch of clients witch means the same amount of db since I have 1 > db/client. > > Can I hope having several hundred of db on 1 db server? Like 250 dbs = > 250 client = 360 000 tables !!! > So is there a limit for the number of db in the db server ?(this spec is > not on the website) I'll take a stab at this question. Each table and database are referenced by an OID. So the sum(tables) + sum(database) << max-size(OID). In my case max-size of OID (I believe) is 9223372036854775807. So if you limited yourself to 1% of the OIDs for use as tables and databases then you could potentially have 92233720368547758 table or database. Each database create produces a directory with the database OID: ./data/base/10792 ./data/base/10793 ./data/base/16814 ... ... since the creation of a new db produces a directory, one limitation would come from your file-systems limitation on the number of sub-directories that are allowed. Each table with-in the database is assigned an OID and is located inside the DB directory. So if there is a file-system limitation on the number of files with-in a given directory it would also be a limit to the number of tables that could be created for each database. The only other limitation that I am aware of is the storage capacity of you DB server. If there are additional limitations beyond these I would be interested in knowing about them and adding them to the http://www.postgresql.org/about/ we be helpful also. Regards, Richard Broersma Jr. ---(end of broadcast)--- TIP 3: Have you checked our extensive FAQ? http://www.postgresql.org/docs/faq
Re: [PERFORM] Is postgresql ca do the job for software deployed in ASP ou SaaS mode?
On 7/3/06, David Gagnon <[EMAIL PROTECTED]> wrote: Can I hope having several hundred of db on 1 db server? Like 250 dbs =250 client = 360 000 tables !!!So is there a limit for the number of db in the db server ?(this spec isnot on the website)What about the performance? Can I expect to have the same performance? I am running a similar environment. Each of our customers has a seperate database with serveral hundred tables per database. One of our servers is running over 200 customer databases with absolutely no problems. HTH,chris
Re: [PERFORM] Is postgresql ca do the job for software deployed in
Richard Broersma Jr wrote: Each table with-in the database is assigned an OID and is located inside the DB directory. So if there is a file-system limitation on the number of files with-in a given directory it would also be a limit to the number of tables that could be created for each database. You could handle this with tablespaces. For example, create ten tablespaces, and then assign customer databases to them in round-robin fashion. This also lets you assign databases to different disks to balance the I/O load. Craig ---(end of broadcast)--- TIP 5: don't forget to increase your free space map settings
Re: [PERFORM] [BUGS] query very slow when enable_seqscan=on
On Mon, 2006-07-03 at 22:31 +0200, Tomasz Ostrowski wrote: > I have a very slow query when enable_seqscan=on and very fast when > enable_seqscan=off. My schema looks like this (relevant columns > only): > PS. Actual table and column names are different (they're in Polish) > but I've translated them for better readability for english-speaking. Thanks > PS. I wonder if it makes sense to "enable_seqscan=off" for every client > if a database is small enough to fit in OS cache. You can set this for individual statements if you choose to. > -> Seq Scan on organization (cost=0.00..480.95 rows=1 > width=4) (actual time=0.071..69.702 rows=1892 loops=1) >Filter: ((organization_location)::text ~* > 'warszawa'::text) The issue is caused by the under-estimation of the number of rows in the table as a result of the regular expression comparison. As a result the planner thinks it can choose a nested loops scan, though ends up doing 1892 seq scans of persons, when it thought it would do only one. The under estimation is a known issue. Posting to -perform for the record. -- Simon Riggs EnterpriseDB http://www.enterprisedb.com ---(end of broadcast)--- TIP 5: don't forget to increase your free space map settings
Re: [PERFORM] Is postgresql ca do the job for software deployed in ASP ou SaaS mode?
Hi, Chris, In your deployment, can you put a bit more detail if available? Many thanks! My questions are: a) How do you resolve the connection pool issue? b) If each client application has many threads of connections to the remote server, what is the likely performance penalty with compare to the DBMS hosted at the same host as the client? Indeed, the application requirements may be quite different for us, but above two are my main concerns prior to doing a porting work for a large application (from other vendor DBMS). We have several idential applications on different servers, each has 250+ database connections, currently they are having a DBMS on each server but we want them to share one DBMS at a dedicate DBMS server (in the same LAN) if performance penalty is little. I wonder if anyone there can provide your comments and experience on this. Many thanks. Regards, Guoping -Original Message-From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED]On Behalf Of Chris HooverSent: 2006Äê7ÔÂ4ÈÕ 3:20To: David GagnonCc: pgsql-performance@postgresql.orgSubject: Re: [PERFORM] Is postgresql ca do the job for software deployed in ASP ou SaaS mode?On 7/3/06, David Gagnon <[EMAIL PROTECTED]> wrote: Can I hope having several hundred of db on 1 db server? Like 250 dbs =250 client = 360 000 tables !!!So is there a limit for the number of db in the db server ?(this spec isnot on the website)What about the performance? Can I expect to have the same performance? I am running a similar environment. Each of our customers has a seperate database with serveral hundred tables per database. One of our servers is running over 200 customer databases with absolutely no problems. HTH,chris