[PERFORM] Slow update statement

2005-08-06 Thread Patrick Hatcher

[Reposted from General section with updated information]
Pg 7.4.5

I'm running an update statement on about 12 million records using the
following query:

Update table_A
set F1 = b.new_data
from table_B b
where b.keyfield = table_A.keyfield

both keyfields are indexed, all other keys in table_A were dropped, yet this 
job has been running over 15 hours.  Is
this normal?

I stopped the process the first time after 3 hours of running due to excessive 
log rotation and reset the conf file to these settings:


wal_buffers = 64# min 4, 8KB each

# - Checkpoints -

checkpoint_segments = 128   # in logfile segments, min 1, 16MB each
checkpoint_timeout = 1800   # range 30-3600, in seconds
#checkpoint_warning = 30# 0 is off, in seconds
#commit_delay = 0   # range 0-10, in microseconds
#commit_siblings = 5# range 1-1000


Would it just be quicker to run a JOIN statement to a temp file and then reinsert?  

TIA 
Patrick



---(end of broadcast)---
TIP 9: In versions below 8.0, the planner will ignore your desire to
  choose an index scan if your joining column's datatypes do not
  match


Re: [PERFORM] Slow update statement

2005-08-06 Thread John A Meinel
Patrick Hatcher wrote:
> [Reposted from General section with updated information]
> Pg 7.4.5
>
> I'm running an update statement on about 12 million records using the
> following query:
>
> Update table_A
> set F1 = b.new_data
> from table_B b
> where b.keyfield = table_A.keyfield
>
> both keyfields are indexed, all other keys in table_A were dropped, yet
> this job has been running over 15 hours.  Is
> this normal?

Can you do an EXPLAIN UPDATE so that we can have an idea what the
planner is trying to do?

My personal concern is if it doing something like pulling in all rows
from b, and then one by one updating table_A, but as it is going, it
can't retire any dead rows, because you are still in a transaction. So
you are getting a lot of old rows, which it has to pull in to realize it
was old.

How many rows are in table_B?

I can see that possibly doing it in smaller chunks might be faster, as
would inserting into another table. But I would do more of a test and
see what happens.

John
=:->

>
> I stopped the process the first time after 3 hours of running due to
> excessive log rotation and reset the conf file to these settings:
>
>
> wal_buffers = 64# min 4, 8KB each
>
> # - Checkpoints -
>
> checkpoint_segments = 128   # in logfile segments, min 1, 16MB each
> checkpoint_timeout = 1800   # range 30-3600, in seconds
> #checkpoint_warning = 30# 0 is off, in seconds
> #commit_delay = 0   # range 0-10, in microseconds
> #commit_siblings = 5# range 1-1000
>
>
> Would it just be quicker to run a JOIN statement to a temp file and then
> reinsert?
> TIA Patrick
>
>
> ---(end of broadcast)---
> TIP 9: In versions below 8.0, the planner will ignore your desire to
>   choose an index scan if your joining column's datatypes do not
>   match
>



signature.asc
Description: OpenPGP digital signature


Re: [PERFORM] Slow update statement

2005-08-06 Thread Tom Lane
Patrick Hatcher <[EMAIL PROTECTED]> writes:
> I'm running an update statement on about 12 million records using the
> following query:

> Update table_A
> set F1 = b.new_data
> from table_B b
> where b.keyfield = table_A.keyfield

What does EXPLAIN show for this?

Do you have any foreign key references to table_A from elsewhere?

regards, tom lane

---(end of broadcast)---
TIP 9: In versions below 8.0, the planner will ignore your desire to
   choose an index scan if your joining column's datatypes do not
   match