Michael Fuhr <[EMAIL PROTECTED]> writes:
>  Limit  (cost=0.00..25.79 rows=1 width=8) (actual time=631.964..631.964 
> rows=0 loops=1)
>    ->  Index Scan using foo_value_idx on foo  (cost=0.00..2552.75 rows=99 
> width=8) (actual time=631.942..631.942 rows=0 loops=1)
>          Filter: (id = -1)
>  Total runtime: 632.135 ms
> (4 rows)

> Maybe I don't understand something about what EXPLAIN is showing,
> but why does Limit have an estimated cost of 0.00..25.79 when the
> thing it's limiting has a cost of 0.00..2552.75?

This represents the planner assuming that the indexscan will only need
to be run 1/99th of the way to completion.  That is, having estimated
that there were 99 matching rows to be found, it assumes those are
uniformly distributed in the index-by-value, and that the scan can stop
as soon as the first one is found.

Since in reality there aren't *any* matching rows, the index scan has to
go all the way to the end :-(.  Even if there were matching rows, they
might be much further out in the index order than the
uniform-distribution hypothesis predicts, because the id and value
columns might have been correlated.

Basically, what you're looking at here is that the planner is thinking
it should go for a fast-start plan in a scenario where that bet loses.
It's still a good bet though.  I'm not sure how to formulate the notion
that there's too much risk of a slow result in this scenario.

                        regards, tom lane

---------------------------(end of broadcast)---------------------------
TIP 8: explain analyze is your friend

Reply via email to