On Fri, Apr 28, 2017 at 3:24 PM, David Rowley <david.row...@2ndquadrant.com> wrote:
> On 29 April 2017 at 07:59, Dave Vitek <dvi...@grammatech.com> wrote: > > Is what I want in the query planner's vocabulary? It would need to > exploit > > the fact that the _id columns are not nullable, and either exploit the > > uniqueness of the id columns or do an extra LIMIT step after the join. I > > think I want it to effectively give the same result I expect (haven't > > tested) it would give for: > > Unfortunately, it's not a plan that the current planner will consider. > > > SELECT D.val, B.val, C.val FROM > > (SELECT * FROM A ORDER BY A.rank LIMIT 100) AS D > > JOIN B ON D.b_id = B.id > > JOIN C ON B.c_id = C.id > > LIMIT 100; > > > > Perhaps there are reasons why this optimization is not safe that I > haven't > > thought about? > > Yeah, I think so. What happens if an A row cannot find a match in B or > C? This version of the query will end up returning fewer rows due to > that, but the original version would consider other rows with a higher > rank. > > We've danced around a bit with using foreign keys as proofs that rows > will exist for other optimisations in the past, but it's tricky ground > since foreign keys are not updated immediately, so there are windows > where they may not actually hold true to their word. > I read this query as having a relation cardinality of one-to-one mandatory - which precludes the scenario described. Is the above saying that, today, there is no planning benefit to setting up two deferrable references constraints to enforce the non-optional requirement? I know I'm guilty of not enforcing the non-optional part of the constraint. Mostly due to not really realizing it but also having to deal the added syntax to perform inserts. ORMs I suspect generally would be unaccommodating here as well... David J.