> On May 28, 2020, at 8:54 PM, John Naylor <john.nay...@2ndquadrant.com> wrote:
> 
> On Fri, May 29, 2020 at 5:59 AM Mark Dilger
> <mark.dil...@enterprisedb.com> wrote:
>> 
>>> On May 21, 2020, at 12:12 AM, John Naylor <john.nay...@2ndquadrant.com> 
>>> wrote:
> 
>>> very picky in general. As a test, it also successfully finds a
>>> function for the OS "words" file, the "D" sets of codepoints, and for
>>> sets of the first n built-in OIDs, where n > 5.
>> 
>> Prior to this patch, src/tools/gen_keywordlist.pl is the only script that 
>> uses PerfectHash.  Your patch adds a second.  I'm not convinced that 
>> modifying the PerfectHash code directly each time a new caller needs 
>> different multipliers is the right way to go.

I forgot in my first round of code review to mention, "thanks for the patch".  
I generally like what you are doing here, and am trying to review it so it gets 
committed.

> Calling it "each time" with a sample size of two is a bit of a
> stretch. The first implementation made a reasonable attempt to suit
> future uses and I simply made it a bit more robust. In the text quoted
> above you can see I tested some scenarios beyond the current use
> cases, with key set sizes as low as 6 and as high as 250k.

I don't really have an objection to what you did in the patch.  I'm not going 
to lose any sleep if it gets committed this way.

The reason I gave this feedback is that I saved the *kwlist_d.h files generated 
before applying the patch, and compared them with the same files generated 
after applying the patch, and noticed a very slight degradation.  Most of the 
files changed without any expansion, but the largest of them, 
src/common/kwlist_d.h, changed from

  static const int16 h[901]

to

  static const int16 h[902]

suggesting that even with your reworking of the parameters for PerfectHash, you 
weren't able to find a single set of numbers that worked for the two datasets 
quite as well as different sets of numbers each tailored for a particular data 
set.  I started to imagine that if we wanted to use PerfectHash for yet more 
stuff, the problem of finding numbers that worked across all N data sets (even 
if N is only 3 or 4) might be harder still.  That's all I was referring to.  
901 -> 902 is such a small expansion that it might not be worth worrying about.

—
Mark Dilger
EnterpriseDB: http://www.enterprisedb.com
The Enterprise PostgreSQL Company





Reply via email to