Of course. Since I intend to implement VPP as a deterministic CGN gateway I have some parameters regarding to nat config, for example: translation hash buckets, translation hash memory , user hash buckets and user hash memory to be configured in startup.conf.
In this context I would like to know how do I give the right value to those parameters. Thanks Marcos -----Mensagem original----- De: vpp-dev@lists.fd.io <vpp-dev@lists.fd.io> Em nome de Damjan Marion via lists.fd.io Enviada em: quinta-feira, 26 de novembro de 2020 16:17 Para: Marcos - Mgiga <mar...@mgiga.com.br> Cc: Elias Rudberg <elias.rudb...@bahnhof.net>; vpp-dev@lists.fd.io Assunto: Re: RES: [vpp-dev] NAT memory usage problem for VPP 20.09 compared to 20.05 due to larger translation_buckets value Sorry, I don’t understand your question. Can you elaborate further? -- Damjan > On 26.11.2020., at 20:05, Marcos - Mgiga <mar...@mgiga.com.br> wrote: > > Hello, > > Taking benefit of the topic, how you suggest to monitor if translation hash > bucket value has an appropriate value? What about translation hash memory, > user hash buckets and user hash memory ? > > How do I know if I increase or decrease those values? > > Best Regards > > Marcos > > -----Mensagem original----- > De: vpp-dev@lists.fd.io <vpp-dev@lists.fd.io> Em nome de Damjan Marion > via lists.fd.io Enviada em: quinta-feira, 26 de novembro de 2020 14:53 > Para: Elias Rudberg <elias.rudb...@bahnhof.net> > Cc: vpp-dev@lists.fd.io > Assunto: Re: [vpp-dev] NAT memory usage problem for VPP 20.09 compared > to 20.05 due to larger translation_buckets value > > > Dear Elias, > > Let me try to explain a bit underlying mechanics. > Let’s assume your target number of sessions is 10M and we are talking about > 16byte key size. > That means each hash entry (KV) is 24 bytes (16 bytes key and 8 bytes value). > > In the setup you were mentioning, with 1<<20 buckets, your will need to fit > 10 KVs into each bucket. > Initial bihash bucket holds 4 KVs and to accomodate 10 keys (assuming that > hash function gives us equal distribution) you will need to grow each bucket > 2 times. Growing means doubling bucket size. > So at the end you will have 1<<20 buckets where each holds 16 KVs. > > Math is: > 1<<20 * (16 * 24 /* KV size in bytes */ + 8 /*bucket header size*/) Which > means 392 MB of memory. > > If you keep target number of 10M sessions, but you increase number of buckets > to 1 << 22 (which is roughly what formula bellow is trying to do) you end up > with the following math: > > Math is: > 1<<22 * (4 * 24 /* KV size in bytes */ + 8 /*bucket header size*/) Which > means 416 MB of memory. > > So why 2nd one is better. Several reasons: > > - in first case you need to grow each bucket twice, that means > allocating memory for the new bucket, copying existing data from the > old bucket and putting old bucket to the free list. This operation > increases key insertion time and lowers performance > > - growing will likely result in significant amount of old buckets > sitting in the free list and they are effectively wasted memory > (bihash tries to reuse that memory but at some point there is no > demand anymore for smaller buckets) > > - performance-wise original bucket (one which first 4 KVs) is collocated with > bucket header. > This is new behaviour Dave introduced earlier this year (and I think it is > present in 20.09). > Bucket collocated with header means that there is no dependant > prefetch needed as both header and at least part of data sits in the same > cacheline. This significantly improveslookup performance. > > So in general, for best performance and optimal memory usage, number of > buckets should be big enough so it unlikely grow with your target number of > KVs. rule of thumb would be rounding target number of entries to closer > power-of-2 value and then dividing that number with 2. > For example, for 10M entries first lower power-of-2 number is 1<<23 (8M) and > first higher is 1<<24 (16M). > 1<<23 is closer, when we divide that by 2 we got 1<<22 (4M) buckets. > > Hope this explains…. > > — > Damjan > > >> On 26.11.2020., at 17:54, Elias Rudberg <elias.rudb...@bahnhof.net> wrote: >> >> Hello VPP experts, >> >> We are using VPP for NAT44 and are currently looking at how to move >> from VPP 20.05 to 20.09. There are some differences in the way the >> NAT plugin is configured. >> >> One difficulty for us is the maximum number of sessions allowed, we >> need to handle large numbers of sessions so that limit can be >> important for us. For VPP 20.05 we have used "translation hash >> buckets 1048576" and then the maximum number of sessions per thread >> becomes 10 times that because of this line in the source code in >> snat_config(): >> >> sm->max_translations = 10 * translation_buckets; >> >> So then we got a limit of about 10 million sessions per thread, which >> we have been happy with so far. >> >> With VPP 20.09 however, things have changed so that the maximum >> number of sessions is now configured explicitly, and the relationship >> between max_translations_per_thread and translation_buckets is no >> longer a factor of 10 but instead given by the >> nat_calc_bihash_buckets() >> function: >> >> static u32 >> nat_calc_bihash_buckets (u32 n_elts) >> { >> return 1 << (max_log2 (n_elts >> 1) + 1); } >> >> The above function corresponds to a factor of somewhere between 1 and >> 2 instead of 10. So, if I understood this correctly, for a given >> maximum number of sessions, the corresponding translation_buckets >> value will be something like 5 to 10 times larger in VPP 20.09 >> compared to how it was in VPP 20.05, leading to significantly >> increased memory requirement given that we want to have the same >> maximum number of sessions as before. >> >> It seems a little strange that the translation_buckets value would >> change so much between VPP versions, was that change intentional? The >> old relationship "max_translations = 10 * translation_buckets" seems >> to have worked well in practice, at least for our use case. >> >> What could we do to get around this, if we want to switch to VPP >> 20.09 but without reducing the maximum number of sessions? If we were >> to simply divide the nat_calc_bihash_buckets() value by 8 or so to >> make it more similar to how it was earlier, would that lead to other >> problems? >> >> Best regards, >> Elias >> >> >> > > > > >
-=-=-=-=-=-=-=-=-=-=-=- Links: You receive all messages sent to this group. View/Reply Online (#18165): https://lists.fd.io/g/vpp-dev/message/18165 Mute This Topic: https://lists.fd.io/mt/78535814/21656 Group Owner: vpp-dev+ow...@lists.fd.io Unsubscribe: https://lists.fd.io/g/vpp-dev/unsub [arch...@mail-archive.com] -=-=-=-=-=-=-=-=-=-=-=-