Will leave that to NAT folks to comment… They have multiple tables and they are two per thread…
— Damjan > On 26.11.2020., at 20:27, Marcos - Mgiga <mar...@mgiga.com.br> wrote: > > Of course. > > Since I intend to implement VPP as a deterministic CGN gateway I have some > parameters regarding to nat config, for example: translation hash buckets, > translation hash memory , user hash buckets and user hash memory to be > configured in startup.conf. > > In this context I would like to know how do I give the right value to those > parameters. > > > Thanks > > Marcos > > > -----Mensagem original----- > De: vpp-dev@lists.fd.io <vpp-dev@lists.fd.io> Em nome de Damjan Marion via > lists.fd.io > Enviada em: quinta-feira, 26 de novembro de 2020 16:17 > Para: Marcos - Mgiga <mar...@mgiga.com.br> > Cc: Elias Rudberg <elias.rudb...@bahnhof.net>; vpp-dev@lists.fd.io > Assunto: Re: RES: [vpp-dev] NAT memory usage problem for VPP 20.09 compared > to 20.05 due to larger translation_buckets value > > > Sorry, I don’t understand your question. Can you elaborate further? > > -- > Damjan > >> On 26.11.2020., at 20:05, Marcos - Mgiga <mar...@mgiga.com.br> wrote: >> >> Hello, >> >> Taking benefit of the topic, how you suggest to monitor if translation hash >> bucket value has an appropriate value? What about translation hash memory, >> user hash buckets and user hash memory ? >> >> How do I know if I increase or decrease those values? >> >> Best Regards >> >> Marcos >> >> -----Mensagem original----- >> De: vpp-dev@lists.fd.io <vpp-dev@lists.fd.io> Em nome de Damjan Marion >> via lists.fd.io Enviada em: quinta-feira, 26 de novembro de 2020 14:53 >> Para: Elias Rudberg <elias.rudb...@bahnhof.net> >> Cc: vpp-dev@lists.fd.io >> Assunto: Re: [vpp-dev] NAT memory usage problem for VPP 20.09 compared >> to 20.05 due to larger translation_buckets value >> >> >> Dear Elias, >> >> Let me try to explain a bit underlying mechanics. >> Let’s assume your target number of sessions is 10M and we are talking about >> 16byte key size. >> That means each hash entry (KV) is 24 bytes (16 bytes key and 8 bytes value). >> >> In the setup you were mentioning, with 1<<20 buckets, your will need to fit >> 10 KVs into each bucket. >> Initial bihash bucket holds 4 KVs and to accomodate 10 keys (assuming that >> hash function gives us equal distribution) you will need to grow each bucket >> 2 times. Growing means doubling bucket size. >> So at the end you will have 1<<20 buckets where each holds 16 KVs. >> >> Math is: >> 1<<20 * (16 * 24 /* KV size in bytes */ + 8 /*bucket header size*/) Which >> means 392 MB of memory. >> >> If you keep target number of 10M sessions, but you increase number of >> buckets to 1 << 22 (which is roughly what formula bellow is trying to do) >> you end up with the following math: >> >> Math is: >> 1<<22 * (4 * 24 /* KV size in bytes */ + 8 /*bucket header size*/) Which >> means 416 MB of memory. >> >> So why 2nd one is better. Several reasons: >> >> - in first case you need to grow each bucket twice, that means >> allocating memory for the new bucket, copying existing data from the >> old bucket and putting old bucket to the free list. This operation >> increases key insertion time and lowers performance >> >> - growing will likely result in significant amount of old buckets >> sitting in the free list and they are effectively wasted memory >> (bihash tries to reuse that memory but at some point there is no >> demand anymore for smaller buckets) >> >> - performance-wise original bucket (one which first 4 KVs) is collocated >> with bucket header. >> This is new behaviour Dave introduced earlier this year (and I think it is >> present in 20.09). >> Bucket collocated with header means that there is no dependant >> prefetch needed as both header and at least part of data sits in the same >> cacheline. This significantly improveslookup performance. >> >> So in general, for best performance and optimal memory usage, number of >> buckets should be big enough so it unlikely grow with your target number of >> KVs. rule of thumb would be rounding target number of entries to closer >> power-of-2 value and then dividing that number with 2. >> For example, for 10M entries first lower power-of-2 number is 1<<23 (8M) and >> first higher is 1<<24 (16M). >> 1<<23 is closer, when we divide that by 2 we got 1<<22 (4M) buckets. >> >> Hope this explains…. >> >> — >> Damjan >> >> >>> On 26.11.2020., at 17:54, Elias Rudberg <elias.rudb...@bahnhof.net> wrote: >>> >>> Hello VPP experts, >>> >>> We are using VPP for NAT44 and are currently looking at how to move >>> from VPP 20.05 to 20.09. There are some differences in the way the >>> NAT plugin is configured. >>> >>> One difficulty for us is the maximum number of sessions allowed, we >>> need to handle large numbers of sessions so that limit can be >>> important for us. For VPP 20.05 we have used "translation hash >>> buckets 1048576" and then the maximum number of sessions per thread >>> becomes 10 times that because of this line in the source code in >>> snat_config(): >>> >>> sm->max_translations = 10 * translation_buckets; >>> >>> So then we got a limit of about 10 million sessions per thread, which >>> we have been happy with so far. >>> >>> With VPP 20.09 however, things have changed so that the maximum >>> number of sessions is now configured explicitly, and the relationship >>> between max_translations_per_thread and translation_buckets is no >>> longer a factor of 10 but instead given by the >>> nat_calc_bihash_buckets() >>> function: >>> >>> static u32 >>> nat_calc_bihash_buckets (u32 n_elts) >>> { >>> return 1 << (max_log2 (n_elts >> 1) + 1); } >>> >>> The above function corresponds to a factor of somewhere between 1 and >>> 2 instead of 10. So, if I understood this correctly, for a given >>> maximum number of sessions, the corresponding translation_buckets >>> value will be something like 5 to 10 times larger in VPP 20.09 >>> compared to how it was in VPP 20.05, leading to significantly >>> increased memory requirement given that we want to have the same >>> maximum number of sessions as before. >>> >>> It seems a little strange that the translation_buckets value would >>> change so much between VPP versions, was that change intentional? The >>> old relationship "max_translations = 10 * translation_buckets" seems >>> to have worked well in practice, at least for our use case. >>> >>> What could we do to get around this, if we want to switch to VPP >>> 20.09 but without reducing the maximum number of sessions? If we were >>> to simply divide the nat_calc_bihash_buckets() value by 8 or so to >>> make it more similar to how it was earlier, would that lead to other >>> problems? >>> >>> Best regards, >>> Elias >>> >>> >>> >> >> >> >> >> >
-=-=-=-=-=-=-=-=-=-=-=- Links: You receive all messages sent to this group. View/Reply Online (#18166): https://lists.fd.io/g/vpp-dev/message/18166 Mute This Topic: https://lists.fd.io/mt/78535814/21656 Group Owner: vpp-dev+ow...@lists.fd.io Unsubscribe: https://lists.fd.io/g/vpp-dev/unsub [arch...@mail-archive.com] -=-=-=-=-=-=-=-=-=-=-=-