On Tuesday, 15 November 2016 07:09:13 GMT Jerry Lundström wrote:
> On 11/15/16 06:18, Sara Dickinson wrote:
> > - It requires around an order of magnitude less CPU to compress the C-DNS
> > file.
> Can you clarify this, do you mean that it takes longer to compress PCAP
> then C-DNS? If so, do you have numbers? And how did you get them?
We looked at compressing raw PCAP versus the same data encoded as C-DNS as
part of the C-DNS design process. We ran
$ /usr/bin/time xz -9 <input>
and logged the reported user and system CPU times and maximum resident set
size. The idea was to get an idea of the resource consumption of the
compression, and look at the obvious question of whether it was better than
just compressing PCAP.
We used 'xz -9' to get some idea of the upper bound of achievable compression.
We looked at 5 minute capture samples for 3 different servers, with loadings I
might roughly describe as light, medium and heavy.
You'll need fixed font to comfortably view the following table of results:
File size CPU
Traffic Format Input Output VM space User System
Light PCAP 47.92M 4.14M 490.56M 8.27 7.29
Light C-DNS 5.12M 1.78M 113.44M 3.39 1.19
Medium PCAP 245.77M 17.96M 690.05M 102.53 15.91
Medium C-DNS 23.57M 6.88M 276.47M 9.38 0.11
Heavy PCAP 677.61M 45.25M 690.05M 180.15 6.61
Heavy C-DNS 62.85M 16.83M 620.37M 29.36 0.95
This is really what one would expect. C-DNS presents the compression with much
smaller input, so less to chew through, and groups data likely to be similar
together in its header tables. So the compressor is given an easier job, and
needs fewer resources.
--
Jim Hague - [email protected] Never trust a computer you can't lift.
_______________________________________________
DNSOP mailing list
[email protected]
https://www.ietf.org/mailman/listinfo/dnsop