Re: [go-nuts] Sort a huge slice of data around 2GB

2017-11-29 Thread Michael Jones
agree! On Wed, Nov 29, 2017 at 11:37 AM, 'Axel Wagner' via golang-nuts < golang-nuts@googlegroups.com> wrote: > BTW, depending on what the problem is we are actually talking about, even > the numbers from the initial post might be fine. With those numbers, > sorting the 2GB should take roughly ha

Re: [go-nuts] Sort a huge slice of data around 2GB

2017-11-29 Thread 'Axel Wagner' via golang-nuts
BTW, depending on what the problem is we are actually talking about, even the numbers from the initial post might be fine. With those numbers, sorting the 2GB should take roughly half an hour. A significant amount of time, yes, but if you only need to do it once (or once a day or week or whatever)

Re: [go-nuts] Sort a huge slice of data around 2GB

2017-11-29 Thread Michael Jones
here's a little more detail on system / direct-type / and parallel direct-type sorting: celeste:double mtj$ go test -run=NONE -bench Sort goos: darwin goarch: amd64 pkg: double BenchmarkSSort250-8 10 11285 ns/op BenchmarkSSort500-8 5 31001 ns/op BenchmarkSSort

Re: [go-nuts] Sort a huge slice of data around 2GB

2017-11-29 Thread Michael Jones
Using a personal sort to avoid the standard sort's abstraction... // 50331648 bytes,6291456 8-byte elements, 0.687317 seconds (48MB of uint64) // 50331648 bytes, 50331648 1-byte elements, 1.605258 seconds (48MB of uint8) // 2147483648 bytes, 268435456 8-byte elements, 35.262593 seconds (2

Re: [go-nuts] Sort a huge slice of data around 2GB

2017-11-29 Thread 'Axel Wagner' via golang-nuts
I'm not sure that the sort package is your your problem, here. First, the actual size (in bytes) of the data set matters, but is not *that* important. The number of elements is much more important. And I wrote a naive program to measure how long it takes to sort different in-memory data sets and ge

Re: [go-nuts] Sort a huge slice of data around 2GB

2017-11-29 Thread Subramanian Karunanithi
Hi, Yes measuring it now shall get back, on a side note do we have any big datafile parsing package part of go or some library? The basic sort.Sort() what I have is taking 44sec for 48MB of data, I have to parse 2G of data. Regards, Subu K On Wed, Nov 29, 2017 at 8:03 PM, Jan Mercl <0xj...@gmai

Re: [go-nuts] Sort a huge slice of data around 2GB

2017-11-29 Thread Jan Mercl
On Wed, Nov 29, 2017 at 3:19 PM Subramanian K wrote: > To run 2GB of data it takes really long time, I am trying to split these to buckets and make it run concurrently, finally need to collate results of all these small sorted buckets. Have you measured and detected where the bottleneck is? If i

[go-nuts] Sort a huge slice of data around 2GB

2017-11-29 Thread Subramanian K
Hi I am using native sort provided part of the package and to process 48MB of slice data, it takes ~45sec. To run 2GB of data it takes really long time, I am trying to split these to buckets and make it run concurrently, finally need to collate results of all these small sorted buckets. Do we