Thanks a ton! We (at Laserlike) gave this a try and it basically cured our problem in one fell swoop. The GC cycles now take longer to run, especially under load, but memory growth isn't too bad. I know that Go's philosophy is not to add knobs to the GC, but there really needs to be a way to tell it "please don't confiscate 95% of my CPU for seconds at a time".
And for that matter, the fact that it can do this should really be documented somewhere. We spent a great deal of time just trying to diagnose the problem. On Thursday, July 27, 2017 at 8:06:39 AM UTC-7, Marko Kevac wrote: > > Hi. We have made a simple one line patch to Go that turns off GC assist. > Unfortunately, it was the only way. I feel your pain. > > $ git diff > diff --git a/src/runtime/malloc.go b/src/runtime/malloc.go > index 8850659748..a2c8697bf5 100644 > --- a/src/runtime/malloc.go > +++ b/src/runtime/malloc.go > @@ -602,7 +602,7 @@ func mallocgc(size uintptr, typ *_type, needzero bool) > unsafe.Pointer { > } > // Charge the allocation against the G. We'll account > // for internal fragmentation at the end of mallocgc. > - assistG.gcAssistBytes -= int64(size) > + assistG.gcAssistBytes = 1024 > > if assistG.gcAssistBytes < 0 { > // This G is in debt. Assist the GC to correct > > > Our CPU usage went down 6x: > https://twitter.com/mkevac/status/882289083132903424 > Latency is now negligible: > https://twitter.com/mkevac/status/882289349328596992 > https://twitter.com/mkevac/status/882289514437332992 > And memory usage increase is not so big: > https://twitter.com/mkevac/status/882289742901018625 > > On Tuesday, July 25, 2017 at 3:44:10 AM UTC+3, stb...@laserlike.com wrote: >> >> Hi, >> >> We are experiencing a problem that I believe may be related to issue >> 14812 <https://github.com/golang/go/issues/14812> but I wanted to ask >> here before adding to that case or filing a new issue. Of course, we’d also >> greatly appreciate any advice about how to make our program performant. >> >> Here is what we observe: at Laserlike one of our core user-facing >> services (the “leaf”) typically responds to a particular rpc in <400ms. >> During GC we see spikes in latency to >5s on some simple requests. The >> stop the world periods are short, so the GC spikes appear to happen at >> other times. >> >> We have been unable to replicate this in a simple program, but we did run >> our code in a test mode that repros it. In our test environment the server >> loads ~10 GB of persistent data (never deleted so doesn’t really need to be >> GCed), and we ask for 8 processors. We are running go version 1.8.3 on >> kubernetes on GCP machine of type n1-highmem-64. To create the problem we >> send the server a single request with >500 tasks.. >> >> >> This google drive folder has leaf-logs-full.redacted.txt as well as other >> performance tooling files >> <https://drive.google.com/drive/folders/0BypqYg6r4ebqb0lQNnM5eWtISms?usp=sharing>. >> >> A snippet from that log here shows normal query responses and timing: >> >> I0719 22:50:22.467367 leaf.go:363] Worker #5 done search for >> '[redacted]', took 0.013 seconds >> >> I0719 22:50:22.467406 leaf.go:225] Worker #5 starting search for >> '[redacted]' >> >> I0719 22:50:22.467486 leaf.go:363] Worker #6 done search for >> '[redacted]', took 0.001 seconds >> >> I0719 22:50:22.467520 leaf.go:225] Worker #6 starting search for >> '[redacted]' >> >> I0719 22:50:22.468050 leaf.go:363] Worker #9 done search for >> '[redacted]', took 0.005 seconds >> >> >> We have observed that if a GC cycle happens to start while serving >> traffic (which is often) AND there is a large amount of time spent in >> assist, then our serving latency skyrockets by 10x. In the log the >> slowdown commences roughly when pacer assist starts at I0719 >> 22:50:31.079283 and then reverts to normal latencies shortly after the gc >> cycle completes at I0719 22:50:36.806085. >> >> Below I copy parts of the log where we see latencies of up to 729ms on >> tasks. I also bold the line that shows 32929ms spent on alloc gc assist. >> >> We have captured an attached cpu profile during this time which seems to >> confirm a large amount of time spent in runtime.gcAssistAlloc.func1. >> >> >> Pardon our ignorance about GC in golang, but our hypothesis about what >> may be going wrong is that our large in-memory data structures are causing >> gc to often go into assist mode, and that for reasons we don’t understand >> malloc becomes expensive in that mode. Since we also create ~100k new data >> objects when processing user requests, we are guessing those allocs become >> very slow. Another piece of evidence for this hypothesis is that we have >> another (prototype) implementation of this kind of service that makes more >> use of object pools and doesn’t seem to have as much of slowdown during GC. >> >> Note on large in-memory data-structures: >> >> The principal data structures can be thought of as: >> >> Map[uint64][]byte (about 10M map entries, the slice lengths between 5K to >> 50K) (around ~10G total memory usage) >> >> Map[uint64][]uint64 (about 50M map entries, the slice lengths vary >> between 10 and 100K, in a zipfian distribution, about 3G total memory usage) >> >> These data structures mostly stay as is over the life of the program. >> >> We are trying to figure out how to solve this so would appreciate any >> advice. An engineer on our team wrote up the following ideas, none of which >> are that great: >> >> 1. >> >> Figure out a simple way to prevent our goroutines slowing down during >> GC. I had some hopes LockOSThread() could be made to work, but it didn't >> seem to help in my experiments. I'm not ruling this solution out >> entirely, >> but if it's the write barriers that are the main problem, I don't have >> much >> hope. >> 2. >> >> Run at least 2 replicas of all our servers. Manage their GC cycles >> ourselves, synchronized so that at most one replica is in GC at any given >> time. The clients should either send all requests to both replicas (and >> cancel when one replies), or use some more complicated Kubernetes and >> client logic so a GCing replica is never sent requests. This is the >> simplest solution that is likely to work, and doesn't require us to >> change >> our habits too much. It just costs more. :) >> 3. >> >> Refactor our current servers and program future servers with the >> explicit goal of reducing GC burden. Give our servers more memory and >> increase SetGCPercent() so that Garbage Collection happens less >> frequently. >> Use simpler data structures with fewer pointers to reduce the length of >> time the GC cycle lasts when it does run. This isn't a perfect solution, >> because the GC will still hit us on occasion, but it reduces the harm. >> 4. >> >> Throw out much of the convenience of Go and write our own >> allocators/shared pointers that index into a flat byte buffer, making no >> use of the GC. (Either do this for particular objects, or just write a >> general solution.) Make GC cycles so infrequent that we can either turn >> GC >> off completely without going OOM, or at least run it only once per day, >> during off hours. This seems very hard, and we'll be fighting with Go a >> lot. >> 5. >> >> Implement the core serving data structures of our service in C++, >> paired off with another server written in Go that manages populating that >> server’s data -- the Go server would manage the data in the C++ server >> via >> rpc, and query it over rpc. >> 6. >> >> The nuclear option: decide that Go is simply not a suitable language >> for our serving stack. Rewrite major servers in C++. >> >> >> >> Right now #5 is our (terrible but only feasible-seeming) choice. >> >> Appreciate any advice! >> >> Steve Baker >> >> >> >> >> Logs with gc log lines: >> >> >> pacer: assist ratio=+3.155887e+001 (scan 2669 MB in 10072->10157 MB) >> workers=2+0 >> >> I0719 22:50:31.097707 leaf.go:363] Worker #3 done search for >> '[redacted]', took 0.017 seconds >> >> I0719 22:50:31.098036 leaf.go:225] Worker #3 starting search for >> '[redacted]' >> >> I0719 22:50:31.133959 leaf.go:363] Worker #6 done search for >> '[redacted]', took 0.069 seconds >> >> I0719 22:50:31.134300 leaf.go:225] Worker #6 starting search for >> '[redacted]' >> >> I0719 22:50:31.142191 leaf.go:363] Worker #9 done search for >> '[redacted]', took 0.066 seconds >> >> I0719 22:50:31.142275 leaf.go:225] Worker #9 starting search for >> '[redacted]' >> >> I0719 22:50:31.173921 leaf.go:363] Worker #10 done search for >> '[redacted]', took 0.098 seconds >> >> I0719 22:50:31.174009 leaf.go:225] Worker #10 starting search for >> '[redacted]' >> >> I0719 22:50:31.328796 leaf.go:363] Worker #15 done search for >> '[redacted]', took 0.260 seconds >> >> I0719 22:50:31.330194 leaf.go:225] Worker #15 starting search for >> '[redacted]' >> >> I0719 22:50:31.395579 leaf.go:363] Worker #11 done search for >> '[redacted]', took 0.374 seconds >> >> >> <SNIP> >> >> I0719 22:50:36.317004 leaf.go:363] Worker #5 done search for >> '[redacted]', took 0.729 seconds >> >> I0719 22:50:36.317190 leaf.go:225] Worker #5 starting search for >> '[redacted]' >> >> I0719 22:50:36.370191 leaf.go:363] Worker #15 done search for >> '[redacted]', took 0.192 seconds >> >> I0719 22:50:36.371446 leaf.go:225] Worker #15 starting search for >> '[redacted]' >> >> I0719 22:50:36.421953 leaf.go:363] Worker #7 done search for >> '[redacted]', took 0.116 seconds >> >> I0719 22:50:36.422092 leaf.go:225] Worker #7 starting search for >> '[redacted]' >> >> I0719 22:50:36.570008 leaf.go:363] Worker #12 done search for >> '[redacted]', took 0.778 seconds >> >> I0719 22:50:36.572970 leaf.go:225] Worker #12 starting search for >> '[redacted]' >> >> I0719 22:50:36.573571 leaf.go:363] Worker #6 done search for >> '[redacted]', took 0.710 seconds >> >> I0719 22:50:36.573721 leaf.go:225] Worker #6 starting search for >> '[redacted]' >> >> pacer: H_m_prev=8875421544 h_t=+1.900000e-001 H_T=10561751637 >> h_a=+1.983366e-001 H_a=10635742360 h_g=+2.000000e-001 H_g=10650505852 >> u_a=+9.735779e-001 u_g=+2.500000e-001 W_a=2354135120 goalΔ=+1.000000e-002 >> actualΔ=+8.336587e-003 u_a/u_g=+3.894312e+000 >> >> gc 1072 @4337.737s 39%: 0.37+5688+0.32 ms clock, 2.2+32929/11310/21+1.9 >> ms cpu, 10072->10143->8661 MB, 10157 MB goal, 8 P >> >> >> -- You received this message because you are subscribed to the Google Groups "golang-nuts" group. To unsubscribe from this group and stop receiving emails from it, send an email to golang-nuts+unsubscr...@googlegroups.com. For more options, visit https://groups.google.com/d/optout.