// Copyright 2013 The Go Authors. All rights reserved. // Use of this source code is governed by a BSD-style // license that can be found in the LICENSE file. // Pool is no-op under race detector, so all these tests do not work. // +build !race package sync_test import ( "runtime" "runtime/debug" "sort" . "sync" "sync/atomic" "testing" "time" ) func TestPool(t *testing.T) { // disable GC so we can control when it happens. defer debug.SetGCPercent(debug.SetGCPercent(-1)) var p Pool if p.Get() != nil { t.Fatal("expected empty") } // Make sure that the goroutine doesn't migrate to another P // between Put and Get calls. Runtime_procPin() p.Put("a") p.Put("b") if g := p.Get(); g != "a" { t.Fatalf("got %#v; want a", g) } if g := p.Get(); g != "b" { t.Fatalf("got %#v; want b", g) } if g := p.Get(); g != nil { t.Fatalf("got %#v; want nil", g) } Runtime_procUnpin() // Put in a large number of objects so they spill into // stealable space. for i := 0; i < 100; i++ { p.Put("c") } // After one GC, the victim cache should keep them alive. runtime.GC() if g := p.Get(); g != "c" { t.Fatalf("got %#v; want c after GC", g) } // A second GC should drop the victim cache. runtime.GC() if g := p.Get(); g != nil { t.Fatalf("got %#v; want nil after second GC", g) } } func TestPoolNew(t *testing.T) { // disable GC so we can control when it happens. defer debug.SetGCPercent(debug.SetGCPercent(-1)) i := 0 p := Pool{ New: func() interface{} { i++ return i }, } if v := p.Get(); v != 1 { t.Fatalf("got %v; want 1", v) } if v := p.Get(); v != 2 { t.Fatalf("got %v; want 2", v) } // Make sure that the goroutine doesn't migrate to another P // between Put and Get calls. Runtime_procPin() p.Put(42) if v := p.Get(); v != 42 { t.Fatalf("got %v; want 42", v) } Runtime_procUnpin() if v := p.Get(); v != 3 { t.Fatalf("got %v; want 3", v) } } // Test that Pool does not hold pointers to previously cached resources. func TestPoolGC(t *testing.T) { testPool(t, true) } // Test that Pool releases resources on GC. func TestPoolRelease(t *testing.T) { testPool(t, false) } func testPool(t *testing.T, drain bool) { t.Skip("gccgo imprecise GC breaks this test") var p Pool const N = 100 loop: for try := 0; try < 3; try++ { if try == 1 && testing.Short() { break } var fin, fin1 uint32 for i := 0; i < N; i++ { v := new(string) runtime.SetFinalizer(v, func(vv *string) { atomic.AddUint32(&fin, 1) }) p.Put(v) } if drain { for i := 0; i < N; i++ { p.Get() } } for i := 0; i < 5; i++ { runtime.GC() time.Sleep(time.Duration(i*100+10) * time.Millisecond) // 1 pointer can remain on stack or elsewhere if fin1 = atomic.LoadUint32(&fin); fin1 >= N-1 { continue loop } } t.Fatalf("only %v out of %v resources are finalized on try %v", fin1, N, try) } } func TestPoolStress(t *testing.T) { const P = 10 N := int(1e6) if testing.Short() { N /= 100 } var p Pool done := make(chan bool) for i := 0; i < P; i++ { go func() { var v interface{} = 0 for j := 0; j < N; j++ { if v == nil { v = 0 } p.Put(v) v = p.Get() if v != nil && v.(int) != 0 { t.Errorf("expect 0, got %v", v) break } } done <- true }() } for i := 0; i < P; i++ { <-done } } func TestPoolDequeue(t *testing.T) { testPoolDequeue(t, NewPoolDequeue(16)) } func TestPoolChain(t *testing.T) { testPoolDequeue(t, NewPoolChain()) } func testPoolDequeue(t *testing.T, d PoolDequeue) { const P = 10 var N int = 2e6 if testing.Short() { N = 1e3 } have := make([]int32, N) var stop int32 var wg WaitGroup record := func(val int) { atomic.AddInt32(&have[val], 1) if val == N-1 { atomic.StoreInt32(&stop, 1) } } // Start P-1 consumers. for i := 1; i < P; i++ { wg.Add(1) go func() { fail := 0 for atomic.LoadInt32(&stop) == 0 { val, ok := d.PopTail() if ok { fail = 0 record(val.(int)) } else { // Speed up the test by // allowing the pusher to run. if fail++; fail%100 == 0 { runtime.Gosched() } } } wg.Done() }() } // Start 1 producer. nPopHead := 0 wg.Add(1) go func() { for j := 0; j < N; j++ { for !d.PushHead(j) { // Allow a popper to run. runtime.Gosched() } if j%10 == 0 { val, ok := d.PopHead() if ok { nPopHead++ record(val.(int)) } } } wg.Done() }() wg.Wait() // Check results. for i, count := range have { if count != 1 { t.Errorf("expected have[%d] = 1, got %d", i, count) } } // Check that at least some PopHeads succeeded. We skip this // check in short mode because it's common enough that the // queue will stay nearly empty all the time and a PopTail // will happen during the window between every PushHead and // PopHead. if !testing.Short() && nPopHead == 0 { t.Errorf("popHead never succeeded") } } func BenchmarkPool(b *testing.B) { var p Pool b.RunParallel(func(pb *testing.PB) { for pb.Next() { p.Put(1) p.Get() } }) } func BenchmarkPoolOverflow(b *testing.B) { var p Pool b.RunParallel(func(pb *testing.PB) { for pb.Next() { for b := 0; b < 100; b++ { p.Put(1) } for b := 0; b < 100; b++ { p.Get() } } }) } var globalSink interface{} func BenchmarkPoolSTW(b *testing.B) { // Take control of GC. defer debug.SetGCPercent(debug.SetGCPercent(-1)) var mstats runtime.MemStats var pauses []uint64 var p Pool for i := 0; i < b.N; i++ { // Put a large number of items into a pool. const N = 100000 var item interface{} = 42 for i := 0; i < N; i++ { p.Put(item) } // Do a GC. runtime.GC() // Record pause time. runtime.ReadMemStats(&mstats) pauses = append(pauses, mstats.PauseNs[(mstats.NumGC+255)%256]) } // Get pause time stats. sort.Slice(pauses, func(i, j int) bool { return pauses[i] < pauses[j] }) var total uint64 for _, ns := range pauses { total += ns } // ns/op for this benchmark is average STW time. b.ReportMetric(float64(total)/float64(b.N), "ns/op") b.ReportMetric(float64(pauses[len(pauses)*95/100]), "p95-ns/STW") b.ReportMetric(float64(pauses[len(pauses)*50/100]), "p50-ns/STW") } func BenchmarkPoolExpensiveNew(b *testing.B) { // Populate a pool with items that are expensive to construct // to stress pool cleanup and subsequent reconstruction. // Create a ballast so the GC has a non-zero heap size and // runs at reasonable times. globalSink = make([]byte, 8<<20) defer func() { globalSink = nil }() // Create a pool that's "expensive" to fill. var p Pool var nNew uint64 p.New = func() interface{} { atomic.AddUint64(&nNew, 1) time.Sleep(time.Millisecond) return 42 } var mstats1, mstats2 runtime.MemStats runtime.ReadMemStats(&mstats1) b.RunParallel(func(pb *testing.PB) { // Simulate 100X the number of goroutines having items // checked out from the Pool simultaneously. items := make([]interface{}, 100) var sink []byte for pb.Next() { // Stress the pool. for i := range items { items[i] = p.Get() // Simulate doing some work with this // item checked out. sink = make([]byte, 32<<10) } for i, v := range items { p.Put(v) items[i] = nil } } _ = sink }) runtime.ReadMemStats(&mstats2) b.ReportMetric(float64(mstats2.NumGC-mstats1.NumGC)/float64(b.N), "GCs/op") b.ReportMetric(float64(nNew)/float64(b.N), "New/op") }