c2047754c3
Compiler changes: * Change map assignment to use mapassign and assign value directly. * Change string iteration to use decoderune, faster for ASCII strings. * Change makeslice to take int, and use makeslice64 for larger values. * Add new noverflow field to hmap struct used for maps. Unresolved problems, to be fixed later: * Commented out test in go/types/sizes_test.go that doesn't compile. * Commented out reflect.TestStructOf test for padding after zero-sized field. Reviewed-on: https://go-review.googlesource.com/35231 gotools/: Updates for Go 1.8rc1. * Makefile.am (go_cmd_go_files): Add bug.go. (s-zdefaultcc): Write defaultPkgConfig. * Makefile.in: Rebuild. From-SVN: r244456
282 lines
5.1 KiB
Go
282 lines
5.1 KiB
Go
// Copyright 2009 The Go Authors. All rights reserved.
|
|
// Use of this source code is governed by a BSD-style
|
|
// license that can be found in the LICENSE file.
|
|
|
|
// GOMAXPROCS=10 go test
|
|
|
|
package sync_test
|
|
|
|
import (
|
|
"fmt"
|
|
"internal/testenv"
|
|
"os"
|
|
"os/exec"
|
|
"runtime"
|
|
"strings"
|
|
. "sync"
|
|
"testing"
|
|
)
|
|
|
|
func HammerSemaphore(s *uint32, loops int, cdone chan bool) {
|
|
for i := 0; i < loops; i++ {
|
|
Runtime_Semacquire(s)
|
|
Runtime_Semrelease(s)
|
|
}
|
|
cdone <- true
|
|
}
|
|
|
|
func TestSemaphore(t *testing.T) {
|
|
s := new(uint32)
|
|
*s = 1
|
|
c := make(chan bool)
|
|
for i := 0; i < 10; i++ {
|
|
go HammerSemaphore(s, 1000, c)
|
|
}
|
|
for i := 0; i < 10; i++ {
|
|
<-c
|
|
}
|
|
}
|
|
|
|
func BenchmarkUncontendedSemaphore(b *testing.B) {
|
|
s := new(uint32)
|
|
*s = 1
|
|
HammerSemaphore(s, b.N, make(chan bool, 2))
|
|
}
|
|
|
|
func BenchmarkContendedSemaphore(b *testing.B) {
|
|
b.StopTimer()
|
|
s := new(uint32)
|
|
*s = 1
|
|
c := make(chan bool)
|
|
defer runtime.GOMAXPROCS(runtime.GOMAXPROCS(2))
|
|
b.StartTimer()
|
|
|
|
go HammerSemaphore(s, b.N/2, c)
|
|
go HammerSemaphore(s, b.N/2, c)
|
|
<-c
|
|
<-c
|
|
}
|
|
|
|
func HammerMutex(m *Mutex, loops int, cdone chan bool) {
|
|
for i := 0; i < loops; i++ {
|
|
m.Lock()
|
|
m.Unlock()
|
|
}
|
|
cdone <- true
|
|
}
|
|
|
|
func TestMutex(t *testing.T) {
|
|
if n := runtime.SetMutexProfileFraction(1); n != 0 {
|
|
t.Logf("got mutexrate %d expected 0", n)
|
|
}
|
|
defer runtime.SetMutexProfileFraction(0)
|
|
m := new(Mutex)
|
|
c := make(chan bool)
|
|
for i := 0; i < 10; i++ {
|
|
go HammerMutex(m, 1000, c)
|
|
}
|
|
for i := 0; i < 10; i++ {
|
|
<-c
|
|
}
|
|
}
|
|
|
|
var misuseTests = []struct {
|
|
name string
|
|
f func()
|
|
}{
|
|
{
|
|
"Mutex.Unlock",
|
|
func() {
|
|
var mu Mutex
|
|
mu.Unlock()
|
|
},
|
|
},
|
|
{
|
|
"Mutex.Unlock2",
|
|
func() {
|
|
var mu Mutex
|
|
mu.Lock()
|
|
mu.Unlock()
|
|
mu.Unlock()
|
|
},
|
|
},
|
|
{
|
|
"RWMutex.Unlock",
|
|
func() {
|
|
var mu RWMutex
|
|
mu.Unlock()
|
|
},
|
|
},
|
|
{
|
|
"RWMutex.Unlock2",
|
|
func() {
|
|
var mu RWMutex
|
|
mu.RLock()
|
|
mu.Unlock()
|
|
},
|
|
},
|
|
{
|
|
"RWMutex.Unlock3",
|
|
func() {
|
|
var mu RWMutex
|
|
mu.Lock()
|
|
mu.Unlock()
|
|
mu.Unlock()
|
|
},
|
|
},
|
|
{
|
|
"RWMutex.RUnlock",
|
|
func() {
|
|
var mu RWMutex
|
|
mu.RUnlock()
|
|
},
|
|
},
|
|
{
|
|
"RWMutex.RUnlock2",
|
|
func() {
|
|
var mu RWMutex
|
|
mu.Lock()
|
|
mu.RUnlock()
|
|
},
|
|
},
|
|
{
|
|
"RWMutex.RUnlock3",
|
|
func() {
|
|
var mu RWMutex
|
|
mu.RLock()
|
|
mu.RUnlock()
|
|
mu.RUnlock()
|
|
},
|
|
},
|
|
}
|
|
|
|
func init() {
|
|
if len(os.Args) == 3 && os.Args[1] == "TESTMISUSE" {
|
|
for _, test := range misuseTests {
|
|
if test.name == os.Args[2] {
|
|
test.f()
|
|
fmt.Printf("test completed\n")
|
|
os.Exit(0)
|
|
}
|
|
}
|
|
fmt.Printf("unknown test\n")
|
|
os.Exit(0)
|
|
}
|
|
}
|
|
|
|
func TestMutexMisuse(t *testing.T) {
|
|
testenv.MustHaveExec(t)
|
|
for _, test := range misuseTests {
|
|
out, err := exec.Command(os.Args[0], "TESTMISUSE", test.name).CombinedOutput()
|
|
if err == nil || !strings.Contains(string(out), "unlocked") {
|
|
t.Errorf("%s: did not find failure with message about unlocked lock: %s\n%s\n", test.name, err, out)
|
|
}
|
|
}
|
|
}
|
|
|
|
func BenchmarkMutexUncontended(b *testing.B) {
|
|
type PaddedMutex struct {
|
|
Mutex
|
|
pad [128]uint8
|
|
}
|
|
b.RunParallel(func(pb *testing.PB) {
|
|
var mu PaddedMutex
|
|
for pb.Next() {
|
|
mu.Lock()
|
|
mu.Unlock()
|
|
}
|
|
})
|
|
}
|
|
|
|
func benchmarkMutex(b *testing.B, slack, work bool) {
|
|
var mu Mutex
|
|
if slack {
|
|
b.SetParallelism(10)
|
|
}
|
|
b.RunParallel(func(pb *testing.PB) {
|
|
foo := 0
|
|
for pb.Next() {
|
|
mu.Lock()
|
|
mu.Unlock()
|
|
if work {
|
|
for i := 0; i < 100; i++ {
|
|
foo *= 2
|
|
foo /= 2
|
|
}
|
|
}
|
|
}
|
|
_ = foo
|
|
})
|
|
}
|
|
|
|
func BenchmarkMutex(b *testing.B) {
|
|
benchmarkMutex(b, false, false)
|
|
}
|
|
|
|
func BenchmarkMutexSlack(b *testing.B) {
|
|
benchmarkMutex(b, true, false)
|
|
}
|
|
|
|
func BenchmarkMutexWork(b *testing.B) {
|
|
benchmarkMutex(b, false, true)
|
|
}
|
|
|
|
func BenchmarkMutexWorkSlack(b *testing.B) {
|
|
benchmarkMutex(b, true, true)
|
|
}
|
|
|
|
func BenchmarkMutexNoSpin(b *testing.B) {
|
|
// This benchmark models a situation where spinning in the mutex should be
|
|
// non-profitable and allows to confirm that spinning does not do harm.
|
|
// To achieve this we create excess of goroutines most of which do local work.
|
|
// These goroutines yield during local work, so that switching from
|
|
// a blocked goroutine to other goroutines is profitable.
|
|
// As a matter of fact, this benchmark still triggers some spinning in the mutex.
|
|
var m Mutex
|
|
var acc0, acc1 uint64
|
|
b.SetParallelism(4)
|
|
b.RunParallel(func(pb *testing.PB) {
|
|
c := make(chan bool)
|
|
var data [4 << 10]uint64
|
|
for i := 0; pb.Next(); i++ {
|
|
if i%4 == 0 {
|
|
m.Lock()
|
|
acc0 -= 100
|
|
acc1 += 100
|
|
m.Unlock()
|
|
} else {
|
|
for i := 0; i < len(data); i += 4 {
|
|
data[i]++
|
|
}
|
|
// Elaborate way to say runtime.Gosched
|
|
// that does not put the goroutine onto global runq.
|
|
go func() {
|
|
c <- true
|
|
}()
|
|
<-c
|
|
}
|
|
}
|
|
})
|
|
}
|
|
|
|
func BenchmarkMutexSpin(b *testing.B) {
|
|
// This benchmark models a situation where spinning in the mutex should be
|
|
// profitable. To achieve this we create a goroutine per-proc.
|
|
// These goroutines access considerable amount of local data so that
|
|
// unnecessary rescheduling is penalized by cache misses.
|
|
var m Mutex
|
|
var acc0, acc1 uint64
|
|
b.RunParallel(func(pb *testing.PB) {
|
|
var data [16 << 10]uint64
|
|
for i := 0; pb.Next(); i++ {
|
|
m.Lock()
|
|
acc0 -= 100
|
|
acc1 += 100
|
|
m.Unlock()
|
|
for i := 0; i < len(data); i += 4 {
|
|
data[i]++
|
|
}
|
|
}
|
|
})
|
|
}
|