| // Copyright 2013 The Go Authors. All rights reserved. | 
 | // Use of this source code is governed by a BSD-style | 
 | // license that can be found in the LICENSE file. | 
 |  | 
 | package runtime_test | 
 |  | 
 | import ( | 
 | 	"crypto/rand" | 
 | 	"encoding/binary" | 
 | 	"fmt" | 
 | 	"internal/race" | 
 | 	"internal/testenv" | 
 | 	. "runtime" | 
 | 	"sync/atomic" | 
 | 	"testing" | 
 | 	"unsafe" | 
 | ) | 
 |  | 
 | func TestMemmove(t *testing.T) { | 
 | 	if *flagQuick { | 
 | 		t.Skip("-quick") | 
 | 	} | 
 | 	t.Parallel() | 
 | 	size := 256 | 
 | 	if testing.Short() { | 
 | 		size = 128 + 16 | 
 | 	} | 
 | 	src := make([]byte, size) | 
 | 	dst := make([]byte, size) | 
 | 	for i := 0; i < size; i++ { | 
 | 		src[i] = byte(128 + (i & 127)) | 
 | 	} | 
 | 	for i := 0; i < size; i++ { | 
 | 		dst[i] = byte(i & 127) | 
 | 	} | 
 | 	for n := 0; n <= size; n++ { | 
 | 		for x := 0; x <= size-n; x++ { // offset in src | 
 | 			for y := 0; y <= size-n; y++ { // offset in dst | 
 | 				copy(dst[y:y+n], src[x:x+n]) | 
 | 				for i := 0; i < y; i++ { | 
 | 					if dst[i] != byte(i&127) { | 
 | 						t.Fatalf("prefix dst[%d] = %d", i, dst[i]) | 
 | 					} | 
 | 				} | 
 | 				for i := y; i < y+n; i++ { | 
 | 					if dst[i] != byte(128+((i-y+x)&127)) { | 
 | 						t.Fatalf("copied dst[%d] = %d", i, dst[i]) | 
 | 					} | 
 | 					dst[i] = byte(i & 127) // reset dst | 
 | 				} | 
 | 				for i := y + n; i < size; i++ { | 
 | 					if dst[i] != byte(i&127) { | 
 | 						t.Fatalf("suffix dst[%d] = %d", i, dst[i]) | 
 | 					} | 
 | 				} | 
 | 			} | 
 | 		} | 
 | 	} | 
 | } | 
 |  | 
 | func TestMemmoveAlias(t *testing.T) { | 
 | 	if *flagQuick { | 
 | 		t.Skip("-quick") | 
 | 	} | 
 | 	t.Parallel() | 
 | 	size := 256 | 
 | 	if testing.Short() { | 
 | 		size = 128 + 16 | 
 | 	} | 
 | 	buf := make([]byte, size) | 
 | 	for i := 0; i < size; i++ { | 
 | 		buf[i] = byte(i) | 
 | 	} | 
 | 	for n := 0; n <= size; n++ { | 
 | 		for x := 0; x <= size-n; x++ { // src offset | 
 | 			for y := 0; y <= size-n; y++ { // dst offset | 
 | 				copy(buf[y:y+n], buf[x:x+n]) | 
 | 				for i := 0; i < y; i++ { | 
 | 					if buf[i] != byte(i) { | 
 | 						t.Fatalf("prefix buf[%d] = %d", i, buf[i]) | 
 | 					} | 
 | 				} | 
 | 				for i := y; i < y+n; i++ { | 
 | 					if buf[i] != byte(i-y+x) { | 
 | 						t.Fatalf("copied buf[%d] = %d", i, buf[i]) | 
 | 					} | 
 | 					buf[i] = byte(i) // reset buf | 
 | 				} | 
 | 				for i := y + n; i < size; i++ { | 
 | 					if buf[i] != byte(i) { | 
 | 						t.Fatalf("suffix buf[%d] = %d", i, buf[i]) | 
 | 					} | 
 | 				} | 
 | 			} | 
 | 		} | 
 | 	} | 
 | } | 
 |  | 
 | func TestMemmoveLarge0x180000(t *testing.T) { | 
 | 	if testing.Short() && testenv.Builder() == "" { | 
 | 		t.Skip("-short") | 
 | 	} | 
 |  | 
 | 	t.Parallel() | 
 | 	if race.Enabled { | 
 | 		t.Skip("skipping large memmove test under race detector") | 
 | 	} | 
 | 	testSize(t, 0x180000) | 
 | } | 
 |  | 
 | func TestMemmoveOverlapLarge0x120000(t *testing.T) { | 
 | 	if testing.Short() && testenv.Builder() == "" { | 
 | 		t.Skip("-short") | 
 | 	} | 
 |  | 
 | 	t.Parallel() | 
 | 	if race.Enabled { | 
 | 		t.Skip("skipping large memmove test under race detector") | 
 | 	} | 
 | 	testOverlap(t, 0x120000) | 
 | } | 
 |  | 
 | func testSize(t *testing.T, size int) { | 
 | 	src := make([]byte, size) | 
 | 	dst := make([]byte, size) | 
 | 	_, _ = rand.Read(src) | 
 | 	_, _ = rand.Read(dst) | 
 |  | 
 | 	ref := make([]byte, size) | 
 | 	copyref(ref, dst) | 
 |  | 
 | 	for n := size - 50; n > 1; n >>= 1 { | 
 | 		for x := 0; x <= size-n; x = x*7 + 1 { // offset in src | 
 | 			for y := 0; y <= size-n; y = y*9 + 1 { // offset in dst | 
 | 				copy(dst[y:y+n], src[x:x+n]) | 
 | 				copyref(ref[y:y+n], src[x:x+n]) | 
 | 				p := cmpb(dst, ref) | 
 | 				if p >= 0 { | 
 | 					t.Fatalf("Copy failed, copying from src[%d:%d] to dst[%d:%d].\nOffset %d is different, %v != %v", x, x+n, y, y+n, p, dst[p], ref[p]) | 
 | 				} | 
 | 			} | 
 | 		} | 
 | 	} | 
 | } | 
 |  | 
 | func testOverlap(t *testing.T, size int) { | 
 | 	src := make([]byte, size) | 
 | 	test := make([]byte, size) | 
 | 	ref := make([]byte, size) | 
 | 	_, _ = rand.Read(src) | 
 |  | 
 | 	for n := size - 50; n > 1; n >>= 1 { | 
 | 		for x := 0; x <= size-n; x = x*7 + 1 { // offset in src | 
 | 			for y := 0; y <= size-n; y = y*9 + 1 { // offset in dst | 
 | 				// Reset input | 
 | 				copyref(test, src) | 
 | 				copyref(ref, src) | 
 | 				copy(test[y:y+n], test[x:x+n]) | 
 | 				if y <= x { | 
 | 					copyref(ref[y:y+n], ref[x:x+n]) | 
 | 				} else { | 
 | 					copybw(ref[y:y+n], ref[x:x+n]) | 
 | 				} | 
 | 				p := cmpb(test, ref) | 
 | 				if p >= 0 { | 
 | 					t.Fatalf("Copy failed, copying from src[%d:%d] to dst[%d:%d].\nOffset %d is different, %v != %v", x, x+n, y, y+n, p, test[p], ref[p]) | 
 | 				} | 
 | 			} | 
 | 		} | 
 | 	} | 
 |  | 
 | } | 
 |  | 
 | // Forward copy. | 
 | func copyref(dst, src []byte) { | 
 | 	for i, v := range src { | 
 | 		dst[i] = v | 
 | 	} | 
 | } | 
 |  | 
 | // Backwards copy | 
 | func copybw(dst, src []byte) { | 
 | 	if len(src) == 0 { | 
 | 		return | 
 | 	} | 
 | 	for i := len(src) - 1; i >= 0; i-- { | 
 | 		dst[i] = src[i] | 
 | 	} | 
 | } | 
 |  | 
 | // Returns offset of difference | 
 | func matchLen(a, b []byte, max int) int { | 
 | 	a = a[:max] | 
 | 	b = b[:max] | 
 | 	for i, av := range a { | 
 | 		if b[i] != av { | 
 | 			return i | 
 | 		} | 
 | 	} | 
 | 	return max | 
 | } | 
 |  | 
 | func cmpb(a, b []byte) int { | 
 | 	l := matchLen(a, b, len(a)) | 
 | 	if l == len(a) { | 
 | 		return -1 | 
 | 	} | 
 | 	return l | 
 | } | 
 |  | 
 | // Ensure that memmove writes pointers atomically, so the GC won't | 
 | // observe a partially updated pointer. | 
 | func TestMemmoveAtomicity(t *testing.T) { | 
 | 	if race.Enabled { | 
 | 		t.Skip("skip under the race detector -- this test is intentionally racy") | 
 | 	} | 
 |  | 
 | 	var x int | 
 |  | 
 | 	for _, backward := range []bool{true, false} { | 
 | 		for _, n := range []int{3, 4, 5, 6, 7, 8, 9, 10, 15, 25, 49} { | 
 | 			n := n | 
 |  | 
 | 			// test copying [N]*int. | 
 | 			sz := uintptr(n * PtrSize) | 
 | 			name := fmt.Sprint(sz) | 
 | 			if backward { | 
 | 				name += "-backward" | 
 | 			} else { | 
 | 				name += "-forward" | 
 | 			} | 
 | 			t.Run(name, func(t *testing.T) { | 
 | 				// Use overlapping src and dst to force forward/backward copy. | 
 | 				var s [100]*int | 
 | 				src := s[n-1 : 2*n-1] | 
 | 				dst := s[:n] | 
 | 				if backward { | 
 | 					src, dst = dst, src | 
 | 				} | 
 | 				for i := range src { | 
 | 					src[i] = &x | 
 | 				} | 
 | 				for i := range dst { | 
 | 					dst[i] = nil | 
 | 				} | 
 |  | 
 | 				var ready uint32 | 
 | 				go func() { | 
 | 					sp := unsafe.Pointer(&src[0]) | 
 | 					dp := unsafe.Pointer(&dst[0]) | 
 | 					atomic.StoreUint32(&ready, 1) | 
 | 					for i := 0; i < 10000; i++ { | 
 | 						Memmove(dp, sp, sz) | 
 | 						MemclrNoHeapPointers(dp, sz) | 
 | 					} | 
 | 					atomic.StoreUint32(&ready, 2) | 
 | 				}() | 
 |  | 
 | 				for atomic.LoadUint32(&ready) == 0 { | 
 | 					Gosched() | 
 | 				} | 
 |  | 
 | 				for atomic.LoadUint32(&ready) != 2 { | 
 | 					for i := range dst { | 
 | 						p := dst[i] | 
 | 						if p != nil && p != &x { | 
 | 							t.Fatalf("got partially updated pointer %p at dst[%d], want either nil or %p", p, i, &x) | 
 | 						} | 
 | 					} | 
 | 				} | 
 | 			}) | 
 | 		} | 
 | 	} | 
 | } | 
 |  | 
 | func benchmarkSizes(b *testing.B, sizes []int, fn func(b *testing.B, n int)) { | 
 | 	for _, n := range sizes { | 
 | 		b.Run(fmt.Sprint(n), func(b *testing.B) { | 
 | 			b.SetBytes(int64(n)) | 
 | 			fn(b, n) | 
 | 		}) | 
 | 	} | 
 | } | 
 |  | 
 | var bufSizes = []int{ | 
 | 	0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, | 
 | 	32, 64, 128, 256, 512, 1024, 2048, 4096, | 
 | } | 
 | var bufSizesOverlap = []int{ | 
 | 	32, 64, 128, 256, 512, 1024, 2048, 4096, | 
 | } | 
 |  | 
 | func BenchmarkMemmove(b *testing.B) { | 
 | 	benchmarkSizes(b, bufSizes, func(b *testing.B, n int) { | 
 | 		x := make([]byte, n) | 
 | 		y := make([]byte, n) | 
 | 		for i := 0; i < b.N; i++ { | 
 | 			copy(x, y) | 
 | 		} | 
 | 	}) | 
 | } | 
 |  | 
 | func BenchmarkMemmoveOverlap(b *testing.B) { | 
 | 	benchmarkSizes(b, bufSizesOverlap, func(b *testing.B, n int) { | 
 | 		x := make([]byte, n+16) | 
 | 		for i := 0; i < b.N; i++ { | 
 | 			copy(x[16:n+16], x[:n]) | 
 | 		} | 
 | 	}) | 
 | } | 
 |  | 
 | func BenchmarkMemmoveUnalignedDst(b *testing.B) { | 
 | 	benchmarkSizes(b, bufSizes, func(b *testing.B, n int) { | 
 | 		x := make([]byte, n+1) | 
 | 		y := make([]byte, n) | 
 | 		for i := 0; i < b.N; i++ { | 
 | 			copy(x[1:], y) | 
 | 		} | 
 | 	}) | 
 | } | 
 |  | 
 | func BenchmarkMemmoveUnalignedDstOverlap(b *testing.B) { | 
 | 	benchmarkSizes(b, bufSizesOverlap, func(b *testing.B, n int) { | 
 | 		x := make([]byte, n+16) | 
 | 		for i := 0; i < b.N; i++ { | 
 | 			copy(x[16:n+16], x[1:n+1]) | 
 | 		} | 
 | 	}) | 
 | } | 
 |  | 
 | func BenchmarkMemmoveUnalignedSrc(b *testing.B) { | 
 | 	benchmarkSizes(b, bufSizes, func(b *testing.B, n int) { | 
 | 		x := make([]byte, n) | 
 | 		y := make([]byte, n+1) | 
 | 		for i := 0; i < b.N; i++ { | 
 | 			copy(x, y[1:]) | 
 | 		} | 
 | 	}) | 
 | } | 
 |  | 
 | func BenchmarkMemmoveUnalignedSrcOverlap(b *testing.B) { | 
 | 	benchmarkSizes(b, bufSizesOverlap, func(b *testing.B, n int) { | 
 | 		x := make([]byte, n+1) | 
 | 		for i := 0; i < b.N; i++ { | 
 | 			copy(x[1:n+1], x[:n]) | 
 | 		} | 
 | 	}) | 
 | } | 
 |  | 
 | func TestMemclr(t *testing.T) { | 
 | 	size := 512 | 
 | 	if testing.Short() { | 
 | 		size = 128 + 16 | 
 | 	} | 
 | 	mem := make([]byte, size) | 
 | 	for i := 0; i < size; i++ { | 
 | 		mem[i] = 0xee | 
 | 	} | 
 | 	for n := 0; n < size; n++ { | 
 | 		for x := 0; x <= size-n; x++ { // offset in mem | 
 | 			MemclrBytes(mem[x : x+n]) | 
 | 			for i := 0; i < x; i++ { | 
 | 				if mem[i] != 0xee { | 
 | 					t.Fatalf("overwrite prefix mem[%d] = %d", i, mem[i]) | 
 | 				} | 
 | 			} | 
 | 			for i := x; i < x+n; i++ { | 
 | 				if mem[i] != 0 { | 
 | 					t.Fatalf("failed clear mem[%d] = %d", i, mem[i]) | 
 | 				} | 
 | 				mem[i] = 0xee | 
 | 			} | 
 | 			for i := x + n; i < size; i++ { | 
 | 				if mem[i] != 0xee { | 
 | 					t.Fatalf("overwrite suffix mem[%d] = %d", i, mem[i]) | 
 | 				} | 
 | 			} | 
 | 		} | 
 | 	} | 
 | } | 
 |  | 
 | func BenchmarkMemclr(b *testing.B) { | 
 | 	for _, n := range []int{5, 16, 64, 256, 4096, 65536} { | 
 | 		x := make([]byte, n) | 
 | 		b.Run(fmt.Sprint(n), func(b *testing.B) { | 
 | 			b.SetBytes(int64(n)) | 
 | 			for i := 0; i < b.N; i++ { | 
 | 				MemclrBytes(x) | 
 | 			} | 
 | 		}) | 
 | 	} | 
 | 	for _, m := range []int{1, 4, 8, 16, 64} { | 
 | 		x := make([]byte, m<<20) | 
 | 		b.Run(fmt.Sprint(m, "M"), func(b *testing.B) { | 
 | 			b.SetBytes(int64(m << 20)) | 
 | 			for i := 0; i < b.N; i++ { | 
 | 				MemclrBytes(x) | 
 | 			} | 
 | 		}) | 
 | 	} | 
 | } | 
 |  | 
 | func BenchmarkGoMemclr(b *testing.B) { | 
 | 	benchmarkSizes(b, []int{5, 16, 64, 256}, func(b *testing.B, n int) { | 
 | 		x := make([]byte, n) | 
 | 		for i := 0; i < b.N; i++ { | 
 | 			for j := range x { | 
 | 				x[j] = 0 | 
 | 			} | 
 | 		} | 
 | 	}) | 
 | } | 
 |  | 
 | func BenchmarkClearFat8(b *testing.B) { | 
 | 	for i := 0; i < b.N; i++ { | 
 | 		var x [8 / 4]uint32 | 
 | 		_ = x | 
 | 	} | 
 | } | 
 | func BenchmarkClearFat12(b *testing.B) { | 
 | 	for i := 0; i < b.N; i++ { | 
 | 		var x [12 / 4]uint32 | 
 | 		_ = x | 
 | 	} | 
 | } | 
 | func BenchmarkClearFat16(b *testing.B) { | 
 | 	for i := 0; i < b.N; i++ { | 
 | 		var x [16 / 4]uint32 | 
 | 		_ = x | 
 | 	} | 
 | } | 
 | func BenchmarkClearFat24(b *testing.B) { | 
 | 	for i := 0; i < b.N; i++ { | 
 | 		var x [24 / 4]uint32 | 
 | 		_ = x | 
 | 	} | 
 | } | 
 | func BenchmarkClearFat32(b *testing.B) { | 
 | 	for i := 0; i < b.N; i++ { | 
 | 		var x [32 / 4]uint32 | 
 | 		_ = x | 
 | 	} | 
 | } | 
 | func BenchmarkClearFat40(b *testing.B) { | 
 | 	for i := 0; i < b.N; i++ { | 
 | 		var x [40 / 4]uint32 | 
 | 		_ = x | 
 | 	} | 
 | } | 
 | func BenchmarkClearFat48(b *testing.B) { | 
 | 	for i := 0; i < b.N; i++ { | 
 | 		var x [48 / 4]uint32 | 
 | 		_ = x | 
 | 	} | 
 | } | 
 | func BenchmarkClearFat56(b *testing.B) { | 
 | 	for i := 0; i < b.N; i++ { | 
 | 		var x [56 / 4]uint32 | 
 | 		_ = x | 
 | 	} | 
 | } | 
 | func BenchmarkClearFat64(b *testing.B) { | 
 | 	for i := 0; i < b.N; i++ { | 
 | 		var x [64 / 4]uint32 | 
 | 		_ = x | 
 | 	} | 
 | } | 
 | func BenchmarkClearFat128(b *testing.B) { | 
 | 	for i := 0; i < b.N; i++ { | 
 | 		var x [128 / 4]uint32 | 
 | 		_ = x | 
 | 	} | 
 | } | 
 | func BenchmarkClearFat256(b *testing.B) { | 
 | 	for i := 0; i < b.N; i++ { | 
 | 		var x [256 / 4]uint32 | 
 | 		_ = x | 
 | 	} | 
 | } | 
 | func BenchmarkClearFat512(b *testing.B) { | 
 | 	for i := 0; i < b.N; i++ { | 
 | 		var x [512 / 4]uint32 | 
 | 		_ = x | 
 | 	} | 
 | } | 
 | func BenchmarkClearFat1024(b *testing.B) { | 
 | 	for i := 0; i < b.N; i++ { | 
 | 		var x [1024 / 4]uint32 | 
 | 		_ = x | 
 | 	} | 
 | } | 
 |  | 
 | func BenchmarkCopyFat8(b *testing.B) { | 
 | 	var x [8 / 4]uint32 | 
 | 	for i := 0; i < b.N; i++ { | 
 | 		y := x | 
 | 		_ = y | 
 | 	} | 
 | } | 
 | func BenchmarkCopyFat12(b *testing.B) { | 
 | 	var x [12 / 4]uint32 | 
 | 	for i := 0; i < b.N; i++ { | 
 | 		y := x | 
 | 		_ = y | 
 | 	} | 
 | } | 
 | func BenchmarkCopyFat16(b *testing.B) { | 
 | 	var x [16 / 4]uint32 | 
 | 	for i := 0; i < b.N; i++ { | 
 | 		y := x | 
 | 		_ = y | 
 | 	} | 
 | } | 
 | func BenchmarkCopyFat24(b *testing.B) { | 
 | 	var x [24 / 4]uint32 | 
 | 	for i := 0; i < b.N; i++ { | 
 | 		y := x | 
 | 		_ = y | 
 | 	} | 
 | } | 
 | func BenchmarkCopyFat32(b *testing.B) { | 
 | 	var x [32 / 4]uint32 | 
 | 	for i := 0; i < b.N; i++ { | 
 | 		y := x | 
 | 		_ = y | 
 | 	} | 
 | } | 
 | func BenchmarkCopyFat64(b *testing.B) { | 
 | 	var x [64 / 4]uint32 | 
 | 	for i := 0; i < b.N; i++ { | 
 | 		y := x | 
 | 		_ = y | 
 | 	} | 
 | } | 
 | func BenchmarkCopyFat128(b *testing.B) { | 
 | 	var x [128 / 4]uint32 | 
 | 	for i := 0; i < b.N; i++ { | 
 | 		y := x | 
 | 		_ = y | 
 | 	} | 
 | } | 
 | func BenchmarkCopyFat256(b *testing.B) { | 
 | 	var x [256 / 4]uint32 | 
 | 	for i := 0; i < b.N; i++ { | 
 | 		y := x | 
 | 		_ = y | 
 | 	} | 
 | } | 
 | func BenchmarkCopyFat512(b *testing.B) { | 
 | 	var x [512 / 4]uint32 | 
 | 	for i := 0; i < b.N; i++ { | 
 | 		y := x | 
 | 		_ = y | 
 | 	} | 
 | } | 
 | func BenchmarkCopyFat520(b *testing.B) { | 
 | 	var x [520 / 4]uint32 | 
 | 	for i := 0; i < b.N; i++ { | 
 | 		y := x | 
 | 		_ = y | 
 | 	} | 
 | } | 
 | func BenchmarkCopyFat1024(b *testing.B) { | 
 | 	var x [1024 / 4]uint32 | 
 | 	for i := 0; i < b.N; i++ { | 
 | 		y := x | 
 | 		_ = y | 
 | 	} | 
 | } | 
 |  | 
 | // BenchmarkIssue18740 ensures that memmove uses 4 and 8 byte load/store to move 4 and 8 bytes. | 
 | // It used to do 2 2-byte load/stores, which leads to a pipeline stall | 
 | // when we try to read the result with one 4-byte load. | 
 | func BenchmarkIssue18740(b *testing.B) { | 
 | 	benchmarks := []struct { | 
 | 		name  string | 
 | 		nbyte int | 
 | 		f     func([]byte) uint64 | 
 | 	}{ | 
 | 		{"2byte", 2, func(buf []byte) uint64 { return uint64(binary.LittleEndian.Uint16(buf)) }}, | 
 | 		{"4byte", 4, func(buf []byte) uint64 { return uint64(binary.LittleEndian.Uint32(buf)) }}, | 
 | 		{"8byte", 8, func(buf []byte) uint64 { return binary.LittleEndian.Uint64(buf) }}, | 
 | 	} | 
 |  | 
 | 	var g [4096]byte | 
 | 	for _, bm := range benchmarks { | 
 | 		buf := make([]byte, bm.nbyte) | 
 | 		b.Run(bm.name, func(b *testing.B) { | 
 | 			for j := 0; j < b.N; j++ { | 
 | 				for i := 0; i < 4096; i += bm.nbyte { | 
 | 					copy(buf[:], g[i:]) | 
 | 					sink += bm.f(buf[:]) | 
 | 				} | 
 | 			} | 
 | 		}) | 
 | 	} | 
 | } |