| // Copyright 2020 The Go Authors. All rights reserved. | 
 | // Use of this source code is governed by a BSD-style | 
 | // license that can be found in the LICENSE file. | 
 |  | 
 | // +build goexperiment.staticlockranking | 
 |  | 
 | package runtime | 
 |  | 
 | import ( | 
 | 	"runtime/internal/atomic" | 
 | 	"unsafe" | 
 | ) | 
 |  | 
 | // worldIsStopped is accessed atomically to track world-stops. 1 == world | 
 | // stopped. | 
 | var worldIsStopped uint32 | 
 |  | 
 | // lockRankStruct is embedded in mutex | 
 | type lockRankStruct struct { | 
 | 	// static lock ranking of the lock | 
 | 	rank lockRank | 
 | 	// pad field to make sure lockRankStruct is a multiple of 8 bytes, even on | 
 | 	// 32-bit systems. | 
 | 	pad int | 
 | } | 
 |  | 
 | // init checks that the partial order in lockPartialOrder fits within the total | 
 | // order determined by the order of the lockRank constants. | 
 | func init() { | 
 | 	for rank, list := range lockPartialOrder { | 
 | 		for _, entry := range list { | 
 | 			if entry > lockRank(rank) { | 
 | 				println("lockPartial order row", lockRank(rank).String(), "entry", entry.String()) | 
 | 				throw("lockPartialOrder table is inconsistent with total lock ranking order") | 
 | 			} | 
 | 		} | 
 | 	} | 
 | } | 
 |  | 
 | func lockInit(l *mutex, rank lockRank) { | 
 | 	l.rank = rank | 
 | } | 
 |  | 
 | func getLockRank(l *mutex) lockRank { | 
 | 	return l.rank | 
 | } | 
 |  | 
 | // lockWithRank is like lock(l), but allows the caller to specify a lock rank | 
 | // when acquiring a non-static lock. | 
 | // | 
 | // Note that we need to be careful about stack splits: | 
 | // | 
 | // This function is not nosplit, thus it may split at function entry. This may | 
 | // introduce a new edge in the lock order, but it is no different from any | 
 | // other (nosplit) call before this call (including the call to lock() itself). | 
 | // | 
 | // However, we switch to the systemstack to record the lock held to ensure that | 
 | // we record an accurate lock ordering. e.g., without systemstack, a stack | 
 | // split on entry to lock2() would record stack split locks as taken after l, | 
 | // even though l is not actually locked yet. | 
 | func lockWithRank(l *mutex, rank lockRank) { | 
 | 	if l == &debuglock || l == &paniclk { | 
 | 		// debuglock is only used for println/printlock(). Don't do lock | 
 | 		// rank recording for it, since print/println are used when | 
 | 		// printing out a lock ordering problem below. | 
 | 		// | 
 | 		// paniclk has an ordering problem, since it can be acquired | 
 | 		// during a panic with any other locks held (especially if the | 
 | 		// panic is because of a directed segv), and yet also allg is | 
 | 		// acquired after paniclk in tracebackothers()). This is a genuine | 
 | 		// problem, so for now we don't do lock rank recording for paniclk | 
 | 		// either. | 
 | 		lock2(l) | 
 | 		return | 
 | 	} | 
 | 	if rank == 0 { | 
 | 		rank = lockRankLeafRank | 
 | 	} | 
 | 	gp := getg() | 
 | 	// Log the new class. | 
 | 	systemstack(func() { | 
 | 		i := gp.m.locksHeldLen | 
 | 		if i >= len(gp.m.locksHeld) { | 
 | 			throw("too many locks held concurrently for rank checking") | 
 | 		} | 
 | 		gp.m.locksHeld[i].rank = rank | 
 | 		gp.m.locksHeld[i].lockAddr = uintptr(unsafe.Pointer(l)) | 
 | 		gp.m.locksHeldLen++ | 
 |  | 
 | 		// i is the index of the lock being acquired | 
 | 		if i > 0 { | 
 | 			checkRanks(gp, gp.m.locksHeld[i-1].rank, rank) | 
 | 		} | 
 | 		lock2(l) | 
 | 	}) | 
 | } | 
 |  | 
 | // nosplit to ensure it can be called in as many contexts as possible. | 
 | //go:nosplit | 
 | func printHeldLocks(gp *g) { | 
 | 	if gp.m.locksHeldLen == 0 { | 
 | 		println("<none>") | 
 | 		return | 
 | 	} | 
 |  | 
 | 	for j, held := range gp.m.locksHeld[:gp.m.locksHeldLen] { | 
 | 		println(j, ":", held.rank.String(), held.rank, unsafe.Pointer(gp.m.locksHeld[j].lockAddr)) | 
 | 	} | 
 | } | 
 |  | 
 | // acquireLockRank acquires a rank which is not associated with a mutex lock | 
 | // | 
 | // This function may be called in nosplit context and thus must be nosplit. | 
 | //go:nosplit | 
 | func acquireLockRank(rank lockRank) { | 
 | 	gp := getg() | 
 | 	// Log the new class. See comment on lockWithRank. | 
 | 	systemstack(func() { | 
 | 		i := gp.m.locksHeldLen | 
 | 		if i >= len(gp.m.locksHeld) { | 
 | 			throw("too many locks held concurrently for rank checking") | 
 | 		} | 
 | 		gp.m.locksHeld[i].rank = rank | 
 | 		gp.m.locksHeld[i].lockAddr = 0 | 
 | 		gp.m.locksHeldLen++ | 
 |  | 
 | 		// i is the index of the lock being acquired | 
 | 		if i > 0 { | 
 | 			checkRanks(gp, gp.m.locksHeld[i-1].rank, rank) | 
 | 		} | 
 | 	}) | 
 | } | 
 |  | 
 | // checkRanks checks if goroutine g, which has mostly recently acquired a lock | 
 | // with rank 'prevRank', can now acquire a lock with rank 'rank'. | 
 | // | 
 | //go:systemstack | 
 | func checkRanks(gp *g, prevRank, rank lockRank) { | 
 | 	rankOK := false | 
 | 	if rank < prevRank { | 
 | 		// If rank < prevRank, then we definitely have a rank error | 
 | 		rankOK = false | 
 | 	} else if rank == lockRankLeafRank { | 
 | 		// If new lock is a leaf lock, then the preceding lock can | 
 | 		// be anything except another leaf lock. | 
 | 		rankOK = prevRank < lockRankLeafRank | 
 | 	} else { | 
 | 		// We've now verified the total lock ranking, but we | 
 | 		// also enforce the partial ordering specified by | 
 | 		// lockPartialOrder as well. Two locks with the same rank | 
 | 		// can only be acquired at the same time if explicitly | 
 | 		// listed in the lockPartialOrder table. | 
 | 		list := lockPartialOrder[rank] | 
 | 		for _, entry := range list { | 
 | 			if entry == prevRank { | 
 | 				rankOK = true | 
 | 				break | 
 | 			} | 
 | 		} | 
 | 	} | 
 | 	if !rankOK { | 
 | 		printlock() | 
 | 		println(gp.m.procid, " ======") | 
 | 		printHeldLocks(gp) | 
 | 		throw("lock ordering problem") | 
 | 	} | 
 | } | 
 |  | 
 | // See comment on lockWithRank regarding stack splitting. | 
 | func unlockWithRank(l *mutex) { | 
 | 	if l == &debuglock || l == &paniclk { | 
 | 		// See comment at beginning of lockWithRank. | 
 | 		unlock2(l) | 
 | 		return | 
 | 	} | 
 | 	gp := getg() | 
 | 	systemstack(func() { | 
 | 		found := false | 
 | 		for i := gp.m.locksHeldLen - 1; i >= 0; i-- { | 
 | 			if gp.m.locksHeld[i].lockAddr == uintptr(unsafe.Pointer(l)) { | 
 | 				found = true | 
 | 				copy(gp.m.locksHeld[i:gp.m.locksHeldLen-1], gp.m.locksHeld[i+1:gp.m.locksHeldLen]) | 
 | 				gp.m.locksHeldLen-- | 
 | 				break | 
 | 			} | 
 | 		} | 
 | 		if !found { | 
 | 			println(gp.m.procid, ":", l.rank.String(), l.rank, l) | 
 | 			throw("unlock without matching lock acquire") | 
 | 		} | 
 | 		unlock2(l) | 
 | 	}) | 
 | } | 
 |  | 
 | // releaseLockRank releases a rank which is not associated with a mutex lock | 
 | // | 
 | // This function may be called in nosplit context and thus must be nosplit. | 
 | //go:nosplit | 
 | func releaseLockRank(rank lockRank) { | 
 | 	gp := getg() | 
 | 	systemstack(func() { | 
 | 		found := false | 
 | 		for i := gp.m.locksHeldLen - 1; i >= 0; i-- { | 
 | 			if gp.m.locksHeld[i].rank == rank && gp.m.locksHeld[i].lockAddr == 0 { | 
 | 				found = true | 
 | 				copy(gp.m.locksHeld[i:gp.m.locksHeldLen-1], gp.m.locksHeld[i+1:gp.m.locksHeldLen]) | 
 | 				gp.m.locksHeldLen-- | 
 | 				break | 
 | 			} | 
 | 		} | 
 | 		if !found { | 
 | 			println(gp.m.procid, ":", rank.String(), rank) | 
 | 			throw("lockRank release without matching lockRank acquire") | 
 | 		} | 
 | 	}) | 
 | } | 
 |  | 
 | // See comment on lockWithRank regarding stack splitting. | 
 | func lockWithRankMayAcquire(l *mutex, rank lockRank) { | 
 | 	gp := getg() | 
 | 	if gp.m.locksHeldLen == 0 { | 
 | 		// No possibilty of lock ordering problem if no other locks held | 
 | 		return | 
 | 	} | 
 |  | 
 | 	systemstack(func() { | 
 | 		i := gp.m.locksHeldLen | 
 | 		if i >= len(gp.m.locksHeld) { | 
 | 			throw("too many locks held concurrently for rank checking") | 
 | 		} | 
 | 		// Temporarily add this lock to the locksHeld list, so | 
 | 		// checkRanks() will print out list, including this lock, if there | 
 | 		// is a lock ordering problem. | 
 | 		gp.m.locksHeld[i].rank = rank | 
 | 		gp.m.locksHeld[i].lockAddr = uintptr(unsafe.Pointer(l)) | 
 | 		gp.m.locksHeldLen++ | 
 | 		checkRanks(gp, gp.m.locksHeld[i-1].rank, rank) | 
 | 		gp.m.locksHeldLen-- | 
 | 	}) | 
 | } | 
 |  | 
 | // nosplit to ensure it can be called in as many contexts as possible. | 
 | //go:nosplit | 
 | func checkLockHeld(gp *g, l *mutex) bool { | 
 | 	for i := gp.m.locksHeldLen - 1; i >= 0; i-- { | 
 | 		if gp.m.locksHeld[i].lockAddr == uintptr(unsafe.Pointer(l)) { | 
 | 			return true | 
 | 		} | 
 | 	} | 
 | 	return false | 
 | } | 
 |  | 
 | // assertLockHeld throws if l is not held by the caller. | 
 | // | 
 | // nosplit to ensure it can be called in as many contexts as possible. | 
 | //go:nosplit | 
 | func assertLockHeld(l *mutex) { | 
 | 	gp := getg() | 
 |  | 
 | 	held := checkLockHeld(gp, l) | 
 | 	if held { | 
 | 		return | 
 | 	} | 
 |  | 
 | 	// Crash from system stack to avoid splits that may cause | 
 | 	// additional issues. | 
 | 	systemstack(func() { | 
 | 		printlock() | 
 | 		print("caller requires lock ", l, " (rank ", l.rank.String(), "), holding:\n") | 
 | 		printHeldLocks(gp) | 
 | 		throw("not holding required lock!") | 
 | 	}) | 
 | } | 
 |  | 
 | // assertRankHeld throws if a mutex with rank r is not held by the caller. | 
 | // | 
 | // This is less precise than assertLockHeld, but can be used in places where a | 
 | // pointer to the exact mutex is not available. | 
 | // | 
 | // nosplit to ensure it can be called in as many contexts as possible. | 
 | //go:nosplit | 
 | func assertRankHeld(r lockRank) { | 
 | 	gp := getg() | 
 |  | 
 | 	for i := gp.m.locksHeldLen - 1; i >= 0; i-- { | 
 | 		if gp.m.locksHeld[i].rank == r { | 
 | 			return | 
 | 		} | 
 | 	} | 
 |  | 
 | 	// Crash from system stack to avoid splits that may cause | 
 | 	// additional issues. | 
 | 	systemstack(func() { | 
 | 		printlock() | 
 | 		print("caller requires lock with rank ", r.String(), "), holding:\n") | 
 | 		printHeldLocks(gp) | 
 | 		throw("not holding required lock!") | 
 | 	}) | 
 | } | 
 |  | 
 | // worldStopped notes that the world is stopped. | 
 | // | 
 | // Caller must hold worldsema. | 
 | // | 
 | // nosplit to ensure it can be called in as many contexts as possible. | 
 | //go:nosplit | 
 | func worldStopped() { | 
 | 	if stopped := atomic.Xadd(&worldIsStopped, 1); stopped != 1 { | 
 | 		systemstack(func() { | 
 | 			print("world stop count=", stopped, "\n") | 
 | 			throw("recursive world stop") | 
 | 		}) | 
 | 	} | 
 | } | 
 |  | 
 | // worldStarted that the world is starting. | 
 | // | 
 | // Caller must hold worldsema. | 
 | // | 
 | // nosplit to ensure it can be called in as many contexts as possible. | 
 | //go:nosplit | 
 | func worldStarted() { | 
 | 	if stopped := atomic.Xadd(&worldIsStopped, -1); stopped != 0 { | 
 | 		systemstack(func() { | 
 | 			print("world stop count=", stopped, "\n") | 
 | 			throw("released non-stopped world stop") | 
 | 		}) | 
 | 	} | 
 | } | 
 |  | 
 | // nosplit to ensure it can be called in as many contexts as possible. | 
 | //go:nosplit | 
 | func checkWorldStopped() bool { | 
 | 	stopped := atomic.Load(&worldIsStopped) | 
 | 	if stopped > 1 { | 
 | 		systemstack(func() { | 
 | 			print("inconsistent world stop count=", stopped, "\n") | 
 | 			throw("inconsistent world stop count") | 
 | 		}) | 
 | 	} | 
 |  | 
 | 	return stopped == 1 | 
 | } | 
 |  | 
 | // assertWorldStopped throws if the world is not stopped. It does not check | 
 | // which M stopped the world. | 
 | // | 
 | // nosplit to ensure it can be called in as many contexts as possible. | 
 | //go:nosplit | 
 | func assertWorldStopped() { | 
 | 	if checkWorldStopped() { | 
 | 		return | 
 | 	} | 
 |  | 
 | 	throw("world not stopped") | 
 | } | 
 |  | 
 | // assertWorldStoppedOrLockHeld throws if the world is not stopped and the | 
 | // passed lock is not held. | 
 | // | 
 | // nosplit to ensure it can be called in as many contexts as possible. | 
 | //go:nosplit | 
 | func assertWorldStoppedOrLockHeld(l *mutex) { | 
 | 	if checkWorldStopped() { | 
 | 		return | 
 | 	} | 
 |  | 
 | 	gp := getg() | 
 | 	held := checkLockHeld(gp, l) | 
 | 	if held { | 
 | 		return | 
 | 	} | 
 |  | 
 | 	// Crash from system stack to avoid splits that may cause | 
 | 	// additional issues. | 
 | 	systemstack(func() { | 
 | 		printlock() | 
 | 		print("caller requires world stop or lock ", l, " (rank ", l.rank.String(), "), holding:\n") | 
 | 		println("<no world stop>") | 
 | 		printHeldLocks(gp) | 
 | 		throw("no world stop or required lock!") | 
 | 	}) | 
 | } |