From f3b5a2bc1983ddb83d72e741b176993d9b800faf Mon Sep 17 00:00:00 2001 From: Austin Clements Date: Thu, 6 Jul 2017 11:55:39 -0400 Subject: [PATCH] runtime: prevent descheduling while holding rwmutex read lock Currently only the rwmutex write lock prevents descheduling. The read lock does not. This leads to the following situation: 1. A reader acquires the lock and gets descheduled. 2. GOMAXPROCS writers attempt to acquire the lock (or at least one writer does, followed by readers). This blocks all of the Ps. 3. There is no 3. The descheduled reader never gets to run again because there are no Ps, so it never releases the lock and the system deadlocks. Fix this by preventing descheduling while holding the read lock. This requires also rewriting TestParallelRWMutexReaders to always create enough GOMAXPROCS and to use non-blocking operations for synchronization. Fixes #20903. Change-Id: Ibd460663a7e5a555be5490e13b2eaaa295fac39f Reviewed-on: https://go-review.googlesource.com/47632 Run-TryBot: Austin Clements TryBot-Result: Gobot Gobot Reviewed-by: Ian Lance Taylor --- src/runtime/rwmutex.go | 8 +++++++- src/runtime/rwmutex_test.go | 25 ++++++++++++------------- 2 files changed, 19 insertions(+), 14 deletions(-) diff --git a/src/runtime/rwmutex.go b/src/runtime/rwmutex.go index bca29d15d0..7eeb559adb 100644 --- a/src/runtime/rwmutex.go +++ b/src/runtime/rwmutex.go @@ -31,6 +31,11 @@ const rwmutexMaxReaders = 1 << 30 // rlock locks rw for reading. func (rw *rwmutex) rlock() { + // The reader must not be allowed to lose its P or else other + // things blocking on the lock may consume all of the Ps and + // deadlock (issue #20903). Alternatively, we could drop the P + // while sleeping. + acquirem() if int32(atomic.Xadd(&rw.readerCount, 1)) < 0 { // A writer is pending. Park on the reader queue. systemstack(func() { @@ -70,11 +75,12 @@ func (rw *rwmutex) runlock() { unlock(&rw.rLock) } } + releasem(getg().m) } // lock locks rw for writing. func (rw *rwmutex) lock() { - // Resolve competition with other writers. + // Resolve competition with other writers and stick to our P. lock(&rw.wLock) m := getg().m // Announce that there is a pending writer. diff --git a/src/runtime/rwmutex_test.go b/src/runtime/rwmutex_test.go index b78a8e7987..a69eca1511 100644 --- a/src/runtime/rwmutex_test.go +++ b/src/runtime/rwmutex_test.go @@ -16,30 +16,29 @@ import ( "testing" ) -func parallelReader(m *RWMutex, clocked, cunlock, cdone chan bool) { +func parallelReader(m *RWMutex, clocked chan bool, cunlock *uint32, cdone chan bool) { m.RLock() clocked <- true - <-cunlock + for atomic.LoadUint32(cunlock) == 0 { + } m.RUnlock() cdone <- true } -func doTestParallelReaders(numReaders, gomaxprocs int) { - GOMAXPROCS(gomaxprocs) +func doTestParallelReaders(numReaders int) { + GOMAXPROCS(numReaders + 1) var m RWMutex - clocked := make(chan bool) - cunlock := make(chan bool) + clocked := make(chan bool, numReaders) + var cunlock uint32 cdone := make(chan bool) for i := 0; i < numReaders; i++ { - go parallelReader(&m, clocked, cunlock, cdone) + go parallelReader(&m, clocked, &cunlock, cdone) } // Wait for all parallel RLock()s to succeed. for i := 0; i < numReaders; i++ { <-clocked } - for i := 0; i < numReaders; i++ { - cunlock <- true - } + atomic.StoreUint32(&cunlock, 1) // Wait for the goroutines to finish. for i := 0; i < numReaders; i++ { <-cdone @@ -48,9 +47,9 @@ func doTestParallelReaders(numReaders, gomaxprocs int) { func TestParallelRWMutexReaders(t *testing.T) { defer GOMAXPROCS(GOMAXPROCS(-1)) - doTestParallelReaders(1, 4) - doTestParallelReaders(3, 4) - doTestParallelReaders(4, 2) + doTestParallelReaders(1) + doTestParallelReaders(3) + doTestParallelReaders(4) } func reader(rwm *RWMutex, num_iterations int, activity *int32, cdone chan bool) {