* [PATCH 0/4] tsan fixes @ 2023-01-08 16:39 Emilio Cota 2023-01-08 16:39 ` [PATCH 1/4] cpu: free cpu->tb_jmp_cache with RCU Emilio Cota ` (3 more replies) 0 siblings, 4 replies; 12+ messages in thread From: Emilio Cota @ 2023-01-08 16:39 UTC (permalink / raw) To: qemu-devel Cc: Alex Bennée, Paolo Bonzini, Richard Henderson, Emilio Cota Hi, Here are some fixes for tsan issues that I've encountered. The most important patch is 3/4, which allows us to run tsan for non-trivial workloads. Thanks, Emilio ^ permalink raw reply [flat|nested] 12+ messages in thread
* [PATCH 1/4] cpu: free cpu->tb_jmp_cache with RCU 2023-01-08 16:39 [PATCH 0/4] tsan fixes Emilio Cota @ 2023-01-08 16:39 ` Emilio Cota 2023-01-08 19:19 ` Richard Henderson 2023-01-08 16:39 ` [PATCH 2/4] util/qht: add missing atomic_set(hashes[i]) Emilio Cota ` (2 subsequent siblings) 3 siblings, 1 reply; 12+ messages in thread From: Emilio Cota @ 2023-01-08 16:39 UTC (permalink / raw) To: qemu-devel Cc: Alex Bennée, Paolo Bonzini, Richard Henderson, Emilio Cota Fixes the appended use-after-free. The root cause is that during tb invalidation we use CPU_FOREACH, and therefore to safely free a vCPU we must wait for an RCU grace period to elapse. $ x86_64-linux-user/qemu-x86_64 tests/tcg/x86_64-linux-user/munmap-pthread ================================================================= ==1800604==ERROR: AddressSanitizer: heap-use-after-free on address 0x62d0005f7418 at pc 0x5593da6704eb bp 0x7f4961a7ac70 sp 0x7f4961a7ac60 READ of size 8 at 0x62d0005f7418 thread T2 #0 0x5593da6704ea in tb_jmp_cache_inval_tb ../accel/tcg/tb-maint.c:244 #1 0x5593da6704ea in do_tb_phys_invalidate ../accel/tcg/tb-maint.c:290 #2 0x5593da670631 in tb_phys_invalidate__locked ../accel/tcg/tb-maint.c:306 #3 0x5593da670631 in tb_invalidate_phys_page_range__locked ../accel/tcg/tb-maint.c:542 #4 0x5593da67106d in tb_invalidate_phys_range ../accel/tcg/tb-maint.c:614 #5 0x5593da6a64d4 in target_munmap ../linux-user/mmap.c:766 #6 0x5593da6dba05 in do_syscall1 ../linux-user/syscall.c:10105 #7 0x5593da6f564c in do_syscall ../linux-user/syscall.c:13329 #8 0x5593da49e80c in cpu_loop ../linux-user/x86_64/../i386/cpu_loop.c:233 #9 0x5593da6be28c in clone_func ../linux-user/syscall.c:6633 #10 0x7f496231cb42 in start_thread nptl/pthread_create.c:442 #11 0x7f49623ae9ff (/lib/x86_64-linux-gnu/libc.so.6+0x1269ff) 0x62d0005f7418 is located 28696 bytes inside of 32768-byte region [0x62d0005f0400,0x62d0005f8400) freed by thread T148 here: #0 0x7f49627b6460 in __interceptor_free ../../../../src/libsanitizer/asan/asan_malloc_linux.cpp:52 #1 0x5593da5ac057 in cpu_exec_unrealizefn ../cpu.c:180 #2 0x5593da81f851 (/home/cota/src/qemu/build/qemu-x86_64+0x484851) Signed-off-by: Emilio Cota <cota@braap.org> --- accel/tcg/cpu-exec.c | 1 - cpu.c | 7 +++++++ include/hw/core/cpu.h | 3 +++ 3 files changed, 10 insertions(+), 1 deletion(-) diff --git a/accel/tcg/cpu-exec.c b/accel/tcg/cpu-exec.c index 356fe348de..ca95d21528 100644 --- a/accel/tcg/cpu-exec.c +++ b/accel/tcg/cpu-exec.c @@ -1070,7 +1070,6 @@ void tcg_exec_unrealizefn(CPUState *cpu) #endif /* !CONFIG_USER_ONLY */ tlb_destroy(cpu); - g_free(cpu->tb_jmp_cache); } #ifndef CONFIG_USER_ONLY diff --git a/cpu.c b/cpu.c index 4a7d865427..564200559f 100644 --- a/cpu.c +++ b/cpu.c @@ -164,6 +164,12 @@ void cpu_exec_realizefn(CPUState *cpu, Error **errp) #endif /* CONFIG_USER_ONLY */ } +static void cpu_free_rcu(CPUState *cpu) +{ + /* .tb_jmp_cache is NULL except under TCG */ + g_free(cpu->tb_jmp_cache); +} + void cpu_exec_unrealizefn(CPUState *cpu) { #ifndef CONFIG_USER_ONLY @@ -181,6 +187,7 @@ void cpu_exec_unrealizefn(CPUState *cpu) } cpu_list_remove(cpu); + call_rcu(cpu, cpu_free_rcu, rcu); } /* diff --git a/include/hw/core/cpu.h b/include/hw/core/cpu.h index 8830546121..9fefad5656 100644 --- a/include/hw/core/cpu.h +++ b/include/hw/core/cpu.h @@ -266,6 +266,7 @@ struct qemu_work_item; /** * CPUState: + * @rcu: Used for safe deferred memory reclamation. * @cpu_index: CPU index (informative). * @cluster_index: Identifies which cluster this CPU is in. * For boards which don't define clusters or for "loose" CPUs not assigned @@ -321,6 +322,8 @@ struct qemu_work_item; * State of one CPU core or thread. */ struct CPUState { + struct rcu_head rcu; + /*< private >*/ DeviceState parent_obj; /* cache to avoid expensive CPU_GET_CLASS */ -- 2.34.1 ^ permalink raw reply related [flat|nested] 12+ messages in thread
* Re: [PATCH 1/4] cpu: free cpu->tb_jmp_cache with RCU 2023-01-08 16:39 ` [PATCH 1/4] cpu: free cpu->tb_jmp_cache with RCU Emilio Cota @ 2023-01-08 19:19 ` Richard Henderson 2023-01-09 21:52 ` Emilio Cota 0 siblings, 1 reply; 12+ messages in thread From: Richard Henderson @ 2023-01-08 19:19 UTC (permalink / raw) To: Emilio Cota, qemu-devel; +Cc: Alex Bennée, Paolo Bonzini On 1/8/23 08:39, Emilio Cota wrote: > Fixes the appended use-after-free. The root cause is that > during tb invalidation we use CPU_FOREACH, and therefore > to safely free a vCPU we must wait for an RCU grace period > to elapse. > > $ x86_64-linux-user/qemu-x86_64 tests/tcg/x86_64-linux-user/munmap-pthread > ================================================================= > ==1800604==ERROR: AddressSanitizer: heap-use-after-free on address 0x62d0005f7418 at pc 0x5593da6704eb bp 0x7f4961a7ac70 sp 0x7f4961a7ac60 > READ of size 8 at 0x62d0005f7418 thread T2 > #0 0x5593da6704ea in tb_jmp_cache_inval_tb ../accel/tcg/tb-maint.c:244 > #1 0x5593da6704ea in do_tb_phys_invalidate ../accel/tcg/tb-maint.c:290 > #2 0x5593da670631 in tb_phys_invalidate__locked ../accel/tcg/tb-maint.c:306 > #3 0x5593da670631 in tb_invalidate_phys_page_range__locked ../accel/tcg/tb-maint.c:542 > #4 0x5593da67106d in tb_invalidate_phys_range ../accel/tcg/tb-maint.c:614 > #5 0x5593da6a64d4 in target_munmap ../linux-user/mmap.c:766 > #6 0x5593da6dba05 in do_syscall1 ../linux-user/syscall.c:10105 > #7 0x5593da6f564c in do_syscall ../linux-user/syscall.c:13329 > #8 0x5593da49e80c in cpu_loop ../linux-user/x86_64/../i386/cpu_loop.c:233 > #9 0x5593da6be28c in clone_func ../linux-user/syscall.c:6633 > #10 0x7f496231cb42 in start_thread nptl/pthread_create.c:442 > #11 0x7f49623ae9ff (/lib/x86_64-linux-gnu/libc.so.6+0x1269ff) > > 0x62d0005f7418 is located 28696 bytes inside of 32768-byte region [0x62d0005f0400,0x62d0005f8400) > freed by thread T148 here: > #0 0x7f49627b6460 in __interceptor_free ../../../../src/libsanitizer/asan/asan_malloc_linux.cpp:52 > #1 0x5593da5ac057 in cpu_exec_unrealizefn ../cpu.c:180 > #2 0x5593da81f851 (/home/cota/src/qemu/build/qemu-x86_64+0x484851) > > Signed-off-by: Emilio Cota <cota@braap.org> > --- > accel/tcg/cpu-exec.c | 1 - > cpu.c | 7 +++++++ > include/hw/core/cpu.h | 3 +++ > 3 files changed, 10 insertions(+), 1 deletion(-) > > diff --git a/accel/tcg/cpu-exec.c b/accel/tcg/cpu-exec.c > index 356fe348de..ca95d21528 100644 > --- a/accel/tcg/cpu-exec.c > +++ b/accel/tcg/cpu-exec.c > @@ -1070,7 +1070,6 @@ void tcg_exec_unrealizefn(CPUState *cpu) > #endif /* !CONFIG_USER_ONLY */ > > tlb_destroy(cpu); > - g_free(cpu->tb_jmp_cache); Can you simply use g_free_rcu here? > } > > #ifndef CONFIG_USER_ONLY > diff --git a/cpu.c b/cpu.c > index 4a7d865427..564200559f 100644 > --- a/cpu.c > +++ b/cpu.c > @@ -164,6 +164,12 @@ void cpu_exec_realizefn(CPUState *cpu, Error **errp) > #endif /* CONFIG_USER_ONLY */ > } > > +static void cpu_free_rcu(CPUState *cpu) > +{ > + /* .tb_jmp_cache is NULL except under TCG */ > + g_free(cpu->tb_jmp_cache); > +} > + > void cpu_exec_unrealizefn(CPUState *cpu) > { > #ifndef CONFIG_USER_ONLY > @@ -181,6 +187,7 @@ void cpu_exec_unrealizefn(CPUState *cpu) > } > > cpu_list_remove(cpu); > + call_rcu(cpu, cpu_free_rcu, rcu); Certainly this seems wrong, exposing tb_jmp_cache beyond tcg. r~ ^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: [PATCH 1/4] cpu: free cpu->tb_jmp_cache with RCU 2023-01-08 19:19 ` Richard Henderson @ 2023-01-09 21:52 ` Emilio Cota 0 siblings, 0 replies; 12+ messages in thread From: Emilio Cota @ 2023-01-09 21:52 UTC (permalink / raw) To: Richard Henderson; +Cc: qemu-devel, Alex Bennée, Paolo Bonzini On Sun, Jan 08, 2023 at 11:19:53 -0800, Richard Henderson wrote: > On 1/8/23 08:39, Emilio Cota wrote: (snip) > > diff --git a/accel/tcg/cpu-exec.c b/accel/tcg/cpu-exec.c > > index 356fe348de..ca95d21528 100644 > > --- a/accel/tcg/cpu-exec.c > > +++ b/accel/tcg/cpu-exec.c > > @@ -1070,7 +1070,6 @@ void tcg_exec_unrealizefn(CPUState *cpu) > > #endif /* !CONFIG_USER_ONLY */ > > tlb_destroy(cpu); > > - g_free(cpu->tb_jmp_cache); > > Can you simply use g_free_rcu here? Yes, although we must have removed the CPU from the RCU list before doing so. > > diff --git a/cpu.c b/cpu.c > > index 4a7d865427..564200559f 100644 > > --- a/cpu.c > > +++ b/cpu.c > > @@ -164,6 +164,12 @@ void cpu_exec_realizefn(CPUState *cpu, Error **errp) > > #endif /* CONFIG_USER_ONLY */ > > } > > +static void cpu_free_rcu(CPUState *cpu) > > +{ > > + /* .tb_jmp_cache is NULL except under TCG */ > > + g_free(cpu->tb_jmp_cache); > > +} > > + > > void cpu_exec_unrealizefn(CPUState *cpu) > > { > > #ifndef CONFIG_USER_ONLY > > @@ -181,6 +187,7 @@ void cpu_exec_unrealizefn(CPUState *cpu) > > } > > cpu_list_remove(cpu); > > + call_rcu(cpu, cpu_free_rcu, rcu); > > Certainly this seems wrong, exposing tb_jmp_cache beyond tcg. I've changed this in v2 to call tcg_exec_unrealizefn after cpu_list_remove. An alternative would be to call the whole cpu_exec_unrealizefn after an RCU grace period, but I think that might be more trouble than it's worth. Thanks, Emilio ^ permalink raw reply [flat|nested] 12+ messages in thread
* [PATCH 2/4] util/qht: add missing atomic_set(hashes[i]) 2023-01-08 16:39 [PATCH 0/4] tsan fixes Emilio Cota 2023-01-08 16:39 ` [PATCH 1/4] cpu: free cpu->tb_jmp_cache with RCU Emilio Cota @ 2023-01-08 16:39 ` Emilio Cota 2023-01-08 19:20 ` Richard Henderson 2023-01-08 19:54 ` Philippe Mathieu-Daudé 2023-01-08 16:39 ` [PATCH 3/4] util/qht: use striped locks under TSAN Emilio Cota 2023-01-08 16:39 ` [PATCH 4/4] plugins: make qemu_plugin_user_exit's locking order consistent with fork_start's Emilio Cota 3 siblings, 2 replies; 12+ messages in thread From: Emilio Cota @ 2023-01-08 16:39 UTC (permalink / raw) To: qemu-devel Cc: Alex Bennée, Paolo Bonzini, Richard Henderson, Emilio Cota We forgot to add this one in "a890643958 util/qht: atomically set b->hashes". Detected with tsan. Signed-off-by: Emilio Cota <cota@braap.org> --- util/qht.c | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/util/qht.c b/util/qht.c index 065fc501f4..15866299e6 100644 --- a/util/qht.c +++ b/util/qht.c @@ -688,7 +688,7 @@ static inline void qht_bucket_remove_entry(struct qht_bucket *orig, int pos) int i; if (qht_entry_is_last(orig, pos)) { - orig->hashes[pos] = 0; + qatomic_set(&orig->hashes[pos], 0); qatomic_set(&orig->pointers[pos], NULL); return; } -- 2.34.1 ^ permalink raw reply related [flat|nested] 12+ messages in thread
* Re: [PATCH 2/4] util/qht: add missing atomic_set(hashes[i]) 2023-01-08 16:39 ` [PATCH 2/4] util/qht: add missing atomic_set(hashes[i]) Emilio Cota @ 2023-01-08 19:20 ` Richard Henderson 2023-01-08 19:54 ` Philippe Mathieu-Daudé 1 sibling, 0 replies; 12+ messages in thread From: Richard Henderson @ 2023-01-08 19:20 UTC (permalink / raw) To: Emilio Cota, qemu-devel; +Cc: Alex Bennée, Paolo Bonzini On 1/8/23 08:39, Emilio Cota wrote: > We forgot to add this one in "a890643958 util/qht: atomically set b->hashes". > > Detected with tsan. > > Signed-off-by: Emilio Cota<cota@braap.org> > --- > util/qht.c | 2 +- > 1 file changed, 1 insertion(+), 1 deletion(-) Reviewed-by: Richard Henderson <richard.henderson@linaro.org> r~ ^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: [PATCH 2/4] util/qht: add missing atomic_set(hashes[i]) 2023-01-08 16:39 ` [PATCH 2/4] util/qht: add missing atomic_set(hashes[i]) Emilio Cota 2023-01-08 19:20 ` Richard Henderson @ 2023-01-08 19:54 ` Philippe Mathieu-Daudé 1 sibling, 0 replies; 12+ messages in thread From: Philippe Mathieu-Daudé @ 2023-01-08 19:54 UTC (permalink / raw) To: Emilio Cota, qemu-devel Cc: Alex Bennée, Paolo Bonzini, Richard Henderson On 8/1/23 17:39, Emilio Cota wrote: > We forgot to add this one in "a890643958 util/qht: atomically set b->hashes". > > Detected with tsan. > > Signed-off-by: Emilio Cota <cota@braap.org> > --- > util/qht.c | 2 +- > 1 file changed, 1 insertion(+), 1 deletion(-) Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org> ^ permalink raw reply [flat|nested] 12+ messages in thread
* [PATCH 3/4] util/qht: use striped locks under TSAN 2023-01-08 16:39 [PATCH 0/4] tsan fixes Emilio Cota 2023-01-08 16:39 ` [PATCH 1/4] cpu: free cpu->tb_jmp_cache with RCU Emilio Cota 2023-01-08 16:39 ` [PATCH 2/4] util/qht: add missing atomic_set(hashes[i]) Emilio Cota @ 2023-01-08 16:39 ` Emilio Cota 2023-01-08 19:51 ` Richard Henderson 2023-01-08 16:39 ` [PATCH 4/4] plugins: make qemu_plugin_user_exit's locking order consistent with fork_start's Emilio Cota 3 siblings, 1 reply; 12+ messages in thread From: Emilio Cota @ 2023-01-08 16:39 UTC (permalink / raw) To: qemu-devel Cc: Alex Bennée, Paolo Bonzini, Richard Henderson, Emilio Cota Fixes this tsan crash, easy to reproduce with any large enough program: $ tests/unit/test-qht 1..2 ThreadSanitizer: CHECK failed: sanitizer_deadlock_detector.h:67 "((n_all_locks_)) < (((sizeof(all_locks_with_contexts_)/sizeof((all_locks_with_contexts_)[0]))))" (0x40, 0x40) (tid=1821568) #0 __tsan::CheckUnwind() ../../../../src/libsanitizer/tsan/tsan_rtl.cpp:353 (libtsan.so.2+0x90034) #1 __sanitizer::CheckFailed(char const*, int, char const*, unsigned long long, unsigned long long) ../../../../src/libsanitizer/sanitizer_common/sanitizer_termination.cpp:86 (libtsan.so.2+0xca555) #2 __sanitizer::DeadlockDetectorTLS<__sanitizer::TwoLevelBitVector<1ul, __sanitizer::BasicBitVector<unsigned long> > >::addLock(unsigned long, unsigned long, unsigned int) ../../../../src/libsanitizer/sanitizer_common/sanitizer_deadlock_detector.h:67 (libtsan.so.2+0xb3616) #3 __sanitizer::DeadlockDetectorTLS<__sanitizer::TwoLevelBitVector<1ul, __sanitizer::BasicBitVector<unsigned long> > >::addLock(unsigned long, unsigned long, unsigned int) ../../../../src/libsanitizer/sanitizer_common/sanitizer_deadlock_detector.h:59 (libtsan.so.2+0xb3616) #4 __sanitizer::DeadlockDetector<__sanitizer::TwoLevelBitVector<1ul, __sanitizer::BasicBitVector<unsigned long> > >::onLockAfter(__sanitizer::DeadlockDetectorTLS<__sanitizer::TwoLevelBitVector<1ul, __sanitizer::BasicBitVector<unsigned long> > >*, unsigned long, unsigned int) ../../../../src/libsanitizer/sanitizer_common/sanitizer_deadlock_detector.h:216 (libtsan.so.2+0xb3616) #5 __sanitizer::DD::MutexAfterLock(__sanitizer::DDCallback*, __sanitizer::DDMutex*, bool, bool) ../../../../src/libsanitizer/sanitizer_common/sanitizer_deadlock_detector1.cpp:169 (libtsan.so.2+0xb3616) #6 __tsan::MutexPostLock(__tsan::ThreadState*, unsigned long, unsigned long, unsigned int, int) ../../../../src/libsanitizer/tsan/tsan_rtl_mutex.cpp:200 (libtsan.so.2+0xa3382) #7 __tsan_mutex_post_lock ../../../../src/libsanitizer/tsan/tsan_interface_ann.cpp:384 (libtsan.so.2+0x76bc3) #8 qemu_spin_lock /home/cota/src/qemu/include/qemu/thread.h:259 (test-qht+0x44a97) #9 qht_map_lock_buckets ../util/qht.c:253 (test-qht+0x44a97) #10 do_qht_iter ../util/qht.c:809 (test-qht+0x45f33) #11 qht_iter ../util/qht.c:821 (test-qht+0x45f33) #12 iter_check ../tests/unit/test-qht.c:121 (test-qht+0xe473) #13 qht_do_test ../tests/unit/test-qht.c:202 (test-qht+0xe473) #14 qht_test ../tests/unit/test-qht.c:240 (test-qht+0xe7c1) #15 test_default ../tests/unit/test-qht.c:246 (test-qht+0xe828) #16 <null> <null> (libglib-2.0.so.0+0x7daed) #17 <null> <null> (libglib-2.0.so.0+0x7d80a) #18 <null> <null> (libglib-2.0.so.0+0x7d80a) #19 g_test_run_suite <null> (libglib-2.0.so.0+0x7dfe9) #20 g_test_run <null> (libglib-2.0.so.0+0x7e055) #21 main ../tests/unit/test-qht.c:259 (test-qht+0xd2c6) #22 __libc_start_call_main ../sysdeps/nptl/libc_start_call_main.h:58 (libc.so.6+0x29d8f) #23 __libc_start_main_impl ../csu/libc-start.c:392 (libc.so.6+0x29e3f) #24 _start <null> (test-qht+0xdb44) Signed-off-by: Emilio Cota <cota@braap.org> --- util/qht.c | 107 ++++++++++++++++++++++++++++++++++++++++++++++------- 1 file changed, 93 insertions(+), 14 deletions(-) diff --git a/util/qht.c b/util/qht.c index 15866299e6..6174533f10 100644 --- a/util/qht.c +++ b/util/qht.c @@ -151,6 +151,22 @@ struct qht_bucket { QEMU_BUILD_BUG_ON(sizeof(struct qht_bucket) > QHT_BUCKET_ALIGN); +/* + * Under TSAN, we use striped locks instead of one lock per bucket chain. + * This avoids crashing under TSAN, since TSAN aborts the program if more than + * 64 locks are held (this is a hardcoded limit in TSAN). + * When resizing a QHT we grab all the buckets' locks, which can easily + * go over TSAN's limit. By using striped locks, we avoid this problem. + * + * Note: this number must be a power of two for easy index computation. + */ +#define QHT_TSAN_BUCKET_LOCKS_BITS 4 +#define QHT_TSAN_BUCKET_LOCKS (1 << QHT_TSAN_BUCKET_LOCKS_BITS) + +struct qht_tsan_lock { + QemuSpin lock; +} QEMU_ALIGNED(QHT_BUCKET_ALIGN); + /** * struct qht_map - structure to track an array of buckets * @rcu: used by RCU. Keep it as the top field in the struct to help valgrind @@ -160,6 +176,7 @@ QEMU_BUILD_BUG_ON(sizeof(struct qht_bucket) > QHT_BUCKET_ALIGN); * @n_added_buckets: number of added (i.e. "non-head") buckets * @n_added_buckets_threshold: threshold to trigger an upward resize once the * number of added buckets surpasses it. + * @tsan_bucket_locks: Array of striped locks to be used only under TSAN. * * Buckets are tracked in what we call a "map", i.e. this structure. */ @@ -169,6 +186,9 @@ struct qht_map { size_t n_buckets; size_t n_added_buckets; size_t n_added_buckets_threshold; +#ifdef CONFIG_TSAN + struct qht_tsan_lock tsan_bucket_locks[QHT_TSAN_BUCKET_LOCKS]; +#endif }; /* trigger a resize when n_added_buckets > n_buckets / div */ @@ -229,10 +249,68 @@ static inline size_t qht_elems_to_buckets(size_t n_elems) return pow2ceil(n_elems / QHT_BUCKET_ENTRIES); } -static inline void qht_head_init(struct qht_bucket *b) +/* + * When using striped locks (i.e. under TSAN), we have to be careful not + * to operate on the same lock twice (e.g. when iterating through all buckets). + * We achieve this by operating only on each stripe's first matching lock. + */ +static inline void qht_do_if_first_in_stripe(const struct qht_map *map, + struct qht_bucket *b, + void (*func)(QemuSpin *spin)) +{ +#ifdef CONFIG_TSAN + unsigned long bucket_idx = b - map->buckets; + bool is_first_in_stripe = (bucket_idx >> QHT_TSAN_BUCKET_LOCKS_BITS) == 0; + if (is_first_in_stripe) { + unsigned long lock_idx = bucket_idx & (QHT_TSAN_BUCKET_LOCKS - 1); + func(&map->tsan_bucket_locks[lock_idx]); + } +#else + func(&b->lock); +#endif +} + +static inline void qht_bucket_lock_init(const struct qht_map *map, + struct qht_bucket *b) +{ + qht_do_if_first_in_stripe(map, b, qemu_spin_init); +} + +static inline void qht_bucket_lock_destroy(const struct qht_map *map, + struct qht_bucket *b) +{ + qht_do_if_first_in_stripe(map, b, (void (*)(QemuSpin *))qemu_spin_destroy); +} + +static inline void qht_bucket_lock_do(const struct qht_map *map, + struct qht_bucket *b, + void (*func)(QemuSpin *lock)) +{ +#ifdef CONFIG_TSAN + unsigned long bucket_idx = b - map->buckets; + unsigned long lock_idx = bucket_idx & (QHT_TSAN_BUCKET_LOCKS - 1); + func(&map->tsan_bucket_locks[lock_idx]); +#else + func(&b->lock); +#endif +} + +static inline void qht_bucket_lock(const struct qht_map *map, + struct qht_bucket *b) +{ + qht_bucket_lock_do(map, b, qemu_spin_lock); +} + +static inline void qht_bucket_unlock(const struct qht_map *map, + struct qht_bucket *b) +{ + qht_bucket_lock_do(map, b, qemu_spin_unlock); +} + +static inline void qht_head_init(struct qht_map *map, struct qht_bucket *b) { memset(b, 0, sizeof(*b)); - qemu_spin_init(&b->lock); + qht_bucket_lock_init(map, b); seqlock_init(&b->sequence); } @@ -250,7 +328,7 @@ static void qht_map_lock_buckets(struct qht_map *map) for (i = 0; i < map->n_buckets; i++) { struct qht_bucket *b = &map->buckets[i]; - qemu_spin_lock(&b->lock); + qht_do_if_first_in_stripe(map, b, qemu_spin_lock); } } @@ -261,7 +339,7 @@ static void qht_map_unlock_buckets(struct qht_map *map) for (i = 0; i < map->n_buckets; i++) { struct qht_bucket *b = &map->buckets[i]; - qemu_spin_unlock(&b->lock); + qht_do_if_first_in_stripe(map, b, qemu_spin_unlock); } } @@ -308,7 +386,7 @@ void qht_map_lock_buckets__no_stale(struct qht *ht, struct qht_map **pmap) * Get a head bucket and lock it, making sure its parent map is not stale. * @pmap is filled with a pointer to the bucket's parent map. * - * Unlock with qemu_spin_unlock(&b->lock). + * Unlock with qht_bucket_unlock. * * Note: callers cannot have ht->lock held. */ @@ -322,18 +400,18 @@ struct qht_bucket *qht_bucket_lock__no_stale(struct qht *ht, uint32_t hash, map = qatomic_rcu_read(&ht->map); b = qht_map_to_bucket(map, hash); - qemu_spin_lock(&b->lock); + qht_bucket_lock(map, b); if (likely(!qht_map_is_stale__locked(ht, map))) { *pmap = map; return b; } - qemu_spin_unlock(&b->lock); + qht_bucket_unlock(map, b); /* we raced with a resize; acquire ht->lock to see the updated ht->map */ qht_lock(ht); map = ht->map; b = qht_map_to_bucket(map, hash); - qemu_spin_lock(&b->lock); + qht_bucket_lock(map, b); qht_unlock(ht); *pmap = map; return b; @@ -345,12 +423,13 @@ static inline bool qht_map_needs_resize(const struct qht_map *map) map->n_added_buckets_threshold; } -static inline void qht_chain_destroy(const struct qht_bucket *head) +static inline void qht_chain_destroy(const struct qht_map *map, + struct qht_bucket *head) { struct qht_bucket *curr = head->next; struct qht_bucket *prev; - qemu_spin_destroy(&head->lock); + qht_bucket_lock_destroy(map, head); while (curr) { prev = curr; curr = curr->next; @@ -364,7 +443,7 @@ static void qht_map_destroy(struct qht_map *map) size_t i; for (i = 0; i < map->n_buckets; i++) { - qht_chain_destroy(&map->buckets[i]); + qht_chain_destroy(map, &map->buckets[i]); } qemu_vfree(map->buckets); g_free(map); @@ -390,7 +469,7 @@ static struct qht_map *qht_map_create(size_t n_buckets) map->buckets = qemu_memalign(QHT_BUCKET_ALIGN, sizeof(*map->buckets) * n_buckets); for (i = 0; i < n_buckets; i++) { - qht_head_init(&map->buckets[i]); + qht_head_init(map, &map->buckets[i]); } return map; } @@ -638,7 +717,7 @@ bool qht_insert(struct qht *ht, void *p, uint32_t hash, void **existing) b = qht_bucket_lock__no_stale(ht, hash, &map); prev = qht_insert__locked(ht, map, b, p, hash, &needs_resize); qht_bucket_debug__locked(b); - qemu_spin_unlock(&b->lock); + qht_bucket_unlock(map, b); if (unlikely(needs_resize) && ht->mode & QHT_MODE_AUTO_RESIZE) { qht_grow_maybe(ht); @@ -749,7 +828,7 @@ bool qht_remove(struct qht *ht, const void *p, uint32_t hash) b = qht_bucket_lock__no_stale(ht, hash, &map); ret = qht_remove__locked(b, p, hash); qht_bucket_debug__locked(b); - qemu_spin_unlock(&b->lock); + qht_bucket_unlock(map, b); return ret; } -- 2.34.1 ^ permalink raw reply related [flat|nested] 12+ messages in thread
* Re: [PATCH 3/4] util/qht: use striped locks under TSAN 2023-01-08 16:39 ` [PATCH 3/4] util/qht: use striped locks under TSAN Emilio Cota @ 2023-01-08 19:51 ` Richard Henderson 2023-01-09 22:10 ` Emilio Cota 0 siblings, 1 reply; 12+ messages in thread From: Richard Henderson @ 2023-01-08 19:51 UTC (permalink / raw) To: Emilio Cota, qemu-devel; +Cc: Alex Bennée, Paolo Bonzini On 1/8/23 08:39, Emilio Cota wrote: > +static inline void qht_bucket_lock_init(const struct qht_map *map, > + struct qht_bucket *b) > +{ > + qht_do_if_first_in_stripe(map, b, qemu_spin_init); > +} > + > +static inline void qht_bucket_lock_destroy(const struct qht_map *map, > + struct qht_bucket *b) > +{ > + qht_do_if_first_in_stripe(map, b, (void (*)(QemuSpin *))qemu_spin_destroy); > +} I think it's clear the declaration of qemu_spin_destroy is wrong vs const. Why do you create two wrappers for qht_do_if_first_in_stripe, but... > +static inline void qht_head_init(struct qht_map *map, struct qht_bucket *b) > { > memset(b, 0, sizeof(*b)); > - qemu_spin_init(&b->lock); > + qht_bucket_lock_init(map, b); > seqlock_init(&b->sequence); > } > > @@ -250,7 +328,7 @@ static void qht_map_lock_buckets(struct qht_map *map) > for (i = 0; i < map->n_buckets; i++) { > struct qht_bucket *b = &map->buckets[i]; > > - qemu_spin_lock(&b->lock); > + qht_do_if_first_in_stripe(map, b, qemu_spin_lock); > } ... not others? The actual logic of the stripe seems ok. r~ ^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: [PATCH 3/4] util/qht: use striped locks under TSAN 2023-01-08 19:51 ` Richard Henderson @ 2023-01-09 22:10 ` Emilio Cota 0 siblings, 0 replies; 12+ messages in thread From: Emilio Cota @ 2023-01-09 22:10 UTC (permalink / raw) To: Richard Henderson; +Cc: qemu-devel, Alex Bennée, Paolo Bonzini On Sun, Jan 08, 2023 at 11:51:44 -0800, Richard Henderson wrote: > On 1/8/23 08:39, Emilio Cota wrote: > > +static inline void qht_bucket_lock_init(const struct qht_map *map, > > + struct qht_bucket *b) > > +{ > > + qht_do_if_first_in_stripe(map, b, qemu_spin_init); > > +} > > + > > +static inline void qht_bucket_lock_destroy(const struct qht_map *map, > > + struct qht_bucket *b) > > +{ > > + qht_do_if_first_in_stripe(map, b, (void (*)(QemuSpin *))qemu_spin_destroy); > > +} > > I think it's clear the declaration of qemu_spin_destroy is wrong vs const. Done, added a separate patch for this. > Why do you create two wrappers for qht_do_if_first_in_stripe, but... > > > +static inline void qht_head_init(struct qht_map *map, struct qht_bucket *b) > > { > > memset(b, 0, sizeof(*b)); > > - qemu_spin_init(&b->lock); > > + qht_bucket_lock_init(map, b); > > seqlock_init(&b->sequence); > > } > > @@ -250,7 +328,7 @@ static void qht_map_lock_buckets(struct qht_map *map) > > for (i = 0; i < map->n_buckets; i++) { > > struct qht_bucket *b = &map->buckets[i]; > > - qemu_spin_lock(&b->lock); > > + qht_do_if_first_in_stripe(map, b, qemu_spin_lock); > > } > > ... not others? Thought the remaining two cases were obvious enough. I've now removed the wrappers. > The actual logic of the stripe seems ok. Thanks! Emilio ^ permalink raw reply [flat|nested] 12+ messages in thread
* [PATCH 4/4] plugins: make qemu_plugin_user_exit's locking order consistent with fork_start's 2023-01-08 16:39 [PATCH 0/4] tsan fixes Emilio Cota ` (2 preceding siblings ...) 2023-01-08 16:39 ` [PATCH 3/4] util/qht: use striped locks under TSAN Emilio Cota @ 2023-01-08 16:39 ` Emilio Cota 2023-01-08 19:31 ` Richard Henderson 3 siblings, 1 reply; 12+ messages in thread From: Emilio Cota @ 2023-01-08 16:39 UTC (permalink / raw) To: qemu-devel Cc: Alex Bennée, Paolo Bonzini, Richard Henderson, Emilio Cota To fix potential deadlocks as reported by tsan. Signed-off-by: Emilio Cota <cota@braap.org> --- plugins/core.c | 16 +++++++++++----- 1 file changed, 11 insertions(+), 5 deletions(-) diff --git a/plugins/core.c b/plugins/core.c index ccb770a485..728bacef95 100644 --- a/plugins/core.c +++ b/plugins/core.c @@ -500,10 +500,17 @@ void qemu_plugin_user_exit(void) enum qemu_plugin_event ev; CPUState *cpu; - QEMU_LOCK_GUARD(&plugin.lock); - + /* + * Locking order: we must acquire locks in an order that is consistent + * with the one in fork_start(). That is: + * - start_exclusive(), which acquires qemu_cpu_list_lock, + * must be called before acquiring plugin.lock. + * - tb_flush(), which acquires mmap_lock(), must be called + * while plugin.lock is not held. + */ start_exclusive(); + qemu_rec_mutex_lock(&plugin.lock); /* un-register all callbacks except the final AT_EXIT one */ for (ev = 0; ev < QEMU_PLUGIN_EV_MAX; ev++) { if (ev != QEMU_PLUGIN_EV_ATEXIT) { @@ -513,13 +520,12 @@ void qemu_plugin_user_exit(void) } } } - - tb_flush(current_cpu); - CPU_FOREACH(cpu) { qemu_plugin_disable_mem_helpers(cpu); } + qemu_rec_mutex_unlock(&plugin.lock); + tb_flush(current_cpu); end_exclusive(); /* now it's safe to handle the exit case */ -- 2.34.1 ^ permalink raw reply related [flat|nested] 12+ messages in thread
* Re: [PATCH 4/4] plugins: make qemu_plugin_user_exit's locking order consistent with fork_start's 2023-01-08 16:39 ` [PATCH 4/4] plugins: make qemu_plugin_user_exit's locking order consistent with fork_start's Emilio Cota @ 2023-01-08 19:31 ` Richard Henderson 0 siblings, 0 replies; 12+ messages in thread From: Richard Henderson @ 2023-01-08 19:31 UTC (permalink / raw) To: Emilio Cota, qemu-devel; +Cc: Alex Bennée, Paolo Bonzini On 1/8/23 08:39, Emilio Cota wrote: > To fix potential deadlocks as reported by tsan. > > Signed-off-by: Emilio Cota<cota@braap.org> > --- > plugins/core.c | 16 +++++++++++----- > 1 file changed, 11 insertions(+), 5 deletions(-) Reviewed-by: Richard Henderson <richard.henderson@linaro.org> r~ ^ permalink raw reply [flat|nested] 12+ messages in thread
end of thread, other threads:[~2023-01-09 22:21 UTC | newest] Thread overview: 12+ messages (download: mbox.gz follow: Atom feed -- links below jump to the message on this page -- 2023-01-08 16:39 [PATCH 0/4] tsan fixes Emilio Cota 2023-01-08 16:39 ` [PATCH 1/4] cpu: free cpu->tb_jmp_cache with RCU Emilio Cota 2023-01-08 19:19 ` Richard Henderson 2023-01-09 21:52 ` Emilio Cota 2023-01-08 16:39 ` [PATCH 2/4] util/qht: add missing atomic_set(hashes[i]) Emilio Cota 2023-01-08 19:20 ` Richard Henderson 2023-01-08 19:54 ` Philippe Mathieu-Daudé 2023-01-08 16:39 ` [PATCH 3/4] util/qht: use striped locks under TSAN Emilio Cota 2023-01-08 19:51 ` Richard Henderson 2023-01-09 22:10 ` Emilio Cota 2023-01-08 16:39 ` [PATCH 4/4] plugins: make qemu_plugin_user_exit's locking order consistent with fork_start's Emilio Cota 2023-01-08 19:31 ` Richard Henderson
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox; as well as URLs for NNTP newsgroup(s).