public inbox for linux-kernel@vger.kernel.org
 help / color / mirror / Atom feed
* osq_lock on arm64
@ 2026-03-15 20:11 David Laight
  2026-03-15 22:02 ` Marc Zyngier
  0 siblings, 1 reply; 2+ messages in thread
From: David Laight @ 2026-03-15 20:11 UTC (permalink / raw)
  To: Boqun Feng, Catalin Marinas, Ingo Molnar, linux-arm-kernel,
	linux-kernel, Lorenzo Pieralisi, Marc Zyngier, Peter Zijlstra,
	Thomas Huth, Waiman Long, Will Deacon, Linus Torvalds

Commit f5bfdc8e3947 change the osq_lock (optimistic spin queue) code to
use smp_cond_load_relaxed() to wait for another cpu to relase the lock.

It waits for 'VAL || need_resched()' with a comment that the IPI that
sets need_resched() will break any monitor/wait code.
However arm64 implements it as (03e3c2b7edbe and 1cfc63b5ae6):

#define smp_cond_load_acquire(ptr, cond_expr)                          \
({                                                                     \
       typeof(ptr) __PTR = (ptr);                                      \
       typeof(*ptr) VAL;                                               \
       for (;;) {                                                      \
               VAL = smp_load_acquire(__PTR);                          \
               if (cond_expr)                                          \
                       break;                                          \
               __cmpwait_relaxed(__PTR, VAL);                          \
       }                                                               \
       VAL;                                                            \
})

#define __CMPWAIT_CASE(w, sz, name)                                    \
static inline void __cmpwait_case_##name(volatile void *ptr,           \
                                        unsigned long val)             \
{                                                                      \
       unsigned long tmp;                                              \
                                                                       \
       asm volatile(                                                   \
       "       sevl\n"                                                 \
       "       wfe\n"                                                  \
       "       ldxr" #sz "\t%" #w "[tmp], %[v]\n"                      \
       "       eor     %" #w "[tmp], %" #w "[tmp], %" #w "[val]\n"     \
       "       cbnz    %" #w "[tmp], 1f\n"                             \
       "       wfe\n"                                                  \
       "1:"                                                            \
       : [tmp] "=&r" (tmp), [v] "+Q" (*(unsigned long *)ptr)           \
       : [val] "r" (val));                                             \
}

The "eor" and "cbnz" seem to be there to skip the "wfe" if the value has
already changed, but this only works if the condition only depends on VAL.
This means it can miss the IPI that makes need_resched() return non-zero.

To work I think it would need the cmp_load_acquire() to be "ldxr" with
a "wfe" instead of __cmpwait_relaxed().
But that wouldn't work if 'cond_expr' issued another "ldx".

The documentation I found for the arm64 instruction set doesn't mention that
losing the exclusive access requested by ldx will wake wfe, I assume that
is an omission from the docs?

	David 


^ permalink raw reply	[flat|nested] 2+ messages in thread

end of thread, other threads:[~2026-03-15 22:02 UTC | newest]

Thread overview: 2+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2026-03-15 20:11 osq_lock on arm64 David Laight
2026-03-15 22:02 ` Marc Zyngier

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox