2 * include/asm-arm/mutex.h
4 * ARM optimized mutex locking primitives
6 * Please look into asm-generic/mutex-xchg.h for a formal definition.
11 #if __LINUX_ARM_ARCH__ < 6
12 /* On pre-ARMv6 hardware the swp based implementation is the most efficient. */
13 # include <asm-generic/mutex-xchg.h>
17 * Attempting to lock a mutex on ARMv6+ can be done with a bastardized
18 * atomic decrement (it is not a reliable atomic decrement but it satisfies
19 * the defined semantics for our purpose, while being smaller and faster
20 * than a real atomic decrement or atomic swap. The idea is to attempt
21 * decrementing the lock value only once. If once decremented it isn't zero,
22 * or if its store-back fails due to a dispute on the exclusive store, we
23 * simply bail out immediately through the slow path where the lock will be
24 * reattempted until it succeeds.
26 #define __mutex_fastpath_lock(count, fail_fn) \
28 int __ex_flag, __res; \
30 typecheck(atomic_t *, count); \
31 typecheck_fn(fastcall void (*)(atomic_t *), fail_fn); \
36 "strex %1, %0, [%2] \n" \
38 : "=&r" (__res), "=&r" (__ex_flag) \
39 : "r" (&(count)->counter) \
42 if (unlikely(__res || __ex_flag)) \
46 #define __mutex_fastpath_lock_retval(count, fail_fn) \
48 int __ex_flag, __res; \
50 typecheck(atomic_t *, count); \
51 typecheck_fn(fastcall int (*)(atomic_t *), fail_fn); \
56 "strex %1, %0, [%2] \n" \
58 : "=&r" (__res), "=&r" (__ex_flag) \
59 : "r" (&(count)->counter) \
63 if (unlikely(__res != 0)) \
64 __res = fail_fn(count); \
69 * Same trick is used for the unlock fast path. However the original value,
70 * rather than the result, is used to test for success in order to have
71 * better generated assembly.
73 #define __mutex_fastpath_unlock(count, fail_fn) \
75 int __ex_flag, __res, __orig; \
77 typecheck(atomic_t *, count); \
78 typecheck_fn(fastcall void (*)(atomic_t *), fail_fn); \
83 "strex %2, %1, [%3] \n" \
85 : "=&r" (__orig), "=&r" (__res), "=&r" (__ex_flag) \
86 : "r" (&(count)->counter) \
89 if (unlikely(__orig || __ex_flag)) \
94 * If the unlock was done on a contended lock, or if the unlock simply fails
95 * then the mutex remains locked.
97 #define __mutex_slowpath_needs_to_unlock() 1
100 * For __mutex_fastpath_trylock we use another construct which could be
101 * described as a "single value cmpxchg".
103 * This provides the needed trylock semantics like cmpxchg would, but it is
104 * lighter and less generic than a true cmpxchg implementation.
107 __mutex_fastpath_trylock(atomic_t
*count
, int (*fail_fn
)(atomic_t
*))
109 int __ex_flag
, __res
, __orig
;
113 "1: ldrex %0, [%3] \n"
115 "strexeq %2, %1, [%3] \n"
120 : "=&r" (__orig
), "=&r" (__res
), "=&r" (__ex_flag
)
121 : "r" (&count
->counter
)