| lh | 9ed821d | 2023-04-07 01:36:19 -0700 | [diff] [blame] | 1 | /* | 
|  | 2 | * include/asm-generic/mutex-xchg.h | 
|  | 3 | * | 
|  | 4 | * Generic implementation of the mutex fastpath, based on xchg(). | 
|  | 5 | * | 
|  | 6 | * NOTE: An xchg based implementation might be less optimal than an atomic | 
|  | 7 | *       decrement/increment based implementation. If your architecture | 
|  | 8 | *       has a reasonable atomic dec/inc then you should probably use | 
|  | 9 | *	 asm-generic/mutex-dec.h instead, or you could open-code an | 
|  | 10 | *	 optimized version in asm/mutex.h. | 
|  | 11 | */ | 
|  | 12 | #ifndef _ASM_GENERIC_MUTEX_XCHG_H | 
|  | 13 | #define _ASM_GENERIC_MUTEX_XCHG_H | 
|  | 14 |  | 
|  | 15 | /** | 
|  | 16 | *  __mutex_fastpath_lock - try to take the lock by moving the count | 
|  | 17 | *                          from 1 to a 0 value | 
|  | 18 | *  @count: pointer of type atomic_t | 
|  | 19 | *  @fail_fn: function to call if the original value was not 1 | 
|  | 20 | * | 
|  | 21 | * Change the count from 1 to a value lower than 1, and call <fail_fn> if it | 
|  | 22 | * wasn't 1 originally. This function MUST leave the value lower than 1 | 
|  | 23 | * even when the "1" assertion wasn't true. | 
|  | 24 | */ | 
|  | 25 | static inline void | 
|  | 26 | __mutex_fastpath_lock(atomic_t *count, void (*fail_fn)(atomic_t *)) | 
|  | 27 | { | 
|  | 28 | if (unlikely(atomic_xchg(count, 0) != 1)) | 
|  | 29 | /* | 
|  | 30 | * We failed to acquire the lock, so mark it contended | 
|  | 31 | * to ensure that any waiting tasks are woken up by the | 
|  | 32 | * unlock slow path. | 
|  | 33 | */ | 
|  | 34 | if (likely(atomic_xchg(count, -1) != 1)) | 
|  | 35 | fail_fn(count); | 
|  | 36 | } | 
|  | 37 |  | 
|  | 38 | /** | 
|  | 39 | *  __mutex_fastpath_lock_retval - try to take the lock by moving the count | 
|  | 40 | *                                 from 1 to a 0 value | 
|  | 41 | *  @count: pointer of type atomic_t | 
|  | 42 | *  @fail_fn: function to call if the original value was not 1 | 
|  | 43 | * | 
|  | 44 | * Change the count from 1 to a value lower than 1, and call <fail_fn> if it | 
|  | 45 | * wasn't 1 originally. This function returns 0 if the fastpath succeeds, | 
|  | 46 | * or anything the slow path function returns | 
|  | 47 | */ | 
|  | 48 | static inline int | 
|  | 49 | __mutex_fastpath_lock_retval(atomic_t *count, int (*fail_fn)(atomic_t *)) | 
|  | 50 | { | 
|  | 51 | if (unlikely(atomic_xchg(count, 0) != 1)) | 
|  | 52 | if (likely(atomic_xchg(count, -1) != 1)) | 
|  | 53 | return fail_fn(count); | 
|  | 54 | return 0; | 
|  | 55 | } | 
|  | 56 |  | 
|  | 57 | /** | 
|  | 58 | *  __mutex_fastpath_unlock - try to promote the mutex from 0 to 1 | 
|  | 59 | *  @count: pointer of type atomic_t | 
|  | 60 | *  @fail_fn: function to call if the original value was not 0 | 
|  | 61 | * | 
|  | 62 | * try to promote the mutex from 0 to 1. if it wasn't 0, call <function> | 
|  | 63 | * In the failure case, this function is allowed to either set the value to | 
|  | 64 | * 1, or to set it to a value lower than one. | 
|  | 65 | * If the implementation sets it to a value of lower than one, the | 
|  | 66 | * __mutex_slowpath_needs_to_unlock() macro needs to return 1, it needs | 
|  | 67 | * to return 0 otherwise. | 
|  | 68 | */ | 
|  | 69 | static inline void | 
|  | 70 | __mutex_fastpath_unlock(atomic_t *count, void (*fail_fn)(atomic_t *)) | 
|  | 71 | { | 
|  | 72 | if (unlikely(atomic_xchg(count, 1) != 0)) | 
|  | 73 | fail_fn(count); | 
|  | 74 | } | 
|  | 75 |  | 
|  | 76 | #define __mutex_slowpath_needs_to_unlock()		0 | 
|  | 77 |  | 
|  | 78 | /** | 
|  | 79 | * __mutex_fastpath_trylock - try to acquire the mutex, without waiting | 
|  | 80 | * | 
|  | 81 | *  @count: pointer of type atomic_t | 
|  | 82 | *  @fail_fn: spinlock based trylock implementation | 
|  | 83 | * | 
|  | 84 | * Change the count from 1 to a value lower than 1, and return 0 (failure) | 
|  | 85 | * if it wasn't 1 originally, or return 1 (success) otherwise. This function | 
|  | 86 | * MUST leave the value lower than 1 even when the "1" assertion wasn't true. | 
|  | 87 | * Additionally, if the value was < 0 originally, this function must not leave | 
|  | 88 | * it to 0 on failure. | 
|  | 89 | * | 
|  | 90 | * If the architecture has no effective trylock variant, it should call the | 
|  | 91 | * <fail_fn> spinlock-based trylock variant unconditionally. | 
|  | 92 | */ | 
|  | 93 | static inline int | 
|  | 94 | __mutex_fastpath_trylock(atomic_t *count, int (*fail_fn)(atomic_t *)) | 
|  | 95 | { | 
|  | 96 | int prev = atomic_xchg(count, 0); | 
|  | 97 |  | 
|  | 98 | if (unlikely(prev < 0)) { | 
|  | 99 | /* | 
|  | 100 | * The lock was marked contended so we must restore that | 
|  | 101 | * state. If while doing so we get back a prev value of 1 | 
|  | 102 | * then we just own it. | 
|  | 103 | * | 
|  | 104 | * [ In the rare case of the mutex going to 1, to 0, to -1 | 
|  | 105 | *   and then back to 0 in this few-instructions window, | 
|  | 106 | *   this has the potential to trigger the slowpath for the | 
|  | 107 | *   owner's unlock path needlessly, but that's not a problem | 
|  | 108 | *   in practice. ] | 
|  | 109 | */ | 
|  | 110 | prev = atomic_xchg(count, prev); | 
|  | 111 | if (prev < 0) | 
|  | 112 | prev = 0; | 
|  | 113 | } | 
|  | 114 |  | 
|  | 115 | return prev; | 
|  | 116 | } | 
|  | 117 |  | 
|  | 118 | #endif |