mutex-xchg.h 3.8 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121
  1. /*
  2. * include/asm-generic/mutex-xchg.h
  3. *
  4. * Generic implementation of the mutex fastpath, based on xchg().
  5. *
  6. * NOTE: An xchg based implementation might be less optimal than an atomic
  7. * decrement/increment based implementation. If your architecture
  8. * has a reasonable atomic dec/inc then you should probably use
  9. * asm-generic/mutex-dec.h instead, or you could open-code an
  10. * optimized version in asm/mutex.h.
  11. */
  12. #ifndef _ASM_GENERIC_MUTEX_XCHG_H
  13. #define _ASM_GENERIC_MUTEX_XCHG_H
  14. /**
  15. * __mutex_fastpath_lock - try to take the lock by moving the count
  16. * from 1 to a 0 value
  17. * @count: pointer of type atomic_t
  18. * @fail_fn: function to call if the original value was not 1
  19. *
  20. * Change the count from 1 to a value lower than 1, and call <fail_fn> if it
  21. * wasn't 1 originally. This function MUST leave the value lower than 1
  22. * even when the "1" assertion wasn't true.
  23. */
  24. static inline void
  25. __mutex_fastpath_lock(atomic_t *count, void (*fail_fn)(atomic_t *))
  26. {
  27. if (unlikely(atomic_xchg(count, 0) != 1))
  28. /*
  29. * We failed to acquire the lock, so mark it contended
  30. * to ensure that any waiting tasks are woken up by the
  31. * unlock slow path.
  32. */
  33. if (likely(atomic_xchg_acquire(count, -1) != 1))
  34. fail_fn(count);
  35. }
  36. /**
  37. * __mutex_fastpath_lock_retval - try to take the lock by moving the count
  38. * from 1 to a 0 value
  39. * @count: pointer of type atomic_t
  40. *
  41. * Change the count from 1 to a value lower than 1. This function returns 0
  42. * if the fastpath succeeds, or -1 otherwise.
  43. */
  44. static inline int
  45. __mutex_fastpath_lock_retval(atomic_t *count)
  46. {
  47. if (unlikely(atomic_xchg_acquire(count, 0) != 1))
  48. if (likely(atomic_xchg(count, -1) != 1))
  49. return -1;
  50. return 0;
  51. }
  52. /**
  53. * __mutex_fastpath_unlock - try to promote the mutex from 0 to 1
  54. * @count: pointer of type atomic_t
  55. * @fail_fn: function to call if the original value was not 0
  56. *
  57. * try to promote the mutex from 0 to 1. if it wasn't 0, call <function>
  58. * In the failure case, this function is allowed to either set the value to
  59. * 1, or to set it to a value lower than one.
  60. * If the implementation sets it to a value of lower than one, the
  61. * __mutex_slowpath_needs_to_unlock() macro needs to return 1, it needs
  62. * to return 0 otherwise.
  63. */
  64. static inline void
  65. __mutex_fastpath_unlock(atomic_t *count, void (*fail_fn)(atomic_t *))
  66. {
  67. if (unlikely(atomic_xchg_release(count, 1) != 0))
  68. fail_fn(count);
  69. }
  70. #define __mutex_slowpath_needs_to_unlock() 0
  71. /**
  72. * __mutex_fastpath_trylock - try to acquire the mutex, without waiting
  73. *
  74. * @count: pointer of type atomic_t
  75. * @fail_fn: spinlock based trylock implementation
  76. *
  77. * Change the count from 1 to a value lower than 1, and return 0 (failure)
  78. * if it wasn't 1 originally, or return 1 (success) otherwise. This function
  79. * MUST leave the value lower than 1 even when the "1" assertion wasn't true.
  80. * Additionally, if the value was < 0 originally, this function must not leave
  81. * it to 0 on failure.
  82. *
  83. * If the architecture has no effective trylock variant, it should call the
  84. * <fail_fn> spinlock-based trylock variant unconditionally.
  85. */
  86. static inline int
  87. __mutex_fastpath_trylock(atomic_t *count, int (*fail_fn)(atomic_t *))
  88. {
  89. int prev;
  90. if (atomic_read(count) != 1)
  91. return 0;
  92. prev = atomic_xchg_acquire(count, 0);
  93. if (unlikely(prev < 0)) {
  94. /*
  95. * The lock was marked contended so we must restore that
  96. * state. If while doing so we get back a prev value of 1
  97. * then we just own it.
  98. *
  99. * [ In the rare case of the mutex going to 1, to 0, to -1
  100. * and then back to 0 in this few-instructions window,
  101. * this has the potential to trigger the slowpath for the
  102. * owner's unlock path needlessly, but that's not a problem
  103. * in practice. ]
  104. */
  105. prev = atomic_xchg_acquire(count, prev);
  106. if (prev < 0)
  107. prev = 0;
  108. }
  109. return prev;
  110. }
  111. #endif