Searched refs:URW_WRITE_LOCKED (Results 1 – 6 of 6) sorted by relevance
40 ASSERT(!((readers) & URW_WRITE_LOCKED) || \41 ((readers) & ~URW_HAS_WAITERS) == URW_WRITE_LOCKED)150 if (!(readers & URW_WRITE_LOCKED) && in rw_read_held()191 rval = ((readers & URW_WRITE_LOCKED) && in rw_write_held()286 URW_WRITE_LOCKED : (URW_HAS_WAITERS | URW_WRITE_LOCKED); in read_lock_try()330 (URW_WRITE_LOCKED | URW_READERS_MASK) : in write_lock_try()397 atomic_and_32(rwstate, ~URW_WRITE_LOCKED); in rw_queue_release()460 atomic_or_32(rwstate, URW_WRITE_LOCKED); in rw_queue_release()614 if ((readers & URW_WRITE_LOCKED) || in rwlock_lock()682 (*rwstate & URW_WRITE_LOCKED) && in rwlock_lock()[all …]
300 if (rwstate & URW_WRITE_LOCKED) { in rwlock_error()
76 #define URW_WRITE_LOCKED 0x40000000 macro
2270 rwstate |= URW_WRITE_LOCKED; in lwp_rwlock_release()2489 if (!(rwstate & URW_WRITE_LOCKED)) { in lwp_rwlock_lock()2505 rwstate |= URW_WRITE_LOCKED; in lwp_rwlock_lock()2772 if (rwstate & URW_WRITE_LOCKED) in lwp_rwlock_unlock()
2576 if (rwstate & URW_WRITE_LOCKED) { in sync_get_info_common()2962 *rwstate |= URW_WRITE_LOCKED; in __td_sync_setstate()