Searched refs:cur_state (Results 1 – 5 of 5) sorted by relevance
/art/runtime/base/ |
D | mutex-inl.h | 192 int32_t cur_state = state_.load(std::memory_order_relaxed); in SharedLock() local 193 if (LIKELY(cur_state >= 0)) { in SharedLock() 195 done = state_.CompareAndSetWeakAcquire(cur_state, cur_state + 1); in SharedLock() 197 HandleSharedLockContention(self, cur_state); in SharedLock() 216 int32_t cur_state = state_.load(std::memory_order_relaxed); in SharedUnlock() local 217 if (LIKELY(cur_state > 0)) { in SharedUnlock() 220 done = state_.CompareAndSetWeakSequentiallyConsistent(cur_state, cur_state - 1); in SharedUnlock() 221 if (done && (cur_state - 1) == 0) { // Weak CAS may fail spuriously. in SharedUnlock() 228 LOG(FATAL) << "Unexpected state_:" << cur_state << " for " << name_; in SharedUnlock()
|
D | mutex.cc | 461 int32_t cur_state = state_and_contenders_.load(std::memory_order_relaxed); in ExclusiveLock() local 462 if (LIKELY((cur_state & kHeldMask) == 0) /* lock not held */) { in ExclusiveLock() 463 done = state_and_contenders_.CompareAndSetWeakAcquire(cur_state, cur_state | kHeldMask); in ExclusiveLock() 475 cur_state += kContenderIncrement; in ExclusiveLock() 489 if (futex(state_and_contenders_.Address(), FUTEX_WAIT_PRIVATE, cur_state, in ExclusiveLock() 509 cur_state = state_and_contenders_.load(std::memory_order_relaxed); in ExclusiveLock() 510 } while ((cur_state & kHeldMask) != 0); in ExclusiveLock() 601 int32_t cur_state = state_and_contenders_.load(std::memory_order_relaxed); in ExclusiveTryLock() local 602 if ((cur_state & kHeldMask) == 0) { in ExclusiveTryLock() 604 done = state_and_contenders_.CompareAndSetWeakAcquire(cur_state, cur_state | kHeldMask); in ExclusiveTryLock() [all …]
|
D | mutex.h | 410 void HandleSharedLockContention(Thread* self, int32_t cur_state);
|
/art/test/924-threads/src/art/ |
D | Test924.java | 425 int cur_state = getThreadState(t); in printMajorityThreadState() local 426 states.put(cur_state, states.getOrDefault(cur_state, 0) + 1); in printMajorityThreadState()
|
/art/runtime/gc/collector/ |
D | mark_compact.cc | 2563 uint32_t cur_state = moving_pages_status_[arr_idx].load(std::memory_order_acquire); in MapMovingSpacePages() local 2570 DCHECK_EQ((cur_state & ~kPageStateMask) + (i - arr_idx) * gPageSize, s & ~kPageStateMask); in MapMovingSpacePages() 2575 bool mapped = GetPageStateFromWord(cur_state) == PageState::kProcessedAndMapped; in MapMovingSpacePages() 2588 uint32_t from_space_offset = cur_state & ~kPageStateMask; in MapMovingSpacePages() 3626 uint32_t cur_state = moving_pages_status_[idx].load(std::memory_order_acquire); in ConcurrentlyProcessMovingPage() local 3627 if (cur_state != static_cast<uint8_t>(PageState::kUnprocessed)) { in ConcurrentlyProcessMovingPage() 3628 DCHECK_EQ(cur_state, static_cast<uint8_t>(PageState::kProcessedAndMapped)); in ConcurrentlyProcessMovingPage() 3764 for (Atomic<PageState>* cur_state = state; in MapUpdatedLinearAllocPages() local 3765 map_len < length && cur_state->load(std::memory_order_acquire) == PageState::kProcessed; in MapUpdatedLinearAllocPages() 3766 map_len += gPageSize, cur_state++) { in MapUpdatedLinearAllocPages()
|