Home
last modified time | relevance | path

Searched refs:cur_state (Results 1 – 5 of 5) sorted by relevance

/art/runtime/base/
Dmutex-inl.h192 int32_t cur_state = state_.load(std::memory_order_relaxed); in SharedLock() local
193 if (LIKELY(cur_state >= 0)) { in SharedLock()
195 done = state_.CompareAndSetWeakAcquire(cur_state, cur_state + 1); in SharedLock()
197 HandleSharedLockContention(self, cur_state); in SharedLock()
216 int32_t cur_state = state_.load(std::memory_order_relaxed); in SharedUnlock() local
217 if (LIKELY(cur_state > 0)) { in SharedUnlock()
220 done = state_.CompareAndSetWeakSequentiallyConsistent(cur_state, cur_state - 1); in SharedUnlock()
221 if (done && (cur_state - 1) == 0) { // Weak CAS may fail spuriously. in SharedUnlock()
228 LOG(FATAL) << "Unexpected state_:" << cur_state << " for " << name_; in SharedUnlock()
Dmutex.cc461 int32_t cur_state = state_and_contenders_.load(std::memory_order_relaxed); in ExclusiveLock() local
462 if (LIKELY((cur_state & kHeldMask) == 0) /* lock not held */) { in ExclusiveLock()
463 done = state_and_contenders_.CompareAndSetWeakAcquire(cur_state, cur_state | kHeldMask); in ExclusiveLock()
475 cur_state += kContenderIncrement; in ExclusiveLock()
489 if (futex(state_and_contenders_.Address(), FUTEX_WAIT_PRIVATE, cur_state, in ExclusiveLock()
509 cur_state = state_and_contenders_.load(std::memory_order_relaxed); in ExclusiveLock()
510 } while ((cur_state & kHeldMask) != 0); in ExclusiveLock()
601 int32_t cur_state = state_and_contenders_.load(std::memory_order_relaxed); in ExclusiveTryLock() local
602 if ((cur_state & kHeldMask) == 0) { in ExclusiveTryLock()
604 done = state_and_contenders_.CompareAndSetWeakAcquire(cur_state, cur_state | kHeldMask); in ExclusiveTryLock()
[all …]
Dmutex.h410 void HandleSharedLockContention(Thread* self, int32_t cur_state);
/art/test/924-threads/src/art/
DTest924.java425 int cur_state = getThreadState(t); in printMajorityThreadState() local
426 states.put(cur_state, states.getOrDefault(cur_state, 0) + 1); in printMajorityThreadState()
/art/runtime/gc/collector/
Dmark_compact.cc2563 uint32_t cur_state = moving_pages_status_[arr_idx].load(std::memory_order_acquire); in MapMovingSpacePages() local
2570 DCHECK_EQ((cur_state & ~kPageStateMask) + (i - arr_idx) * gPageSize, s & ~kPageStateMask); in MapMovingSpacePages()
2575 bool mapped = GetPageStateFromWord(cur_state) == PageState::kProcessedAndMapped; in MapMovingSpacePages()
2588 uint32_t from_space_offset = cur_state & ~kPageStateMask; in MapMovingSpacePages()
3626 uint32_t cur_state = moving_pages_status_[idx].load(std::memory_order_acquire); in ConcurrentlyProcessMovingPage() local
3627 if (cur_state != static_cast<uint8_t>(PageState::kUnprocessed)) { in ConcurrentlyProcessMovingPage()
3628 DCHECK_EQ(cur_state, static_cast<uint8_t>(PageState::kProcessedAndMapped)); in ConcurrentlyProcessMovingPage()
3764 for (Atomic<PageState>* cur_state = state; in MapUpdatedLinearAllocPages() local
3765 map_len < length && cur_state->load(std::memory_order_acquire) == PageState::kProcessed; in MapUpdatedLinearAllocPages()
3766 map_len += gPageSize, cur_state++) { in MapUpdatedLinearAllocPages()