Searched refs:atomic_load (Results 1 – 17 of 17) sorted by relevance
41 ASSERT_EQ(123, atomic_load(&v)); in TEST()44 ASSERT_EQ(456, atomic_load(&v)); in TEST()92 ASSERT_EQ(123, atomic_load(&i)); in TEST()151 ASSERT_EQ(125, atomic_load(&i)); in TEST()158 ASSERT_EQ(121, atomic_load(&i)); in TEST()165 ASSERT_EQ(0x123, atomic_load(&i)); in TEST()172 ASSERT_EQ(0x123, atomic_load(&i)); in TEST()179 ASSERT_EQ(0x002, atomic_load(&i)); in TEST()212 if (i >= BIG - 1000 && !atomic_load(&read_enough)) { in writer()273 EXPECT_EQ(atomic_load(&a.z), BIG + 1); in TEST()
228 if (atomic_load(&gHeapprofdState) == kIncompatibleHooks) { in HandleHeapprofdSignal()297 atomic_load(&gPreviousDefaultDispatchTable); in HandleHeapprofdSignal()350 void* impl_handle = atomic_load(&gHeapprofdHandle); in CommonInstallHooks()403 const MallocDispatch* previous_dispatch = atomic_load(&gPreviousDefaultDispatchTable); in MallocInitHeapprofdHook()431 const MallocDispatch* previous_dispatch = atomic_load(&gPreviousDefaultDispatchTable); in MallocInitHeapprofdHook()449 if (atomic_load(&gHeapprofdState) == kInitialState) { in DispatchReset()459 const MallocDispatch* previous_dispatch = atomic_load(&gPreviousDefaultDispatchTable); in DispatchReset()
59 if (crash_detail_t* head = atomic_load(&free_head)) { in android_crash_detail_register()103 crash_detail_t* prev = atomic_load(&free_head); in android_crash_detail_unregister()
182 if (!atomic_load(&__libc_globals->memtag)) return; in __pthread_internal_remap_stack_with_mte()256 if (atomic_load(&t->terminating)) continue; in android_run_on_all_threads()274 return atomic_load(&g_retval); in android_run_on_all_threads()
42 if (!atomic_load(&__android_fdtrack_hook)) { in track_fds()
308 if (__get_tcb_dtv(tcb)->generation == atomic_load(&modules.generation)) { in update_tls_dtv()357 dtv->generation = atomic_load(&modules.generation); in update_tls_dtv()410 size_t generation = atomic_load(&__libc_tls_generation_copy); in TLS_GET_ADDR()
64 FdTableOverflow* local_overflow = atomic_load(&overflow); in at()136 auto error_level = atomic_load(&fd_table.error_level); in fdsan_error()
223 if (atomic_load(&__libc_memtag_stack)) { in memtag_handle_longjmp()
356 *reinterpret_cast<bool*>(arg) = atomic_load(&__libc_memtag_stack); in android_mallopt()364 *reinterpret_cast<bool*>(arg) = atomic_load(&__libc_globals->decay_time_enabled); in android_mallopt()
94 if (atomic_load(&__libc_memtag_stack)) { in __init_alternate_signal_stack()240 if (atomic_load(&__libc_memtag_stack)) { in __allocate_thread_mapping()
545 *reinterpret_cast<bool*>(arg) = atomic_load(&__libc_memtag_stack); in android_mallopt()553 *reinterpret_cast<bool*>(arg) = atomic_load(&__libc_globals->decay_time_enabled); in android_mallopt()
322 int val = SEMCOUNT_TO_VALUE(atomic_load(sem_count_ptr)); in sem_getvalue()
103 if (atomic_load(&timer->deleted) == true) { in __timer_thread_start()
203 if (atomic_load(&thread->join_state) == THREAD_DETACHED) { in pthread_getattr_np()
63 atomic_load (&__android_fdtrack_hook)(&event); \92 atomic_load (&__android_fdtrack_hook)(&event); \
61 using std::atomic_load;
254 #define atomic_load(object) \ macro