/art/compiler/optimizing/ |
D | pc_relative_fixups_x86.cc | 32 base_(nullptr) {} in PCRelativeHandlerVisitor() 35 if (base_ != nullptr) { in MoveBaseIfNeeded() 39 base_->MoveBeforeFirstUserAndOutOfLoops(); in MoveBaseIfNeeded() 166 if (base_ != nullptr) { in GetPCRelativeBasePointer() 167 return base_; in GetPCRelativeBasePointer() 179 base_ = method_address; in GetPCRelativeBasePointer() 266 HX86ComputeBaseMethodAddress* base_; member in art::x86::PCRelativeHandlerVisitor
|
D | nodes_vector.h | 32 Alignment(size_t base, size_t offset) : base_(base), offset_(offset) { in Alignment() 42 return ((offset_ | base_) & (base - 1u)) == 0; in IsAlignedAt() 45 size_t Base() const { return base_; } in Base() 50 return "ALIGN(" + std::to_string(base_) + "," + std::to_string(offset_) + ")"; in ToString() 54 return base_ == other.base_ && offset_ == other.offset_; 58 size_t base_;
|
D | intrinsics_riscv64.cc | 1330 base_(base), in ReadBarrierCasSlowPathRISCV64() 1401 __ Add(tmp_ptr, base_, offset_); in EmitNativeCode() 1465 XRegister base_; member in art::riscv64::ReadBarrierCasSlowPathRISCV64
|
D | intrinsics_arm_vixl.cc | 3446 base_(base), in ReadBarrierCasSlowPathARMVIXL() 3514 __ Add(tmp_ptr, base_, offset_); in EmitNativeCode() 3555 vixl32::Register base_; member in art::arm::ReadBarrierCasSlowPathARMVIXL
|
D | intrinsics_arm64.cc | 1323 base_(base), in ReadBarrierCasSlowPathARM64() 1390 __ Add(tmp_ptr, base_.X(), Operand(offset_)); in EmitNativeCode() 1435 Register base_; member in art::arm64::ReadBarrierCasSlowPathARM64
|
D | code_generator_riscv64.cc | 270 base_(base), in CompileOptimizedSlowPathRISCV64() 281 __ Sh(counter, base_, imm12_); in EmitNativeCode() 300 XRegister base_; member in art::riscv64::CompileOptimizedSlowPathRISCV64
|
/art/libartbase/base/ |
D | transform_iterator.h | 70 ++data_.base_; 84 --data_.base_; 128 return data_.base_; in base() 140 Data(BaseIterator base, Function fn) : Function(fn), base_(base) { } in Data() 142 BaseIterator base_; member
|
D | transform_array_ref.h | 126 return data_.base_; in base() 129 return ArrayRef<const BaseType>(data_.base_); in base() 136 Data(ArrayRef<BaseType> base, Function fn) : Function(fn), base_(base) { } in Data() 138 ArrayRef<BaseType> base_; member
|
/art/runtime/gc/allocator/ |
D | rosalloc.cc | 59 : base_(reinterpret_cast<uint8_t*>(base)), footprint_(capacity), in RosAlloc() 73 memset(base_, 0, max_capacity); in RosAlloc() 75 CHECK_EQ(madvise(base_, max_capacity, MADV_DONTNEED), 0); in RosAlloc() 80 << std::hex << (intptr_t)base_ << ", end=" in RosAlloc() 81 << std::hex << (intptr_t)(base_ + capacity_) in RosAlloc() 104 FreePageRun* free_pages = reinterpret_cast<FreePageRun*>(base_); in RosAlloc() 126 MEMORY_TOOL_MAKE_DEFINED(base_, capacity_); in ~RosAlloc() 180 … if (it != free_page_runs_.rend() && (last_free_page_run = *it)->End(this) == base_ + footprint_) { in AllocPages() 208 DCHECK_EQ(last_free_page_run->End(this), base_ + new_footprint); in AllocPages() 211 FreePageRun* new_free_page_run = reinterpret_cast<FreePageRun*>(base_ + footprint_); in AllocPages() [all …]
|
D | rosalloc.h | 84 …return reinterpret_cast<uint8_t*>(this) + ByteSize(rosalloc) == rosalloc->base_ + rosalloc->footpr… in IsAtEndOfSpace() 608 DCHECK_LE(base_, addr); in ToPageMapIndex() 609 DCHECK_LT(addr, base_ + capacity_); in ToPageMapIndex() 610 size_t byte_offset = reinterpret_cast<const uint8_t*>(addr) - base_; in ToPageMapIndex() 616 DCHECK(base_ <= addr && addr < reinterpret_cast<uint8_t*>(base_) + capacity_); in RoundDownToPageMapIndex() 617 return DivideByPageSize(reinterpret_cast<uintptr_t>(addr) - reinterpret_cast<uintptr_t>(base_)); in RoundDownToPageMapIndex() 695 uint8_t* base_; 773 uint8_t* Begin() { return base_; } in Begin() 775 uint8_t* End() { return base_ + capacity_; } in End()
|