/art/compiler/optimizing/ |
D | code_generator_arm_vixl.cc | 517 __ Mov(calling_convention.GetRegisterAt(0), type_index.index_); in EmitNativeCode() local 571 __ Mov(calling_convention.GetRegisterAt(0), string_index.index_); in EmitNativeCode() local 657 __ Mov(calling_convention.GetRegisterAt(0), in EmitNativeCode() local 797 __ Mov(free_reg, index_reg); in EmitNativeCode() local 855 __ Mov(calling_convention.GetRegisterAt(2), offset_); in EmitNativeCode() local 959 __ Mov(vixl::aarch32::Register(R2), arm_codegen->GetFrameSize()); in EmitNativeCode() local 988 __ Mov(tmp, ProfilingInfo::GetOptimizeThreshold()); in EmitNativeCode() local 1132 __ Mov(out, in); in GenerateDataProcInstruction() local 1433 __ Mov(LeaveFlags, operand_low.GetBaseRegister(), value_low); in GenerateLongTestConstant() local 1588 __ Mov(LeaveFlags, out, 0); in GenerateConditionGeneric() local [all …]
|
D | code_generator_arm64.cc | 296 __ Mov(calling_convention.GetRegisterAt(0).W(), type_index.index_); in EmitNativeCode() local 351 __ Mov(calling_convention.GetRegisterAt(0).W(), string_index.index_); in EmitNativeCode() local 501 __ Mov(calling_convention.GetRegisterAt(0), in EmitNativeCode() local 664 __ Mov(free_reg.W(), index_reg); in EmitNativeCode() local 800 __ Mov(calling_convention.GetRegisterAt(0), XRegisterFrom(out_)); in EmitNativeCode() local 834 __ Mov(vixl::aarch64::x4, arm64_codegen->GetFrameSize()); in EmitNativeCode() local 862 __ Mov(counter, ProfilingInfo::GetOptimizeThreshold()); in EmitNativeCode() local 1236 __ Mov(addr, address + offset.Int32Value()); in GenerateMethodEntryExitHook() local 1505 __ Mov(RegisterFrom(location, DataType::Type::kInt32), value); in MoveConstant() local 1643 __ Mov(Register(destination), constant->AsIntConstant()->GetValue()); in MoveConstant() local [all …]
|
D | intrinsics_arm_vixl.cc | 677 __ Mov(temp1, value_offset); in GenerateStringCompareToLoop() local 781 __ Mov(temp1, str); in GenerateStringCompareToLoop() local 782 __ Mov(temp2, arg); in GenerateStringCompareToLoop() local 1016 __ Mov(temp1, value_offset); in VisitStringEquals() local 1033 __ Mov(out, 1); in VisitStringEquals() local 1038 __ Mov(out, 0); in VisitStringEquals() local 1082 __ Mov(tmp_reg, 0); in GenerateVisitStringIndexOf() local 2001 __ Mov(temp, 0x80000000); // Modified immediate. in GenHighestOneBit() local 2014 __ Mov(LeaveFlags, temp, 0); in GenHighestOneBit() local 2026 __ Mov(temp, 0x80000000); // Modified immediate. in GenHighestOneBit() local [all …]
|
D | code_generator_vector_arm64_sve.cc | 915 __ Mov(dst.V16B(), 0, InputRegisterAt(instruction, 0)); in VisitVecSetScalars() local 919 __ Mov(dst.V8H(), 0, InputRegisterAt(instruction, 0)); in VisitVecSetScalars() local 922 __ Mov(dst.V4S(), 0, InputRegisterAt(instruction, 0)); in VisitVecSetScalars() local 925 __ Mov(dst.V2D(), 0, InputRegisterAt(instruction, 0)); in VisitVecSetScalars() local 1361 __ Mov(ZRegisterFrom(destination), ZRegisterFrom(source)); in MoveSIMDRegToSIMDReg() local
|
D | code_generator_vector_arm64_neon.cc | 1045 __ Mov(dst.V16B(), 0, InputRegisterAt(instruction, 0)); in VisitVecSetScalars() local 1050 __ Mov(dst.V8H(), 0, InputRegisterAt(instruction, 0)); in VisitVecSetScalars() local 1054 __ Mov(dst.V4S(), 0, InputRegisterAt(instruction, 0)); in VisitVecSetScalars() local 1058 __ Mov(dst.V2D(), 0, InputRegisterAt(instruction, 0)); in VisitVecSetScalars() local 1585 __ Mov(QRegisterFrom(destination), QRegisterFrom(source)); in MoveSIMDRegToSIMDReg() local
|
D | intrinsics_arm64.cc | 445 __ Mov(dst, UINT64_C(1) << high_bit); // MOV (bitmask immediate) in GenHighestOneBit() local 976 __ Mov(temp.W(), value.W()); in GenUnsafePut() local 1407 __ Mov(old_value_, expected_); in EmitNativeCode() local 1919 __ Mov(temp1, value_offset); in VisitStringCompareTo() local 2222 __ Mov(temp1, value_offset); in VisitStringEquals() local 2241 __ Mov(out, 1); in VisitStringEquals() local 2246 __ Mov(out, 0); in VisitStringEquals() local 2284 __ Mov(tmp_reg, 0); in GenerateVisitStringIndexOf() local 2671 __ Mov(num_chr, tmp1); in VisitStringGetCharsNoCheck() local 2706 __ Mov(num_chr, tmp1); in VisitStringGetCharsNoCheck() local [all …]
|
/art/compiler/utils/arm64/ |
D | jni_macro_assembler_arm64.cc | 68 ___ Mov(reg_x(dest.AsArm64().AsXRegister()), reg_x(TR)); in GetCurrentThread() local 206 ___ Mov(scratch, reg_x(SP)); in StoreStackPointerToThread() local 216 ___ Mov(reg_x(dest), value); in LoadImmediate() local 224 ___ Mov(temp, value); in LoadImmediate() local 561 ___ Mov(reg_w(dst.AsOverlappingWRegister()), reg_w(src.AsWRegister())); in Move() local 564 ___ Mov(reg_x(dst.AsXRegister()), reg_x(src.AsXRegister())); in Move() local 566 ___ Mov(reg_x(dst.AsXRegister()), reg_x(src.AsOverlappingXRegister())); in Move() local 571 ___ Mov(reg_w(dst.AsWRegister()), reg_w(src.AsWRegister())); in Move() local 586 ___ Mov(reg_x(dst.AsXRegister()), value); in Move() local 726 ___ Mov(scratch2, kNativeStateValue); in TryToTransitionFromRunnableToNative() local [all …]
|
/art/compiler/utils/arm/ |
D | jni_macro_assembler_arm_vixl.cc | 735 ___ Mov(AsVIXLRegister(dst), AsVIXLRegister(src)); in Move() local 761 ___ Mov(AsVIXLRegisterPairLow(dst), AsVIXLRegisterPairLow(src)); in Move() local 762 ___ Mov(AsVIXLRegisterPairHigh(dst), AsVIXLRegisterPairHigh(src)); in Move() local 764 ___ Mov(AsVIXLRegisterPairHigh(dst), AsVIXLRegisterPairHigh(src)); in Move() local 765 ___ Mov(AsVIXLRegisterPairLow(dst), AsVIXLRegisterPairLow(src)); in Move() local 777 ___ Mov(AsVIXLRegister(dst), static_cast<uint32_t>(value)); in Move() local 890 ___ Mov(AsVIXLRegister(dest.AsArm()), tr); in GetCurrentThread() local 913 ___ Mov(scratch2, kNativeStateValue); in TryToTransitionFromRunnableToNative() local 1064 ___ Mov(scratch, static_cast<uint32_t>(address)); in TestByteAndJumpIfNotZero() local
|
D | assembler_arm_vixl.cc | 102 ___ Mov(rd, value); in LoadImmediate() local 155 ___ Mov(temp, offset); in AdjustLoadStoreOffset() local 423 ___ Mov(rd, rn); in AddConstant() local
|