/art/runtime/interpreter/mterp/x86ng/ |
D | invoke.S | 24 movl (rFP, %ecx, 4), %ecx 30 movl rSELF:THREAD_SELF_OFFSET, ARG0 31 movl 0(%esp), ARG1 32 movl rPC, ARG2 59 movl (rFP, %ecx, 4), %ecx 70 movl (rFP, %ecx, 4), %ecx 87 movl (rFP, %ecx, 4), %ecx 88 movl MIRROR_OBJECT_CLASS_OFFSET(%ecx), %edx 99 movl MIRROR_CLASS_IMT_PTR_OFFSET_32(%edx), %edx 100 movl (%edx, %eax, 4), %eax [all …]
|
D | main.S | 79 movl VREG_ADDRESS(\_vreg), \_reg 83 movl VREG_REF_ADDRESS(\_vreg), \_reg 92 movl \_reg, VREG_ADDRESS(\_vreg) 93 movl MACRO_LITERAL(0), VREG_REF_ADDRESS(\_vreg) 104 movl \_reg, VREG_ADDRESS(\_vreg) 105 movl \_reg, VREG_REF_ADDRESS(\_vreg) 109 movl VREG_HIGH_ADDRESS(\_vreg), \_reg 113 movl \_reg, VREG_HIGH_ADDRESS(\_vreg) 114 movl MACRO_LITERAL(0), VREG_REF_HIGH_ADDRESS(\_vreg) 118 movl MACRO_LITERAL(0), VREG_REF_ADDRESS(\_vreg) [all …]
|
D | object.S | 17 movl rSELF:THREAD_SELF_OFFSET, ARG0 18 movl 0(%esp), ARG1 19 movl rPC, ARG2 21 movl %eax, %ecx 54 movl $$1, %eax 61 movl rSELF:THREAD_SELF_OFFSET, ARG0 62 movl 0(%esp), ARG1 63 movl rPC, ARG2 65 movl %eax, %ecx 76 movl MIRROR_OBJECT_CLASS_OFFSET(%eax), %eax [all …]
|
D | arithmetic.S | 14 movl %eax, %edx 26 movl $special, $result 39 movl %eax, %edx 71 movl $special, $result 103 movl $special, %eax 130 movl $special, %eax 240 movl rIBASE, LOCAL0(%esp) # save rIBASE 246 movl LOCAL0(%esp), rIBASE # restore rIBASE 295 movl $$0x80000000, %eax 323 movl $$0, VREG_ADDRESS(%ecx) [all …]
|
D | control_flow.S | 10 movl rINST, %ecx # rcx <- A+ 67 movl 2(rPC), rINST # rINST <- AAAAAAAA 117 movl 2(rPC), ARG0 # eax <- BBBBbbbb 126 movl %eax, rINST 140 movl -4(rREFS), %esp 154 movl -4(rREFS), %esp 169 movl -4(rREFS), %esp 183 movl rSELF:THREAD_SELF_OFFSET, ARG1
|
D | other.S | 6 movl 2(rPC), %eax # grab all 32 bits at once 45 movl rSELF:THREAD_SELF_OFFSET, ARG0 46 movl 0(%esp), ARG1 47 movl rPC, ARG2 74 movl 2(rPC), %eax # eax <- lsw 76 movl 6(rPC), rINST # rINST <- msw 84 movl rIBASE, %ecx # preserve rIBASE (cdq trashes it) 88 movl %ecx, rIBASE # restore rIBASE 93 movl 2(rPC), %eax # eax <- BBBBbbbb 94 movl rIBASE, %ecx # preserve rIBASE (cdq trashes it) [all …]
|
D | array.S | 117 movl rINST, %eax # eax <- BA 123 movl MIRROR_ARRAY_LENGTH_OFFSET(%ecx), rINST 130 movl 2(rPC), %ecx # ecx <- BBBBbbbb 146 movl rSELF:THREAD_SELF_OFFSET, ARG0 147 movl (%esp), ARG1 148 movl rFP, ARG2 149 movl rPC, ARG3
|
/art/runtime/arch/x86/ |
D | jni_entrypoints_x86.S | 99 movl (%esp), %eax // Thread* self 100 movl THREAD_TOP_QUICK_FRAME_OFFSET(%eax), %eax // uintptr_t tagged_quick_frame 102 movl (%eax), %eax // ArtMethod* method 132 movl (%esp), %ecx 147 movl (%esp), %edx 155 movl %eax, %ecx 182 movl %edx, FRAME_SIZE_SAVE_REFS_AND_ARGS - __SIZEOF_POINTER__(%eax) 186 movl RUNTIME_SAVE_REFS_AND_ARGS_METHOD_OFFSET(%ebx), %ebx 188 movl %eax, %ecx // Prepare untagged managed SP for the runtime method. 192 movl %ebx, (%eax) [all …]
|
D | quick_entrypoints_x86.S | 42 movl %esp, %fs:THREAD_TOP_QUICK_FRAME_OFFSET 44 movl 12(%esp), REG_VAR(temp_reg) 66 movl %esp, %fs:THREAD_TOP_QUICK_FRAME_OFFSET 79 movl %esp, %fs:THREAD_TOP_QUICK_FRAME_OFFSET 135 movl %esp, %fs:THREAD_TOP_QUICK_FRAME_OFFSET 270 movl (FRAME_SIZE_SAVE_EVERYTHING - 2 * __SIZEOF_POINTER__)(%esp), %eax 271 movl %edi, (FRAME_SIZE_SAVE_EVERYTHING - 2 * __SIZEOF_POINTER__)(%esp) 327 movl %esp, %edx // remember SP 336 movl %edx, %edi // save code pointer in EDI 471 movl LITERAL(0), (%esp) // store null for method* [all …]
|
D | asm_support_x86.S | 261 movl SYMBOL(_ZN3art7Runtime9instance_E) - 0b(REG_VAR(reg)), REG_VAR(reg) 293 movl %esp, %fs:THREAD_TOP_QUICK_FRAME_OFFSET 325 movl %esp, %fs:THREAD_TOP_QUICK_FRAME_OFFSET 411 movl MIRROR_OBJECT_LOCK_WORD_OFFSET(REG_VAR(obj)), %eax // EAX := lock word 412 movl %fs:THREAD_ID_OFFSET, REG_VAR(tmp) // tmp: thread id. 421 movl REG_VAR(saved_eax), %eax // Restore EAX. 438 movl REG_VAR(saved_eax), %eax // Restore EAX. 446 movl MIRROR_OBJECT_LOCK_WORD_OFFSET(REG_VAR(obj)), %eax // EAX := lock word 447 movl %fs:THREAD_ID_OFFSET, REG_VAR(tmp) // tmp := thread id 453 movl REG_VAR(tmp), MIRROR_OBJECT_LOCK_WORD_OFFSET(REG_VAR(obj)) [all …]
|
/art/runtime/interpreter/mterp/x86_64ng/ |
D | invoke.S | 24 movl (rFP, %r11, 4), %esi 26 movl (%esi), %eax 59 movl (rFP, %r11, 4), %esi 61 movl (%esi), %eax 69 movl (rFP, %r11, 4), %esi 71 movl (%esi), %eax 85 movl (rFP, %r11, 4), %esi 86 movl MIRROR_OBJECT_CLASS_OFFSET(%esi), %edx 159 movl (rFP, %r11, 4), %esi 161 movl MIRROR_OBJECT_CLASS_OFFSET(%esi), %edx [all …]
|
D | main.S | 145 movl VREG_ADDRESS(\_vreg), \_reg 149 movl VREG_REF_ADDRESS(\_vreg), \_reg 158 movl \_reg, VREG_ADDRESS(\_vreg) 159 movl MACRO_LITERAL(0), VREG_REF_ADDRESS(\_vreg) 170 movl \_reg, VREG_ADDRESS(\_vreg) 171 movl \_reg, VREG_REF_ADDRESS(\_vreg) 175 movl VREG_HIGH_ADDRESS(\_vreg), \_reg 179 movl \_reg, VREG_HIGH_ADDRESS(\_vreg) 180 movl MACRO_LITERAL(0), VREG_REF_HIGH_ADDRESS(\_vreg) 184 movl MACRO_LITERAL(0), VREG_REF_ADDRESS(\_vreg) [all …]
|
D | object.S | 28 movl MIRROR_OBJECT_CLASS_OFFSET(%edi), %eax 33 movl MIRROR_CLASS_SUPER_CLASS_OFFSET(%eax), %eax 51 movl MIRROR_CLASS_COMPONENT_TYPE_OFFSET(%eax), %eax 56 movl MIRROR_CLASS_COMPONENT_TYPE_OFFSET(%esi), %ecx 103 movl $$1, %edi 117 movl MIRROR_OBJECT_CLASS_OFFSET(%edi), %eax 122 movl MIRROR_CLASS_SUPER_CLASS_OFFSET(%eax), %eax 132 movl MIRROR_CLASS_COMPONENT_TYPE_OFFSET(%eax), %eax 137 movl MIRROR_CLASS_COMPONENT_TYPE_OFFSET(%esi), %ecx
|
D | arithmetic.S | 66 movl rINST, %ecx # rcx <- BA 128 movl rINST, %eax # rax <- 000000BA 226 movl rINST, %ecx # rcx <- A+ 245 movl rINST, %eax # rax <- 000000BA 290 movl rINST, %ecx # rcx <- A+ 305 movl rINST, %ecx # rcx <- A+ 331 movl rINST, %ecx # ecx <- BA 352 movl rINST, %ecx # rcx <- A+ 417 movl $$-1, %esi 470 movl rINST, %ecx # rcx <- A+ [all …]
|
D | other.S | 6 movl 2(rPC), %eax # grab all 32 bits at once 124 movl rINST, %eax # eax <- BA 150 movl rSELF:THREAD_EXCEPTION_OFFSET, %eax 152 movl $$0, rSELF:THREAD_EXCEPTION_OFFSET 197 movl rINST, %ecx # ecx <- BA
|
D | floating_point.S | 32 movl $$-1, %eax 42 movl rINST, %ecx # rcx <- A+ 73 movl rINST, %ecx # ecx <- A+
|
/art/compiler/utils/x86/ |
D | jni_macro_assembler_x86.cc | 147 __ movl(Address(base.AsCpuRegister(), offs), src.AsCpuRegister()); in Store() local 150 __ movl(Address(base.AsCpuRegister(), offs), src.AsRegisterPairLow()); in Store() local 151 __ movl(Address(base.AsCpuRegister(), FrameOffset(offs.Int32Value()+4)), in Store() local 172 __ movl(Address(ESP, dest), src.AsCpuRegister()); in StoreRawPtr() local 179 __ movl(Address(ESP, -32), scratch); in StoreStackPointerToThread() local 180 __ movl(scratch, ESP); in StoreStackPointerToThread() local 182 __ fs()->movl(Address::Absolute(thr_offs), scratch); in StoreStackPointerToThread() 183 __ movl(scratch, Address(ESP, -32)); in StoreStackPointerToThread() local 185 __ fs()->movl(Address::Absolute(thr_offs), ESP); in StoreStackPointerToThread() 203 __ movl(dest.AsCpuRegister(), Address(base.AsCpuRegister(), offs)); in Load() local [all …]
|
/art/runtime/arch/x86_64/ |
D | quick_entrypoints_x86_64.S | 389 movl (%r11), REG_VAR(gpr_reg32) 445 movl %edx, %r10d 458 movl %r10d, %ecx // Place size of args in rcx. 467 movl (%r11), %esi // rsi := this pointer 539 movl %edx, %r10d 552 movl %r10d, %ecx // Place size of args in rcx. 696 movl %eax, %edi // pass the index of the constant as arg0 821 movl MIRROR_CLASS_OBJECT_SIZE_ALLOC_FAST_PATH_OFFSET(%rdi), %eax 846 movl %eax, (%rcx) 862 movl %edi, MIRROR_OBJECT_CLASS_OFFSET(%rax) [all …]
|
D | asm_support_x86_64.S | 494 movl MIRROR_OBJECT_LOCK_WORD_OFFSET(REG_VAR(obj)), %eax // EAX := lock word 495 movl %gs:THREAD_ID_OFFSET, REG_VAR(tmp) // tmp: thread id. 522 movl MIRROR_OBJECT_LOCK_WORD_OFFSET(REG_VAR(obj)), %eax // EAX := lock word 523 movl %gs:THREAD_ID_OFFSET, REG_VAR(tmp) // tmp := thread id 529 movl REG_VAR(tmp), MIRROR_OBJECT_LOCK_WORD_OFFSET(REG_VAR(obj)) 548 movl REG_VAR(tmp), MIRROR_OBJECT_LOCK_WORD_OFFSET(REG_VAR(obj))
|
/art/compiler/optimizing/ |
D | code_generator_x86.cc | 136 __ movl(reg_, Immediate(0)); in EmitNativeCode() local 177 __ movl(length_arg.AsRegister<Register>(), in EmitNativeCode() local 183 __ movl(length_arg.AsRegister<Register>(), in EmitNativeCode() local 195 __ movl(length_arg.AsRegister<Register>(), in EmitNativeCode() local 281 __ movl(calling_convention.GetRegisterAt(0), Immediate(string_index.index_)); in EmitNativeCode() local 322 __ movl(calling_convention.GetRegisterAt(0), Immediate(type_index.index_)); in EmitNativeCode() local 621 __ movl(temp_, ref_reg); in EmitNativeCode() local 668 __ movl(EAX, temp_); in EmitNativeCode() local 682 __ movl(value, base); in EmitNativeCode() local 818 __ movl(free_reg, index_reg); in EmitNativeCode() local [all …]
|
D | intrinsics_x86.cc | 123 __ movl(value, Address(src_curr_addr, 0)); in EmitNativeCode() local 136 __ movl(Address(dst_curr_addr, 0), value); in EmitNativeCode() local 294 __ movl(output_lo, input_hi); in VisitLongReverseBytes() local 295 __ movl(output_hi, input_lo); in VisitLongReverseBytes() local 436 __ movl(out, Immediate(kPrimIntMax)); in VisitMathRoundFloat() local 440 __ movl(out, Immediate(0)); // does not change flags in VisitMathRoundFloat() local 541 __ movl(out_lo, src_lo); in GenLowestOneBit() local 542 __ movl(out_hi, src_hi); in GenLowestOneBit() local 558 __ movl(out, src.AsRegister<Register>()); in GenLowestOneBit() local 561 __ movl(out, Address(ESP, src.GetStackIndex())); in GenLowestOneBit() local [all …]
|
D | code_generator_x86_64.cc | 228 __ movl(length_arg.AsRegister<CpuRegister>(), array_len); in EmitNativeCode() local 233 __ movl(length_arg.AsRegister<CpuRegister>(), array_len); in EmitNativeCode() local 237 __ movl(CpuRegister(TMP), array_len); in EmitNativeCode() local 240 __ movl(length_arg.AsRegister<CpuRegister>(), CpuRegister(TMP)); in EmitNativeCode() local 287 __ movl(CpuRegister(RAX), Immediate(proto_index.index_)); in EmitNativeCode() local 331 __ movl(CpuRegister(RAX), Immediate(type_index.index_)); in EmitNativeCode() local 383 __ movl(CpuRegister(RAX), Immediate(string_index.index_)); in EmitNativeCode() local 667 __ movl(temp1_, ref_cpu_reg); in EmitNativeCode() local 715 __ movl(CpuRegister(RAX), temp1_); in EmitNativeCode() local 730 __ movl(CpuRegister(value_reg), base); in EmitNativeCode() local [all …]
|
D | intrinsics_x86_64.cc | 119 __ movl(CpuRegister(TMP), Address(src_curr_addr, 0)); in EmitNativeCode() local 130 __ movl(Address(dst_curr_addr, 0), CpuRegister(TMP)); in EmitNativeCode() local 263 __ movl(output, Immediate(1)); in GenIsInfinite() local 269 __ movl(output, Immediate(1)); in GenIsInfinite() local 403 __ movl(out, Immediate(0)); // does not change flags in VisitMathRoundFloat() local 444 __ movl(out, Immediate(0)); // does not change flags, implicit zero extension to 64-bit in VisitMathRoundDouble() local 710 __ movl(temp, Address(array, length_offset)); in CheckSystemArrayCopyPosition() local 732 __ movl(temp, Address(array, length_offset)); in CheckSystemArrayCopyPosition() local 803 __ movl(count, Immediate(length.GetConstant()->AsIntConstant()->GetValue())); in SystemArrayCopyPrimitive() local 805 __ movl(count, length.AsRegister<CpuRegister>()); in SystemArrayCopyPrimitive() local [all …]
|
/art/test/521-regression-integer-field-set/ |
D | info.txt | 3 a `movw` instruction instead of a `movl` instruction.
|
/art/compiler/utils/x86_64/ |
D | jni_macro_assembler_x86_64.cc | 177 __ movl(Address(base.AsCpuRegister(), offs), src.AsCpuRegister()); in Store() local 230 __ movl(dest.AsCpuRegister(), Address(base.AsCpuRegister(), offs)); in Load() local 417 __ movl(scratch, Address(CpuRegister(RSP), src)); in Copy() local 418 __ movl(Address(CpuRegister(RSP), dest), scratch); in Copy() local 432 __ movl(in_reg.AsCpuRegister(), Address(CpuRegister(RSP), spilled_reference_offset)); in CreateJObject() local 457 __ movl(scratch, Address(CpuRegister(RSP), spilled_reference_offset)); in CreateJObject() local 478 __ movl(reg.AsX86_64().AsCpuRegister(), Address(reg.AsX86_64().AsCpuRegister(), /*disp=*/ 0)); in DecodeJNITransitionOrLocalJObject() local 531 __ movl(scratch, Immediate(kNativeStateValue)); in TryToTransitionFromRunnableToNative() local 568 __ movl(rax, Immediate(kNativeStateValue)); in TryToTransitionFromNativeToRunnable() local
|