/art/runtime/arch/x86/ |
D | memcmp16_x86.S | 57 movl BLK2(%esp), %edx 61 add %ecx, %edx 79 movdqu (%edx), %xmm0 81 movl %edx, %esi 83 pmovmskb %xmm3, %edx 86 sub $0xffff, %edx 89 mov %edi, %edx 90 and $0xf, %edx 91 xor %edx, %edi 92 sub %edx, %esi [all …]
|
D | quick_entrypoints_x86.S | 97 POP edx 115 PUSH edx 180 POP edx 327 movl %esp, %edx // remember SP 330 PUSH edx // pass SP 336 movl %edx, %edi // save code pointer in EDI 349 POP edx 491 movl (%edi), %edx 501 movl (%edi), %edx 515 mov %edx, 4(%ecx) // store the other half of the result [all …]
|
D | jni_entrypoints_x86.S | 28 PUSH_ARG edx 47 POP_ARG edx 68 PUSH_ARG edx 85 POP_ARG edx 147 movl (%esp), %edx 148 CFI_REGISTER(%eip, %edx) 182 movl %edx, FRAME_SIZE_SAVE_REFS_AND_ARGS - __SIZEOF_POINTER__(%eax) 211 movl %edx, __SIZEOF_POINTER__(%ebx) 228 movl __SIZEOF_POINTER__(%ebx), %edx 229 CFI_REGISTER(%eip, %edx) [all …]
|
D | asm_support_x86.S | 342 PUSH_ARG edx 370 POP_ARG edx
|
/art/runtime/interpreter/mterp/x86ng/ |
D | main.S | 41 #define ARG2 %edx 57 #define rIBASE %edx 219 PUSH_ARG edx 225 POP_ARG edx 235 movl %edx, LOCAL2(%esp) 272 PUSH_ARG edx 333 movzwl COMPACT_CODE_ITEM_FIELDS_OFFSET(\code_item), %edx 334 movl %edx, \registers 337 movl %edx, \outs 341 movl %edx, \ins [all …]
|
D | invoke.S | 88 movl MIRROR_OBJECT_CLASS_OFFSET(%ecx), %edx 89 UNPOISON_HEAP_REF edx 99 movl MIRROR_CLASS_IMT_PTR_OFFSET_32(%edx), %edx 100 movl (%edx, %eax, 4), %eax 166 movl MIRROR_OBJECT_CLASS_OFFSET(%ecx), %edx 167 UNPOISON_HEAP_REF edx 168 movl MIRROR_CLASS_VTABLE_OFFSET_32(%edx, %eax, 4), %eax
|
D | arithmetic.S | 14 movl %eax, %edx 15 orl %ecx, %edx 16 testl $$0xFFFFFF00, %edx # If both arguments are less 19 testl $$0xFFFF0000, %edx # If both arguments are less 39 movl %eax, %edx 40 shr $$8, %edx 46 xorl %edx, %edx # Clear %edx before divide 440 mov rIBASE, LOCAL0(%esp) # save rIBASE/%edx 445 movl %ecx, %edx 455 mov LOCAL0(%esp), rIBASE # restore rIBASE/%edx [all …]
|
D | control_flow.S | 167 GET_VREG_HIGH %edx, rINST # edx <- vAA
|
D | array.S | 108 GET_VREG %edx, rINST 110 RESTORE_IBASE # edx got overwritten, restore it
|
/art/runtime/interpreter/mterp/x86_64ng/ |
D | other.S | 127 GET_VREG %edx, rINSTq 129 SET_VREG_OBJECT %edx, %rax # fp[A] <- fp[B] 131 SET_VREG %edx, %rax # fp[A] <- fp[B] 140 GET_VREG %edx, %rcx 142 SET_VREG_OBJECT %edx, %rax # fp[A] <- fp[B] 144 SET_VREG %edx, %rax # fp[A] <- fp[B] 159 GET_VREG %edx, %rax # edx <- fp[BBBB] 161 SET_VREG_OBJECT %edx, rINSTq # fp[A] <- fp[B] 163 SET_VREG %edx, rINSTq # fp[A] <- fp[B]
|
D | invoke.S | 86 movl MIRROR_OBJECT_CLASS_OFFSET(%esi), %edx 87 UNPOISON_HEAP_REF edx 95 movq MIRROR_CLASS_IMT_PTR_OFFSET_64(%edx), %rdx 161 movl MIRROR_OBJECT_CLASS_OFFSET(%esi), %edx 162 UNPOISON_HEAP_REF edx 163 movq MIRROR_CLASS_VTABLE_OFFSET_64(%edx, %edi, 8), %rdi
|
D | main.S | 55 #define OUT_32_ARG2 %edx 930 GET_VREG_OBJECT %edx, %rdi 931 movl %edx, (rNEW_FP, %r10, 4) 932 GET_VREG %edx, %rdi 933 movl %edx, (%r11, %r10, 4) 938 GET_VREG_OBJECT %edx, %rax 939 movl %edx, (rNEW_FP, %r10, 4) 940 GET_VREG %edx, %rax 941 movl %edx, (%r11, %r10, 4) 947 GET_VREG_OBJECT %edx, %rax [all …]
|
D | array.S | 107 GET_VREG %edx, rINSTq
|
D | arithmetic.S | 411 movzbq 2(rPC), %rdx # edx <- BB
|
/art/runtime/arch/x86_64/ |
D | quick_entrypoints_x86_64.S | 445 movl %edx, %r10d 446 addl LITERAL(100), %edx // Reserve space for return addr, StackReference<method>, rbp, 448 andl LITERAL(0xFFFFFFF0), %edx // Align frame size to 16 bytes. 449 subl LITERAL(72), %edx // Remove space for return address, rbp, r8, r9, rbx, r12, 469 LOOP_OVER_SHORTY_LOADING_GPRS rdx, edx, .Lgpr_setup_finished 539 movl %edx, %r10d 540 addl LITERAL(100), %edx // Reserve space for return addr, StackReference<method>, rbp, 542 andl LITERAL(0xFFFFFFF0), %edx // Align frame size to 16 bytes. 543 subl LITERAL(72), %edx // Remove space for return address, rbp, r8, r9, rbx, r12, 562 LOOP_OVER_SHORTY_LOADING_GPRS rdx, edx, .Lgpr_setup_finished2 [all …]
|
/art/runtime/interpreter/mterp/ |
D | README.txt | 63 (r8 for ARM, edx for x86) holds a mainHandlerTable. If we need to switch
|
/art/compiler/jni/ |
D | jni_cfi_test_expected.inc | 228 // 0x00000011: mov [esp + 44], edx 303 // 0x00000038: mov [rsp + 112], edx
|
/art/compiler/optimizing/ |
D | code_generator_x86.cc | 3898 Register edx = locations->GetTemp(1).AsRegister<Register>(); in VisitMul() local 3901 DCHECK_EQ(EDX, edx); in VisitMul() 3929 __ addl(in1_hi, edx); in VisitMul() 3948 __ addl(in1_hi, edx); in VisitMul() 3968 __ addl(in1_hi, edx); in VisitMul() 4193 Register edx; in GenerateDivRemWithAnyConstant() local 4196 edx = locations->GetTemp(0).AsRegister<Register>(); in GenerateDivRemWithAnyConstant() 4199 edx = locations->Out().AsRegister<Register>(); in GenerateDivRemWithAnyConstant() 4204 DCHECK_EQ(EDX, edx); in GenerateDivRemWithAnyConstant() 4226 __ addl(edx, num); in GenerateDivRemWithAnyConstant() [all …]
|
D | code_generator_x86_64.cc | 4364 CpuRegister edx = instruction->IsDiv() ? locations->GetTemp(0).AsRegister<CpuRegister>() in GenerateDivRemWithAnyConstant() local 4369 DCHECK_EQ(RDX, edx.AsRegister()); in GenerateDivRemWithAnyConstant() 4391 __ addl(edx, numerator); in GenerateDivRemWithAnyConstant() 4393 __ subl(edx, numerator); in GenerateDivRemWithAnyConstant() 4397 __ sarl(edx, Immediate(shift)); in GenerateDivRemWithAnyConstant() 4400 __ movl(eax, edx); in GenerateDivRemWithAnyConstant() 4401 __ shrl(edx, Immediate(31)); in GenerateDivRemWithAnyConstant() 4402 __ addl(edx, eax); in GenerateDivRemWithAnyConstant() 4406 __ imull(edx, Immediate(imm)); in GenerateDivRemWithAnyConstant() 4407 __ subl(eax, edx); in GenerateDivRemWithAnyConstant() [all …]
|
D | intrinsics_x86.cc | 3651 Register edx = locations->GetTemp(0).AsRegister<Register>(); in VisitIntegerDivideUnsigned() local 3656 DCHECK_EQ(EDX, edx); in VisitIntegerDivideUnsigned() 3664 __ xorl(edx, edx); in VisitIntegerDivideUnsigned()
|
/art/runtime/arch/ |
D | stub_test.cc | 104 PUSH(%%edx) in Invoke3WithReferrerAndHidden() 119 POP(%%edx) in Invoke3WithReferrerAndHidden() 132 POP(%%edx) in Invoke3WithReferrerAndHidden()
|