/art/compiler/optimizing/ |
D | instruction_simplifier_x86_64.cc | 38 return codegen_->GetInstructionSetFeatures().HasAVX2(); in HasAVX2()
|
D | instruction_simplifier_x86.cc | 38 return (codegen_->GetInstructionSetFeatures().HasAVX2()); in HasAVX2()
|
D | intrinsics_arm64.cc | 3677 if (!codegen_->GetInstructionSetFeatures().HasCRC()) { in VisitCRC32Update() 3692 DCHECK(codegen_->GetInstructionSetFeatures().HasCRC()); in VisitCRC32Update() 3823 if (!codegen_->GetInstructionSetFeatures().HasCRC()) { in VisitCRC32UpdateBytes() 3844 DCHECK(codegen_->GetInstructionSetFeatures().HasCRC()); in VisitCRC32UpdateBytes() 3879 if (!codegen_->GetInstructionSetFeatures().HasCRC()) { in VisitCRC32UpdateByteBuffer() 3906 DCHECK(codegen_->GetInstructionSetFeatures().HasCRC()); in VisitCRC32UpdateByteBuffer() 3922 if (!codegen_->GetInstructionSetFeatures().HasFP16()) { in VisitFP16ToFloat() 3934 DCHECK(codegen_->GetInstructionSetFeatures().HasFP16()); in VisitFP16ToFloat() 3945 if (!codegen_->GetInstructionSetFeatures().HasFP16()) { in VisitFP16ToHalf() 3957 DCHECK(codegen_->GetInstructionSetFeatures().HasFP16()); in VisitFP16ToHalf() [all …]
|
D | scheduler_arm.cc | 1031 bool atomic_ldrd_strd = codegen_->GetInstructionSetFeatures().HasAtomicLdrdAndStrd(); in HandleFieldGetLatencies() 1091 bool atomic_ldrd_strd = codegen_->GetInstructionSetFeatures().HasAtomicLdrdAndStrd(); in HandleFieldSetLatencies()
|
D | optimizing_compiler.cc | 415 const InstructionSetFeatures* features = compiler_options.GetInstructionSetFeatures(); in DumpInstructionSetFeaturesToCfg() 1508 const InstructionSetFeatures* features = compiler_options.GetInstructionSetFeatures(); in GenerateJitDebugInfo()
|
D | code_generator_x86_64.h | 474 const X86_64InstructionSetFeatures& GetInstructionSetFeatures() const;
|
D | intrinsics_x86.cc | 331 if (!codegen->GetInstructionSetFeatures().HasSSE4_1()) { in CreateSSE41FPToFPLocations() 372 if (!codegen_->GetInstructionSetFeatures().HasSSE4_1()) { in VisitMathRoundFloat() 551 if (codegen->GetInstructionSetFeatures().HasAVX2() && src.IsRegister()) { in GenLowestOneBit() 2768 if (!codegen->GetInstructionSetFeatures().HasPopCnt()) { in CreateBitCountLocations() 4943 DCHECK(codegen_->GetInstructionSetFeatures().HasAVX2()); in VisitMathFmaDouble() 4948 if (codegen_->GetInstructionSetFeatures().HasAVX2()) { in VisitMathFmaDouble() 4954 DCHECK(codegen_->GetInstructionSetFeatures().HasAVX2()); in VisitMathFmaFloat() 4959 if (codegen_->GetInstructionSetFeatures().HasAVX2()) { in VisitMathFmaFloat()
|
D | intrinsics_x86_64.cc | 313 if (!codegen->GetInstructionSetFeatures().HasSSE4_1()) { in CreateSSE41FPToFPLocations() 356 if (!codegen->GetInstructionSetFeatures().HasSSE4_1()) { in CreateSSE41FPToIntLocations() 2847 if (!codegen->GetInstructionSetFeatures().HasPopCnt()) { in CreateBitCountLocations() 2948 if (!is_high && codegen->GetInstructionSetFeatures().HasAVX2() && in GenOneBit() 3583 DCHECK(codegen_->GetInstructionSetFeatures().HasAVX2()); in VisitMathFmaDouble() 3588 if (codegen_->GetInstructionSetFeatures().HasAVX2()) { in VisitMathFmaDouble() 3594 DCHECK(codegen_->GetInstructionSetFeatures().HasAVX2()); in VisitMathFmaFloat() 3599 if (codegen_->GetInstructionSetFeatures().HasAVX2()) { in VisitMathFmaFloat()
|
D | code_generator_x86.h | 492 const X86InstructionSetFeatures& GetInstructionSetFeatures() const;
|
D | code_generator_riscv64.h | 498 const Riscv64InstructionSetFeatures& GetInstructionSetFeatures() const;
|
D | code_generator_arm_vixl.cc | 1082 const ArmInstructionSetFeatures& CodeGeneratorARMVIXL::GetInstructionSetFeatures() const { in GetInstructionSetFeatures() function in art::arm::CodeGeneratorARMVIXL 1083 return *GetCompilerOptions().GetInstructionSetFeatures()->AsArmInstructionSetFeatures(); in GetInstructionSetFeatures() 4697 !codegen_->GetInstructionSetFeatures().HasDivideInstruction()) { in VisitDiv() 4724 } else if (codegen_->GetInstructionSetFeatures().HasDivideInstruction()) { in VisitDiv() 4768 } else if (codegen_->GetInstructionSetFeatures().HasDivideInstruction()) { in VisitDiv() 4815 && codegen_->GetInstructionSetFeatures().HasDivideInstruction()) { in VisitRem() 4839 } else if (codegen_->GetInstructionSetFeatures().HasDivideInstruction()) { in VisitRem() 4898 } else if (codegen_->GetInstructionSetFeatures().HasDivideInstruction()) { in VisitRem() 5942 && !codegen_->GetInstructionSetFeatures().HasAtomicLdrdAndStrd(); in HandleFieldSet() 5984 bool atomic_ldrd_strd = codegen_->GetInstructionSetFeatures().HasAtomicLdrdAndStrd(); in HandleFieldSet() [all …]
|
D | code_generator_x86_64.cc | 1514 const X86_64InstructionSetFeatures& CodeGeneratorX86_64::GetInstructionSetFeatures() const { in GetInstructionSetFeatures() function in art::x86_64::CodeGeneratorX86_64 1515 return *GetCompilerOptions().GetInstructionSetFeatures()->AsX86_64InstructionSetFeatures(); in GetInstructionSetFeatures() 1615 compiler_options.GetInstructionSetFeatures()->AsX86_64InstructionSetFeatures()), in CodeGeneratorX86_64() 7685 DCHECK(codegen_->GetInstructionSetFeatures().HasAVX2()); in VisitX86AndNot() 7695 DCHECK(codegen_->GetInstructionSetFeatures().HasAVX2()); in VisitX86MaskOrResetLeastSetBit() 8549 return codegen_->GetInstructionSetFeatures().HasAVX(); in CpuHasAvxFeatureFlag() 8553 return codegen_->GetInstructionSetFeatures().HasAVX2(); in CpuHasAvx2FeatureFlag() 8557 return codegen_->GetInstructionSetFeatures().HasAVX(); in CpuHasAvxFeatureFlag() 8561 return codegen_->GetInstructionSetFeatures().HasAVX2(); in CpuHasAvx2FeatureFlag()
|
D | intrinsics_arm_vixl.cc | 153 features_(codegen->GetInstructionSetFeatures()) {} in IntrinsicLocationsBuilderARMVIXL() 361 DCHECK(codegen_->GetInstructionSetFeatures().HasARMv8AInstructions()); in VisitMathRint() 377 DCHECK(codegen_->GetInstructionSetFeatures().HasARMv8AInstructions()); in VisitMathRoundFloat() 2292 DCHECK(codegen_->GetInstructionSetFeatures().HasARMv8AInstructions()); in VisitMathCeil() 2304 DCHECK(codegen_->GetInstructionSetFeatures().HasARMv8AInstructions()); in VisitMathFloor() 2568 return atomic && !codegen->GetInstructionSetFeatures().HasAtomicLdrdAndStrd(); in Use64BitExclusiveLoadStore()
|
D | code_generator_arm_vixl.h | 592 const ArmInstructionSetFeatures& GetInstructionSetFeatures() const;
|
D | code_generator_x86.cc | 1074 const X86InstructionSetFeatures& CodeGeneratorX86::GetInstructionSetFeatures() const { in GetInstructionSetFeatures() function in art::x86::CodeGeneratorX86 1075 return *GetCompilerOptions().GetInstructionSetFeatures()->AsX86InstructionSetFeatures(); in GetInstructionSetFeatures() 1172 compiler_options.GetInstructionSetFeatures()->AsX86InstructionSetFeatures()), in CodeGeneratorX86() 8334 DCHECK(codegen_->GetInstructionSetFeatures().HasAVX2()); in VisitX86AndNot() 8363 DCHECK(codegen_->GetInstructionSetFeatures().HasAVX2()); in VisitX86MaskOrResetLeastSetBit() 9293 return codegen_->GetInstructionSetFeatures().HasAVX(); in CpuHasAvxFeatureFlag() 9296 return codegen_->GetInstructionSetFeatures().HasAVX2(); in CpuHasAvx2FeatureFlag() 9299 return codegen_->GetInstructionSetFeatures().HasAVX(); in CpuHasAvxFeatureFlag() 9302 return codegen_->GetInstructionSetFeatures().HasAVX2(); in CpuHasAvx2FeatureFlag()
|
D | code_generator_arm64.h | 706 const Arm64InstructionSetFeatures& GetInstructionSetFeatures() const;
|
D | code_generator_arm64.cc | 1010 compiler_options.GetInstructionSetFeatures()->AsArm64InstructionSetFeatures()), in CodeGeneratorARM64() 1041 return GetInstructionSetFeatures().HasSVE(); in ShouldUseSVE() 1046 ? GetInstructionSetFeatures().GetSVEVectorLength() / kBitsPerByte in GetSIMDRegisterWidth() 1637 const Arm64InstructionSetFeatures& CodeGeneratorARM64::GetInstructionSetFeatures() const { in GetInstructionSetFeatures() function in art::arm64::CodeGeneratorARM64 1638 return *GetCompilerOptions().GetInstructionSetFeatures()->AsArm64InstructionSetFeatures(); in GetInstructionSetFeatures() 2678 codegen_->GetInstructionSetFeatures().NeedFixCortexA53_835769()) { in VisitMultiplyAccumulate()
|
D | code_generator_riscv64.cc | 5409 DCHECK(codegen_->GetInstructionSetFeatures().HasZba()); in VisitRiscv64ShiftAdd() 5444 DCHECK(codegen_->GetInstructionSetFeatures().HasZbb()); in VisitBitwiseNegatedRight() 5854 compiler_options.GetInstructionSetFeatures()->AsRiscv64InstructionSetFeatures()), in CodeGeneratorRISCV64() 6351 const Riscv64InstructionSetFeatures& CodeGeneratorRISCV64::GetInstructionSetFeatures() const { in GetInstructionSetFeatures() function in art::riscv64::CodeGeneratorRISCV64 6352 return *GetCompilerOptions().GetInstructionSetFeatures()->AsRiscv64InstructionSetFeatures(); in GetInstructionSetFeatures()
|
/art/compiler/jit/ |
D | jit_compiler.cc | 142 const InstructionSetFeatures* features = compiler_options.GetInstructionSetFeatures(); in TypesLoaded()
|
/art/compiler/driver/ |
D | compiler_options.h | 274 const InstructionSetFeatures* GetInstructionSetFeatures() const { in GetInstructionSetFeatures() function
|
/art/dex2oat/linker/ |
D | elf_writer_quick.cc | 254 compiler_options_.GetInstructionSetFeatures(), in PrepareDebugInfo()
|
D | image_test.h | 294 compiler_options_->GetInstructionSetFeatures(), in DoCompile()
|
D | oat_writer_test.cc | 198 compiler_options_->GetInstructionSetFeatures(), in DoWriteElf()
|
/art/compiler/jni/quick/ |
D | jni_compiler.cc | 91 compiler_options.GetInstructionSetFeatures(); in ArtJniCompileMethodInternal()
|
/art/dex2oat/ |
D | dex2oat.cc | 833 if (!compiler_options_->GetInstructionSetFeatures()->Equals(runtime_features.get())) { in ProcessOptions() 835 << *compiler_options_->GetInstructionSetFeatures() in ProcessOptions() 2118 compiler_options_->GetInstructionSetFeatures(), in WriteOutputFiles()
|