Lines Matching refs:dst
504 if ((args.dst & args.nf) != 0) { in OpVector()
510 state_->cpu.v[args.dst + index] = ptr[index]; in OpVector()
931 args.dst, src, [&indexes](size_t index) { return indexes[index]; }); in OpVector()
951 args.dst, src, [stride](size_t index) { return stride * index; }); in OpVector()
978 args.dst, src, [](size_t index) { return kSegmentSize * sizeof(ElementType) * index; }); in OpVector()
986 args.dst, src, [](size_t index) { return kSegmentSize * sizeof(ElementType) * index; }); in OpVector()
996 args.dst, src, [](size_t index) { return index; }); in OpVector()
1037 void OpVectorLoad(uint8_t dst, Register src, GetElementOffsetLambdaType GetElementOffset) {
1039 if (!IsAligned<kNumRegistersInGroup>(dst)) {
1042 if (dst + kNumRegistersInGroup * kSegmentSize > 32) {
1090 result[field].Set(state_->cpu.v[dst + within_group_id + field * kNumRegistersInGroup]);
1195 state_->cpu.v[dst + within_group_id + field * kNumRegistersInGroup] =
1212 void OpVectorGather(uint8_t dst, uint8_t src1, GetElementIndexLambdaType GetElementIndex) { in OpVectorGather() argument
1214 if (!IsAligned<kRegistersInvolved>(dst | src1)) { in OpVectorGather()
1218 if (dst < (src1 + kRegistersInvolved) && src1 < (dst + kRegistersInvolved)) { in OpVectorGather()
1240 SIMD128Register result{state_->cpu.v[dst + index]}; in OpVectorGather()
1257 state_->cpu.v[dst + index] = result.Get<__uint128_t>(); in OpVectorGather()
1275 kFrm>(args.dst, args.src1, arg2); in OpVector()
1282 kFrm>(args.dst, args.src1, arg2); in OpVector()
1289 kFrm>(args.dst, args.src1, arg2); in OpVector()
1296 kFrm>(args.dst, args.src1, arg2); in OpVector()
1303 kFrm>(args.dst, args.src1, arg2); in OpVector()
1310 kFrm>(args.dst, args.src1, arg2); in OpVector()
1317 kFrm>(args.dst, args.src1, arg2); in OpVector()
1324 kFrm>(args.dst, args.src1, arg2); in OpVector()
1331 kFrm>(args.dst, args.src1, arg2); in OpVector()
1340 args.dst, args.src1, arg2); in OpVector()
1343 args.dst, args.src1, arg2); in OpVector()
1346 args.dst, args.src1, arg2); in OpVector()
1349 args.dst, args.src1, arg2); in OpVector()
1352 args.dst, args.src1, arg2); in OpVector()
1354 return OpVectorslide1up<ElementType, vlmul, vta, vma>(args.dst, args.src1, arg2); in OpVector()
1356 return OpVectorslide1down<ElementType, vlmul, vta, vma>(args.dst, args.src1, arg2); in OpVector()
1364 return OpVectorVmvsx<ElementType, vta>(args.dst, arg2); in OpVector()
1370 return OpVectorx<intrinsics::Vcopyx<ElementType>, ElementType, vlmul, vta, vma>(args.dst, in OpVector()
1379 args.dst, arg2, /*dst_mask=*/args.src1); in OpVector()
1383 args.dst, args.src1, arg2); in OpVector()
1386 args.dst, args.src1, arg2); in OpVector()
1389 args.dst, args.src1, arg2); in OpVector()
1392 args.dst, args.src1, arg2); in OpVector()
1395 args.dst, args.src1, arg2); in OpVector()
1398 args.dst, args.src1, arg2); in OpVector()
1405 kFrm>(args.dst, Vec<SignedType{}>{args.src1}, arg2); in OpVector()
1413 args.dst, in OpVector()
1424 kFrm>(args.dst, Vec<SignedType{}>{args.src1}, arg2); in OpVector()
1431 kFrm>(args.dst, Vec<SignedType{}>{args.src1}, arg2); in OpVector()
1438 kFrm>(args.dst, Vec<SignedType{}>{args.src1}, arg2); in OpVector()
1445 kFrm>(args.dst, Vec<SignedType{}>{args.src1}, arg2); in OpVector()
1448 args.dst, args.src1, arg2); in OpVector()
1451 args.dst, args.src1, arg2); in OpVector()
1454 args.dst, args.src1, arg2); in OpVector()
1457 args.dst, args.src1, arg2); in OpVector()
1460 args.dst, args.src1, arg2); in OpVector()
1463 args.dst, args.src1, arg2); in OpVector()
1466 args.dst, args.src1, arg2); in OpVector()
1469 args.dst, args.src1, arg2); in OpVector()
1495 kFrm>(args.dst, args.src1); in OpVector()
1504 kFrm>(args.dst, args.src1); in OpVector()
1513 kFrm>(args.dst, args.src1); in OpVector()
1522 kFrm>(args.dst, args.src1); in OpVector()
1531 kFrm>(args.dst, args.src1); in OpVector()
1540 kFrm>(args.dst, args.src1); in OpVector()
1564 kFrm>(args.dst, args.src1, args.src2); in OpVector()
1571 kFrm>(args.dst, args.src1, args.src2); in OpVector()
1578 kFrm>(args.dst, args.src1, args.src2); in OpVector()
1585 kFrm>(args.dst, args.src1, args.src2); in OpVector()
1592 kFrm>(args.dst, args.src1, args.src2); in OpVector()
1599 kFrm>(args.dst, args.src1, args.src2); in OpVector()
1606 kFrm>(args.dst, args.src1, args.src2); in OpVector()
1613 kFrm>(args.dst, args.src1, args.src2); in OpVector()
1620 kFrm>(args.dst, args.src1, args.src2); in OpVector()
1631 kFrm>(args.dst, args.src1); in OpVector()
1640 kFrm>(args.dst, args.src1); in OpVector()
1649 kFrm>(args.dst, args.src1); in OpVector()
1658 kFrm>(args.dst, args.src1); in OpVector()
1667 kFrm>(args.dst, args.src1); in OpVector()
1676 kFrm>(args.dst, args.src1); in OpVector()
1685 kFrm>(args.dst, args.src1); in OpVector()
1694 kFrm>(args.dst, args.src1); in OpVector()
1726 kFrm>(args.dst, Vec<kNegativeZero>{args.src1}, args.src2); in OpVector()
1733 kFrm>(args.dst, Vec<kPositiveZero>{args.src1}, args.src2); in OpVector()
1744 kFrm>(args.dst, Vec<kNegativeZero>{args.src1}, args.src2); in OpVector()
1751 kFrm>(args.dst, Vec<kPositiveZero>{args.src1}, args.src2); in OpVector()
1755 args.dst, args.src1, args.src2); in OpVector()
1759 args.dst, in OpVector()
1766 args.dst, args.src1, args.src2); in OpVector()
1770 args.dst, in OpVector()
1777 args.dst, args.src1, args.src2); in OpVector()
1780 args.dst, args.src1, args.src2); in OpVector()
1783 args.dst, args.src1, args.src2); in OpVector()
1794 kFrm>(args.dst, args.src1); in OpVector()
1803 kFrm>(args.dst, args.src1); in OpVector()
1812 kFrm>(args.dst, args.src1); in OpVector()
1821 kFrm>(args.dst, args.src1); in OpVector()
1830 kFrm>(args.dst, args.src1); in OpVector()
1839 kFrm>(args.dst, args.src1); in OpVector()
1852 kFrm>(args.dst, args.src1); in OpVector()
1856 args.dst, args.src1); in OpVector()
1860 args.dst, args.src1); in OpVector()
1873 return OpVectorVmvfs<ElementType>(args.dst, args.src1); in OpVector()
1876 args.dst, args.src1, args.src2); in OpVector()
1879 args.dst, args.src1, args.src2); in OpVector()
1882 args.dst, args.src1, args.src2); in OpVector()
1885 args.dst, args.src1, args.src2); in OpVector()
1893 args.dst, in OpVector()
1905 args.dst, Vec<SignedType{}>{args.src1}, Vec<SignedType{}>{args.src2}); in OpVector()
1913 args.dst, Vec<SignedType{}>{args.src1}, Vec<SignedType{}>{args.src2}); in OpVector()
1921 args.dst, Vec<SignedType{}>{args.src1}, Vec<SignedType{}>{args.src2}); in OpVector()
1924 args.dst, args.src1, args.src2); in OpVector()
1927 args.dst, args.src1, args.src2); in OpVector()
1930 args.dst, args.src1, args.src2); in OpVector()
1933 args.dst, args.src1, args.src2); in OpVector()
1940 kFrm>(args.dst, args.src1, args.src2); in OpVector()
1947 kFrm>(args.dst, args.src1, args.src2); in OpVector()
1954 kFrm>(args.dst, args.src1, args.src2); in OpVector()
1961 kFrm>(args.dst, args.src1, args.src2); in OpVector()
1979 args.dst, args.src, SignedType{args.imm}); in OpVector()
1982 args.dst, args.src, SignedType{args.imm}); in OpVector()
1985 args.dst, args.src, SignedType{args.imm}); in OpVector()
1988 args.dst, args.src, SignedType{args.imm}); in OpVector()
1991 args.dst, args.src, SignedType{args.imm}); in OpVector()
1994 args.dst, args.src, [&args](size_t /*index*/) { return ElementType{args.uimm}; }); in OpVector()
1997 args.dst, args.src, SignedType{args.imm}); in OpVector()
2000 args.dst, args.src, SignedType{args.imm}); in OpVector()
2005 args.dst, args.src, BitCastToUnsigned(SignedType{args.imm})); in OpVector()
2008 args.dst, args.src, SignedType{args.imm}); in OpVector()
2013 args.dst, args.src, BitCastToUnsigned(SignedType{args.imm})); in OpVector()
2016 args.dst, args.src, SignedType{args.imm}); in OpVector()
2025 args.dst, args.src, BitCastToUnsigned(SaturatingSignedType{args.imm})); in OpVector()
2031 vma>(args.dst, args.src, SaturatingSignedType{args.imm}); in OpVector()
2034 args.dst, args.src, UnsignedType{args.uimm}); in OpVector()
2037 args.dst, args.src, UnsignedType{args.uimm}); in OpVector()
2042 args.dst, args.src, BitCastToSigned(UnsignedType{args.uimm})); in OpVector()
2049 args.dst, SignedType{args.imm}); in OpVector()
2057 args.dst, SignedType{args.imm}, /*dst_mask=*/args.src); in OpVector()
2064 return OpVectorVmvXrv<ElementType, 1>(args.dst, args.src); in OpVector()
2066 return OpVectorVmvXrv<ElementType, 2>(args.dst, args.src); in OpVector()
2068 return OpVectorVmvXrv<ElementType, 4>(args.dst, args.src); in OpVector()
2070 return OpVectorVmvXrv<ElementType, 8>(args.dst, args.src); in OpVector()
2081 args.dst, args.src, BitCastToSigned(UnsignedType{args.uimm})); in OpVector()
2084 args.dst, args.src, UnsignedType{args.uimm}); in OpVector()
2087 args.dst, args.src, UnsignedType{args.uimm}); in OpVector()
2090 args.dst, args.src, UnsignedType{args.uimm}); in OpVector()
2097 kVxrm>(args.dst, args.src, UnsignedType{args.uimm}); in OpVector()
2104 kVxrm>(args.dst, args.src, UnsignedType{args.uimm}); in OpVector()
2107 args.dst, args.src, UnsignedType{args.uimm}); in OpVector()
2110 args.dst, args.src, BitCastToSigned(UnsignedType{args.uimm})); in OpVector()
2126 args.dst, args.src1, args.src2); in OpVector()
2129 args.dst, args.src1, args.src2); in OpVector()
2132 args.dst, args.src1, args.src2); in OpVector()
2135 args.dst, args.src1, args.src2); in OpVector()
2138 args.dst, args.src1, args.src2); in OpVector()
2148 args.dst, args.src1, [&indexes](size_t index) { return indexes[index]; }); in OpVector()
2152 args.dst, args.src1, args.src2); in OpVector()
2155 args.dst, args.src1, args.src2); in OpVector()
2158 args.dst, args.src1, args.src2); in OpVector()
2161 args.dst, args.src1, args.src2); in OpVector()
2164 args.dst, args.src1, args.src2); in OpVector()
2167 args.dst, args.src1, args.src2); in OpVector()
2173 vma>(args.dst, args.src1, args.src2); in OpVector()
2179 vma>(args.dst, args.src1, args.src2); in OpVector()
2185 vma>(args.dst, args.src1, args.src2); in OpVector()
2191 vma>(args.dst, args.src1, args.src2); in OpVector()
2194 args.dst, args.src1, args.src2); in OpVector()
2197 args.dst, args.src1, args.src2); in OpVector()
2200 args.dst, args.src1, args.src2); in OpVector()
2203 args.dst, args.src1, args.src2); in OpVector()
2206 args.dst, args.src1, args.src2); in OpVector()
2209 args.dst, args.src1, args.src2); in OpVector()
2212 args.dst, args.src1, args.src2); in OpVector()
2219 args.dst, args.src2); in OpVector()
2227 args.dst, args.src2, /*dst_mask=*/args.src1); in OpVector()
2231 args.dst, args.src1, args.src2); in OpVector()
2234 args.dst, args.src1, args.src2); in OpVector()
2241 kVxrm>(args.dst, args.src1, args.src2); in OpVector()
2244 args.dst, args.src1, args.src2); in OpVector()
2247 args.dst, args.src1, args.src2); in OpVector()
2254 kVxrm>(args.dst, args.src1, args.src2); in OpVector()
2261 kVxrm>(args.dst, args.src1, args.src2); in OpVector()
2277 args.dst, args.src1, MaybeTruncateTo<ElementType>(arg2)); in OpVector()
2280 args.dst, args.src1, MaybeTruncateTo<ElementType>(arg2)); in OpVector()
2283 args.dst, args.src1, MaybeTruncateTo<ElementType>(arg2)); in OpVector()
2286 args.dst, args.src1, MaybeTruncateTo<ElementType>(arg2)); in OpVector()
2289 args.dst, args.src1, MaybeTruncateTo<ElementType>(arg2)); in OpVector()
2292 args.dst, args.src1, MaybeTruncateTo<ElementType>(arg2)); in OpVector()
2295 args.dst, args.src1, [&arg2](size_t /*index*/) { in OpVector()
2300 args.dst, args.src1, MaybeTruncateTo<ElementType>(arg2)); in OpVector()
2303 args.dst, args.src1, MaybeTruncateTo<ElementType>(arg2)); in OpVector()
2306 args.dst, args.src1, MaybeTruncateTo<UnsignedType>(arg2)); in OpVector()
2309 args.dst, args.src1, MaybeTruncateTo<SignedType>(arg2)); in OpVector()
2312 args.dst, args.src1, MaybeTruncateTo<UnsignedType>(arg2)); in OpVector()
2315 args.dst, args.src1, MaybeTruncateTo<SignedType>(arg2)); in OpVector()
2318 args.dst, args.src1, MaybeTruncateTo<UnsignedType>(arg2)); in OpVector()
2321 args.dst, args.src1, MaybeTruncateTo<SignedType>(arg2)); in OpVector()
2327 vma>(args.dst, args.src1, MaybeTruncateTo<ElementType>(arg2)); in OpVector()
2333 vma>(args.dst, args.src1, MaybeTruncateTo<ElementType>(arg2)); in OpVector()
2339 vma>(args.dst, args.src1, MaybeTruncateTo<ElementType>(arg2)); in OpVector()
2345 vma>(args.dst, args.src1, MaybeTruncateTo<ElementType>(arg2)); in OpVector()
2348 args.dst, args.src1, MaybeTruncateTo<ElementType>(arg2)); in OpVector()
2351 args.dst, args.src1, MaybeTruncateTo<UnsignedType>(arg2)); in OpVector()
2354 args.dst, args.src1, MaybeTruncateTo<SignedType>(arg2)); in OpVector()
2357 args.dst, args.src1, MaybeTruncateTo<UnsignedType>(arg2)); in OpVector()
2360 args.dst, args.src1, MaybeTruncateTo<SignedType>(arg2)); in OpVector()
2363 args.dst, args.src1, MaybeTruncateTo<UnsignedType>(arg2)); in OpVector()
2366 args.dst, args.src1, MaybeTruncateTo<SignedType>(arg2)); in OpVector()
2373 args.dst, MaybeTruncateTo<ElementType>(arg2)); in OpVector()
2381 args.dst, MaybeTruncateTo<ElementType>(arg2), /*dst_mask=*/args.src1); in OpVector()
2385 args.dst, args.src1, MaybeTruncateTo<SignedType>(arg2)); in OpVector()
2388 args.dst, args.src1, MaybeTruncateTo<UnsignedType>(arg2)); in OpVector()
2391 args.dst, args.src1, MaybeTruncateTo<UnsignedType>(arg2)); in OpVector()
2394 args.dst, args.src1, MaybeTruncateTo<UnsignedType>(arg2)); in OpVector()
2401 kVxrm>(args.dst, args.src1, MaybeTruncateTo<SignedType>(arg2)); in OpVector()
2404 args.dst, args.src1, MaybeTruncateTo<UnsignedType>(arg2)); in OpVector()
2407 args.dst, args.src1, MaybeTruncateTo<SignedType>(arg2)); in OpVector()
2414 kVxrm>(args.dst, args.src1, MaybeTruncateTo<UnsignedType>(arg2)); in OpVector()
2421 kVxrm>(args.dst, args.src1, MaybeTruncateTo<SignedType>(arg2)); in OpVector()
2436 args.dst, args.src1, args.src2); in OpVector()
2439 args.dst, args.src1, args.src2); in OpVector()
2442 args.dst, args.src1, args.src2); in OpVector()
2445 args.dst, args.src1, args.src2); in OpVector()
2448 args.dst, args.src1, args.src2); in OpVector()
2451 args.dst, args.src1, args.src2); in OpVector()
2454 args.dst, args.src1, args.src2); in OpVector()
2457 args.dst, args.src1, args.src2); in OpVector()
2465 args.dst, Vec<ElementType{}>{args.src1}, args.src2); in OpVector()
2468 args.dst, Vec<~ElementType{}>{args.src1}, args.src2); in OpVector()
2471 args.dst, Vec<ElementType{}>{args.src1}, args.src2); in OpVector()
2474 args.dst, Vec<ElementType{}>{args.src1}, args.src2); in OpVector()
2477 args.dst, in OpVector()
2483 args.dst, in OpVector()
2488 args.dst, Vec<UnsignedType{}>{args.src1}, args.src2); in OpVector()
2491 args.dst, in OpVector()
2496 args.dst, args.src1, args.src2); in OpVector()
2499 args.dst, args.src1, args.src2); in OpVector()
2502 args.dst, args.src1, args.src2); in OpVector()
2505 args.dst, args.src1, args.src2); in OpVector()
2512 return OpVectorVmvxs<SignedType>(args.dst, args.src1); in OpVector()
2514 return OpVectorVWXUnary0<intrinsics::Vcpopm<>, vma>(args.dst, args.src1); in OpVector()
2516 return OpVectorVWXUnary0<intrinsics::Vfirstm<>, vma>(args.dst, args.src1); in OpVector()
2529 vma>(args.dst, args.src1); in OpVector()
2539 vma>(args.dst, args.src1); in OpVector()
2549 vma>(args.dst, args.src1); in OpVector()
2559 vma>(args.dst, args.src1); in OpVector()
2569 vma>(args.dst, args.src1); in OpVector()
2579 vma>(args.dst, args.src1); in OpVector()
2589 return OpVectorVMUnary0<intrinsics::Vmsbfm<>, vma>(args.dst, args.src1); in OpVector()
2591 return OpVectorVMUnary0<intrinsics::Vmsofm<>, vma>(args.dst, args.src1); in OpVector()
2593 return OpVectorVMUnary0<intrinsics::Vmsifm<>, vma>(args.dst, args.src1); in OpVector()
2595 return OpVectorViotam<ElementType, vlmul, vta, vma>(args.dst, args.src1); in OpVector()
2600 return OpVectorVidv<ElementType, vlmul, vta, vma>(args.dst); in OpVector()
2606 args.dst, args.src1, args.src2); in OpVector()
2609 args.dst, args.src1, args.src2); in OpVector()
2612 args.dst, args.src1, args.src2); in OpVector()
2615 args.dst, args.src1, args.src2); in OpVector()
2618 args.dst, args.src1, args.src2); in OpVector()
2621 args.dst, args.src1, args.src2); in OpVector()
2624 args.dst, args.src1, args.src2); in OpVector()
2627 args.dst, args.src1, args.src2); in OpVector()
2630 args.dst, args.src1, args.src2); in OpVector()
2633 args.dst, args.src1, args.src2); in OpVector()
2636 args.dst, args.src1, args.src2); in OpVector()
2639 args.dst, args.src1, args.src2); in OpVector()
2642 args.dst, args.src1, args.src2); in OpVector()
2645 args.dst, args.src1, args.src2); in OpVector()
2648 args.dst, args.src1, args.src2); in OpVector()
2651 args.dst, args.src1, args.src2); in OpVector()
2654 args.dst, args.src1, args.src2); in OpVector()
2657 args.dst, args.src1, args.src2); in OpVector()
2660 args.dst, args.src1, args.src2); in OpVector()
2663 args.dst, args.src1, args.src2); in OpVector()
2666 args.dst, args.src1, args.src2); in OpVector()
2669 args.dst, args.src1, args.src2); in OpVector()
2672 args.dst, args.src1, args.src2); in OpVector()
2675 args.dst, args.src1, args.src2); in OpVector()
2678 args.dst, args.src1, args.src2); in OpVector()
2681 args.dst, args.src1, args.src2); in OpVector()
2695 args.dst, args.src1, MaybeTruncateTo<UnsignedType>(arg2)); in OpVector()
2698 args.dst, args.src1, MaybeTruncateTo<SignedType>(arg2)); in OpVector()
2701 args.dst, args.src1, MaybeTruncateTo<UnsignedType>(arg2)); in OpVector()
2704 args.dst, args.src1, MaybeTruncateTo<SignedType>(arg2)); in OpVector()
2707 args.dst, args.src1, MaybeTruncateTo<SignedType>(arg2)); in OpVector()
2710 args.dst, args.src1, MaybeTruncateTo<SignedType>(arg2)); in OpVector()
2717 return OpVectorVmvsx<SignedType, vta>(args.dst, MaybeTruncateTo<SignedType>(arg2)); in OpVector()
2723 args.dst, args.src1, MaybeTruncateTo<UnsignedType>(arg2)); in OpVector()
2726 args.dst, args.src1, MaybeTruncateTo<SignedType>(arg2)); in OpVector()
2729 args.dst, args.src1, MaybeTruncateTo<UnsignedType>(arg2)); in OpVector()
2732 args.dst, args.src1, MaybeTruncateTo<SignedType>(arg2)); in OpVector()
2735 args.dst, args.src1, MaybeTruncateTo<UnsignedType>(arg2)); in OpVector()
2738 args.dst, args.src1, MaybeTruncateTo<SignedType>(arg2)); in OpVector()
2741 args.dst, args.src1, MaybeTruncateTo<UnsignedType>(arg2)); in OpVector()
2744 args.dst, args.src1, MaybeTruncateTo<SignedType>(arg2)); in OpVector()
2747 args.dst, args.src1, MaybeTruncateTo<ElementType>(arg2)); in OpVector()
2750 args.dst, args.src1, MaybeTruncateTo<ElementType>(arg2)); in OpVector()
2753 args.dst, args.src1, MaybeTruncateTo<ElementType>(arg2)); in OpVector()
2756 args.dst, args.src1, MaybeTruncateTo<ElementType>(arg2)); in OpVector()
2759 args.dst, args.src1, MaybeTruncateTo<UnsignedType>(arg2)); in OpVector()
2762 args.dst, args.src1, MaybeTruncateTo<SignedType>(arg2)); in OpVector()
2765 args.dst, args.src1, MaybeTruncateTo<UnsignedType>(arg2)); in OpVector()
2768 args.dst, args.src1, MaybeTruncateTo<SignedType>(arg2)); in OpVector()
2771 args.dst, args.src1, MaybeTruncateTo<UnsignedType>(arg2)); in OpVector()
2774 args.dst, args.src1, MaybeTruncateTo<SignedType>(arg2)); in OpVector()
2777 args.dst, args.src1, MaybeTruncateTo<UnsignedType>(arg2)); in OpVector()
2780 args.dst, args.src1, MaybeTruncateTo<SignedType>(arg2)); in OpVector()
2783 args.dst, args.src1, MaybeTruncateTo<UnsignedType>(arg2)); in OpVector()
2786 args.dst, args.src1, MaybeTruncateTo<ElementType>(arg2)); in OpVector()
2789 args.dst, args.src1, MaybeTruncateTo<SignedType>(arg2)); in OpVector()
2792 args.dst, args.src1, MaybeTruncateTo<UnsignedType>(arg2)); in OpVector()
2795 args.dst, args.src1, MaybeTruncateTo<SignedType>(arg2)); in OpVector()
2798 args.dst, args.src1, MaybeTruncateTo<ElementType>(arg2)); in OpVector()
2801 args.dst, args.src1, MaybeTruncateTo<ElementType>(arg2)); in OpVector()
2968 void OpVectorViotam(uint8_t dst, uint8_t src1) { in OpVectorViotam() argument
2969 return OpVectorViotam<ElementType, NumberOfRegistersInvolved(vlmul), vta, vma>(dst, src1); in OpVectorViotam()
2973 void OpVectorViotam(uint8_t dst, uint8_t src1) { in OpVectorViotam() argument
2993 SIMD128Register result{state_->cpu.v[dst + index]}; in OpVectorViotam()
3001 state_->cpu.v[dst + index] = result.Get<__uint128_t>(); in OpVectorViotam()
3006 void OpVectorVidv(uint8_t dst) { in OpVectorVidv() argument
3007 return OpVectorVidv<ElementType, NumberOfRegistersInvolved(vlmul), vta, vma>(dst); in OpVectorVidv()
3011 void OpVectorVidv(uint8_t dst) { in OpVectorVidv() argument
3012 if (!IsAligned<kRegistersInvolved>(dst)) { in OpVectorVidv()
3025 SIMD128Register result{state_->cpu.v[dst + index]}; in OpVectorVidv()
3028 state_->cpu.v[dst + index] = result.Get<__uint128_t>(); in OpVectorVidv()
3033 void OpVectorVmvfs(uint8_t dst, uint8_t src) { in OpVectorVmvfs() argument
3039 NanBoxAndSetFpReg<ElementType>(dst, SIMD128Register{state_->cpu.v[src]}.Get<FpRegister>(0)); in OpVectorVmvfs()
3044 void OpVectorVmvsx(uint8_t dst, ElementType element) { in OpVectorVmvsx() argument
3056 result.Set(state_->cpu.v[dst]); in OpVectorVmvsx()
3059 state_->cpu.v[dst] = result.Get<Int128>(); in OpVectorVmvsx()
3065 void OpVectorVmvxs(uint8_t dst, uint8_t src1) { in OpVectorVmvxs() argument
3069 SetRegOrIgnore(dst, element); in OpVectorVmvxs()
3074 void OpVectorVWXUnary0(uint8_t dst, uint8_t src1) { in OpVectorVWXUnary0() argument
3090 SetRegOrIgnore(dst, TruncateTo<UInt64>(BitCastToUnsigned(result.Get<Int128>()))); in OpVectorVWXUnary0()
3094 void OpVectormm(uint8_t dst, uint8_t src1, uint8_t src2) { in OpVectormm() argument
3108 result.Set(state_->cpu.v[dst]); in OpVectormm()
3115 state_->cpu.v[dst] = result.Get<__uint128_t>(); in OpVectormm()
3119 void OpVectorVMUnary0(uint8_t dst, uint8_t src1) { in OpVectorVMUnary0() argument
3142 result = (result & mask) | (SIMD128Register(state_->cpu.v[dst]) & ~mask); in OpVectorVMUnary0()
3148 state_->cpu.v[dst] = result.Get<__uint128_t>(); in OpVectorVMUnary0()
3152 void OpVectorVmvXrv(uint8_t dst, uint8_t src) { in OpVectorVmvXrv() argument
3153 if (!IsAligned<kRegistersInvolved>(dst | src)) { in OpVectorVmvXrv()
3166 state_->cpu.v[dst + index] = state_->cpu.v[src + index]; in OpVectorVmvXrv()
3171 SIMD128Register destination{state_->cpu.v[dst + index]}; in OpVectorVmvXrv()
3177 state_->cpu.v[dst + index] = destination.Get<__uint128_t>(); in OpVectorVmvXrv()
3179 state_->cpu.v[dst + index] = state_->cpu.v[src + index]; in OpVectorVmvXrv()
3188 void OpVectorToMaskvv(uint8_t dst, uint8_t src1, uint8_t src2) { in OpVectorToMaskvv() argument
3193 kExtraCsrs...>(dst, Vec{src1}, Vec{src2}); in OpVectorToMaskvv()
3201 void OpVectorToMaskvx(uint8_t dst, uint8_t src1, ElementType arg2) { in OpVectorToMaskvx() argument
3206 kExtraCsrs...>(dst, Vec{src1}, arg2); in OpVectorToMaskvx()
3215 void OpVectorToMask(uint8_t dst, Args... args) { in OpVectorToMask() argument
3221 SIMD128Register original_result(state_->cpu.v[dst]); in OpVectorToMask()
3252 state_->cpu.v[dst] = (result_before_vl_masking | tail_mask).Get<__uint128_t>(); in OpVectorToMask()
3262 void OpVectorv(uint8_t dst, uint8_t src1, DstMaskType... dst_mask) { in OpVectorv() argument
3268 kExtraCsrs...>(dst, src1, dst_mask...); in OpVectorv()
3278 void OpVectorv(uint8_t dst, uint8_t src, DstMaskType... dst_mask) { in OpVectorv() argument
3280 if (!IsAligned<kRegistersInvolved>(dst | src | (dst_mask | ... | 0))) { in OpVectorv()
3293 SIMD128Register result{state_->cpu.v[dst + index]}; in OpVectorv()
3296 result_mask.Set(state_->cpu.v[dst + index]); in OpVectorv()
3310 state_->cpu.v[dst + index] = result.Get<__uint128_t>(); in OpVectorv()
3321 void OpVectorvs(uint8_t dst, Vec<kDefaultElement> src1, uint8_t src2) { in OpVectorvs() argument
3327 kExtraCsrs...>(dst, src1, src2); in OpVectorvs()
3337 void OpVectorvs(uint8_t dst, Vec<kDefaultElement> src1, uint8_t src2) { in OpVectorvs() argument
3338 if (!IsAligned<kRegistersInvolved>(dst | src1.start_no)) { in OpVectorvs()
3359 SIMD128Register result{state_->cpu.v[dst]}; in OpVectorvs()
3362 state_->cpu.v[dst] = result.Get<__uint128_t>(); in OpVectorvs()
3371 void OpVectorvv(uint8_t dst, uint8_t src1, uint8_t src2) { in OpVectorvv() argument
3377 kExtraCsrs...>(dst, Vec{src1}, Vec{src2}); in OpVectorvv()
3386 void OpVectorvvv(uint8_t dst, uint8_t src1, uint8_t src2) { in OpVectorvvv() argument
3392 kExtraCsrs...>(dst, Vec{src1}, Vec{src2}, Vec{dst}); in OpVectorvvv()
3401 void OpVectorWidenv(uint8_t dst, uint8_t src) { in OpVectorWidenv() argument
3410 kExtraCsrs...>(dst, Vec{src}); in OpVectorWidenv()
3423 void OpVectorWidenvv(uint8_t dst, uint8_t src1, uint8_t src2) { in OpVectorWidenvv() argument
3432 kExtraCsrs...>(dst, Vec{src1}, Vec{src2}); in OpVectorWidenvv()
3444 void OpVectorWidenvvw(uint8_t dst, uint8_t src1, uint8_t src2) { in OpVectorWidenvvw() argument
3453 kExtraCsrs...>(dst, Vec{src1}, Vec{src2}, WideVec{dst}); in OpVectorWidenvvw()
3465 void OpVectorWidenwv(uint8_t dst, uint8_t src1, uint8_t src2) { in OpVectorWidenwv() argument
3474 kExtraCsrs...>(dst, WideVec{src1}, Vec{src2}); in OpVectorWidenwv()
3485 void OpVectorWidenwx(uint8_t dst, uint8_t src1, ElementType arg2) { in OpVectorWidenwx() argument
3494 kExtraCsrs...>(dst, WideVec{src1}, arg2); in OpVectorWidenwx()
3505 void OpVectorWidenvx(uint8_t dst, uint8_t src1, ElementType arg2) { in OpVectorWidenvx() argument
3514 kExtraCsrs...>(dst, Vec{src1}, arg2); in OpVectorWidenvx()
3525 void OpVectorWidenvxw(uint8_t dst, uint8_t src1, ElementType arg2) { in OpVectorWidenvxw() argument
3534 kExtraCsrs...>(dst, Vec{src1}, arg2, WideVec{dst}); in OpVectorWidenvxw()
3547 void OpVectorWiden(uint8_t dst, Args... args) { in OpVectorWiden() argument
3557 uint8_t ored_wide_args = OrValuesOnlyForType<WideVec>(args...) | dst; in OpVectorWiden()
3569 if (OrResultsOnlyForType<Vec>([dst](auto arg) { return arg.start_no == dst; }, args...)) { in OpVectorWiden()
3582 SIMD128Register result(state_->cpu.v[dst + 2 * index]); in OpVectorWiden()
3592 state_->cpu.v[dst + 2 * index] = result.Get<__uint128_t>(); in OpVectorWiden()
3594 result.Set(state_->cpu.v[dst + 2 * index + 1]); in OpVectorWiden()
3604 state_->cpu.v[dst + 2 * index + 1] = result.Get<__uint128_t>(); in OpVectorWiden()
3615 void OpVectorvx(uint8_t dst, uint8_t src1, ElementType arg2) { in OpVectorvx() argument
3621 kExtraCsrs...>(dst, Vec{src1}, arg2); in OpVectorvx()
3631 void OpVectorSameWidth(uint8_t dst, Args... args) { in OpVectorSameWidth() argument
3634 if (!IsAligned<kRegistersInvolved>(OrValuesOnlyForType<Vec>(args...) | dst)) { in OpVectorSameWidth()
3647 SIMD128Register result(state_->cpu.v[dst + index]); in OpVectorSameWidth()
3657 state_->cpu.v[dst + index] = result.Get<__uint128_t>(); in OpVectorSameWidth()
3667 void OpVectorNarroww(uint8_t dst, uint8_t src) { in OpVectorNarroww() argument
3676 kExtraCsrs...>(dst, WideVec{src}); in OpVectorNarroww()
3688 void OpVectorNarrowwx(uint8_t dst, uint8_t src1, ElementType arg2) { in OpVectorNarrowwx() argument
3697 kExtraCsrs...>(dst, WideVec{src1}, arg2); in OpVectorNarrowwx()
3709 void OpVectorNarrowwv(uint8_t dst, uint8_t src1, uint8_t src2) { in OpVectorNarrowwv() argument
3718 kExtraCsrs...>(dst, WideVec{src1}, Vec{src2}); in OpVectorNarrowwv()
3731 void OpVectorNarrow(uint8_t dst, Args... args) { in OpVectorNarrow() argument
3737 uint8_t ored_args = OrValuesOnlyForType<Vec>(args...) | dst; in OpVectorNarrow()
3751 [dst](auto arg) { return arg.start_no == dst + kRegistersInvolved; }, args...)) { in OpVectorNarrow()
3765 SIMD128Register orig_result(state_->cpu.v[dst + index]); in OpVectorNarrow()
3778 state_->cpu.v[dst + index] = result.template Get<__uint128_t>(); in OpVectorNarrow()
3788 void OpVectorVXUnary0(uint8_t dst, uint8_t src) { in OpVectorVXUnary0() argument
3792 if (!IsAligned<kDestRegistersInvolved>(dst) || !IsAligned<kSourceRegistersInvolved>(src)) { in OpVectorVXUnary0()
3807 SIMD128Register result{state_->cpu.v[dst + dst_index]}; in OpVectorVXUnary0()
3812 state_->cpu.v[dst + dst_index] = result.Get<__uint128_t>(); in OpVectorVXUnary0()
3823 void OpVectorvxv(uint8_t dst, uint8_t src1, ElementType arg2) { in OpVectorvxv() argument
3829 kExtraCsrs...>(dst, Vec{src1}, arg2, Vec{dst}); in OpVectorvxv()
3838 void OpVectorx(uint8_t dst, ElementType arg2, DstMaskType... dst_mask) { in OpVectorx() argument
3840 dst, arg2, dst_mask...); in OpVectorx()
3849 void OpVectorx(uint8_t dst, ElementType arg2, DstMaskType... dst_mask) { in OpVectorx() argument
3851 if (!IsAligned<kRegistersInvolved>(dst | (dst_mask | ... | 0))) { in OpVectorx()
3864 SIMD128Register result(state_->cpu.v[dst + index]); in OpVectorx()
3867 result_mask.Set(state_->cpu.v[dst + index]); in OpVectorx()
3874 state_->cpu.v[dst + index] = result.Get<__uint128_t>(); in OpVectorx()
3879 void OpVectorslideup(uint8_t dst, uint8_t src, Register offset) { in OpVectorslideup() argument
3881 dst, src, offset); in OpVectorslideup()
3885 void OpVectorslideup(uint8_t dst, uint8_t src, Register offset) { in OpVectorslideup() argument
3887 if (!IsAligned<kRegistersInvolved>(dst | src)) { in OpVectorslideup()
3891 if (dst < (src + kRegistersInvolved) && src < (dst + kRegistersInvolved)) { in OpVectorslideup()
3915 SIMD128Register result(state_->cpu.v[dst + index]); in OpVectorslideup()
3934 state_->cpu.v[dst + index] = result.Get<__uint128_t>(); in OpVectorslideup()
3939 void OpVectorslide1up(uint8_t dst, uint8_t src, ElementType xval) { in OpVectorslide1up() argument
3943 OpVectorslideup<ElementType, NumberOfRegistersInvolved(vlmul), vta, vma>(dst, src, 1); in OpVectorslide1up()
3964 SIMD128Register result = state_->cpu.v[dst]; in OpVectorslide1up()
3966 state_->cpu.v[dst] = result.Get<__uint128_t>(); in OpVectorslide1up()
3970 void OpVectorslidedown(uint8_t dst, uint8_t src, Register offset) { in OpVectorslidedown() argument
3972 dst, src, offset); in OpVectorslidedown()
3976 void OpVectorslidedown(uint8_t dst, uint8_t src, Register offset) { in OpVectorslidedown() argument
3978 if (!IsAligned<kRegistersInvolved>(dst | src)) { in OpVectorslidedown()
3992 SIMD128Register result(state_->cpu.v[dst + index]); in OpVectorslidedown()
4010 state_->cpu.v[dst + index] = result.Get<__uint128_t>(); in OpVectorslidedown()
4015 void OpVectorslide1down(uint8_t dst, uint8_t src, ElementType xval) { in OpVectorslide1down() argument
4044 SIMD128Register original = state_->cpu.v[dst + last_elem_register]; in OpVectorslide1down()
4051 OpVectorslidedown<ElementType, NumberOfRegistersInvolved(vlmul), vta, vma>(dst, src, 1); in OpVectorslide1down()
4059 SIMD128Register result = state_->cpu.v[dst + last_elem_register]; in OpVectorslide1down()
4061 state_->cpu.v[dst + last_elem_register] = result.Get<__uint128_t>(); in OpVectorslide1down()