/aosp_15_r20/external/vixl/test/aarch64/ |
H A D | test-assembler-neon-aarch64.cc | 6678 __ Fcmla(v31.V2S(), v1.V2S(), v2.V2S(), 90); in TEST() local 6679 __ Fcmla(v31.V2S(), v1.V2S(), v2.V2S(), 0); in TEST() local 6681 __ Fcmla(v30.V2S(), v3.V2S(), v4.V2S(), 0); in TEST() local 6682 __ Fcmla(v30.V2S(), v3.V2S(), v4.V2S(), 90); in TEST() local 6684 __ Fcmla(v29.V2S(), v5.V2S(), v6.V2S(), 90); in TEST() local 6685 __ Fcmla(v29.V2S(), v5.V2S(), v6.V2S(), 0); in TEST() local 6687 __ Fcmla(v28.V2D(), v9.V2D(), v10.V2D(), 0); in TEST() local 6688 __ Fcmla(v28.V2D(), v9.V2D(), v10.V2D(), 90); in TEST() local 6691 __ Fcmla(v27.V2S(), v1.V2S(), v2.V2S(), 0); in TEST() local 6692 __ Fcmla(v26.V2S(), v2.V2S(), v1.V2S(), 0); in TEST() local [all …]
|
H A D | test-assembler-sve-aarch64.cc | 14639 __ Fcmla(z4.VnH(), z0.VnH(), z3.VnH(), 2, 0); in TEST_SVE() local 14640 __ Fcmla(z4.VnH(), z0.VnH(), z3.VnH(), 2, 90); in TEST_SVE() local 14646 __ Fcmla(z5.VnH(), z0.VnH(), z3.VnH(), 1, 180); in TEST_SVE() local 14647 __ Fcmla(z5.VnH(), z0.VnH(), z3.VnH(), 1, 270); in TEST_SVE() local 14652 __ Fcmla(z6.VnH(), p0.Merging(), z6.VnH(), z0.VnH(), z2.VnH(), 0); in TEST_SVE() local 14653 __ Fcmla(z6.VnH(), p0.Merging(), z6.VnH(), z0.VnH(), z2.VnH(), 90); in TEST_SVE() local 14669 __ Fcmla(z7.VnS(), z0.VnS(), z3.VnS(), 1, 0); in TEST_SVE() local 14670 __ Fcmla(z7.VnS(), z0.VnS(), z3.VnS(), 1, 90); in TEST_SVE() local 14673 __ Fcmla(z8.VnS(), z0.VnS(), z3.VnS(), 0, 180); in TEST_SVE() local 14674 __ Fcmla(z8.VnS(), z0.VnS(), z3.VnS(), 0, 270); in TEST_SVE() local [all …]
|
/aosp_15_r20/external/vixl/src/aarch64/ |
H A D | macro-assembler-sve-aarch64.cc | 2122 void MacroAssembler::Fcmla(const ZRegister& zd, in Fcmla() function in vixl::aarch64::MacroAssembler
|
H A D | macro-assembler-aarch64.h | 3153 void Fcmla(const VRegister& vd, in Fcmla() function 3162 void Fcmla(const VRegister& vd, in Fcmla() function 4268 void Fcmla(const ZRegister& zda, in Fcmla() function
|