Lines Matching full:le
3 …vector -triple powerpc64le-unknown-unknown -emit-llvm %s -o - | FileCheck %s -check-prefix=CHECK-LE
73 // CHECK-LE: call <2 x i64> @llvm.ppc.altivec.vmaxsd(<2 x i64> %{{[0-9]*}}, <2 x i64> in test1()
79 // CHECK-LE: call <2 x double> @llvm.fabs.v2f64(<2 x double> %{{.*}}) in test1()
80 // CHECK-LE: store <2 x double> %{{.*}}, <2 x double>* @res_vd in test1()
86 // CHECK-LE: add <2 x i64> in test1()
91 // CHECK-LE: add <2 x i64> in test1()
97 // CHECK-LE: @llvm.ppc.altivec.vaddcuw in test1()
101 // CHECK-LE: @llvm.ppc.altivec.vaddcuw in test1()
105 // CHECK-LE: @llvm.ppc.altivec.vaddcuq in test1()
109 // CHECK-LE: @llvm.ppc.altivec.vaddcuq in test1()
114 // CHECK-LE: @llvm.ppc.altivec.vaddeuqm in test1()
118 // CHECK-LE: @llvm.ppc.altivec.vaddeuqm in test1()
123 // CHECK-LE: @llvm.ppc.altivec.vaddecuq in test1()
128 // CHECK-LE: @llvm.ppc.altivec.vperm in test1()
132 // CHECK-LE: @llvm.ppc.altivec.vperm in test1()
136 // CHECK-LE: @llvm.ppc.altivec.vperm in test1()
142 // CHECK-LE: @llvm.ppc.altivec.vperm in test1()
146 // CHECK-LE: @llvm.ppc.altivec.vperm in test1()
150 // CHECK-LE: @llvm.ppc.altivec.vperm in test1()
156 // CHECK-LE: @llvm.ppc.altivec.vcmpequd in test1()
161 // CHECK-LE: @llvm.ppc.altivec.vcmpequd in test1()
167 // CHECK-LE: @llvm.ppc.altivec.vcmpgtsd in test1()
172 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud in test1()
178 // CHECK-LE: @llvm.ppc.altivec.vcmpgtsd in test1()
183 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud in test1()
189 // CHECK-LE: @llvm.ppc.altivec.vcmpgtsd in test1()
194 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud in test1()
200 // CHECK-LE: call <2 x i64> @llvm.ppc.altivec.vcmpgtsd(<2 x i64> %{{[0-9]*}}, <2 x i64> %{{[0-9]*}}) in test1()
205 // CHECK-LE: call <2 x i64> @llvm.ppc.altivec.vcmpgtud(<2 x i64> %{{[0-9]*}}, <2 x i64> %{{[0-9]*}}) in test1()
223 // CHECK-LE: [[T1:%.+]] = bitcast <16 x i8> {{.+}} to <4 x i32> in test1()
224 // CHECK-LE: [[T2:%.+]] = bitcast <16 x i8> {{.+}} to <4 x i32> in test1()
225 // CHECK-LE: [[T3:%.+]] = call <4 x i32> @llvm.ppc.vsx.xxleqv(<4 x i32> [[T1]], <4 x i32> [[T2]]) in test1()
226 // CHECK-LE: bitcast <4 x i32> [[T3]] to <16 x i8> in test1()
234 // CHECK-LE: [[T1:%.+]] = bitcast <16 x i8> {{.+}} to <4 x i32> in test1()
235 // CHECK-LE: [[T2:%.+]] = bitcast <16 x i8> {{.+}} to <4 x i32> in test1()
236 // CHECK-LE: [[T3:%.+]] = call <4 x i32> @llvm.ppc.vsx.xxleqv(<4 x i32> [[T1]], <4 x i32> [[T2]]) in test1()
237 // CHECK-LE: bitcast <4 x i32> [[T3]] to <16 x i8> in test1()
245 // CHECK-LE: [[T1:%.+]] = bitcast <16 x i8> {{.+}} to <4 x i32> in test1()
246 // CHECK-LE: [[T2:%.+]] = bitcast <16 x i8> {{.+}} to <4 x i32> in test1()
247 // CHECK-LE: [[T3:%.+]] = call <4 x i32> @llvm.ppc.vsx.xxleqv(<4 x i32> [[T1]], <4 x i32> [[T2]]) in test1()
248 // CHECK-LE: bitcast <4 x i32> [[T3]] to <16 x i8> in test1()
256 // CHECK-LE: [[T1:%.+]] = bitcast <8 x i16> {{.+}} to <4 x i32> in test1()
257 // CHECK-LE: [[T2:%.+]] = bitcast <8 x i16> {{.+}} to <4 x i32> in test1()
258 // CHECK-LE: [[T3:%.+]] = call <4 x i32> @llvm.ppc.vsx.xxleqv(<4 x i32> [[T1]], <4 x i32> [[T2]]) in test1()
259 // CHECK-LE: bitcast <4 x i32> [[T3]] to <8 x i16> in test1()
267 // CHECK-LE: [[T1:%.+]] = bitcast <8 x i16> {{.+}} to <4 x i32> in test1()
268 // CHECK-LE: [[T2:%.+]] = bitcast <8 x i16> {{.+}} to <4 x i32> in test1()
269 // CHECK-LE: [[T3:%.+]] = call <4 x i32> @llvm.ppc.vsx.xxleqv(<4 x i32> [[T1]], <4 x i32> [[T2]]) in test1()
270 // CHECK-LE: bitcast <4 x i32> [[T3]] to <8 x i16> in test1()
278 // CHECK-LE: [[T1:%.+]] = bitcast <8 x i16> {{.+}} to <4 x i32> in test1()
279 // CHECK-LE: [[T2:%.+]] = bitcast <8 x i16> {{.+}} to <4 x i32> in test1()
280 // CHECK-LE: [[T3:%.+]] = call <4 x i32> @llvm.ppc.vsx.xxleqv(<4 x i32> [[T1]], <4 x i32> [[T2]]) in test1()
281 // CHECK-LE: bitcast <4 x i32> [[T3]] to <8 x i16> in test1()
286 // CHECK-LE: call <4 x i32> @llvm.ppc.vsx.xxleqv(<4 x i32> {{.*}}, <4 x i32> {{.+}}) in test1()
291 // CHECK-LE: call <4 x i32> @llvm.ppc.vsx.xxleqv(<4 x i32> {{.*}}, <4 x i32> {{.+}}) in test1()
296 // CHECK-LE: call <4 x i32> @llvm.ppc.vsx.xxleqv(<4 x i32> {{.*}}, <4 x i32> {{.+}}) in test1()
304 // CHECK-LE: [[T1:%.+]] = bitcast <2 x i64> {{.+}} to <4 x i32> in test1()
305 // CHECK-LE: [[T2:%.+]] = bitcast <2 x i64> {{.+}} to <4 x i32> in test1()
306 // CHECK-LE: [[T3:%.+]] = call <4 x i32> @llvm.ppc.vsx.xxleqv(<4 x i32> [[T1]], <4 x i32> [[T2]]) in test1()
307 // CHECK-LE: bitcast <4 x i32> [[T3]] to <2 x i64> in test1()
315 // CHECK-LE: [[T1:%.+]] = bitcast <2 x i64> {{.+}} to <4 x i32> in test1()
316 // CHECK-LE: [[T2:%.+]] = bitcast <2 x i64> {{.+}} to <4 x i32> in test1()
317 // CHECK-LE: [[T3:%.+]] = call <4 x i32> @llvm.ppc.vsx.xxleqv(<4 x i32> [[T1]], <4 x i32> [[T2]]) in test1()
318 // CHECK-LE: bitcast <4 x i32> [[T3]] to <2 x i64> in test1()
326 // CHECK-LE: [[T1:%.+]] = bitcast <2 x i64> {{.+}} to <4 x i32> in test1()
327 // CHECK-LE: [[T2:%.+]] = bitcast <2 x i64> {{.+}} to <4 x i32> in test1()
328 // CHECK-LE: [[T3:%.+]] = call <4 x i32> @llvm.ppc.vsx.xxleqv(<4 x i32> [[T1]], <4 x i32> [[T2]]) in test1()
329 // CHECK-LE: bitcast <4 x i32> [[T3]] to <2 x i64> in test1()
337 // CHECK-LE: [[T1:%.+]] = bitcast <4 x float> {{.+}} to <4 x i32> in test1()
338 // CHECK-LE: [[T2:%.+]] = bitcast <4 x float> {{.+}} to <4 x i32> in test1()
339 // CHECK-LE: [[T3:%.+]] = call <4 x i32> @llvm.ppc.vsx.xxleqv(<4 x i32> [[T1]], <4 x i32> [[T2]]) in test1()
340 // CHECK-LE: bitcast <4 x i32> [[T3]] to <4 x float> in test1()
348 // CHECK-LE: [[T1:%.+]] = bitcast <2 x double> {{.+}} to <4 x i32> in test1()
349 // CHECK-LE: [[T2:%.+]] = bitcast <2 x double> {{.+}} to <4 x i32> in test1()
350 // CHECK-LE: [[T3:%.+]] = call <4 x i32> @llvm.ppc.vsx.xxleqv(<4 x i32> [[T1]], <4 x i32> [[T2]]) in test1()
351 // CHECK-LE: bitcast <4 x i32> [[T3]] to <2 x double> in test1()
357 // CHECK-LE: extractelement <2 x i64> in test1()
361 // CHECK-LE: extractelement <2 x i64> in test1()
365 // CHECK-LE: extractelement <2 x i64> in test1()
369 // CHECK-LE: extractelement <2 x double> in test1()
374 // CHECK-LE: insertelement <2 x i64> in test1()
378 // CHECK-LE: insertelement <2 x i64> in test1()
382 // CHECK-LE: insertelement <2 x i64> in test1()
386 // CHECK-LE: insertelement <2 x double> in test1()
391 // CHECK-LE: call <16 x i8> @llvm.ctlz.v16i8(<16 x i8> %{{.+}}, i1 false) in test1()
396 // CHECK-LE: call <16 x i8> @llvm.ctlz.v16i8(<16 x i8> %{{.+}}, i1 false) in test1()
400 // CHECK-LE: call <8 x i16> @llvm.ctlz.v8i16(<8 x i16> %{{.+}}, i1 false) in test1()
404 // CHECK-LE: call <8 x i16> @llvm.ctlz.v8i16(<8 x i16> %{{.+}}, i1 false) in test1()
408 // CHECK-LE: call <4 x i32> @llvm.ctlz.v4i32(<4 x i32> %{{.+}}, i1 false) in test1()
412 // CHECK-LE: call <4 x i32> @llvm.ctlz.v4i32(<4 x i32> %{{.+}}, i1 false) in test1()
416 // CHECK-LE: call <2 x i64> @llvm.ctlz.v2i64(<2 x i64> %{{.+}}, i1 false) in test1()
420 // CHECK-LE: call <2 x i64> @llvm.ctlz.v2i64(<2 x i64> %{{.+}}, i1 false) in test1()
426 // CHECK-LE: @llvm.ppc.altivec.vcmpequd.p in test1()
431 // CHECK-LE: @llvm.ppc.altivec.vcmpequd.p in test1()
436 // CHECK-LE: @llvm.ppc.altivec.vcmpequd.p in test1()
441 // CHECK-LE: @llvm.ppc.altivec.vcmpequd.p in test1()
446 // CHECK-LE: @llvm.ppc.altivec.vcmpequd.p in test1()
451 // CHECK-LE: @llvm.ppc.altivec.vcmpequd.p in test1()
456 // CHECK-LE: @llvm.ppc.altivec.vcmpequd.p in test1()
461 // CHECK-LE: @llvm.ppc.vsx.xvcmpeqdp.p in test1()
466 // CHECK-LE: @llvm.ppc.altivec.vcmpequd.p in test1()
471 // CHECK-LE: @llvm.ppc.altivec.vcmpequd.p in test1()
476 // CHECK-LE: @llvm.ppc.altivec.vcmpequd.p in test1()
481 // CHECK-LE: @llvm.ppc.altivec.vcmpequd.p in test1()
486 // CHECK-LE: @llvm.ppc.altivec.vcmpequd.p in test1()
491 // CHECK-LE: @llvm.ppc.altivec.vcmpequd.p in test1()
496 // CHECK-LE: @llvm.ppc.altivec.vcmpequd.p in test1()
504 // CHECK-LE: @llvm.ppc.vsx.xvcmpeqdp.p in test1()
511 // CHECK-LE: @llvm.ppc.vsx.xvcmpgedp.p in test1()
515 // CHECK-LE: @llvm.ppc.vsx.xvcmpgtdp.p in test1()
520 // CHECK-LE: @llvm.ppc.altivec.vcmpequd.p in test1()
525 // CHECK-LE: @llvm.ppc.altivec.vcmpequd.p in test1()
530 // CHECK-LE: @llvm.ppc.altivec.vcmpequd.p in test1()
535 // CHECK-LE: @llvm.ppc.altivec.vcmpequd.p in test1()
540 // CHECK-LE: @llvm.ppc.altivec.vcmpequd.p in test1()
545 // CHECK-LE: @llvm.ppc.altivec.vcmpequd.p in test1()
550 // CHECK-LE: @llvm.ppc.altivec.vcmpequd.p in test1()
555 // CHECK-LE: @llvm.ppc.vsx.xvcmpeqdp.p in test1()
560 // CHECK-LE: @llvm.ppc.altivec.vcmpequd.p in test1()
565 // CHECK-LE: @llvm.ppc.altivec.vcmpequd.p in test1()
570 // CHECK-LE: @llvm.ppc.altivec.vcmpequd.p in test1()
575 // CHECK-LE: @llvm.ppc.altivec.vcmpequd.p in test1()
580 // CHECK-LE: @llvm.ppc.altivec.vcmpequd.p in test1()
585 // CHECK-LE: @llvm.ppc.altivec.vcmpequd.p in test1()
590 // CHECK-LE: @llvm.ppc.altivec.vcmpequd.p in test1()
595 // CHECK-LE: @llvm.ppc.vsx.xvcmpeqdp.p in test1()
600 // CHECK-LE: @llvm.ppc.altivec.vcmpgtsd.p in test1()
605 // CHECK-LE: @llvm.ppc.altivec.vcmpgtsd.p in test1()
610 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
615 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
620 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
625 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
630 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
635 // CHECK-LE: @llvm.ppc.vsx.xvcmpgedp.p in test1()
640 // CHECK-LE: @llvm.ppc.altivec.vcmpgtsd.p in test1()
645 // CHECK-LE: @llvm.ppc.altivec.vcmpgtsd.p in test1()
650 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
655 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
660 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
665 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
670 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
675 // CHECK-LE: @llvm.ppc.vsx.xvcmpgtdp.p in test1()
680 // CHECK-LE: @llvm.ppc.altivec.vcmpgtsd.p in test1()
685 // CHECK-LE: @llvm.ppc.altivec.vcmpgtsd.p in test1()
690 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
695 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
700 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
705 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
710 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
715 // CHECK-LE: @llvm.ppc.vsx.xvcmpgedp.p in test1()
720 // CHECK-LE: @llvm.ppc.altivec.vcmpgtsd.p in test1()
725 // CHECK-LE: @llvm.ppc.altivec.vcmpgtsd.p in test1()
730 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
735 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
740 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
745 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
750 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
755 // CHECK-LE: @llvm.ppc.vsx.xvcmpgtdp.p in test1()
759 // CHECK-LE: @llvm.ppc.vsx.xvcmpeqdp.p in test1()
764 // CHECK-LE: @llvm.ppc.altivec.vcmpgtsd.p in test1()
769 // CHECK-LE: @llvm.ppc.altivec.vcmpgtsd.p in test1()
774 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
779 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
784 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
789 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
794 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
799 // CHECK-LE: @llvm.ppc.vsx.xvcmpgedp.p in test1()
804 // CHECK-LE: @llvm.ppc.altivec.vcmpgtsd.p in test1()
809 // CHECK-LE: @llvm.ppc.altivec.vcmpgtsd.p in test1()
814 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
819 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
824 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
829 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
834 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
839 // CHECK-LE: @llvm.ppc.vsx.xvcmpgtdp.p in test1()
844 // CHECK-LE: @llvm.ppc.altivec.vcmpgtsd.p in test1()
849 // CHECK-LE: @llvm.ppc.altivec.vcmpgtsd.p in test1()
854 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
859 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
864 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
869 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
874 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
879 // CHECK-LE: @llvm.ppc.vsx.xvcmpgedp.p in test1()
884 // CHECK-LE: @llvm.ppc.altivec.vcmpgtsd.p in test1()
889 // CHECK-LE: @llvm.ppc.altivec.vcmpgtsd.p in test1()
894 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
899 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
904 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
909 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
914 // CHECK-LE: @llvm.ppc.altivec.vcmpgtud.p in test1()
919 // CHECK-LE: @llvm.ppc.vsx.xvcmpgtdp.p in test1()
924 // CHECK-LE: @llvm.ppc.altivec.vmaxsd in test1()
929 // CHECK-LE: @llvm.ppc.altivec.vmaxsd in test1()
934 // CHECK-LE: @llvm.ppc.altivec.vmaxsd in test1()
939 // CHECK-LE: @llvm.ppc.altivec.vmaxud in test1()
944 // CHECK-LE: @llvm.ppc.altivec.vmaxud in test1()
949 // CHECK-LE: @llvm.ppc.altivec.vmaxud in test1()
955 // CHECK-LE: @llvm.ppc.altivec.vperm in test1()
959 // CHECK-LE: @llvm.ppc.altivec.vperm in test1()
964 // CHECK-LE: @llvm.ppc.altivec.vminsd in test1()
969 // CHECK-LE: @llvm.ppc.altivec.vminsd in test1()
974 // CHECK-LE: @llvm.ppc.altivec.vminsd in test1()
979 // CHECK-LE: @llvm.ppc.altivec.vminud in test1()
984 // CHECK-LE: @llvm.ppc.altivec.vminud in test1()
989 // CHECK-LE: @llvm.ppc.altivec.vminud in test1()
995 // CHECK-LE: @llvm.ppc.altivec.vmulosw in test1()
1000 // CHECK-LE: @llvm.ppc.altivec.vmulouw in test1()
1006 // CHECK-LE: @llvm.ppc.altivec.vmulesw in test1()
1011 // CHECK-LE: @llvm.ppc.altivec.vmuleuw in test1()
1017 // CHECK-LE: @llvm.ppc.altivec.vpksdss in test1()
1022 // CHECK-LE: @llvm.ppc.altivec.vpkudus in test1()
1028 // CHECK-LE: @llvm.ppc.altivec.vpksdus in test1()
1033 // CHECK-LE: @llvm.ppc.altivec.vpkudus in test1()
1039 // CHECK-LE: @llvm.ppc.altivec.vrld in test1()
1044 // CHECK-LE: @llvm.ppc.altivec.vrld in test1()
1050 // CHECK-LE: shl <2 x i64> in test1()
1055 // CHECK-LE: shl <2 x i64> in test1()
1061 // CHECK-LE: lshr <2 x i64> in test1()
1066 // CHECK-LE: lshr <2 x i64> in test1()
1072 // CHECK-LE: ashr <2 x i64> in test1()
1077 // CHECK-LE: ashr <2 x i64> in test1()
1083 // CHECK-LE: insertelement <2 x i64> in test1()
1087 // CHECK-LE: insertelement <2 x i64> in test1()
1091 // CHECK-LE: insertelement <1 x i128> in test1()
1095 // CHECK-LE: insertelement <1 x i128> in test1()
1099 // CHECK-LE: insertelement <2 x double> in test1()
1105 // CHECK-LE: llvm.ppc.altivec.vupklsw in test1()
1110 // CHECK-LE: llvm.ppc.altivec.vupklsw in test1()
1116 // CHECK-LE: llvm.ppc.altivec.vupkhsw in test1()
1121 // CHECK-LE: llvm.ppc.altivec.vupkhsw in test1()
1127 // CHECK-LE: llvm.ppc.altivec.vpksdss in test1()
1133 // CHECK-LE: llvm.ppc.altivec.vpksdus in test1()
1139 // CHECK-LE: vperm in test1()
1144 // CHECK-LE: vperm in test1()
1148 // CHECK-LE: llvm.ppc.altivec.vpkudus in test1()
1154 // CHECK-LE: llvm.ppc.altivec.vupklsw in test1()
1159 // CHECK-LE: llvm.ppc.altivec.vupklsw in test1()
1164 // CHECK-LE: llvm.ppc.altivec.vupkhsw in test1()
1169 // CHECK-LE: llvm.ppc.altivec.vupkhsw in test1()
1174 // CHECK-LE: @llvm.ppc.altivec.vmaxsd in test1()
1178 // CHECK-LE: @llvm.ppc.altivec.vmaxsd in test1()
1182 // CHECK-LE: @llvm.ppc.altivec.vmaxsd in test1()
1186 // CHECK-LE: @llvm.ppc.altivec.vmaxud in test1()
1190 // CHECK-LE: @llvm.ppc.altivec.vmaxud in test1()
1195 // CHECK-LE: @llvm.ppc.altivec.vminsd in test1()
1199 // CHECK-LE: @llvm.ppc.altivec.vminsd in test1()
1203 // CHECK-LE: @llvm.ppc.altivec.vminsd in test1()
1207 // CHECK-LE: @llvm.ppc.altivec.vminud in test1()
1211 // CHECK-LE: @llvm.ppc.altivec.vminud in test1()
1217 // CHECK-LE: [[T1:%.+]] = and <16 x i8> in test1()
1218 // CHECK-LE: xor <16 x i8> [[T1]], <i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, … in test1()
1224 // CHECK-LE: [[T1:%.+]] = and <16 x i8> in test1()
1225 // CHECK-LE: xor <16 x i8> [[T1]], <i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, … in test1()
1230 // CHECK-LE: [[T1:%.+]] = and <16 x i8> in test1()
1231 // CHECK-LE: xor <16 x i8> [[T1]], <i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, … in test1()
1236 // CHECK-LE: [[T1:%.+]] = and <8 x i16> in test1()
1237 // CHECK-LE: xor <8 x i16> [[T1]], <i16 -1, i16 -1, i16 -1, i16 -1, i16 -1, i16 -1, i16 -1, i16 -1> in test1()
1242 // CHECK-LE: [[T1:%.+]] = and <8 x i16> in test1()
1243 // CHECK-LE: xor <8 x i16> [[T1]], <i16 -1, i16 -1, i16 -1, i16 -1, i16 -1, i16 -1, i16 -1, i16 -1> in test1()
1248 // CHECK-LE: [[T1:%.+]] = and <8 x i16> in test1()
1249 // CHECK-LE: xor <8 x i16> [[T1]], <i16 -1, i16 -1, i16 -1, i16 -1, i16 -1, i16 -1, i16 -1, i16 -1> in test1()
1254 // CHECK-LE: [[T1:%.+]] = and <4 x i32> in test1()
1255 // CHECK-LE: xor <4 x i32> [[T1]], <i32 -1, i32 -1, i32 -1, i32 -1> in test1()
1260 // CHECK-LE: [[T1:%.+]] = and <4 x i32> in test1()
1261 // CHECK-LE: xor <4 x i32> [[T1]], <i32 -1, i32 -1, i32 -1, i32 -1> in test1()
1266 // CHECK-LE: [[T1:%.+]] = and <4 x i32> in test1()
1267 // CHECK-LE: xor <4 x i32> [[T1]], <i32 -1, i32 -1, i32 -1, i32 -1> in test1()
1272 // CHECK-LE: [[T1:%.+]] = and <2 x i64> in test1()
1273 // CHECK-LE: xor <2 x i64> [[T1]], <i64 -1, i64 -1> in test1()
1278 // CHECK-LE: [[T1:%.+]] = and <2 x i64> in test1()
1279 // CHECK-LE: xor <2 x i64> [[T1]], <i64 -1, i64 -1> in test1()
1284 // CHECK-LE: [[T1:%.+]] = and <2 x i64> in test1()
1285 // CHECK-LE: xor <2 x i64> [[T1]], <i64 -1, i64 -1> in test1()
1291 // CHECK-LE: [[T1:%.+]] = xor <16 x i8> {{%.+}}, <i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, … in test1()
1292 // CHECK-LE: or <16 x i8> {{%.+}}, [[T1]] in test1()
1298 // CHECK-LE: [[T1:%.+]] = xor <16 x i8> {{%.+}}, <i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, … in test1()
1299 // CHECK-LE: or <16 x i8> {{%.+}}, [[T1]] in test1()
1304 // CHECK-LE: [[T1:%.+]] = xor <16 x i8> {{%.+}}, <i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, … in test1()
1305 // CHECK-LE: or <16 x i8> {{%.+}}, [[T1]] in test1()
1310 // CHECK-LE: [[T1:%.+]] = xor <16 x i8> {{%.+}}, <i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, … in test1()
1311 // CHECK-LE: or <16 x i8> {{%.+}}, [[T1]] in test1()
1316 // CHECK-LE: [[T1:%.+]] = xor <16 x i8> {{%.+}}, <i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, … in test1()
1317 // CHECK-LE: or <16 x i8> {{%.+}}, [[T1]] in test1()
1322 // CHECK-LE: [[T1:%.+]] = xor <16 x i8> {{%.+}}, <i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, … in test1()
1323 // CHECK-LE: or <16 x i8> {{%.+}}, [[T1]] in test1()
1328 // CHECK-LE: [[T1:%.+]] = xor <16 x i8> {{%.+}}, <i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, … in test1()
1329 // CHECK-LE: or <16 x i8> {{%.+}}, [[T1]] in test1()
1334 // CHECK-LE: [[T1:%.+]] = xor <8 x i16> {{%.+}}, <i16 -1, i16 -1, i16 -1, i16 -1, i16 -1, i16 -1, i… in test1()
1335 // CHECK-LE: or <8 x i16> {{%.+}}, [[T1]] in test1()
1340 // CHECK-LE: [[T1:%.+]] = xor <8 x i16> {{%.+}}, <i16 -1, i16 -1, i16 -1, i16 -1, i16 -1, i16 -1, i… in test1()
1341 // CHECK-LE: or <8 x i16> {{%.+}}, [[T1]] in test1()
1346 // CHECK-LE: [[T1:%.+]] = xor <8 x i16> {{%.+}}, <i16 -1, i16 -1, i16 -1, i16 -1, i16 -1, i16 -1, i… in test1()
1347 // CHECK-LE: or <8 x i16> {{%.+}}, [[T1]] in test1()
1352 // CHECK-LE: [[T1:%.+]] = xor <8 x i16> {{%.+}}, <i16 -1, i16 -1, i16 -1, i16 -1, i16 -1, i16 -1, i… in test1()
1353 // CHECK-LE: or <8 x i16> {{%.+}}, [[T1]] in test1()
1358 // CHECK-LE: [[T1:%.+]] = xor <8 x i16> {{%.+}}, <i16 -1, i16 -1, i16 -1, i16 -1, i16 -1, i16 -1, i… in test1()
1359 // CHECK-LE: or <8 x i16> {{%.+}}, [[T1]] in test1()
1364 // CHECK-LE: [[T1:%.+]] = xor <8 x i16> {{%.+}}, <i16 -1, i16 -1, i16 -1, i16 -1, i16 -1, i16 -1, i… in test1()
1365 // CHECK-LE: or <8 x i16> {{%.+}}, [[T1]] in test1()
1370 // CHECK-LE: [[T1:%.+]] = xor <8 x i16> {{%.+}}, <i16 -1, i16 -1, i16 -1, i16 -1, i16 -1, i16 -1, i… in test1()
1371 // CHECK-LE: or <8 x i16> {{%.+}}, [[T1]] in test1()
1376 // CHECK-LE: [[T1:%.+]] = xor <4 x i32> {{%.+}}, <i32 -1, i32 -1, i32 -1, i32 -1> in test1()
1377 // CHECK-LE: or <4 x i32> {{%.+}}, [[T1]] in test1()
1382 // CHECK-LE: [[T1:%.+]] = xor <4 x i32> {{%.+}}, <i32 -1, i32 -1, i32 -1, i32 -1> in test1()
1383 // CHECK-LE: or <4 x i32> {{%.+}}, [[T1]] in test1()
1388 // CHECK-LE: [[T1:%.+]] = xor <4 x i32> {{%.+}}, <i32 -1, i32 -1, i32 -1, i32 -1> in test1()
1389 // CHECK-LE: or <4 x i32> {{%.+}}, [[T1]] in test1()
1394 // CHECK-LE: [[T1:%.+]] = xor <4 x i32> {{%.+}}, <i32 -1, i32 -1, i32 -1, i32 -1> in test1()
1395 // CHECK-LE: or <4 x i32> {{%.+}}, [[T1]] in test1()
1400 // CHECK-LE: [[T1:%.+]] = xor <4 x i32> {{%.+}}, <i32 -1, i32 -1, i32 -1, i32 -1> in test1()
1401 // CHECK-LE: or <4 x i32> {{%.+}}, [[T1]] in test1()
1406 // CHECK-LE: [[T1:%.+]] = xor <4 x i32> {{%.+}}, <i32 -1, i32 -1, i32 -1, i32 -1> in test1()
1407 // CHECK-LE: or <4 x i32> {{%.+}}, [[T1]] in test1()
1412 // CHECK-LE: [[T1:%.+]] = xor <4 x i32> {{%.+}}, <i32 -1, i32 -1, i32 -1, i32 -1> in test1()
1413 // CHECK-LE: or <4 x i32> {{%.+}}, [[T1]] in test1()
1418 // CHECK-LE: [[T1:%.+]] = xor <2 x i64> {{%.+}}, <i64 -1, i64 -1> in test1()
1419 // CHECK-LE: or <2 x i64> {{%.+}}, [[T1]] in test1()
1424 // CHECK-LE: [[T1:%.+]] = xor <2 x i64> {{%.+}}, <i64 -1, i64 -1> in test1()
1425 // CHECK-LE: or <2 x i64> {{%.+}}, [[T1]] in test1()
1430 // CHECK-LE: [[T1:%.+]] = xor <2 x i64> {{%.+}}, <i64 -1, i64 -1> in test1()
1431 // CHECK-LE: or <2 x i64> {{%.+}}, [[T1]] in test1()
1436 // CHECK-LE: [[T1:%.+]] = xor <2 x i64> {{%.+}}, <i64 -1, i64 -1> in test1()
1437 // CHECK-LE: or <2 x i64> {{%.+}}, [[T1]] in test1()
1442 // CHECK-LE: [[T1:%.+]] = xor <2 x i64> {{%.+}}, <i64 -1, i64 -1> in test1()
1443 // CHECK-LE: or <2 x i64> {{%.+}}, [[T1]] in test1()
1448 // CHECK-LE: [[T1:%.+]] = xor <2 x i64> {{%.+}}, <i64 -1, i64 -1> in test1()
1449 // CHECK-LE: or <2 x i64> {{%.+}}, [[T1]] in test1()
1454 // CHECK-LE: [[T1:%.+]] = xor <2 x i64> {{%.+}}, <i64 -1, i64 -1> in test1()
1455 // CHECK-LE: or <2 x i64> {{%.+}}, [[T1]] in test1()
1460 // CHECK-LE: sub <2 x i64> in test1()
1464 // CHECK-LE: sub <2 x i64> in test1()
1468 // CHECK-LE: fsub <2 x double> in test1()
1472 // CHECK-LE: sub <1 x i128> in test1()
1476 // CHECK-LE: sub <1 x i128> in test1()
1481 // CHECK-LE: llvm.ppc.altivec.vbpermq in test1()
1485 // CHECK-LE: llvm.ppc.altivec.vbpermq in test1()
1491 // CHECK-LE: llvm.ppc.altivec.vgbbd in test1()
1495 // CHECK-LE: llvm.ppc.altivec.vgbbd in test1()
1500 // CHECK-LE: llvm.ppc.altivec.vgbbd in test1()
1505 // CHECK-LE: llvm.ppc.altivec.vbpermq in test1()