Lines Matching refs:vD
170 : VAForm_1a<xo, (outs VRRC:$vD), (ins VRRC:$vA, VRRC:$vB, VRRC:$vC),
171 !strconcat(opc, " $vD, $vA, $vB, $vC"), VecFP,
172 [(set VRRC:$vD, (IntID VRRC:$vA, VRRC:$vB, VRRC:$vC))]>;
176 : VXForm_1<xo, (outs VRRC:$vD), (ins VRRC:$vA, VRRC:$vB),
177 !strconcat(opc, " $vD, $vA, $vB"), VecFP,
178 [(set VRRC:$vD, (IntID VRRC:$vA, VRRC:$vB))]>;
182 : VXForm_2<xo, (outs VRRC:$vD), (ins VRRC:$vB),
183 !strconcat(opc, " $vD, $vB"), VecFP,
184 [(set VRRC:$vD, (IntID VRRC:$vB))]>;
221 def MFVSCR : VXForm_4<1540, (outs VRRC:$vD), (ins),
222 "mfvscr $vD", LdStStore,
223 [(set VRRC:$vD, (int_ppc_altivec_mfvscr))]>;
229 def LVEBX: XForm_1<31, 7, (outs VRRC:$vD), (ins memrr:$src),
230 "lvebx $vD, $src", LdStLoad,
231 [(set VRRC:$vD, (int_ppc_altivec_lvebx xoaddr:$src))]>;
232 def LVEHX: XForm_1<31, 39, (outs VRRC:$vD), (ins memrr:$src),
233 "lvehx $vD, $src", LdStLoad,
234 [(set VRRC:$vD, (int_ppc_altivec_lvehx xoaddr:$src))]>;
235 def LVEWX: XForm_1<31, 71, (outs VRRC:$vD), (ins memrr:$src),
236 "lvewx $vD, $src", LdStLoad,
237 [(set VRRC:$vD, (int_ppc_altivec_lvewx xoaddr:$src))]>;
238 def LVX : XForm_1<31, 103, (outs VRRC:$vD), (ins memrr:$src),
239 "lvx $vD, $src", LdStLoad,
240 [(set VRRC:$vD, (int_ppc_altivec_lvx xoaddr:$src))]>;
241 def LVXL : XForm_1<31, 359, (outs VRRC:$vD), (ins memrr:$src),
242 "lvxl $vD, $src", LdStLoad,
243 [(set VRRC:$vD, (int_ppc_altivec_lvxl xoaddr:$src))]>;
246 def LVSL : XForm_1<31, 6, (outs VRRC:$vD), (ins memrr:$src),
247 "lvsl $vD, $src", LdStLoad,
248 [(set VRRC:$vD, (int_ppc_altivec_lvsl xoaddr:$src))]>,
250 def LVSR : XForm_1<31, 38, (outs VRRC:$vD), (ins memrr:$src),
251 "lvsr $vD, $src", LdStLoad,
252 [(set VRRC:$vD, (int_ppc_altivec_lvsr xoaddr:$src))]>,
275 def VMADDFP : VAForm_1<46, (outs VRRC:$vD), (ins VRRC:$vA, VRRC:$vC, VRRC:$vB),
276 "vmaddfp $vD, $vA, $vC, $vB", VecFP,
277 [(set VRRC:$vD, (fadd (fmul VRRC:$vA, VRRC:$vC),
280 def VNMSUBFP: VAForm_1<47, (outs VRRC:$vD), (ins VRRC:$vA, VRRC:$vC, VRRC:$vB),
281 "vnmsubfp $vD, $vA, $vC, $vB", VecFP,
282 [(set VRRC:$vD, (fsub V_immneg0,
294 def VSLDOI : VAForm_2<44, (outs VRRC:$vD), (ins VRRC:$vA, VRRC:$vB, u5imm:$SH),
295 "vsldoi $vD, $vA, $vB, $SH", VecFP,
296 [(set VRRC:$vD,
300 def VADDFP : VXForm_1<10, (outs VRRC:$vD), (ins VRRC:$vA, VRRC:$vB),
301 "vaddfp $vD, $vA, $vB", VecFP,
302 [(set VRRC:$vD, (fadd VRRC:$vA, VRRC:$vB))]>;
304 def VADDUBM : VXForm_1<0, (outs VRRC:$vD), (ins VRRC:$vA, VRRC:$vB),
305 "vaddubm $vD, $vA, $vB", VecGeneral,
306 [(set VRRC:$vD, (add (v16i8 VRRC:$vA), VRRC:$vB))]>;
307 def VADDUHM : VXForm_1<64, (outs VRRC:$vD), (ins VRRC:$vA, VRRC:$vB),
308 "vadduhm $vD, $vA, $vB", VecGeneral,
309 [(set VRRC:$vD, (add (v8i16 VRRC:$vA), VRRC:$vB))]>;
310 def VADDUWM : VXForm_1<128, (outs VRRC:$vD), (ins VRRC:$vA, VRRC:$vB),
311 "vadduwm $vD, $vA, $vB", VecGeneral,
312 [(set VRRC:$vD, (add (v4i32 VRRC:$vA), VRRC:$vB))]>;
323 def VAND : VXForm_1<1028, (outs VRRC:$vD), (ins VRRC:$vA, VRRC:$vB),
324 "vand $vD, $vA, $vB", VecFP,
325 [(set VRRC:$vD, (and (v4i32 VRRC:$vA), VRRC:$vB))]>;
326 def VANDC : VXForm_1<1092, (outs VRRC:$vD), (ins VRRC:$vA, VRRC:$vB),
327 "vandc $vD, $vA, $vB", VecFP,
328 [(set VRRC:$vD, (and (v4i32 VRRC:$vA),
331 def VCFSX : VXForm_1<842, (outs VRRC:$vD), (ins u5imm:$UIMM, VRRC:$vB),
332 "vcfsx $vD, $vB, $UIMM", VecFP,
333 [(set VRRC:$vD,
335 def VCFUX : VXForm_1<778, (outs VRRC:$vD), (ins u5imm:$UIMM, VRRC:$vB),
336 "vcfux $vD, $vB, $UIMM", VecFP,
337 [(set VRRC:$vD,
339 def VCTSXS : VXForm_1<970, (outs VRRC:$vD), (ins u5imm:$UIMM, VRRC:$vB),
340 "vctsxs $vD, $vB, $UIMM", VecFP,
341 [(set VRRC:$vD,
343 def VCTUXS : VXForm_1<906, (outs VRRC:$vD), (ins u5imm:$UIMM, VRRC:$vB),
344 "vctuxs $vD, $vB, $UIMM", VecFP,
345 [(set VRRC:$vD,
372 def VMRGHB : VXForm_1< 12, (outs VRRC:$vD), (ins VRRC:$vA, VRRC:$vB),
373 "vmrghb $vD, $vA, $vB", VecFP,
374 [(set VRRC:$vD, (vmrghb_shuffle VRRC:$vA, VRRC:$vB))]>;
375 def VMRGHH : VXForm_1< 76, (outs VRRC:$vD), (ins VRRC:$vA, VRRC:$vB),
376 "vmrghh $vD, $vA, $vB", VecFP,
377 [(set VRRC:$vD, (vmrghh_shuffle VRRC:$vA, VRRC:$vB))]>;
378 def VMRGHW : VXForm_1<140, (outs VRRC:$vD), (ins VRRC:$vA, VRRC:$vB),
379 "vmrghw $vD, $vA, $vB", VecFP,
380 [(set VRRC:$vD, (vmrghw_shuffle VRRC:$vA, VRRC:$vB))]>;
381 def VMRGLB : VXForm_1<268, (outs VRRC:$vD), (ins VRRC:$vA, VRRC:$vB),
382 "vmrglb $vD, $vA, $vB", VecFP,
383 [(set VRRC:$vD, (vmrglb_shuffle VRRC:$vA, VRRC:$vB))]>;
384 def VMRGLH : VXForm_1<332, (outs VRRC:$vD), (ins VRRC:$vA, VRRC:$vB),
385 "vmrglh $vD, $vA, $vB", VecFP,
386 [(set VRRC:$vD, (vmrglh_shuffle VRRC:$vA, VRRC:$vB))]>;
387 def VMRGLW : VXForm_1<396, (outs VRRC:$vD), (ins VRRC:$vA, VRRC:$vB),
388 "vmrglw $vD, $vA, $vB", VecFP,
389 [(set VRRC:$vD, (vmrglw_shuffle VRRC:$vA, VRRC:$vB))]>;
416 def VSUBFP : VXForm_1<74, (outs VRRC:$vD), (ins VRRC:$vA, VRRC:$vB),
417 "vsubfp $vD, $vA, $vB", VecGeneral,
418 [(set VRRC:$vD, (fsub VRRC:$vA, VRRC:$vB))]>;
419 def VSUBUBM : VXForm_1<1024, (outs VRRC:$vD), (ins VRRC:$vA, VRRC:$vB),
420 "vsububm $vD, $vA, $vB", VecGeneral,
421 [(set VRRC:$vD, (sub (v16i8 VRRC:$vA), VRRC:$vB))]>;
422 def VSUBUHM : VXForm_1<1088, (outs VRRC:$vD), (ins VRRC:$vA, VRRC:$vB),
423 "vsubuhm $vD, $vA, $vB", VecGeneral,
424 [(set VRRC:$vD, (sub (v8i16 VRRC:$vA), VRRC:$vB))]>;
425 def VSUBUWM : VXForm_1<1152, (outs VRRC:$vD), (ins VRRC:$vA, VRRC:$vB),
426 "vsubuwm $vD, $vA, $vB", VecGeneral,
427 [(set VRRC:$vD, (sub (v4i32 VRRC:$vA), VRRC:$vB))]>;
441 def VNOR : VXForm_1<1284, (outs VRRC:$vD), (ins VRRC:$vA, VRRC:$vB),
442 "vnor $vD, $vA, $vB", VecFP,
443 [(set VRRC:$vD, (vnot_ppc (or (v4i32 VRRC:$vA),
445 def VOR : VXForm_1<1156, (outs VRRC:$vD), (ins VRRC:$vA, VRRC:$vB),
446 "vor $vD, $vA, $vB", VecFP,
447 [(set VRRC:$vD, (or (v4i32 VRRC:$vA), VRRC:$vB))]>;
448 def VXOR : VXForm_1<1220, (outs VRRC:$vD), (ins VRRC:$vA, VRRC:$vB),
449 "vxor $vD, $vA, $vB", VecFP,
450 [(set VRRC:$vD, (xor (v4i32 VRRC:$vA), VRRC:$vB))]>;
462 def VSPLTB : VXForm_1<524, (outs VRRC:$vD), (ins u5imm:$UIMM, VRRC:$vB),
463 "vspltb $vD, $vB, $UIMM", VecPerm,
464 [(set VRRC:$vD,
466 def VSPLTH : VXForm_1<588, (outs VRRC:$vD), (ins u5imm:$UIMM, VRRC:$vB),
467 "vsplth $vD, $vB, $UIMM", VecPerm,
468 [(set VRRC:$vD,
470 def VSPLTW : VXForm_1<652, (outs VRRC:$vD), (ins u5imm:$UIMM, VRRC:$vB),
471 "vspltw $vD, $vB, $UIMM", VecPerm,
472 [(set VRRC:$vD,
485 def VSPLTISB : VXForm_3<780, (outs VRRC:$vD), (ins s5imm:$SIMM),
486 "vspltisb $vD, $SIMM", VecPerm,
487 [(set VRRC:$vD, (v16i8 vecspltisb:$SIMM))]>;
488 def VSPLTISH : VXForm_3<844, (outs VRRC:$vD), (ins s5imm:$SIMM),
489 "vspltish $vD, $SIMM", VecPerm,
490 [(set VRRC:$vD, (v8i16 vecspltish:$SIMM))]>;
491 def VSPLTISW : VXForm_3<908, (outs VRRC:$vD), (ins s5imm:$SIMM),
492 "vspltisw $vD, $SIMM", VecPerm,
493 [(set VRRC:$vD, (v4i32 vecspltisw:$SIMM))]>;
501 def VPKUHUM : VXForm_1<14, (outs VRRC:$vD), (ins VRRC:$vA, VRRC:$vB),
502 "vpkuhum $vD, $vA, $vB", VecFP,
503 [(set VRRC:$vD,
506 def VPKUWUM : VXForm_1<78, (outs VRRC:$vD), (ins VRRC:$vA, VRRC:$vB),
507 "vpkuwum $vD, $vA, $vB", VecFP,
508 [(set VRRC:$vD,
524 : VXRForm_1<xo, (outs VRRC:$vD), (ins VRRC:$vA, VRRC:$vB),asmstr,VecFPCompare,
525 [(set VRRC:$vD, (Ty (PPCvcmp VRRC:$vA, VRRC:$vB, xo)))]>;
527 : VXRForm_1<xo, (outs VRRC:$vD), (ins VRRC:$vA, VRRC:$vB),asmstr,VecFPCompare,
528 [(set VRRC:$vD, (Ty (PPCvcmp_o VRRC:$vA, VRRC:$vB, xo)))]> {
534 def VCMPBFP : VCMP <966, "vcmpbfp $vD, $vA, $vB" , v4f32>;
535 def VCMPBFPo : VCMPo<966, "vcmpbfp. $vD, $vA, $vB" , v4f32>;
536 def VCMPEQFP : VCMP <198, "vcmpeqfp $vD, $vA, $vB" , v4f32>;
537 def VCMPEQFPo : VCMPo<198, "vcmpeqfp. $vD, $vA, $vB", v4f32>;
538 def VCMPGEFP : VCMP <454, "vcmpgefp $vD, $vA, $vB" , v4f32>;
539 def VCMPGEFPo : VCMPo<454, "vcmpgefp. $vD, $vA, $vB", v4f32>;
540 def VCMPGTFP : VCMP <710, "vcmpgtfp $vD, $vA, $vB" , v4f32>;
541 def VCMPGTFPo : VCMPo<710, "vcmpgtfp. $vD, $vA, $vB", v4f32>;
544 def VCMPEQUB : VCMP < 6, "vcmpequb $vD, $vA, $vB" , v16i8>;
545 def VCMPEQUBo : VCMPo< 6, "vcmpequb. $vD, $vA, $vB", v16i8>;
546 def VCMPGTSB : VCMP <774, "vcmpgtsb $vD, $vA, $vB" , v16i8>;
547 def VCMPGTSBo : VCMPo<774, "vcmpgtsb. $vD, $vA, $vB", v16i8>;
548 def VCMPGTUB : VCMP <518, "vcmpgtub $vD, $vA, $vB" , v16i8>;
549 def VCMPGTUBo : VCMPo<518, "vcmpgtub. $vD, $vA, $vB", v16i8>;
552 def VCMPEQUH : VCMP < 70, "vcmpequh $vD, $vA, $vB" , v8i16>;
553 def VCMPEQUHo : VCMPo< 70, "vcmpequh. $vD, $vA, $vB", v8i16>;
554 def VCMPGTSH : VCMP <838, "vcmpgtsh $vD, $vA, $vB" , v8i16>;
555 def VCMPGTSHo : VCMPo<838, "vcmpgtsh. $vD, $vA, $vB", v8i16>;
556 def VCMPGTUH : VCMP <582, "vcmpgtuh $vD, $vA, $vB" , v8i16>;
557 def VCMPGTUHo : VCMPo<582, "vcmpgtuh. $vD, $vA, $vB", v8i16>;
560 def VCMPEQUW : VCMP <134, "vcmpequw $vD, $vA, $vB" , v4i32>;
561 def VCMPEQUWo : VCMPo<134, "vcmpequw. $vD, $vA, $vB", v4i32>;
562 def VCMPGTSW : VCMP <902, "vcmpgtsw $vD, $vA, $vB" , v4i32>;
563 def VCMPGTSWo : VCMPo<902, "vcmpgtsw. $vD, $vA, $vB", v4i32>;
564 def VCMPGTUW : VCMP <646, "vcmpgtuw $vD, $vA, $vB" , v4i32>;
565 def VCMPGTUWo : VCMPo<646, "vcmpgtuw. $vD, $vA, $vB", v4i32>;
567 def V_SET0 : VXForm_setzero<1220, (outs VRRC:$vD), (ins),
568 "vxor $vD, $vD, $vD", VecFP,
569 [(set VRRC:$vD, (v4i32 immAllZerosV))]>;