Home
last modified time | relevance | path

Searched refs:vsx (Results 1 – 25 of 279) sorted by relevance

12345678910>>...12

/external/llvm-project/llvm/test/CodeGen/PowerPC/
Dbuiltins-ppc-elf2-abi.ll22 %2 = call <2 x double> @llvm.ppc.vsx.xvdivdp(<2 x double> %0, <2 x double> %1)
34 %2 = call <4 x float> @llvm.ppc.vsx.xvdivsp(<4 x float> %0, <4 x float> %1)
70 %2 = call <2 x i64> @llvm.ppc.vsx.xvcmpeqdp(<2 x double> %0, <2 x double> %1)
82 %2 = call <4 x i32> @llvm.ppc.vsx.xvcmpeqsp(<4 x float> %0, <4 x float> %1)
94 %2 = call <2 x i64> @llvm.ppc.vsx.xvcmpgedp(<2 x double> %0, <2 x double> %1)
106 %2 = call <4 x i32> @llvm.ppc.vsx.xvcmpgesp(<4 x float> %0, <4 x float> %1)
118 %2 = call <2 x i64> @llvm.ppc.vsx.xvcmpgtdp(<2 x double> %0, <2 x double> %1)
130 %2 = call <4 x i32> @llvm.ppc.vsx.xvcmpgtsp(<4 x float> %0, <4 x float> %1)
143 %1 = call <4 x float> @llvm.ppc.vsx.xvresp(<4 x float> %0)
155 %1 = call <2 x double> @llvm.ppc.vsx.xvredp(<2 x double> %0)
[all …]
Dvsx-minmax.ll1 ; RUN: llc -verify-machineinstrs -mcpu=pwr7 -mattr=+vsx -O0 -fast-isel=0 -mtriple=powerpc64-unknown…
23 %2 = tail call <4 x float> @llvm.ppc.vsx.xvmaxsp(<4 x float> %0, <4 x float> %1)
27 %4 = tail call <2 x double> @llvm.ppc.vsx.xvmaxdp(<2 x double> %3, <2 x double> %3)
32 %7 = tail call <4 x float> @llvm.ppc.vsx.xvmaxsp(<4 x float> %5, <4 x float> %6)
37 %10 = tail call <4 x float> @llvm.ppc.vsx.xvminsp(<4 x float> %8, <4 x float> %9)
41 %12 = tail call <2 x double> @llvm.ppc.vsx.xvmindp(<2 x double> %11, <2 x double> %11)
46 %15 = tail call <4 x float> @llvm.ppc.vsx.xvminsp(<4 x float> %13, <4 x float> %14)
50 %17 = tail call double @llvm.ppc.vsx.xsmaxdp(double %16, double %16)
53 %18 = tail call double @llvm.ppc.vsx.xsmindp(double %16, double %16)
60 declare double @llvm.ppc.vsx.xsmaxdp(double, double)
[all …]
Dvsx_builtins.ll13 %0 = tail call <4 x i32> @llvm.ppc.vsx.lxvw4x.be(i8* %a)
17 declare <4 x i32> @llvm.ppc.vsx.lxvw4x.be(i8*)
26 %0 = tail call <2 x double> @llvm.ppc.vsx.lxvd2x.be(i8* %a)
30 declare <2 x double> @llvm.ppc.vsx.lxvd2x.be(i8*)
39 tail call void @llvm.ppc.vsx.stxvw4x.be(<4 x i32> %a, i8* %b)
43 declare void @llvm.ppc.vsx.stxvw4x.be(<4 x i32>, i8*)
52 tail call void @llvm.ppc.vsx.stxvd2x.be(<2 x double> %a, i8* %b)
56 declare void @llvm.ppc.vsx.stxvd2x.be(<2 x double>, i8*)
66 %0 = tail call i32 @llvm.ppc.vsx.xvtdivdp(<2 x double> %a, <2 x double> %b)
69 declare i32 @llvm.ppc.vsx.xvtdivdp(<2 x double>, <2 x double>)
[all …]
Di64_fp.ll4 ; RUN: llc -verify-machineinstrs < %s -mattr=-vsx -mtriple=ppc32-- -mattr=+64bit | \
6 ; RUN: llc -verify-machineinstrs < %s -mattr=-vsx -mtriple=ppc32-- -mattr=+64bit | \
8 ; RUN: llc -verify-machineinstrs < %s -mattr=-vsx -mtriple=ppc32-- -mcpu=g5 | \
10 ; RUN: llc -verify-machineinstrs < %s -mattr=-vsx -mtriple=ppc32-- -mcpu=g5 | \
12 ; RUN: llc -verify-machineinstrs < %s -mattr=-vsx -mtriple=ppc32-- -mattr=-64bit | \
14 ; RUN: llc -verify-machineinstrs < %s -mattr=-vsx -mtriple=ppc32-- -mattr=-64bit | \
16 ; RUN: llc -verify-machineinstrs < %s -mattr=-vsx -mtriple=ppc32-- -mcpu=g4 | \
18 ; RUN: llc -verify-machineinstrs < %s -mattr=-vsx -mtriple=ppc32-- -mcpu=g4 | \
Dvsx-ldst-builtin-le.ll1 ; RUN: llc -verify-machineinstrs -mcpu=pwr8 -mattr=+vsx -O2 \
30 %0 = call <4 x i32> @llvm.ppc.vsx.lxvw4x(i8* bitcast (<4 x i32>* @vsi to i8*))
36 %1 = call <4 x i32> @llvm.ppc.vsx.lxvw4x(i8* bitcast (<4 x i32>* @vui to i8*))
42 %2 = call <4 x i32> @llvm.ppc.vsx.lxvw4x(i8* bitcast (<4 x float>* @vf to i8*))
49 %4 = call <2 x double> @llvm.ppc.vsx.lxvd2x(i8* bitcast (<2 x i64>* @vsll to i8*))
56 %6 = call <2 x double> @llvm.ppc.vsx.lxvd2x(i8* bitcast (<2 x i64>* @vull to i8*))
63 %8 = call <2 x double> @llvm.ppc.vsx.lxvd2x(i8* bitcast (<2 x double>* @vd to i8*))
72 call void @llvm.ppc.vsx.stxvw4x(<4 x i32> %9, i8* bitcast (<4 x i32>* @res_vsi to i8*))
78 call void @llvm.ppc.vsx.stxvw4x(<4 x i32> %10, i8* bitcast (<4 x i32>* @res_vui to i8*))
85 call void @llvm.ppc.vsx.stxvw4x(<4 x i32> %12, i8* bitcast (<4 x float>* @res_vf to i8*))
[all …]
Dp10-vsx-pcv.ll8 declare <16 x i8> @llvm.ppc.vsx.xxgenpcvbm(<16 x i8>, i32)
9 declare <8 x i16> @llvm.ppc.vsx.xxgenpcvhm(<8 x i16>, i32)
10 declare <4 x i32> @llvm.ppc.vsx.xxgenpcvwm(<4 x i32>, i32)
11 declare <2 x i64> @llvm.ppc.vsx.xxgenpcvdm(<2 x i64>, i32)
19 %gen = tail call <16 x i8> @llvm.ppc.vsx.xxgenpcvbm(<16 x i8> %a, i32 1)
29 %gen = tail call <8 x i16> @llvm.ppc.vsx.xxgenpcvhm(<8 x i16> %a, i32 1)
39 %gen = tail call <4 x i32> @llvm.ppc.vsx.xxgenpcvwm(<4 x i32> %a, i32 1)
49 %gen = tail call <2 x i64> @llvm.ppc.vsx.xxgenpcvdm(<2 x i64> %a, i32 1)
Dppc-passname.ll25 ; Test pass name: ppc-vsx-copy.
26 ; RUN: llc -mtriple=powerpc64le-unknown-unknown < %s -debug-pass=Structure -stop-before=ppc-vsx-cop…
27 ; STOP-BEFORE-VSX-COPY-NOT: -ppc-vsx-copy
28 ; STOP-BEFORE-VSX-COPY-NOT: "ppc-vsx-copy" pass is not registered.
31 ; RUN: llc -mtriple=powerpc64le-unknown-unknown < %s -debug-pass=Structure -stop-after=ppc-vsx-copy…
32 ; STOP-AFTER-VSX-COPY: -ppc-vsx-copy
33 ; STOP-AFTER-VSX-COPY-NOT: "ppc-vsx-copy" pass is not registered.
49 ; Test pass name: ppc-vsx-fma-mutate.
50 ; RUN: llc -mtriple=powerpc64le-unknown-unknown < %s -debug-pass=Structure -stop-before=ppc-vsx-fma…
51 ; STOP-BEFORE-VSX-FMA-MUTATE-NOT: -ppc-vsx-fma-mutate
[all …]
Dvsx-p9.ll149 %0 = tail call <4 x float> @llvm.ppc.vsx.xviexpsp(<4 x i32> %a, <4 x i32> %b)
156 declare <4 x float> @llvm.ppc.vsx.xviexpsp(<4 x i32>, <4 x i32>)
161 %0 = tail call <2 x double> @llvm.ppc.vsx.xviexpdp(<2 x i64> %a, <2 x i64> %b)
168 declare <2 x double> @llvm.ppc.vsx.xviexpdp(<2 x i64>, <2 x i64>)
199 %0 = tail call <4 x float> @llvm.ppc.vsx.xvcvsphp(<4 x float> %a)
249 declare <4 x float> @llvm.ppc.vsx.xvcvsphp(<4 x float>)
265 %0 = tail call <4 x i32> @llvm.ppc.vsx.xvxexpsp(<4 x float> %a)
272 declare <4 x i32> @llvm.ppc.vsx.xvxexpsp(<4 x float>)
277 %0 = tail call <2 x i64> @llvm.ppc.vsx.xvxexpdp(<2 x double> %a)
284 declare <2 x i64>@llvm.ppc.vsx.xvxexpdp(<2 x double>)
[all …]
Dvsx-div.ll1 ; RUN: llc -verify-machineinstrs -mcpu=pwr7 -mattr=+vsx -O1 -mtriple=powerpc64-unknown-linux-gnu < …
11 %1 = tail call <4 x float> @llvm.ppc.vsx.xvdivsp(<4 x float> %0, <4 x float> %0)
21 %1 = tail call <2 x double> @llvm.ppc.vsx.xvdivdp(<2 x double> %0, <2 x double> %0)
28 declare <2 x double> @llvm.ppc.vsx.xvdivdp(<2 x double>, <2 x double>)
29 declare <4 x float> @llvm.ppc.vsx.xvdivsp(<4 x float>, <4 x float>)
Dmcm-4.ll2 ; RUN: -fast-isel=false -mattr=-vsx <%s | FileCheck -check-prefix=MEDIUM %s
4 ; RUN: -fast-isel=false -mattr=+vsx <%s | FileCheck -check-prefix=MEDIUM-VSX %s
6 ; RUN: -fast-isel=false -mattr=-vsx <%s | FileCheck -check-prefix=LARGE %s
8 ; RUN: -fast-isel=false -mattr=+vsx <%s | FileCheck -check-prefix=LARGE-VSX %s
10 ; RUN: -fast-isel=false -mattr=+vsx <%s | FileCheck -check-prefix=MEDIUM-P9 %s
12 ; RUN: -fast-isel=false -mattr=+vsx <%s | FileCheck -check-prefix=LARGE-P9 %s
/external/llvm/test/CodeGen/PowerPC/
Dvsx-ldst-builtin-le.ll1 ; RUN: llc -mcpu=pwr8 -mattr=+vsx -O2 -mtriple=powerpc64le-unknown-linux-gnu < %s | FileCheck %s
20 %0 = call <4 x i32> @llvm.ppc.vsx.lxvw4x(i8* bitcast (<4 x i32>* @vsi to i8*))
24 %1 = call <4 x i32> @llvm.ppc.vsx.lxvw4x(i8* bitcast (<4 x i32>* @vui to i8*))
28 %2 = call <4 x i32> @llvm.ppc.vsx.lxvw4x(i8* bitcast (<4 x float>* @vf to i8*))
33 %4 = call <2 x double> @llvm.ppc.vsx.lxvd2x(i8* bitcast (<2 x i64>* @vsll to i8*))
38 %6 = call <2 x double> @llvm.ppc.vsx.lxvd2x(i8* bitcast (<2 x i64>* @vull to i8*))
43 %8 = call <2 x double> @llvm.ppc.vsx.lxvd2x(i8* bitcast (<2 x double>* @vd to i8*))
49 call void @llvm.ppc.vsx.stxvw4x(<4 x i32> %9, i8* bitcast (<4 x i32>* @res_vsi to i8*))
53 call void @llvm.ppc.vsx.stxvw4x(<4 x i32> %10, i8* bitcast (<4 x i32>* @res_vui to i8*))
58 call void @llvm.ppc.vsx.stxvw4x(<4 x i32> %12, i8* bitcast (<4 x float>* @res_vf to i8*))
[all …]
Di64_fp.ll4 ; RUN: llc < %s -mattr=-vsx -march=ppc32 -mattr=+64bit | \
6 ; RUN: llc < %s -mattr=-vsx -march=ppc32 -mattr=+64bit | \
8 ; RUN: llc < %s -mattr=-vsx -march=ppc32 -mcpu=g5 | \
10 ; RUN: llc < %s -mattr=-vsx -march=ppc32 -mcpu=g5 | \
12 ; RUN: llc < %s -mattr=-vsx -march=ppc32 -mattr=-64bit | \
14 ; RUN: llc < %s -mattr=-vsx -march=ppc32 -mattr=-64bit | \
16 ; RUN: llc < %s -mattr=-vsx -march=ppc32 -mcpu=g4 | \
18 ; RUN: llc < %s -mattr=-vsx -march=ppc32 -mcpu=g4 | \
Dvsx-minmax.ll1 ; RUN: llc -mcpu=pwr7 -mattr=+vsx -O0 -fast-isel=0 -mtriple=powerpc64-unknown-linux-gnu < %s | File…
23 %2 = tail call <4 x float> @llvm.ppc.vsx.xvmaxsp(<4 x float> %0, <4 x float> %1)
27 %4 = tail call <2 x double> @llvm.ppc.vsx.xvmaxdp(<2 x double> %3, <2 x double> %3)
32 %7 = tail call <4 x float> @llvm.ppc.vsx.xvmaxsp(<4 x float> %5, <4 x float> %6)
37 %10 = tail call <4 x float> @llvm.ppc.vsx.xvminsp(<4 x float> %8, <4 x float> %9)
41 %12 = tail call <2 x double> @llvm.ppc.vsx.xvmindp(<2 x double> %11, <2 x double> %11)
46 %15 = tail call <4 x float> @llvm.ppc.vsx.xvminsp(<4 x float> %13, <4 x float> %14)
50 %17 = tail call double @llvm.ppc.vsx.xsmaxdp(double %16, double %16)
53 %18 = tail call double @llvm.ppc.vsx.xsmindp(double %16, double %16)
60 declare double @llvm.ppc.vsx.xsmaxdp(double, double)
[all …]
Dbuiltins-ppc-elf2-abi.ll22 %2 = call <2 x double> @llvm.ppc.vsx.xvdivdp(<2 x double> %0, <2 x double> %1)
34 %2 = call <4 x float> @llvm.ppc.vsx.xvdivsp(<4 x float> %0, <4 x float> %1)
70 %2 = call <2 x i64> @llvm.ppc.vsx.xvcmpeqdp(<2 x double> %0, <2 x double> %1)
82 %2 = call <4 x i32> @llvm.ppc.vsx.xvcmpeqsp(<4 x float> %0, <4 x float> %1)
94 %2 = call <2 x i64> @llvm.ppc.vsx.xvcmpgedp(<2 x double> %0, <2 x double> %1)
106 %2 = call <4 x i32> @llvm.ppc.vsx.xvcmpgesp(<4 x float> %0, <4 x float> %1)
118 %2 = call <2 x i64> @llvm.ppc.vsx.xvcmpgtdp(<2 x double> %0, <2 x double> %1)
130 %2 = call <4 x i32> @llvm.ppc.vsx.xvcmpgtsp(<4 x float> %0, <4 x float> %1)
143 %1 = call <4 x float> @llvm.ppc.vsx.xvresp(<4 x float> %0)
155 %1 = call <2 x double> @llvm.ppc.vsx.xvredp(<2 x double> %0)
[all …]
Dvsx-div.ll1 ; RUN: llc -mcpu=pwr7 -mattr=+vsx -O1 -mtriple=powerpc64-unknown-linux-gnu < %s | FileCheck %s
11 %1 = tail call <4 x float> @llvm.ppc.vsx.xvdivsp(<4 x float> %0, <4 x float> %0)
21 %1 = tail call <2 x double> @llvm.ppc.vsx.xvdivdp(<2 x double> %0, <2 x double> %0)
28 declare <2 x double> @llvm.ppc.vsx.xvdivdp(<2 x double>, <2 x double>)
29 declare <4 x float> @llvm.ppc.vsx.xvdivsp(<4 x float>, <4 x float>)
Dfsqrt.ll4 ; RUN: llc < %s -mattr=-vsx -march=ppc32 -mtriple=powerpc-apple-darwin8 -mattr=+fsqrt | \
6 ; RUN: llc < %s -mattr=-vsx -march=ppc32 -mtriple=powerpc-apple-darwin8 -mcpu=g5 | \
8 ; RUN: llc < %s -mattr=-vsx -march=ppc32 -mtriple=powerpc-apple-darwin8 -mattr=-fsqrt | \
10 ; RUN: llc < %s -mattr=-vsx -march=ppc32 -mtriple=powerpc-apple-darwin8 -mcpu=g4 | \
/external/linux-kselftest/tools/testing/selftests/powerpc/ptrace/
Dptrace-vsx.c18 void vsx(void) in vsx() function
39 unsigned long vsx[VSX_MAX]; in trace_vsx() local
43 FAIL_IF(show_vsx(child, vsx)); in trace_vsx()
44 FAIL_IF(validate_vsx(vsx, fp_load)); in trace_vsx()
48 memset(vsx, 0, sizeof(vsx)); in trace_vsx()
50 load_vsx_vmx(fp_load_new, vsx, vmx); in trace_vsx()
52 FAIL_IF(write_vsx(child, vsx)); in trace_vsx()
79 vsx(); in ptrace_vsx()
Dptrace-tm-vsx.c86 unsigned long vsx[VSX_MAX]; in trace_tm_vsx() local
90 FAIL_IF(show_vsx(child, vsx)); in trace_tm_vsx()
91 FAIL_IF(validate_vsx(vsx, fp_load)); in trace_tm_vsx()
94 FAIL_IF(show_vsx_ckpt(child, vsx)); in trace_tm_vsx()
95 FAIL_IF(validate_vsx(vsx, fp_load_ckpt)); in trace_tm_vsx()
98 memset(vsx, 0, sizeof(vsx)); in trace_tm_vsx()
101 load_vsx_vmx(fp_load_ckpt_new, vsx, vmx); in trace_tm_vsx()
103 FAIL_IF(write_vsx_ckpt(child, vsx)); in trace_tm_vsx()
Dptrace-tm-spd-vsx.c98 unsigned long vsx[VSX_MAX]; in trace_tm_spd_vsx() local
102 FAIL_IF(show_vsx(child, vsx)); in trace_tm_spd_vsx()
103 FAIL_IF(validate_vsx(vsx, fp_load)); in trace_tm_spd_vsx()
106 FAIL_IF(show_vsx_ckpt(child, vsx)); in trace_tm_spd_vsx()
107 FAIL_IF(validate_vsx(vsx, fp_load_ckpt)); in trace_tm_spd_vsx()
111 memset(vsx, 0, sizeof(vsx)); in trace_tm_spd_vsx()
114 load_vsx_vmx(fp_load_ckpt_new, vsx, vmx); in trace_tm_spd_vsx()
116 FAIL_IF(write_vsx_ckpt(child, vsx)); in trace_tm_spd_vsx()
Dptrace-vsx.h13 int validate_vsx(unsigned long *vsx, unsigned long *load) in validate_vsx() argument
18 if (vsx[i] != load[2 * i + 1]) { in validate_vsx()
20 i, vsx[i], 2 * i + 1, load[2 * i + 1]); in validate_vsx()
108 void load_vsx_vmx(unsigned long *load, unsigned long *vsx, in load_vsx_vmx() argument
114 vsx[i] = load[1 + 2 * i]; in load_vsx_vmx()
D.gitignore7 ptrace-vsx
8 ptrace-tm-vsx
9 ptrace-tm-spd-vsx
/external/libaom/libaom/av1/common/ppc/
Dcfl_ppc.c113 CFL_SUB_AVG_X(vsx, 8, 4, 16, 5)
114 CFL_SUB_AVG_X(vsx, 8, 8, 32, 6)
115 CFL_SUB_AVG_X(vsx, 8, 16, 64, 7)
116 CFL_SUB_AVG_X(vsx, 8, 32, 128, 8)
117 CFL_SUB_AVG_X(vsx, 16, 4, 32, 6)
118 CFL_SUB_AVG_X(vsx, 16, 8, 64, 7)
119 CFL_SUB_AVG_X(vsx, 16, 16, 128, 8)
120 CFL_SUB_AVG_X(vsx, 16, 32, 256, 9)
121 CFL_SUB_AVG_X(vsx, 32, 8, 128, 8)
122 CFL_SUB_AVG_X(vsx, 32, 16, 256, 9)
[all …]
/external/llvm-project/llvm/test/Transforms/InstCombine/PowerPC/
Dvsx-unaligned.ll19 %2 = call <4 x i32> @llvm.ppc.vsx.lxvw4x(i8* %1)
23 call void @llvm.ppc.vsx.stxvw4x(<4 x i32> %2, i8* %4)
27 %7 = call <2 x double> @llvm.ppc.vsx.lxvd2x(i8* %6)
31 call void @llvm.ppc.vsx.stxvd2x(<2 x double> %7, i8* %9)
41 declare <4 x i32> @llvm.ppc.vsx.lxvw4x(i8*)
42 declare void @llvm.ppc.vsx.stxvw4x(<4 x i32>, i8*)
43 declare <2 x double> @llvm.ppc.vsx.lxvd2x(i8*)
44 declare void @llvm.ppc.vsx.stxvd2x(<2 x double>, i8*)
/external/llvm/test/Transforms/InstCombine/
Dvsx-unaligned.ll19 %2 = call <4 x i32> @llvm.ppc.vsx.lxvw4x(i8* %1)
23 call void @llvm.ppc.vsx.stxvw4x(<4 x i32> %2, i8* %4)
27 %7 = call <2 x double> @llvm.ppc.vsx.lxvd2x(i8* %6)
31 call void @llvm.ppc.vsx.stxvd2x(<2 x double> %7, i8* %9)
41 declare <4 x i32> @llvm.ppc.vsx.lxvw4x(i8*)
42 declare void @llvm.ppc.vsx.stxvw4x(<4 x i32>, i8*)
43 declare <2 x double> @llvm.ppc.vsx.lxvd2x(i8*)
44 declare void @llvm.ppc.vsx.stxvd2x(<2 x double>, i8*)
/external/linux-kselftest/tools/testing/selftests/powerpc/tm/
Dtm-signal-context-chk-vsx.c60 uint8_t vsx[sizeof(vector int)]; in signal_usr1() local
106 memcpy(vsx, &ucp->uc_mcontext.fp_regs[FPR20 + i], 8); in signal_usr1()
107 memcpy(vsx + 8, &vsx_ptr[VSX20 + i], 8); in signal_usr1()
109 fail = memcmp(vsx, &vsxs[i], sizeof(vector int)); in signal_usr1()
115 printf("%02x", vsx[j]); in signal_usr1()

12345678910>>...12