/external/vixl/examples/aarch64/ |
D | add2-vectors.cc | 56 __ Ld1(v0.V16B(), MemOperand(x0)); in GenerateAdd2Vectors() local 57 __ Ld1(v1.V16B(), MemOperand(x1, 16, PostIndex)); in GenerateAdd2Vectors() local
|
D | neon-matrix-multiply.cc | 70 __ Ld1(v4.V4S(), v5.V4S(), v6.V4S(), v7.V4S(), MemOperand(x1)); in GenerateNEONMatrixMultiply() local 72 __ Ld1(v16.V4S(), v17.V4S(), v18.V4S(), v19.V4S(), MemOperand(x2)); in GenerateNEONMatrixMultiply() local
|
/external/vixl/test/aarch64/ |
D | test-disasm-aarch64.cc | 3115 COMPARE_MACRO(Ld1(v0.M, MemOperand(x15)), "ld1 {v0." S "}, [x15]"); \ in TEST() 3116 COMPARE_MACRO(Ld1(v1.M, v2.M, MemOperand(x16)), \ in TEST() 3118 COMPARE_MACRO(Ld1(v3.M, v4.M, v5.M, MemOperand(x17)), \ in TEST() 3120 COMPARE_MACRO(Ld1(v6.M, v7.M, v8.M, v9.M, MemOperand(x18)), \ in TEST() 3122 COMPARE_MACRO(Ld1(v30.M, v31.M, v0.M, v1.M, MemOperand(sp)), \ in TEST() 3136 COMPARE_MACRO(Ld1(v0.M, MemOperand(x15, x20, PostIndex)), \ in TEST() 3138 COMPARE_MACRO(Ld1(v1.M, v2.M, MemOperand(x16, x21, PostIndex)), \ in TEST() 3140 COMPARE_MACRO(Ld1(v3.M, v4.M, v5.M, MemOperand(x17, x22, PostIndex)), \ in TEST() 3142 COMPARE_MACRO(Ld1(v6.M, v7.M, v8.M, v9.M, MemOperand(x18, x23, PostIndex)), \ in TEST() 3144 COMPARE_MACRO(Ld1(v30.M, v31.M, v0.M, v1.M, MemOperand(sp, x24, PostIndex)), \ in TEST() [all …]
|
D | test-assembler-aarch64.cc | 2999 __ Ld1(v2.V8B(), MemOperand(x17)); in TEST() local 3001 __ Ld1(v3.V8B(), v4.V8B(), MemOperand(x17)); in TEST() local 3003 __ Ld1(v5.V4H(), v6.V4H(), v7.V4H(), MemOperand(x17)); in TEST() local 3005 __ Ld1(v16.V2S(), v17.V2S(), v18.V2S(), v19.V2S(), MemOperand(x17)); in TEST() local 3007 __ Ld1(v30.V2S(), v31.V2S(), v0.V2S(), v1.V2S(), MemOperand(x17)); in TEST() local 3009 __ Ld1(v20.V1D(), v21.V1D(), v22.V1D(), v23.V1D(), MemOperand(x17)); in TEST() local 3055 __ Ld1(v2.V8B(), MemOperand(x17, x23, PostIndex)); in TEST() local 3056 __ Ld1(v3.V8B(), v4.V8B(), MemOperand(x18, 16, PostIndex)); in TEST() local 3057 __ Ld1(v5.V4H(), v6.V4H(), v7.V4H(), MemOperand(x19, 24, PostIndex)); in TEST() local 3058 __ Ld1(v16.V2S(), in TEST() local [all …]
|
/external/llvm/test/Transforms/LoopVectorize/AArch64/ |
D | reduction-small-size.ll | 141 ; CHECK: [[Ld1:%[a-zA-Z0-9.]+]] = load <8 x i8> 142 ; CHECK: zext <8 x i8> [[Ld1]] to <8 x i16>
|
/external/vixl/src/aarch64/ |
D | macro-assembler-aarch64.h | 2511 void Ld1(const VRegister& vt, const MemOperand& src) { in Ld1() function 2516 void Ld1(const VRegister& vt, const VRegister& vt2, const MemOperand& src) { in Ld1() function 2521 void Ld1(const VRegister& vt, in Ld1() function 2529 void Ld1(const VRegister& vt, in Ld1() function 2538 void Ld1(const VRegister& vt, int lane, const MemOperand& src) { in Ld1() function
|