/external/XNNPACK/src/x32-fill/ |
D | wasmsimd.c | 27 const v128_t vfill = wasm_v32x4_load_splat(fill_value); in xnn_x32_fill_ukernel__wasmsimd() local 31 wasm_v128_store(output, vfill); in xnn_x32_fill_ukernel__wasmsimd() 32 wasm_v128_store(output + 4, vfill); in xnn_x32_fill_ukernel__wasmsimd() 33 wasm_v128_store(output + 8, vfill); in xnn_x32_fill_ukernel__wasmsimd() 34 wasm_v128_store(output + 12, vfill); in xnn_x32_fill_ukernel__wasmsimd() 38 wasm_v128_store(output, vfill); in xnn_x32_fill_ukernel__wasmsimd() 43 *((double*) output) = wasm_f64x2_extract_lane(vfill, 0); in xnn_x32_fill_ukernel__wasmsimd() 47 *((float*) output) = wasm_f32x4_extract_lane(vfill, 0); in xnn_x32_fill_ukernel__wasmsimd()
|
D | sse.c | 27 const __m128 vfill = _mm_load1_ps((const float*) fill_value); in xnn_x32_fill_ukernel__sse() local 32 _mm_storeu_ps(o, vfill); in xnn_x32_fill_ukernel__sse() 33 _mm_storeu_ps(o + 4, vfill); in xnn_x32_fill_ukernel__sse() 34 _mm_storeu_ps(o + 8, vfill); in xnn_x32_fill_ukernel__sse() 35 _mm_storeu_ps(o + 12, vfill); in xnn_x32_fill_ukernel__sse() 39 _mm_storeu_ps(o, vfill); in xnn_x32_fill_ukernel__sse() 44 _mm_storel_pi((__m64*) o, vfill); in xnn_x32_fill_ukernel__sse() 48 _mm_store_ss(o, vfill); in xnn_x32_fill_ukernel__sse()
|
D | neon.c | 27 const uint32x4_t vfill = vld1q_dup_u32(fill_value); in xnn_x32_fill_ukernel__neon() local 31 vst1q_u32(output, vfill); output += 4; in xnn_x32_fill_ukernel__neon() 32 vst1q_u32(output, vfill); output += 4; in xnn_x32_fill_ukernel__neon() 33 vst1q_u32(output, vfill); output += 4; in xnn_x32_fill_ukernel__neon() 34 vst1q_u32(output, vfill); output += 4; in xnn_x32_fill_ukernel__neon() 37 vst1q_u32(output, vfill); output += 4; in xnn_x32_fill_ukernel__neon() 41 vst1_u32(output, vget_low_u32(vfill)); output += 2; in xnn_x32_fill_ukernel__neon() 44 vst1q_lane_u32(output, vfill, 0); output += 1; in xnn_x32_fill_ukernel__neon()
|
D | scalar-int.c | 25 const uint32_t vfill = *fill_value; in xnn_x32_fill_ukernel__scalar_int() local 29 output[0] = vfill; in xnn_x32_fill_ukernel__scalar_int() 30 output[1] = vfill; in xnn_x32_fill_ukernel__scalar_int() 31 output[2] = vfill; in xnn_x32_fill_ukernel__scalar_int() 32 output[3] = vfill; in xnn_x32_fill_ukernel__scalar_int() 37 *output++ = vfill; in xnn_x32_fill_ukernel__scalar_int()
|
D | scalar-float.c | 25 const float vfill = *((const float*) fill_value); in xnn_x32_fill_ukernel__scalar_float() local 30 o[0] = vfill; in xnn_x32_fill_ukernel__scalar_float() 31 o[1] = vfill; in xnn_x32_fill_ukernel__scalar_float() 32 o[2] = vfill; in xnn_x32_fill_ukernel__scalar_float() 33 o[3] = vfill; in xnn_x32_fill_ukernel__scalar_float() 38 *o++ = vfill; in xnn_x32_fill_ukernel__scalar_float()
|
/external/XNNPACK/src/x32-pad/ |
D | wasmsimd.c | 32 const v128_t vfill = wasm_v32x4_load_splat(fill_value); in xnn_x32_pad_ukernel__wasmsimd() local 38 wasm_v128_store(output, vfill); in xnn_x32_pad_ukernel__wasmsimd() 42 *((double*) output) = wasm_f64x2_extract_lane(vfill, 0); in xnn_x32_pad_ukernel__wasmsimd() 46 *((float*) output) = wasm_f32x4_extract_lane(vfill, 0); in xnn_x32_pad_ukernel__wasmsimd() 79 wasm_v128_store(output, vfill); in xnn_x32_pad_ukernel__wasmsimd() 83 *((double*) output) = wasm_f64x2_extract_lane(vfill, 0); in xnn_x32_pad_ukernel__wasmsimd() 87 *((float*) output) = wasm_f32x4_extract_lane(vfill, 0); in xnn_x32_pad_ukernel__wasmsimd()
|
D | sse.c | 32 const __m128 vfill = _mm_load1_ps((const float*) fill_value); in xnn_x32_pad_ukernel__sse() local 40 _mm_storeu_ps(o, vfill); in xnn_x32_pad_ukernel__sse() 44 _mm_storel_pi((__m64*) o, vfill); in xnn_x32_pad_ukernel__sse() 48 _mm_store_ss(o, vfill); in xnn_x32_pad_ukernel__sse() 81 _mm_storeu_ps(o, vfill); in xnn_x32_pad_ukernel__sse() 85 _mm_storel_pi((__m64*) o, vfill); in xnn_x32_pad_ukernel__sse() 89 _mm_store_ss(o, vfill); in xnn_x32_pad_ukernel__sse()
|
D | neon.c | 32 const uint32x4_t vfill = vld1q_dup_u32(fill_value); in xnn_x32_pad_ukernel__neon() local 38 vst1q_u32(output, vfill); output += 4; in xnn_x32_pad_ukernel__neon() 41 vst1_u32(output, vget_low_u32(vfill)); output += 2; in xnn_x32_pad_ukernel__neon() 44 vst1q_lane_u32(output, vfill, 0); output += 1; in xnn_x32_pad_ukernel__neon() 72 vst1q_u32(output, vfill); output += 4; in xnn_x32_pad_ukernel__neon() 75 vst1_u32(output, vget_low_u32(vfill)); output += 2; in xnn_x32_pad_ukernel__neon() 78 vst1q_lane_u32(output, vfill, 0); output += 1; in xnn_x32_pad_ukernel__neon()
|
D | scalar-float.c | 30 const float vfill = *((const float*) fill_value); in xnn_x32_pad_ukernel__scalar_float() local 38 *o++ = vfill; in xnn_x32_pad_ukernel__scalar_float() 64 *o++ = vfill; in xnn_x32_pad_ukernel__scalar_float()
|
D | scalar-int.c | 30 const uint32_t vfill = *fill_value; in xnn_x32_pad_ukernel__scalar_int() local 36 *output++ = vfill; in xnn_x32_pad_ukernel__scalar_int() 62 *output++ = vfill; in xnn_x32_pad_ukernel__scalar_int()
|
/external/XNNPACK/src/x32-unpool/ |
D | neon.c | 22 const uint32x4_t vfill = vdupq_n_u32(fill); in xnn_x32_unpool_ukernel__neon() local 28 vst1q_u32(o, vfill); o += 4; in xnn_x32_unpool_ukernel__neon() 32 vst1_u32(o, vget_low_u32(vfill)); o += 2; in xnn_x32_unpool_ukernel__neon() 35 vst1q_lane_u32(o, vfill, 0); in xnn_x32_unpool_ukernel__neon()
|
D | sse2.c | 22 const __m128i vfill = _mm_set1_epi32((int) fill); in xnn_x32_unpool_ukernel__sse2() local 28 _mm_storeu_si128((__m128i*) o, vfill); in xnn_x32_unpool_ukernel__sse2() 33 _mm_storel_epi64((__m128i*) o, vfill); in xnn_x32_unpool_ukernel__sse2() 37 *((int*) o) = _mm_cvtsi128_si32(vfill); in xnn_x32_unpool_ukernel__sse2()
|
D | wasmsimd.c | 22 const v128_t vfill = wasm_i32x4_splat(fill); in xnn_x32_unpool_ukernel__wasmsimd() local 28 wasm_v128_store(o, vfill); in xnn_x32_unpool_ukernel__wasmsimd() 33 *((double*) o) = wasm_f64x2_extract_lane(vfill, 0); in xnn_x32_unpool_ukernel__wasmsimd() 37 *o = wasm_f32x4_extract_lane(vfill, 0); in xnn_x32_unpool_ukernel__wasmsimd()
|
/external/llvm-project/clang/test/CodeGen/X86/ |
D | mmx-inline-asm.c | 8 __m64 vfill = _mm_cvtsi64_m64(fill); in foo() local 21 : "y" (vfill)); in foo()
|
/external/clang/test/CodeGen/ |
D | mmx-inline-asm.c | 8 __m64 vfill = _mm_cvtsi64_m64(fill); in foo() local 21 : "y" (vfill)); in foo()
|
/external/mtools/ |
D | texinfo.tex | 652 % explicit \vfill so that the extra space is at the bottom. The 764 \def\page{\par\vfill\supereject} 5942 \ifdim\pagetotal>0.8\vsize\vfill\eject\fi 6121 \setbox2=\vbox to \ht1 {\unvbox3\vfill}% 6122 \setbox0=\vbox to \ht1 {\unvbox1\vfill}% 6521 \def\chappager{\par\vfill\supereject} 6950 \vfill \eject 6989 \vfill \eject
|
/external/e2fsprogs/doc/ |
D | texinfo.tex | 683 % explicit \vfill so that the extra space is at the bottom. The 799 \def\page{\par\vfill\supereject} 4331 \def\chappager{\par\vfill\supereject} 4661 \vfill \eject 4700 \vfill \eject
|
/external/e2fsprogs/lib/et/ |
D | texinfo.tex | 683 % explicit \vfill so that the extra space is at the bottom. The 799 \def\page{\par\vfill\supereject} 4331 \def\chappager{\par\vfill\supereject} 4661 \vfill \eject 4700 \vfill \eject
|
/external/libffi/ |
D | texinfo.tex | 643 % explicit \vfill so that the extra space is at the bottom. The 750 \def\page{\par\vfill\supereject} 5629 \def\chappager{\par\vfill\supereject} 6067 \vfill \eject 6107 \vfill \eject
|
/external/python/cpython2/Modules/_ctypes/libffi/ |
D | texinfo.tex | 643 % explicit \vfill so that the extra space is at the bottom. The 750 \def\page{\par\vfill\supereject} 5629 \def\chappager{\par\vfill\supereject} 6067 \vfill \eject 6107 \vfill \eject
|