Home
last modified time | relevance | path

Searched refs:vfill (Results 1 – 20 of 20) sorted by relevance

/external/XNNPACK/src/x32-fill/
Dwasmsimd.c27 const v128_t vfill = wasm_v32x4_load_splat(fill_value); in xnn_x32_fill_ukernel__wasmsimd() local
31 wasm_v128_store(output, vfill); in xnn_x32_fill_ukernel__wasmsimd()
32 wasm_v128_store(output + 4, vfill); in xnn_x32_fill_ukernel__wasmsimd()
33 wasm_v128_store(output + 8, vfill); in xnn_x32_fill_ukernel__wasmsimd()
34 wasm_v128_store(output + 12, vfill); in xnn_x32_fill_ukernel__wasmsimd()
38 wasm_v128_store(output, vfill); in xnn_x32_fill_ukernel__wasmsimd()
43 *((double*) output) = wasm_f64x2_extract_lane(vfill, 0); in xnn_x32_fill_ukernel__wasmsimd()
47 *((float*) output) = wasm_f32x4_extract_lane(vfill, 0); in xnn_x32_fill_ukernel__wasmsimd()
Dsse.c27 const __m128 vfill = _mm_load1_ps((const float*) fill_value); in xnn_x32_fill_ukernel__sse() local
32 _mm_storeu_ps(o, vfill); in xnn_x32_fill_ukernel__sse()
33 _mm_storeu_ps(o + 4, vfill); in xnn_x32_fill_ukernel__sse()
34 _mm_storeu_ps(o + 8, vfill); in xnn_x32_fill_ukernel__sse()
35 _mm_storeu_ps(o + 12, vfill); in xnn_x32_fill_ukernel__sse()
39 _mm_storeu_ps(o, vfill); in xnn_x32_fill_ukernel__sse()
44 _mm_storel_pi((__m64*) o, vfill); in xnn_x32_fill_ukernel__sse()
48 _mm_store_ss(o, vfill); in xnn_x32_fill_ukernel__sse()
Dneon.c27 const uint32x4_t vfill = vld1q_dup_u32(fill_value); in xnn_x32_fill_ukernel__neon() local
31 vst1q_u32(output, vfill); output += 4; in xnn_x32_fill_ukernel__neon()
32 vst1q_u32(output, vfill); output += 4; in xnn_x32_fill_ukernel__neon()
33 vst1q_u32(output, vfill); output += 4; in xnn_x32_fill_ukernel__neon()
34 vst1q_u32(output, vfill); output += 4; in xnn_x32_fill_ukernel__neon()
37 vst1q_u32(output, vfill); output += 4; in xnn_x32_fill_ukernel__neon()
41 vst1_u32(output, vget_low_u32(vfill)); output += 2; in xnn_x32_fill_ukernel__neon()
44 vst1q_lane_u32(output, vfill, 0); output += 1; in xnn_x32_fill_ukernel__neon()
Dscalar-int.c25 const uint32_t vfill = *fill_value; in xnn_x32_fill_ukernel__scalar_int() local
29 output[0] = vfill; in xnn_x32_fill_ukernel__scalar_int()
30 output[1] = vfill; in xnn_x32_fill_ukernel__scalar_int()
31 output[2] = vfill; in xnn_x32_fill_ukernel__scalar_int()
32 output[3] = vfill; in xnn_x32_fill_ukernel__scalar_int()
37 *output++ = vfill; in xnn_x32_fill_ukernel__scalar_int()
Dscalar-float.c25 const float vfill = *((const float*) fill_value); in xnn_x32_fill_ukernel__scalar_float() local
30 o[0] = vfill; in xnn_x32_fill_ukernel__scalar_float()
31 o[1] = vfill; in xnn_x32_fill_ukernel__scalar_float()
32 o[2] = vfill; in xnn_x32_fill_ukernel__scalar_float()
33 o[3] = vfill; in xnn_x32_fill_ukernel__scalar_float()
38 *o++ = vfill; in xnn_x32_fill_ukernel__scalar_float()
/external/XNNPACK/src/x32-pad/
Dwasmsimd.c32 const v128_t vfill = wasm_v32x4_load_splat(fill_value); in xnn_x32_pad_ukernel__wasmsimd() local
38 wasm_v128_store(output, vfill); in xnn_x32_pad_ukernel__wasmsimd()
42 *((double*) output) = wasm_f64x2_extract_lane(vfill, 0); in xnn_x32_pad_ukernel__wasmsimd()
46 *((float*) output) = wasm_f32x4_extract_lane(vfill, 0); in xnn_x32_pad_ukernel__wasmsimd()
79 wasm_v128_store(output, vfill); in xnn_x32_pad_ukernel__wasmsimd()
83 *((double*) output) = wasm_f64x2_extract_lane(vfill, 0); in xnn_x32_pad_ukernel__wasmsimd()
87 *((float*) output) = wasm_f32x4_extract_lane(vfill, 0); in xnn_x32_pad_ukernel__wasmsimd()
Dsse.c32 const __m128 vfill = _mm_load1_ps((const float*) fill_value); in xnn_x32_pad_ukernel__sse() local
40 _mm_storeu_ps(o, vfill); in xnn_x32_pad_ukernel__sse()
44 _mm_storel_pi((__m64*) o, vfill); in xnn_x32_pad_ukernel__sse()
48 _mm_store_ss(o, vfill); in xnn_x32_pad_ukernel__sse()
81 _mm_storeu_ps(o, vfill); in xnn_x32_pad_ukernel__sse()
85 _mm_storel_pi((__m64*) o, vfill); in xnn_x32_pad_ukernel__sse()
89 _mm_store_ss(o, vfill); in xnn_x32_pad_ukernel__sse()
Dneon.c32 const uint32x4_t vfill = vld1q_dup_u32(fill_value); in xnn_x32_pad_ukernel__neon() local
38 vst1q_u32(output, vfill); output += 4; in xnn_x32_pad_ukernel__neon()
41 vst1_u32(output, vget_low_u32(vfill)); output += 2; in xnn_x32_pad_ukernel__neon()
44 vst1q_lane_u32(output, vfill, 0); output += 1; in xnn_x32_pad_ukernel__neon()
72 vst1q_u32(output, vfill); output += 4; in xnn_x32_pad_ukernel__neon()
75 vst1_u32(output, vget_low_u32(vfill)); output += 2; in xnn_x32_pad_ukernel__neon()
78 vst1q_lane_u32(output, vfill, 0); output += 1; in xnn_x32_pad_ukernel__neon()
Dscalar-float.c30 const float vfill = *((const float*) fill_value); in xnn_x32_pad_ukernel__scalar_float() local
38 *o++ = vfill; in xnn_x32_pad_ukernel__scalar_float()
64 *o++ = vfill; in xnn_x32_pad_ukernel__scalar_float()
Dscalar-int.c30 const uint32_t vfill = *fill_value; in xnn_x32_pad_ukernel__scalar_int() local
36 *output++ = vfill; in xnn_x32_pad_ukernel__scalar_int()
62 *output++ = vfill; in xnn_x32_pad_ukernel__scalar_int()
/external/XNNPACK/src/x32-unpool/
Dneon.c22 const uint32x4_t vfill = vdupq_n_u32(fill); in xnn_x32_unpool_ukernel__neon() local
28 vst1q_u32(o, vfill); o += 4; in xnn_x32_unpool_ukernel__neon()
32 vst1_u32(o, vget_low_u32(vfill)); o += 2; in xnn_x32_unpool_ukernel__neon()
35 vst1q_lane_u32(o, vfill, 0); in xnn_x32_unpool_ukernel__neon()
Dsse2.c22 const __m128i vfill = _mm_set1_epi32((int) fill); in xnn_x32_unpool_ukernel__sse2() local
28 _mm_storeu_si128((__m128i*) o, vfill); in xnn_x32_unpool_ukernel__sse2()
33 _mm_storel_epi64((__m128i*) o, vfill); in xnn_x32_unpool_ukernel__sse2()
37 *((int*) o) = _mm_cvtsi128_si32(vfill); in xnn_x32_unpool_ukernel__sse2()
Dwasmsimd.c22 const v128_t vfill = wasm_i32x4_splat(fill); in xnn_x32_unpool_ukernel__wasmsimd() local
28 wasm_v128_store(o, vfill); in xnn_x32_unpool_ukernel__wasmsimd()
33 *((double*) o) = wasm_f64x2_extract_lane(vfill, 0); in xnn_x32_unpool_ukernel__wasmsimd()
37 *o = wasm_f32x4_extract_lane(vfill, 0); in xnn_x32_unpool_ukernel__wasmsimd()
/external/llvm-project/clang/test/CodeGen/X86/
Dmmx-inline-asm.c8 __m64 vfill = _mm_cvtsi64_m64(fill); in foo() local
21 : "y" (vfill)); in foo()
/external/clang/test/CodeGen/
Dmmx-inline-asm.c8 __m64 vfill = _mm_cvtsi64_m64(fill); in foo() local
21 : "y" (vfill)); in foo()
/external/mtools/
Dtexinfo.tex652 % explicit \vfill so that the extra space is at the bottom. The
764 \def\page{\par\vfill\supereject}
5942 \ifdim\pagetotal>0.8\vsize\vfill\eject\fi
6121 \setbox2=\vbox to \ht1 {\unvbox3\vfill}%
6122 \setbox0=\vbox to \ht1 {\unvbox1\vfill}%
6521 \def\chappager{\par\vfill\supereject}
6950 \vfill \eject
6989 \vfill \eject
/external/e2fsprogs/doc/
Dtexinfo.tex683 % explicit \vfill so that the extra space is at the bottom. The
799 \def\page{\par\vfill\supereject}
4331 \def\chappager{\par\vfill\supereject}
4661 \vfill \eject
4700 \vfill \eject
/external/e2fsprogs/lib/et/
Dtexinfo.tex683 % explicit \vfill so that the extra space is at the bottom. The
799 \def\page{\par\vfill\supereject}
4331 \def\chappager{\par\vfill\supereject}
4661 \vfill \eject
4700 \vfill \eject
/external/libffi/
Dtexinfo.tex643 % explicit \vfill so that the extra space is at the bottom. The
750 \def\page{\par\vfill\supereject}
5629 \def\chappager{\par\vfill\supereject}
6067 \vfill \eject
6107 \vfill \eject
/external/python/cpython2/Modules/_ctypes/libffi/
Dtexinfo.tex643 % explicit \vfill so that the extra space is at the bottom. The
750 \def\page{\par\vfill\supereject}
5629 \def\chappager{\par\vfill\supereject}
6067 \vfill \eject
6107 \vfill \eject