| /external/selinux/python/sepolgen/tests/ |
| D | perm_map | 19 # w = Write 47 transition w 5 48 sigchld w 1 49 sigkill w 1 50 sigstop w 1 52 signal w 5 55 setsched w 1 58 setpgid w 5 60 setcap w 1 63 setexec w 1 [all …]
|
| /external/selinux/python/sepolgen/src/share/ |
| D | perm_map | 19 # w = Write 47 transition w 5 48 sigchld w 1 49 sigkill w 1 50 sigstop w 1 52 signal w 5 55 setsched w 1 58 setpgid w 5 60 setcap w 1 63 setexec w 1 [all …]
|
| /external/llvm/test/CodeGen/AArch64/ |
| D | arm64-neon-mul-div.ll | 104 ; CHECK: sdiv {{w[0-9]+}}, {{w[0-9]+}}, {{w[0-9]+}} 111 ; CHECK: sdiv {{w[0-9]+}}, {{w[0-9]+}}, {{w[0-9]+}} 112 ; CHECK: sdiv {{w[0-9]+}}, {{w[0-9]+}}, {{w[0-9]+}} 113 ; CHECK: sdiv {{w[0-9]+}}, {{w[0-9]+}}, {{w[0-9]+}} 114 ; CHECK: sdiv {{w[0-9]+}}, {{w[0-9]+}}, {{w[0-9]+}} 115 ; CHECK: sdiv {{w[0-9]+}}, {{w[0-9]+}}, {{w[0-9]+}} 116 ; CHECK: sdiv {{w[0-9]+}}, {{w[0-9]+}}, {{w[0-9]+}} 117 ; CHECK: sdiv {{w[0-9]+}}, {{w[0-9]+}}, {{w[0-9]+}} 118 ; CHECK: sdiv {{w[0-9]+}}, {{w[0-9]+}}, {{w[0-9]+}} 125 ; CHECK: sdiv {{w[0-9]+}}, {{w[0-9]+}}, {{w[0-9]+}} [all …]
|
| /external/grpc-grpc/ |
| D | grpc.gemspec | 18 s.files = %w( Makefile .yardopts ) 19 s.files += %w( etc/roots.pem ) 29 s.require_paths = %w( src/ruby/lib src/ruby/bin src/ruby/pb ) 47 s.extensions = %w(src/ruby/ext/grpc/extconf.rb) 49 s.files += %w( third_party/address_sorting/address_sorting_internal.h ) 50 s.files += %w( third_party/address_sorting/include/address_sorting/address_sorting.h ) 51 s.files += %w( third_party/address_sorting/address_sorting.c ) 52 s.files += %w( third_party/address_sorting/address_sorting_posix.c ) 53 s.files += %w( third_party/address_sorting/address_sorting_windows.c ) 54 s.files += %w( include/grpc/support/alloc.h ) [all …]
|
| /external/oss-fuzz/projects/exprtk/ |
| D | exprtk_test_expressions.dict | 82 "x*y*z+w" 83 "x*y*z-w" 85 "x*y+z*w" 86 "x*y+z+w" 87 "x*y+z-w" 88 "x*y+z/w" 90 "x*y-z*w" 91 "x*y-z+w" 92 "x*y-z-w" 93 "x*y-z/w" [all …]
|
| /external/rust/crates/grpcio-sys/grpc/ |
| D | grpc.gemspec | 18 s.files = %w( Makefile .yardopts ) 19 s.files += %w( etc/roots.pem ) 32 s.require_paths = %w( src/ruby/lib src/ruby/bin src/ruby/pb ) 50 s.extensions = %w(src/ruby/ext/grpc/extconf.rb) 52 s.files += %w( include/grpc/byte_buffer.h ) 53 s.files += %w( include/grpc/byte_buffer_reader.h ) 54 s.files += %w( include/grpc/census.h ) 55 s.files += %w( include/grpc/compression.h ) 56 s.files += %w( include/grpc/event_engine/channel_args.h ) 57 s.files += %w( include/grpc/event_engine/event_engine.h ) [all …]
|
| /external/XNNPACK/src/qs8-gemm/gen/ |
| D | 1x16c2s4-minmax-rndnu-neon-mlal.c | 24 const void* restrict w, in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2s4__neon_mlal() argument 36 assert(w != NULL); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2s4__neon_mlal() 44 int32x4_t vacc0x0123 = vld1q_s32(w); w = (const int32_t*) w + 4; in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2s4__neon_mlal() 45 int32x4_t vacc0x4567 = vld1q_s32(w); w = (const int32_t*) w + 4; in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2s4__neon_mlal() 46 int32x4_t vacc0x89AB = vld1q_s32(w); w = (const int32_t*) w + 4; in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2s4__neon_mlal() 47 int32x4_t vacc0xCDEF = vld1q_s32(w); w = (const int32_t*) w + 4; in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2s4__neon_mlal() 54 const int8x8_t vb0123c0x0 = vld1_s8(w); w = (const int8_t*) w + 8; in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2s4__neon_mlal() 55 const int8x8_t vb4567c0x0 = vld1_s8(w); w = (const int8_t*) w + 8; in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2s4__neon_mlal() 56 const int8x8_t vb89ABc0x0 = vld1_s8(w); w = (const int8_t*) w + 8; in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2s4__neon_mlal() 57 const int8x8_t vbCDEFc0x0 = vld1_s8(w); w = (const int8_t*) w + 8; in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2s4__neon_mlal() [all …]
|
| D | 1x16c4s2-minmax-rndnu-neon-mlal.c | 24 const void* restrict w, in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c4s2__neon_mlal() argument 36 assert(w != NULL); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c4s2__neon_mlal() 44 int32x4_t vacc0x01 = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const int32_t*) w + 2; in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c4s2__neon_mlal() 45 int32x4_t vacc0x23 = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const int32_t*) w + 2; in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c4s2__neon_mlal() 46 int32x4_t vacc0x45 = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const int32_t*) w + 2; in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c4s2__neon_mlal() 47 int32x4_t vacc0x67 = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const int32_t*) w + 2; in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c4s2__neon_mlal() 48 int32x4_t vacc0x89 = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const int32_t*) w + 2; in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c4s2__neon_mlal() 49 int32x4_t vacc0xAB = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const int32_t*) w + 2; in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c4s2__neon_mlal() 50 int32x4_t vacc0xCD = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const int32_t*) w + 2; in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c4s2__neon_mlal() 51 int32x4_t vacc0xEF = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const int32_t*) w + 2; in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c4s2__neon_mlal() [all …]
|
| D | 1x16c4-minmax-rndnu-neon-mlal-ld1r.c | 24 const void* restrict w, in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c4__neon_mlal_ld1r() argument 36 assert(w != NULL); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c4__neon_mlal_ld1r() 44 …x4_t vacc0x01 = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const void*) ((uintptr_t) w + … in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c4__neon_mlal_ld1r() 45 …x4_t vacc0x23 = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const void*) ((uintptr_t) w + … in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c4__neon_mlal_ld1r() 46 …x4_t vacc0x45 = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const void*) ((uintptr_t) w + … in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c4__neon_mlal_ld1r() 47 …x4_t vacc0x67 = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const void*) ((uintptr_t) w + … in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c4__neon_mlal_ld1r() 48 …x4_t vacc0x89 = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const void*) ((uintptr_t) w + … in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c4__neon_mlal_ld1r() 49 …x4_t vacc0xAB = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const void*) ((uintptr_t) w + … in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c4__neon_mlal_ld1r() 50 …x4_t vacc0xCD = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const void*) ((uintptr_t) w + … in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c4__neon_mlal_ld1r() 51 …x4_t vacc0xEF = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const void*) ((uintptr_t) w + … in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c4__neon_mlal_ld1r() [all …]
|
| D | 1x16c2-minmax-rndnu-neon-mlal-dup.c | 23 const void* restrict w, in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_dup() argument 35 assert(w != NULL); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_dup() 43 int32x4_t vacc0x0123 = vld1q_s32(w); w = (const void*) ((uintptr_t) w + 4 * sizeof(int32_t)); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_dup() 44 int32x4_t vacc0x4567 = vld1q_s32(w); w = (const void*) ((uintptr_t) w + 4 * sizeof(int32_t)); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_dup() 45 int32x4_t vacc0x89AB = vld1q_s32(w); w = (const void*) ((uintptr_t) w + 4 * sizeof(int32_t)); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_dup() 46 int32x4_t vacc0xCDEF = vld1q_s32(w); w = (const void*) ((uintptr_t) w + 4 * sizeof(int32_t)); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_dup() 53 … const int8x8_t vb0123c0x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof(int8_t)); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_dup() 54 … const int8x8_t vb4567c0x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof(int8_t)); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_dup() 55 … const int8x8_t vb89ABc0x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof(int8_t)); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_dup() 56 … const int8x8_t vbCDEFc0x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof(int8_t)); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_dup() [all …]
|
| D | 1x16c8-minmax-rndnu-neon-mlal.c | 24 const void* restrict w, in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c8__neon_mlal() argument 36 assert(w != NULL); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c8__neon_mlal() 44 …int32x4_t vacc0x0 = vld1q_lane_s32(w, vmovq_n_s32(0), 0); w = (const void*) ((uintptr_t) w + sizeo… in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c8__neon_mlal() 45 …int32x4_t vacc0x1 = vld1q_lane_s32(w, vmovq_n_s32(0), 0); w = (const void*) ((uintptr_t) w + sizeo… in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c8__neon_mlal() 46 …int32x4_t vacc0x2 = vld1q_lane_s32(w, vmovq_n_s32(0), 0); w = (const void*) ((uintptr_t) w + sizeo… in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c8__neon_mlal() 47 …int32x4_t vacc0x3 = vld1q_lane_s32(w, vmovq_n_s32(0), 0); w = (const void*) ((uintptr_t) w + sizeo… in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c8__neon_mlal() 48 …int32x4_t vacc0x4 = vld1q_lane_s32(w, vmovq_n_s32(0), 0); w = (const void*) ((uintptr_t) w + sizeo… in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c8__neon_mlal() 49 …int32x4_t vacc0x5 = vld1q_lane_s32(w, vmovq_n_s32(0), 0); w = (const void*) ((uintptr_t) w + sizeo… in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c8__neon_mlal() 50 …int32x4_t vacc0x6 = vld1q_lane_s32(w, vmovq_n_s32(0), 0); w = (const void*) ((uintptr_t) w + sizeo… in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c8__neon_mlal() 51 …int32x4_t vacc0x7 = vld1q_lane_s32(w, vmovq_n_s32(0), 0); w = (const void*) ((uintptr_t) w + sizeo… in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c8__neon_mlal() [all …]
|
| D | 1x16c2-minmax-rndnu-neon-mlal-ld2r.c | 23 const void* restrict w, in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld2r() argument 35 assert(w != NULL); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld2r() 43 int32x4_t vacc0x0123 = vld1q_s32(w); w = (const void*) ((uintptr_t) w + 4 * sizeof(int32_t)); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld2r() 44 int32x4_t vacc0x4567 = vld1q_s32(w); w = (const void*) ((uintptr_t) w + 4 * sizeof(int32_t)); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld2r() 45 int32x4_t vacc0x89AB = vld1q_s32(w); w = (const void*) ((uintptr_t) w + 4 * sizeof(int32_t)); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld2r() 46 int32x4_t vacc0xCDEF = vld1q_s32(w); w = (const void*) ((uintptr_t) w + 4 * sizeof(int32_t)); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld2r() 55 … const int8x8_t vb0123c0x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof(int8_t)); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld2r() 56 … const int8x8_t vb4567c0x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof(int8_t)); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld2r() 57 … const int8x8_t vb89ABc0x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof(int8_t)); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld2r() 58 … const int8x8_t vbCDEFc0x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof(int8_t)); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld2r() [all …]
|
| D | 1x16c2-minmax-rndnu-neon-mlal-ld1r.c | 23 const void* restrict w, in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld1r() argument 35 assert(w != NULL); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld1r() 43 int32x4_t vacc0x0123 = vld1q_s32(w); w = (const void*) ((uintptr_t) w + 4 * sizeof(int32_t)); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld1r() 44 int32x4_t vacc0x4567 = vld1q_s32(w); w = (const void*) ((uintptr_t) w + 4 * sizeof(int32_t)); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld1r() 45 int32x4_t vacc0x89AB = vld1q_s32(w); w = (const void*) ((uintptr_t) w + 4 * sizeof(int32_t)); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld1r() 46 int32x4_t vacc0xCDEF = vld1q_s32(w); w = (const void*) ((uintptr_t) w + 4 * sizeof(int32_t)); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld1r() 59 … const int8x8_t vb0123c0x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof(int8_t)); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld1r() 60 … const int8x8_t vb4567c0x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof(int8_t)); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld1r() 61 … const int8x8_t vb89ABc0x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof(int8_t)); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld1r() 62 … const int8x8_t vbCDEFc0x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof(int8_t)); in xnn_qs8_gemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld1r() [all …]
|
| /external/boringssl/src/crypto/fipsmodule/bn/ |
| D | miller_rabin_tests.txt | 2 # witness for W. W must be odd and B must satisfy 1 <= B <= W-1. 6 # def is_miller_rabin_witness(w, b): 9 # m = w - 1 14 # # b is a composite witness for w iff the following are true: 15 # # - b^m != 1 (mod w) 16 # # - b^(m*2^j) != -1 (mod w), for 0 <= j < a 17 # z = pow(b, m, w) 19 # # b^m = 1 (mod w) 22 # if z == w-1: 23 # # b^(m*2^j) = -1 (mod w) [all …]
|
| /external/libpcap/ |
| D | scanner.l | 161 W ([0-9A-Fa-f][0-9A-Fa-f]?[0-9A-Fa-f]?[0-9A-Fa-f]?) 170 V680 {W}:{W}:{W}:{W}:{W}:{W}:{W}:{W} 172 V670 ::{W}:{W}:{W}:{W}:{W}:{W}:{W} 173 V671 {W}::{W}:{W}:{W}:{W}:{W}:{W} 174 V672 {W}:{W}::{W}:{W}:{W}:{W}:{W} 175 V673 {W}:{W}:{W}::{W}:{W}:{W}:{W} 176 V674 {W}:{W}:{W}:{W}::{W}:{W}:{W} 177 V675 {W}:{W}:{W}:{W}:{W}::{W}:{W} 178 V676 {W}:{W}:{W}:{W}:{W}:{W}::{W} 179 V677 {W}:{W}:{W}:{W}:{W}:{W}:{W}:: [all …]
|
| /external/ImageMagick/PerlMagick/t/reference/write/read/ |
| D | gradient.miff | 9 …view��_.���\�XYZ L VPW�meas… 10 #(-27;@EJOTY^chmrw|�������������������������… 12 +:IXgw��������'7HYj{�������+=Oat�������2FZ… 34 …4Vx���&Il����Ae����@e���� Ek���*Qw���;c���… 35 #8#f#�#�#�$$M$|$�$�% %8%h%�%�%�&'&W&�&�&�''I'z'�'�( 37 …P�QQPQ�Q�R1R|R�SS_S�S�TBT�T�U(UuU�VV\V�V�WDW�W�X/X}X�YYiY�ZZVZ�Z�[E[�[�\5\�\�]']x]�^^l^�__a… 38 �k�͂0����W�������G����r�ׇ;����i�Ή3�����d�ʋ0�����c�ʍ1�����f�Ώ6����n�֑?����z��M��� �����_�ɖ4��� 41 …W���w����)���K���m�����������������������������������������…
|
| /external/XNNPACK/src/qs8-igemm/gen/ |
| D | 1x16c4s2-minmax-rndnu-neon-mlal.c | 24 const void* restrict w, in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c4s2__neon_mlal() argument 40 assert(w != NULL); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c4s2__neon_mlal() 47 int32x4_t vacc0x01 = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const int32_t*) w + 2; in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c4s2__neon_mlal() 48 int32x4_t vacc0x23 = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const int32_t*) w + 2; in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c4s2__neon_mlal() 49 int32x4_t vacc0x45 = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const int32_t*) w + 2; in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c4s2__neon_mlal() 50 int32x4_t vacc0x67 = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const int32_t*) w + 2; in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c4s2__neon_mlal() 51 int32x4_t vacc0x89 = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const int32_t*) w + 2; in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c4s2__neon_mlal() 52 int32x4_t vacc0xAB = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const int32_t*) w + 2; in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c4s2__neon_mlal() 53 int32x4_t vacc0xCD = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const int32_t*) w + 2; in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c4s2__neon_mlal() 54 int32x4_t vacc0xEF = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const int32_t*) w + 2; in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c4s2__neon_mlal() [all …]
|
| D | 1x16c2s4-minmax-rndnu-neon-mlal.c | 24 const void* restrict w, in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2s4__neon_mlal() argument 40 assert(w != NULL); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2s4__neon_mlal() 47 int32x4_t vacc0x0123 = vld1q_s32(w); w = (const int32_t*) w + 4; in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2s4__neon_mlal() 48 int32x4_t vacc0x4567 = vld1q_s32(w); w = (const int32_t*) w + 4; in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2s4__neon_mlal() 49 int32x4_t vacc0x89AB = vld1q_s32(w); w = (const int32_t*) w + 4; in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2s4__neon_mlal() 50 int32x4_t vacc0xCDEF = vld1q_s32(w); w = (const int32_t*) w + 4; in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2s4__neon_mlal() 65 const int8x8_t vb0123c0x0 = vld1_s8(w); w = (const int8_t*) w + 8; in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2s4__neon_mlal() 66 const int8x8_t vb4567c0x0 = vld1_s8(w); w = (const int8_t*) w + 8; in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2s4__neon_mlal() 67 const int8x8_t vb89ABc0x0 = vld1_s8(w); w = (const int8_t*) w + 8; in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2s4__neon_mlal() 68 const int8x8_t vbCDEFc0x0 = vld1_s8(w); w = (const int8_t*) w + 8; in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2s4__neon_mlal() [all …]
|
| D | 1x16c4-minmax-rndnu-neon-mlal-ld1r.c | 24 const void* restrict w, in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c4__neon_mlal_ld1r() argument 40 assert(w != NULL); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c4__neon_mlal_ld1r() 47 …x4_t vacc0x01 = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const void*) ((uintptr_t) w + … in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c4__neon_mlal_ld1r() 48 …x4_t vacc0x23 = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const void*) ((uintptr_t) w + … in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c4__neon_mlal_ld1r() 49 …x4_t vacc0x45 = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const void*) ((uintptr_t) w + … in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c4__neon_mlal_ld1r() 50 …x4_t vacc0x67 = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const void*) ((uintptr_t) w + … in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c4__neon_mlal_ld1r() 51 …x4_t vacc0x89 = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const void*) ((uintptr_t) w + … in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c4__neon_mlal_ld1r() 52 …x4_t vacc0xAB = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const void*) ((uintptr_t) w + … in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c4__neon_mlal_ld1r() 53 …x4_t vacc0xCD = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const void*) ((uintptr_t) w + … in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c4__neon_mlal_ld1r() 54 …x4_t vacc0xEF = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const void*) ((uintptr_t) w + … in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c4__neon_mlal_ld1r() [all …]
|
| D | 1x16c2-minmax-rndnu-neon-mlal-ld1r.c | 23 const void* restrict w, in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld1r() argument 39 assert(w != NULL); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld1r() 46 int32x4_t vacc0x0123 = vld1q_s32(w); w = (const void*) ((uintptr_t) w + 4 * sizeof(int32_t)); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld1r() 47 int32x4_t vacc0x4567 = vld1q_s32(w); w = (const void*) ((uintptr_t) w + 4 * sizeof(int32_t)); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld1r() 48 int32x4_t vacc0x89AB = vld1q_s32(w); w = (const void*) ((uintptr_t) w + 4 * sizeof(int32_t)); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld1r() 49 int32x4_t vacc0xCDEF = vld1q_s32(w); w = (const void*) ((uintptr_t) w + 4 * sizeof(int32_t)); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld1r() 71 … const int8x8_t vb0123c0x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof(int8_t)); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld1r() 72 … const int8x8_t vb4567c0x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof(int8_t)); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld1r() 73 … const int8x8_t vb89ABc0x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof(int8_t)); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld1r() 74 … const int8x8_t vbCDEFc0x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof(int8_t)); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld1r() [all …]
|
| D | 1x16c2-minmax-rndnu-neon-mlal-ld4r.c | 23 const void* restrict w, in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld4r() argument 39 assert(w != NULL); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld4r() 46 int32x4_t vacc0x0123 = vld1q_s32(w); w = (const void*) ((uintptr_t) w + 4 * sizeof(int32_t)); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld4r() 47 int32x4_t vacc0x4567 = vld1q_s32(w); w = (const void*) ((uintptr_t) w + 4 * sizeof(int32_t)); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld4r() 48 int32x4_t vacc0x89AB = vld1q_s32(w); w = (const void*) ((uintptr_t) w + 4 * sizeof(int32_t)); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld4r() 49 int32x4_t vacc0xCDEF = vld1q_s32(w); w = (const void*) ((uintptr_t) w + 4 * sizeof(int32_t)); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld4r() 65 … const int8x8_t vb0123c0x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof(int8_t)); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld4r() 66 … const int8x8_t vb4567c0x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof(int8_t)); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld4r() 67 … const int8x8_t vb89ABc0x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof(int8_t)); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld4r() 68 … const int8x8_t vbCDEFc0x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof(int8_t)); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld4r() [all …]
|
| D | 1x16c8-minmax-rndnu-neon-mlal.c | 24 const void* restrict w, in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c8__neon_mlal() argument 40 assert(w != NULL); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c8__neon_mlal() 47 …int32x4_t vacc0x0 = vld1q_lane_s32(w, vmovq_n_s32(0), 0); w = (const void*) ((uintptr_t) w + sizeo… in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c8__neon_mlal() 48 …int32x4_t vacc0x1 = vld1q_lane_s32(w, vmovq_n_s32(0), 0); w = (const void*) ((uintptr_t) w + sizeo… in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c8__neon_mlal() 49 …int32x4_t vacc0x2 = vld1q_lane_s32(w, vmovq_n_s32(0), 0); w = (const void*) ((uintptr_t) w + sizeo… in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c8__neon_mlal() 50 …int32x4_t vacc0x3 = vld1q_lane_s32(w, vmovq_n_s32(0), 0); w = (const void*) ((uintptr_t) w + sizeo… in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c8__neon_mlal() 51 …int32x4_t vacc0x4 = vld1q_lane_s32(w, vmovq_n_s32(0), 0); w = (const void*) ((uintptr_t) w + sizeo… in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c8__neon_mlal() 52 …int32x4_t vacc0x5 = vld1q_lane_s32(w, vmovq_n_s32(0), 0); w = (const void*) ((uintptr_t) w + sizeo… in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c8__neon_mlal() 53 …int32x4_t vacc0x6 = vld1q_lane_s32(w, vmovq_n_s32(0), 0); w = (const void*) ((uintptr_t) w + sizeo… in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c8__neon_mlal() 54 …int32x4_t vacc0x7 = vld1q_lane_s32(w, vmovq_n_s32(0), 0); w = (const void*) ((uintptr_t) w + sizeo… in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c8__neon_mlal() [all …]
|
| D | 1x16c2-minmax-rndnu-neon-mlal-dup.c | 23 const void* restrict w, in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_dup() argument 39 assert(w != NULL); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_dup() 46 int32x4_t vacc0x0123 = vld1q_s32(w); w = (const void*) ((uintptr_t) w + 4 * sizeof(int32_t)); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_dup() 47 int32x4_t vacc0x4567 = vld1q_s32(w); w = (const void*) ((uintptr_t) w + 4 * sizeof(int32_t)); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_dup() 48 int32x4_t vacc0x89AB = vld1q_s32(w); w = (const void*) ((uintptr_t) w + 4 * sizeof(int32_t)); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_dup() 49 int32x4_t vacc0xCDEF = vld1q_s32(w); w = (const void*) ((uintptr_t) w + 4 * sizeof(int32_t)); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_dup() 65 … const int8x8_t vb0123c0x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof(int8_t)); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_dup() 66 … const int8x8_t vb4567c0x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof(int8_t)); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_dup() 67 … const int8x8_t vb89ABc0x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof(int8_t)); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_dup() 68 … const int8x8_t vbCDEFc0x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof(int8_t)); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_dup() [all …]
|
| D | 1x16c4-minmax-rndnu-neon-mlal-ld2r.c | 24 const void* restrict w, in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c4__neon_mlal_ld2r() argument 40 assert(w != NULL); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c4__neon_mlal_ld2r() 47 …x4_t vacc0x01 = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const void*) ((uintptr_t) w + … in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c4__neon_mlal_ld2r() 48 …x4_t vacc0x23 = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const void*) ((uintptr_t) w + … in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c4__neon_mlal_ld2r() 49 …x4_t vacc0x45 = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const void*) ((uintptr_t) w + … in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c4__neon_mlal_ld2r() 50 …x4_t vacc0x67 = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const void*) ((uintptr_t) w + … in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c4__neon_mlal_ld2r() 51 …x4_t vacc0x89 = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const void*) ((uintptr_t) w + … in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c4__neon_mlal_ld2r() 52 …x4_t vacc0xAB = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const void*) ((uintptr_t) w + … in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c4__neon_mlal_ld2r() 53 …x4_t vacc0xCD = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const void*) ((uintptr_t) w + … in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c4__neon_mlal_ld2r() 54 …x4_t vacc0xEF = vreinterpretq_s32_u64(vmovl_u32(vld1_u32(w))); w = (const void*) ((uintptr_t) w + … in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c4__neon_mlal_ld2r() [all …]
|
| D | 1x16c2-minmax-rndnu-neon-mlal-ld2r.c | 23 const void* restrict w, in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld2r() argument 39 assert(w != NULL); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld2r() 46 int32x4_t vacc0x0123 = vld1q_s32(w); w = (const void*) ((uintptr_t) w + 4 * sizeof(int32_t)); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld2r() 47 int32x4_t vacc0x4567 = vld1q_s32(w); w = (const void*) ((uintptr_t) w + 4 * sizeof(int32_t)); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld2r() 48 int32x4_t vacc0x89AB = vld1q_s32(w); w = (const void*) ((uintptr_t) w + 4 * sizeof(int32_t)); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld2r() 49 int32x4_t vacc0xCDEF = vld1q_s32(w); w = (const void*) ((uintptr_t) w + 4 * sizeof(int32_t)); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld2r() 67 … const int8x8_t vb0123c0x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof(int8_t)); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld2r() 68 … const int8x8_t vb4567c0x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof(int8_t)); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld2r() 69 … const int8x8_t vb89ABc0x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof(int8_t)); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld2r() 70 … const int8x8_t vbCDEFc0x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof(int8_t)); in xnn_qs8_igemm_minmax_rndnu_ukernel_1x16c2__neon_mlal_ld2r() [all …]
|