/external/openssh/regress/unittests/sshkey/ |
D | test_sshkey.c | 179 struct sshkey *k1, *k2, *k3, *k4, *kr, *kd, *kf; in sshkey_tests() local 186 k1 = sshkey_new(-42); in sshkey_tests() 187 ASSERT_PTR_EQ(k1, NULL); in sshkey_tests() 191 k1 = sshkey_new(KEY_UNSPEC); in sshkey_tests() 192 ASSERT_PTR_NE(k1, NULL); in sshkey_tests() 193 sshkey_free(k1); in sshkey_tests() 197 k1 = sshkey_new(KEY_RSA1); in sshkey_tests() 198 ASSERT_PTR_NE(k1, NULL); in sshkey_tests() 199 ASSERT_PTR_NE(k1->rsa, NULL); in sshkey_tests() 200 ASSERT_PTR_NE(k1->rsa->n, NULL); in sshkey_tests() [all …]
|
D | test_fuzz.c | 51 struct sshkey *k1; in public_fuzz() local 63 &k1), 0); in public_fuzz() 64 sshkey_free(k1); in public_fuzz() 68 if (sshkey_from_blob(fuzz_ptr(fuzz), fuzz_len(fuzz), &k1) == 0) in public_fuzz() 69 sshkey_free(k1); in public_fuzz() 102 struct sshkey *k1; in sshkey_fuzz_tests() local 113 ASSERT_INT_EQ(sshkey_parse_private_fileblob(buf, "", &k1, NULL), 0); in sshkey_fuzz_tests() 114 sshkey_free(k1); in sshkey_fuzz_tests() 121 if (sshkey_parse_private_fileblob(fuzzed, "", &k1, NULL) == 0) in sshkey_fuzz_tests() 122 sshkey_free(k1); in sshkey_fuzz_tests() [all …]
|
D | test_file.c | 45 struct sshkey *k1, *k2; in sshkey_file_tests() local 57 ASSERT_INT_EQ(sshkey_parse_private_fileblob(buf, "", &k1, NULL), 0); in sshkey_file_tests() 59 ASSERT_PTR_NE(k1, NULL); in sshkey_file_tests() 61 ASSERT_BIGNUM_EQ(k1->rsa->n, a); in sshkey_file_tests() 71 ASSERT_INT_EQ(sshkey_equal(k1, k2), 1); in sshkey_file_tests() 79 ASSERT_INT_EQ(sshkey_equal(k1, k2), 1); in sshkey_file_tests() 85 cp = sshkey_fingerprint(k1, SSH_DIGEST_SHA256, SSH_FP_BASE64); in sshkey_file_tests() 94 cp = sshkey_fingerprint(k1, SSH_DIGEST_SHA1, SSH_FP_BUBBLEBABBLE); in sshkey_file_tests() 101 sshkey_free(k1); in sshkey_file_tests() 106 ASSERT_INT_EQ(sshkey_parse_private_fileblob(buf, "", &k1, NULL), 0); in sshkey_file_tests() [all …]
|
/external/llvm/test/MC/X86/ |
D | intel-syntax-x86-64-avx512f_vl.s | 177 vcmpps k4{k1},ymm19,ymm18,0xab 229 vgatherdpd xmm17 {k1}, xmmword ptr [r14 + 8*xmm31 + 123] 233 vgatherdpd xmm17 {k1}, xmmword ptr [r9 + xmm31 + 256] 237 vgatherdpd xmm17 {k1}, xmmword ptr [rcx + 4*xmm31 + 1024] 241 vgatherdpd ymm23 {k1}, ymmword ptr [r14 + 8*xmm31 + 123] 245 vgatherdpd ymm23 {k1}, ymmword ptr [r9 + xmm31 + 256] 249 vgatherdpd ymm23 {k1}, ymmword ptr [rcx + 4*xmm31 + 1024] 253 vgatherdpd xmm23 {k1}, xmmword ptr [r14 + 8*xmm31 - 123] 257 vgatherdpd xmm23 {k1}, xmmword ptr [r9 + xmm31 + 256] 261 vgatherdpd xmm23 {k1}, xmmword ptr [rcx + 4*xmm31 + 1024] [all …]
|
D | avx512vl-encoding.s | 121 vblendmps %xmm20, %xmm20, %xmm24 {%k1} 125 vblendmps %xmm20, %xmm20, %xmm24 {%k1} {z} 869 vptestnmq %xmm19, %xmm28,%k2 {%k1} 921 vptestnmq %ymm17, %ymm26,%k4 {%k1} 1358 vexpandps (%rcx), %xmm18 {%k1} 1362 vexpandps (%rcx), %xmm18 {%k1} {z} 1666 vpgatherdd 123(%r14,%xmm31,8), %xmm17 {%k1} 1670 vpgatherdd 256(%r9,%xmm31), %xmm17 {%k1} 1674 vpgatherdd 1024(%rcx,%xmm31,4), %xmm17 {%k1} 1678 vpgatherdd 123(%r14,%ymm31,8), %ymm19 {%k1} [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/MC/X86/ |
D | intel-syntax-x86-64-avx512f_vl.s | 177 vcmpps k4{k1},ymm19,ymm18,0xab 229 vgatherdpd xmm17 {k1}, xmmword ptr [r14 + 8*xmm31 + 123] 233 vgatherdpd xmm17 {k1}, xmmword ptr [r9 + xmm31 + 256] 237 vgatherdpd xmm17 {k1}, xmmword ptr [rcx + 4*xmm31 + 1024] 241 vgatherdpd ymm23 {k1}, ymmword ptr [r14 + 8*xmm31 + 123] 245 vgatherdpd ymm23 {k1}, ymmword ptr [r9 + xmm31 + 256] 249 vgatherdpd ymm23 {k1}, ymmword ptr [rcx + 4*xmm31 + 1024] 253 vgatherdpd xmm23 {k1}, xmmword ptr [r14 + 8*xmm31 - 123] 257 vgatherdpd xmm23 {k1}, xmmword ptr [r9 + xmm31 + 256] 261 vgatherdpd xmm23 {k1}, xmmword ptr [rcx + 4*xmm31 + 1024] [all …]
|
D | avx512vl-encoding.s | 121 vblendmps %xmm20, %xmm20, %xmm24 {%k1} 125 vblendmps %xmm20, %xmm20, %xmm24 {%k1} {z} 869 vptestnmq %xmm19, %xmm28,%k2 {%k1} 921 vptestnmq %ymm17, %ymm26,%k4 {%k1} 1358 vexpandps (%rcx), %xmm18 {%k1} 1362 vexpandps (%rcx), %xmm18 {%k1} {z} 1666 vpgatherdd 123(%r14,%xmm31,8), %xmm17 {%k1} 1670 vpgatherdd 256(%r9,%xmm31), %xmm17 {%k1} 1674 vpgatherdd 1024(%rcx,%xmm31,4), %xmm17 {%k1} 1678 vpgatherdd 123(%r14,%ymm31,8), %ymm19 {%k1} [all …]
|
D | avx512vl_bitalg-encoding.s | 229 vpshufbitqmb %xmm2, %xmm23, %k1 233 vpshufbitqmb %xmm2, %xmm23, %k1 {%k2} 237 vpshufbitqmb (%rcx), %xmm23, %k1 241 vpshufbitqmb -64(%rsp), %xmm23, %k1 245 vpshufbitqmb 64(%rsp), %xmm23, %k1 249 vpshufbitqmb 268435456(%rcx,%r14,8), %xmm23, %k1 253 vpshufbitqmb -536870912(%rcx,%r14,8), %xmm23, %k1 257 vpshufbitqmb -536870910(%rcx,%r14,8), %xmm23, %k1 261 vpshufbitqmb (%rcx), %xmm23, %k1 {%k2} 265 vpshufbitqmb -64(%rsp), %xmm23, %k1 {%k2} [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/tools/llvm-mca/X86/Generic/ |
D | resources-avx512vl.s | 6 vaddpd %xmm16, %xmm17, %xmm19 {k1} 7 vaddpd (%rax), %xmm17, %xmm19 {k1} 8 vaddpd %xmm16, %xmm17, %xmm19 {z}{k1} 9 vaddpd (%rax), %xmm17, %xmm19 {z}{k1} 13 vaddpd %ymm16, %ymm17, %ymm19 {k1} 14 vaddpd (%rax), %ymm17, %ymm19 {k1} 15 vaddpd %ymm16, %ymm17, %ymm19 {z}{k1} 16 vaddpd (%rax), %ymm17, %ymm19 {z}{k1} 20 vaddps %xmm16, %xmm17, %xmm19 {k1} 21 vaddps (%rax), %xmm17, %xmm19 {k1} [all …]
|
D | resources-avx512.s | 6 vaddpd %zmm16, %zmm17, %zmm19 {k1} 7 vaddpd (%rax), %zmm17, %zmm19 {k1} 8 vaddpd %zmm16, %zmm17, %zmm19 {z}{k1} 9 vaddpd (%rax), %zmm17, %zmm19 {z}{k1} 13 vaddps %zmm16, %zmm17, %zmm19 {k1} 14 vaddps (%rax), %zmm17, %zmm19 {k1} 15 vaddps %zmm16, %zmm17, %zmm19 {z}{k1} 16 vaddps (%rax), %zmm17, %zmm19 {z}{k1} 20 vdivpd %zmm16, %zmm17, %zmm19 {k1} 21 vdivpd (%rax), %zmm17, %zmm19 {k1} [all …]
|
/external/llvm/test/CodeGen/X86/ |
D | avx512-vec-cmp.ll | 8 ; CHECK-NEXT: vcmpleps %zmm1, %zmm0, %k1 9 ; CHECK-NEXT: vblendmps %zmm0, %zmm1, %zmm0 {%k1} 19 ; CHECK-NEXT: vcmplepd %zmm1, %zmm0, %k1 20 ; CHECK-NEXT: vblendmpd %zmm0, %zmm1, %zmm0 {%k1} 30 ; CHECK-NEXT: vpcmpeqd (%rdi), %zmm0, %k1 31 ; CHECK-NEXT: vpblendmd %zmm0, %zmm1, %zmm0 {%k1} 42 ; CHECK-NEXT: vpcmpnltud %zmm1, %zmm0, %k1 43 ; CHECK-NEXT: vpblendmd %zmm2, %zmm1, %zmm0 {%k1} 53 ; CHECK-NEXT: vpcmpeqq %zmm1, %zmm0, %k1 54 ; CHECK-NEXT: vpblendmq %zmm0, %zmm1, %zmm0 {%k1} [all …]
|
D | avx512-mask-op.ll | 137 ; CHECK-NEXT: kmovw (%rsi), %k1 138 ; CHECK-NEXT: kandw %k1, %k0, %k2 139 ; CHECK-NEXT: kxorw %k1, %k0, %k0 262 ; SKX-NEXT: knotw %k0, %k1 263 ; SKX-NEXT: vpcmpgtq %ymm1, %ymm0, %k0 {%k1} 284 ; SKX-NEXT: knotw %k0, %k1 285 ; SKX-NEXT: vpcmpgtq %xmm3, %xmm2, %k0 {%k1} 313 ; KNL-NEXT: kmovw %eax, %k1 314 ; KNL-NEXT: korw %k1, %k0, %k0 324 ; SKX-NEXT: kmovb %eax, %k1 [all …]
|
D | masked_gather_scatter.ll | 22 ; KNL_64-NEXT: kxnorw %k0, %k0, %k1 23 ; KNL_64-NEXT: vgatherdps (%rdi,%zmm0,4), %zmm1 {%k1} 30 ; KNL_32-NEXT: kxnorw %k0, %k0, %k1 31 ; KNL_32-NEXT: vgatherdps (%eax,%zmm0,4), %zmm1 {%k1} 37 ; SKX-NEXT: kxnorw %k0, %k0, %k1 38 ; SKX-NEXT: vgatherdps (%rdi,%zmm0,4), %zmm1 {%k1} 71 ; KNL_64-NEXT: kmovw %esi, %k1 72 ; KNL_64-NEXT: vgatherdps (%rdi,%zmm0,4), %zmm1 {%k1} 79 ; KNL_32-NEXT: kmovw {{[0-9]+}}(%esp), %k1 80 ; KNL_32-NEXT: vgatherdps (%eax,%zmm0,4), %zmm1 {%k1} [all …]
|
/external/python/cpython3/Lib/test/ |
D | test_keywordonlyarg.py | 10 def keywordonly_sum(*, k1=0, k2): argument 11 return k1 + k2 12 def keywordonly_nodefaults_sum(*, k1, k2): argument 13 return k1 + k2 14 def keywordonly_and_kwarg_sum(*, k1, k2, **kwarg): argument 15 return k1 + k2 + sum(kwarg.values()) 16 def mixedargs_sum(a, b=0, *arg, k1, k2=0): argument 17 return a + b + k1 + k2 + sum(arg) 18 def mixedargs_sum2(a, b=0, *arg, k1, k2=0, **kwargs): argument 19 return a + b + k1 + k2 + sum(arg) + sum(kwargs.values()) [all …]
|
/external/jemalloc_new/include/jemalloc/internal/ |
D | hash.h | 88 uint32_t k1 = hash_get_block_32(blocks, i); in hash_x86_32() local 90 k1 *= c1; in hash_x86_32() 91 k1 = hash_rotl_32(k1, 15); in hash_x86_32() 92 k1 *= c2; in hash_x86_32() 94 h1 ^= k1; in hash_x86_32() 104 uint32_t k1 = 0; in hash_x86_32() local 107 case 3: k1 ^= tail[2] << 16; in hash_x86_32() 108 case 2: k1 ^= tail[1] << 8; in hash_x86_32() 109 case 1: k1 ^= tail[0]; k1 *= c1; k1 = hash_rotl_32(k1, 15); in hash_x86_32() 110 k1 *= c2; h1 ^= k1; in hash_x86_32() [all …]
|
/external/jemalloc/include/jemalloc/internal/ |
D | hash.h | 121 uint32_t k1 = hash_get_block_32(blocks, i); in hash_x86_32() local 123 k1 *= c1; in hash_x86_32() 124 k1 = hash_rotl_32(k1, 15); in hash_x86_32() 125 k1 *= c2; in hash_x86_32() 127 h1 ^= k1; in hash_x86_32() 137 uint32_t k1 = 0; in hash_x86_32() local 140 case 3: k1 ^= tail[2] << 16; in hash_x86_32() 141 case 2: k1 ^= tail[1] << 8; in hash_x86_32() 142 case 1: k1 ^= tail[0]; k1 *= c1; k1 = hash_rotl_32(k1, 15); in hash_x86_32() 143 k1 *= c2; h1 ^= k1; in hash_x86_32() [all …]
|
/external/guava/guava/src/com/google/common/hash/ |
D | Murmur3_32HashFunction.java | 85 int k1 = mixK1(input); in hashInt() local 86 int h1 = mixH1(seed, k1); in hashInt() 95 int k1 = mixK1(low); in hashLong() local 96 int h1 = mixH1(seed, k1); in hashLong() 98 k1 = mixK1(high); in hashLong() 99 h1 = mixH1(h1, k1); in hashLong() 110 int k1 = input.charAt(i - 1) | (input.charAt(i) << 16); in hashUnencodedChars() local 111 k1 = mixK1(k1); in hashUnencodedChars() 112 h1 = mixH1(h1, k1); in hashUnencodedChars() 117 int k1 = input.charAt(input.length() - 1); in hashUnencodedChars() local [all …]
|
D | Murmur3_128HashFunction.java | 94 long k1 = bb.getLong(); in process() local 96 bmix64(k1, k2); in process() 100 private void bmix64(long k1, long k2) { in bmix64() argument 101 h1 ^= mixK1(k1); in bmix64() 115 long k1 = 0; in processRemaining() local 134 k1 ^= bb.getLong(); in processRemaining() 137 k1 ^= (long) toInt(bb.get(6)) << 48; // fall through in processRemaining() 139 k1 ^= (long) toInt(bb.get(5)) << 40; // fall through in processRemaining() 141 k1 ^= (long) toInt(bb.get(4)) << 32; // fall through in processRemaining() 143 k1 ^= (long) toInt(bb.get(3)) << 24; // fall through in processRemaining() [all …]
|
/external/grpc-grpc/src/core/lib/gpr/ |
D | murmur_hash.cc | 36 uint32_t k1; in gpr_murmur_hash3() local 42 const size_t bsize = sizeof(k1); in gpr_murmur_hash3() 47 memcpy(&k1, keyptr, bsize); in gpr_murmur_hash3() 49 k1 *= c1; in gpr_murmur_hash3() 50 k1 = ROTL32(k1, 15); in gpr_murmur_hash3() 51 k1 *= c2; in gpr_murmur_hash3() 53 h1 ^= k1; in gpr_murmur_hash3() 58 k1 = 0; in gpr_murmur_hash3() 63 k1 ^= (static_cast<uint32_t>(keyptr[2])) << 16; in gpr_murmur_hash3() 66 k1 ^= (static_cast<uint32_t>(keyptr[1])) << 8; in gpr_murmur_hash3() [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/X86/ |
D | avx512-masked-memop-64-32.ll | 8 ; AVX512-NEXT: vptestnmd %zmm0, %zmm0, %k1 9 ; AVX512-NEXT: vmovdqu32 (%rdi), %zmm0 {%k1} {z} 19 ; AVX512-NEXT: vptestnmd %zmm0, %zmm0, %k1 20 ; AVX512-NEXT: vmovdqu32 (%rdi), %zmm0 {%k1} {z} 30 ; AVX512-NEXT: vptestnmd %zmm0, %zmm0, %k1 31 ; AVX512-NEXT: vmovdqu32 %zmm1, (%rdi) {%k1} 42 ; AVX512-NEXT: vptestnmd %zmm0, %zmm0, %k1 43 ; AVX512-NEXT: vblendmps (%rdi), %zmm1, %zmm0 {%k1} 53 ; AVX512-NEXT: vptestnmd %zmm0, %zmm0, %k1 54 ; AVX512-NEXT: vmovups %zmm1, (%rdi) {%k1} [all …]
|
D | avx512vl-vec-cmp.ll | 8 ; VLX-NEXT: vpcmpeqq %ymm1, %ymm0, %k1 9 ; VLX-NEXT: vpblendmq %ymm0, %ymm1, %ymm0 {%k1} 16 ; NoVLX-NEXT: vpcmpeqq %zmm1, %zmm0, %k1 17 ; NoVLX-NEXT: vpblendmq %zmm0, %zmm1, %zmm0 {%k1} 28 ; VLX-NEXT: vpcmpgtq %ymm1, %ymm0, %k1 29 ; VLX-NEXT: vpblendmq %ymm2, %ymm1, %ymm0 {%k1} 37 ; NoVLX-NEXT: vpcmpgtq %zmm1, %zmm0, %k1 38 ; NoVLX-NEXT: vpblendmq %zmm2, %zmm1, %zmm0 {%k1} 49 ; VLX-NEXT: vpcmpnltd %ymm1, %ymm0, %k1 50 ; VLX-NEXT: vpblendmd %ymm2, %ymm1, %ymm0 {%k1} [all …]
|
D | avx512vpopcntdq-schedule.ll | 9 ; GENERIC-NEXT: kmovw %esi, %k1 # sched: [1:0.33] 12 ; GENERIC-NEXT: vpopcntd %zmm1, %zmm0 {%k1} # sched: [1:0.50] 13 ; GENERIC-NEXT: vpopcntd %zmm1, %zmm0 {%k1} {z} # sched: [1:0.50] 15 ; GENERIC-NEXT: vpopcntd (%rdi), %zmm0 {%k1} # sched: [8:0.50] 16 ; GENERIC-NEXT: vpopcntd (%rdi), %zmm0 {%k1} {z} # sched: [8:0.50] 18 ; GENERIC-NEXT: vpopcntd (%rdi){1to16}, %zmm0 {%k1} # sched: [8:0.50] 19 ; GENERIC-NEXT: vpopcntd (%rdi){1to16}, %zmm0 {%k1} {z} # sched: [8:0.50] 26 ; ICELAKE-NEXT: kmovd %esi, %k1 # sched: [1:1.00] 29 ; ICELAKE-NEXT: vpopcntd %zmm1, %zmm0 {%k1} # sched: [1:1.00] 30 ; ICELAKE-NEXT: vpopcntd %zmm1, %zmm0 {%k1} {z} # sched: [1:1.00] [all …]
|
D | avx512-bugfix-26264.ll | 8 ; AVX512BW-NEXT: vpmovb2m %zmm0, %k1 9 ; AVX512BW-NEXT: vblendmpd (%rdi), %zmm1, %zmm0 {%k1} 10 ; AVX512BW-NEXT: kshiftrw $8, %k1, %k2 12 ; AVX512BW-NEXT: kshiftrd $16, %k1, %k1 13 ; AVX512BW-NEXT: vblendmpd 128(%rdi), %zmm3, %zmm2 {%k1} 14 ; AVX512BW-NEXT: kshiftrw $8, %k1, %k1 15 ; AVX512BW-NEXT: vblendmpd 192(%rdi), %zmm4, %zmm3 {%k1} 25 ; AVX512BW-NEXT: vpmovb2m %zmm0, %k1 26 ; AVX512BW-NEXT: vpblendmq (%rdi), %zmm1, %zmm0 {%k1} 27 ; AVX512BW-NEXT: kshiftrw $8, %k1, %k2 [all …]
|
D | avx512vbmi2-intrinsics-fast-isel.ll | 10 ; X86-NEXT: kmovd {{[0-9]+}}(%esp), %k1 11 ; X86-NEXT: vpcompressw %zmm1, %zmm0 {%k1} 16 ; X64-NEXT: kmovd %edi, %k1 17 ; X64-NEXT: vpcompressw %zmm1, %zmm0 {%k1} 30 ; X86-NEXT: kmovd {{[0-9]+}}(%esp), %k1 31 ; X86-NEXT: vpcompressw %zmm0, %zmm0 {%k1} {z} 36 ; X64-NEXT: kmovd %edi, %k1 37 ; X64-NEXT: vpcompressw %zmm0, %zmm0 {%k1} {z} 50 ; X86-NEXT: kmovd {{[0-9]+}}(%esp), %k1 51 ; X86-NEXT: kunpckdq %k1, %k0, %k1 [all …]
|
/external/webrtc/webrtc/modules/audio_coding/codecs/isac/fix/source/ |
D | fft.c | 74 int16_t k0, k1, k2, k3, k4, kk; in WebRtcIsacfix_FftRadix16Fastest() local 99 k1 = k0 + 60; in WebRtcIsacfix_FftRadix16Fastest() 100 k2 = k1 + 60; in WebRtcIsacfix_FftRadix16Fastest() 105 ajpQx = RexQx[k1] + RexQx[k3]; in WebRtcIsacfix_FftRadix16Fastest() 106 ajmQx = RexQx[k1] - RexQx[k3]; in WebRtcIsacfix_FftRadix16Fastest() 109 bjpQx = ImxQx[k1] + ImxQx[k3]; in WebRtcIsacfix_FftRadix16Fastest() 110 bjmQx = ImxQx[k1] - ImxQx[k3]; in WebRtcIsacfix_FftRadix16Fastest() 148 RexQx[k1] = (int16_t)WEBRTC_SPL_MUL_16_16_RSFT(ccc1Q14, akpQx, 14) - in WebRtcIsacfix_FftRadix16Fastest() 154 ImxQx[k1] = (int16_t)WEBRTC_SPL_MUL_16_16_RSFT(sss1Q14, akpQx, 14) + in WebRtcIsacfix_FftRadix16Fastest() 168 k1=20; in WebRtcIsacfix_FftRadix16Fastest() [all …]
|