/third_party/glslang/Test/ |
D | spv.int8.frag | 19 int8_t i8; 42 int8_t i8 = i8Const[index]; 115 int8_t i8; 123 i8--; 124 ++i8; 129 i8 = +i8; 133 i8 += i8; 135 i *= i8; 137 uv %= i8; 140 i16 = i8 - i16; [all …]
|
D | spv.subgroupExtendedTypesArithmeticNeg.comp | 13 i8vec4 i8; 26 data[invocation].i8.x = subgroupAdd(data[0].i8.x); 27 data[invocation].i8.xy = subgroupAdd(data[1].i8.xy); 28 data[invocation].i8.xyz = subgroupAdd(data[2].i8.xyz); 29 data[invocation].i8 = subgroupAdd(data[3].i8); 31 data[invocation].i8.x = subgroupMul(data[0].i8.x); 32 data[invocation].i8.xy = subgroupMul(data[1].i8.xy); 33 data[invocation].i8.xyz = subgroupMul(data[2].i8.xyz); 34 data[invocation].i8 = subgroupMul(data[3].i8); 36 data[invocation].i8.x = subgroupMin(data[0].i8.x); [all …]
|
D | spv.subgroupExtendedTypesArithmetic.comp | 13 i8vec4 i8; 26 data[invocation].i8.x = subgroupAdd(data[0].i8.x); 27 data[invocation].i8.xy = subgroupAdd(data[1].i8.xy); 28 data[invocation].i8.xyz = subgroupAdd(data[2].i8.xyz); 29 data[invocation].i8 = subgroupAdd(data[3].i8); 31 data[invocation].i8.x = subgroupMul(data[0].i8.x); 32 data[invocation].i8.xy = subgroupMul(data[1].i8.xy); 33 data[invocation].i8.xyz = subgroupMul(data[2].i8.xyz); 34 data[invocation].i8 = subgroupMul(data[3].i8); 36 data[invocation].i8.x = subgroupMin(data[0].i8.x); [all …]
|
D | spv.subgroupExtendedTypesClustered.comp | 13 i8vec4 i8; 26 data[invocation].i8.x = subgroupClusteredAdd(data[0].i8.x, 1); 27 data[invocation].i8.xy = subgroupClusteredAdd(data[1].i8.xy, 1); 28 data[invocation].i8.xyz = subgroupClusteredAdd(data[2].i8.xyz, 1); 29 data[invocation].i8 = subgroupClusteredAdd(data[3].i8, 1); 31 data[invocation].i8.x = subgroupClusteredMul(data[0].i8.x, 1); 32 data[invocation].i8.xy = subgroupClusteredMul(data[1].i8.xy, 1); 33 data[invocation].i8.xyz = subgroupClusteredMul(data[2].i8.xyz, 1); 34 data[invocation].i8 = subgroupClusteredMul(data[3].i8, 1); 36 data[invocation].i8.x = subgroupClusteredMin(data[0].i8.x, 1); [all …]
|
D | spv.subgroupExtendedTypesClusteredNeg.comp | 13 i8vec4 i8; 26 data[invocation].i8.x = subgroupClusteredAdd(data[0].i8.x, 1); 27 data[invocation].i8.xy = subgroupClusteredAdd(data[1].i8.xy, 1); 28 data[invocation].i8.xyz = subgroupClusteredAdd(data[2].i8.xyz, 1); 29 data[invocation].i8 = subgroupClusteredAdd(data[3].i8, 1); 31 data[invocation].i8.x = subgroupClusteredMul(data[0].i8.x, 1); 32 data[invocation].i8.xy = subgroupClusteredMul(data[1].i8.xy, 1); 33 data[invocation].i8.xyz = subgroupClusteredMul(data[2].i8.xyz, 1); 34 data[invocation].i8 = subgroupClusteredMul(data[3].i8, 1); 36 data[invocation].i8.x = subgroupClusteredMin(data[0].i8.x, 1); [all …]
|
D | spv.subgroupExtendedTypesPartitionedNeg.comp | 13 i8vec4 i8; 27 ballot = subgroupPartitionNV(data[0].i8.x); 28 ballot = subgroupPartitionNV(data[1].i8.xy); 29 ballot = subgroupPartitionNV(data[2].i8.xyz); 30 ballot = subgroupPartitionNV(data[3].i8); 62 data[invocation].i8.x = subgroupPartitionedAddNV(data[0].i8.x, ballot); 63 data[invocation].i8.xy = subgroupPartitionedAddNV(data[1].i8.xy, ballot); 64 data[invocation].i8.xyz = subgroupPartitionedAddNV(data[2].i8.xyz, ballot); 65 data[invocation].i8 = subgroupPartitionedAddNV(data[3].i8, ballot); 67 data[invocation].i8.x = subgroupPartitionedMulNV(data[0].i8.x, ballot); [all …]
|
D | spv.subgroupExtendedTypesPartitioned.comp | 13 i8vec4 i8; 27 ballot = subgroupPartitionNV(data[0].i8.x); 28 ballot = subgroupPartitionNV(data[1].i8.xy); 29 ballot = subgroupPartitionNV(data[2].i8.xyz); 30 ballot = subgroupPartitionNV(data[3].i8); 62 data[invocation].i8.x = subgroupPartitionedAddNV(data[0].i8.x, ballot); 63 data[invocation].i8.xy = subgroupPartitionedAddNV(data[1].i8.xy, ballot); 64 data[invocation].i8.xyz = subgroupPartitionedAddNV(data[2].i8.xyz, ballot); 65 data[invocation].i8 = subgroupPartitionedAddNV(data[3].i8, ballot); 67 data[invocation].i8.x = subgroupPartitionedMulNV(data[0].i8.x, ballot); [all …]
|
D | spv.subgroupExtendedTypesQuad.comp | 13 i8vec4 i8; 26 data[invocation].i8.x = subgroupQuadBroadcast(data[0].i8.x, 1); 27 data[invocation].i8.xy = subgroupQuadBroadcast(data[1].i8.xy, 1); 28 data[invocation].i8.xyz = subgroupQuadBroadcast(data[2].i8.xyz, 1); 29 data[invocation].i8 = subgroupQuadBroadcast(data[3].i8, 1); 31 data[invocation].i8.x = subgroupQuadSwapHorizontal(data[0].i8.x); 32 data[invocation].i8.xy = subgroupQuadSwapHorizontal(data[1].i8.xy); 33 data[invocation].i8.xyz = subgroupQuadSwapHorizontal(data[2].i8.xyz); 34 data[invocation].i8 = subgroupQuadSwapHorizontal(data[3].i8); 36 data[invocation].i8.x = subgroupQuadSwapVertical(data[0].i8.x); [all …]
|
D | spv.subgroupExtendedTypesQuadNeg.comp | 13 i8vec4 i8; 26 data[invocation].i8.x = subgroupQuadBroadcast(data[0].i8.x, 1); 27 data[invocation].i8.xy = subgroupQuadBroadcast(data[1].i8.xy, 1); 28 data[invocation].i8.xyz = subgroupQuadBroadcast(data[2].i8.xyz, 1); 29 data[invocation].i8 = subgroupQuadBroadcast(data[3].i8, 1); 31 data[invocation].i8.x = subgroupQuadSwapHorizontal(data[0].i8.x); 32 data[invocation].i8.xy = subgroupQuadSwapHorizontal(data[1].i8.xy); 33 data[invocation].i8.xyz = subgroupQuadSwapHorizontal(data[2].i8.xyz); 34 data[invocation].i8 = subgroupQuadSwapHorizontal(data[3].i8); 36 data[invocation].i8.x = subgroupQuadSwapVertical(data[0].i8.x); [all …]
|
D | spv.subgroupExtendedTypesBallotNeg.comp | 13 i8vec4 i8; 26 data[invocation].i8.x = subgroupBroadcast(data[0].i8.x, 3); 27 data[invocation].i8.xy = subgroupBroadcast(data[1].i8.xy, 3); 28 data[invocation].i8.xyz = subgroupBroadcast(data[2].i8.xyz, 3); 29 data[invocation].i8 = subgroupBroadcast(data[3].i8, 3); 30 data[invocation].i8.x = subgroupBroadcastFirst(data[0].i8.x); 31 data[invocation].i8.xy = subgroupBroadcastFirst(data[1].i8.xy); 32 data[invocation].i8.xyz = subgroupBroadcastFirst(data[2].i8.xyz); 33 data[invocation].i8 = subgroupBroadcastFirst(data[3].i8);
|
D | spv.subgroupExtendedTypesBallot.comp | 13 i8vec4 i8; 26 data[invocation].i8.x = subgroupBroadcast(data[0].i8.x, 3); 27 data[invocation].i8.xy = subgroupBroadcast(data[1].i8.xy, 3); 28 data[invocation].i8.xyz = subgroupBroadcast(data[2].i8.xyz, 3); 29 data[invocation].i8 = subgroupBroadcast(data[3].i8, 3); 30 data[invocation].i8.x = subgroupBroadcastFirst(data[0].i8.x); 31 data[invocation].i8.xy = subgroupBroadcastFirst(data[1].i8.xy); 32 data[invocation].i8.xyz = subgroupBroadcastFirst(data[2].i8.xyz); 33 data[invocation].i8 = subgroupBroadcastFirst(data[3].i8);
|
/third_party/skia/third_party/externals/swiftshader/third_party/subzero/crosstest/ |
D | test_vector_ops_ll.ll | 102 define <16 x i8> @insertelement_v16i1(<16 x i8> %arg_vec, i64 %elt_arg, i32 %idx) { 103 %vec = trunc <16 x i8> %arg_vec to <16 x i1> 125 %res0 = zext <16 x i1> %res0_i1 to <16 x i8> 126 ret <16 x i8> %res0 129 %res1 = zext <16 x i1> %res1_i1 to <16 x i8> 130 ret <16 x i8> %res1 133 %res2 = zext <16 x i1> %res2_i1 to <16 x i8> 134 ret <16 x i8> %res2 137 %res3 = zext <16 x i1> %res3_i1 to <16 x i8> 138 ret <16 x i8> %res3 [all …]
|
D | test_arith_ll.ll | 11 %a.trunc = trunc i32 %a to i8 12 %b.trunc = trunc i32 %b to i8 13 %result.trunc = add i8 %a.trunc, %b.trunc 14 %result = zext i8 %result.trunc to i32 58 define internal <16 x i8> @_Z7testAddDv16_hS_(<16 x i8> %a, <16 x i8> %b) { 61 %result = add <16 x i8> %a, %b 63 ret <16 x i8> %result 75 %a.trunc = trunc i32 %a to i8 76 %b.trunc = trunc i32 %b to i8 77 %result.trunc = sub i8 %a.trunc, %b.trunc [all …]
|
D | test_icmp_i1vec.ll | 1 define <16 x i8> @_Z8icmpi1EqDv16_aS_(<16 x i8> %a, <16 x i8> %b) { 3 %a.trunc = trunc <16 x i8> %a to <16 x i1> 4 %b.trunc = trunc <16 x i8> %b to <16 x i1> 6 %cmp.sext = sext <16 x i1> %cmp to <16 x i8> 7 ret <16 x i8> %cmp.sext 10 define <16 x i8> @_Z8icmpi1NeDv16_aS_(<16 x i8> %a, <16 x i8> %b) { 12 %a.trunc = trunc <16 x i8> %a to <16 x i1> 13 %b.trunc = trunc <16 x i8> %b to <16 x i1> 15 %cmp.sext = sext <16 x i1> %cmp to <16 x i8> 16 ret <16 x i8> %cmp.sext [all …]
|
D | test_select.ll | 36 define <16 x i8> @_Z6selectDv16_aS_S_(<16 x i8> %cond.ext, <16 x i8> %arg1, <16 x i8> %arg2) { 38 %cond = trunc <16 x i8> %cond.ext to <16 x i1> 39 %res = select <16 x i1> %cond, <16 x i8> %arg1, <16 x i8> %arg2 40 ret <16 x i8> %res 43 define <16 x i8> @_Z6selectDv16_aDv16_hS0_(<16 x i8> %cond.ext, <16 x i8> %arg1, <16 x i8> %arg2) { 45 %cond = trunc <16 x i8> %cond.ext to <16 x i1> 46 %res = select <16 x i1> %cond, <16 x i8> %arg1, <16 x i8> %arg2 47 ret <16 x i8> %res 70 define <16 x i8> @_Z9select_i1Dv16_aS_S_(<16 x i8> %cond.ext, <16 x i8> %arg1.ext, <16 x i8> %arg2.… 72 %cond = trunc <16 x i8> %cond.ext to <16 x i1> [all …]
|
/third_party/skia/third_party/externals/swiftshader/third_party/subzero/unittest/AssemblerX8664/ |
D | LowLevel.cpp | 311 TestRegReg(cmp, eax, ecx, i8, 2, 0x3A, 0xC1); in TEST_F() 312 TestRegReg(cmp, ecx, edx, i8, 2, 0x3A, 0xCA); in TEST_F() 313 TestRegReg(cmp, edx, ebx, i8, 2, 0x3A, 0xD3); in TEST_F() 314 TestRegReg(cmp, ebx, esp, i8, 2, 0x3A, 0xDC); // emit: cmp bl, ah in TEST_F() 315 TestRegReg(cmp, esp, ebp, i8, 3, 0x40, 0x3A, 0xE5); in TEST_F() 316 TestRegReg(cmp, ebp, esi, i8, 3, 0x40, 0x3A, 0xEE); in TEST_F() 317 TestRegReg(cmp, esi, edi, i8, 3, 0x40, 0x3A, 0xF7); in TEST_F() 318 TestRegReg(cmp, edi, r8, i8, 3, 0x41, 0x3A, 0xF8); in TEST_F() 319 TestRegReg(cmp, r8, r9, i8, 3, 0x45, 0x3A, 0xC1); in TEST_F() 320 TestRegReg(cmp, r9, r10, i8, 3, 0x45, 0x3A, 0xCA); in TEST_F() [all …]
|
/third_party/ffmpeg/libavcodec/ |
D | h264_direct.c | 211 int i8, i4; in pred_spatial_direct_motion() local 360 for (i8 = 0; i8 < 4; i8++) { in pred_spatial_direct_motion() 361 int x8 = i8 & 1; in pred_spatial_direct_motion() 362 int y8 = i8 >> 1; in pred_spatial_direct_motion() 367 if (is_b8x8 && !IS_DIRECT(sl->sub_mb_type[i8])) in pred_spatial_direct_motion() 369 sl->sub_mb_type[i8] = sub_mb_type; in pred_spatial_direct_motion() 371 fill_rectangle(&sl->ref_cache[0][scan8[i8 * 4]], 2, 2, 8, in pred_spatial_direct_motion() 373 fill_rectangle(&sl->ref_cache[1][scan8[i8 * 4]], 2, 2, 8, in pred_spatial_direct_motion() 394 fill_rectangle(&sl->mv_cache[0][scan8[i8 * 4]], 2, 2, 8, a, 4); in pred_spatial_direct_motion() 395 fill_rectangle(&sl->mv_cache[1][scan8[i8 * 4]], 2, 2, 8, b, 4); in pred_spatial_direct_motion() [all …]
|
/third_party/skia/third_party/externals/swiftshader/third_party/subzero/unittest/AssemblerX8632/ |
D | LowLevel.cpp | 254 TestRegReg(cmp, eax, ecx, i8, 2, 0x3A, 0xC1); in TEST_F() 255 TestRegReg(cmp, ecx, edx, i8, 2, 0x3A, 0xCA); in TEST_F() 256 TestRegReg(cmp, edx, ebx, i8, 2, 0x3A, 0xD3); in TEST_F() 257 TestRegReg(cmp, ebx, esp, i8, 2, 0x3A, 0xDC); in TEST_F() 258 TestRegReg(cmp, esp, ebp, i8, 2, 0x3A, 0xE5); in TEST_F() 259 TestRegReg(cmp, ebp, esi, i8, 2, 0x3A, 0xEE); in TEST_F() 260 TestRegReg(cmp, esi, edi, i8, 2, 0x3A, 0xF7); in TEST_F() 261 TestRegReg(cmp, edi, eax, i8, 2, 0x3A, 0xF8); in TEST_F() 282 TestRegImm(cmp, eax, 5, i8, 2, 0x3C, 0x05); in TEST_F() 283 TestRegImm(cmp, ecx, 5, i8, 3, 0x80, 0xF9, 0x05); in TEST_F() [all …]
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/PowerPC/ |
D | README_ALTIVEC.txt | 218 define <16 x i8> @test_vpmsumb() #0 { 220 %a = alloca <16 x i8>, align 16 221 %b = alloca <16 x i8>, align 16 222 …store <16 x i8> <i8 1, i8 2, i8 3, i8 4, i8 5, i8 6, i8 7, i8 8, i8 9, i8 10, i8 11, i8 12, i8 13,… 223 …16 x i8> <i8 113, i8 114, i8 115, i8 116, i8 117, i8 118, i8 119, i8 120, i8 121, i8 122, i8 123, … 224 %0 = load <16 x i8>* %a, align 16 225 %1 = load <16 x i8>* %b, align 16 226 %2 = call <16 x i8> @llvm.ppc.altivec.crypto.vpmsumb(<16 x i8> %0, <16 x i8> %1) 227 ret <16 x i8> %2 232 declare <16 x i8> @llvm.ppc.altivec.crypto.vpmsumb(<16 x i8>, <16 x i8>) #1
|
/third_party/glslang/Test/baseResults/ |
D | spv.int8.frag.out | 32 Name 16 "i8" 52 Name 197 "i8" 58 Name 341 "i8" 70 MemberName 513(Block) 0 "i8" 198 16(i8): 15(ptr) Variable Function 207 Store 16(i8) 35 364 197(i8): 15(ptr) Variable Function 373 198: 14(int8_t) Load 197(i8) 375 Store 197(i8) 199 376 200: 14(int8_t) Load 197(i8) [all …]
|
/third_party/skia/third_party/externals/spirv-cross/reference/opt/shaders/vulkan/frag/ |
D | shader-arithmetic-8bit.nocompat.vk.frag.vk | 7 int8_t i8[16]; 13 int8_t i8; 19 int8_t i8; 30 ssbo.i8[0] = _199.x; 31 ssbo.i8[1] = _199.y; 32 ssbo.i8[2] = _199.z; 33 ssbo.i8[3] = _199.w; 40 …(registers.i8)) + i8vec4(-40)) + i8vec4(-50)) + i8vec4(int8_t(10), int8_t(20), int8_t(30), int8_t(…
|
/third_party/skia/third_party/externals/spirv-cross/reference/opt/shaders-msl/frag/ |
D | shader-arithmetic-8bit.frag | 8 char i8[16]; 14 char i8; 20 char i8; 39 ssbo.i8[0] = _199.x; 40 ssbo.i8[1] = _199.y; 41 ssbo.i8[2] = _199.z; 42 ssbo.i8[3] = _199.w; 49 …+ char4(registers.i8)) + char4(-40)) + char4(-50)) + char4(char(10), char(20), char(30), char(40))…
|
/third_party/typescript/tests/cases/compiler/ |
D | intTypeCheck.ts | 43 interface i8 extends i4 { } interface 195 var obj77: i8; 196 var obj78: i8 = {}; 197 var obj79: i8 = new Object(); 198 var obj80: i8 = new obj77; 199 var obj81: i8 = new Base; 200 var obj82: i8 = null; 201 var obj83: i8 = function () { }; 203 var obj85: i8 = <i8> anyVar; 204 var obj86: i8 = new <i8> anyVar; [all …]
|
/third_party/skia/third_party/externals/spirv-cross/shaders/vulkan/frag/ |
D | shader-arithmetic-8bit.nocompat.vk.frag | 11 int8_t i8; 17 int8_t i8; 23 int8_t i8[16]; 36 ssbo.i8[0] = i8_4.x; 37 ssbo.i8[1] = i8_4.y; 38 ssbo.i8[2] = i8_4.z; 39 ssbo.i8[3] = i8_4.w; 61 tmp += registers.i8; 65 tmp += ssbo.i8[4]; 66 tmp += ubo.i8;
|
/third_party/skia/third_party/externals/spirv-cross/shaders-msl/frag/ |
D | shader-arithmetic-8bit.frag | 11 int8_t i8; 17 int8_t i8; 23 int8_t i8[16]; 36 ssbo.i8[0] = i8_4.x; 37 ssbo.i8[1] = i8_4.y; 38 ssbo.i8[2] = i8_4.z; 39 ssbo.i8[3] = i8_4.w; 61 tmp += registers.i8; 65 tmp += ssbo.i8[4]; 66 tmp += ubo.i8;
|