/external/valgrind/memcheck/tests/ |
D | sh-mem.c | 15 typedef unsigned long long U8; typedef 24 U8 a [SZB_OF_a / 8]; // Type is U8 to ensure it's 8-aligned 25 U8 b [SZB_OF_a / 8]; // same size as a[] 38 U8 build(int size, U1 byte) in build() 41 U8 mask = 0; in build() 42 U8 shres; in build() 43 U8 res = 0xffffffffffffffffULL, res2; in build() 49 mask |= (U8)byte; in build() 109 assert(8 == sizeof(U8)); in main() 153 U8 tmp = build(NNN, j); \ in main() [all …]
|
D | sh-mem-random.c | 16 typedef unsigned long long U8; typedef 37 U8 build(int size, U1 byte) in build() 40 U8 mask = 0; in build() 41 U8 shres; in build() 42 U8 res = 0xffffffffffffffffULL, res2; in build() 48 mask |= (U8)byte; in build() 75 U8 sum = 0; in check() 85 sum += (U8)arr_i; in check() 159 *(U8*)(arr+dst) = *(U8*)(arr+src); in do_test_at()
|
D | sh-mem.stderr.exp | 16 -- NNN: 8 U8 U8 ------------------------ 25 -- NNN: 8 F8 U8 ------------------------
|
/external/libhevc/decoder/arm/ |
D | ihevcd_fmt_conv_420sp_to_rgba8888.s | 182 VSUBL.U8 Q2,D2,D1 @//(U-128) 183 VSUBL.U8 Q3,D3,D1 @//(V-128) 219 VADDW.U8 Q7,Q4,D30 @//Q7 - HAS Y + B 220 VADDW.U8 Q8,Q5,D30 @//Q8 - HAS Y + R 221 VADDW.U8 Q9,Q6,D30 @//Q9 - HAS Y + G 223 VADDW.U8 Q10,Q4,D31 @//Q10 - HAS Y + B 224 VADDW.U8 Q11,Q5,D31 @//Q11 - HAS Y + R 225 VADDW.U8 Q12,Q6,D31 @//Q12 - HAS Y + G 260 VADDW.U8 Q7,Q4,D28 @//Q7 - HAS Y + B 261 VADDW.U8 Q8,Q5,D28 @//Q2 - HAS Y + R [all …]
|
/external/libhevc/common/arm/ |
D | ihevc_sao_edge_offset_class0_chroma.s | 174 VCGT.U8 Q8,Q6,Q7 @vcgtq_u8(pu1_cur_row, pu1_cur_row_tmp) 178 VCLT.U8 Q9,Q6,Q7 @vcltq_u8(pu1_cur_row, pu1_cur_row_tmp) 181 … VSUB.U8 Q10,Q9,Q8 @sign_left = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 188 VCGT.U8 Q13,Q15,Q14 @II vcgtq_u8(pu1_cur_row, pu1_cur_row_tmp) 191 VCLT.U8 Q12,Q15,Q14 @II vcltq_u8(pu1_cur_row, pu1_cur_row_tmp) 202 VCGT.U8 Q8,Q6,Q7 @vcgtq_u8(pu1_cur_row, pu1_cur_row_tmp) 205 VCLT.U8 Q9,Q6,Q7 @vcltq_u8(pu1_cur_row, pu1_cur_row_tmp) 208 …VSUB.U8 Q11,Q9,Q8 @sign_right = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 211 VADD.U8 Q7,Q1,Q10 @edge_idx = vaddq_s8(const_2, sign_left) 213 VADD.U8 Q7,Q7,Q11 @edge_idx = vaddq_s8(edge_idx, sign_right) [all …]
|
D | ihevc_sao_edge_offset_class1.s | 142 VCGT.U8 Q6,Q5,Q4 @vcgtq_u8(pu1_cur_row, pu1_top_row) 145 VCLT.U8 Q7,Q5,Q4 @vcltq_u8(pu1_cur_row, pu1_top_row) 147 VSUB.U8 Q8,Q7,Q6 @sign_up = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 157 VCGT.U8 Q6,Q5,Q9 @vcgtq_u8(pu1_cur_row, pu1_top_row) 162 VCLT.U8 Q7,Q5,Q9 @vcltq_u8(pu1_cur_row, pu1_top_row) 165 … VSUB.U8 Q10,Q7,Q6 @sign_down = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 166 …VMOVL.U8 Q13,D18 @II pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u… 169 …VMOVL.U8 Q14,D19 @II pi2_tmp_cur_row.val[1] = vreinterpretq_s16_u16(vmovl_u… 172 VCGT.U8 Q11,Q9,Q15 @II vcgtq_u8(pu1_cur_row, pu1_top_row) 176 VCLT.U8 Q12,Q9,Q15 @II vcltq_u8(pu1_cur_row, pu1_top_row) [all …]
|
D | ihevc_sao_edge_offset_class1_chroma.s | 146 VCGT.U8 Q6,Q5,Q14 @vcgtq_u8(pu1_cur_row, pu1_top_row) 149 VCLT.U8 Q7,Q5,Q14 @vcltq_u8(pu1_cur_row, pu1_top_row) 151 VSUB.U8 Q8,Q7,Q6 @sign_up = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 161 VCGT.U8 Q6,Q5,Q9 @vcgtq_u8(pu1_cur_row, pu1_top_row) 166 VCLT.U8 Q7,Q5,Q9 @vcltq_u8(pu1_cur_row, pu1_top_row) 169 … VSUB.U8 Q10,Q7,Q6 @sign_down = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 170 …VMOVL.U8 Q13,D18 @II pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u… 173 …VMOVL.U8 Q14,D19 @II pi2_tmp_cur_row.val[1] = vreinterpretq_s16_u16(vmovl_u… 176 VCGT.U8 Q11,Q9,Q15 @II vcgtq_u8(pu1_cur_row, pu1_top_row) 180 VCLT.U8 Q12,Q9,Q15 @II vcltq_u8(pu1_cur_row, pu1_top_row) [all …]
|
D | ihevc_sao_band_offset_chroma.s | 166 VCLE.U8 D13,D4,D30 @vcle_u8(band_table.val[3], vdup_n_u8(16)) 169 VORR.U8 D4,D4,D13 @band_table.val[3] = vorr_u8(band_table.val[3], au1_cmp) 175 VCLE.U8 D14,D3,D30 @vcle_u8(band_table.val[2], vdup_n_u8(16)) 177 VORR.U8 D3,D3,D14 @band_table.val[2] = vorr_u8(band_table.val[2], au1_cmp) 179 VAND.U8 D4,D4,D13 @band_table.val[3] = vand_u8(band_table.val[3], au1_cmp) 185 VCLE.U8 D15,D2,D30 @vcle_u8(band_table.val[1], vdup_n_u8(16)) 187 VORR.U8 D2,D2,D15 @band_table.val[1] = vorr_u8(band_table.val[1], au1_cmp) 189 VAND.U8 D3,D3,D14 @band_table.val[2] = vand_u8(band_table.val[2], au1_cmp) 195 VCLE.U8 D16,D1,D30 @vcle_u8(band_table.val[0], vdup_n_u8(16)) 196 VORR.U8 D1,D1,D16 @band_table.val[0] = vorr_u8(band_table.val[0], au1_cmp) [all …]
|
D | ihevc_sao_band_offset_luma.s | 150 VCLE.U8 D12,D4,D29 @vcle_u8(band_table.val[3], vdup_n_u8(16)) 153 VORR.U8 D4,D4,D12 @band_table.val[3] = vorr_u8(band_table.val[3], au1_cmp) 158 VCLE.U8 D11,D3,D29 @vcle_u8(band_table.val[2], vdup_n_u8(16)) 161 VORR.U8 D3,D3,D11 @band_table.val[2] = vorr_u8(band_table.val[2], au1_cmp) 163 VAND.U8 D4,D4,D12 @band_table.val[3] = vand_u8(band_table.val[3], au1_cmp) 168 VCLE.U8 D10,D2,D29 @vcle_u8(band_table.val[1], vdup_n_u8(16)) 171 VORR.U8 D2,D2,D10 @band_table.val[1] = vorr_u8(band_table.val[1], au1_cmp) 173 VAND.U8 D3,D3,D11 @band_table.val[2] = vand_u8(band_table.val[2], au1_cmp) 180 VCLE.U8 D9,D1,D29 @vcle_u8(band_table.val[0], vdup_n_u8(16)) 181 VORR.U8 D1,D1,D9 @band_table.val[0] = vorr_u8(band_table.val[0], au1_cmp) [all …]
|
D | ihevc_sao_edge_offset_class0.s | 167 VCGT.U8 Q8,Q6,Q7 @vcgtq_u8(pu1_cur_row, pu1_cur_row_tmp) 172 VCLT.U8 Q9,Q6,Q7 @vcltq_u8(pu1_cur_row, pu1_cur_row_tmp) 184 VCGT.U8 Q15,Q13,Q14 @II vcgtq_u8(pu1_cur_row, pu1_cur_row_tmp) 190 VCLT.U8 Q0,Q13,Q14 @II vcltq_u8(pu1_cur_row, pu1_cur_row_tmp) 195 VCGT.U8 Q8,Q6,Q7 @vcgtq_u8(pu1_cur_row, pu1_cur_row_tmp) 198 VCLT.U8 Q9,Q6,Q7 @vcltq_u8(pu1_cur_row, pu1_cur_row_tmp) 209 …VMOVL.U8 Q9,D12 @pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(v… 213 VCGT.U8 Q15,Q13,Q14 @II vcgtq_u8(pu1_cur_row, pu1_cur_row_tmp) 215 VCLT.U8 Q0,Q13,Q14 @II vcltq_u8(pu1_cur_row, pu1_cur_row_tmp) 221 …VMOVL.U8 Q0,D26 @II pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u… [all …]
|
D | ihevc_sao_edge_offset_class2.s | 257 VCGT.U8 Q7,Q6,Q5 @vcgtq_u8(pu1_cur_row, pu1_top_row) 261 VCLT.U8 Q8,Q6,Q5 @vcltq_u8(pu1_cur_row, pu1_top_row) 265 VSUB.U8 Q7,Q8,Q7 @sign_up = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 302 VCGT.U8 Q5,Q6,Q9 @I vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) 305 VCLT.U8 Q9,Q6,Q9 @I vcltq_u8(pu1_cur_row, pu1_next_row_tmp) 306 …VSUB.U8 Q5,Q9,Q5 @I sign_down = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt… 318 …VMOVL.U8 Q10,D12 @I pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8… 323 …VMOVL.U8 Q11,D13 @I pi2_tmp_cur_row.val[1] = vreinterpretq_s16_u16(vmovl_u8… 369 VCGT.U8 Q12,Q6,Q11 @II vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) 373 VCLT.U8 Q11,Q6,Q11 @II vcltq_u8(pu1_cur_row, pu1_next_row_tmp) [all …]
|
D | ihevc_sao_edge_offset_class3.s | 272 VCGT.U8 Q7,Q6,Q5 @vcgtq_u8(pu1_cur_row, pu1_top_row) 282 VCLT.U8 Q8,Q6,Q5 @vcltq_u8(pu1_cur_row, pu1_top_row) 285 VSUB.U8 Q7,Q8,Q7 @sign_up = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 317 VCGT.U8 Q5,Q6,Q9 @I vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) 318 VCLT.U8 Q9,Q6,Q9 @I vcltq_u8(pu1_cur_row, pu1_next_row_tmp) 319 …VSUB.U8 Q5,Q9,Q5 @I sign_down = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt… 329 …VMOVL.U8 Q10,D12 @I pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8… 334 …VMOVL.U8 Q11,D13 @I pi2_tmp_cur_row.val[1] = vreinterpretq_s16_u16(vmovl_u8… 394 VCGT.U8 Q12,Q6,Q9 @II vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) 398 VCLT.U8 Q13,Q6,Q9 @II vcltq_u8(pu1_cur_row, pu1_next_row_tmp) [all …]
|
D | ihevc_sao_edge_offset_class3_chroma.s | 340 VCGT.U8 Q7,Q6,Q5 @vcgtq_u8(pu1_cur_row, pu1_top_row) 344 VCLT.U8 Q8,Q6,Q5 @vcltq_u8(pu1_cur_row, pu1_top_row) 348 VSUB.U8 Q7,Q8,Q7 @sign_up = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 404 VCGT.U8 Q10,Q6,Q9 @I vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) 406 VCLT.U8 Q11,Q6,Q9 @I vcltq_u8(pu1_cur_row, pu1_next_row_tmp) 407 …VSUB.U8 Q11,Q11,Q10 @I sign_down = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt… 417 …VMOVL.U8 Q10,D12 @I pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8… 425 …VMOVL.U8 Q9,D13 @I pi2_tmp_cur_row.val[1] = vreinterpretq_s16_u16(vmovl_u8… 487 VCGT.U8 Q11,Q6,Q14 @II vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) 497 VCLT.U8 Q12,Q6,Q14 @II vcltq_u8(pu1_cur_row, pu1_next_row_tmp) [all …]
|
D | ihevc_sao_edge_offset_class2_chroma.s | 350 VCGT.U8 Q7,Q6,Q5 @vcgtq_u8(pu1_cur_row, pu1_top_row) 354 VCLT.U8 Q8,Q6,Q5 @vcltq_u8(pu1_cur_row, pu1_top_row) 366 VSUB.U8 Q7,Q8,Q7 @sign_up = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 412 VCGT.U8 Q10,Q6,Q9 @I vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) 414 VCLT.U8 Q11,Q6,Q9 @I vcltq_u8(pu1_cur_row, pu1_next_row_tmp) 415 …VSUB.U8 Q11,Q11,Q10 @I sign_down = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt… 426 …VMOVL.U8 Q10,D12 @I pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8… 429 …VMOVL.U8 Q9,D13 @I pi2_tmp_cur_row.val[1] = vreinterpretq_s16_u16(vmovl_u8… 486 VCGT.U8 Q11,Q6,Q14 @II vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) 494 VCLT.U8 Q12,Q6,Q14 @II vcltq_u8(pu1_cur_row, pu1_next_row_tmp) [all …]
|
/external/syslinux/gpxe/src/drivers/net/phantom/ |
D | nxhal_nic_interface.h | 316 U8 reserved[128]; /* future expansion */ 327 U8 phys_port; /* Physical id of port */ 328 U8 virt_port; /* Virtual/Logical id of port */ 330 U8 reserved[128]; /* future expansion */ 403 U8 reserved[128]; /* reserve space for future expansion*/ 430 U8 phys_port; /* Physical id of port */ 431 U8 virt_port; /* Virtual/Logical id of port */ 432 U8 reserved[128]; /* save space for future expansion */
|
D | phantom.h | 42 typedef uint8_t U8; typedef
|
/external/compiler-rt/lib/asan/tests/ |
D | asan_asm_test.cc | 68 DECLARE_ASM_WRITE(U8, "8", "movq", "r"); 69 DECLARE_ASM_READ(U8, "8", "movq", "=r"); 70 DECLARE_ASM_REP_MOVS(U8, "movsq"); 222 TestAsmWrite<U8>("WRITE of size 8"); in TEST() 230 TestAsmRead<U8>("READ of size 8"); in TEST() 265 TestAsmRepMovs<U8>("READ of size 8", "WRITE of size 8"); in TEST()
|
D | asan_test_utils.h | 61 typedef uint64_t U8; typedef
|
/external/skia/src/jumper/ |
D | SkJumper_stages_lowp.cpp | 28 using U8 = V<uint8_t>; typedef 183 SI F from_byte(U8 bytes) { in from_byte() 196 SI U8 to_byte(F v) { in to_byte() 204 return unaligned_load<U8>(&packed); in to_byte() 287 a = from_byte(load<U8>(ptr, tail)); in STAGE() 292 da = from_byte(load<U8>(ptr, tail)); in STAGE() 301 r = g = b = from_byte(load<U8>(ptr, tail)); in STAGE() 307 dr = dg = db = from_byte(load<U8>(ptr, tail)); in STAGE() 335 U8 scales = load<U8>(ptr, tail); in STAGE() 355 U8 scales = load<U8>(ptr, tail); in STAGE()
|
D | SkJumper_vectors.h | 28 using U8 = uint8_t ; variable 40 SI U8 pack(U16 v) { return (U8)v; } in pack() 88 using U8 = V<uint8_t >; variable 101 SI U8 pack(U16 v) { return __builtin_convertvector(v, U8); } in pack() 183 using U8 = V<uint8_t >; variable 193 SI U8 pack(U16 v) { return __builtin_convertvector(v, U8); } in pack() 288 using U8 = V<uint8_t >; variable 311 SI U8 pack(U16 v) { in pack() 313 return unaligned_load<U8>(&r); in pack() 501 using U8 = V<uint8_t >; variable [all …]
|
D | SkJumper_stages.cpp | 186 SI F from_byte(U8 b) { in from_byte() 673 auto scales = load<U8>(ptr, tail); in STAGE() 697 auto scales = load<U8>(ptr, tail); in STAGE() 814 a = from_byte(load<U8>(ptr, tail)); in STAGE() 820 da = from_byte(load<U8>(ptr, tail)); in STAGE() 831 U8 packed = pack(pack(round(a, 255.0f))); in STAGE() 838 r = g = b = from_byte(load<U8>(ptr, tail)); in STAGE() 844 dr = dg = db = from_byte(load<U8>(ptr, tail)); in STAGE()
|
/external/compiler-rt/lib/msan/tests/ |
D | msan_test.cc | 127 typedef unsigned long long U8; // NOLINT typedef 184 static U8 poisoned_array[100]; 468 EXPECT_UMR(break_optimization((void*)(U8)a[GetPoisonedZero()])); in TEST() 1540 TestOverlapMemmove<U8, 4>(); in TEST() 1541 TestOverlapMemmove<U8, 1000>(); in TEST() 2707 U8 b; 2783 U8 a8, b8; 3681 U8 y8 = 0; in TEST() 3750 U8 y = 0; in TEST() 3782 U8 y = 0; in TEST() [all …]
|
/external/ImageMagick/PerlMagick/t/reference/write/filter/ |
D | Colorize.miff | 43 …U8�U7�X8�\7�U8�[L�__�%�%�%�&�&�&�&�(�(�&�#�.8�CZ�CS�:8�%*�(�#��������& …
|
/external/svox/pico_resources/tools/LingwareBuilding/PicoLingware_source_files/pkb/de-DE/ |
D | de-DE_gl0_kdt_mgc3.pkb | 70 ���U8�� '��zu�0@�8��X4L4��6
|
/external/valgrind/VEX/priv/ |
D | ir_opt.c | 1099 case Ico_U8: return toBool( c1->Ico.U8 == c2->Ico.U8 ); in sameIRExprs_aux2() 1222 case Ico_U8: return toBool( e->Iex.Const.con->Ico.U8 == 0); in isZeroU() 1236 case Ico_U8: return toBool( e->Iex.Const.con->Ico.U8 == 0xFF); in isOnesU() 1419 UInt u32 = e->Iex.Unop.arg->Iex.Const.con->Ico.U8; in fold_Expr() 1434 0xFFULL & e->Iex.Unop.arg->Iex.Const.con->Ico.U8)); in fold_Expr() 1442 0xFF & e->Iex.Unop.arg->Iex.Const.con->Ico.U8)); in fold_Expr() 1445 UShort u16 = e->Iex.Unop.arg->Iex.Const.con->Ico.U8; in fold_Expr() 1453 0xFF & e->Iex.Unop.arg->Iex.Const.con->Ico.U8)); in fold_Expr() 1496 ~ (e->Iex.Unop.arg->Iex.Const.con->Ico.U8)))); in fold_Expr() 1565 (0xFF & e->Iex.Unop.arg->Iex.Const.con->Ico.U8) in fold_Expr() [all …]
|