/external/rust/crates/uuid/src/adapter/ |
D | compact.rs | 60 serde_test::Token::U8(uuid_bytes[0]), in test_serialize_compact() 61 serde_test::Token::U8(uuid_bytes[1]), in test_serialize_compact() 62 serde_test::Token::U8(uuid_bytes[2]), in test_serialize_compact() 63 serde_test::Token::U8(uuid_bytes[3]), in test_serialize_compact() 64 serde_test::Token::U8(uuid_bytes[4]), in test_serialize_compact() 65 serde_test::Token::U8(uuid_bytes[5]), in test_serialize_compact() 66 serde_test::Token::U8(uuid_bytes[6]), in test_serialize_compact() 67 serde_test::Token::U8(uuid_bytes[7]), in test_serialize_compact() 68 serde_test::Token::U8(uuid_bytes[8]), in test_serialize_compact() 69 serde_test::Token::U8(uuid_bytes[9]), in test_serialize_compact() [all …]
|
/external/llvm-project/clang/test/Analysis/ |
D | conversion.c | 3 unsigned char U8; variable 8 U8 = S; // expected-warning {{Loss of sign in implicit conversion}} in assign() 12 U8 = S; // no-warning in assign() 20 U8 += L; // expected-warning {{Loss of precision in implicit conversion}} in addAssign() 27 U8 -= L; // expected-warning {{Loss of precision in implicit conversion}} in subAssign() 34 U8 *= L; // expected-warning {{Loss of precision in implicit conversion}} in mulAssign() 43 U8 /= L; // no-warning in divAssign() 50 U8 %= L; // no-warning in remAssign() 57 U8 &= L; // no-warning in andAssign() 64 U8 |= L; // expected-warning {{Loss of precision in implicit conversion}} in orAssign() [all …]
|
/external/libhevc/decoder/arm/ |
D | ihevcd_fmt_conv_420sp_to_rgba8888.s | 182 VSUBL.U8 Q2,D2,D1 @//(U-128) 183 VSUBL.U8 Q3,D3,D1 @//(V-128) 219 VADDW.U8 Q7,Q4,D30 @//Q7 - HAS Y + B 220 VADDW.U8 Q8,Q5,D30 @//Q8 - HAS Y + R 221 VADDW.U8 Q9,Q6,D30 @//Q9 - HAS Y + G 223 VADDW.U8 Q10,Q4,D31 @//Q10 - HAS Y + B 224 VADDW.U8 Q11,Q5,D31 @//Q11 - HAS Y + R 225 VADDW.U8 Q12,Q6,D31 @//Q12 - HAS Y + G 260 VADDW.U8 Q7,Q4,D28 @//Q7 - HAS Y + B 261 VADDW.U8 Q8,Q5,D28 @//Q2 - HAS Y + R [all …]
|
/external/libhevc/common/arm/ |
D | ihevc_sao_edge_offset_class0_chroma.s | 174 VCGT.U8 Q8,Q6,Q7 @vcgtq_u8(pu1_cur_row, pu1_cur_row_tmp) 178 VCLT.U8 Q9,Q6,Q7 @vcltq_u8(pu1_cur_row, pu1_cur_row_tmp) 181 … VSUB.U8 Q10,Q9,Q8 @sign_left = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 188 VCGT.U8 Q13,Q15,Q14 @II vcgtq_u8(pu1_cur_row, pu1_cur_row_tmp) 191 VCLT.U8 Q12,Q15,Q14 @II vcltq_u8(pu1_cur_row, pu1_cur_row_tmp) 202 VCGT.U8 Q8,Q6,Q7 @vcgtq_u8(pu1_cur_row, pu1_cur_row_tmp) 205 VCLT.U8 Q9,Q6,Q7 @vcltq_u8(pu1_cur_row, pu1_cur_row_tmp) 208 …VSUB.U8 Q11,Q9,Q8 @sign_right = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 211 VADD.U8 Q7,Q1,Q10 @edge_idx = vaddq_s8(const_2, sign_left) 213 VADD.U8 Q7,Q7,Q11 @edge_idx = vaddq_s8(edge_idx, sign_right) [all …]
|
D | ihevc_sao_edge_offset_class1.s | 142 VCGT.U8 Q6,Q5,Q4 @vcgtq_u8(pu1_cur_row, pu1_top_row) 145 VCLT.U8 Q7,Q5,Q4 @vcltq_u8(pu1_cur_row, pu1_top_row) 147 VSUB.U8 Q8,Q7,Q6 @sign_up = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 157 VCGT.U8 Q6,Q5,Q9 @vcgtq_u8(pu1_cur_row, pu1_top_row) 162 VCLT.U8 Q7,Q5,Q9 @vcltq_u8(pu1_cur_row, pu1_top_row) 165 … VSUB.U8 Q10,Q7,Q6 @sign_down = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 166 …VMOVL.U8 Q13,D18 @II pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u… 169 …VMOVL.U8 Q14,D19 @II pi2_tmp_cur_row.val[1] = vreinterpretq_s16_u16(vmovl_u… 172 VCGT.U8 Q11,Q9,Q15 @II vcgtq_u8(pu1_cur_row, pu1_top_row) 176 VCLT.U8 Q12,Q9,Q15 @II vcltq_u8(pu1_cur_row, pu1_top_row) [all …]
|
D | ihevc_sao_edge_offset_class1_chroma.s | 146 VCGT.U8 Q6,Q5,Q14 @vcgtq_u8(pu1_cur_row, pu1_top_row) 149 VCLT.U8 Q7,Q5,Q14 @vcltq_u8(pu1_cur_row, pu1_top_row) 151 VSUB.U8 Q8,Q7,Q6 @sign_up = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 161 VCGT.U8 Q6,Q5,Q9 @vcgtq_u8(pu1_cur_row, pu1_top_row) 166 VCLT.U8 Q7,Q5,Q9 @vcltq_u8(pu1_cur_row, pu1_top_row) 169 … VSUB.U8 Q10,Q7,Q6 @sign_down = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 170 …VMOVL.U8 Q13,D18 @II pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u… 173 …VMOVL.U8 Q14,D19 @II pi2_tmp_cur_row.val[1] = vreinterpretq_s16_u16(vmovl_u… 176 VCGT.U8 Q11,Q9,Q15 @II vcgtq_u8(pu1_cur_row, pu1_top_row) 180 VCLT.U8 Q12,Q9,Q15 @II vcltq_u8(pu1_cur_row, pu1_top_row) [all …]
|
D | ihevc_sao_band_offset_chroma.s | 166 VCLE.U8 D13,D4,D30 @vcle_u8(band_table.val[3], vdup_n_u8(16)) 169 VORR.U8 D4,D4,D13 @band_table.val[3] = vorr_u8(band_table.val[3], au1_cmp) 175 VCLE.U8 D14,D3,D30 @vcle_u8(band_table.val[2], vdup_n_u8(16)) 177 VORR.U8 D3,D3,D14 @band_table.val[2] = vorr_u8(band_table.val[2], au1_cmp) 179 VAND.U8 D4,D4,D13 @band_table.val[3] = vand_u8(band_table.val[3], au1_cmp) 185 VCLE.U8 D15,D2,D30 @vcle_u8(band_table.val[1], vdup_n_u8(16)) 187 VORR.U8 D2,D2,D15 @band_table.val[1] = vorr_u8(band_table.val[1], au1_cmp) 189 VAND.U8 D3,D3,D14 @band_table.val[2] = vand_u8(band_table.val[2], au1_cmp) 195 VCLE.U8 D16,D1,D30 @vcle_u8(band_table.val[0], vdup_n_u8(16)) 196 VORR.U8 D1,D1,D16 @band_table.val[0] = vorr_u8(band_table.val[0], au1_cmp) [all …]
|
D | ihevc_sao_band_offset_luma.s | 150 VCLE.U8 D12,D4,D29 @vcle_u8(band_table.val[3], vdup_n_u8(16)) 153 VORR.U8 D4,D4,D12 @band_table.val[3] = vorr_u8(band_table.val[3], au1_cmp) 158 VCLE.U8 D11,D3,D29 @vcle_u8(band_table.val[2], vdup_n_u8(16)) 161 VORR.U8 D3,D3,D11 @band_table.val[2] = vorr_u8(band_table.val[2], au1_cmp) 163 VAND.U8 D4,D4,D12 @band_table.val[3] = vand_u8(band_table.val[3], au1_cmp) 168 VCLE.U8 D10,D2,D29 @vcle_u8(band_table.val[1], vdup_n_u8(16)) 171 VORR.U8 D2,D2,D10 @band_table.val[1] = vorr_u8(band_table.val[1], au1_cmp) 173 VAND.U8 D3,D3,D11 @band_table.val[2] = vand_u8(band_table.val[2], au1_cmp) 180 VCLE.U8 D9,D1,D29 @vcle_u8(band_table.val[0], vdup_n_u8(16)) 181 VORR.U8 D1,D1,D9 @band_table.val[0] = vorr_u8(band_table.val[0], au1_cmp) [all …]
|
D | ihevc_sao_edge_offset_class0.s | 167 VCGT.U8 Q8,Q6,Q7 @vcgtq_u8(pu1_cur_row, pu1_cur_row_tmp) 172 VCLT.U8 Q9,Q6,Q7 @vcltq_u8(pu1_cur_row, pu1_cur_row_tmp) 184 VCGT.U8 Q15,Q13,Q14 @II vcgtq_u8(pu1_cur_row, pu1_cur_row_tmp) 190 VCLT.U8 Q0,Q13,Q14 @II vcltq_u8(pu1_cur_row, pu1_cur_row_tmp) 195 VCGT.U8 Q8,Q6,Q7 @vcgtq_u8(pu1_cur_row, pu1_cur_row_tmp) 198 VCLT.U8 Q9,Q6,Q7 @vcltq_u8(pu1_cur_row, pu1_cur_row_tmp) 209 …VMOVL.U8 Q9,D12 @pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(v… 213 VCGT.U8 Q15,Q13,Q14 @II vcgtq_u8(pu1_cur_row, pu1_cur_row_tmp) 215 VCLT.U8 Q0,Q13,Q14 @II vcltq_u8(pu1_cur_row, pu1_cur_row_tmp) 221 …VMOVL.U8 Q0,D26 @II pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u… [all …]
|
D | ihevc_sao_edge_offset_class2.s | 257 VCGT.U8 Q7,Q6,Q5 @vcgtq_u8(pu1_cur_row, pu1_top_row) 261 VCLT.U8 Q8,Q6,Q5 @vcltq_u8(pu1_cur_row, pu1_top_row) 265 VSUB.U8 Q7,Q8,Q7 @sign_up = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 302 VCGT.U8 Q5,Q6,Q9 @I vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) 305 VCLT.U8 Q9,Q6,Q9 @I vcltq_u8(pu1_cur_row, pu1_next_row_tmp) 306 …VSUB.U8 Q5,Q9,Q5 @I sign_down = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt… 318 …VMOVL.U8 Q10,D12 @I pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8… 323 …VMOVL.U8 Q11,D13 @I pi2_tmp_cur_row.val[1] = vreinterpretq_s16_u16(vmovl_u8… 369 VCGT.U8 Q12,Q6,Q11 @II vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) 373 VCLT.U8 Q11,Q6,Q11 @II vcltq_u8(pu1_cur_row, pu1_next_row_tmp) [all …]
|
D | ihevc_sao_edge_offset_class3.s | 272 VCGT.U8 Q7,Q6,Q5 @vcgtq_u8(pu1_cur_row, pu1_top_row) 282 VCLT.U8 Q8,Q6,Q5 @vcltq_u8(pu1_cur_row, pu1_top_row) 285 VSUB.U8 Q7,Q8,Q7 @sign_up = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 317 VCGT.U8 Q5,Q6,Q9 @I vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) 318 VCLT.U8 Q9,Q6,Q9 @I vcltq_u8(pu1_cur_row, pu1_next_row_tmp) 319 …VSUB.U8 Q5,Q9,Q5 @I sign_down = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt… 329 …VMOVL.U8 Q10,D12 @I pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8… 334 …VMOVL.U8 Q11,D13 @I pi2_tmp_cur_row.val[1] = vreinterpretq_s16_u16(vmovl_u8… 394 VCGT.U8 Q12,Q6,Q9 @II vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) 398 VCLT.U8 Q13,Q6,Q9 @II vcltq_u8(pu1_cur_row, pu1_next_row_tmp) [all …]
|
D | ihevc_resi_trans.s | 276 VABDL.U8 q0,d2,d3 @ Absolute differences of rows 1 and 2 in d0 279 VSUBL.U8 q5,d2,d3 @ R1:[d10[3] d10[2] d10[1] d10[0]] => Row 1 of residue 284 VSUBL.U8 q6,d8,d9 @ R3:[d12[3] d12[2] d12[1] d12[0]] => Row 3 of residue 286 VABAL.U8 q0,d8,d9 @ Absolute differences of rows 3 and 4 in d1 447 VABDL.U8 q15,d1,d0 @ Row 1 of absolute difference in q15 449 VSUBL.U8 q0,d1,d0 @ Row 1 of residue in q0 452 VABDL.U8 q9,d3,d2 @ Row 2 of absolute difference in q9 454 VSUBL.U8 q1,d3,d2 @ Row 2 of residue in q1 457 VABAL.U8 q15,d5,d4 @ Row 3 of absolute difference accumulated in q15 459 VSUBL.U8 q2,d5,d4 @ Row 3 of residue in q2 [all …]
|
D | ihevc_sao_edge_offset_class3_chroma.s | 340 VCGT.U8 Q7,Q6,Q5 @vcgtq_u8(pu1_cur_row, pu1_top_row) 344 VCLT.U8 Q8,Q6,Q5 @vcltq_u8(pu1_cur_row, pu1_top_row) 348 VSUB.U8 Q7,Q8,Q7 @sign_up = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 404 VCGT.U8 Q10,Q6,Q9 @I vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) 406 VCLT.U8 Q11,Q6,Q9 @I vcltq_u8(pu1_cur_row, pu1_next_row_tmp) 407 …VSUB.U8 Q11,Q11,Q10 @I sign_down = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt… 417 …VMOVL.U8 Q10,D12 @I pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8… 425 …VMOVL.U8 Q9,D13 @I pi2_tmp_cur_row.val[1] = vreinterpretq_s16_u16(vmovl_u8… 487 VCGT.U8 Q11,Q6,Q14 @II vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) 497 VCLT.U8 Q12,Q6,Q14 @II vcltq_u8(pu1_cur_row, pu1_next_row_tmp) [all …]
|
D | ihevc_sao_edge_offset_class2_chroma.s | 350 VCGT.U8 Q7,Q6,Q5 @vcgtq_u8(pu1_cur_row, pu1_top_row) 354 VCLT.U8 Q8,Q6,Q5 @vcltq_u8(pu1_cur_row, pu1_top_row) 366 VSUB.U8 Q7,Q8,Q7 @sign_up = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 412 VCGT.U8 Q10,Q6,Q9 @I vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) 414 VCLT.U8 Q11,Q6,Q9 @I vcltq_u8(pu1_cur_row, pu1_next_row_tmp) 415 …VSUB.U8 Q11,Q11,Q10 @I sign_down = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt… 426 …VMOVL.U8 Q10,D12 @I pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8… 429 …VMOVL.U8 Q9,D13 @I pi2_tmp_cur_row.val[1] = vreinterpretq_s16_u16(vmovl_u8… 486 VCGT.U8 Q11,Q6,Q14 @II vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) 494 VCLT.U8 Q12,Q6,Q14 @II vcltq_u8(pu1_cur_row, pu1_next_row_tmp) [all …]
|
/external/llvm-project/compiler-rt/test/builtins/Unit/ |
D | atomic_test.c | 117 #define U8(value) ((uint8_t)(value)) macro 148 a8 = U8(value); in set_a_values() 158 b8 = U8(value); in set_b_values() 195 if (__atomic_load_1(&a8, model) != U8(V + m)) in test_loads() 236 __atomic_store_1(&a8, U8(V + m), model); in test_stores() 237 if (a8 != U8(V + m)) in test_stores() 296 if (__atomic_exchange_1(&a8, U8(V + m + 1), model) != U8(V + m)) in test_exchanges() 361 if (__atomic_compare_exchange_1(&a8, &b8, U8(V + m1), m_succ, m_fail)) in test_compare_exchanges() 363 if (a8 != U8(ONES) || b8 != U8(ONES)) in test_compare_exchanges() 365 if (!__atomic_compare_exchange_1(&a8, &b8, U8(V + m1), m_succ, m_fail)) in test_compare_exchanges() [all …]
|
/external/compiler-rt/lib/asan/tests/ |
D | asan_asm_test.cc | 68 DECLARE_ASM_WRITE(U8, "8", "movq", "r"); 69 DECLARE_ASM_READ(U8, "8", "movq", "=r"); 70 DECLARE_ASM_REP_MOVS(U8, "movsq"); 222 TestAsmWrite<U8>("WRITE of size 8"); in TEST() 230 TestAsmRead<U8>("READ of size 8"); in TEST() 265 TestAsmRepMovs<U8>("READ of size 8", "WRITE of size 8"); in TEST()
|
/external/tensorflow/tensorflow/compiler/xla/service/cpu/ |
D | xfeed_manager_test.cc | 36 : shape_(ShapeUtil::MakeShape(U8, {length})), in TestInfeedBuffer() 66 auto shape = ShapeUtil::MakeShape(U8, {length}); in ProcessNextBuffer() 136 ProcessNextOutfeedBuffer(32, ShapeUtil::MakeShape(U8, {32})); in TEST_F() 144 ProcessNextOutfeedBuffer(0, ShapeUtil::MakeShape(U8, {0})); in TEST_F() 152 ProcessNextOutfeedBuffer(32, ShapeUtil::MakeShape(U8, {33})); in TEST_F()
|
/external/tensorflow/tensorflow/compiler/xla/service/ |
D | operand_upcaster_test.cc | 85 std::make_tuple(U8, U8, U16))); 93 ::testing::Combine(::testing::Values(U8, U16), 94 ::testing::Values(U8, U16),
|
/external/rust/crates/bindgen/src/ir/ |
D | int.rs | 52 U8, enumerator 96 Bool | UChar | UShort | UInt | ULong | ULongLong | U8 | U16 | in is_signed() 114 Bool | UChar | SChar | U8 | I8 | Char { .. } => 1, in known_size()
|
/external/rust/cxx/syntax/ |
D | atom.rs | 9 U8, enumerator 35 "u8" => Some(U8), in from_str() 66 U8 => "u8", in as_ref()
|
/external/skia/src/opts/ |
D | SkBlitRow_opts.h | 227 using U8 = skvx::Vec<4*N, uint8_t>; in blit_row_color32() local 236 U8 s = skvx::bit_pun<U8>(src), in blit_row_color32() 237 a = U8(invA); in blit_row_color32() 238 U16 c = skvx::cast<uint16_t>(skvx::bit_pun<U8>(U32(color))), in blit_row_color32()
|
/external/rust/crates/unicode-bidi/src/ |
D | lib.rs | 867 Token::U8(0), in test_levels() 869 Token::U8(0), in test_levels() 871 Token::U8(0), in test_levels() 873 Token::U8(0), in test_levels() 875 Token::U8(1), in test_levels() 877 Token::U8(1), in test_levels() 879 Token::U8(1), in test_levels() 881 Token::U8(1), in test_levels() 883 Token::U8(1), in test_levels() 885 Token::U8(1), in test_levels()
|
/external/llvm-project/compiler-rt/lib/msan/tests/ |
D | msan_test.cpp | 136 typedef unsigned long long U8; typedef 199 static U8 poisoned_array[100]; 483 EXPECT_UMR(break_optimization((void*)(U8)a[GetPoisonedZero()])); in TEST() 1641 TestOverlapMemmove<U8, 4>(); in TEST() 1642 TestOverlapMemmove<U8, 1000>(); in TEST() 2915 U8 b; 2991 U8 a8, b8; 3940 U8 y8 = 0; in TEST() 4009 U8 y = 0; in TEST() 4041 U8 y = 0; in TEST() [all …]
|
/external/tensorflow/tensorflow/compiler/xla/ |
D | primitive_util.cc | 89 return type == U8 || type == U16 || type == U32 || type == U64; in IsUnsignedIntegralType() 102 case U8: in BitWidth() 139 return xla::U8; in UnsignedIntegralTypeForBitWidth()
|
/external/flatbuffers/tests/MyGame/Example/ |
D | TypeAliases.cs | 24 …public byte U8 { get { int o = __p.__offset(6); return o != 0 ? __p.bb.Get(o + __p.bb_pos) : (byte… property 120 _o.U8 = this.U8; in UnPackTo() 149 _o.U8, in Pack() 168 public byte U8 { get; set; } property in MyGame.Example.TypeAliasesT 192 this.U8 = 0; in TypeAliasesT()
|