/external/llvm-project/compiler-rt/test/tsan/ |
D | unaligned_race.cpp | 39 template<int off, int off2> 45 p += off2; in access3() 57 access2(bool main, int sz1, int off2, bool rw, volatile char *obj) { in access2() argument 58 if (off2 == 0) in access2() 60 else if (off2 == 1) in access2() 62 else if (off2 == 2) in access2() 64 else if (off2 == 3) in access2() 66 else if (off2 == 4) in access2() 68 else if (off2 == 5) in access2() 70 else if (off2 == 6) in access2() [all …]
|
D | bench_shadow_flush.cpp | 39 for (uptr off2 = 0; off2 < stride2; off2++) in main() local 40 mem[off + off2 * kStride] = 42; in main()
|
/external/compiler-rt/test/tsan/ |
D | unaligned_race.cc | 48 template<int off, int off2> 54 p += off2; in access3() 66 access2(bool main, int sz1, int off2, bool rw, volatile char *obj) { in access2() argument 67 if (off2 == 0) in access2() 69 else if (off2 == 1) in access2() 71 else if (off2 == 2) in access2() 73 else if (off2 == 3) in access2() 75 else if (off2 == 4) in access2() 77 else if (off2 == 5) in access2() 79 else if (off2 == 6) in access2() [all …]
|
D | bench_shadow_flush.cc | 39 for (uptr off2 = 0; off2 < stride2; off2++) in main() local 40 mem[off + off2 * kStride] = 42; in main()
|
/external/libvpx/libvpx/vp9/encoder/ppc/ |
D | vp9_quantize_vsx.c | 100 int off2 = 64; in vp9_quantize_fp_vsx() local 105 coeff2 = vec_vsx_ld(off2, coeff_ptr); in vp9_quantize_fp_vsx() 108 scan2 = vec_vsx_ld(off2, iscan); in vp9_quantize_fp_vsx() 127 vec_vsx_st(qcoeff2, off2, qcoeff_ptr); in vp9_quantize_fp_vsx() 129 vec_vsx_st(dqcoeff2, off2, dqcoeff_ptr); in vp9_quantize_fp_vsx() 138 off2 += 48; in vp9_quantize_fp_vsx() 180 int off2 = 64; in vp9_quantize_fp_32x32_vsx() local 239 coeff2 = vec_vsx_ld(off2, coeff_ptr); in vp9_quantize_fp_32x32_vsx() 242 scan2 = vec_vsx_ld(off2, iscan); in vp9_quantize_fp_32x32_vsx() 270 vec_vsx_st(qcoeff2, off2, qcoeff_ptr); in vp9_quantize_fp_32x32_vsx() [all …]
|
/external/libvpx/libvpx/vpx_dsp/ppc/ |
D | quantize_vsx.c | 151 int off2 = 64; in vpx_quantize_b_vsx() local 157 coeff2 = vec_vsx_ld(off2, coeff_ptr); in vpx_quantize_b_vsx() 172 vec_vsx_st(qcoeff2, off2, qcoeff_ptr); in vpx_quantize_b_vsx() 180 vec_vsx_st(dqcoeff2, off2, dqcoeff_ptr); in vpx_quantize_b_vsx() 185 nonzero_scanindex(qcoeff2, zero_mask2, iscan_ptr, off2)); in vpx_quantize_b_vsx() 191 off2 += 48; in vpx_quantize_b_vsx() 212 int off2 = 64; in vpx_quantize_b_32x32_vsx() local 266 coeff2 = vec_vsx_ld(off2, coeff_ptr); in vpx_quantize_b_32x32_vsx() 285 vec_vsx_st(qcoeff2, off2, qcoeff_ptr); in vpx_quantize_b_32x32_vsx() 289 vec_vsx_st(dequantize_coeff_32(qcoeff2, dequant), off2, dqcoeff_ptr); in vpx_quantize_b_32x32_vsx() [all …]
|
/external/arm-optimized-routines/string/aarch64/ |
D | strcmp-mte.S | 36 #define off2 x10 macro 56 sub off2, src2, src1 59 tst off2, 7 66 ldr data2, [src1, off2] 111 ldr data2, [src1, off2] 153 ldr data2, [src1, off2]
|
/external/libavc/common/ |
D | ih264_inter_pred_filters.c | 538 WORD32 off1, off2, off3; in ih264_inter_pred_luma_vert_qpel() local 544 off2 = src_strd << 1; in ih264_inter_pred_luma_vert_qpel() 545 off3 = off1 + off2; in ih264_inter_pred_luma_vert_qpel() 558 i2_temp = pu1_src[-off2] + pu1_src[off3] in ih264_inter_pred_luma_vert_qpel() 559 - (pu1_src[-off1] + pu1_src[off2]) in ih264_inter_pred_luma_vert_qpel() 560 + ((pu1_src[0] + pu1_src[off1] - pu1_src[-off1] - pu1_src[off2]) << 2) in ih264_inter_pred_luma_vert_qpel() 623 WORD32 off1, off2, off3; in ih264_inter_pred_luma_horz_qpel_vert_qpel() local 629 off2 = src_strd << 1; in ih264_inter_pred_luma_horz_qpel_vert_qpel() 630 off3 = off1 + off2; in ih264_inter_pred_luma_horz_qpel_vert_qpel() 645 i2_temp_vert = pu1_pred_vert[-off2] + pu1_pred_vert[off3] in ih264_inter_pred_luma_horz_qpel_vert_qpel() [all …]
|
/external/tpm2-tss/src/util/ |
D | log.cpp | 106 unsigned int i, y, x, off, off2; in doLogBlob() local 111 for (i = 1, off = 0, off2 = 0; i <= size; i++) { in doLogBlob() 137 if (isgraph(blob[off2 + y])) { in doLogBlob() 138 sprintf(&buffer[y + off], "%c", blob[off2 + y]); in doLogBlob() 145 off2 = i; in doLogBlob()
|
/external/llvm-project/llvm/test/Transforms/IRCE/ |
D | empty_ranges.ll | 26 ; CHECK-NEXT: %off2 = add i32 %idx, -60 27 ; CHECK-NEXT: %c3 = icmp slt i32 %off2, 40 56 %off2 = add i32 %idx, -60 57 %c3 = icmp slt i32 %off2, 40
|
/external/ltp/testcases/kernel/syscalls/copy_file_range/ |
D | copy_file_range01.c | 37 loff_t *off1, loff_t *off2, size_t len) in check_file_content() argument 48 if (off2 && fseek(fp2, *off2, SEEK_SET)) in check_file_content()
|
/external/llvm-project/llvm/test/CodeGen/ARM/ |
D | vldm-liveness.ll | 18 %off2 = getelementptr float, float* %ptr, i32 2 19 %val2 = load float, float* %off2
|
/external/freetype/src/cid/ |
D | cidgload.c | 108 FT_ULong off1, off2; in cid_load_glyph() local 120 off2 = cid_get_offset( &p, (FT_Byte)cid->gd_bytes ); in cid_load_glyph() 124 off2 > stream->size || in cid_load_glyph() 125 off1 > off2 ) in cid_load_glyph() 132 glyph_length = off2 - off1; in cid_load_glyph()
|
/external/flatbuffers/swift/Sources/FlatBuffers/ |
D | Table.swift | 109 static public func compare(_ off1: Int32, _ off2: Int32, fbb: ByteBuffer) -> Int32 { in compare() 112 let _off2 = off2 + fbb.read(def: Int32.self, position: Int(off2)) in compare()
|
/external/llvm/test/CodeGen/ARM/ |
D | vldm-liveness.ll | 31 %off2 = getelementptr float, float* %ptr, i32 2 32 %val2 = load float, float* %off2
|
/external/llvm-project/llvm/test/CodeGen/SystemZ/ |
D | unaligned-01.ll | 13 %off2 = getelementptr i8, i8 *%ptr, i64 2 17 store i8 2, i8 *%off2
|
/external/llvm/test/CodeGen/SystemZ/ |
D | unaligned-01.ll | 16 %off2 = getelementptr i8, i8 *%ptr, i64 2 20 store i8 2, i8 *%off2
|
/external/flatbuffers/lua/flatbuffers/ |
D | builder.lua | 230 local off2 = self:Offset() - off + 4 231 self:Place(off2, SOffsetT) 238 local off2 = soffset - off + 4 239 self:Place(off2, UOffsetT)
|
/external/llvm-project/llvm/test/CodeGen/RISCV/ |
D | fold-addi-loadstore.ll | 8 ; (load (addi base, off1), off2) -> (load base, off1+off2) 9 ; (store val, (addi base, off1), off2) -> (store val, base, off1+off2) 10 ; This is possible when the off1+off2 continues to fit the 12-bit immediate.
|
/external/freetype/src/cff/ |
D | cffload.c | 516 FT_ULong off1, off2 = 0; in cff_index_access_element() local 537 off2 = cff_index_read_offset( idx, &error ); in cff_index_access_element() 539 } while ( off2 == 0 && element < idx->count ); in cff_index_access_element() 550 off2 = idx->offsets[element]; in cff_index_access_element() 552 } while ( off2 == 0 && element < idx->count ); in cff_index_access_element() 558 if ( off2 > stream->size + 1 || in cff_index_access_element() 559 idx->data_offset > stream->size - off2 + 1 ) in cff_index_access_element() 564 off2, stream->size - idx->data_offset + 1 )); in cff_index_access_element() 565 off2 = stream->size - idx->data_offset + 1; in cff_index_access_element() 569 if ( off1 && off2 > off1 ) in cff_index_access_element() [all …]
|
/external/elfutils/src/ |
D | elfcmp.c | 430 size_t off2 = 0; in main() local 440 if (off2 >= data2->d_size) in main() 448 off2 = gelf_getnote (data2, off2, ¬e2, in main() 450 if (off2 == 0) in main() 510 if (off2 < data2->d_size) in main()
|
/external/libabigail/src/ |
D | abg-comparison-priv.h | 692 size_t off2 = get_data_member_offset(second_dm); in operator() local 694 if (off1 != off2) in operator() 695 return off1 < off2; in operator() 717 off2 = get_data_member_offset(second_dm); in operator() 719 if (off1 != off2) in operator() 720 return off1 < off2; in operator()
|
/external/flatbuffers/python/flatbuffers/ |
D | builder.py | 345 off2 = self.Offset() - off + N.SOffsetTFlags.bytewidth 346 self.PlaceSOffsetT(off2) 359 off2 = self.Offset() - off + N.UOffsetTFlags.bytewidth 360 self.PlaceUOffsetT(off2)
|
/external/llvm-project/clang/test/CodeGenCXX/ |
D | no-unique-address.cpp | 38 Empty1 HasEmptyDuplicates::*off2 = &HasEmptyDuplicates::e2; variable
|
/external/flatbuffers/go/ |
D | builder.go | 273 off2 := SOffsetT(b.Offset()) - off + SOffsetT(SizeSOffsetT) 274 b.PlaceSOffsetT(off2) 283 off2 := b.Offset() - off + UOffsetT(SizeUOffsetT) 284 b.PlaceUOffsetT(off2)
|