/third_party/boost/boost/lockfree/ |
D | spsc_queue.hpp | 74 static size_t read_available(size_t write_index, size_t read_index, size_t max_size) in read_available() argument 76 if (write_index >= read_index) in read_available() 77 return write_index - read_index; in read_available() 79 const size_t ret = write_index + max_size - read_index; in read_available() 83 static size_t write_available(size_t write_index, size_t read_index, size_t max_size) in write_available() argument 85 size_t ret = read_index - write_index - 1; in write_available() 86 if (write_index >= read_index) in write_available() 94 const size_t read_index = read_index_.load(memory_order_relaxed); in read_available() local 95 return read_available(write_index, read_index, max_size); in read_available() 101 const size_t read_index = read_index_.load(memory_order_acquire); in write_available() local [all …]
|
/third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/gpu/cuda_impl/ |
D | scatter_nd_functor_impl.cu | 24 for (size_t read_index = blockIdx.x * blockDim.x + threadIdx.x; read_index < (updates_size); in ScatterNdUpdate() local 25 read_index += blockDim.x * gridDim.x) { in ScatterNdUpdate() 29 i = read_index / unit_size; in ScatterNdUpdate() 30 j = read_index % unit_size; in ScatterNdUpdate() 41 input[write_index] = updates[read_index]; in ScatterNdUpdate() 50 for (size_t read_index = blockIdx.x * blockDim.x + threadIdx.x; read_index < (updates_size); in ScatterNdAdd() local 51 read_index += blockDim.x * gridDim.x) { in ScatterNdAdd() 55 i = read_index / unit_size; in ScatterNdAdd() 56 j = read_index % unit_size; in ScatterNdAdd() 67 MsAtomicAdd(&input[write_index], updates[read_index]); in ScatterNdAdd() [all …]
|
D | cumsum_impl.cu | 39 size_t read_index = j * stride2 + offset; in LeftMoveSum() local 41 output[read_index] = 0; in LeftMoveSum() 44 output[read_index] = input[read_index2]; in LeftMoveSum() 61 size_t read_index = j * stride2 + offset; in RightMoveSum() local 63 output[read_index] = 0; in RightMoveSum() 66 output[read_index] = input[read_index2]; in RightMoveSum() 82 size_t read_index = j * stride2 + offset; in CumSumKernelReverse() local 84 output[read_index] = input[read_index]; in CumSumKernelReverse() 87 output[read_index] = output[read_index2] + input[read_index]; in CumSumKernelReverse() 104 size_t read_index = j * stride2 + offset; in CumSumKernel() local [all …]
|
D | cumprod_impl.cu | 39 size_t read_index = j * stride2 + offset; in LeftMoveProd() local 41 output[read_index] = 1; in LeftMoveProd() 44 output[read_index] = input[read_index2]; in LeftMoveProd() 61 size_t read_index = j * stride2 + offset; in RightMoveProd() local 63 output[read_index] = 1; in RightMoveProd() 66 output[read_index] = input[read_index2]; in RightMoveProd() 82 size_t read_index = j * stride2 + offset; in CumProdKernelReverse() local 84 output[read_index] = input[read_index]; in CumProdKernelReverse() 87 output[read_index] = output[read_index2] * input[read_index]; in CumProdKernelReverse() 104 size_t read_index = j * stride2 + offset; in CumProdKernel() local [all …]
|
D | tensor_scatter_add.cu | 26 for (size_t read_index = blockIdx.x * blockDim.x + threadIdx.x; read_index < input_size; in TensorScatterAddKernel() local 27 read_index += blockDim.x * gridDim.x) { in TensorScatterAddKernel() 31 i = read_index / block_size; in TensorScatterAddKernel() 32 j = read_index % block_size; in TensorScatterAddKernel() 44 MsAtomicAdd(&output[write_index], update[read_index]); in TensorScatterAddKernel()
|
D | tensor_scatter_update.cu | 26 for (size_t read_index = blockIdx.x * blockDim.x + threadIdx.x; read_index < input_size; in TensorScatterUpdateKernel() local 27 read_index += blockDim.x * gridDim.x) { in TensorScatterUpdateKernel() 31 i = read_index / block_size; in TensorScatterUpdateKernel() 32 j = read_index % block_size; in TensorScatterUpdateKernel() 44 output[write_index] = update[read_index]; in TensorScatterUpdateKernel()
|
D | tensor_scatter_min.cu | 26 for (size_t read_index = blockIdx.x * blockDim.x + threadIdx.x; read_index < input_size; in TensorScatterMinKernel() local 27 read_index += blockDim.x * gridDim.x) { in TensorScatterMinKernel() 31 i = read_index / block_size; in TensorScatterMinKernel() 32 j = read_index % block_size; in TensorScatterMinKernel() 44 MsAtomicMin(&output[write_index], update[read_index]); in TensorScatterMinKernel()
|
D | tensor_scatter_max.cu | 26 for (size_t read_index = blockIdx.x * blockDim.x + threadIdx.x; read_index < input_size; in TensorScatterMaxKernel() local 27 read_index += blockDim.x * gridDim.x) { in TensorScatterMaxKernel() 31 i = read_index / block_size; in TensorScatterMaxKernel() 32 j = read_index % block_size; in TensorScatterMaxKernel() 44 MsAtomicMax(&output[write_index], update[read_index]); in TensorScatterMaxKernel()
|
D | tensor_scatter_sub.cu | 26 for (size_t read_index = blockIdx.x * blockDim.x + threadIdx.x; read_index < input_size; in TensorScatterSubKernel() local 27 read_index += blockDim.x * gridDim.x) { in TensorScatterSubKernel() 31 i = read_index / block_size; in TensorScatterSubKernel() 32 j = read_index % block_size; in TensorScatterSubKernel() 44 MsAtomicSub(&output[write_index], update[read_index]); in TensorScatterSubKernel()
|
D | scatter_nd.cu | 26 for (size_t read_index = blockIdx.x * blockDim.x + threadIdx.x; read_index < input_size; in ScatterNdKernel() local 27 read_index += blockDim.x * gridDim.x) { in ScatterNdKernel() 31 i = read_index / block_size; in ScatterNdKernel() 32 j = read_index % block_size; in ScatterNdKernel() 44 MsAtomicAdd(&output[write_index], update[read_index]); in ScatterNdKernel()
|
D | gathernd.cu | 30 int read_index = 0; in GatherNdKernel() local 36 read_index += indices_i * batch_strides[k]; in GatherNdKernel() 38 read_index += j; in GatherNdKernel() 41 output[write_index] = input[read_index]; in GatherNdKernel()
|
D | gatherv2.cu | 32 size_t read_index = i * input_dim1 * output_dim2 + indices[j] * output_dim2 + k; in GatherV2Kernel() local 33 output[write_index] = input[read_index]; in GatherV2Kernel()
|
/third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/cpu/ |
D | cumsum_cpu_kernel.cc | 120 size_t read_index = j * stride2 + offset; in LeftMove() local 122 output[read_index] = (T)0; in LeftMove() 125 output[read_index] = input[read_index2]; in LeftMove() 139 size_t read_index = j * stride2 + offset; in RightMove() local 141 output[read_index] = (T)0; in RightMove() 144 output[read_index] = input[read_index2]; in RightMove() 158 size_t read_index = j * stride2 + offset; in Copy() local 159 input[read_index] = output[read_index]; in Copy() 172 size_t read_index = j * stride2 + offset; in CumSumKernelReverse() local 174 output[read_index] = input[read_index]; in CumSumKernelReverse() [all …]
|
D | gathernd_cpu_kernel.cc | 95 int read_index = 0; in Launch() local 99 read_index += indices_i * batch_indices_[k]; in Launch() 101 read_index += j; in Launch() 102 output_addr[write_index] = input_addr[read_index]; in Launch()
|
/third_party/pulseaudio/src/pulsecore/ |
D | memblockq.c | 52 int64_t read_index, write_index; member 82 bq->read_index = bq->write_index = idx; in pa_memblockq_new() 135 while (PA_UNLIKELY(bq->current_read->index > bq->read_index)) in fix_current_read() 143 … PA_UNLIKELY(bq->current_read->index + (int64_t) bq->current_read->chunk.length <= bq->read_index)) in fix_current_read() 217 boundary = bq->read_index - (int64_t) bq->maxrewind; in drop_backlog() 228 if (bq->read_index > bq->write_index) { in can_push() 229 int64_t d = bq->read_index - bq->write_index; in can_push() 241 if (bq->write_index + (int64_t) l - bq->read_index > (int64_t) bq->maxlength) in can_push() 269 delta = bq->read_index - old_read_index; in read_index_changed() 450 return bq->prebuf > 0 && bq->read_index >= bq->write_index; in pa_memblockq_prebuf_active() [all …]
|
/third_party/pulseaudio/src/modules/ |
D | module-esound-sink.c | 98 size_t read_length, read_index; member 425 u->read_index = 0; in handle_response() 455 u->read_index = u->read_length = 0; in handle_response() 479 pa_assert(u->read_index < u->read_length); in do_read() 481 … pa_iochannel_read(u->io, (uint8_t*) u->read_data + u->read_index, u->read_length - u->read_index)… in do_read() 486 u->read_index += (size_t) r; in do_read() 487 pa_assert(u->read_index <= u->read_length); in do_read() 489 if (u->read_index == u->read_length) in do_read() 591 u->read_index = u->write_index = u->read_length = u->write_length = 0; in pa__init()
|
D | module-tunnel.c | 778 int64_t write_index, read_index; in stream_get_latency_callback() local 800 pa_tagstruct_gets64(t, &read_index) < 0) { in stream_get_latency_callback() 849 if (write_index >= read_index) in stream_get_latency_callback() 850 delay += (int64_t) pa_bytes_to_usec((uint64_t) (write_index-read_index), ss); in stream_get_latency_callback() 852 delay -= (int64_t) pa_bytes_to_usec((uint64_t) (read_index-write_index), ss); in stream_get_latency_callback()
|
/third_party/ffmpeg/libavformat/ |
D | ifv.c | 52 static int read_index(AVFormatContext *s, in read_index() function 168 ret = read_index(s, AVMEDIA_TYPE_VIDEO, 0); in ifv_read_header() 176 ret = read_index(s, AVMEDIA_TYPE_AUDIO, 0); in ifv_read_header() 232 ret = read_index(s, AVMEDIA_TYPE_VIDEO, ifv->next_video_index); in ifv_read_packet() 236 ret = read_index(s, AVMEDIA_TYPE_AUDIO, ifv->next_audio_index); in ifv_read_packet() 257 ret = read_index(s, AVMEDIA_TYPE_VIDEO, ifv->next_video_index); in ifv_read_packet()
|
D | mvdec.c | 269 static void read_index(AVIOContext *pb, AVStream *st) in read_index() function 422 read_index(pb, ast); in mv_read_header() 425 read_index(pb, vst); in mv_read_header()
|
/third_party/pulseaudio/src/pulse/ |
D | simple.c | 45 size_t read_index, read_length; member 341 p->read_index = 0; in pa_simple_read() 345 memcpy(data, (const uint8_t*) p->read_data+p->read_index, l); in pa_simple_read() 350 p->read_index += l; in pa_simple_read() 359 p->read_index = 0; in pa_simple_read() 479 … already_read = pa_bytes_to_usec(p->read_index, pa_stream_get_sample_spec(p->stream)); in pa_simple_get_latency()
|
D | def.h | 730 int64_t read_index; member
|
D | stream.c | 1667 s->timing_info.read_index += (int64_t) s->peek_memchunk.length; in pa_stream_drop() 1748 …usec = pa_bytes_to_usec(s->timing_info.read_index < 0 ? 0 : (uint64_t) s->timing_info.read_index, … in calc_time() 1827 pa_tagstruct_gets64(t, &i->read_index) < 0) { in stream_get_timing_info_callback() 1929 i->read_index -= (int64_t) pa_memblockq_get_length(o->stream->record_memblockq); in stream_get_timing_info_callback() 2528 cindex = s->timing_info.read_index; in pa_stream_get_latency()
|
/third_party/mesa3d/src/gallium/drivers/nouveau/nvc0/ |
D | nvc0_state_validate.c | 141 unsigned read_index = pixel_y % grid_height * hw_grid_width; in gm200_validate_sample_locations() local 143 read_index += pixel_x % grid_width; in gm200_validate_sample_locations() 144 read_index = read_index * ms + sample; in gm200_validate_sample_locations() 145 x = sample_locations[read_index][0]; in gm200_validate_sample_locations() 146 y = sample_locations[read_index][1]; in gm200_validate_sample_locations()
|
/third_party/pulseaudio/src/modules/bluetooth/ |
D | module-bluez5-device.c | 129 uint64_t read_index; member 386 u->read_index += (uint64_t) l; in sco_process_push() 404 …pa_smoother_put(u->read_smoother, tstamp, pa_bytes_to_usec(u->read_index, &u->decoder_sample_spec)… in sco_process_push() 626 u->read_index += (uint64_t) memchunk.length; in a2dp_process_push() 627 …pa_smoother_put(u->read_smoother, tstamp, pa_bytes_to_usec(u->read_index, &u->decoder_sample_spec)… in a2dp_process_push() 841 u->read_index = u->write_index = 0; in setup_stream() 881 ri = pa_bytes_to_usec(u->read_index, &u->decoder_sample_spec); in source_process_msg() 1412 if (have_sink && u->write_index == 0 && u->read_index <= 0) in thread_func()
|
/third_party/mesa3d/src/compiler/nir/ |
D | nir_opt_find_array_copies.c | 421 unsigned write_index, unsigned read_index, in handle_write() argument 484 dst_node->first_src_read = MIN2(dst_node->first_src_read, read_index); in handle_write()
|