/external/bcc/tools/ |
D | runqlen.py | 210 queued = {} variable 217 queued[c] = 0 222 queued[k.cpu] += v.value 224 samples = idle[c] + queued[c] 226 runqocc = float(queued[c]) / samples 234 queued = 0 variable 239 queued += v.value 240 samples = idle + queued 242 runqocc = float(queued) / samples
|
D | biolatency_example.txt | 31 completion. A -Q option can be used to include time queued in the kernel. 105 queued in the kernel, and includes queuing latency: 215 -Q, --queued include OS queued time in I/O time 223 ./biolatency -Q # include OS queued time in I/O time
|
D | cpuunclaimed_example.txt | 5 idle CPUs, yet queued threads waiting their turn. It reports the amount of idle 56 queued threads waiting to run on CPU. This is an artificial situation caused 141 means one thread running and one thread queued, while CPU 6 has a run queue 271 some CPUs were idle and others had queued work. The tool will detect this, 278 had queued work, so it prints an error, and exits. Eg:
|
D | runqlen_example.txt | 34 Now there is often threads queued, with one sample reaching a queue length 134 to CPUs, leaving other CPUs idle while work is queued. 205 was work queued waiting its turn. Eg:
|
/external/mesa3d/src/gallium/drivers/radeonsi/ |
D | si_cp_dma.c | 532 cik_prefetch_shader_async(sctx, sctx->queued.named.hs); in cik_emit_prefetch_L2() 536 cik_prefetch_shader_async(sctx, sctx->queued.named.gs); in cik_emit_prefetch_L2() 538 cik_prefetch_shader_async(sctx, sctx->queued.named.vs); in cik_emit_prefetch_L2() 541 cik_prefetch_shader_async(sctx, sctx->queued.named.gs); in cik_emit_prefetch_L2() 545 cik_prefetch_shader_async(sctx, sctx->queued.named.vs); in cik_emit_prefetch_L2() 548 cik_prefetch_shader_async(sctx, sctx->queued.named.vs); in cik_emit_prefetch_L2() 557 cik_prefetch_shader_async(sctx, sctx->queued.named.ls); in cik_emit_prefetch_L2() 561 cik_prefetch_shader_async(sctx, sctx->queued.named.hs); in cik_emit_prefetch_L2() 563 cik_prefetch_shader_async(sctx, sctx->queued.named.es); in cik_emit_prefetch_L2() 565 cik_prefetch_shader_async(sctx, sctx->queued.named.gs); in cik_emit_prefetch_L2() [all …]
|
D | si_hw_context.c | 207 if (ctx->queued.named.ls) in si_begin_new_cs() 209 if (ctx->queued.named.hs) in si_begin_new_cs() 211 if (ctx->queued.named.es) in si_begin_new_cs() 213 if (ctx->queued.named.gs) in si_begin_new_cs() 215 if (ctx->queued.named.vs) in si_begin_new_cs() 217 if (ctx->queued.named.ps) in si_begin_new_cs()
|
D | si_state_viewport.c | 43 if (!ctx->queued.named.rasterizer || in si_set_scissor_states() 44 !ctx->queued.named.rasterizer->scissor_enable) in si_set_scissor_states() 184 ctx->queued.named.rasterizer) { in si_emit_guardband() 187 const struct si_state_rasterizer *rs = ctx->queued.named.rasterizer; in si_emit_guardband() 224 if (ctx->queued.named.rasterizer) in si_emit_scissors() 225 scissor_enabled = ctx->queued.named.rasterizer->scissor_enable; in si_emit_scissors() 352 if (ctx->queued.named.rasterizer) in si_emit_depth_ranges() 353 clip_halfz = ctx->queued.named.rasterizer->clip_halfz; in si_emit_depth_ranges()
|
D | si_state.h | 300 ((sctx)->queued.named.member != (sctx)->emitted.named.member) 303 ((sctx)->queued.named.member && si_pm4_state_changed(sctx, member)) 307 (sctx)->queued.named.member = (value); \ 313 if ((sctx)->queued.named.member == (value)) { \ 314 (sctx)->queued.named.member = NULL; \
|
D | si_state_binning.c | 194 struct si_state_dsa *dsa = sctx->queued.named.dsa; in si_get_depth_bin_size() 341 struct si_state_blend *blend = sctx->queued.named.blend; in si_emit_dpbb_state() 342 struct si_state_dsa *dsa = sctx->queued.named.dsa; in si_emit_dpbb_state()
|
D | si_state_shaders.c | 1127 if (sctx->queued.named.dsa) in si_get_alpha_test_func() 1128 return sctx->queued.named.dsa->alpha_func; in si_get_alpha_test_func() 1163 sctx->queued.named.rasterizer->clip_plane_enable == 0 && in si_shader_selector_key_hw_vs() 1171 const struct si_state_blend *blend = sctx->queued.named.blend; in si_shader_selector_key_hw_vs() 1181 sctx->queued.named.blend->cb_target_mask; in si_shader_selector_key_hw_vs() 1185 ps_disabled = sctx->queued.named.rasterizer->rasterizer_discard || in si_shader_selector_key_hw_vs() 1305 struct si_state_rasterizer *rs = sctx->queued.named.rasterizer; in si_shader_selector_key() 1306 struct si_state_blend *blend = sctx->queued.named.blend; in si_shader_selector_key() 1368 if (sctx->queued.named.blend) { in si_shader_selector_key() 1369 key->part.ps.epilog.alpha_to_one = sctx->queued.named.blend->alpha_to_one && in si_shader_selector_key() [all …]
|
/external/perfetto/src/traced/probes/ftrace/test/data/android_hammerhead_MRA59G_3.4.0/events/kgsl/adreno_cmdbatch_queued/ |
D | format | 12 field:unsigned int queued; offset:20; size:4; signed:0; 15 print fmt: "ctx=%u ts=%u queued=%u flags=%s", REC->id, REC->timestamp, REC->queued, REC->flags ? __…
|
/external/perfetto/src/traced/probes/ftrace/test/data/android_seed_N2F62_3.10.49/events/kgsl/adreno_cmdbatch_queued/ |
D | format | 11 field:unsigned int queued; offset:16; size:4; signed:0; 14 print fmt: "ctx=%u ts=%u queued=%u flags=%s", REC->id, REC->timestamp, REC->queued, REC->flags ? __…
|
/external/perfetto/src/traced/probes/ftrace/test/data/android_walleye_OPM5.171019.017.A1_4.4.88/events/kgsl/adreno_cmdbatch_queued/ |
D | format | 11 field:unsigned int queued; offset:16; size:4; signed:0; 15 print fmt: "ctx=%u ctx_prio=%u ts=%u queued=%u flags=%s", REC->id, REC->prio, REC->timestamp, REC->…
|
/external/adhd/cras/src/tests/ |
D | hfp_info_unittest.cc | 64 unsigned buffer_frames, buffer_frames2, queued; in TEST() local 86 queued = hfp_buf_queued(info, &dev); in TEST() 89 ASSERT_GE(info->playback_buf->used_size / 2, buffer_frames + queued); in TEST() 92 buf_increment_read(info->playback_buf, queued * 2); in TEST() 94 queued = hfp_buf_queued(info, &dev); in TEST() 95 ASSERT_EQ(0, queued); in TEST()
|
/external/google-breakpad/src/client/windows/crash_generation/ |
D | ReadMe.txt | 14 | If connection is successfully queued up asynchronously, 27 | If read request is successfully queued up asynchronously, 37 | If write request is successfully queued up asynchronously, 46 | If read request is successfully queued up asynchonously,
|
/external/python/cpython2/Doc/library/ |
D | termios.rst | 47 immediately, :const:`TCSADRAIN` to change after transmitting all queued output, 48 or :const:`TCSAFLUSH` to change after transmitting all queued output and 49 discarding all queued input. 65 Discard queued data on file descriptor *fd*. The *queue* selector specifies
|
/external/python/cpython3/Doc/library/ |
D | termios.rst | 47 immediately, :const:`TCSADRAIN` to change after transmitting all queued output, 48 or :const:`TCSAFLUSH` to change after transmitting all queued output and 49 discarding all queued input. 65 Discard queued data on file descriptor *fd*. The *queue* selector specifies
|
/external/v8/src/compiler/ |
D | loop-variable-optimizer.cc | 38 NodeMarker<bool> queued(graph(), 2); in Run() local 42 queued.Set(node, false); in Run() 68 } else if (!queued.Get(use)) { in Run() 70 queued.Set(use, true); in Run()
|
/external/autotest/scheduler/ |
D | prejob_task.py | 159 queued = models.SpecialTask.objects.filter( 162 queued = queued.exclude(id=self.task.id) 163 return queued.count() == 0
|
/external/perfetto/src/traced/probes/ftrace/test/data/android_walleye_OPM5.171019.017.A1_4.4.88/events/vb2/vb2_qbuf/ |
D | format | 16 print fmt: "owner = %p, queued = %u, owned_by_drv = %d, index = %u, type = %u, bytesused = %u", REC…
|
/external/perfetto/src/traced/probes/ftrace/test/data/android_walleye_OPM5.171019.017.A1_4.4.88/events/vb2/vb2_buf_done/ |
D | format | 16 print fmt: "owner = %p, queued = %u, owned_by_drv = %d, index = %u, type = %u, bytesused = %u", REC…
|
/external/perfetto/src/traced/probes/ftrace/test/data/android_walleye_OPM5.171019.017.A1_4.4.88/events/vb2/vb2_dqbuf/ |
D | format | 16 print fmt: "owner = %p, queued = %u, owned_by_drv = %d, index = %u, type = %u, bytesused = %u", REC…
|
/external/perfetto/src/traced/probes/ftrace/test/data/android_walleye_OPM5.171019.017.A1_4.4.88/events/vb2/vb2_buf_queue/ |
D | format | 16 print fmt: "owner = %p, queued = %u, owned_by_drv = %d, index = %u, type = %u, bytesused = %u", REC…
|
/external/tensorflow/tensorflow/core/api_def/base_api/ |
D | api_def_IdentityReaderV2.pbtxt | 26 summary: "A Reader that outputs the queued work as both the key and value."
|
D | api_def_IdentityReader.pbtxt | 24 summary: "A Reader that outputs the queued work as both the key and value."
|