/drivers/crypto/chelsio/ |
D | chcr_core.c | 56 if (atomic_read(&dev->inflight)) { in detach_work_fn() 60 atomic_read(&dev->inflight)); in detach_work_fn() 65 atomic_read(&dev->inflight)); in detach_work_fn() 103 atomic_set(&dev->inflight, 0); in chcr_dev_add() 122 atomic_set(&dev->inflight, 0); in chcr_dev_init() 232 if (atomic_read(&dev->inflight) != 0) { in chcr_detach_device()
|
D | chcr_core.h | 99 atomic_t inflight; member
|
D | chcr_algo.c | 200 atomic_inc(&dev->inflight); in chcr_inc_wrcount() 206 atomic_dec(&dev->inflight); in chcr_dec_wrcount()
|
/drivers/gpu/drm/i915/gt/ |
D | intel_context_types.h | 78 struct intel_engine_cs *inflight; member 82 __intel_context_inflight(READ_ONCE((ce)->inflight)) 84 __intel_context_inflight_count(READ_ONCE((ce)->inflight))
|
D | intel_execlists_submission.c | 533 old = ce->inflight; in execlists_schedule_in() 536 WRITE_ONCE(ce->inflight, ptr_inc(old)); in execlists_schedule_in() 597 GEM_BUG_ON(ce->inflight != engine); in __execlists_schedule_out() 644 WRITE_ONCE(ce->inflight, NULL); in __execlists_schedule_out() 654 GEM_BUG_ON(!ce->inflight); in execlists_schedule_out() 655 ce->inflight = ptr_dec(ce->inflight); in execlists_schedule_out() 656 if (!__intel_context_inflight_count(ce->inflight)) in execlists_schedule_out() 983 const struct intel_engine_cs *inflight; in virtual_matches() local 1000 inflight = intel_context_inflight(&ve->context); in virtual_matches() 1001 if (inflight && inflight != engine) in virtual_matches() [all …]
|
D | intel_timeline.c | 420 unsigned long count, ready, inflight; in intel_gt_show_timelines() local 437 inflight = 0; in intel_gt_show_timelines() 446 inflight++; in intel_gt_show_timelines() 451 count, ready, inflight); in intel_gt_show_timelines()
|
D | intel_engine_types.h | 221 struct i915_request *inflight[EXECLIST_MAX_PORTS + 1 /* sentinel */]; member
|
D | intel_engine_cs.c | 641 memset(execlists->inflight, 0, sizeof(execlists->inflight)); in intel_engine_init_execlists()
|
/drivers/vhost/ |
D | scsi.c | 99 struct vhost_scsi_inflight *inflight; member 210 struct vhost_scsi_inflight *inflight; member 235 struct vhost_scsi_inflight *inflight; in vhost_scsi_done_inflight() local 237 inflight = container_of(kref, struct vhost_scsi_inflight, kref); in vhost_scsi_done_inflight() 238 complete(&inflight->comp); in vhost_scsi_done_inflight() 271 struct vhost_scsi_inflight *inflight; in vhost_scsi_get_inflight() local 275 inflight = &svq->inflights[svq->inflight_idx]; in vhost_scsi_get_inflight() 276 kref_get(&inflight->kref); in vhost_scsi_get_inflight() 278 return inflight; in vhost_scsi_get_inflight() 281 static void vhost_scsi_put_inflight(struct vhost_scsi_inflight *inflight) in vhost_scsi_put_inflight() argument [all …]
|
/drivers/infiniband/ulp/rtrs/ |
D | rtrs-clt-stats.c | 100 atomic_read(&stats->inflight), sum.failover_cnt); in rtrs_clt_stats_rdma_to_str() 157 atomic_set(&s->inflight, 0); in rtrs_clt_reset_all_stats() 184 atomic_inc(&stats->inflight); in rtrs_clt_update_all_stats()
|
D | README | 100 corresponding path is disconnected, all the inflight IO are failed over to a 131 inflight IO and for the error code. 149 inflight IO and for the error code. The new rkey is sent back using 171 outstanding inflight IO and the error code. 192 outstanding inflight IO and the error code. The new rkey is sent back using
|
D | rtrs-clt.h | 68 atomic_t inflight; member
|
D | rtrs-clt.c | 442 atomic_dec(&clt_path->stats->inflight); in complete_rdma_req() 821 int inflight; in get_next_path_min_inflight() local 830 inflight = atomic_read(&clt_path->stats->inflight); in get_next_path_min_inflight() 832 if (inflight < min_inflight) { in get_next_path_min_inflight() 833 min_inflight = inflight; in get_next_path_min_inflight() 1162 atomic_dec(&clt_path->stats->inflight); in rtrs_clt_write_req() 1268 atomic_dec(&clt_path->stats->inflight); in rtrs_clt_read_req()
|
/drivers/crypto/cavium/cpt/ |
D | cpt_hw_types.h | 443 u64 inflight:8; member 445 u64 inflight:8;
|
D | cptvf_main.c | 390 vqx_inprg.s.inflight = val; in cptvf_write_vq_inprog()
|
/drivers/crypto/marvell/octeontx/ |
D | otx_cpt_hw_types.h | 581 u64 inflight:8; member 583 u64 inflight:8;
|
D | otx_cptvf_main.c | 371 vqx_inprg.s.inflight = val; in cptvf_write_vq_inprog()
|
/drivers/iommu/amd/ |
D | iommu_v2.c | 30 atomic_t inflight; member 417 if (atomic_dec_and_test(&pasid_state->pri[tag].inflight) && in finish_pri_tag() 558 atomic_inc(&pasid_state->pri[tag].inflight); in ppr_notifier()
|
/drivers/crypto/marvell/octeontx2/ |
D | otx2_cpt_hw_types.h | 403 u64 inflight:9; member
|
D | otx2_cptlf.h | 209 if (!lf_inprog.s.inflight) in otx2_cptlf_do_disable_iqueue()
|
/drivers/block/xen-blkback/ |
D | blkback.c | 697 if (atomic_dec_and_test(&ring->inflight) && atomic_read(&blkif->drain)) { in xen_blkbk_unmap_and_respond_callback() 1025 if (atomic_read(&ring->inflight) == 0) in xen_blk_drain_io() 1321 atomic_inc(&ring->inflight); in dispatch_rw_block_io()
|
D | common.h | 261 atomic_t inflight; member
|
D | xenbus.c | 289 if (atomic_read(&ring->inflight) > 0) { in xen_blkif_disconnect()
|
/drivers/gpu/drm/i915/ |
D | i915_request.c | 392 bool inflight = false; in __request_in_flight() local 441 inflight = i915_seqno_passed(rq->fence.seqno, in __request_in_flight() 448 return inflight; in __request_in_flight()
|
/drivers/scsi/ |
D | scsi_transport_fc.c | 4051 int err = 0, inflight = 0; in fc_bsg_job_timeout() local 4056 inflight = bsg_job_get(job); in fc_bsg_job_timeout() 4058 if (inflight && i->f->bsg_timeout) { in fc_bsg_job_timeout() 4070 if (inflight) in fc_bsg_job_timeout()
|