/drivers/gpu/host1x/ |
D | job.c | 38 struct host1x_job *job = NULL; in host1x_job_alloc() local 54 mem = job = kzalloc(total, GFP_KERNEL); in host1x_job_alloc() 55 if (!job) in host1x_job_alloc() 58 kref_init(&job->ref); in host1x_job_alloc() 59 job->channel = ch; in host1x_job_alloc() 63 job->relocarray = num_relocs ? mem : NULL; in host1x_job_alloc() 65 job->unpins = num_unpins ? mem : NULL; in host1x_job_alloc() 67 job->waitchk = num_waitchks ? mem : NULL; in host1x_job_alloc() 69 job->gathers = num_cmdbufs ? mem : NULL; in host1x_job_alloc() 71 job->addr_phys = num_unpins ? mem : NULL; in host1x_job_alloc() [all …]
|
D | cdma.c | 176 struct host1x_job *job) in cdma_start_timer_locked() argument 185 cdma->timeout.client = job->client; in cdma_start_timer_locked() 186 cdma->timeout.syncpt = host1x_syncpt_get(host, job->syncpt_id); in cdma_start_timer_locked() 187 cdma->timeout.syncpt_val = job->syncpt_end; in cdma_start_timer_locked() 191 msecs_to_jiffies(job->timeout)); in cdma_start_timer_locked() 218 struct host1x_job *job, *n; in update_cdma_locked() local 228 list_for_each_entry_safe(job, n, &cdma->sync_queue, list) { in update_cdma_locked() 230 host1x_syncpt_get(host1x, job->syncpt_id); in update_cdma_locked() 233 if (!host1x_syncpt_is_expired(sp, job->syncpt_end)) { in update_cdma_locked() 235 if (job->timeout) in update_cdma_locked() [all …]
|
D | channel.c | 40 int host1x_job_submit(struct host1x_job *job) in host1x_job_submit() argument 42 struct host1x *host = dev_get_drvdata(job->channel->dev->parent); in host1x_job_submit() 44 return host1x_hw_channel_submit(host, job); in host1x_job_submit()
|
D | cdma.h | 90 int host1x_cdma_begin(struct host1x_cdma *cdma, struct host1x_job *job); 92 void host1x_cdma_end(struct host1x_cdma *cdma, struct host1x_job *job);
|
D | dev.h | 41 int (*submit)(struct host1x_job *job); 220 struct host1x_job *job) in host1x_hw_channel_submit() argument 222 return host->channel_op->submit(job); in host1x_hw_channel_submit()
|
D | Makefile | 8 job.o \
|
D | job.h | 52 void host1x_job_dump(struct device *dev, struct host1x_job *job);
|
/drivers/md/ |
D | dm-kcopyd.c | 392 struct kcopyd_job *job = NULL; in pop() local 398 job = list_entry(jobs->next, struct kcopyd_job, list); in pop() 399 list_del(&job->list); in pop() 403 return job; in pop() 406 static void push(struct list_head *jobs, struct kcopyd_job *job) in push() argument 409 struct dm_kcopyd_client *kc = job->kc; in push() 412 list_add_tail(&job->list, jobs); in push() 417 static void push_head(struct list_head *jobs, struct kcopyd_job *job) in push_head() argument 420 struct dm_kcopyd_client *kc = job->kc; in push_head() 423 list_add(&job->list, jobs); in push_head() [all …]
|
/drivers/gpu/host1x/hw/ |
D | channel_hw.c | 56 static void submit_gathers(struct host1x_job *job) in submit_gathers() argument 58 struct host1x_cdma *cdma = &job->channel->cdma; in submit_gathers() 61 for (i = 0; i < job->num_gathers; i++) { in submit_gathers() 62 struct host1x_job_gather *g = &job->gathers[i]; in submit_gathers() 70 static inline void synchronize_syncpt_base(struct host1x_job *job) in synchronize_syncpt_base() argument 72 struct host1x *host = dev_get_drvdata(job->channel->dev->parent); in synchronize_syncpt_base() 73 struct host1x_syncpt *sp = host->syncpt + job->syncpt_id; in synchronize_syncpt_base() 79 host1x_cdma_push(&job->channel->cdma, in synchronize_syncpt_base() 86 static int channel_submit(struct host1x_job *job) in channel_submit() argument 88 struct host1x_channel *ch = job->channel; in channel_submit() [all …]
|
D | debug_hw.c | 143 struct host1x_job *job; in show_channel_gathers() local 145 list_for_each_entry(job, &cdma->sync_queue, list) { in show_channel_gathers() 148 job, job->syncpt_id, job->syncpt_end, in show_channel_gathers() 149 job->first_get, job->timeout, in show_channel_gathers() 150 job->num_slots, job->num_unpins); in show_channel_gathers() 152 for (i = 0; i < job->num_gathers; i++) { in show_channel_gathers() 153 struct host1x_job_gather *g = &job->gathers[i]; in show_channel_gathers() 156 if (job->gather_copy_mapped) in show_channel_gathers() 157 mapped = (u32 *)job->gather_copy_mapped; in show_channel_gathers() 172 if (!job->gather_copy_mapped) in show_channel_gathers()
|
/drivers/scsi/lpfc/ |
D | lpfc_bsg.c | 300 struct fc_bsg_job *job; in lpfc_bsg_send_mgmt_cmd_cmp() local 313 job = dd_data->set_job; in lpfc_bsg_send_mgmt_cmd_cmp() 314 if (job) { in lpfc_bsg_send_mgmt_cmd_cmp() 316 job->dd_data = NULL; in lpfc_bsg_send_mgmt_cmd_cmp() 334 if (job) { in lpfc_bsg_send_mgmt_cmd_cmp() 353 job->reply->reply_payload_rcv_len = in lpfc_bsg_send_mgmt_cmd_cmp() 354 lpfc_bsg_copy_data(rmp, &job->reply_payload, in lpfc_bsg_send_mgmt_cmd_cmp() 369 if (job) { in lpfc_bsg_send_mgmt_cmd_cmp() 370 job->reply->result = rc; in lpfc_bsg_send_mgmt_cmd_cmp() 371 job->job_done(job); in lpfc_bsg_send_mgmt_cmd_cmp() [all …]
|
/drivers/s390/scsi/ |
D | zfcp_fc.c | 842 struct fc_bsg_job *job = data; in zfcp_fc_ct_els_job_handler() local 843 struct zfcp_fsf_ct_els *zfcp_ct_els = job->dd_data; in zfcp_fc_ct_els_job_handler() 844 struct fc_bsg_reply *jr = job->reply; in zfcp_fc_ct_els_job_handler() 846 jr->reply_payload_rcv_len = job->reply_payload.payload_len; in zfcp_fc_ct_els_job_handler() 849 job->job_done(job); in zfcp_fc_ct_els_job_handler() 852 static struct zfcp_fc_wka_port *zfcp_fc_job_wka_port(struct fc_bsg_job *job) in zfcp_fc_job_wka_port() argument 858 preamble_word1 = job->request->rqst_data.r_ct.preamble_word1; in zfcp_fc_job_wka_port() 861 adapter = (struct zfcp_adapter *) job->shost->hostdata[0]; in zfcp_fc_job_wka_port() 881 struct fc_bsg_job *job = data; in zfcp_fc_ct_job_handler() local 884 wka_port = zfcp_fc_job_wka_port(job); in zfcp_fc_ct_job_handler() [all …]
|
/drivers/scsi/libfc/ |
D | fc_lport.c | 152 struct fc_bsg_job *job; member 1906 struct fc_bsg_job *job = info->job; in fc_lport_bsg_resp() local 1913 job->reply->result = (PTR_ERR(fp) == -FC_EX_CLOSED) ? in fc_lport_bsg_resp() 1915 job->reply_len = sizeof(uint32_t); in fc_lport_bsg_resp() 1916 job->state_flags |= FC_RQST_STATE_DONE; in fc_lport_bsg_resp() 1917 job->job_done(job); in fc_lport_bsg_resp() 1934 job->reply->reply_data.ctels_reply.status = in fc_lport_bsg_resp() 1939 job->reply->reply_payload_rcv_len += in fc_lport_bsg_resp() 1946 if (job->reply->reply_payload_rcv_len > in fc_lport_bsg_resp() 1947 job->reply_payload.payload_len) in fc_lport_bsg_resp() [all …]
|
/drivers/scsi/ |
D | scsi_transport_fc.c | 3561 fc_destroy_bsgjob(struct fc_bsg_job *job) in fc_destroy_bsgjob() argument 3565 spin_lock_irqsave(&job->job_lock, flags); in fc_destroy_bsgjob() 3566 if (job->ref_cnt) { in fc_destroy_bsgjob() 3567 spin_unlock_irqrestore(&job->job_lock, flags); in fc_destroy_bsgjob() 3570 spin_unlock_irqrestore(&job->job_lock, flags); in fc_destroy_bsgjob() 3572 put_device(job->dev); /* release reference for the request */ in fc_destroy_bsgjob() 3574 kfree(job->request_payload.sg_list); in fc_destroy_bsgjob() 3575 kfree(job->reply_payload.sg_list); in fc_destroy_bsgjob() 3576 kfree(job); in fc_destroy_bsgjob() 3585 fc_bsg_jobdone(struct fc_bsg_job *job) in fc_bsg_jobdone() argument [all …]
|
D | scsi_transport_iscsi.c | 1476 static int iscsi_bsg_host_dispatch(struct bsg_job *job) in iscsi_bsg_host_dispatch() argument 1478 struct Scsi_Host *shost = iscsi_job_to_shost(job); in iscsi_bsg_host_dispatch() 1479 struct iscsi_bsg_request *req = job->request; in iscsi_bsg_host_dispatch() 1480 struct iscsi_bsg_reply *reply = job->reply; in iscsi_bsg_host_dispatch() 1486 if (job->request_len < sizeof(uint32_t)) { in iscsi_bsg_host_dispatch() 1508 if (job->request_len < cmdlen) { in iscsi_bsg_host_dispatch() 1513 ret = i->iscsi_transport->bsg_request(job); in iscsi_bsg_host_dispatch() 1519 BUG_ON(job->reply_len < sizeof(uint32_t)); in iscsi_bsg_host_dispatch() 1522 job->reply_len = sizeof(uint32_t); in iscsi_bsg_host_dispatch() 1523 bsg_job_done(job, ret, 0); in iscsi_bsg_host_dispatch()
|
/drivers/gpu/drm/tegra/ |
D | drm.c | 188 struct host1x_job *job; in tegra_drm_submit() local 195 job = host1x_job_alloc(context->channel, args->num_cmdbufs, in tegra_drm_submit() 197 if (!job) in tegra_drm_submit() 200 job->num_relocs = args->num_relocs; in tegra_drm_submit() 201 job->num_waitchk = args->num_waitchks; in tegra_drm_submit() 202 job->client = (u32)args->context; in tegra_drm_submit() 203 job->class = context->client->base.class; in tegra_drm_submit() 204 job->serialize = true; in tegra_drm_submit() 221 host1x_job_add_gather(job, bo, cmdbuf.words, cmdbuf.offset); in tegra_drm_submit() 228 err = host1x_reloc_copy_from_user(&job->relocarray[num_relocs], in tegra_drm_submit() [all …]
|
/drivers/scsi/bfa/ |
D | bfad_bsg.c | 3132 bfad_im_bsg_vendor_request(struct fc_bsg_job *job) in bfad_im_bsg_vendor_request() argument 3134 uint32_t vendor_cmd = job->request->rqst_data.h_vendor.vendor_cmd[0]; in bfad_im_bsg_vendor_request() 3136 (struct bfad_im_port_s *) job->shost->hostdata[0]; in bfad_im_bsg_vendor_request() 3138 struct request_queue *request_q = job->req->q; in bfad_im_bsg_vendor_request() 3149 payload_kbuf = kzalloc(job->request_payload.payload_len, GFP_KERNEL); in bfad_im_bsg_vendor_request() 3156 sg_copy_to_buffer(job->request_payload.sg_list, in bfad_im_bsg_vendor_request() 3157 job->request_payload.sg_cnt, payload_kbuf, in bfad_im_bsg_vendor_request() 3158 job->request_payload.payload_len); in bfad_im_bsg_vendor_request() 3162 job->request_payload.payload_len); in bfad_im_bsg_vendor_request() 3167 sg_copy_from_buffer(job->reply_payload.sg_list, in bfad_im_bsg_vendor_request() [all …]
|
D | bfad_im.h | 168 int bfad_im_bsg_request(struct fc_bsg_job *job); 169 int bfad_im_bsg_timeout(struct fc_bsg_job *job);
|
/drivers/crypto/caam/ |
D | Kconfig | 7 This module creates job ring devices, and configures h/w 21 and Assurance Module (CAAM). This module adds a job ring operation 64 equal or greater than the job ring size will force timeouts. 85 stack) to the SEC4 via job ring. 97 scatterlist crypto API to the SEC4 via job ring.
|
/drivers/scsi/ibmvscsi/ |
D | ibmvfc.c | 1714 static int ibmvfc_bsg_timeout(struct fc_bsg_job *job) in ibmvfc_bsg_timeout() argument 1716 struct ibmvfc_host *vhost = shost_priv(job->shost); in ibmvfc_bsg_timeout() 1717 unsigned long port_id = (unsigned long)job->dd_data; in ibmvfc_bsg_timeout() 1827 static int ibmvfc_bsg_request(struct fc_bsg_job *job) in ibmvfc_bsg_request() argument 1829 struct ibmvfc_host *vhost = shost_priv(job->shost); in ibmvfc_bsg_request() 1830 struct fc_rport *rport = job->rport; in ibmvfc_bsg_request() 1835 unsigned int code = job->request->msgcode; in ibmvfc_bsg_request() 1840 job->reply->reply_payload_rcv_len = 0; in ibmvfc_bsg_request() 1846 port_id = (job->request->rqst_data.h_els.port_id[0] << 16) | in ibmvfc_bsg_request() 1847 (job->request->rqst_data.h_els.port_id[1] << 8) | in ibmvfc_bsg_request() [all …]
|
/drivers/scsi/be2iscsi/ |
D | be_mgmt.c | 483 struct bsg_job *job, in mgmt_vendor_specific_fw_cmd() argument 490 struct iscsi_bsg_request *bsg_req = job->request; in mgmt_vendor_specific_fw_cmd() 494 nonemb_cmd->size = job->request_payload.payload_len; in mgmt_vendor_specific_fw_cmd() 511 sg_copy_to_buffer(job->request_payload.sg_list, in mgmt_vendor_specific_fw_cmd() 512 job->request_payload.sg_cnt, in mgmt_vendor_specific_fw_cmd() 513 nonemb_cmd->va + offset, job->request_len); in mgmt_vendor_specific_fw_cmd() 537 job->request_payload.sg_cnt); in mgmt_vendor_specific_fw_cmd()
|
D | be_mgmt.h | 114 struct bsg_job *job,
|
D | be_main.c | 5110 static int beiscsi_bsg_request(struct bsg_job *job) in beiscsi_bsg_request() argument 5114 struct iscsi_bsg_request *bsg_req = job->request; in beiscsi_bsg_request() 5119 struct iscsi_bsg_reply *bsg_reply = job->reply; in beiscsi_bsg_request() 5122 shost = iscsi_job_to_shost(job); in beiscsi_bsg_request() 5128 job->request_payload.payload_len, in beiscsi_bsg_request() 5136 tag = mgmt_vendor_specific_fw_cmd(&phba->ctrl, phba, job, in beiscsi_bsg_request() 5156 sg_copy_from_buffer(job->reply_payload.sg_list, in beiscsi_bsg_request() 5157 job->reply_payload.sg_cnt, in beiscsi_bsg_request() 5162 bsg_job_done(job, bsg_reply->result, in beiscsi_bsg_request()
|