/drivers/s390/char/ |
D | tape_std.c | 37 struct tape_request * request; in tape_std_assign_timeout() local 41 request = (struct tape_request *) data; in tape_std_assign_timeout() 42 device = request->device; in tape_std_assign_timeout() 47 rc = tape_cancel_io(device, request); in tape_std_assign_timeout() 58 struct tape_request *request; in tape_std_assign() local 60 request = tape_alloc_request(2, 11); in tape_std_assign() 61 if (IS_ERR(request)) in tape_std_assign() 62 return PTR_ERR(request); in tape_std_assign() 64 request->op = TO_ASSIGN; in tape_std_assign() 65 tape_ccw_cc(request->cpaddr, ASSIGN, 11, request->cpdata); in tape_std_assign() [all …]
|
D | tape_34xx.c | 55 static void __tape_34xx_medium_sense(struct tape_request *request) in __tape_34xx_medium_sense() argument 57 struct tape_device *device = request->device; in __tape_34xx_medium_sense() 60 if (request->rc == 0) { in __tape_34xx_medium_sense() 61 sense = request->cpdata; in __tape_34xx_medium_sense() 80 request->rc); in __tape_34xx_medium_sense() 81 tape_free_request(request); in __tape_34xx_medium_sense() 86 struct tape_request *request; in tape_34xx_medium_sense() local 89 request = tape_alloc_request(1, 32); in tape_34xx_medium_sense() 90 if (IS_ERR(request)) { in tape_34xx_medium_sense() 92 return PTR_ERR(request); in tape_34xx_medium_sense() [all …]
|
D | tape_3590.c | 203 struct tape_request *request; in tape_3592_kekl_query() local 212 request = tape_alloc_request(2, sizeof(*order)); in tape_3592_kekl_query() 213 if (IS_ERR(request)) { in tape_3592_kekl_query() 214 rc = PTR_ERR(request); in tape_3592_kekl_query() 217 order = request->cpdata; in tape_3592_kekl_query() 221 request->op = TO_KEKL_QUERY; in tape_3592_kekl_query() 222 tape_ccw_cc(request->cpaddr, PERF_SUBSYS_FUNC, sizeof(*order), order); in tape_3592_kekl_query() 223 tape_ccw_end(request->cpaddr + 1, READ_SS_DATA, sizeof(*int_kekls), in tape_3592_kekl_query() 225 rc = tape_do_io(device, request); in tape_3592_kekl_query() 232 tape_free_request(request); in tape_3592_kekl_query() [all …]
|
D | tape_core.c | 291 __tape_cancel_io(struct tape_device *device, struct tape_request *request) in __tape_cancel_io() argument 297 if (request->callback == NULL) in __tape_cancel_io() 302 rc = ccw_device_clear(device->cdev, (long) request); in __tape_cancel_io() 306 request->status = TAPE_REQUEST_DONE; in __tape_cancel_io() 309 request->status = TAPE_REQUEST_CANCEL; in __tape_cancel_io() 641 struct tape_request * request; in __tape_discard_requests() local 645 request = list_entry(l, struct tape_request, list); in __tape_discard_requests() 646 if (request->status == TAPE_REQUEST_IN_IO) in __tape_discard_requests() 647 request->status = TAPE_REQUEST_DONE; in __tape_discard_requests() 648 list_del(&request->list); in __tape_discard_requests() [all …]
|
D | sclp_vt220.c | 97 static int __sclp_vt220_emit(struct sclp_vt220_request *request); 115 sclp_vt220_process_queue(struct sclp_vt220_request *request) in sclp_vt220_process_queue() argument 122 page = request->sclp_req.sccb; in sclp_vt220_process_queue() 125 list_del(&request->list); in sclp_vt220_process_queue() 128 request = NULL; in sclp_vt220_process_queue() 130 request = list_entry(sclp_vt220_outqueue.next, in sclp_vt220_process_queue() 132 if (!request || sclp_vt220_suspended) { in sclp_vt220_process_queue() 138 } while (__sclp_vt220_emit(request)); in sclp_vt220_process_queue() 139 if (request == NULL && sclp_vt220_flush_later) in sclp_vt220_process_queue() 151 sclp_vt220_callback(struct sclp_req *request, void *data) in sclp_vt220_callback() argument [all …]
|
D | sclp_sdias.c | 86 static void sdias_callback(struct sclp_req *request, void *data) in sdias_callback() argument 140 struct sclp_req request; in sclp_sdias_blk_count() local 146 memset(&request, 0, sizeof(request)); in sclp_sdias_blk_count() 156 request.sccb = &sccb; in sclp_sdias_blk_count() 157 request.command = SCLP_CMDW_WRITE_EVENT_DATA; in sclp_sdias_blk_count() 158 request.status = SCLP_REQ_FILLED; in sclp_sdias_blk_count() 159 request.callback = sdias_callback; in sclp_sdias_blk_count() 161 rc = sdias_sclp_send(&request); in sclp_sdias_blk_count() 199 struct sclp_req request; in sclp_sdias_copy() local 205 memset(&request, 0, sizeof(request)); in sclp_sdias_copy() [all …]
|
D | sclp_async.c | 25 static struct sclp_req *request; variable 132 request->command = SCLP_CMDW_WRITE_EVENT_DATA; in sclp_async_send_wait() 133 request->sccb = sccb; in sclp_async_send_wait() 134 request->status = SCLP_REQ_FILLED; in sclp_async_send_wait() 144 rc = sclp_add_request(request); in sclp_async_send_wait() 148 while (request->status != SCLP_REQ_DONE && in sclp_async_send_wait() 149 request->status != SCLP_REQ_FAILED) { in sclp_async_send_wait() 153 if (request->status != SCLP_REQ_DONE) in sclp_async_send_wait() 156 request->sccb)->header.response_code; in sclp_async_send_wait() 178 request = kzalloc(sizeof(struct sclp_req), GFP_KERNEL); in sclp_async_init() [all …]
|
D | tape_char.c | 131 struct tape_request *request; in tapechar_read() local 165 request = device->discipline->read_block(device, block_size); in tapechar_read() 166 if (IS_ERR(request)) in tapechar_read() 167 return PTR_ERR(request); in tapechar_read() 169 rc = tape_do_io(device, request); in tapechar_read() 171 rc = block_size - request->rescnt; in tapechar_read() 178 tape_free_request(request); in tapechar_read() 189 struct tape_request *request; in tapechar_write() local 218 request = device->discipline->write_block(device, block_size); in tapechar_write() 219 if (IS_ERR(request)) in tapechar_write() [all …]
|
/drivers/gpu/drm/radeon/ |
D | radeon_ioc32.c | 137 drm_radeon_stipple_t __user *request; in compat_radeon_cp_stipple() local 143 request = compat_alloc_user_space(sizeof(*request)); in compat_radeon_cp_stipple() 144 if (!access_ok(VERIFY_WRITE, request, sizeof(*request)) in compat_radeon_cp_stipple() 146 &request->mask)) in compat_radeon_cp_stipple() 149 return drm_ioctl(file, DRM_IOCTL_RADEON_STIPPLE, (unsigned long)request); in compat_radeon_cp_stipple() 171 drm_radeon_texture_t __user *request; in compat_radeon_cp_texture() local 183 request = compat_alloc_user_space(sizeof(*request) + sizeof(*image)); in compat_radeon_cp_texture() 184 if (!access_ok(VERIFY_WRITE, request, in compat_radeon_cp_texture() 185 sizeof(*request) + sizeof(*image))) in compat_radeon_cp_texture() 187 image = (drm_radeon_tex_image_t __user *) (request + 1); in compat_radeon_cp_texture() [all …]
|
/drivers/usb/musb/ |
D | musb_gadget.c | 56 static inline void map_dma_buffer(struct musb_request *request, in map_dma_buffer() argument 62 request->map_state = UN_MAPPED; in map_dma_buffer() 73 musb_ep->packet_sz, request->request.buf, in map_dma_buffer() 74 request->request.length); in map_dma_buffer() 78 if (request->request.dma == DMA_ADDR_INVALID) { in map_dma_buffer() 79 request->request.dma = dma_map_single( in map_dma_buffer() 81 request->request.buf, in map_dma_buffer() 82 request->request.length, in map_dma_buffer() 83 request->tx in map_dma_buffer() 86 request->map_state = MUSB_MAPPED; in map_dma_buffer() [all …]
|
D | musb_gadget_ep0.c | 261 struct musb_request *request; in service_zero_data_request() local 304 request = next_request(musb_ep); in service_zero_data_request() 305 if (!musb_ep->busy && request) { in service_zero_data_request() 307 musb_ep_restart(musb, request); in service_zero_data_request() 488 struct musb_request *request; in ep0_rxstate() local 492 request = next_ep0_request(musb); in ep0_rxstate() 493 req = &request->request; in ep0_rxstate() 546 struct usb_request *request; in ep0_txstate() local 557 request = &req->request; in ep0_txstate() 560 fifo_src = (u8 *) request->buf + request->actual; in ep0_txstate() [all …]
|
/drivers/gpu/drm/ |
D | drm_ioc32.c | 439 struct drm_buf_info __user *request; in compat_drm_infobufs() local 456 nbytes = sizeof(*request) + count * sizeof(struct drm_buf_desc); in compat_drm_infobufs() 457 request = compat_alloc_user_space(nbytes); in compat_drm_infobufs() 458 if (!access_ok(VERIFY_WRITE, request, nbytes)) in compat_drm_infobufs() 460 list = (struct drm_buf_desc *) (request + 1); in compat_drm_infobufs() 462 if (__put_user(count, &request->count) in compat_drm_infobufs() 463 || __put_user(list, &request->list)) in compat_drm_infobufs() 466 err = drm_ioctl(file, DRM_IOCTL_INFO_BUFS, (unsigned long)request); in compat_drm_infobufs() 470 if (__get_user(actual, &request->count)) in compat_drm_infobufs() 503 struct drm_buf_map __user *request; in compat_drm_mapbufs() local [all …]
|
D | drm_bufs.c | 519 struct drm_map *request = data; in drm_rmmap_ioctl() local 527 r_list->user_token == (unsigned long)request->handle && in drm_rmmap_ioctl() 601 int drm_addbufs_agp(struct drm_device * dev, struct drm_buf_desc * request) in drm_addbufs_agp() argument 622 count = request->count; in drm_addbufs_agp() 623 order = drm_order(request->size); in drm_addbufs_agp() 626 alignment = (request->flags & _DRM_PAGE_ALIGN) in drm_addbufs_agp() 632 agp_offset = dev->agp->base + request->agp_start; in drm_addbufs_agp() 753 request->count = entry->buf_count; in drm_addbufs_agp() 754 request->size = size; in drm_addbufs_agp() 764 int drm_addbufs_pci(struct drm_device * dev, struct drm_buf_desc * request) in drm_addbufs_pci() argument [all …]
|
D | drm_agpsupport.c | 198 int drm_agp_alloc(struct drm_device *dev, struct drm_agp_buffer *request) in drm_agp_alloc() argument 212 pages = (request->size + PAGE_SIZE - 1) / PAGE_SIZE; in drm_agp_alloc() 213 type = (u32) request->type; in drm_agp_alloc() 225 request->handle = entry->handle; in drm_agp_alloc() 226 request->physical = memory->physical; in drm_agp_alloc() 236 struct drm_agp_buffer *request = data; in drm_agp_alloc_ioctl() local 238 return drm_agp_alloc(dev, request); in drm_agp_alloc_ioctl() 274 int drm_agp_unbind(struct drm_device *dev, struct drm_agp_binding *request) in drm_agp_unbind() argument 281 if (!(entry = drm_agp_lookup_entry(dev, request->handle))) in drm_agp_unbind() 296 struct drm_agp_binding *request = data; in drm_agp_unbind_ioctl() local [all …]
|
/drivers/gpu/drm/i915/ |
D | i915_ioc32.c | 116 drm_i915_irq_emit_t __user *request; in compat_i915_irq_emit() local 121 request = compat_alloc_user_space(sizeof(*request)); in compat_i915_irq_emit() 122 if (!access_ok(VERIFY_WRITE, request, sizeof(*request)) in compat_i915_irq_emit() 124 &request->irq_seq)) in compat_i915_irq_emit() 128 (unsigned long)request); in compat_i915_irq_emit() 139 drm_i915_getparam_t __user *request; in compat_i915_getparam() local 144 request = compat_alloc_user_space(sizeof(*request)); in compat_i915_getparam() 145 if (!access_ok(VERIFY_WRITE, request, sizeof(*request)) in compat_i915_getparam() 146 || __put_user(req32.param, &request->param) in compat_i915_getparam() 148 &request->value)) in compat_i915_getparam() [all …]
|
/drivers/net/hyperv/ |
D | rndis_filter.c | 86 struct rndis_request *request; in get_rndis_request() local 91 request = kzalloc(sizeof(struct rndis_request), GFP_KERNEL); in get_rndis_request() 92 if (!request) in get_rndis_request() 95 init_completion(&request->wait_event); in get_rndis_request() 97 rndis_msg = &request->request_msg; in get_rndis_request() 111 list_add_tail(&request->list_ent, &dev->req_list); in get_rndis_request() 114 return request; in get_rndis_request() 250 struct rndis_request *request = NULL; in rndis_filter_receive_response() local 258 list_for_each_entry(request, &dev->req_list, list_ent) { in rndis_filter_receive_response() 263 if (request->request_msg.msg.init_req.req_id in rndis_filter_receive_response() [all …]
|
/drivers/staging/rtl8712/ |
D | usb_ops.c | 39 u8 request; in usb_read8() local 47 request = 0x05; in usb_read8() 52 r8712_usbctrl_vendorreq(pintfpriv, request, wvalue, index, &data, len, in usb_read8() 59 u8 request; in usb_read16() local 67 request = 0x05; in usb_read16() 72 r8712_usbctrl_vendorreq(pintfpriv, request, wvalue, index, &data, len, in usb_read16() 79 u8 request; in usb_read32() local 87 request = 0x05; in usb_read32() 92 r8712_usbctrl_vendorreq(pintfpriv, request, wvalue, index, &data, len, in usb_read32() 99 u8 request; in usb_write8() local [all …]
|
/drivers/firewire/ |
D | core-transaction.c | 633 struct fw_request *request; in free_response_callback() local 635 request = container_of(packet, struct fw_request, response); in free_response_callback() 636 kfree(request); in free_response_callback() 751 struct fw_request *request; in allocate_request() local 784 request = kmalloc(sizeof(*request) + length, GFP_ATOMIC); in allocate_request() 785 if (request == NULL) in allocate_request() 788 request->response.speed = p->speed; in allocate_request() 789 request->response.timestamp = in allocate_request() 791 request->response.generation = p->generation; in allocate_request() 792 request->response.ack = 0; in allocate_request() [all …]
|
/drivers/s390/cio/ |
D | chsc_sch.c | 56 struct chsc_request *request = private->request; in chsc_subchannel_irq() local 64 if (!request) { in chsc_subchannel_irq() 69 private->request = NULL; in chsc_subchannel_irq() 70 memcpy(&request->irb, irb, sizeof(*irb)); in chsc_subchannel_irq() 72 complete(&request->completion); in chsc_subchannel_irq() 110 if (private->request) { in chsc_subchannel_remove() 111 complete(&private->request->completion); in chsc_subchannel_remove() 240 struct chsc_request *request) in chsc_async() argument 252 if (private->request) { in chsc_async() 270 private->request = request; in chsc_async() [all …]
|
/drivers/infiniband/hw/mthca/ |
D | mthca_profile.c | 65 struct mthca_profile *request, in mthca_make_profile() argument 100 profile[MTHCA_RES_UARC].size = request->uarc_size; in mthca_make_profile() 102 profile[MTHCA_RES_QP].num = request->num_qp; in mthca_make_profile() 103 profile[MTHCA_RES_SRQ].num = request->num_srq; in mthca_make_profile() 104 profile[MTHCA_RES_EQP].num = request->num_qp; in mthca_make_profile() 105 profile[MTHCA_RES_RDB].num = request->num_qp * request->rdb_per_qp; in mthca_make_profile() 106 profile[MTHCA_RES_CQ].num = request->num_cq; in mthca_make_profile() 108 profile[MTHCA_RES_MCG].num = request->num_mcg; in mthca_make_profile() 109 profile[MTHCA_RES_MPT].num = request->num_mpt; in mthca_make_profile() 110 profile[MTHCA_RES_MTT].num = request->num_mtt; in mthca_make_profile() [all …]
|
/drivers/isdn/hardware/eicon/ |
D | dadapter.c | 112 if (d->request) { in diva_didd_add_descriptor() 113 MAdapter.request = d->request; in diva_didd_add_descriptor() 114 dprintf = (DIVA_DI_PRINTF)d->request; in diva_didd_add_descriptor() 116 DBG_TRC(("DIMAINT registered, dprintf=%08x", d->request)) in diva_didd_add_descriptor() 120 MAdapter.request = (IDI_CALL)no_printf; in diva_didd_add_descriptor() 132 DBG_TRC(("Add adapter[%d], request=%08x", (i + 1), d->request)) in diva_didd_add_descriptor() 145 static int diva_didd_remove_descriptor(IDI_CALL request) { in diva_didd_remove_descriptor() argument 148 if (request == MAdapter.request) { in diva_didd_remove_descriptor() 152 MAdapter.request = (IDI_CALL)no_printf; in diva_didd_remove_descriptor() 156 if (HandleTable[i].request == request) { in diva_didd_remove_descriptor() [all …]
|
/drivers/input/misc/ |
D | uinput.c | 61 struct uinput_request *request) in uinput_request_alloc_id() argument 70 request->id = id; in uinput_request_alloc_id() 71 udev->requests[id] = request; in uinput_request_alloc_id() 92 struct uinput_request *request) in uinput_request_reserve_slot() argument 96 uinput_request_alloc_id(udev, request)); in uinput_request_reserve_slot() 100 struct uinput_request *request) in uinput_request_done() argument 103 udev->requests[request->id] = NULL; in uinput_request_done() 106 complete(&request->done); in uinput_request_done() 110 struct uinput_request *request) in uinput_request_send() argument 123 init_completion(&request->done); in uinput_request_send() [all …]
|
/drivers/s390/block/ |
D | scm_blk_cluster.c | 64 firstA = ((u64) blk_rq_pos(A->request) << 9) / CLUSTER_SIZE; in clusters_intersect() 65 lastA = (((u64) blk_rq_pos(A->request) << 9) + in clusters_intersect() 66 blk_rq_bytes(A->request) - 1) / CLUSTER_SIZE; in clusters_intersect() 68 firstB = ((u64) blk_rq_pos(B->request) << 9) / CLUSTER_SIZE; in clusters_intersect() 69 lastB = (((u64) blk_rq_pos(B->request) << 9) + in clusters_intersect() 70 blk_rq_bytes(B->request) - 1) / CLUSTER_SIZE; in clusters_intersect() 86 (rq_data_dir(scmrq->request) == WRITE || in scm_reserve_cluster() 87 rq_data_dir(iter->request) == WRITE)) { in scm_reserve_cluster() 121 struct request *req = scmrq->request; in scm_prepare_cluster_request() 134 scmrq->aob->request.msb_count = 1; in scm_prepare_cluster_request() [all …]
|
/drivers/scsi/ |
D | scsi_tgt_lib.c | 50 struct request *rq; 82 struct request *rq; in scsi_host_get_command() 109 cmd->request = rq; in scsi_host_get_command() 148 struct request *rq = cmd->request; in scsi_host_put_command() 164 struct request_queue *q = cmd->request->q; in cmd_hashlist_del() 167 struct scsi_tgt_cmd *tcmd = cmd->request->end_io_data; in cmd_hashlist_del() 186 rq_data_dir(cmd->request)); in scsi_tgt_cmd_destroy() 192 static void init_scsi_tgt_cmd(struct request *rq, struct scsi_tgt_cmd *tcmd, in init_scsi_tgt_cmd() 297 struct scsi_tgt_queuedata *queue = cmd->request->q->queuedata; in scsi_tgt_cmd_to_host() 311 struct scsi_tgt_cmd *tcmd = cmd->request->end_io_data; in scsi_tgt_queue_command() [all …]
|
/drivers/media/usb/pwc/ |
D | pwc-ctrl.c | 108 u8 request, u16 value, int recv_count) in recv_control_msg() argument 113 request, in recv_control_msg() 119 rc, request, value); in recv_control_msg() 144 u8 request, u16 value, void *buf, int buflen) in send_control_msg() argument 147 request, in send_control_msg() 412 int pwc_get_u8_ctrl(struct pwc_device *pdev, u8 request, u16 value, int *data) in pwc_get_u8_ctrl() argument 416 ret = recv_control_msg(pdev, request, value, 1); in pwc_get_u8_ctrl() 424 int pwc_set_u8_ctrl(struct pwc_device *pdev, u8 request, u16 value, u8 data) in pwc_set_u8_ctrl() argument 429 ret = send_control_msg(pdev, request, value, pdev->ctrl_buf, 1); in pwc_set_u8_ctrl() 436 int pwc_get_s8_ctrl(struct pwc_device *pdev, u8 request, u16 value, int *data) in pwc_get_s8_ctrl() argument [all …]
|