Home
last modified time | relevance | path

Searched refs:request (Results 1 – 25 of 479) sorted by relevance

12345678910>>...20

/drivers/s390/char/
Dtape_std.c37 struct tape_request * request; in tape_std_assign_timeout() local
41 request = (struct tape_request *) data; in tape_std_assign_timeout()
42 device = request->device; in tape_std_assign_timeout()
47 rc = tape_cancel_io(device, request); in tape_std_assign_timeout()
58 struct tape_request *request; in tape_std_assign() local
60 request = tape_alloc_request(2, 11); in tape_std_assign()
61 if (IS_ERR(request)) in tape_std_assign()
62 return PTR_ERR(request); in tape_std_assign()
64 request->op = TO_ASSIGN; in tape_std_assign()
65 tape_ccw_cc(request->cpaddr, ASSIGN, 11, request->cpdata); in tape_std_assign()
[all …]
Dtape_34xx.c55 static void __tape_34xx_medium_sense(struct tape_request *request) in __tape_34xx_medium_sense() argument
57 struct tape_device *device = request->device; in __tape_34xx_medium_sense()
60 if (request->rc == 0) { in __tape_34xx_medium_sense()
61 sense = request->cpdata; in __tape_34xx_medium_sense()
80 request->rc); in __tape_34xx_medium_sense()
81 tape_free_request(request); in __tape_34xx_medium_sense()
86 struct tape_request *request; in tape_34xx_medium_sense() local
89 request = tape_alloc_request(1, 32); in tape_34xx_medium_sense()
90 if (IS_ERR(request)) { in tape_34xx_medium_sense()
92 return PTR_ERR(request); in tape_34xx_medium_sense()
[all …]
Dtape_3590.c203 struct tape_request *request; in tape_3592_kekl_query() local
212 request = tape_alloc_request(2, sizeof(*order)); in tape_3592_kekl_query()
213 if (IS_ERR(request)) { in tape_3592_kekl_query()
214 rc = PTR_ERR(request); in tape_3592_kekl_query()
217 order = request->cpdata; in tape_3592_kekl_query()
221 request->op = TO_KEKL_QUERY; in tape_3592_kekl_query()
222 tape_ccw_cc(request->cpaddr, PERF_SUBSYS_FUNC, sizeof(*order), order); in tape_3592_kekl_query()
223 tape_ccw_end(request->cpaddr + 1, READ_SS_DATA, sizeof(*int_kekls), in tape_3592_kekl_query()
225 rc = tape_do_io(device, request); in tape_3592_kekl_query()
232 tape_free_request(request); in tape_3592_kekl_query()
[all …]
Dtape_core.c291 __tape_cancel_io(struct tape_device *device, struct tape_request *request) in __tape_cancel_io() argument
297 if (request->callback == NULL) in __tape_cancel_io()
302 rc = ccw_device_clear(device->cdev, (long) request); in __tape_cancel_io()
306 request->status = TAPE_REQUEST_DONE; in __tape_cancel_io()
309 request->status = TAPE_REQUEST_CANCEL; in __tape_cancel_io()
641 struct tape_request * request; in __tape_discard_requests() local
645 request = list_entry(l, struct tape_request, list); in __tape_discard_requests()
646 if (request->status == TAPE_REQUEST_IN_IO) in __tape_discard_requests()
647 request->status = TAPE_REQUEST_DONE; in __tape_discard_requests()
648 list_del(&request->list); in __tape_discard_requests()
[all …]
Dsclp_vt220.c97 static int __sclp_vt220_emit(struct sclp_vt220_request *request);
115 sclp_vt220_process_queue(struct sclp_vt220_request *request) in sclp_vt220_process_queue() argument
122 page = request->sclp_req.sccb; in sclp_vt220_process_queue()
125 list_del(&request->list); in sclp_vt220_process_queue()
128 request = NULL; in sclp_vt220_process_queue()
130 request = list_entry(sclp_vt220_outqueue.next, in sclp_vt220_process_queue()
132 if (!request || sclp_vt220_suspended) { in sclp_vt220_process_queue()
138 } while (__sclp_vt220_emit(request)); in sclp_vt220_process_queue()
139 if (request == NULL && sclp_vt220_flush_later) in sclp_vt220_process_queue()
151 sclp_vt220_callback(struct sclp_req *request, void *data) in sclp_vt220_callback() argument
[all …]
Dsclp_sdias.c86 static void sdias_callback(struct sclp_req *request, void *data) in sdias_callback() argument
140 struct sclp_req request; in sclp_sdias_blk_count() local
146 memset(&request, 0, sizeof(request)); in sclp_sdias_blk_count()
156 request.sccb = &sccb; in sclp_sdias_blk_count()
157 request.command = SCLP_CMDW_WRITE_EVENT_DATA; in sclp_sdias_blk_count()
158 request.status = SCLP_REQ_FILLED; in sclp_sdias_blk_count()
159 request.callback = sdias_callback; in sclp_sdias_blk_count()
161 rc = sdias_sclp_send(&request); in sclp_sdias_blk_count()
199 struct sclp_req request; in sclp_sdias_copy() local
205 memset(&request, 0, sizeof(request)); in sclp_sdias_copy()
[all …]
Dsclp_async.c25 static struct sclp_req *request; variable
132 request->command = SCLP_CMDW_WRITE_EVENT_DATA; in sclp_async_send_wait()
133 request->sccb = sccb; in sclp_async_send_wait()
134 request->status = SCLP_REQ_FILLED; in sclp_async_send_wait()
144 rc = sclp_add_request(request); in sclp_async_send_wait()
148 while (request->status != SCLP_REQ_DONE && in sclp_async_send_wait()
149 request->status != SCLP_REQ_FAILED) { in sclp_async_send_wait()
153 if (request->status != SCLP_REQ_DONE) in sclp_async_send_wait()
156 request->sccb)->header.response_code; in sclp_async_send_wait()
178 request = kzalloc(sizeof(struct sclp_req), GFP_KERNEL); in sclp_async_init()
[all …]
Dtape_char.c131 struct tape_request *request; in tapechar_read() local
165 request = device->discipline->read_block(device, block_size); in tapechar_read()
166 if (IS_ERR(request)) in tapechar_read()
167 return PTR_ERR(request); in tapechar_read()
169 rc = tape_do_io(device, request); in tapechar_read()
171 rc = block_size - request->rescnt; in tapechar_read()
178 tape_free_request(request); in tapechar_read()
189 struct tape_request *request; in tapechar_write() local
218 request = device->discipline->write_block(device, block_size); in tapechar_write()
219 if (IS_ERR(request)) in tapechar_write()
[all …]
/drivers/gpu/drm/radeon/
Dradeon_ioc32.c137 drm_radeon_stipple_t __user *request; in compat_radeon_cp_stipple() local
143 request = compat_alloc_user_space(sizeof(*request)); in compat_radeon_cp_stipple()
144 if (!access_ok(VERIFY_WRITE, request, sizeof(*request)) in compat_radeon_cp_stipple()
146 &request->mask)) in compat_radeon_cp_stipple()
149 return drm_ioctl(file, DRM_IOCTL_RADEON_STIPPLE, (unsigned long)request); in compat_radeon_cp_stipple()
171 drm_radeon_texture_t __user *request; in compat_radeon_cp_texture() local
183 request = compat_alloc_user_space(sizeof(*request) + sizeof(*image)); in compat_radeon_cp_texture()
184 if (!access_ok(VERIFY_WRITE, request, in compat_radeon_cp_texture()
185 sizeof(*request) + sizeof(*image))) in compat_radeon_cp_texture()
187 image = (drm_radeon_tex_image_t __user *) (request + 1); in compat_radeon_cp_texture()
[all …]
/drivers/usb/musb/
Dmusb_gadget.c56 static inline void map_dma_buffer(struct musb_request *request, in map_dma_buffer() argument
62 request->map_state = UN_MAPPED; in map_dma_buffer()
73 musb_ep->packet_sz, request->request.buf, in map_dma_buffer()
74 request->request.length); in map_dma_buffer()
78 if (request->request.dma == DMA_ADDR_INVALID) { in map_dma_buffer()
79 request->request.dma = dma_map_single( in map_dma_buffer()
81 request->request.buf, in map_dma_buffer()
82 request->request.length, in map_dma_buffer()
83 request->tx in map_dma_buffer()
86 request->map_state = MUSB_MAPPED; in map_dma_buffer()
[all …]
Dmusb_gadget_ep0.c261 struct musb_request *request; in service_zero_data_request() local
304 request = next_request(musb_ep); in service_zero_data_request()
305 if (!musb_ep->busy && request) { in service_zero_data_request()
307 musb_ep_restart(musb, request); in service_zero_data_request()
488 struct musb_request *request; in ep0_rxstate() local
492 request = next_ep0_request(musb); in ep0_rxstate()
493 req = &request->request; in ep0_rxstate()
546 struct usb_request *request; in ep0_txstate() local
557 request = &req->request; in ep0_txstate()
560 fifo_src = (u8 *) request->buf + request->actual; in ep0_txstate()
[all …]
/drivers/gpu/drm/
Ddrm_ioc32.c439 struct drm_buf_info __user *request; in compat_drm_infobufs() local
456 nbytes = sizeof(*request) + count * sizeof(struct drm_buf_desc); in compat_drm_infobufs()
457 request = compat_alloc_user_space(nbytes); in compat_drm_infobufs()
458 if (!access_ok(VERIFY_WRITE, request, nbytes)) in compat_drm_infobufs()
460 list = (struct drm_buf_desc *) (request + 1); in compat_drm_infobufs()
462 if (__put_user(count, &request->count) in compat_drm_infobufs()
463 || __put_user(list, &request->list)) in compat_drm_infobufs()
466 err = drm_ioctl(file, DRM_IOCTL_INFO_BUFS, (unsigned long)request); in compat_drm_infobufs()
470 if (__get_user(actual, &request->count)) in compat_drm_infobufs()
503 struct drm_buf_map __user *request; in compat_drm_mapbufs() local
[all …]
Ddrm_bufs.c519 struct drm_map *request = data; in drm_rmmap_ioctl() local
527 r_list->user_token == (unsigned long)request->handle && in drm_rmmap_ioctl()
601 int drm_addbufs_agp(struct drm_device * dev, struct drm_buf_desc * request) in drm_addbufs_agp() argument
622 count = request->count; in drm_addbufs_agp()
623 order = drm_order(request->size); in drm_addbufs_agp()
626 alignment = (request->flags & _DRM_PAGE_ALIGN) in drm_addbufs_agp()
632 agp_offset = dev->agp->base + request->agp_start; in drm_addbufs_agp()
753 request->count = entry->buf_count; in drm_addbufs_agp()
754 request->size = size; in drm_addbufs_agp()
764 int drm_addbufs_pci(struct drm_device * dev, struct drm_buf_desc * request) in drm_addbufs_pci() argument
[all …]
Ddrm_agpsupport.c198 int drm_agp_alloc(struct drm_device *dev, struct drm_agp_buffer *request) in drm_agp_alloc() argument
212 pages = (request->size + PAGE_SIZE - 1) / PAGE_SIZE; in drm_agp_alloc()
213 type = (u32) request->type; in drm_agp_alloc()
225 request->handle = entry->handle; in drm_agp_alloc()
226 request->physical = memory->physical; in drm_agp_alloc()
236 struct drm_agp_buffer *request = data; in drm_agp_alloc_ioctl() local
238 return drm_agp_alloc(dev, request); in drm_agp_alloc_ioctl()
274 int drm_agp_unbind(struct drm_device *dev, struct drm_agp_binding *request) in drm_agp_unbind() argument
281 if (!(entry = drm_agp_lookup_entry(dev, request->handle))) in drm_agp_unbind()
296 struct drm_agp_binding *request = data; in drm_agp_unbind_ioctl() local
[all …]
/drivers/gpu/drm/i915/
Di915_ioc32.c116 drm_i915_irq_emit_t __user *request; in compat_i915_irq_emit() local
121 request = compat_alloc_user_space(sizeof(*request)); in compat_i915_irq_emit()
122 if (!access_ok(VERIFY_WRITE, request, sizeof(*request)) in compat_i915_irq_emit()
124 &request->irq_seq)) in compat_i915_irq_emit()
128 (unsigned long)request); in compat_i915_irq_emit()
139 drm_i915_getparam_t __user *request; in compat_i915_getparam() local
144 request = compat_alloc_user_space(sizeof(*request)); in compat_i915_getparam()
145 if (!access_ok(VERIFY_WRITE, request, sizeof(*request)) in compat_i915_getparam()
146 || __put_user(req32.param, &request->param) in compat_i915_getparam()
148 &request->value)) in compat_i915_getparam()
[all …]
/drivers/net/hyperv/
Drndis_filter.c86 struct rndis_request *request; in get_rndis_request() local
91 request = kzalloc(sizeof(struct rndis_request), GFP_KERNEL); in get_rndis_request()
92 if (!request) in get_rndis_request()
95 init_completion(&request->wait_event); in get_rndis_request()
97 rndis_msg = &request->request_msg; in get_rndis_request()
111 list_add_tail(&request->list_ent, &dev->req_list); in get_rndis_request()
114 return request; in get_rndis_request()
250 struct rndis_request *request = NULL; in rndis_filter_receive_response() local
258 list_for_each_entry(request, &dev->req_list, list_ent) { in rndis_filter_receive_response()
263 if (request->request_msg.msg.init_req.req_id in rndis_filter_receive_response()
[all …]
/drivers/staging/rtl8712/
Dusb_ops.c39 u8 request; in usb_read8() local
47 request = 0x05; in usb_read8()
52 r8712_usbctrl_vendorreq(pintfpriv, request, wvalue, index, &data, len, in usb_read8()
59 u8 request; in usb_read16() local
67 request = 0x05; in usb_read16()
72 r8712_usbctrl_vendorreq(pintfpriv, request, wvalue, index, &data, len, in usb_read16()
79 u8 request; in usb_read32() local
87 request = 0x05; in usb_read32()
92 r8712_usbctrl_vendorreq(pintfpriv, request, wvalue, index, &data, len, in usb_read32()
99 u8 request; in usb_write8() local
[all …]
/drivers/firewire/
Dcore-transaction.c633 struct fw_request *request; in free_response_callback() local
635 request = container_of(packet, struct fw_request, response); in free_response_callback()
636 kfree(request); in free_response_callback()
751 struct fw_request *request; in allocate_request() local
784 request = kmalloc(sizeof(*request) + length, GFP_ATOMIC); in allocate_request()
785 if (request == NULL) in allocate_request()
788 request->response.speed = p->speed; in allocate_request()
789 request->response.timestamp = in allocate_request()
791 request->response.generation = p->generation; in allocate_request()
792 request->response.ack = 0; in allocate_request()
[all …]
/drivers/s390/cio/
Dchsc_sch.c56 struct chsc_request *request = private->request; in chsc_subchannel_irq() local
64 if (!request) { in chsc_subchannel_irq()
69 private->request = NULL; in chsc_subchannel_irq()
70 memcpy(&request->irb, irb, sizeof(*irb)); in chsc_subchannel_irq()
72 complete(&request->completion); in chsc_subchannel_irq()
110 if (private->request) { in chsc_subchannel_remove()
111 complete(&private->request->completion); in chsc_subchannel_remove()
240 struct chsc_request *request) in chsc_async() argument
252 if (private->request) { in chsc_async()
270 private->request = request; in chsc_async()
[all …]
/drivers/infiniband/hw/mthca/
Dmthca_profile.c65 struct mthca_profile *request, in mthca_make_profile() argument
100 profile[MTHCA_RES_UARC].size = request->uarc_size; in mthca_make_profile()
102 profile[MTHCA_RES_QP].num = request->num_qp; in mthca_make_profile()
103 profile[MTHCA_RES_SRQ].num = request->num_srq; in mthca_make_profile()
104 profile[MTHCA_RES_EQP].num = request->num_qp; in mthca_make_profile()
105 profile[MTHCA_RES_RDB].num = request->num_qp * request->rdb_per_qp; in mthca_make_profile()
106 profile[MTHCA_RES_CQ].num = request->num_cq; in mthca_make_profile()
108 profile[MTHCA_RES_MCG].num = request->num_mcg; in mthca_make_profile()
109 profile[MTHCA_RES_MPT].num = request->num_mpt; in mthca_make_profile()
110 profile[MTHCA_RES_MTT].num = request->num_mtt; in mthca_make_profile()
[all …]
/drivers/isdn/hardware/eicon/
Ddadapter.c112 if (d->request) { in diva_didd_add_descriptor()
113 MAdapter.request = d->request; in diva_didd_add_descriptor()
114 dprintf = (DIVA_DI_PRINTF)d->request; in diva_didd_add_descriptor()
116 DBG_TRC(("DIMAINT registered, dprintf=%08x", d->request)) in diva_didd_add_descriptor()
120 MAdapter.request = (IDI_CALL)no_printf; in diva_didd_add_descriptor()
132 DBG_TRC(("Add adapter[%d], request=%08x", (i + 1), d->request)) in diva_didd_add_descriptor()
145 static int diva_didd_remove_descriptor(IDI_CALL request) { in diva_didd_remove_descriptor() argument
148 if (request == MAdapter.request) { in diva_didd_remove_descriptor()
152 MAdapter.request = (IDI_CALL)no_printf; in diva_didd_remove_descriptor()
156 if (HandleTable[i].request == request) { in diva_didd_remove_descriptor()
[all …]
/drivers/input/misc/
Duinput.c61 struct uinput_request *request) in uinput_request_alloc_id() argument
70 request->id = id; in uinput_request_alloc_id()
71 udev->requests[id] = request; in uinput_request_alloc_id()
92 struct uinput_request *request) in uinput_request_reserve_slot() argument
96 uinput_request_alloc_id(udev, request)); in uinput_request_reserve_slot()
100 struct uinput_request *request) in uinput_request_done() argument
103 udev->requests[request->id] = NULL; in uinput_request_done()
106 complete(&request->done); in uinput_request_done()
110 struct uinput_request *request) in uinput_request_send() argument
123 init_completion(&request->done); in uinput_request_send()
[all …]
/drivers/s390/block/
Dscm_blk_cluster.c64 firstA = ((u64) blk_rq_pos(A->request) << 9) / CLUSTER_SIZE; in clusters_intersect()
65 lastA = (((u64) blk_rq_pos(A->request) << 9) + in clusters_intersect()
66 blk_rq_bytes(A->request) - 1) / CLUSTER_SIZE; in clusters_intersect()
68 firstB = ((u64) blk_rq_pos(B->request) << 9) / CLUSTER_SIZE; in clusters_intersect()
69 lastB = (((u64) blk_rq_pos(B->request) << 9) + in clusters_intersect()
70 blk_rq_bytes(B->request) - 1) / CLUSTER_SIZE; in clusters_intersect()
86 (rq_data_dir(scmrq->request) == WRITE || in scm_reserve_cluster()
87 rq_data_dir(iter->request) == WRITE)) { in scm_reserve_cluster()
121 struct request *req = scmrq->request; in scm_prepare_cluster_request()
134 scmrq->aob->request.msb_count = 1; in scm_prepare_cluster_request()
[all …]
/drivers/scsi/
Dscsi_tgt_lib.c50 struct request *rq;
82 struct request *rq; in scsi_host_get_command()
109 cmd->request = rq; in scsi_host_get_command()
148 struct request *rq = cmd->request; in scsi_host_put_command()
164 struct request_queue *q = cmd->request->q; in cmd_hashlist_del()
167 struct scsi_tgt_cmd *tcmd = cmd->request->end_io_data; in cmd_hashlist_del()
186 rq_data_dir(cmd->request)); in scsi_tgt_cmd_destroy()
192 static void init_scsi_tgt_cmd(struct request *rq, struct scsi_tgt_cmd *tcmd, in init_scsi_tgt_cmd()
297 struct scsi_tgt_queuedata *queue = cmd->request->q->queuedata; in scsi_tgt_cmd_to_host()
311 struct scsi_tgt_cmd *tcmd = cmd->request->end_io_data; in scsi_tgt_queue_command()
[all …]
/drivers/media/usb/pwc/
Dpwc-ctrl.c108 u8 request, u16 value, int recv_count) in recv_control_msg() argument
113 request, in recv_control_msg()
119 rc, request, value); in recv_control_msg()
144 u8 request, u16 value, void *buf, int buflen) in send_control_msg() argument
147 request, in send_control_msg()
412 int pwc_get_u8_ctrl(struct pwc_device *pdev, u8 request, u16 value, int *data) in pwc_get_u8_ctrl() argument
416 ret = recv_control_msg(pdev, request, value, 1); in pwc_get_u8_ctrl()
424 int pwc_set_u8_ctrl(struct pwc_device *pdev, u8 request, u16 value, u8 data) in pwc_set_u8_ctrl() argument
429 ret = send_control_msg(pdev, request, value, pdev->ctrl_buf, 1); in pwc_set_u8_ctrl()
436 int pwc_get_s8_ctrl(struct pwc_device *pdev, u8 request, u16 value, int *data) in pwc_get_s8_ctrl() argument
[all …]

12345678910>>...20