Home
last modified time | relevance | path

Searched refs:iob (Results 1 – 24 of 24) sorted by relevance

/drivers/rtc/
Drtc-gamecube.c92 void __iomem *iob; member
99 void __iomem *iob = d->iob; in exi_read() local
106 iowrite32be(RTC_EXICSR, iob + EXICSR); in exi_read()
107 iowrite32be(reg << 8, iob + EXIDATA); in exi_read()
108 iowrite32be(RTC_EXICR_W, iob + EXICR); in exi_read()
109 while (!(ioread32be(iob + EXICSR) & EXICSR_INTSET)) in exi_read()
113 iowrite32be(RTC_EXICSR, iob + EXICSR); in exi_read()
114 iowrite32be(RTC_EXICR_R, iob + EXICR); in exi_read()
115 while (!(ioread32be(iob + EXICSR) & EXICSR_INTSET)) in exi_read()
117 *data = ioread32be(iob + EXIDATA); in exi_read()
[all …]
/drivers/s390/net/
Dqeth_core_main.c69 struct qeth_cmd_buffer *iob,
416 static void qeth_put_cmd(struct qeth_cmd_buffer *iob) in qeth_put_cmd() argument
418 if (refcount_dec_and_test(&iob->ref_count)) { in qeth_put_cmd()
419 kfree(iob->data); in qeth_put_cmd()
420 kfree(iob); in qeth_put_cmd()
434 struct qeth_cmd_buffer *iob = card->read_cmd; in __qeth_issue_next_read() local
435 struct qeth_channel *channel = iob->channel; in __qeth_issue_next_read()
436 struct ccw1 *ccw = __ccw_from_cmd(iob); in __qeth_issue_next_read()
443 memset(iob->data, 0, iob->length); in __qeth_issue_next_read()
444 qeth_setup_ccw(ccw, CCW_CMD_READ, 0, iob->length, iob->data); in __qeth_issue_next_read()
[all …]
Dqeth_l2_main.c78 struct qeth_cmd_buffer *iob; in qeth_l2_send_setdelmac() local
81 iob = qeth_ipa_alloc_cmd(card, ipacmd, QETH_PROT_IPV4, in qeth_l2_send_setdelmac()
83 if (!iob) in qeth_l2_send_setdelmac()
85 cmd = __ipa_cmd(iob); in qeth_l2_send_setdelmac()
88 return qeth_send_ipa_cmd(card, iob, qeth_l2_send_setdelmac_cb, NULL); in qeth_l2_send_setdelmac()
233 struct qeth_cmd_buffer *iob; in qeth_l2_send_setdelvlan() local
236 iob = qeth_ipa_alloc_cmd(card, ipacmd, QETH_PROT_IPV4, in qeth_l2_send_setdelvlan()
238 if (!iob) in qeth_l2_send_setdelvlan()
240 cmd = __ipa_cmd(iob); in qeth_l2_send_setdelvlan()
242 return qeth_send_ipa_cmd(card, iob, qeth_l2_send_setdelvlan_cb, NULL); in qeth_l2_send_setdelvlan()
[all …]
Dqeth_l3_main.c318 struct qeth_cmd_buffer *iob; in qeth_l3_send_setdelmc() local
323 iob = qeth_ipa_alloc_cmd(card, ipacmd, addr->proto, in qeth_l3_send_setdelmc()
325 if (!iob) in qeth_l3_send_setdelmc()
327 cmd = __ipa_cmd(iob); in qeth_l3_send_setdelmc()
336 return qeth_send_ipa_cmd(card, iob, qeth_l3_setdelip_cb, NULL); in qeth_l3_send_setdelmc()
369 struct qeth_cmd_buffer *iob; in qeth_l3_send_setdelip() local
375 iob = qeth_ipa_alloc_cmd(card, ipacmd, addr->proto, in qeth_l3_send_setdelip()
377 if (!iob) in qeth_l3_send_setdelip()
379 cmd = __ipa_cmd(iob); in qeth_l3_send_setdelip()
395 return qeth_send_ipa_cmd(card, iob, qeth_l3_setdelip_cb, NULL); in qeth_l3_send_setdelip()
[all …]
Dqeth_core.h624 void (*finalize)(struct qeth_card *card, struct qeth_cmd_buffer *iob);
625 bool (*match)(struct qeth_cmd_buffer *iob,
627 void (*callback)(struct qeth_card *card, struct qeth_cmd_buffer *iob,
632 static inline void qeth_get_cmd(struct qeth_cmd_buffer *iob) in qeth_get_cmd() argument
634 refcount_inc(&iob->ref_count); in qeth_get_cmd()
637 static inline struct qeth_ipa_cmd *__ipa_reply(struct qeth_cmd_buffer *iob) in __ipa_reply() argument
639 if (!IS_IPA(iob->data)) in __ipa_reply()
642 return (struct qeth_ipa_cmd *) PDU_ENCAPSULATION(iob->data); in __ipa_reply()
645 static inline struct qeth_ipa_cmd *__ipa_cmd(struct qeth_cmd_buffer *iob) in __ipa_cmd() argument
647 return (struct qeth_ipa_cmd *)(iob->data + IPA_PDU_HEADER_SIZE); in __ipa_cmd()
[all …]
Dlcs.c110 channel->iob[cnt].data = in lcs_alloc_channel()
112 if (channel->iob[cnt].data == NULL) in lcs_alloc_channel()
114 channel->iob[cnt].state = LCS_BUF_STATE_EMPTY; in lcs_alloc_channel()
120 kfree(channel->iob[cnt].data); in lcs_alloc_channel()
136 kfree(channel->iob[cnt].data); in lcs_free_channel()
137 channel->iob[cnt].data = NULL; in lcs_free_channel()
226 (__u32)virt_to_phys(card->read.iob[cnt].data); in lcs_setup_read_ccws()
228 card->read.iob[cnt].data)->offset = LCS_ILLEGAL_OFFSET; in lcs_setup_read_ccws()
229 card->read.iob[cnt].callback = lcs_get_frames_cb; in lcs_setup_read_ccws()
230 card->read.iob[cnt].state = LCS_BUF_STATE_READY; in lcs_setup_read_ccws()
[all …]
Dlcs.h294 struct lcs_buffer iob[LCS_NUM_BUFFS]; member
/drivers/s390/block/
Ddasd_diag.c53 struct dasd_diag_rw_io iob; member
70 static inline int __dia250(void *iob, int cmd) in __dia250() argument
72 union register_pair rx = { .even = (unsigned long)iob, }; in __dia250()
86 : [cc] "+&d" (cc), [rx] "+&d" (rx.pair), "+m" (*(addr_type *)iob) in __dia250()
92 static inline int dia250(void *iob, int cmd) in dia250() argument
95 return __dia250(iob, cmd); in dia250()
180 private->iob.dev_nr = private->dev_id.devno; in dasd_start_diag()
181 private->iob.key = 0; in dasd_start_diag()
182 private->iob.flags = DASD_DIAG_RWFLAG_ASYNC; in dasd_start_diag()
183 private->iob.block_count = dreq->block_count; in dasd_start_diag()
[all …]
/drivers/cpufreq/
Ds3c24xx-cpufreq-debugfs.c101 union s3c_iobank *iob; in io_show() local
125 iob = &iot->bank[bank]; in io_show()
129 if (!iob->io_2410) { in io_show()
134 show_bank(seq, cfg, iob); in io_show()
/drivers/nvme/host/
Dapple.c564 static void apple_nvme_complete_batch(struct io_comp_batch *iob) in apple_nvme_complete_batch() argument
566 nvme_complete_batch(iob, apple_nvme_unmap_rq); in apple_nvme_complete_batch()
586 struct io_comp_batch *iob, u16 idx) in apple_nvme_handle_cqe() argument
602 !blk_mq_add_to_batch(req, iob, nvme_req(req)->status, in apple_nvme_handle_cqe()
620 struct io_comp_batch *iob) in apple_nvme_poll_cq() argument
632 apple_nvme_handle_cqe(q, iob, q->cq_head); in apple_nvme_poll_cq()
645 DEFINE_IO_COMP_BATCH(iob); in apple_nvme_handle_cq()
650 found = apple_nvme_poll_cq(q, &iob); in apple_nvme_handle_cq()
652 if (!rq_list_empty(iob.req_list)) in apple_nvme_handle_cq()
653 apple_nvme_complete_batch(&iob); in apple_nvme_handle_cq()
[all …]
Dnvme.h732 static __always_inline void nvme_complete_batch(struct io_comp_batch *iob, in nvme_complete_batch() argument
737 rq_list_for_each(&iob->req_list, req) { in nvme_complete_batch()
741 blk_mq_end_request_batch(iob); in nvme_complete_batch()
865 struct io_comp_batch *iob, unsigned int poll_flags);
867 struct io_comp_batch *iob, unsigned int poll_flags);
Dpci.c1030 static void nvme_pci_complete_batch(struct io_comp_batch *iob) in nvme_pci_complete_batch() argument
1032 nvme_complete_batch(iob, nvme_pci_unmap_rq); in nvme_pci_complete_batch()
1060 struct io_comp_batch *iob, u16 idx) in nvme_handle_cqe() argument
1088 !blk_mq_add_to_batch(req, iob, nvme_req(req)->status, in nvme_handle_cqe()
1106 struct io_comp_batch *iob) in nvme_poll_cq() argument
1117 nvme_handle_cqe(nvmeq, iob, nvmeq->cq_head); in nvme_poll_cq()
1129 DEFINE_IO_COMP_BATCH(iob); in nvme_irq()
1131 if (nvme_poll_cq(nvmeq, &iob)) { in nvme_irq()
1132 if (!rq_list_empty(iob.req_list)) in nvme_irq()
1133 nvme_pci_complete_batch(&iob); in nvme_irq()
[all …]
Dioctl.c708 struct io_comp_batch *iob, in nvme_ns_chr_uring_cmd_iopoll() argument
722 ret = bio_poll(bio, iob, poll_flags); in nvme_ns_chr_uring_cmd_iopoll()
809 struct io_comp_batch *iob, in nvme_ns_head_chr_uring_cmd_iopoll() argument
826 ret = bio_poll(bio, iob, poll_flags); in nvme_ns_head_chr_uring_cmd_iopoll()
Dcore.c2004 u32 iob; in nvme_set_chunk_sectors() local
2008 iob = ctrl->max_hw_sectors; in nvme_set_chunk_sectors()
2010 iob = nvme_lba_to_sect(ns, le16_to_cpu(id->noiob)); in nvme_set_chunk_sectors()
2012 if (!iob) in nvme_set_chunk_sectors()
2015 if (!is_power_of_2(iob)) { in nvme_set_chunk_sectors()
2018 ns->disk->disk_name, iob); in nvme_set_chunk_sectors()
2029 blk_queue_chunk_sectors(ns->queue, iob); in nvme_set_chunk_sectors()
Drdma.c2094 static int nvme_rdma_poll(struct blk_mq_hw_ctx *hctx, struct io_comp_batch *iob) in nvme_rdma_poll() argument
Dtcp.c2474 static int nvme_tcp_poll(struct blk_mq_hw_ctx *hctx, struct io_comp_batch *iob) in nvme_tcp_poll() argument
/drivers/block/
Dvirtio_blk.c840 static void virtblk_complete_batch(struct io_comp_batch *iob) in virtblk_complete_batch() argument
844 rq_list_for_each(&iob->req_list, req) { in virtblk_complete_batch()
848 blk_mq_end_request_batch(iob); in virtblk_complete_batch()
851 static int virtblk_poll(struct blk_mq_hw_ctx *hctx, struct io_comp_batch *iob) in virtblk_poll() argument
866 if (!blk_mq_add_to_batch(req, iob, vbr->status, in virtblk_poll()
/drivers/tty/serial/
Ddz.c553 iob(); in dz_reset()
822 iob(); in dz_console_putchar()
834 iob(); in dz_console_putchar()
/drivers/staging/octeon/
Docteon-stubs.h1009 uint64_t iob:1; member
1042 uint64_t iob:1; member
1077 uint64_t iob:1; member
1111 uint64_t iob:1; member
/drivers/net/ethernet/amd/
Ddeclance.c299 iob(); in writereg()
384 iob(); in cp_to_buf()
524 iob(); in lance_init_ring()
/drivers/md/
Ddm.c1810 static bool dm_poll_dm_io(struct dm_io *io, struct io_comp_batch *iob, in dm_poll_dm_io() argument
1817 bio_poll(&io->tio.clone, iob, flags); in dm_poll_dm_io()
1823 static int dm_poll_bio(struct bio *bio, struct io_comp_batch *iob, in dm_poll_bio() argument
1850 if (dm_poll_dm_io(curr, iob, flags)) { in dm_poll_bio()
/drivers/block/null_blk/
Dmain.c1595 static int null_poll(struct blk_mq_hw_ctx *hctx, struct io_comp_batch *iob) in null_poll() argument
1617 if (!blk_mq_add_to_batch(req, iob, (__force int) cmd->error, in null_poll()
/drivers/block/rnbd/
Drnbd-clt.c1158 static int rnbd_rdma_poll(struct blk_mq_hw_ctx *hctx, struct io_comp_batch *iob) in rnbd_rdma_poll() argument
/drivers/scsi/
Dscsi_lib.c1853 static int scsi_mq_poll(struct blk_mq_hw_ctx *hctx, struct io_comp_batch *iob) in scsi_mq_poll() argument