/drivers/rtc/ |
D | rtc-gamecube.c | 92 void __iomem *iob; member 99 void __iomem *iob = d->iob; in exi_read() local 106 iowrite32be(RTC_EXICSR, iob + EXICSR); in exi_read() 107 iowrite32be(reg << 8, iob + EXIDATA); in exi_read() 108 iowrite32be(RTC_EXICR_W, iob + EXICR); in exi_read() 109 while (!(ioread32be(iob + EXICSR) & EXICSR_INTSET)) in exi_read() 113 iowrite32be(RTC_EXICSR, iob + EXICSR); in exi_read() 114 iowrite32be(RTC_EXICR_R, iob + EXICR); in exi_read() 115 while (!(ioread32be(iob + EXICSR) & EXICSR_INTSET)) in exi_read() 117 *data = ioread32be(iob + EXIDATA); in exi_read() [all …]
|
/drivers/s390/net/ |
D | qeth_core_main.c | 69 struct qeth_cmd_buffer *iob, 416 static void qeth_put_cmd(struct qeth_cmd_buffer *iob) in qeth_put_cmd() argument 418 if (refcount_dec_and_test(&iob->ref_count)) { in qeth_put_cmd() 419 kfree(iob->data); in qeth_put_cmd() 420 kfree(iob); in qeth_put_cmd() 434 struct qeth_cmd_buffer *iob = card->read_cmd; in __qeth_issue_next_read() local 435 struct qeth_channel *channel = iob->channel; in __qeth_issue_next_read() 436 struct ccw1 *ccw = __ccw_from_cmd(iob); in __qeth_issue_next_read() 443 memset(iob->data, 0, iob->length); in __qeth_issue_next_read() 444 qeth_setup_ccw(ccw, CCW_CMD_READ, 0, iob->length, iob->data); in __qeth_issue_next_read() [all …]
|
D | qeth_l2_main.c | 78 struct qeth_cmd_buffer *iob; in qeth_l2_send_setdelmac() local 81 iob = qeth_ipa_alloc_cmd(card, ipacmd, QETH_PROT_IPV4, in qeth_l2_send_setdelmac() 83 if (!iob) in qeth_l2_send_setdelmac() 85 cmd = __ipa_cmd(iob); in qeth_l2_send_setdelmac() 88 return qeth_send_ipa_cmd(card, iob, qeth_l2_send_setdelmac_cb, NULL); in qeth_l2_send_setdelmac() 233 struct qeth_cmd_buffer *iob; in qeth_l2_send_setdelvlan() local 236 iob = qeth_ipa_alloc_cmd(card, ipacmd, QETH_PROT_IPV4, in qeth_l2_send_setdelvlan() 238 if (!iob) in qeth_l2_send_setdelvlan() 240 cmd = __ipa_cmd(iob); in qeth_l2_send_setdelvlan() 242 return qeth_send_ipa_cmd(card, iob, qeth_l2_send_setdelvlan_cb, NULL); in qeth_l2_send_setdelvlan() [all …]
|
D | qeth_l3_main.c | 318 struct qeth_cmd_buffer *iob; in qeth_l3_send_setdelmc() local 323 iob = qeth_ipa_alloc_cmd(card, ipacmd, addr->proto, in qeth_l3_send_setdelmc() 325 if (!iob) in qeth_l3_send_setdelmc() 327 cmd = __ipa_cmd(iob); in qeth_l3_send_setdelmc() 336 return qeth_send_ipa_cmd(card, iob, qeth_l3_setdelip_cb, NULL); in qeth_l3_send_setdelmc() 369 struct qeth_cmd_buffer *iob; in qeth_l3_send_setdelip() local 375 iob = qeth_ipa_alloc_cmd(card, ipacmd, addr->proto, in qeth_l3_send_setdelip() 377 if (!iob) in qeth_l3_send_setdelip() 379 cmd = __ipa_cmd(iob); in qeth_l3_send_setdelip() 395 return qeth_send_ipa_cmd(card, iob, qeth_l3_setdelip_cb, NULL); in qeth_l3_send_setdelip() [all …]
|
D | qeth_core.h | 624 void (*finalize)(struct qeth_card *card, struct qeth_cmd_buffer *iob); 625 bool (*match)(struct qeth_cmd_buffer *iob, 627 void (*callback)(struct qeth_card *card, struct qeth_cmd_buffer *iob, 632 static inline void qeth_get_cmd(struct qeth_cmd_buffer *iob) in qeth_get_cmd() argument 634 refcount_inc(&iob->ref_count); in qeth_get_cmd() 637 static inline struct qeth_ipa_cmd *__ipa_reply(struct qeth_cmd_buffer *iob) in __ipa_reply() argument 639 if (!IS_IPA(iob->data)) in __ipa_reply() 642 return (struct qeth_ipa_cmd *) PDU_ENCAPSULATION(iob->data); in __ipa_reply() 645 static inline struct qeth_ipa_cmd *__ipa_cmd(struct qeth_cmd_buffer *iob) in __ipa_cmd() argument 647 return (struct qeth_ipa_cmd *)(iob->data + IPA_PDU_HEADER_SIZE); in __ipa_cmd() [all …]
|
D | lcs.c | 110 channel->iob[cnt].data = in lcs_alloc_channel() 112 if (channel->iob[cnt].data == NULL) in lcs_alloc_channel() 114 channel->iob[cnt].state = LCS_BUF_STATE_EMPTY; in lcs_alloc_channel() 120 kfree(channel->iob[cnt].data); in lcs_alloc_channel() 136 kfree(channel->iob[cnt].data); in lcs_free_channel() 137 channel->iob[cnt].data = NULL; in lcs_free_channel() 226 (__u32)virt_to_phys(card->read.iob[cnt].data); in lcs_setup_read_ccws() 228 card->read.iob[cnt].data)->offset = LCS_ILLEGAL_OFFSET; in lcs_setup_read_ccws() 229 card->read.iob[cnt].callback = lcs_get_frames_cb; in lcs_setup_read_ccws() 230 card->read.iob[cnt].state = LCS_BUF_STATE_READY; in lcs_setup_read_ccws() [all …]
|
D | lcs.h | 294 struct lcs_buffer iob[LCS_NUM_BUFFS]; member
|
/drivers/s390/block/ |
D | dasd_diag.c | 53 struct dasd_diag_rw_io iob; member 70 static inline int __dia250(void *iob, int cmd) in __dia250() argument 72 union register_pair rx = { .even = (unsigned long)iob, }; in __dia250() 86 : [cc] "+&d" (cc), [rx] "+&d" (rx.pair), "+m" (*(addr_type *)iob) in __dia250() 92 static inline int dia250(void *iob, int cmd) in dia250() argument 95 return __dia250(iob, cmd); in dia250() 180 private->iob.dev_nr = private->dev_id.devno; in dasd_start_diag() 181 private->iob.key = 0; in dasd_start_diag() 182 private->iob.flags = DASD_DIAG_RWFLAG_ASYNC; in dasd_start_diag() 183 private->iob.block_count = dreq->block_count; in dasd_start_diag() [all …]
|
/drivers/cpufreq/ |
D | s3c24xx-cpufreq-debugfs.c | 101 union s3c_iobank *iob; in io_show() local 125 iob = &iot->bank[bank]; in io_show() 129 if (!iob->io_2410) { in io_show() 134 show_bank(seq, cfg, iob); in io_show()
|
/drivers/nvme/host/ |
D | apple.c | 564 static void apple_nvme_complete_batch(struct io_comp_batch *iob) in apple_nvme_complete_batch() argument 566 nvme_complete_batch(iob, apple_nvme_unmap_rq); in apple_nvme_complete_batch() 586 struct io_comp_batch *iob, u16 idx) in apple_nvme_handle_cqe() argument 602 !blk_mq_add_to_batch(req, iob, nvme_req(req)->status, in apple_nvme_handle_cqe() 620 struct io_comp_batch *iob) in apple_nvme_poll_cq() argument 632 apple_nvme_handle_cqe(q, iob, q->cq_head); in apple_nvme_poll_cq() 645 DEFINE_IO_COMP_BATCH(iob); in apple_nvme_handle_cq() 650 found = apple_nvme_poll_cq(q, &iob); in apple_nvme_handle_cq() 652 if (!rq_list_empty(iob.req_list)) in apple_nvme_handle_cq() 653 apple_nvme_complete_batch(&iob); in apple_nvme_handle_cq() [all …]
|
D | nvme.h | 732 static __always_inline void nvme_complete_batch(struct io_comp_batch *iob, in nvme_complete_batch() argument 737 rq_list_for_each(&iob->req_list, req) { in nvme_complete_batch() 741 blk_mq_end_request_batch(iob); in nvme_complete_batch() 865 struct io_comp_batch *iob, unsigned int poll_flags); 867 struct io_comp_batch *iob, unsigned int poll_flags);
|
D | pci.c | 1030 static void nvme_pci_complete_batch(struct io_comp_batch *iob) in nvme_pci_complete_batch() argument 1032 nvme_complete_batch(iob, nvme_pci_unmap_rq); in nvme_pci_complete_batch() 1060 struct io_comp_batch *iob, u16 idx) in nvme_handle_cqe() argument 1088 !blk_mq_add_to_batch(req, iob, nvme_req(req)->status, in nvme_handle_cqe() 1106 struct io_comp_batch *iob) in nvme_poll_cq() argument 1117 nvme_handle_cqe(nvmeq, iob, nvmeq->cq_head); in nvme_poll_cq() 1129 DEFINE_IO_COMP_BATCH(iob); in nvme_irq() 1131 if (nvme_poll_cq(nvmeq, &iob)) { in nvme_irq() 1132 if (!rq_list_empty(iob.req_list)) in nvme_irq() 1133 nvme_pci_complete_batch(&iob); in nvme_irq() [all …]
|
D | ioctl.c | 708 struct io_comp_batch *iob, in nvme_ns_chr_uring_cmd_iopoll() argument 722 ret = bio_poll(bio, iob, poll_flags); in nvme_ns_chr_uring_cmd_iopoll() 809 struct io_comp_batch *iob, in nvme_ns_head_chr_uring_cmd_iopoll() argument 826 ret = bio_poll(bio, iob, poll_flags); in nvme_ns_head_chr_uring_cmd_iopoll()
|
D | core.c | 2004 u32 iob; in nvme_set_chunk_sectors() local 2008 iob = ctrl->max_hw_sectors; in nvme_set_chunk_sectors() 2010 iob = nvme_lba_to_sect(ns, le16_to_cpu(id->noiob)); in nvme_set_chunk_sectors() 2012 if (!iob) in nvme_set_chunk_sectors() 2015 if (!is_power_of_2(iob)) { in nvme_set_chunk_sectors() 2018 ns->disk->disk_name, iob); in nvme_set_chunk_sectors() 2029 blk_queue_chunk_sectors(ns->queue, iob); in nvme_set_chunk_sectors()
|
D | rdma.c | 2094 static int nvme_rdma_poll(struct blk_mq_hw_ctx *hctx, struct io_comp_batch *iob) in nvme_rdma_poll() argument
|
D | tcp.c | 2474 static int nvme_tcp_poll(struct blk_mq_hw_ctx *hctx, struct io_comp_batch *iob) in nvme_tcp_poll() argument
|
/drivers/block/ |
D | virtio_blk.c | 840 static void virtblk_complete_batch(struct io_comp_batch *iob) in virtblk_complete_batch() argument 844 rq_list_for_each(&iob->req_list, req) { in virtblk_complete_batch() 848 blk_mq_end_request_batch(iob); in virtblk_complete_batch() 851 static int virtblk_poll(struct blk_mq_hw_ctx *hctx, struct io_comp_batch *iob) in virtblk_poll() argument 866 if (!blk_mq_add_to_batch(req, iob, vbr->status, in virtblk_poll()
|
/drivers/tty/serial/ |
D | dz.c | 553 iob(); in dz_reset() 822 iob(); in dz_console_putchar() 834 iob(); in dz_console_putchar()
|
/drivers/staging/octeon/ |
D | octeon-stubs.h | 1009 uint64_t iob:1; member 1042 uint64_t iob:1; member 1077 uint64_t iob:1; member 1111 uint64_t iob:1; member
|
/drivers/net/ethernet/amd/ |
D | declance.c | 299 iob(); in writereg() 384 iob(); in cp_to_buf() 524 iob(); in lance_init_ring()
|
/drivers/md/ |
D | dm.c | 1810 static bool dm_poll_dm_io(struct dm_io *io, struct io_comp_batch *iob, in dm_poll_dm_io() argument 1817 bio_poll(&io->tio.clone, iob, flags); in dm_poll_dm_io() 1823 static int dm_poll_bio(struct bio *bio, struct io_comp_batch *iob, in dm_poll_bio() argument 1850 if (dm_poll_dm_io(curr, iob, flags)) { in dm_poll_bio()
|
/drivers/block/null_blk/ |
D | main.c | 1595 static int null_poll(struct blk_mq_hw_ctx *hctx, struct io_comp_batch *iob) in null_poll() argument 1617 if (!blk_mq_add_to_batch(req, iob, (__force int) cmd->error, in null_poll()
|
/drivers/block/rnbd/ |
D | rnbd-clt.c | 1158 static int rnbd_rdma_poll(struct blk_mq_hw_ctx *hctx, struct io_comp_batch *iob) in rnbd_rdma_poll() argument
|
/drivers/scsi/ |
D | scsi_lib.c | 1853 static int scsi_mq_poll(struct blk_mq_hw_ctx *hctx, struct io_comp_batch *iob) in scsi_mq_poll() argument
|