/drivers/mmc/core/ |
D | queue.c | 27 static inline bool mmc_cqe_dcmd_busy(struct mmc_queue *mq) in mmc_cqe_dcmd_busy() argument 30 return mq->in_flight[MMC_ISSUE_DCMD]; in mmc_cqe_dcmd_busy() 33 void mmc_cqe_check_busy(struct mmc_queue *mq) in mmc_cqe_check_busy() argument 35 if ((mq->cqe_busy & MMC_CQE_DCMD_BUSY) && !mmc_cqe_dcmd_busy(mq)) in mmc_cqe_check_busy() 36 mq->cqe_busy &= ~MMC_CQE_DCMD_BUSY; in mmc_cqe_check_busy() 38 mq->cqe_busy &= ~MMC_CQE_QUEUE_FULL; in mmc_cqe_check_busy() 62 enum mmc_issue_type mmc_issue_type(struct mmc_queue *mq, struct request *req) in mmc_issue_type() argument 64 struct mmc_host *host = mq->card->host; in mmc_issue_type() 66 if (mq->use_cqe && !host->hsq_enabled) in mmc_issue_type() 75 static void __mmc_cqe_recovery_notifier(struct mmc_queue *mq) in __mmc_cqe_recovery_notifier() argument [all …]
|
D | block.c | 176 struct mmc_queue *mq); 239 struct mmc_queue *mq; in power_ro_lock_store() local 250 mq = &md->queue; in power_ro_lock_store() 253 req = blk_get_request(mq->queue, REQ_OP_DRV_OUT, 0); in power_ro_lock_store() 260 blk_execute_rq(mq->queue, NULL, req, 0); in power_ro_lock_store() 638 struct mmc_queue *mq; in mmc_blk_ioctl_cmd() local 658 mq = &md->queue; in mmc_blk_ioctl_cmd() 659 req = blk_get_request(mq->queue, in mmc_blk_ioctl_cmd() 671 blk_execute_rq(mq->queue, NULL, req, 0); in mmc_blk_ioctl_cmd() 689 struct mmc_queue *mq; in mmc_blk_ioctl_multi_cmd() local [all …]
|
D | queue.h | 106 void mmc_cqe_check_busy(struct mmc_queue *mq); 109 enum mmc_issue_type mmc_issue_type(struct mmc_queue *mq, struct request *req); 111 static inline int mmc_tot_in_flight(struct mmc_queue *mq) in mmc_tot_in_flight() argument 113 return mq->in_flight[MMC_ISSUE_SYNC] + in mmc_tot_in_flight() 114 mq->in_flight[MMC_ISSUE_DCMD] + in mmc_tot_in_flight() 115 mq->in_flight[MMC_ISSUE_ASYNC]; in mmc_tot_in_flight() 118 static inline int mmc_cqe_qcnt(struct mmc_queue *mq) in mmc_cqe_qcnt() argument 120 return mq->in_flight[MMC_ISSUE_DCMD] + in mmc_cqe_qcnt() 121 mq->in_flight[MMC_ISSUE_ASYNC]; in mmc_cqe_qcnt()
|
D | block.h | 8 void mmc_blk_cqe_recovery(struct mmc_queue *mq); 12 enum mmc_issued mmc_blk_mq_issue_rq(struct mmc_queue *mq, struct request *req); 14 void mmc_blk_mq_recovery(struct mmc_queue *mq);
|
/drivers/md/ |
D | dm-cache-policy-smq.c | 873 static struct entry *writeback_sentinel(struct smq_policy *mq, unsigned level) in writeback_sentinel() argument 875 return get_sentinel(&mq->writeback_sentinel_alloc, level, mq->current_writeback_sentinels); in writeback_sentinel() 878 static struct entry *demote_sentinel(struct smq_policy *mq, unsigned level) in demote_sentinel() argument 880 return get_sentinel(&mq->demote_sentinel_alloc, level, mq->current_demote_sentinels); in demote_sentinel() 883 static void __update_writeback_sentinels(struct smq_policy *mq) in __update_writeback_sentinels() argument 886 struct queue *q = &mq->dirty; in __update_writeback_sentinels() 890 sentinel = writeback_sentinel(mq, level); in __update_writeback_sentinels() 896 static void __update_demote_sentinels(struct smq_policy *mq) in __update_demote_sentinels() argument 899 struct queue *q = &mq->clean; in __update_demote_sentinels() 903 sentinel = demote_sentinel(mq, level); in __update_demote_sentinels() [all …]
|
/drivers/scsi/arm/ |
D | msgqueue.c | 24 struct msgqueue_entry *mq; in mqe_alloc() local 26 if ((mq = msgq->free) != NULL) in mqe_alloc() 27 msgq->free = mq->next; in mqe_alloc() 29 return mq; in mqe_alloc() 38 static void mqe_free(MsgQueue_t *msgq, struct msgqueue_entry *mq) in mqe_free() argument 40 if (mq) { in mqe_free() 41 mq->next = msgq->free; in mqe_free() 42 msgq->free = mq; in mqe_free() 82 struct msgqueue_entry *mq = msgq->qe; in msgqueue_msglength() local 85 for (mq = msgq->qe; mq; mq = mq->next) in msgqueue_msglength() [all …]
|
/drivers/sh/maple/ |
D | maple.c | 122 void (*callback) (struct mapleq *mq), in maple_getcond_callback() 140 struct mapleq *mq; in maple_release_device() local 143 mq = mdev->mq; in maple_release_device() 144 kmem_cache_free(maple_queue_cache, mq->recvbuf); in maple_release_device() 145 kfree(mq); in maple_release_device() 172 mdev->mq->command = command; in maple_add_packet() 173 mdev->mq->length = length; in maple_add_packet() 176 mdev->mq->sendbuf = sendbuf; in maple_add_packet() 179 list_add_tail(&mdev->mq->list, &maple_waitq); in maple_add_packet() 188 struct mapleq *mq; in maple_allocq() local [all …]
|
/drivers/misc/sgi-xp/ |
D | xpc_uv.c | 112 xpc_get_gru_mq_irq_uv(struct xpc_gru_mq_uv *mq, int cpu, char *irq_name) in xpc_get_gru_mq_irq_uv() argument 114 int mmr_pnode = uv_blade_to_pnode(mq->mmr_blade); in xpc_get_gru_mq_irq_uv() 117 mq->irq = uv_setup_irq(irq_name, cpu, mq->mmr_blade, mq->mmr_offset, in xpc_get_gru_mq_irq_uv() 119 if (mq->irq < 0) in xpc_get_gru_mq_irq_uv() 120 return mq->irq; in xpc_get_gru_mq_irq_uv() 122 mq->mmr_value = uv_read_global_mmr64(mmr_pnode, mq->mmr_offset); in xpc_get_gru_mq_irq_uv() 126 mq->irq = SGI_XPC_ACTIVATE; in xpc_get_gru_mq_irq_uv() 128 mq->irq = SGI_XPC_NOTIFY; in xpc_get_gru_mq_irq_uv() 132 mq->mmr_value = (unsigned long)cpu_physical_id(cpu) << 32 | mq->irq; in xpc_get_gru_mq_irq_uv() 133 uv_write_global_mmr64(mmr_pnode, mq->mmr_offset, mq->mmr_value); in xpc_get_gru_mq_irq_uv() [all …]
|
/drivers/misc/sgi-gru/ |
D | grukservices.c | 134 #define HSTATUS(mq, h) ((mq) + offsetof(struct message_queue, hstatus[h])) argument 548 struct message_queue *mq = p; in gru_create_message_queue() local 552 memset(mq, 0, bytes); in gru_create_message_queue() 553 mq->start = &mq->data; in gru_create_message_queue() 554 mq->start2 = &mq->data + (qlines / 2 - 1) * GRU_CACHE_LINE_BYTES; in gru_create_message_queue() 555 mq->next = &mq->data; in gru_create_message_queue() 556 mq->limit = &mq->data + (qlines - 2) * GRU_CACHE_LINE_BYTES; in gru_create_message_queue() 557 mq->qlines = qlines; in gru_create_message_queue() 558 mq->hstatus[0] = 0; in gru_create_message_queue() 559 mq->hstatus[1] = 1; in gru_create_message_queue() [all …]
|
D | grukservices.h | 32 void *mq; /* message queue vaddress */ member
|
/drivers/mailbox/ |
D | omap-mailbox.c | 257 struct omap_mbox_queue *mq = in mbox_rx_work() local 263 while (kfifo_len(&mq->fifo) >= sizeof(msg)) { in mbox_rx_work() 264 len = kfifo_out(&mq->fifo, (unsigned char *)&msg, sizeof(msg)); in mbox_rx_work() 268 mbox_chan_received_data(mq->mbox->chan, (void *)data); in mbox_rx_work() 269 spin_lock_irq(&mq->lock); in mbox_rx_work() 270 if (mq->full) { in mbox_rx_work() 271 mq->full = false; in mbox_rx_work() 272 _omap_mbox_enable_irq(mq->mbox, IRQ_RX); in mbox_rx_work() 274 spin_unlock_irq(&mq->lock); in mbox_rx_work() 290 struct omap_mbox_queue *mq = mbox->rxq; in __mbox_rx_interrupt() local [all …]
|
/drivers/net/ethernet/netronome/nfp/abm/ |
D | qdisc.c | 304 struct nfp_qdisc *mq = nfp_abm_qdisc_tree_deref_slot(slot); in nfp_abm_qdisc_clear_mq() local 307 if (mq->type != NFP_QDISC_MQ || mq->netdev != netdev) in nfp_abm_qdisc_clear_mq() 309 for (i = 0; i < mq->num_children; i++) in nfp_abm_qdisc_clear_mq() 310 if (mq->children[i] == qdisc) { in nfp_abm_qdisc_clear_mq() 311 mq->children[i] = NULL; in nfp_abm_qdisc_clear_mq() 791 memset(&qdisc->mq.stats, 0, sizeof(qdisc->mq.stats)); in nfp_abm_mq_stats() 792 memset(&qdisc->mq.prev_stats, 0, sizeof(qdisc->mq.prev_stats)); in nfp_abm_mq_stats() 803 nfp_abm_stats_propagate(&qdisc->mq.stats, in nfp_abm_mq_stats() 805 nfp_abm_stats_propagate(&qdisc->mq.prev_stats, in nfp_abm_mq_stats() 810 nfp_abm_stats_calculate(&qdisc->mq.stats, &qdisc->mq.prev_stats, in nfp_abm_mq_stats()
|
D | main.h | 170 } mq; member
|
/drivers/mtd/maps/ |
D | vmu-flash.c | 89 static void vmu_blockread(struct mapleq *mq) in vmu_blockread() argument 94 mdev = mq->dev; in vmu_blockread() 101 memcpy(card->blockread, mq->recvbuf->buf + 12, in vmu_blockread() 191 list_del_init(&(mdev->mq->list)); in maple_vmu_read_block() 192 kfree(mdev->mq->sendbuf); in maple_vmu_read_block() 193 mdev->mq->sendbuf = NULL; in maple_vmu_read_block() 283 kfree(mdev->mq->sendbuf); in maple_vmu_write_block() 284 mdev->mq->sendbuf = NULL; in maple_vmu_write_block() 285 list_del_init(&(mdev->mq->list)); in maple_vmu_write_block() 499 static void vmu_queryblocks(struct mapleq *mq) in vmu_queryblocks() argument [all …]
|
/drivers/net/wireless/intel/iwlwifi/dvm/ |
D | main.c | 2045 int mq = priv->queue_to_mac80211[queue]; in iwl_stop_sw_queue() local 2047 if (WARN_ON_ONCE(mq == IWL_INVALID_MAC80211_QUEUE)) in iwl_stop_sw_queue() 2050 if (atomic_inc_return(&priv->queue_stop_count[mq]) > 1) { in iwl_stop_sw_queue() 2053 queue, mq); in iwl_stop_sw_queue() 2057 set_bit(mq, &priv->transport_queue_stop); in iwl_stop_sw_queue() 2058 ieee80211_stop_queue(priv->hw, mq); in iwl_stop_sw_queue() 2064 int mq = priv->queue_to_mac80211[queue]; in iwl_wake_sw_queue() local 2066 if (WARN_ON_ONCE(mq == IWL_INVALID_MAC80211_QUEUE)) in iwl_wake_sw_queue() 2069 if (atomic_dec_return(&priv->queue_stop_count[mq]) > 0) { in iwl_wake_sw_queue() 2072 queue, mq); in iwl_wake_sw_queue() [all …]
|
D | tx.c | 465 static int iwlagn_alloc_agg_txq(struct iwl_priv *priv, int mq) in iwlagn_alloc_agg_txq() argument 472 priv->queue_to_mac80211[q] = mq; in iwlagn_alloc_agg_txq()
|
/drivers/input/mouse/ |
D | maplemouse.c | 27 static void dc_mouse_callback(struct mapleq *mq) in dc_mouse_callback() argument 30 struct maple_device *mapledev = mq->dev; in dc_mouse_callback() 33 unsigned char *res = mq->recvbuf->buf; in dc_mouse_callback()
|
/drivers/input/joystick/ |
D | maplecontrol.c | 27 static void dc_pad_callback(struct mapleq *mq) in dc_pad_callback() argument 30 struct maple_device *mapledev = mq->dev; in dc_pad_callback() 33 unsigned char *res = mq->recvbuf->buf; in dc_pad_callback()
|
/drivers/infiniband/hw/ocrdma/ |
D | ocrdma_hw.c | 124 (dev->mq.cq.va + (dev->mq.cq.tail * sizeof(struct ocrdma_mcqe))); in ocrdma_get_mcqe() 133 dev->mq.cq.tail = (dev->mq.cq.tail + 1) & (OCRDMA_MQ_CQ_LEN - 1); in ocrdma_mcq_inc_tail() 138 return dev->mq.sq.va + (dev->mq.sq.head * sizeof(struct ocrdma_mqe)); in ocrdma_get_mqe() 143 dev->mq.sq.head = (dev->mq.sq.head + 1) & (OCRDMA_MQ_LEN - 1); in ocrdma_mq_inc_head() 148 return dev->mq.sq.va + (dev->mqe_ctx.tag * sizeof(struct ocrdma_mqe)); in ocrdma_get_mqe_rsp() 325 val |= dev->mq.sq.id & OCRDMA_MQ_ID_MASK; in ocrdma_ring_mq_db() 565 struct ocrdma_queue_info *mq, in ocrdma_mbx_create_mq() argument 574 num_pages = PAGES_4K_SPANNED(mq->va, mq->size); in ocrdma_mbx_create_mq() 589 cmd->async_cqid_ringsize |= (ocrdma_encoded_q_len(mq->len) << in ocrdma_mbx_create_mq() 594 ocrdma_build_q_pages(pa, num_pages, mq->dma, PAGE_SIZE_4K); in ocrdma_mbx_create_mq() [all …]
|
D | ocrdma.h | 277 struct ocrdma_mq mq; member
|
/drivers/input/keyboard/ |
D | maple_keyb.c | 125 static void dc_kbd_callback(struct mapleq *mq) in dc_kbd_callback() argument 127 struct maple_device *mapledev = mq->dev; in dc_kbd_callback() 129 unsigned long *buf = (unsigned long *)(mq->recvbuf->buf); in dc_kbd_callback()
|
/drivers/scsi/qla2xxx/ |
D | qla_dbg.c | 677 struct qla2xxx_mq_chain *mq = ptr; in qla25xx_copy_mq() local 684 mq = ptr; in qla25xx_copy_mq() 685 *last_chain = &mq->type; in qla25xx_copy_mq() 686 mq->type = htonl(DUMP_CHAIN_MQ); in qla25xx_copy_mq() 687 mq->chain_size = htonl(sizeof(struct qla2xxx_mq_chain)); in qla25xx_copy_mq() 691 mq->count = htonl(que_cnt); in qla25xx_copy_mq() 695 mq->qregs[que_idx] = in qla25xx_copy_mq() 697 mq->qregs[que_idx+1] = in qla25xx_copy_mq() 699 mq->qregs[que_idx+2] = in qla25xx_copy_mq() 701 mq->qregs[que_idx+3] = in qla25xx_copy_mq()
|
/drivers/scsi/lpfc/ |
D | lpfc_sli.c | 15639 lpfc_mq_create_fb_init(struct lpfc_hba *phba, struct lpfc_queue *mq, in lpfc_mq_create_fb_init() argument 15653 mq->page_count); in lpfc_mq_create_fb_init() 15657 switch (mq->entry_count) { in lpfc_mq_create_fb_init() 15675 list_for_each_entry(dmabuf, &mq->page_list, list) { in lpfc_mq_create_fb_init() 15705 lpfc_mq_create(struct lpfc_hba *phba, struct lpfc_queue *mq, in lpfc_mq_create() argument 15718 if (!mq || !cq) in lpfc_mq_create() 15735 &mq_create_ext->u.request, mq->page_count); in lpfc_mq_create() 15755 switch (mq->entry_count) { in lpfc_mq_create() 15759 mq->entry_count); in lpfc_mq_create() 15760 if (mq->entry_count < 16) { in lpfc_mq_create() [all …]
|
/drivers/net/ |
D | virtio_net.c | 174 struct virtio_net_ctrl_mq mq; member 1896 vi->ctrl->mq.virtqueue_pairs = cpu_to_virtio16(vi->vdev, queue_pairs); in _virtnet_set_queues() 1897 sg_init_one(&sg, &vi->ctrl->mq, sizeof(vi->ctrl->mq)); in _virtnet_set_queues()
|