Home
last modified time | relevance | path

Searched refs:mq (Results 1 – 25 of 26) sorted by relevance

12

/drivers/md/
Ddm-cache-policy-smq.c840 static struct entry *writeback_sentinel(struct smq_policy *mq, unsigned level) in writeback_sentinel() argument
842 return get_sentinel(&mq->writeback_sentinel_alloc, level, mq->current_writeback_sentinels); in writeback_sentinel()
845 static struct entry *demote_sentinel(struct smq_policy *mq, unsigned level) in demote_sentinel() argument
847 return get_sentinel(&mq->demote_sentinel_alloc, level, mq->current_demote_sentinels); in demote_sentinel()
850 static void __update_writeback_sentinels(struct smq_policy *mq) in __update_writeback_sentinels() argument
853 struct queue *q = &mq->dirty; in __update_writeback_sentinels()
857 sentinel = writeback_sentinel(mq, level); in __update_writeback_sentinels()
863 static void __update_demote_sentinels(struct smq_policy *mq) in __update_demote_sentinels() argument
866 struct queue *q = &mq->clean; in __update_demote_sentinels()
870 sentinel = demote_sentinel(mq, level); in __update_demote_sentinels()
[all …]
DKconfig217 bool "request-based DM: use blk-mq I/O path by default"
220 This option enables the blk-mq based I/O path for request-based
/drivers/mmc/card/
Dqueue.c34 struct mmc_queue *mq = q->queuedata; in mmc_prep_request() local
45 if (mq && (mmc_card_removed(mq->card) || mmc_access_rpmb(mq))) in mmc_prep_request()
55 struct mmc_queue *mq = d; in mmc_queue_thread() local
56 struct request_queue *q = mq->queue; in mmc_queue_thread()
65 down(&mq->thread_sem); in mmc_queue_thread()
72 mq->mqrq_cur->req = req; in mmc_queue_thread()
75 if (req || mq->mqrq_prev->req) { in mmc_queue_thread()
79 mmc_blk_issue_rq(mq, req); in mmc_queue_thread()
81 if (mq->flags & MMC_QUEUE_NEW_REQUEST) { in mmc_queue_thread()
82 mq->flags &= ~MMC_QUEUE_NEW_REQUEST; in mmc_queue_thread()
[all …]
Dblock.c414 struct mmc_queue *mq = &md->queue; in cache_size_show() local
415 int cache_size = atomic_read(&mq->cache_size); in cache_size_show()
421 int speed = atomic_read(&mq->max_write_speed); in cache_size_show()
427 jiffies - mq->cache_jiffies, speed); in cache_size_show()
428 long used = atomic_long_read(&mq->cache_used); in cache_size_show()
466 static long mmc_blk_cache_used(struct mmc_queue *mq, unsigned long waitfor) in mmc_blk_cache_used() argument
469 int speed = atomic_read(&mq->max_write_speed); in mmc_blk_cache_used()
473 waitfor - mq->cache_jiffies, speed); in mmc_blk_cache_used()
474 used = atomic_long_read(&mq->cache_used); in mmc_blk_cache_used()
482 atomic_long_set(&mq->cache_used, used); in mmc_blk_cache_used()
[all …]
Dblock.h1 int mmc_blk_issue_rq(struct mmc_queue *mq, struct request *req);
/drivers/scsi/arm/
Dmsgqueue.c27 struct msgqueue_entry *mq; in mqe_alloc() local
29 if ((mq = msgq->free) != NULL) in mqe_alloc()
30 msgq->free = mq->next; in mqe_alloc()
32 return mq; in mqe_alloc()
41 static void mqe_free(MsgQueue_t *msgq, struct msgqueue_entry *mq) in mqe_free() argument
43 if (mq) { in mqe_free()
44 mq->next = msgq->free; in mqe_free()
45 msgq->free = mq; in mqe_free()
85 struct msgqueue_entry *mq = msgq->qe; in msgqueue_msglength() local
88 for (mq = msgq->qe; mq; mq = mq->next) in msgqueue_msglength()
[all …]
/drivers/sh/maple/
Dmaple.c122 void (*callback) (struct mapleq *mq), in maple_getcond_callback()
140 struct mapleq *mq; in maple_release_device() local
143 mq = mdev->mq; in maple_release_device()
144 kmem_cache_free(maple_queue_cache, mq->recvbuf); in maple_release_device()
145 kfree(mq); in maple_release_device()
172 mdev->mq->command = command; in maple_add_packet()
173 mdev->mq->length = length; in maple_add_packet()
176 mdev->mq->sendbuf = sendbuf; in maple_add_packet()
179 list_add_tail(&mdev->mq->list, &maple_waitq); in maple_add_packet()
188 struct mapleq *mq; in maple_allocq() local
[all …]
/drivers/misc/sgi-xp/
Dxpc_uv.c109 xpc_get_gru_mq_irq_uv(struct xpc_gru_mq_uv *mq, int cpu, char *irq_name) in xpc_get_gru_mq_irq_uv() argument
111 int mmr_pnode = uv_blade_to_pnode(mq->mmr_blade); in xpc_get_gru_mq_irq_uv()
114 mq->irq = uv_setup_irq(irq_name, cpu, mq->mmr_blade, mq->mmr_offset, in xpc_get_gru_mq_irq_uv()
116 if (mq->irq < 0) in xpc_get_gru_mq_irq_uv()
117 return mq->irq; in xpc_get_gru_mq_irq_uv()
119 mq->mmr_value = uv_read_global_mmr64(mmr_pnode, mq->mmr_offset); in xpc_get_gru_mq_irq_uv()
123 mq->irq = SGI_XPC_ACTIVATE; in xpc_get_gru_mq_irq_uv()
125 mq->irq = SGI_XPC_NOTIFY; in xpc_get_gru_mq_irq_uv()
129 mq->mmr_value = (unsigned long)cpu_physical_id(cpu) << 32 | mq->irq; in xpc_get_gru_mq_irq_uv()
130 uv_write_global_mmr64(mmr_pnode, mq->mmr_offset, mq->mmr_value); in xpc_get_gru_mq_irq_uv()
[all …]
/drivers/misc/sgi-gru/
Dgrukservices.c146 #define HSTATUS(mq, h) ((mq) + offsetof(struct message_queue, hstatus[h])) argument
560 struct message_queue *mq = p; in gru_create_message_queue() local
564 memset(mq, 0, bytes); in gru_create_message_queue()
565 mq->start = &mq->data; in gru_create_message_queue()
566 mq->start2 = &mq->data + (qlines / 2 - 1) * GRU_CACHE_LINE_BYTES; in gru_create_message_queue()
567 mq->next = &mq->data; in gru_create_message_queue()
568 mq->limit = &mq->data + (qlines - 2) * GRU_CACHE_LINE_BYTES; in gru_create_message_queue()
569 mq->qlines = qlines; in gru_create_message_queue()
570 mq->hstatus[0] = 0; in gru_create_message_queue()
571 mq->hstatus[1] = 1; in gru_create_message_queue()
[all …]
Dgrukservices.h45 void *mq; /* message queue vaddress */ member
/drivers/isdn/capi/
Dcapilib.c47 struct capilib_msgidqueue *mq; in mq_enqueue() local
48 if ((mq = np->msgidfree) == NULL) in mq_enqueue()
50 np->msgidfree = mq->next; in mq_enqueue()
51 mq->msgid = msgid; in mq_enqueue()
52 mq->next = NULL; in mq_enqueue()
54 np->msgidlast->next = mq; in mq_enqueue()
55 np->msgidlast = mq; in mq_enqueue()
57 np->msgidqueue = mq; in mq_enqueue()
67 struct capilib_msgidqueue *mq = *pp; in mq_dequeue() local
68 *pp = mq->next; in mq_dequeue()
[all …]
/drivers/mailbox/
Domap-mailbox.c261 struct omap_mbox_queue *mq = in mbox_rx_work() local
266 while (kfifo_len(&mq->fifo) >= sizeof(msg)) { in mbox_rx_work()
267 len = kfifo_out(&mq->fifo, (unsigned char *)&msg, sizeof(msg)); in mbox_rx_work()
270 mbox_chan_received_data(mq->mbox->chan, (void *)msg); in mbox_rx_work()
271 spin_lock_irq(&mq->lock); in mbox_rx_work()
272 if (mq->full) { in mbox_rx_work()
273 mq->full = false; in mbox_rx_work()
274 _omap_mbox_enable_irq(mq->mbox, IRQ_RX); in mbox_rx_work()
276 spin_unlock_irq(&mq->lock); in mbox_rx_work()
292 struct omap_mbox_queue *mq = mbox->rxq; in __mbox_rx_interrupt() local
[all …]
/drivers/mtd/maps/
Dvmu-flash.c91 static void vmu_blockread(struct mapleq *mq) in vmu_blockread() argument
96 mdev = mq->dev; in vmu_blockread()
103 memcpy(card->blockread, mq->recvbuf->buf + 12, in vmu_blockread()
193 list_del_init(&(mdev->mq->list)); in maple_vmu_read_block()
194 kfree(mdev->mq->sendbuf); in maple_vmu_read_block()
195 mdev->mq->sendbuf = NULL; in maple_vmu_read_block()
285 kfree(mdev->mq->sendbuf); in maple_vmu_write_block()
286 mdev->mq->sendbuf = NULL; in maple_vmu_write_block()
287 list_del_init(&(mdev->mq->list)); in maple_vmu_write_block()
501 static void vmu_queryblocks(struct mapleq *mq) in vmu_queryblocks() argument
[all …]
/drivers/net/wireless/intel/iwlwifi/dvm/
Dmain.c2061 int mq = priv->queue_to_mac80211[queue]; in iwl_stop_sw_queue() local
2063 if (WARN_ON_ONCE(mq == IWL_INVALID_MAC80211_QUEUE)) in iwl_stop_sw_queue()
2066 if (atomic_inc_return(&priv->queue_stop_count[mq]) > 1) { in iwl_stop_sw_queue()
2069 queue, mq); in iwl_stop_sw_queue()
2073 set_bit(mq, &priv->transport_queue_stop); in iwl_stop_sw_queue()
2074 ieee80211_stop_queue(priv->hw, mq); in iwl_stop_sw_queue()
2080 int mq = priv->queue_to_mac80211[queue]; in iwl_wake_sw_queue() local
2082 if (WARN_ON_ONCE(mq == IWL_INVALID_MAC80211_QUEUE)) in iwl_wake_sw_queue()
2085 if (atomic_dec_return(&priv->queue_stop_count[mq]) > 0) { in iwl_wake_sw_queue()
2088 queue, mq); in iwl_wake_sw_queue()
[all …]
Dtx.c484 static int iwlagn_alloc_agg_txq(struct iwl_priv *priv, int mq) in iwlagn_alloc_agg_txq() argument
491 priv->queue_to_mac80211[q] = mq; in iwlagn_alloc_agg_txq()
/drivers/input/mouse/
Dmaplemouse.c26 static void dc_mouse_callback(struct mapleq *mq) in dc_mouse_callback() argument
29 struct maple_device *mapledev = mq->dev; in dc_mouse_callback()
32 unsigned char *res = mq->recvbuf->buf; in dc_mouse_callback()
/drivers/input/joystick/
Dmaplecontrol.c26 static void dc_pad_callback(struct mapleq *mq) in dc_pad_callback() argument
29 struct maple_device *mapledev = mq->dev; in dc_pad_callback()
32 unsigned char *res = mq->recvbuf->buf; in dc_pad_callback()
/drivers/infiniband/hw/ocrdma/
Docrdma_hw.c123 (dev->mq.cq.va + (dev->mq.cq.tail * sizeof(struct ocrdma_mcqe))); in ocrdma_get_mcqe()
132 dev->mq.cq.tail = (dev->mq.cq.tail + 1) & (OCRDMA_MQ_CQ_LEN - 1); in ocrdma_mcq_inc_tail()
137 return dev->mq.sq.va + (dev->mq.sq.head * sizeof(struct ocrdma_mqe)); in ocrdma_get_mqe()
142 dev->mq.sq.head = (dev->mq.sq.head + 1) & (OCRDMA_MQ_LEN - 1); in ocrdma_mq_inc_head()
147 return dev->mq.sq.va + (dev->mqe_ctx.tag * sizeof(struct ocrdma_mqe)); in ocrdma_get_mqe_rsp()
321 val |= dev->mq.sq.id & OCRDMA_MQ_ID_MASK; in ocrdma_ring_mq_db()
562 struct ocrdma_queue_info *mq, in ocrdma_mbx_create_mq() argument
571 num_pages = PAGES_4K_SPANNED(mq->va, mq->size); in ocrdma_mbx_create_mq()
586 cmd->async_cqid_ringsize |= (ocrdma_encoded_q_len(mq->len) << in ocrdma_mbx_create_mq()
591 ocrdma_build_q_pages(pa, num_pages, mq->dma, PAGE_SIZE_4K); in ocrdma_mbx_create_mq()
[all …]
Docrdma.h279 struct ocrdma_mq mq; member
/drivers/net/wireless/intel/iwlwifi/mvm/
Dops.c980 void iwl_mvm_stop_mac_queues(struct iwl_mvm *mvm, unsigned long mq) in iwl_mvm_stop_mac_queues() argument
984 if (WARN_ON_ONCE(!mq)) in iwl_mvm_stop_mac_queues()
987 for_each_set_bit(q, &mq, IEEE80211_MAX_QUEUES) { in iwl_mvm_stop_mac_queues()
1013 unsigned long mq; in iwl_mvm_stop_sw_queue() local
1016 mq = mvm->queue_info[hw_queue].hw_queue_to_mac80211; in iwl_mvm_stop_sw_queue()
1019 iwl_mvm_stop_mac_queues(mvm, mq); in iwl_mvm_stop_sw_queue()
1022 void iwl_mvm_start_mac_queues(struct iwl_mvm *mvm, unsigned long mq) in iwl_mvm_start_mac_queues() argument
1026 if (WARN_ON_ONCE(!mq)) in iwl_mvm_start_mac_queues()
1029 for_each_set_bit(q, &mq, IEEE80211_MAX_QUEUES) { in iwl_mvm_start_mac_queues()
1043 unsigned long mq; in iwl_mvm_wake_sw_queue() local
[all …]
Dmvm.h1698 void iwl_mvm_start_mac_queues(struct iwl_mvm *mvm, unsigned long mq);
1699 void iwl_mvm_stop_mac_queues(struct iwl_mvm *mvm, unsigned long mq);
Dsta.c542 unsigned long mq; in iwl_mvm_scd_queue_redirect() local
566 mq = mvm->queue_info[queue].hw_queue_to_mac80211; in iwl_mvm_scd_queue_redirect()
574 iwl_mvm_stop_mac_queues(mvm, mq); in iwl_mvm_scd_queue_redirect()
621 iwl_mvm_start_mac_queues(mvm, mq); in iwl_mvm_scd_queue_redirect()
/drivers/input/keyboard/
Dmaple_keyb.c139 static void dc_kbd_callback(struct mapleq *mq) in dc_kbd_callback() argument
141 struct maple_device *mapledev = mq->dev; in dc_kbd_callback()
143 unsigned long *buf = (unsigned long *)(mq->recvbuf->buf); in dc_kbd_callback()
/drivers/scsi/qla2xxx/
Dqla_dbg.c624 struct qla2xxx_mq_chain *mq = ptr; in qla25xx_copy_mq() local
630 mq = ptr; in qla25xx_copy_mq()
631 *last_chain = &mq->type; in qla25xx_copy_mq()
632 mq->type = htonl(DUMP_CHAIN_MQ); in qla25xx_copy_mq()
633 mq->chain_size = htonl(sizeof(struct qla2xxx_mq_chain)); in qla25xx_copy_mq()
637 mq->count = htonl(que_cnt); in qla25xx_copy_mq()
641 mq->qregs[que_idx] = in qla25xx_copy_mq()
643 mq->qregs[que_idx+1] = in qla25xx_copy_mq()
645 mq->qregs[que_idx+2] = in qla25xx_copy_mq()
647 mq->qregs[que_idx+3] = in qla25xx_copy_mq()
/drivers/scsi/lpfc/
Dlpfc_sli.c13403 lpfc_mq_create_fb_init(struct lpfc_hba *phba, struct lpfc_queue *mq, in lpfc_mq_create_fb_init() argument
13417 mq->page_count); in lpfc_mq_create_fb_init()
13421 switch (mq->entry_count) { in lpfc_mq_create_fb_init()
13439 list_for_each_entry(dmabuf, &mq->page_list, list) { in lpfc_mq_create_fb_init()
13469 lpfc_mq_create(struct lpfc_hba *phba, struct lpfc_queue *mq, in lpfc_mq_create() argument
13482 if (!mq || !cq) in lpfc_mq_create()
13499 &mq_create_ext->u.request, mq->page_count); in lpfc_mq_create()
13519 switch (mq->entry_count) { in lpfc_mq_create()
13523 mq->entry_count); in lpfc_mq_create()
13524 if (mq->entry_count < 16) { in lpfc_mq_create()
[all …]

12