Home
last modified time | relevance | path

Searched full:queues (Results 1 – 25 of 2706) sorted by relevance

12345678910>>...109

/kernel/linux/linux-4.19/Documentation/networking/
Dhinic.txt52 Asynchronous Event Queues(AEQs) - The event queues for receiving messages from
66 Completion Event Queues(CEQs) - The completion Event Queues that describe IO
69 Work Queues(WQ) - Contain the memory and operations for use by CMD queues and
74 Command Queues(CMDQ) - The queues for sending commands for IO management and is
79 Queue Pairs(QPs) - The HW Receive and Send queues for Receiving and Transmitting
101 Tx Queues - Logical Tx Queues that use the HW Send Queues for transmit.
105 Rx Queues - Logical Rx Queues that use the HW Receive Queues for receive.
109 hinic_dev - de/constructs the Logical Tx and Rx Queues.
Dscaling.txt23 Contemporary NICs support multiple receive and transmit descriptor queues
25 queues to distribute processing among CPUs. The NIC distributes packets by
43 Some advanced NICs allow steering packets to queues based on
51 module parameter for specifying the number of hardware queues to
54 for each CPU if the device supports enough queues, or otherwise at least
60 default mapping is to distribute the queues evenly in the table, but the
63 indirection table could be done to give different queues different
72 of queues to IRQs can be determined from /proc/interrupts. By default,
85 is to allocate as many queues as there are CPUs in the system (or the
87 is likely the one with the smallest number of receive queues where no
[all …]
/kernel/linux/linux-5.10/Documentation/networking/device_drivers/ethernet/huawei/
Dhinic.rst55 Asynchronous Event Queues(AEQs) - The event queues for receiving messages from
69 Completion Event Queues(CEQs) - The completion Event Queues that describe IO
72 Work Queues(WQ) - Contain the memory and operations for use by CMD queues and
77 Command Queues(CMDQ) - The queues for sending commands for IO management and is
82 Queue Pairs(QPs) - The HW Receive and Send queues for Receiving and Transmitting
104 Tx Queues - Logical Tx Queues that use the HW Send Queues for transmit.
108 Rx Queues - Logical Rx Queues that use the HW Receive Queues for receive.
112 hinic_dev - de/constructs the Logical Tx and Rx Queues.
/kernel/linux/linux-5.10/net/sched/
Dsch_multiq.c25 struct Qdisc **queues; member
54 return q->queues[0]; in multiq_classify()
56 return q->queues[band]; in multiq_classify()
105 qdisc = q->queues[q->curband]; in multiq_dequeue()
137 qdisc = q->queues[curband]; in multiq_peek()
154 qdisc_reset(q->queues[band]); in multiq_reset()
167 qdisc_put(q->queues[band]); in multiq_destroy()
169 kfree(q->queues); in multiq_destroy()
197 if (q->queues[i] != &noop_qdisc) { in multiq_tune()
198 struct Qdisc *child = q->queues[i]; in multiq_tune()
[all …]
Dsch_prio.c26 struct Qdisc *queues[TCQ_PRIO_BANDS]; member
57 return q->queues[q->prio2band[band & TC_PRIO_MAX]]; in prio_classify()
63 return q->queues[q->prio2band[0]]; in prio_classify()
65 return q->queues[band]; in prio_classify()
103 struct Qdisc *qdisc = q->queues[prio]; in prio_peek()
117 struct Qdisc *qdisc = q->queues[prio]; in prio_dequeue()
137 qdisc_reset(q->queues[prio]); in prio_reset()
175 qdisc_put(q->queues[prio]); in prio_destroy()
182 struct Qdisc *queues[TCQ_PRIO_BANDS]; in prio_tune() local
200 queues[i] = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, in prio_tune()
[all …]
/kernel/linux/linux-4.19/Documentation/ABI/testing/
Dsysfs-class-net-queues1 What: /sys/class/<iface>/queues/rx-<queue>/rps_cpus
11 What: /sys/class/<iface>/queues/rx-<queue>/rps_flow_cnt
19 What: /sys/class/<iface>/queues/tx-<queue>/tx_timeout
27 What: /sys/class/<iface>/queues/tx-<queue>/tx_maxrate
35 What: /sys/class/<iface>/queues/tx-<queue>/xps_cpus
45 What: /sys/class/<iface>/queues/tx-<queue>/xps_rxqs
56 What: /sys/class/<iface>/queues/tx-<queue>/byte_queue_limits/hold_time
65 What: /sys/class/<iface>/queues/tx-<queue>/byte_queue_limits/inflight
73 What: /sys/class/<iface>/queues/tx-<queue>/byte_queue_limits/limit
82 What: /sys/class/<iface>/queues/tx-<queue>/byte_queue_limits/limit_max
[all …]
/kernel/linux/linux-5.10/Documentation/ABI/testing/
Dsysfs-class-net-queues1 What: /sys/class/<iface>/queues/rx-<queue>/rps_cpus
11 What: /sys/class/<iface>/queues/rx-<queue>/rps_flow_cnt
19 What: /sys/class/<iface>/queues/tx-<queue>/tx_timeout
27 What: /sys/class/<iface>/queues/tx-<queue>/tx_maxrate
35 What: /sys/class/<iface>/queues/tx-<queue>/xps_cpus
45 What: /sys/class/<iface>/queues/tx-<queue>/xps_rxqs
56 What: /sys/class/<iface>/queues/tx-<queue>/byte_queue_limits/hold_time
65 What: /sys/class/<iface>/queues/tx-<queue>/byte_queue_limits/inflight
73 What: /sys/class/<iface>/queues/tx-<queue>/byte_queue_limits/limit
82 What: /sys/class/<iface>/queues/tx-<queue>/byte_queue_limits/limit_max
[all …]
/kernel/linux/linux-4.19/net/sched/
Dsch_multiq.c36 struct Qdisc **queues; member
65 return q->queues[0]; in multiq_classify()
67 return q->queues[band]; in multiq_classify()
116 qdisc = q->queues[q->curband]; in multiq_dequeue()
148 qdisc = q->queues[curband]; in multiq_peek()
165 qdisc_reset(q->queues[band]); in multiq_reset()
178 qdisc_destroy(q->queues[band]); in multiq_destroy()
180 kfree(q->queues); in multiq_destroy()
202 if (q->queues[i] != &noop_qdisc) { in multiq_tune()
203 struct Qdisc *child = q->queues[i]; in multiq_tune()
[all …]
Dsch_prio.c30 struct Qdisc *queues[TCQ_PRIO_BANDS]; member
61 return q->queues[q->prio2band[band & TC_PRIO_MAX]]; in prio_classify()
67 return q->queues[q->prio2band[0]]; in prio_classify()
69 return q->queues[band]; in prio_classify()
106 struct Qdisc *qdisc = q->queues[prio]; in prio_peek()
120 struct Qdisc *qdisc = q->queues[prio]; in prio_dequeue()
140 qdisc_reset(q->queues[prio]); in prio_reset()
178 qdisc_destroy(q->queues[prio]); in prio_destroy()
185 struct Qdisc *queues[TCQ_PRIO_BANDS]; in prio_tune() local
203 queues[i] = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, in prio_tune()
[all …]
/kernel/linux/linux-4.19/Documentation/devicetree/bindings/soc/ti/
Dkeystone-navigator-qmss.txt9 management of the packet queues. Packets are queued/de-queued by writing or
32 -- managed-queues : the actual queues managed by each queue manager
33 instance, specified as <"base queue #" "# of queues">.
51 - qpend : pool of qpend(interruptible) queues
52 - general-purpose : pool of general queues, primarily used
53 as free descriptor queues or the
54 transmit DMA queues.
55 - accumulator : pool of queues on PDSP accumulator channel
57 -- qrange : number of queues to use per queue range, specified as
58 <"base queue #" "# of queues">.
[all …]
/kernel/linux/linux-5.10/Documentation/devicetree/bindings/soc/ti/
Dkeystone-navigator-qmss.txt9 management of the packet queues. Packets are queued/de-queued by writing or
32 -- managed-queues : the actual queues managed by each queue manager
33 instance, specified as <"base queue #" "# of queues">.
51 - qpend : pool of qpend(interruptible) queues
52 - general-purpose : pool of general queues, primarily used
53 as free descriptor queues or the
54 transmit DMA queues.
55 - accumulator : pool of queues on PDSP accumulator channel
57 -- qrange : number of queues to use per queue range, specified as
58 <"base queue #" "# of queues">.
[all …]
/kernel/linux/linux-4.19/Documentation/block/
Dcfq-iosched.txt17 This specifies how long CFQ should idle for next request on certain cfq queues
22 queues/service trees. This can be very helpful on highly seeky media like
26 Setting slice_idle to 0 will remove all the idling on queues/service tree
76 queues in the group but happens overall on the group and thus still keeps the
78 Not idling on individual queues in the group will dispatch requests from
79 multiple queues in the group at the same time and achieve higher throughput
123 When a queue is selected for execution, the queues IO requests are only
208 it would be better to dispatch multiple requests from multiple cfq queues in
226 dispatch requests from other cfq queues even if requests are pending there.
235 CFQ has following service trees and various queues are put on these trees.
[all …]
/kernel/linux/linux-5.10/Documentation/block/
Dblk-mq.rst37 spawns multiple queues with individual entry points local to the CPU, removing
49 blk-mq has two group of queues: software staging queues and hardware dispatch
50 queues. When the request arrives at the block layer, it will try the shortest
56 Then, after the requests are processed by software queues, they will be placed
62 Software staging queues
65 The block IO subsystem adds requests in the software staging queues
71 the number of queues is defined by a per-CPU or per-node basis.
93 requests from different queues, otherwise there would be cache trashing and a
99 queue (a.k.a. run the hardware queue), the software queues mapped to that
102 Hardware dispatch queues
[all …]
/kernel/linux/linux-5.10/Documentation/networking/
Dscaling.rst27 Contemporary NICs support multiple receive and transmit descriptor queues
29 queues to distribute processing among CPUs. The NIC distributes packets by
47 Some advanced NICs allow steering packets to queues based on
57 module parameter for specifying the number of hardware queues to
60 for each CPU if the device supports enough queues, or otherwise at least
66 default mapping is to distribute the queues evenly in the table, but the
69 indirection table could be done to give different queues different
80 of queues to IRQs can be determined from /proc/interrupts. By default,
95 is to allocate as many queues as there are CPUs in the system (or the
97 is likely the one with the smallest number of receive queues where no
[all …]
/kernel/linux/linux-4.19/Documentation/arm/keystone/
Dknav-qmss.txt13 management of the packet queues. Packets are queued/de-queued by writing or
22 knav qmss driver provides a set of APIs to drivers to open/close qmss queues,
23 allocate descriptor pools, map the descriptors, push/pop to queues etc. For
29 Accumulator QMSS queues using PDSP firmware
32 queue or multiple contiguous queues. drivers/soc/ti/knav_qmss_acc.c is the
35 1 or 32 queues per channel. More description on the firmware is available in
52 Use of accumulated queues requires the firmware image to be present in the
53 file system. The driver doesn't acc queues to the supported queue range if
/kernel/linux/linux-5.10/drivers/staging/wfx/
Dqueue.c234 struct wfx_queue *queues[IEEE80211_NUM_ACS * ARRAY_SIZE(wdev->vif)]; in wfx_tx_queues_get_skb() local
240 // sort the queues in wfx_tx_queues_get_skb()
244 WARN_ON(num_queues >= ARRAY_SIZE(queues)); in wfx_tx_queues_get_skb()
245 queues[num_queues] = &wvif->tx_queue[i]; in wfx_tx_queues_get_skb()
247 if (wfx_tx_queue_get_weight(queues[j]) < in wfx_tx_queues_get_skb()
248 wfx_tx_queue_get_weight(queues[j - 1])) in wfx_tx_queues_get_skb()
249 swap(queues[j - 1], queues[j]); in wfx_tx_queues_get_skb()
259 skb = skb_dequeue(&queues[i]->cab); in wfx_tx_queues_get_skb()
267 WARN_ON(queues[i] != in wfx_tx_queues_get_skb()
269 atomic_inc(&queues[i]->pending_frames); in wfx_tx_queues_get_skb()
[all …]
/kernel/linux/linux-5.10/Documentation/arm/keystone/
Dknav-qmss.rst15 management of the packet queues. Packets are queued/de-queued by writing or
24 knav qmss driver provides a set of APIs to drivers to open/close qmss queues,
25 allocate descriptor pools, map the descriptors, push/pop to queues etc. For
31 Accumulator QMSS queues using PDSP firmware
34 queue or multiple contiguous queues. drivers/soc/ti/knav_qmss_acc.c is the
37 1 or 32 queues per channel. More description on the firmware is available in
56 Use of accumulated queues requires the firmware image to be present in the
57 file system. The driver doesn't acc queues to the supported queue range if
/kernel/linux/linux-5.10/drivers/nvme/target/
Dloop.c30 struct nvme_loop_queue *queues; member
71 return queue - queue->ctrl->queues; in nvme_loop_queue_idx()
176 struct nvme_loop_queue *queue = &ctrl->queues[0]; in nvme_loop_submit_async_event()
198 iod->queue = &ctrl->queues[queue_idx]; in nvme_loop_init_iod()
220 struct nvme_loop_queue *queue = &ctrl->queues[hctx_idx + 1]; in nvme_loop_init_hctx()
240 struct nvme_loop_queue *queue = &ctrl->queues[0]; in nvme_loop_init_admin_hctx()
264 if (!test_and_clear_bit(NVME_LOOP_Q_LIVE, &ctrl->queues[0].flags)) in nvme_loop_destroy_admin_queue()
266 nvmet_sq_destroy(&ctrl->queues[0].nvme_sq); in nvme_loop_destroy_admin_queue()
287 kfree(ctrl->queues); in nvme_loop_free_ctrl()
298 clear_bit(NVME_LOOP_Q_LIVE, &ctrl->queues[i].flags); in nvme_loop_destroy_io_queues()
[all …]
/kernel/linux/linux-5.10/drivers/scsi/snic/
Dvnic_resource.h27 RES_TYPE_WQ, /* Work queues */
28 RES_TYPE_RQ, /* Receive queues */
29 RES_TYPE_CQ, /* Completion queues */
45 RES_TYPE_MQ_WQ, /* MQ Work queues */
46 RES_TYPE_MQ_RQ, /* MQ Receive queues */
47 RES_TYPE_MQ_CQ, /* MQ Completion queues */
/kernel/linux/linux-5.10/drivers/scsi/fnic/
Dvnic_resource.h27 RES_TYPE_WQ, /* Work queues */
28 RES_TYPE_RQ, /* Receive queues */
29 RES_TYPE_CQ, /* Completion queues */
45 RES_TYPE_MQ_WQ, /* MQ Work queues */
46 RES_TYPE_MQ_RQ, /* MQ Receive queues */
47 RES_TYPE_MQ_CQ, /* MQ Completion queues */
/kernel/linux/linux-4.19/drivers/scsi/snic/
Dvnic_resource.h27 RES_TYPE_WQ, /* Work queues */
28 RES_TYPE_RQ, /* Receive queues */
29 RES_TYPE_CQ, /* Completion queues */
45 RES_TYPE_MQ_WQ, /* MQ Work queues */
46 RES_TYPE_MQ_RQ, /* MQ Receive queues */
47 RES_TYPE_MQ_CQ, /* MQ Completion queues */
/kernel/linux/linux-4.19/Documentation/devicetree/bindings/net/
Dstmmac.txt77 - Multiple RX Queues parameters: below the list of all the parameters to
78 configure the multiple RX queues:
79 - snps,rx-queues-to-use: number of RX queues to be used in the driver
95 - Multiple TX Queues parameters: below the list of all the parameters to
96 configure the multiple TX queues:
97 - snps,tx-queues-to-use: number of TX queues to be used in the driver
124 mtl_rx_setup: rx-queues-config {
125 snps,rx-queues-to-use = <1>;
134 mtl_tx_setup: tx-queues-config {
135 snps,tx-queues-to-use = <2>;
/kernel/linux/linux-4.19/drivers/nvme/target/
Dloop.c38 struct nvme_loop_queue *queues; member
80 return queue - queue->ctrl->queues; in nvme_loop_queue_idx()
197 struct nvme_loop_queue *queue = &ctrl->queues[0]; in nvme_loop_submit_async_event()
219 iod->queue = &ctrl->queues[queue_idx]; in nvme_loop_init_iod()
239 struct nvme_loop_queue *queue = &ctrl->queues[hctx_idx + 1]; in nvme_loop_init_hctx()
251 struct nvme_loop_queue *queue = &ctrl->queues[0]; in nvme_loop_init_admin_hctx()
277 clear_bit(NVME_LOOP_Q_LIVE, &ctrl->queues[0].flags); in nvme_loop_destroy_admin_queue()
278 nvmet_sq_destroy(&ctrl->queues[0].nvme_sq); in nvme_loop_destroy_admin_queue()
298 kfree(ctrl->queues); in nvme_loop_free_ctrl()
309 clear_bit(NVME_LOOP_Q_LIVE, &ctrl->queues[i].flags); in nvme_loop_destroy_io_queues()
[all …]
/kernel/linux/linux-5.10/drivers/scsi/aacraid/
Dcomminit.c237 * Fill in addresses of the Comm Area Headers and Queues in aac_alloc_comm()
373 struct aac_entry * queues; in aac_comm_init() local
375 struct aac_queue_block * comm = dev->queues; in aac_comm_init()
394 queues = (struct aac_entry *)(((ulong)headers) + hdrsize); in aac_comm_init()
397 comm->queue[HostNormCmdQueue].base = queues; in aac_comm_init()
399 queues += HOST_NORM_CMD_ENTRIES; in aac_comm_init()
403 comm->queue[HostHighCmdQueue].base = queues; in aac_comm_init()
406 queues += HOST_HIGH_CMD_ENTRIES; in aac_comm_init()
410 comm->queue[AdapNormCmdQueue].base = queues; in aac_comm_init()
413 queues += ADAP_NORM_CMD_ENTRIES; in aac_comm_init()
[all …]
/kernel/linux/linux-4.19/tools/perf/util/
Darm-spe.c31 struct auxtrace_queues queues; member
123 err = auxtrace_queues__add_event(&spe->queues, session, event, in arm_spe_process_auxtrace_event()
159 struct auxtrace_queues *queues = &spe->queues; in arm_spe_free_events() local
162 for (i = 0; i < queues->nr_queues; i++) { in arm_spe_free_events()
163 arm_spe_free_queue(queues->queue_array[i].priv); in arm_spe_free_events()
164 queues->queue_array[i].priv = NULL; in arm_spe_free_events()
166 auxtrace_queues__free(queues); in arm_spe_free_events()
208 err = auxtrace_queues__init(&spe->queues); in arm_spe_process_auxtrace_info()

12345678910>>...109