Home
last modified time | relevance | path

Searched refs:queue_num (Results 1 – 25 of 42) sorted by relevance

12

/drivers/net/ethernet/samsung/sxgbe/
Dsxgbe_mtl.c61 static void sxgbe_mtl_set_txfifosize(void __iomem *ioaddr, int queue_num, in sxgbe_mtl_set_txfifosize() argument
68 reg_val = readl(ioaddr + SXGBE_MTL_TXQ_OPMODE_REG(queue_num)); in sxgbe_mtl_set_txfifosize()
70 writel(reg_val, ioaddr + SXGBE_MTL_TXQ_OPMODE_REG(queue_num)); in sxgbe_mtl_set_txfifosize()
73 static void sxgbe_mtl_set_rxfifosize(void __iomem *ioaddr, int queue_num, in sxgbe_mtl_set_rxfifosize() argument
80 reg_val = readl(ioaddr + SXGBE_MTL_RXQ_OPMODE_REG(queue_num)); in sxgbe_mtl_set_rxfifosize()
82 writel(reg_val, ioaddr + SXGBE_MTL_RXQ_OPMODE_REG(queue_num)); in sxgbe_mtl_set_rxfifosize()
85 static void sxgbe_mtl_enable_txqueue(void __iomem *ioaddr, int queue_num) in sxgbe_mtl_enable_txqueue() argument
89 reg_val = readl(ioaddr + SXGBE_MTL_TXQ_OPMODE_REG(queue_num)); in sxgbe_mtl_enable_txqueue()
91 writel(reg_val, ioaddr + SXGBE_MTL_TXQ_OPMODE_REG(queue_num)); in sxgbe_mtl_enable_txqueue()
94 static void sxgbe_mtl_disable_txqueue(void __iomem *ioaddr, int queue_num) in sxgbe_mtl_disable_txqueue() argument
[all …]
Dsxgbe_mtl.h64 void (*mtl_set_txfifosize)(void __iomem *ioaddr, int queue_num,
67 void (*mtl_set_rxfifosize)(void __iomem *ioaddr, int queue_num,
70 void (*mtl_enable_txqueue)(void __iomem *ioaddr, int queue_num);
72 void (*mtl_disable_txqueue)(void __iomem *ioaddr, int queue_num);
74 void (*set_tx_mtl_mode)(void __iomem *ioaddr, int queue_num,
77 void (*set_rx_mtl_mode)(void __iomem *ioaddr, int queue_num,
82 void (*mtl_fc_active)(void __iomem *ioaddr, int queue_num,
85 void (*mtl_fc_deactive)(void __iomem *ioaddr, int queue_num,
88 void (*mtl_fc_enable)(void __iomem *ioaddr, int queue_num);
90 void (*mtl_fep_enable)(void __iomem *ioaddr, int queue_num);
[all …]
Dsxgbe_main.c563 int queue_num, ret; in init_dma_desc_rings() local
569 SXGBE_FOR_EACH_QUEUE(SXGBE_TX_QUEUES, queue_num) { in init_dma_desc_rings()
570 ret = init_tx_ring(priv->device, queue_num, in init_dma_desc_rings()
571 priv->txq[queue_num], tx_rsize); in init_dma_desc_rings()
580 priv->txq[queue_num]->priv_ptr = priv; in init_dma_desc_rings()
584 SXGBE_FOR_EACH_QUEUE(SXGBE_RX_QUEUES, queue_num) { in init_dma_desc_rings()
585 ret = init_rx_ring(netd, queue_num, in init_dma_desc_rings()
586 priv->rxq[queue_num], rx_rsize); in init_dma_desc_rings()
595 priv->rxq[queue_num]->priv_ptr = priv; in init_dma_desc_rings()
603 while (queue_num--) in init_dma_desc_rings()
[all …]
Dsxgbe_common.h194 #define SXGBE_FOR_EACH_QUEUE(max_queues, queue_num) \ argument
195 for (queue_num = 0; queue_num < max_queues; queue_num++)
358 void (*enable_rxqueue)(void __iomem *ioaddr, int queue_num);
359 void (*disable_rxqueue)(void __iomem *ioaddr, int queue_num);
Dsxgbe_core.c166 static void sxgbe_core_enable_rxqueue(void __iomem *ioaddr, int queue_num) in sxgbe_core_enable_rxqueue() argument
171 reg_val &= ~(SXGBE_CORE_RXQ_ENABLE_MASK << queue_num); in sxgbe_core_enable_rxqueue()
176 static void sxgbe_core_disable_rxqueue(void __iomem *ioaddr, int queue_num) in sxgbe_core_disable_rxqueue() argument
181 reg_val &= ~(SXGBE_CORE_RXQ_ENABLE_MASK << queue_num); in sxgbe_core_disable_rxqueue()
/drivers/net/dsa/
Dbcm_sf2_cfp.c216 unsigned int queue_num, in bcm_sf2_cfp_act_pol_set() argument
228 CHANGE_TC | queue_num << NEW_TC_SHIFT; in bcm_sf2_cfp_act_pol_set()
348 unsigned int queue_num, in bcm_sf2_cfp_ipv4_rule_set() argument
461 queue_num, true); in bcm_sf2_cfp_ipv4_rule_set()
629 unsigned int queue_num, in bcm_sf2_cfp_ipv6_rule_set() argument
763 queue_num, false); in bcm_sf2_cfp_ipv6_rule_set()
821 queue_num, true); in bcm_sf2_cfp_ipv6_rule_set()
853 unsigned int queue_num, port_num; in bcm_sf2_cfp_rule_insert() local
903 queue_num = ring_cookie % SF2_NUM_EGRESS_QUEUES; in bcm_sf2_cfp_rule_insert()
911 queue_num, fs); in bcm_sf2_cfp_rule_insert()
[all …]
/drivers/soc/ti/
Dknav_qmss_acc.c284 cmd->queue_num, cmd->timer_config); in knav_acc_write()
287 writel_relaxed(cmd->queue_num, &pdsp->acc_command->queue_num); in knav_acc_write()
324 cmd->queue_num = info->list_entries << 16; in knav_acc_setup_cmd()
325 cmd->queue_num |= queue_base; in knav_acc_setup_cmd()
Dknav_qmss.h91 u32 queue_num; member
/drivers/net/wireless/broadcom/b43legacy/
Dmain.c2452 int queue_num; in b43legacy_tx_work() local
2462 for (queue_num = 0; queue_num < B43legacy_QOS_QUEUE_NUM; queue_num++) { in b43legacy_tx_work()
2463 while (skb_queue_len(&wl->tx_queue[queue_num])) { in b43legacy_tx_work()
2464 skb = skb_dequeue(&wl->tx_queue[queue_num]); in b43legacy_tx_work()
2470 wl->tx_queue_stopped[queue_num] = 1; in b43legacy_tx_work()
2471 ieee80211_stop_queue(wl->hw, queue_num); in b43legacy_tx_work()
2472 skb_queue_head(&wl->tx_queue[queue_num], skb); in b43legacy_tx_work()
2481 wl->tx_queue_stopped[queue_num] = 0; in b43legacy_tx_work()
2917 int queue_num; in b43legacy_wireless_core_stop() local
2941 for (queue_num = 0; queue_num < B43legacy_QOS_QUEUE_NUM; queue_num++) { in b43legacy_wireless_core_stop()
[all …]
/drivers/dma/
Dfsl-qdma.c488 int queue_num, block_number; in fsl_qdma_alloc_queue_resources() local
492 queue_num = fsl_qdma->n_queues; in fsl_qdma_alloc_queue_resources()
495 if (queue_num > FSL_QDMA_QUEUE_MAX) in fsl_qdma_alloc_queue_resources()
496 queue_num = FSL_QDMA_QUEUE_MAX; in fsl_qdma_alloc_queue_resources()
497 len = sizeof(*queue_head) * queue_num * block_number; in fsl_qdma_alloc_queue_resources()
503 queue_size, queue_num); in fsl_qdma_alloc_queue_resources()
509 for (i = 0; i < queue_num; i++) { in fsl_qdma_alloc_queue_resources()
516 queue_temp = queue_head + i + (j * queue_num); in fsl_qdma_alloc_queue_resources()
/drivers/infiniband/ulp/rtrs/
Drtrs-pri.h301 struct rtrs_iu *rtrs_iu_alloc(u32 queue_num, size_t size, gfp_t t,
304 void rtrs_iu_free(struct rtrs_iu *iu, struct ib_device *dev, u32 queue_num);
Drtrs-clt.h74 u32 queue_num; member
Drtrs.c55 void rtrs_iu_free(struct rtrs_iu *ius, struct ib_device *ibdev, u32 queue_num) in rtrs_iu_free() argument
63 for (i = 0; i < queue_num; i++) { in rtrs_iu_free()
/drivers/net/wireless/broadcom/b43/
Dmain.c3587 int queue_num; in b43_tx_work() local
3597 for (queue_num = 0; queue_num < B43_QOS_QUEUE_NUM; queue_num++) { in b43_tx_work()
3598 while (skb_queue_len(&wl->tx_queue[queue_num])) { in b43_tx_work()
3599 skb = skb_dequeue(&wl->tx_queue[queue_num]); in b43_tx_work()
3605 wl->tx_queue_stopped[queue_num] = true; in b43_tx_work()
3606 ieee80211_stop_queue(wl->hw, queue_num); in b43_tx_work()
3607 skb_queue_head(&wl->tx_queue[queue_num], skb); in b43_tx_work()
3616 wl->tx_queue_stopped[queue_num] = false; in b43_tx_work()
4326 int queue_num; in b43_wireless_core_stop() local
4379 for (queue_num = 0; queue_num < B43_QOS_QUEUE_NUM; queue_num++) { in b43_wireless_core_stop()
[all …]
/drivers/net/ethernet/hisilicon/hns3/
Dhns3_ethtool.c648 int queue_num = h->kinfo.num_tqps; in hns3_get_ringparam() local
659 param->rx_pending = priv->ring[queue_num].desc_num; in hns3_get_ringparam()
1096 u16 queue_num = h->kinfo.num_tqps; in hns3_set_ringparam() local
1107 old_rx_desc_num = priv->ring[queue_num].desc_num; in hns3_set_ringparam()
1373 int queue_num = h->kinfo.num_tqps; in hns3_set_coalesce_per_queue() local
1376 rx_vector = priv->ring[queue_num + queue].tqp_vector; in hns3_set_coalesce_per_queue()
1414 u16 queue_num = h->kinfo.num_tqps; in hns3_set_coalesce() local
1439 for (i = 0; i < queue_num; i++) in hns3_set_coalesce()
/drivers/net/ethernet/ibm/
Dibmvnic.c1665 int queue_num; in ibmvnic_tx_scrq_clean_buffer() local
1672 queue_num = tx_scrq->pool_index; in ibmvnic_tx_scrq_clean_buffer()
1680 tx_pool = &adapter->tso_pool[queue_num]; in ibmvnic_tx_scrq_clean_buffer()
1683 tx_pool = &adapter->tx_pool[queue_num]; in ibmvnic_tx_scrq_clean_buffer()
1692 adapter->tx_stats_buffers[queue_num].packets--; in ibmvnic_tx_scrq_clean_buffer()
1693 adapter->tx_stats_buffers[queue_num].bytes -= in ibmvnic_tx_scrq_clean_buffer()
1704 __netif_subqueue_stopped(adapter->netdev, queue_num)) { in ibmvnic_tx_scrq_clean_buffer()
1708 netif_wake_subqueue(adapter->netdev, queue_num); in ibmvnic_tx_scrq_clean_buffer()
1710 queue_num); in ibmvnic_tx_scrq_clean_buffer()
1744 int queue_num = skb_get_queue_mapping(skb); in ibmvnic_xmit() local
[all …]
/drivers/net/ethernet/chelsio/cxgb3/
Dcxgb3_ioctl.h70 uint32_t queue_num; member
/drivers/net/usb/
Daqc111.c806 u8 queue_num = 0; in aqc111_configure_rx() local
829 queue_num = 1; in aqc111_configure_rx()
851 queue_num = 0; in aqc111_configure_rx()
859 queue_num = 2; /* For Jumbo packet 16KB */ in aqc111_configure_rx()
861 memcpy(buf, &AQC111_BULKIN_SIZE[queue_num], 5); in aqc111_configure_rx()
/drivers/net/ethernet/hisilicon/hns3/hns3pf/
Dhclge_tm.h202 __le16 queue_num; member
/drivers/net/wireless/st/cw1200/
Dwsm.c1664 int queue_num; in wsm_get_tx() local
1691 queue_num = queue - priv->tx_queue; in wsm_get_tx()
1728 if (priv->edca.params[queue_num].txop_limit) in wsm_get_tx()
1736 priv->tx_burst_idx = queue_num; in wsm_get_tx()
/drivers/net/wireless/intel/iwlwifi/fw/api/
Dtx.h870 __le16 queue_num; member
/drivers/infiniband/hw/hns/
Dhns_roce_hw_v2.c5575 irq_work->queue_num, irq_work->sub_type); in hns_roce_irq_work_handle()
5579 irq_work->queue_num); in hns_roce_irq_work_handle()
5583 irq_work->queue_num, irq_work->sub_type); in hns_roce_irq_work_handle()
5595 ibdev_err(ibdev, "CQ 0x%x access err.\n", irq_work->queue_num); in hns_roce_irq_work_handle()
5598 ibdev_warn(ibdev, "CQ 0x%x overflow\n", irq_work->queue_num); in hns_roce_irq_work_handle()
5620 struct hns_roce_eq *eq, u32 queue_num) in hns_roce_v2_init_irq_work() argument
5632 irq_work->queue_num = queue_num; in hns_roce_v2_init_irq_work()
5679 u32 queue_num; in hns_roce_v2_aeq_int() local
5694 queue_num = roce_get_field(aeqe->event.queue_event.num, in hns_roce_v2_aeq_int()
5709 hns_roce_qp_event(hr_dev, queue_num, event_type); in hns_roce_v2_aeq_int()
[all …]
/drivers/net/wireless/intel/iwlwifi/mvm/
Dtx.c2100 int queue_num = le16_to_cpu(queue_info->queue_num); in iwl_mvm_flush_sta_tids() local
2107 tid, queue_num, read_before, read_after); in iwl_mvm_flush_sta_tids()
2109 iwl_mvm_tx_reclaim(mvm, sta_id, tid, queue_num, read_after, in iwl_mvm_flush_sta_tids()
/drivers/block/
Dvirtio_blk.c293 struct virtio_blk_vq *vq = &vblk->vqs[hctx->queue_num]; in virtio_commit_rqs()
312 int qid = hctx->queue_num; in virtio_queue_rq()
/drivers/dma/ti/
Dcppi41.c293 static u32 cppi41_pop_desc(struct cppi41_dd *cdd, unsigned queue_num) in cppi41_pop_desc() argument
297 desc = cppi_readl(cdd->qmgr_mem + QMGR_QUEUE_D(queue_num)); in cppi41_pop_desc()

12