/drivers/net/ethernet/samsung/sxgbe/ |
D | sxgbe_mtl.c | 61 static void sxgbe_mtl_set_txfifosize(void __iomem *ioaddr, int queue_num, in sxgbe_mtl_set_txfifosize() argument 68 reg_val = readl(ioaddr + SXGBE_MTL_TXQ_OPMODE_REG(queue_num)); in sxgbe_mtl_set_txfifosize() 70 writel(reg_val, ioaddr + SXGBE_MTL_TXQ_OPMODE_REG(queue_num)); in sxgbe_mtl_set_txfifosize() 73 static void sxgbe_mtl_set_rxfifosize(void __iomem *ioaddr, int queue_num, in sxgbe_mtl_set_rxfifosize() argument 80 reg_val = readl(ioaddr + SXGBE_MTL_RXQ_OPMODE_REG(queue_num)); in sxgbe_mtl_set_rxfifosize() 82 writel(reg_val, ioaddr + SXGBE_MTL_RXQ_OPMODE_REG(queue_num)); in sxgbe_mtl_set_rxfifosize() 85 static void sxgbe_mtl_enable_txqueue(void __iomem *ioaddr, int queue_num) in sxgbe_mtl_enable_txqueue() argument 89 reg_val = readl(ioaddr + SXGBE_MTL_TXQ_OPMODE_REG(queue_num)); in sxgbe_mtl_enable_txqueue() 91 writel(reg_val, ioaddr + SXGBE_MTL_TXQ_OPMODE_REG(queue_num)); in sxgbe_mtl_enable_txqueue() 94 static void sxgbe_mtl_disable_txqueue(void __iomem *ioaddr, int queue_num) in sxgbe_mtl_disable_txqueue() argument [all …]
|
D | sxgbe_mtl.h | 64 void (*mtl_set_txfifosize)(void __iomem *ioaddr, int queue_num, 67 void (*mtl_set_rxfifosize)(void __iomem *ioaddr, int queue_num, 70 void (*mtl_enable_txqueue)(void __iomem *ioaddr, int queue_num); 72 void (*mtl_disable_txqueue)(void __iomem *ioaddr, int queue_num); 74 void (*set_tx_mtl_mode)(void __iomem *ioaddr, int queue_num, 77 void (*set_rx_mtl_mode)(void __iomem *ioaddr, int queue_num, 82 void (*mtl_fc_active)(void __iomem *ioaddr, int queue_num, 85 void (*mtl_fc_deactive)(void __iomem *ioaddr, int queue_num, 88 void (*mtl_fc_enable)(void __iomem *ioaddr, int queue_num); 90 void (*mtl_fep_enable)(void __iomem *ioaddr, int queue_num); [all …]
|
D | sxgbe_main.c | 563 int queue_num, ret; in init_dma_desc_rings() local 569 SXGBE_FOR_EACH_QUEUE(SXGBE_TX_QUEUES, queue_num) { in init_dma_desc_rings() 570 ret = init_tx_ring(priv->device, queue_num, in init_dma_desc_rings() 571 priv->txq[queue_num], tx_rsize); in init_dma_desc_rings() 580 priv->txq[queue_num]->priv_ptr = priv; in init_dma_desc_rings() 584 SXGBE_FOR_EACH_QUEUE(SXGBE_RX_QUEUES, queue_num) { in init_dma_desc_rings() 585 ret = init_rx_ring(netd, queue_num, in init_dma_desc_rings() 586 priv->rxq[queue_num], rx_rsize); in init_dma_desc_rings() 595 priv->rxq[queue_num]->priv_ptr = priv; in init_dma_desc_rings() 603 while (queue_num--) in init_dma_desc_rings() [all …]
|
D | sxgbe_common.h | 194 #define SXGBE_FOR_EACH_QUEUE(max_queues, queue_num) \ argument 195 for (queue_num = 0; queue_num < max_queues; queue_num++) 358 void (*enable_rxqueue)(void __iomem *ioaddr, int queue_num); 359 void (*disable_rxqueue)(void __iomem *ioaddr, int queue_num);
|
D | sxgbe_core.c | 166 static void sxgbe_core_enable_rxqueue(void __iomem *ioaddr, int queue_num) in sxgbe_core_enable_rxqueue() argument 171 reg_val &= ~(SXGBE_CORE_RXQ_ENABLE_MASK << queue_num); in sxgbe_core_enable_rxqueue() 176 static void sxgbe_core_disable_rxqueue(void __iomem *ioaddr, int queue_num) in sxgbe_core_disable_rxqueue() argument 181 reg_val &= ~(SXGBE_CORE_RXQ_ENABLE_MASK << queue_num); in sxgbe_core_disable_rxqueue()
|
/drivers/net/dsa/ |
D | bcm_sf2_cfp.c | 216 unsigned int queue_num, in bcm_sf2_cfp_act_pol_set() argument 228 CHANGE_TC | queue_num << NEW_TC_SHIFT; in bcm_sf2_cfp_act_pol_set() 348 unsigned int queue_num, in bcm_sf2_cfp_ipv4_rule_set() argument 461 queue_num, true); in bcm_sf2_cfp_ipv4_rule_set() 629 unsigned int queue_num, in bcm_sf2_cfp_ipv6_rule_set() argument 763 queue_num, false); in bcm_sf2_cfp_ipv6_rule_set() 821 queue_num, true); in bcm_sf2_cfp_ipv6_rule_set() 853 unsigned int queue_num, port_num; in bcm_sf2_cfp_rule_insert() local 903 queue_num = ring_cookie % SF2_NUM_EGRESS_QUEUES; in bcm_sf2_cfp_rule_insert() 911 queue_num, fs); in bcm_sf2_cfp_rule_insert() [all …]
|
/drivers/soc/ti/ |
D | knav_qmss_acc.c | 284 cmd->queue_num, cmd->timer_config); in knav_acc_write() 287 writel_relaxed(cmd->queue_num, &pdsp->acc_command->queue_num); in knav_acc_write() 324 cmd->queue_num = info->list_entries << 16; in knav_acc_setup_cmd() 325 cmd->queue_num |= queue_base; in knav_acc_setup_cmd()
|
D | knav_qmss.h | 91 u32 queue_num; member
|
/drivers/net/wireless/broadcom/b43legacy/ |
D | main.c | 2452 int queue_num; in b43legacy_tx_work() local 2462 for (queue_num = 0; queue_num < B43legacy_QOS_QUEUE_NUM; queue_num++) { in b43legacy_tx_work() 2463 while (skb_queue_len(&wl->tx_queue[queue_num])) { in b43legacy_tx_work() 2464 skb = skb_dequeue(&wl->tx_queue[queue_num]); in b43legacy_tx_work() 2470 wl->tx_queue_stopped[queue_num] = 1; in b43legacy_tx_work() 2471 ieee80211_stop_queue(wl->hw, queue_num); in b43legacy_tx_work() 2472 skb_queue_head(&wl->tx_queue[queue_num], skb); in b43legacy_tx_work() 2481 wl->tx_queue_stopped[queue_num] = 0; in b43legacy_tx_work() 2917 int queue_num; in b43legacy_wireless_core_stop() local 2941 for (queue_num = 0; queue_num < B43legacy_QOS_QUEUE_NUM; queue_num++) { in b43legacy_wireless_core_stop() [all …]
|
/drivers/dma/ |
D | fsl-qdma.c | 488 int queue_num, block_number; in fsl_qdma_alloc_queue_resources() local 492 queue_num = fsl_qdma->n_queues; in fsl_qdma_alloc_queue_resources() 495 if (queue_num > FSL_QDMA_QUEUE_MAX) in fsl_qdma_alloc_queue_resources() 496 queue_num = FSL_QDMA_QUEUE_MAX; in fsl_qdma_alloc_queue_resources() 497 len = sizeof(*queue_head) * queue_num * block_number; in fsl_qdma_alloc_queue_resources() 503 queue_size, queue_num); in fsl_qdma_alloc_queue_resources() 509 for (i = 0; i < queue_num; i++) { in fsl_qdma_alloc_queue_resources() 516 queue_temp = queue_head + i + (j * queue_num); in fsl_qdma_alloc_queue_resources()
|
/drivers/infiniband/ulp/rtrs/ |
D | rtrs-pri.h | 301 struct rtrs_iu *rtrs_iu_alloc(u32 queue_num, size_t size, gfp_t t, 304 void rtrs_iu_free(struct rtrs_iu *iu, struct ib_device *dev, u32 queue_num);
|
D | rtrs-clt.h | 74 u32 queue_num; member
|
D | rtrs.c | 55 void rtrs_iu_free(struct rtrs_iu *ius, struct ib_device *ibdev, u32 queue_num) in rtrs_iu_free() argument 63 for (i = 0; i < queue_num; i++) { in rtrs_iu_free()
|
/drivers/net/wireless/broadcom/b43/ |
D | main.c | 3587 int queue_num; in b43_tx_work() local 3597 for (queue_num = 0; queue_num < B43_QOS_QUEUE_NUM; queue_num++) { in b43_tx_work() 3598 while (skb_queue_len(&wl->tx_queue[queue_num])) { in b43_tx_work() 3599 skb = skb_dequeue(&wl->tx_queue[queue_num]); in b43_tx_work() 3605 wl->tx_queue_stopped[queue_num] = true; in b43_tx_work() 3606 ieee80211_stop_queue(wl->hw, queue_num); in b43_tx_work() 3607 skb_queue_head(&wl->tx_queue[queue_num], skb); in b43_tx_work() 3616 wl->tx_queue_stopped[queue_num] = false; in b43_tx_work() 4326 int queue_num; in b43_wireless_core_stop() local 4379 for (queue_num = 0; queue_num < B43_QOS_QUEUE_NUM; queue_num++) { in b43_wireless_core_stop() [all …]
|
/drivers/net/ethernet/hisilicon/hns3/ |
D | hns3_ethtool.c | 648 int queue_num = h->kinfo.num_tqps; in hns3_get_ringparam() local 659 param->rx_pending = priv->ring[queue_num].desc_num; in hns3_get_ringparam() 1096 u16 queue_num = h->kinfo.num_tqps; in hns3_set_ringparam() local 1107 old_rx_desc_num = priv->ring[queue_num].desc_num; in hns3_set_ringparam() 1373 int queue_num = h->kinfo.num_tqps; in hns3_set_coalesce_per_queue() local 1376 rx_vector = priv->ring[queue_num + queue].tqp_vector; in hns3_set_coalesce_per_queue() 1414 u16 queue_num = h->kinfo.num_tqps; in hns3_set_coalesce() local 1439 for (i = 0; i < queue_num; i++) in hns3_set_coalesce()
|
/drivers/net/ethernet/ibm/ |
D | ibmvnic.c | 1665 int queue_num; in ibmvnic_tx_scrq_clean_buffer() local 1672 queue_num = tx_scrq->pool_index; in ibmvnic_tx_scrq_clean_buffer() 1680 tx_pool = &adapter->tso_pool[queue_num]; in ibmvnic_tx_scrq_clean_buffer() 1683 tx_pool = &adapter->tx_pool[queue_num]; in ibmvnic_tx_scrq_clean_buffer() 1692 adapter->tx_stats_buffers[queue_num].packets--; in ibmvnic_tx_scrq_clean_buffer() 1693 adapter->tx_stats_buffers[queue_num].bytes -= in ibmvnic_tx_scrq_clean_buffer() 1704 __netif_subqueue_stopped(adapter->netdev, queue_num)) { in ibmvnic_tx_scrq_clean_buffer() 1708 netif_wake_subqueue(adapter->netdev, queue_num); in ibmvnic_tx_scrq_clean_buffer() 1710 queue_num); in ibmvnic_tx_scrq_clean_buffer() 1744 int queue_num = skb_get_queue_mapping(skb); in ibmvnic_xmit() local [all …]
|
/drivers/net/ethernet/chelsio/cxgb3/ |
D | cxgb3_ioctl.h | 70 uint32_t queue_num; member
|
/drivers/net/usb/ |
D | aqc111.c | 806 u8 queue_num = 0; in aqc111_configure_rx() local 829 queue_num = 1; in aqc111_configure_rx() 851 queue_num = 0; in aqc111_configure_rx() 859 queue_num = 2; /* For Jumbo packet 16KB */ in aqc111_configure_rx() 861 memcpy(buf, &AQC111_BULKIN_SIZE[queue_num], 5); in aqc111_configure_rx()
|
/drivers/net/ethernet/hisilicon/hns3/hns3pf/ |
D | hclge_tm.h | 202 __le16 queue_num; member
|
/drivers/net/wireless/st/cw1200/ |
D | wsm.c | 1664 int queue_num; in wsm_get_tx() local 1691 queue_num = queue - priv->tx_queue; in wsm_get_tx() 1728 if (priv->edca.params[queue_num].txop_limit) in wsm_get_tx() 1736 priv->tx_burst_idx = queue_num; in wsm_get_tx()
|
/drivers/net/wireless/intel/iwlwifi/fw/api/ |
D | tx.h | 870 __le16 queue_num; member
|
/drivers/infiniband/hw/hns/ |
D | hns_roce_hw_v2.c | 5575 irq_work->queue_num, irq_work->sub_type); in hns_roce_irq_work_handle() 5579 irq_work->queue_num); in hns_roce_irq_work_handle() 5583 irq_work->queue_num, irq_work->sub_type); in hns_roce_irq_work_handle() 5595 ibdev_err(ibdev, "CQ 0x%x access err.\n", irq_work->queue_num); in hns_roce_irq_work_handle() 5598 ibdev_warn(ibdev, "CQ 0x%x overflow\n", irq_work->queue_num); in hns_roce_irq_work_handle() 5620 struct hns_roce_eq *eq, u32 queue_num) in hns_roce_v2_init_irq_work() argument 5632 irq_work->queue_num = queue_num; in hns_roce_v2_init_irq_work() 5679 u32 queue_num; in hns_roce_v2_aeq_int() local 5694 queue_num = roce_get_field(aeqe->event.queue_event.num, in hns_roce_v2_aeq_int() 5709 hns_roce_qp_event(hr_dev, queue_num, event_type); in hns_roce_v2_aeq_int() [all …]
|
/drivers/net/wireless/intel/iwlwifi/mvm/ |
D | tx.c | 2100 int queue_num = le16_to_cpu(queue_info->queue_num); in iwl_mvm_flush_sta_tids() local 2107 tid, queue_num, read_before, read_after); in iwl_mvm_flush_sta_tids() 2109 iwl_mvm_tx_reclaim(mvm, sta_id, tid, queue_num, read_after, in iwl_mvm_flush_sta_tids()
|
/drivers/block/ |
D | virtio_blk.c | 293 struct virtio_blk_vq *vq = &vblk->vqs[hctx->queue_num]; in virtio_commit_rqs() 312 int qid = hctx->queue_num; in virtio_queue_rq()
|
/drivers/dma/ti/ |
D | cppi41.c | 293 static u32 cppi41_pop_desc(struct cppi41_dd *cdd, unsigned queue_num) in cppi41_pop_desc() argument 297 desc = cppi_readl(cdd->qmgr_mem + QMGR_QUEUE_D(queue_num)); in cppi41_pop_desc()
|