Searched refs:r_idx (Results 1 – 10 of 10) sorted by relevance
716 u16 r_idx; in qtnf_pcie_pearl_rx_poll() local724 r_idx = priv->rx_bd_r_index; in qtnf_pcie_pearl_rx_poll()725 rxbd = &ps->rx_bd_vbase[r_idx]; in qtnf_pcie_pearl_rx_poll()728 skb = priv->rx_skb[r_idx]; in qtnf_pcie_pearl_rx_poll()733 pr_warn("skip invalid rxbd[%d]\n", r_idx); in qtnf_pcie_pearl_rx_poll()738 pr_warn("skip missing rx_skb[%d]\n", r_idx); in qtnf_pcie_pearl_rx_poll()774 priv->rx_skb[r_idx] = NULL; in qtnf_pcie_pearl_rx_poll()775 if (++r_idx >= priv->rx_bd_num) in qtnf_pcie_pearl_rx_poll()776 r_idx = 0; in qtnf_pcie_pearl_rx_poll()778 priv->rx_bd_r_index = r_idx; in qtnf_pcie_pearl_rx_poll()
621 u16 r_idx; in qtnf_topaz_rx_poll() local629 r_idx = priv->rx_bd_r_index; in qtnf_topaz_rx_poll()630 rxbd = &ts->rx_bd_vbase[r_idx]; in qtnf_topaz_rx_poll()633 skb = priv->rx_skb[r_idx]; in qtnf_topaz_rx_poll()639 pr_warn("skip invalid rxbd[%d]\n", r_idx); in qtnf_topaz_rx_poll()644 pr_warn("skip missing rx_skb[%d]\n", r_idx); in qtnf_topaz_rx_poll()685 priv->rx_skb[r_idx] = NULL; in qtnf_topaz_rx_poll()686 if (++r_idx >= priv->rx_bd_num) in qtnf_topaz_rx_poll()687 r_idx = 0; in qtnf_topaz_rx_poll()689 priv->rx_bd_r_index = r_idx; in qtnf_topaz_rx_poll()
514 unsigned int r_idx = skb->queue_mapping; in fm10k_xmit_frame() local572 if (r_idx >= num_tx_queues) in fm10k_xmit_frame()573 r_idx %= num_tx_queues; in fm10k_xmit_frame()575 err = fm10k_xmit_frame_ring(skb, interface->tx_ring[r_idx]); in fm10k_xmit_frame()
144 unsigned int r_idx = skb->queue_mapping; in alx_tx_queue_mapping() local146 if (r_idx >= alx->num_txq) in alx_tx_queue_mapping()147 r_idx = r_idx % alx->num_txq; in alx_tx_queue_mapping()149 return alx->qnapi[r_idx]->txq; in alx_tx_queue_mapping()
1507 unsigned int r_idx = skb->queue_mapping; in wx_xmit_frame() local1522 if (r_idx >= wx->num_tx_queues) in wx_xmit_frame()1523 r_idx = r_idx % wx->num_tx_queues; in wx_xmit_frame()1524 tx_ring = wx->tx_ring[r_idx]; in wx_xmit_frame()
2953 unsigned int r_idx = smp_processor_id(); in igb_xdp_tx_queue_mapping() local2955 if (r_idx >= adapter->num_tx_queues) in igb_xdp_tx_queue_mapping()2956 r_idx = r_idx % adapter->num_tx_queues; in igb_xdp_tx_queue_mapping()2958 return adapter->tx_ring[r_idx]; in igb_xdp_tx_queue_mapping()6552 unsigned int r_idx = skb->queue_mapping; in igb_tx_queue_mapping() local6554 if (r_idx >= adapter->num_tx_queues) in igb_tx_queue_mapping()6555 r_idx = r_idx % adapter->num_tx_queues; in igb_tx_queue_mapping()6557 return adapter->tx_ring[r_idx]; in igb_tx_queue_mapping()
1686 unsigned int r_idx = skb->queue_mapping; in igc_tx_queue_mapping() local1688 if (r_idx >= adapter->num_tx_queues) in igc_tx_queue_mapping()1689 r_idx = r_idx % adapter->num_tx_queues; in igc_tx_queue_mapping()1691 return adapter->tx_ring[r_idx]; in igc_tx_queue_mapping()
449 iavf_map_vector_to_rxq(struct iavf_adapter *adapter, int v_idx, int r_idx) in iavf_map_vector_to_rxq() argument452 struct iavf_ring *rx_ring = &adapter->rx_rings[r_idx]; in iavf_map_vector_to_rxq()462 q_vector->ring_mask |= BIT(r_idx); in iavf_map_vector_to_rxq()
1290 qe += q->u.r_idx * q->size; in sli_mq_read()1293 if (q->index == q->u.r_idx) { in sli_mq_read()1299 q->u.r_idx = (q->u.r_idx + 1) & (q->length - 1); in sli_mq_read()
3543 u32 r_idx; /* "read" index (MQ only) */ member