/drivers/net/ethernet/intel/fm10k/ |
D | fm10k_main.c | 119 union fm10k_rx_desc *rx_desc; in fm10k_alloc_rx_buffers() local 127 rx_desc = FM10K_RX_DESC(rx_ring, i); in fm10k_alloc_rx_buffers() 138 rx_desc->q.pkt_addr = cpu_to_le64(bi->dma + bi->page_offset); in fm10k_alloc_rx_buffers() 140 rx_desc++; in fm10k_alloc_rx_buffers() 144 rx_desc = FM10K_RX_DESC(rx_ring, 0); in fm10k_alloc_rx_buffers() 150 rx_desc->d.staterr = 0; in fm10k_alloc_rx_buffers() 258 union fm10k_rx_desc *rx_desc, in fm10k_add_rx_frag() argument 305 union fm10k_rx_desc *rx_desc, in fm10k_fetch_rx_buffer() argument 308 unsigned int size = le16_to_cpu(rx_desc->w.length); in fm10k_fetch_rx_buffer() 346 if (fm10k_add_rx_frag(rx_buffer, size, rx_desc, skb)) { in fm10k_fetch_rx_buffer() [all …]
|
/drivers/net/ethernet/intel/ice/ |
D | ice_txrx_lib.c | 56 ice_rx_hash(struct ice_ring *rx_ring, union ice_32b_rx_flex_desc *rx_desc, in ice_rx_hash() argument 65 if (rx_desc->wb.rxdid != ICE_RXDID_FLEX_NIC) in ice_rx_hash() 68 nic_mdid = (struct ice_32b_rx_flex_desc_nic *)rx_desc; in ice_rx_hash() 84 union ice_32b_rx_flex_desc *rx_desc, u8 ptype) in ice_rx_csum() argument 90 rx_status0 = le16_to_cpu(rx_desc->wb.status_error0); in ice_rx_csum() 91 rx_status1 = le16_to_cpu(rx_desc->wb.status_error1); in ice_rx_csum() 168 union ice_32b_rx_flex_desc *rx_desc, in ice_process_skb_fields() argument 171 ice_rx_hash(rx_ring, rx_desc, skb, ptype); in ice_process_skb_fields() 176 ice_rx_csum(rx_ring, skb, rx_desc, ptype); in ice_process_skb_fields()
|
D | ice_xsk.c | 431 union ice_32b_rx_flex_desc *rx_desc; in ice_alloc_rx_bufs_zc() local 440 rx_desc = ICE_RX_DESC(rx_ring, ntu); in ice_alloc_rx_bufs_zc() 451 rx_desc->read.pkt_addr = cpu_to_le64(dma); in ice_alloc_rx_bufs_zc() 452 rx_desc->wb.status_error0 = 0; in ice_alloc_rx_bufs_zc() 454 rx_desc++; in ice_alloc_rx_bufs_zc() 459 rx_desc = ICE_RX_DESC(rx_ring, 0); in ice_alloc_rx_bufs_zc() 467 rx_desc->wb.status_error0 = 0; in ice_alloc_rx_bufs_zc() 592 union ice_32b_rx_flex_desc *rx_desc; in ice_clean_rx_irq_zc() local 606 rx_desc = ICE_RX_DESC(rx_ring, rx_ring->next_to_clean); in ice_clean_rx_irq_zc() 609 if (!ice_test_staterr(rx_desc, stat_err_bits)) in ice_clean_rx_irq_zc() [all …]
|
D | ice_txrx_lib.h | 19 ice_test_staterr(union ice_32b_rx_flex_desc *rx_desc, const u16 stat_err_bits) in ice_test_staterr() argument 21 return !!(rx_desc->wb.status_error0 & cpu_to_le16(stat_err_bits)); in ice_test_staterr() 55 union ice_32b_rx_flex_desc *rx_desc,
|
D | ice_txrx.c | 682 union ice_32b_rx_flex_desc *rx_desc; in ice_alloc_rx_bufs() local 692 rx_desc = ICE_RX_DESC(rx_ring, ntu); in ice_alloc_rx_bufs() 709 rx_desc->read.pkt_addr = cpu_to_le64(bi->dma + bi->page_offset); in ice_alloc_rx_bufs() 711 rx_desc++; in ice_alloc_rx_bufs() 715 rx_desc = ICE_RX_DESC(rx_ring, 0); in ice_alloc_rx_bufs() 721 rx_desc->wb.status_error0 = 0; in ice_alloc_rx_bufs() 1064 ice_is_non_eop(struct ice_ring *rx_ring, union ice_32b_rx_flex_desc *rx_desc, in ice_is_non_eop() argument 1069 if (likely(ice_test_staterr(rx_desc, ICE_RXD_EOF))) in ice_is_non_eop() 1108 union ice_32b_rx_flex_desc *rx_desc; in ice_clean_rx_irq() local 1118 rx_desc = ICE_RX_DESC(rx_ring, rx_ring->next_to_clean); in ice_clean_rx_irq() [all …]
|
/drivers/infiniband/ulp/isert/ |
D | ib_isert.c | 154 struct iser_rx_desc *rx_desc; in isert_alloc_rx_descriptors() local 165 rx_desc = isert_conn->rx_descs; in isert_alloc_rx_descriptors() 167 for (i = 0; i < ISERT_QP_MAX_RECV_DTOS; i++, rx_desc++) { in isert_alloc_rx_descriptors() 168 dma_addr = ib_dma_map_single(ib_dev, rx_desc->buf, in isert_alloc_rx_descriptors() 173 rx_desc->dma_addr = dma_addr; in isert_alloc_rx_descriptors() 175 rx_sg = &rx_desc->rx_sg; in isert_alloc_rx_descriptors() 176 rx_sg->addr = rx_desc->dma_addr + isert_get_hdr_offset(rx_desc); in isert_alloc_rx_descriptors() 179 rx_desc->rx_cqe.done = isert_recv_done; in isert_alloc_rx_descriptors() 185 rx_desc = isert_conn->rx_descs; in isert_alloc_rx_descriptors() 186 for (j = 0; j < i; j++, rx_desc++) { in isert_alloc_rx_descriptors() [all …]
|
/drivers/net/ethernet/seeq/ |
D | sgiseeq.c | 59 (unsigned long)((sp)->rx_desc))) 95 struct sgiseeq_rx_desc *rx_desc; member 202 if (!sp->rx_desc[i].skb) { in seeq_init_ring() 212 sp->rx_desc[i].skb = skb; in seeq_init_ring() 213 sp->rx_desc[i].rdma.pbuf = dma_addr; in seeq_init_ring() 215 sp->rx_desc[i].rdma.cntinfo = RCNTINFO_INIT; in seeq_init_ring() 216 dma_sync_desc_dev(dev, &sp->rx_desc[i]); in seeq_init_ring() 218 sp->rx_desc[i - 1].rdma.cntinfo |= HPCDMA_EOR; in seeq_init_ring() 219 dma_sync_desc_dev(dev, &sp->rx_desc[i - 1]); in seeq_init_ring() 238 if (sp->rx_desc[i].skb) { in seeq_purge_ring() [all …]
|
/drivers/net/wireless/ath/ath11k/ |
D | dp_rx.c | 221 static u32 ath11k_dp_rxdesc_get_mpdulen_err(struct hal_rx_desc *rx_desc) in ath11k_dp_rxdesc_get_mpdulen_err() argument 225 rx_attn = &rx_desc->attention; in ath11k_dp_rxdesc_get_mpdulen_err() 231 static u32 ath11k_dp_rxdesc_get_decap_format(struct hal_rx_desc *rx_desc) in ath11k_dp_rxdesc_get_decap_format() argument 235 rx_msdu_start = &rx_desc->msdu_start; in ath11k_dp_rxdesc_get_decap_format() 241 static u8 *ath11k_dp_rxdesc_get_80211hdr(struct hal_rx_desc *rx_desc) in ath11k_dp_rxdesc_get_80211hdr() argument 245 rx_pkt_hdr = &rx_desc->msdu_payload[0]; in ath11k_dp_rxdesc_get_80211hdr() 250 static bool ath11k_dp_rxdesc_mpdu_valid(struct hal_rx_desc *rx_desc) in ath11k_dp_rxdesc_mpdu_valid() argument 255 __le32_to_cpu(rx_desc->mpdu_start_tag)); in ath11k_dp_rxdesc_mpdu_valid() 260 static u32 ath11k_dp_rxdesc_get_ppduid(struct hal_rx_desc *rx_desc) in ath11k_dp_rxdesc_get_ppduid() argument 262 return __le16_to_cpu(rx_desc->mpdu_start.phy_ppdu_id); in ath11k_dp_rxdesc_get_ppduid() [all …]
|
/drivers/net/ethernet/intel/i40e/ |
D | i40e_xsk.c | 248 union i40e_rx_desc *rx_desc; in i40e_alloc_rx_buffers_zc() local 253 rx_desc = I40E_RX_DESC(rx_ring, ntu); in i40e_alloc_rx_buffers_zc() 263 rx_desc->read.pkt_addr = cpu_to_le64(dma); in i40e_alloc_rx_buffers_zc() 264 rx_desc->read.hdr_addr = 0; in i40e_alloc_rx_buffers_zc() 266 rx_desc++; in i40e_alloc_rx_buffers_zc() 271 rx_desc = I40E_RX_DESC(rx_ring, 0); in i40e_alloc_rx_buffers_zc() 282 rx_desc->wb.qword1.status_error_len = 0; in i40e_alloc_rx_buffers_zc() 353 union i40e_rx_desc *rx_desc; in i40e_clean_rx_irq_zc() local 358 rx_desc = I40E_RX_DESC(rx_ring, rx_ring->next_to_clean); in i40e_clean_rx_irq_zc() 359 qword = le64_to_cpu(rx_desc->wb.qword1.status_error_len); in i40e_clean_rx_irq_zc() [all …]
|
D | i40e_txrx.c | 1555 union i40e_rx_desc *rx_desc; in i40e_alloc_rx_buffers() local 1562 rx_desc = I40E_RX_DESC(rx_ring, ntu); in i40e_alloc_rx_buffers() 1578 rx_desc->read.pkt_addr = cpu_to_le64(bi->dma + bi->page_offset); in i40e_alloc_rx_buffers() 1580 rx_desc++; in i40e_alloc_rx_buffers() 1584 rx_desc = I40E_RX_DESC(rx_ring, 0); in i40e_alloc_rx_buffers() 1590 rx_desc->wb.qword1.status_error_len = 0; in i40e_alloc_rx_buffers() 1618 union i40e_rx_desc *rx_desc) in i40e_rx_checksum() argument 1626 qword = le64_to_cpu(rx_desc->wb.qword1.status_error_len); in i40e_rx_checksum() 1732 union i40e_rx_desc *rx_desc, in i40e_rx_hash() argument 1744 if ((rx_desc->wb.qword1.status_error_len & rss_mask) == rss_mask) { in i40e_rx_hash() [all …]
|
/drivers/infiniband/ulp/iser/ |
D | iser_initiator.c | 244 struct iser_rx_desc *rx_desc; in iser_alloc_rx_descriptors() local 267 rx_desc = iser_conn->rx_descs; in iser_alloc_rx_descriptors() 269 for (i = 0; i < iser_conn->qp_max_recv_dtos; i++, rx_desc++) { in iser_alloc_rx_descriptors() 270 dma_addr = ib_dma_map_single(device->ib_device, (void *)rx_desc, in iser_alloc_rx_descriptors() 275 rx_desc->dma_addr = dma_addr; in iser_alloc_rx_descriptors() 276 rx_desc->cqe.done = iser_task_rsp; in iser_alloc_rx_descriptors() 277 rx_sg = &rx_desc->rx_sg; in iser_alloc_rx_descriptors() 278 rx_sg->addr = rx_desc->dma_addr; in iser_alloc_rx_descriptors() 287 rx_desc = iser_conn->rx_descs; in iser_alloc_rx_descriptors() 288 for (j = 0; j < i; j++, rx_desc++) in iser_alloc_rx_descriptors() [all …]
|
/drivers/net/ethernet/intel/ixgbe/ |
D | ixgbe_xsk.c | 146 union ixgbe_adv_rx_desc *rx_desc; in ixgbe_alloc_rx_buffers_zc() local 156 rx_desc = IXGBE_RX_DESC(rx_ring, i); in ixgbe_alloc_rx_buffers_zc() 172 rx_desc->read.pkt_addr = cpu_to_le64(dma); in ixgbe_alloc_rx_buffers_zc() 174 rx_desc++; in ixgbe_alloc_rx_buffers_zc() 178 rx_desc = IXGBE_RX_DESC(rx_ring, 0); in ixgbe_alloc_rx_buffers_zc() 184 rx_desc->wb.upper.length = 0; in ixgbe_alloc_rx_buffers_zc() 251 union ixgbe_adv_rx_desc *rx_desc; in ixgbe_clean_rx_irq_zc() local 263 rx_desc = IXGBE_RX_DESC(rx_ring, rx_ring->next_to_clean); in ixgbe_clean_rx_irq_zc() 264 size = le16_to_cpu(rx_desc->wb.upper.length); in ixgbe_clean_rx_irq_zc() 276 if (unlikely(!ixgbe_test_staterr(rx_desc, in ixgbe_clean_rx_irq_zc() [all …]
|
/drivers/spi/ |
D | spi-pxa2xx-dma.c | 143 struct dma_async_tx_descriptor *tx_desc, *rx_desc; in pxa2xx_spi_dma_prepare() local 154 rx_desc = pxa2xx_spi_dma_prepare_one(drv_data, DMA_DEV_TO_MEM, xfer); in pxa2xx_spi_dma_prepare() 155 if (!rx_desc) { in pxa2xx_spi_dma_prepare() 163 rx_desc->callback = pxa2xx_spi_dma_callback; in pxa2xx_spi_dma_prepare() 164 rx_desc->callback_param = drv_data; in pxa2xx_spi_dma_prepare() 166 dmaengine_submit(rx_desc); in pxa2xx_spi_dma_prepare()
|
/drivers/net/ethernet/hisilicon/hns3/ |
D | hns3_debugfs.c | 175 struct hns3_desc *rx_desc, *tx_desc; in hns3_dbg_bd_info() local 233 rx_desc = &ring->desc[rx_index]; in hns3_dbg_bd_info() 235 addr = le64_to_cpu(rx_desc->addr); in hns3_dbg_bd_info() 239 le32_to_cpu(rx_desc->rx.l234_info)); in hns3_dbg_bd_info() 240 dev_info(dev, "(RX)pkt_len: %u\n", le16_to_cpu(rx_desc->rx.pkt_len)); in hns3_dbg_bd_info() 241 dev_info(dev, "(RX)size: %u\n", le16_to_cpu(rx_desc->rx.size)); in hns3_dbg_bd_info() 242 dev_info(dev, "(RX)rss_hash: %u\n", le32_to_cpu(rx_desc->rx.rss_hash)); in hns3_dbg_bd_info() 243 dev_info(dev, "(RX)fd_id: %u\n", le16_to_cpu(rx_desc->rx.fd_id)); in hns3_dbg_bd_info() 244 dev_info(dev, "(RX)vlan_tag: %u\n", le16_to_cpu(rx_desc->rx.vlan_tag)); in hns3_dbg_bd_info() 246 le16_to_cpu(rx_desc->rx.o_dm_vlan_id_fb)); in hns3_dbg_bd_info() [all …]
|
/drivers/net/ethernet/intel/iavf/ |
D | iavf_txrx.c | 885 union iavf_rx_desc *rx_desc; in iavf_alloc_rx_buffers() local 892 rx_desc = IAVF_RX_DESC(rx_ring, ntu); in iavf_alloc_rx_buffers() 908 rx_desc->read.pkt_addr = cpu_to_le64(bi->dma + bi->page_offset); in iavf_alloc_rx_buffers() 910 rx_desc++; in iavf_alloc_rx_buffers() 914 rx_desc = IAVF_RX_DESC(rx_ring, 0); in iavf_alloc_rx_buffers() 920 rx_desc->wb.qword1.status_error_len = 0; in iavf_alloc_rx_buffers() 948 union iavf_rx_desc *rx_desc) in iavf_rx_checksum() argument 956 qword = le64_to_cpu(rx_desc->wb.qword1.status_error_len); in iavf_rx_checksum() 1055 union iavf_rx_desc *rx_desc, in iavf_rx_hash() argument 1067 if ((rx_desc->wb.qword1.status_error_len & rss_mask) == rss_mask) { in iavf_rx_hash() [all …]
|
/drivers/net/ethernet/alteon/ |
D | acenic.h | 477 #define RX_STD_RING_SIZE (RX_STD_RING_ENTRIES * sizeof(struct rx_desc)) 480 #define RX_JUMBO_RING_SIZE (RX_JUMBO_RING_ENTRIES *sizeof(struct rx_desc)) 483 #define RX_MINI_RING_SIZE (RX_MINI_RING_ENTRIES *sizeof(struct rx_desc)) 487 sizeof(struct rx_desc)) 489 struct rx_desc{ struct 665 struct rx_desc *rx_std_ring; 666 struct rx_desc *rx_jumbo_ring; 667 struct rx_desc *rx_mini_ring; 668 struct rx_desc *rx_return_ring;
|
/drivers/net/ethernet/ |
D | ec_bhf.c | 87 struct rx_desc { struct 139 struct rx_desc *rx_descs; 194 static int ec_bhf_pkt_received(struct rx_desc *desc) in ec_bhf_pkt_received() 199 static void ec_bhf_add_rx_desc(struct ec_bhf_priv *priv, struct rx_desc *desc) in ec_bhf_add_rx_desc() 207 struct rx_desc *desc = &priv->rx_descs[priv->rx_dnext]; in ec_bhf_process_rx() 373 priv->rx_dcount = priv->rx_buf.len / sizeof(struct rx_desc); in ec_bhf_setup_rx_descs() 374 priv->rx_descs = (struct rx_desc *)priv->rx_buf.buf; in ec_bhf_setup_rx_descs() 378 struct rx_desc *desc = &priv->rx_descs[i]; in ec_bhf_setup_rx_descs() 401 FIFO_SIZE * sizeof(struct rx_desc)); in ec_bhf_open()
|
/drivers/net/wireless/ti/wl1251/ |
D | rx.c | 206 struct wl1251_rx_descriptor *rx_desc; in wl1251_rx() local 211 rx_desc = wl->rx_descriptor; in wl1251_rx() 214 wl1251_rx_header(wl, rx_desc); in wl1251_rx() 217 wl1251_rx_body(wl, rx_desc); in wl1251_rx()
|
/drivers/net/ethernet/google/gve/ |
D | gve_rx.c | 281 static bool gve_rx(struct gve_rx_ring *rx, struct gve_rx_desc *rx_desc, in gve_rx() argument 293 if (unlikely(rx_desc->flags_seq & GVE_RXF_ERR)) { in gve_rx() 300 len = be16_to_cpu(rx_desc->len) - GVE_RX_PAD; in gve_rx() 366 if (rx_desc->csum) in gve_rx() 370 skb->csum = csum_unfold(rx_desc->csum); in gve_rx() 375 gve_needs_rss(rx_desc->flags_seq)) in gve_rx() 376 skb_set_hash(skb, be32_to_cpu(rx_desc->rss_hash), in gve_rx() 377 gve_rss_type(rx_desc->flags_seq)); in gve_rx()
|
/drivers/net/ethernet/marvell/ |
D | pxa168_eth.c | 175 struct rx_desc { struct 207 struct rx_desc *p_rx_desc_area; 307 struct rx_desc *p_used_rx_desc; in rxq_refill() 643 (u32) (pep->rx_desc_dma + rx_curr_desc * sizeof(struct rx_desc))); in eth_port_start() 646 (u32) (pep->rx_desc_dma + rx_curr_desc * sizeof(struct rx_desc))); in eth_port_start() 773 struct rx_desc *rx_desc; in rxq_process() local 781 rx_desc = &pep->p_rx_desc_area[rx_curr_desc]; in rxq_process() 782 cmd_sts = rx_desc->cmd_sts; in rxq_process() 797 dma_unmap_single(&pep->pdev->dev, rx_desc->buf_ptr, in rxq_process() 798 rx_desc->buf_size, in rxq_process() [all …]
|
D | mvneta.c | 884 int rx_desc = rxq->next_desc_to_proc; in mvneta_rxq_next_desc_get() local 886 rxq->next_desc_to_proc = MVNETA_QUEUE_NEXT_DESC(rxq, rx_desc); in mvneta_rxq_next_desc_get() 888 return rxq->descs + rx_desc; in mvneta_rxq_next_desc_get() 1682 static void mvneta_rx_desc_fill(struct mvneta_rx_desc *rx_desc, in mvneta_rx_desc_fill() argument 1688 rx_desc->buf_phys_addr = phys_addr; in mvneta_rx_desc_fill() 1689 i = rx_desc - rxq->descs; in mvneta_rx_desc_fill() 1774 struct mvneta_rx_desc *rx_desc) in mvneta_rx_error() argument 1777 u32 status = rx_desc->status; in mvneta_rx_error() 1787 status, rx_desc->data_size); in mvneta_rx_error() 1791 status, rx_desc->data_size); in mvneta_rx_error() [all …]
|
/drivers/staging/mt7621-dma/ |
D | hsdma-mt7621.c | 235 struct hsdma_desc *rx_desc; in hsdma_dump_desc() local 243 rx_desc = &chan->rx_ring[i]; in hsdma_dump_desc() 248 tx_desc->addr1, rx_desc->addr0, rx_desc->flags); in hsdma_dump_desc() 319 struct hsdma_desc *tx_desc, *rx_desc; in mtk_hsdma_start_transfer() local 362 rx_desc = &chan->rx_ring[rx_idx]; in mtk_hsdma_start_transfer() 368 rx_desc->addr0 = dst; in mtk_hsdma_start_transfer() 369 rx_desc->flags = HSDMA_DESC_PLEN0(tlen); in mtk_hsdma_start_transfer()
|
/drivers/net/ethernet/intel/ixgbevf/ |
D | ixgbevf_main.c | 442 union ixgbe_adv_rx_desc *rx_desc, in ixgbevf_rx_hash() argument 450 rss_type = le16_to_cpu(rx_desc->wb.lower.lo_dword.hs_rss.pkt_info) & in ixgbevf_rx_hash() 456 skb_set_hash(skb, le32_to_cpu(rx_desc->wb.lower.hi_dword.rss), in ixgbevf_rx_hash() 468 union ixgbe_adv_rx_desc *rx_desc, in ixgbevf_rx_checksum() argument 478 if (ixgbevf_test_staterr(rx_desc, IXGBE_RXD_STAT_IPCS) && in ixgbevf_rx_checksum() 479 ixgbevf_test_staterr(rx_desc, IXGBE_RXDADV_ERR_IPE)) { in ixgbevf_rx_checksum() 484 if (!ixgbevf_test_staterr(rx_desc, IXGBE_RXD_STAT_L4CS)) in ixgbevf_rx_checksum() 487 if (ixgbevf_test_staterr(rx_desc, IXGBE_RXDADV_ERR_TCPE)) { in ixgbevf_rx_checksum() 507 union ixgbe_adv_rx_desc *rx_desc, in ixgbevf_process_skb_fields() argument 510 ixgbevf_rx_hash(rx_ring, rx_desc, skb); in ixgbevf_process_skb_fields() [all …]
|
/drivers/net/wireless/ti/wlcore/ |
D | hw_ops.h | 45 wlcore_hw_get_rx_buf_align(struct wl1271 *wl, u32 rx_desc) in wlcore_hw_get_rx_buf_align() argument 51 return wl->ops->get_rx_buf_align(wl, rx_desc); in wlcore_hw_get_rx_buf_align() 55 wlcore_hw_prepare_read(struct wl1271 *wl, u32 rx_desc, u32 len) in wlcore_hw_prepare_read() argument 58 return wl->ops->prepare_read(wl, rx_desc, len); in wlcore_hw_prepare_read()
|
/drivers/net/ethernet/intel/igc/ |
D | igc_dump.c | 117 union igc_adv_rx_desc *rx_desc; in igc_rings_dump() local 260 rx_desc = IGC_RX_DESC(rx_ring, i); in igc_rings_dump() 261 u0 = (struct my_u0 *)rx_desc; in igc_rings_dump() 262 staterr = le32_to_cpu(rx_desc->wb.upper.status_error); in igc_rings_dump()
|