/drivers/net/ethernet/intel/fm10k/ |
D | fm10k_main.c | 129 union fm10k_rx_desc *rx_desc; in fm10k_alloc_rx_buffers() local 137 rx_desc = FM10K_RX_DESC(rx_ring, i); in fm10k_alloc_rx_buffers() 148 rx_desc->q.pkt_addr = cpu_to_le64(bi->dma + bi->page_offset); in fm10k_alloc_rx_buffers() 150 rx_desc++; in fm10k_alloc_rx_buffers() 154 rx_desc = FM10K_RX_DESC(rx_ring, 0); in fm10k_alloc_rx_buffers() 160 rx_desc->d.staterr = 0; in fm10k_alloc_rx_buffers() 266 union fm10k_rx_desc *rx_desc, in fm10k_add_rx_frag() argument 271 unsigned int size = le16_to_cpu(rx_desc->w.length); in fm10k_add_rx_frag() 314 union fm10k_rx_desc *rx_desc, in fm10k_fetch_rx_buffer() argument 357 if (fm10k_add_rx_frag(rx_buffer, rx_desc, skb)) { in fm10k_fetch_rx_buffer() [all …]
|
/drivers/net/ethernet/seeq/ |
D | sgiseeq.c | 58 (unsigned long)((sp)->rx_desc))) 94 struct sgiseeq_rx_desc *rx_desc; member 197 if (!sp->rx_desc[i].skb) { in seeq_init_ring() 207 sp->rx_desc[i].skb = skb; in seeq_init_ring() 208 sp->rx_desc[i].rdma.pbuf = dma_addr; in seeq_init_ring() 210 sp->rx_desc[i].rdma.cntinfo = RCNTINFO_INIT; in seeq_init_ring() 211 dma_sync_desc_dev(dev, &sp->rx_desc[i]); in seeq_init_ring() 213 sp->rx_desc[i - 1].rdma.cntinfo |= HPCDMA_EOR; in seeq_init_ring() 214 dma_sync_desc_dev(dev, &sp->rx_desc[i - 1]); in seeq_init_ring() 233 if (sp->rx_desc[i].skb) { in seeq_purge_ring() [all …]
|
/drivers/infiniband/ulp/isert/ |
D | ib_isert.c | 179 struct iser_rx_desc *rx_desc; in isert_alloc_rx_descriptors() local 189 rx_desc = isert_conn->rx_descs; in isert_alloc_rx_descriptors() 191 for (i = 0; i < ISERT_QP_MAX_RECV_DTOS; i++, rx_desc++) { in isert_alloc_rx_descriptors() 192 dma_addr = ib_dma_map_single(ib_dev, (void *)rx_desc, in isert_alloc_rx_descriptors() 197 rx_desc->dma_addr = dma_addr; in isert_alloc_rx_descriptors() 199 rx_sg = &rx_desc->rx_sg; in isert_alloc_rx_descriptors() 200 rx_sg->addr = rx_desc->dma_addr; in isert_alloc_rx_descriptors() 203 rx_desc->rx_cqe.done = isert_recv_done; in isert_alloc_rx_descriptors() 209 rx_desc = isert_conn->rx_descs; in isert_alloc_rx_descriptors() 210 for (j = 0; j < i; j++, rx_desc++) { in isert_alloc_rx_descriptors() [all …]
|
/drivers/net/ethernet/intel/i40e/ |
D | i40e_txrx.c | 478 union i40e_rx_desc *rx_desc, u8 prog_id) in i40e_fd_handle_status() argument 486 qw = le64_to_cpu(rx_desc->wb.qword1.status_error_len); in i40e_fd_handle_status() 491 pf->fd_inv = le32_to_cpu(rx_desc->wb.qword0.hi_dword.fd_id); in i40e_fd_handle_status() 492 if ((rx_desc->wb.qword0.hi_dword.fd_id != 0) || in i40e_fd_handle_status() 510 if ((rx_desc->wb.qword0.hi_dword.fd_id == 0) && in i40e_fd_handle_status() 536 rx_desc->wb.qword0.hi_dword.fd_id); in i40e_fd_handle_status() 967 union i40e_rx_desc *rx_desc) in i40e_clean_programming_status() argument 972 qw = le64_to_cpu(rx_desc->wb.qword1.status_error_len); in i40e_clean_programming_status() 977 i40e_fd_handle_status(rx_ring, rx_desc, id); in i40e_clean_programming_status() 981 i40e_fcoe_handle_status(rx_ring, rx_desc, id); in i40e_clean_programming_status() [all …]
|
/drivers/net/ethernet/intel/i40evf/ |
D | i40e_txrx.c | 690 union i40e_rx_desc *rx_desc; in i40evf_alloc_rx_buffers() local 697 rx_desc = I40E_RX_DESC(rx_ring, ntu); in i40evf_alloc_rx_buffers() 707 rx_desc->read.pkt_addr = cpu_to_le64(bi->dma + bi->page_offset); in i40evf_alloc_rx_buffers() 708 rx_desc->read.hdr_addr = 0; in i40evf_alloc_rx_buffers() 710 rx_desc++; in i40evf_alloc_rx_buffers() 714 rx_desc = I40E_RX_DESC(rx_ring, 0); in i40evf_alloc_rx_buffers() 720 rx_desc->wb.qword1.status_error_len = 0; in i40evf_alloc_rx_buffers() 750 union i40e_rx_desc *rx_desc) in i40e_rx_checksum() argument 758 qword = le64_to_cpu(rx_desc->wb.qword1.status_error_len); in i40e_rx_checksum() 862 union i40e_rx_desc *rx_desc, in i40e_rx_hash() argument [all …]
|
/drivers/spi/ |
D | spi-pxa2xx-dma.c | 149 struct dma_async_tx_descriptor *tx_desc, *rx_desc; in pxa2xx_spi_dma_prepare() local 160 rx_desc = pxa2xx_spi_dma_prepare_one(drv_data, DMA_DEV_TO_MEM); in pxa2xx_spi_dma_prepare() 161 if (!rx_desc) { in pxa2xx_spi_dma_prepare() 169 rx_desc->callback = pxa2xx_spi_dma_callback; in pxa2xx_spi_dma_prepare() 170 rx_desc->callback_param = drv_data; in pxa2xx_spi_dma_prepare() 172 dmaengine_submit(rx_desc); in pxa2xx_spi_dma_prepare()
|
/drivers/infiniband/ulp/iser/ |
D | iser_initiator.c | 243 struct iser_rx_desc *rx_desc; in iser_alloc_rx_descriptors() local 265 rx_desc = iser_conn->rx_descs; in iser_alloc_rx_descriptors() 267 for (i = 0; i < iser_conn->qp_max_recv_dtos; i++, rx_desc++) { in iser_alloc_rx_descriptors() 268 dma_addr = ib_dma_map_single(device->ib_device, (void *)rx_desc, in iser_alloc_rx_descriptors() 273 rx_desc->dma_addr = dma_addr; in iser_alloc_rx_descriptors() 274 rx_desc->cqe.done = iser_task_rsp; in iser_alloc_rx_descriptors() 275 rx_sg = &rx_desc->rx_sg; in iser_alloc_rx_descriptors() 276 rx_sg->addr = rx_desc->dma_addr; in iser_alloc_rx_descriptors() 285 rx_desc = iser_conn->rx_descs; in iser_alloc_rx_descriptors() 286 for (j = 0; j < i; j++, rx_desc++) in iser_alloc_rx_descriptors() [all …]
|
/drivers/net/ethernet/ibm/emac/ |
D | debug.c | 58 i, p->rx_desc[i].data_ptr, p->rx_skb[i] ? 'V' : ' ', in emac_desc_dump() 59 p->rx_desc[i].ctrl, p->rx_desc[i].data_len, in emac_desc_dump() 61 p->rx_desc[NUM_RX_BUFF / 2 + i].data_ptr, in emac_desc_dump() 63 p->rx_desc[NUM_RX_BUFF / 2 + i].ctrl, in emac_desc_dump() 64 p->rx_desc[NUM_RX_BUFF / 2 + i].data_len); in emac_desc_dump()
|
D | core.c | 1054 if (dev->rx_desc[i].ctrl & MAL_RX_CTRL_FIRST) in emac_resize_rx_ring() 1057 dev->rx_desc[i].data_len = 0; in emac_resize_rx_ring() 1058 dev->rx_desc[i].ctrl = MAL_RX_CTRL_EMPTY | in emac_resize_rx_ring() 1078 dev->rx_desc[i].data_ptr = in emac_resize_rx_ring() 1162 dev->rx_desc[i].ctrl = 0; in emac_clean_rx_ring() 1165 dev->rx_desc[i].data_ptr = 0; in emac_clean_rx_ring() 1182 dev->rx_desc[slot].data_len = 0; in emac_alloc_rx_skb() 1185 dev->rx_desc[slot].data_ptr = in emac_alloc_rx_skb() 1189 dev->rx_desc[slot].ctrl = MAL_RX_CTRL_EMPTY | in emac_alloc_rx_skb() 1662 dev->rx_desc[slot].data_len = 0; in emac_recycle_rx_skb() [all …]
|
/drivers/net/ethernet/alteon/ |
D | acenic.h | 476 #define RX_STD_RING_SIZE (RX_STD_RING_ENTRIES * sizeof(struct rx_desc)) 479 #define RX_JUMBO_RING_SIZE (RX_JUMBO_RING_ENTRIES *sizeof(struct rx_desc)) 482 #define RX_MINI_RING_SIZE (RX_MINI_RING_ENTRIES *sizeof(struct rx_desc)) 486 sizeof(struct rx_desc)) 488 struct rx_desc{ struct 663 struct rx_desc *rx_std_ring; 664 struct rx_desc *rx_jumbo_ring; 665 struct rx_desc *rx_mini_ring; 666 struct rx_desc *rx_return_ring;
|
/drivers/net/ethernet/mellanox/mlx4/ |
D | en_rx.c | 91 struct mlx4_en_rx_desc *rx_desc, in mlx4_en_alloc_frags() argument 120 rx_desc->data[i].addr = cpu_to_be64(dma); in mlx4_en_alloc_frags() 221 struct mlx4_en_rx_desc *rx_desc = ring->buf + ring->stride * index; in mlx4_en_init_rx_desc() local 227 rx_desc->data[i].byte_count = in mlx4_en_init_rx_desc() 229 rx_desc->data[i].lkey = cpu_to_be32(priv->mdev->mr.key); in mlx4_en_init_rx_desc() 237 rx_desc->data[i].byte_count = 0; in mlx4_en_init_rx_desc() 238 rx_desc->data[i].lkey = cpu_to_be32(MLX4_EN_MEMTYPE_PAD); in mlx4_en_init_rx_desc() 239 rx_desc->data[i].addr = 0; in mlx4_en_init_rx_desc() 247 struct mlx4_en_rx_desc *rx_desc = ring->buf + (index * ring->stride); in mlx4_en_prepare_rx_desc() local 253 rx_desc->data[0].addr = cpu_to_be64(frags[0].dma); in mlx4_en_prepare_rx_desc() [all …]
|
/drivers/net/ethernet/ |
D | ec_bhf.c | 96 struct rx_desc { struct 148 struct rx_desc *rx_descs; 203 static int ec_bhf_pkt_received(struct rx_desc *desc) in ec_bhf_pkt_received() 208 static void ec_bhf_add_rx_desc(struct ec_bhf_priv *priv, struct rx_desc *desc) in ec_bhf_add_rx_desc() 216 struct rx_desc *desc = &priv->rx_descs[priv->rx_dnext]; in ec_bhf_process_rx() 382 priv->rx_dcount = priv->rx_buf.len / sizeof(struct rx_desc); in ec_bhf_setup_rx_descs() 383 priv->rx_descs = (struct rx_desc *)priv->rx_buf.buf; in ec_bhf_setup_rx_descs() 387 struct rx_desc *desc = &priv->rx_descs[i]; in ec_bhf_setup_rx_descs() 410 FIFO_SIZE * sizeof(struct rx_desc)); in ec_bhf_open()
|
/drivers/net/wireless/ti/wl1251/ |
D | rx.c | 220 struct wl1251_rx_descriptor *rx_desc; in wl1251_rx() local 225 rx_desc = wl->rx_descriptor; in wl1251_rx() 228 wl1251_rx_header(wl, rx_desc); in wl1251_rx() 231 wl1251_rx_body(wl, rx_desc); in wl1251_rx()
|
/drivers/net/ethernet/marvell/ |
D | pxa168_eth.c | 187 struct rx_desc { struct 218 struct rx_desc *p_rx_desc_area; 320 struct rx_desc *p_used_rx_desc; in rxq_refill() 656 (u32) (pep->rx_desc_dma + rx_curr_desc * sizeof(struct rx_desc))); in eth_port_start() 659 (u32) (pep->rx_desc_dma + rx_curr_desc * sizeof(struct rx_desc))); in eth_port_start() 786 struct rx_desc *rx_desc; in rxq_process() local 794 rx_desc = &pep->p_rx_desc_area[rx_curr_desc]; in rxq_process() 795 cmd_sts = rx_desc->cmd_sts; in rxq_process() 810 dma_unmap_single(NULL, rx_desc->buf_ptr, in rxq_process() 811 rx_desc->buf_size, in rxq_process() [all …]
|
D | mvneta.c | 760 int rx_desc = rxq->next_desc_to_proc; in mvneta_rxq_next_desc_get() local 762 rxq->next_desc_to_proc = MVNETA_QUEUE_NEXT_DESC(rxq, rx_desc); in mvneta_rxq_next_desc_get() 764 return rxq->descs + rx_desc; in mvneta_rxq_next_desc_get() 1580 static void mvneta_rx_desc_fill(struct mvneta_rx_desc *rx_desc, in mvneta_rx_desc_fill() argument 1583 rx_desc->buf_cookie = cookie; in mvneta_rx_desc_fill() 1584 rx_desc->buf_phys_addr = phys_addr; in mvneta_rx_desc_fill() 1668 struct mvneta_rx_desc *rx_desc) in mvneta_rx_error() argument 1670 u32 status = rx_desc->status; in mvneta_rx_error() 1675 status, rx_desc->data_size); in mvneta_rx_error() 1682 status, rx_desc->data_size); in mvneta_rx_error() [all …]
|
D | mv643xx_eth.c | 203 struct rx_desc { struct 219 struct rx_desc { argument 347 struct rx_desc *rx_desc_area; 529 struct rx_desc *rx_desc; in rxq_process() local 534 rx_desc = &rxq->rx_desc_area[rxq->rx_curr_desc]; in rxq_process() 536 cmd_sts = rx_desc->cmd_sts; in rxq_process() 548 dma_unmap_single(mp->dev->dev.parent, rx_desc->buf_ptr, in rxq_process() 549 rx_desc->buf_size, DMA_FROM_DEVICE); in rxq_process() 555 byte_cnt = rx_desc->byte_cnt; in rxq_process() 622 struct rx_desc *rx_desc; in rxq_refill() local [all …]
|
/drivers/net/ethernet/intel/ixgbevf/ |
D | ixgbevf_main.c | 480 union ixgbe_adv_rx_desc *rx_desc, in ixgbevf_rx_hash() argument 488 rss_type = le16_to_cpu(rx_desc->wb.lower.lo_dword.hs_rss.pkt_info) & in ixgbevf_rx_hash() 494 skb_set_hash(skb, le32_to_cpu(rx_desc->wb.lower.hi_dword.rss), in ixgbevf_rx_hash() 506 union ixgbe_adv_rx_desc *rx_desc, in ixgbevf_rx_checksum() argument 516 if (ixgbevf_test_staterr(rx_desc, IXGBE_RXD_STAT_IPCS) && in ixgbevf_rx_checksum() 517 ixgbevf_test_staterr(rx_desc, IXGBE_RXDADV_ERR_IPE)) { in ixgbevf_rx_checksum() 522 if (!ixgbevf_test_staterr(rx_desc, IXGBE_RXD_STAT_L4CS)) in ixgbevf_rx_checksum() 525 if (ixgbevf_test_staterr(rx_desc, IXGBE_RXDADV_ERR_TCPE)) { in ixgbevf_rx_checksum() 545 union ixgbe_adv_rx_desc *rx_desc, in ixgbevf_process_skb_fields() argument 548 ixgbevf_rx_hash(rx_ring, rx_desc, skb); in ixgbevf_process_skb_fields() [all …]
|
/drivers/net/wireless/ti/wlcore/ |
D | hw_ops.h | 59 wlcore_hw_get_rx_buf_align(struct wl1271 *wl, u32 rx_desc) in wlcore_hw_get_rx_buf_align() argument 65 return wl->ops->get_rx_buf_align(wl, rx_desc); in wlcore_hw_get_rx_buf_align() 69 wlcore_hw_prepare_read(struct wl1271 *wl, u32 rx_desc, u32 len) in wlcore_hw_prepare_read() argument 72 return wl->ops->prepare_read(wl, rx_desc, len); in wlcore_hw_prepare_read()
|
/drivers/atm/ |
D | horizon.c | 727 rx_ch_desc * rx_desc = &memmap->rx_descs[channel]; in hrz_open_rx() local 732 channel_type = rd_mem (dev, &rx_desc->wr_buf_type) & BUFFER_PTR_MASK; in hrz_open_rx() 760 wr_mem (dev, &rx_desc->wr_buf_type, in hrz_open_rx() 763 wr_mem (dev, &rx_desc->rd_buf_type, buf_ptr); in hrz_open_rx() 801 rx_ch_desc * rx_desc = &memmap->rx_descs[vc]; in hrz_close_rx() local 806 value = rd_mem (dev, &rx_desc->wr_buf_type) & BUFFER_PTR_MASK; in hrz_close_rx() 820 wr_mem (dev, &rx_desc->wr_buf_type, RX_CHANNEL_DISABLED); in hrz_close_rx() 822 if ((rd_mem (dev, &rx_desc->wr_buf_type) & BUFFER_PTR_MASK) == RX_CHANNEL_DISABLED) in hrz_close_rx() 866 r1 = rd_mem (dev, &rx_desc->rd_buf_type); in hrz_close_rx() 884 r2 = rd_mem (dev, &rx_desc->rd_buf_type); in hrz_close_rx() [all …]
|
/drivers/net/ethernet/intel/ixgb/ |
D | ixgb_main.c | 1938 struct ixgb_rx_desc *rx_desc, in ixgb_rx_checksum() argument 1944 if ((rx_desc->status & IXGB_RX_DESC_STATUS_IXSM) || in ixgb_rx_checksum() 1945 (!(rx_desc->status & IXGB_RX_DESC_STATUS_TCPCS))) { in ixgb_rx_checksum() 1952 if (rx_desc->errors & IXGB_RX_DESC_ERRORS_TCPE) { in ixgb_rx_checksum() 1999 struct ixgb_rx_desc *rx_desc, *next_rxd; in ixgb_clean_rx_irq() local 2007 rx_desc = IXGB_RX_DESC(*rx_ring, i); in ixgb_clean_rx_irq() 2010 while (rx_desc->status & IXGB_RX_DESC_STATUS_DD) { in ixgb_clean_rx_irq() 2019 status = rx_desc->status; in ixgb_clean_rx_irq() 2047 length = le16_to_cpu(rx_desc->length); in ixgb_clean_rx_irq() 2048 rx_desc->length = 0; in ixgb_clean_rx_irq() [all …]
|
/drivers/net/ethernet/intel/e1000/ |
D | e1000_main.c | 3469 struct e1000_rx_desc *rx_desc = E1000_RX_DESC(*rx_ring, i); in e1000_dump() local 3472 struct my_u *u = (struct my_u *)rx_desc; in e1000_dump() 4167 struct e1000_rx_desc *rx_desc, *next_rxd; in e1000_clean_jumbo_rx_irq() local 4176 rx_desc = E1000_RX_DESC(*rx_ring, i); in e1000_clean_jumbo_rx_irq() 4179 while (rx_desc->status & E1000_RXD_STAT_DD) { in e1000_clean_jumbo_rx_irq() 4188 status = rx_desc->status; in e1000_clean_jumbo_rx_irq() 4204 length = le16_to_cpu(rx_desc->length); in e1000_clean_jumbo_rx_irq() 4208 (rx_desc->errors & E1000_RXD_ERR_FRAME_ERR_MASK))) { in e1000_clean_jumbo_rx_irq() 4212 rx_desc->errors, in e1000_clean_jumbo_rx_irq() 4283 status | rx_desc->errors << 24, in e1000_clean_jumbo_rx_irq() [all …]
|
/drivers/scsi/mvsas/ |
D | mv_sas.c | 886 static void mvs_slot_free(struct mvs_info *mvi, u32 rx_desc) in mvs_slot_free() argument 888 u32 slot_idx = rx_desc & RXQ_SLOT_MASK; in mvs_slot_free() 1717 int mvs_slot_complete(struct mvs_info *mvi, u32 rx_desc, u32 flags) in mvs_slot_complete() argument 1719 u32 slot_idx = rx_desc & RXQ_SLOT_MASK; in mvs_slot_complete() 1771 if (unlikely((rx_desc & RXQ_ERR) in mvs_slot_complete() 1776 rx_desc, get_unaligned_le64(slot->response)); in mvs_slot_complete() 1785 if (rx_desc & RXQ_GOOD) { in mvs_slot_complete() 1790 else if (rx_desc & RXQ_RSP) { in mvs_slot_complete() 2068 u32 rx_prod_idx, rx_desc; in mvs_int_rx() local 2094 rx_desc = le32_to_cpu(mvi->rx[rx_prod_idx + 1]); in mvs_int_rx() [all …]
|
/drivers/net/ethernet/intel/e1000e/ |
D | netdev.c | 227 union e1000_rx_desc_extended *rx_desc; in e1000e_dump() local 447 rx_desc = E1000_RX_DESC_EXT(*rx_ring, i); in e1000e_dump() 448 u1 = (struct my_u1 *)rx_desc; in e1000e_dump() 449 staterr = le32_to_cpu(rx_desc->wb.upper.status_error); in e1000e_dump() 667 union e1000_rx_desc_extended *rx_desc; in e1000_alloc_rx_buffers() local 701 rx_desc = E1000_RX_DESC_EXT(*rx_ring, i); in e1000_alloc_rx_buffers() 702 rx_desc->read.buffer_addr = cpu_to_le64(buffer_info->dma); in e1000_alloc_rx_buffers() 735 union e1000_rx_desc_packet_split *rx_desc; in e1000_alloc_rx_buffers_ps() local 745 rx_desc = E1000_RX_DESC_PS(*rx_ring, i); in e1000_alloc_rx_buffers_ps() 751 rx_desc->read.buffer_addr[j + 1] = in e1000_alloc_rx_buffers_ps() [all …]
|
/drivers/net/ethernet/intel/ixgbe/ |
D | ixgbe.h | 555 static inline __le32 ixgbe_test_staterr(union ixgbe_adv_rx_desc *rx_desc, in ixgbe_test_staterr() argument 558 return rx_desc->wb.upper.status_error & cpu_to_le32(stat_err_bits); in ixgbe_test_staterr() 954 union ixgbe_adv_rx_desc *rx_desc, struct sk_buff *skb); 997 union ixgbe_adv_rx_desc *rx_desc, in ixgbe_ptp_rx_hwtstamp() argument 1000 if (unlikely(ixgbe_test_staterr(rx_desc, IXGBE_RXD_STAT_TSIP))) { in ixgbe_ptp_rx_hwtstamp() 1005 if (unlikely(!ixgbe_test_staterr(rx_desc, IXGBE_RXDADV_STAT_TS))) in ixgbe_ptp_rx_hwtstamp()
|
/drivers/net/ethernet/intel/igb/ |
D | igb_main.c | 376 union e1000_adv_rx_desc *rx_desc; in igb_dump() local 524 rx_desc = IGB_RX_DESC(rx_ring, i); in igb_dump() 525 u0 = (struct my_u0 *)rx_desc; in igb_dump() 526 staterr = le32_to_cpu(rx_desc->wb.upper.status_error); in igb_dump() 6888 union e1000_adv_rx_desc *rx_desc, in igb_add_rx_frag() argument 6903 if (unlikely(igb_test_staterr(rx_desc, E1000_RXDADV_STAT_TSIP))) { in igb_add_rx_frag() 6941 union e1000_adv_rx_desc *rx_desc, in igb_fetch_rx_buffer() argument 6944 unsigned int size = le16_to_cpu(rx_desc->wb.upper.length); in igb_fetch_rx_buffer() 6984 if (igb_add_rx_frag(rx_ring, rx_buffer, size, rx_desc, skb)) { in igb_fetch_rx_buffer() 7000 union e1000_adv_rx_desc *rx_desc, in igb_rx_checksum() argument [all …]
|