| /kernel/linux/linux-5.10/drivers/infiniband/ulp/isert/ |
| D | ib_isert.c | 782 isert_login_post_send(struct isert_conn *isert_conn, struct iser_tx_desc *tx_desc) in isert_login_post_send() argument 788 ib_dma_sync_single_for_device(ib_dev, tx_desc->dma_addr, in isert_login_post_send() 791 tx_desc->tx_cqe.done = isert_login_send_done; in isert_login_post_send() 794 send_wr.wr_cqe = &tx_desc->tx_cqe; in isert_login_post_send() 795 send_wr.sg_list = tx_desc->tx_sg; in isert_login_post_send() 796 send_wr.num_sge = tx_desc->num_sge; in isert_login_post_send() 809 struct iser_tx_desc *tx_desc) in __isert_create_send_desc() argument 812 memset(&tx_desc->iser_header, 0, sizeof(struct iser_ctrl)); in __isert_create_send_desc() 813 tx_desc->iser_header.flags = ISCSI_CTRL; in __isert_create_send_desc() 815 tx_desc->num_sge = 1; in __isert_create_send_desc() [all …]
|
| /kernel/linux/linux-6.6/drivers/infiniband/ulp/isert/ |
| D | ib_isert.c | 783 isert_login_post_send(struct isert_conn *isert_conn, struct iser_tx_desc *tx_desc) in isert_login_post_send() argument 789 ib_dma_sync_single_for_device(ib_dev, tx_desc->dma_addr, in isert_login_post_send() 792 tx_desc->tx_cqe.done = isert_login_send_done; in isert_login_post_send() 795 send_wr.wr_cqe = &tx_desc->tx_cqe; in isert_login_post_send() 796 send_wr.sg_list = tx_desc->tx_sg; in isert_login_post_send() 797 send_wr.num_sge = tx_desc->num_sge; in isert_login_post_send() 810 struct iser_tx_desc *tx_desc) in __isert_create_send_desc() argument 813 memset(&tx_desc->iser_header, 0, sizeof(struct iser_ctrl)); in __isert_create_send_desc() 814 tx_desc->iser_header.flags = ISCSI_CTRL; in __isert_create_send_desc() 816 tx_desc->num_sge = 1; in __isert_create_send_desc() [all …]
|
| /kernel/linux/linux-6.6/drivers/crypto/ccp/ |
| D | ccp-dmaengine.c | 111 if (!async_tx_test_ack(&desc->tx_desc)) in ccp_cleanup_desc_resources() 148 desc->tx_desc.cookie, cmd); in ccp_issue_next_cmd() 155 ret, desc->tx_desc.cookie, cmd); in ccp_issue_next_cmd() 170 __func__, desc->tx_desc.cookie, cmd); in ccp_free_active_cmd() 193 struct dma_async_tx_descriptor *tx_desc; in ccp_handle_active_desc() local 212 tx_desc = &desc->tx_desc; in ccp_handle_active_desc() 214 tx_desc = NULL; in ccp_handle_active_desc() 225 desc->tx_desc.cookie, desc->status); in ccp_handle_active_desc() 227 dma_cookie_complete(tx_desc); in ccp_handle_active_desc() 228 dma_descriptor_unmap(tx_desc); in ccp_handle_active_desc() [all …]
|
| /kernel/linux/linux-5.10/drivers/crypto/ccp/ |
| D | ccp-dmaengine.c | 111 if (!async_tx_test_ack(&desc->tx_desc)) in ccp_cleanup_desc_resources() 148 desc->tx_desc.cookie, cmd); in ccp_issue_next_cmd() 155 ret, desc->tx_desc.cookie, cmd); in ccp_issue_next_cmd() 170 __func__, desc->tx_desc.cookie, cmd); in ccp_free_active_cmd() 193 struct dma_async_tx_descriptor *tx_desc; in ccp_handle_active_desc() local 212 tx_desc = &desc->tx_desc; in ccp_handle_active_desc() 214 tx_desc = NULL; in ccp_handle_active_desc() 225 desc->tx_desc.cookie, desc->status); in ccp_handle_active_desc() 227 dma_cookie_complete(tx_desc); in ccp_handle_active_desc() 228 dma_descriptor_unmap(tx_desc); in ccp_handle_active_desc() [all …]
|
| /kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx4/ |
| D | en_tx.c | 241 struct mlx4_en_tx_desc *tx_desc = ring->buf + (index << LOG_TXBB_SIZE); in mlx4_en_stamp_wqe() local 244 __be32 *ptr = (__be32 *)tx_desc; in mlx4_en_stamp_wqe() 248 if (likely((void *)tx_desc + in mlx4_en_stamp_wqe() 281 struct mlx4_en_tx_desc *tx_desc = ring->buf + (index << LOG_TXBB_SIZE); in mlx4_en_free_tx_desc() local 282 struct mlx4_wqe_data_seg *data = (void *) tx_desc + tx_info->data_offset; in mlx4_en_free_tx_desc() 312 if (likely((void *)tx_desc + in mlx4_en_free_tx_desc() 405 struct mlx4_en_tx_desc *tx_desc; in mlx4_en_handle_err_cqe() local 419 tx_desc = ring->buf + (wqe_index << LOG_TXBB_SIZE); in mlx4_en_handle_err_cqe() 420 print_hex_dump(KERN_WARNING, "", DUMP_PREFIX_OFFSET, 16, 1, tx_desc, desc_size, false); in mlx4_en_handle_err_cqe() 690 static void build_inline_wqe(struct mlx4_en_tx_desc *tx_desc, in build_inline_wqe() argument [all …]
|
| /kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx4/ |
| D | en_tx.c | 236 struct mlx4_en_tx_desc *tx_desc = ring->buf + (index << LOG_TXBB_SIZE); in mlx4_en_stamp_wqe() local 239 __be32 *ptr = (__be32 *)tx_desc; in mlx4_en_stamp_wqe() 243 if (likely((void *)tx_desc + in mlx4_en_stamp_wqe() 276 struct mlx4_en_tx_desc *tx_desc = ring->buf + (index << LOG_TXBB_SIZE); in mlx4_en_free_tx_desc() local 277 struct mlx4_wqe_data_seg *data = (void *) tx_desc + tx_info->data_offset; in mlx4_en_free_tx_desc() 307 if (likely((void *)tx_desc + in mlx4_en_free_tx_desc() 400 struct mlx4_en_tx_desc *tx_desc; in mlx4_en_handle_err_cqe() local 414 tx_desc = ring->buf + (wqe_index << LOG_TXBB_SIZE); in mlx4_en_handle_err_cqe() 415 print_hex_dump(KERN_WARNING, "", DUMP_PREFIX_OFFSET, 16, 1, tx_desc, desc_size, false); in mlx4_en_handle_err_cqe() 676 static void build_inline_wqe(struct mlx4_en_tx_desc *tx_desc, in build_inline_wqe() argument [all …]
|
| /kernel/linux/linux-6.6/drivers/net/ethernet/sunplus/ |
| D | spl2sw_desc.c | 37 if (!comm->tx_desc) in spl2sw_tx_descs_clean() 41 comm->tx_desc[i].cmd1 = 0; in spl2sw_tx_descs_clean() 43 comm->tx_desc[i].cmd2 = 0; in spl2sw_tx_descs_clean() 44 comm->tx_desc[i].addr1 = 0; in spl2sw_tx_descs_clean() 45 comm->tx_desc[i].addr2 = 0; in spl2sw_tx_descs_clean() 103 comm->tx_desc = NULL; in spl2sw_descs_free() 119 memset(comm->tx_desc, '\0', sizeof(struct spl2sw_mac_desc) * in spl2sw_tx_descs_init() 187 comm->tx_desc = comm->desc_base; in spl2sw_descs_alloc() 190 comm->rx_desc[0] = &comm->tx_desc[TX_DESC_NUM + MAC_GUARD_DESC_NUM]; in spl2sw_descs_alloc() 214 comm->tx_desc = NULL; in spl2sw_descs_init()
|
| /kernel/linux/linux-5.10/drivers/infiniband/ulp/iser/ |
| D | iser_initiator.c | 164 struct iser_tx_desc *tx_desc) in iser_create_send_desc() argument 169 tx_desc->dma_addr, ISER_HEADERS_LEN, DMA_TO_DEVICE); in iser_create_send_desc() 171 memset(&tx_desc->iser_header, 0, sizeof(struct iser_ctrl)); in iser_create_send_desc() 172 tx_desc->iser_header.flags = ISER_VER; in iser_create_send_desc() 173 tx_desc->num_sge = 1; in iser_create_send_desc() 373 struct iser_tx_desc *tx_desc = &iser_task->desc; in iser_send_command() local 379 tx_desc->type = ISCSI_TX_SCSI_COMMAND; in iser_send_command() 380 tx_desc->cqe.done = iser_cmd_comp; in iser_send_command() 381 iser_create_send_desc(iser_conn, tx_desc); in iser_send_command() 421 err = iser_post_send(&iser_conn->ib_conn, tx_desc, in iser_send_command() [all …]
|
| D | iser_memory.c | 239 struct iser_tx_desc *tx_desc = &iser_task->desc; in iser_reg_sig_mr() local 243 struct ib_reg_wr *wr = &tx_desc->reg_wr; in iser_reg_sig_mr() 254 iser_inv_rkey(&tx_desc->inv_wr, mr, cqe, &wr->wr); in iser_reg_sig_mr() 267 wr->wr.next = &tx_desc->send_wr; in iser_reg_sig_mr() 296 struct iser_tx_desc *tx_desc = &iser_task->desc; in iser_fast_reg_mr() local 299 struct ib_reg_wr *wr = &tx_desc->reg_wr; in iser_fast_reg_mr() 303 iser_inv_rkey(&tx_desc->inv_wr, mr, cqe, &wr->wr); in iser_fast_reg_mr() 314 wr->wr.next = &tx_desc->send_wr; in iser_fast_reg_mr()
|
| /kernel/linux/linux-6.6/drivers/infiniband/ulp/iser/ |
| D | iser_initiator.c | 144 struct iser_tx_desc *tx_desc, enum iser_desc_type type, in iser_create_send_desc() argument 149 tx_desc->type = type; in iser_create_send_desc() 150 tx_desc->cqe.done = done; in iser_create_send_desc() 153 tx_desc->dma_addr, ISER_HEADERS_LEN, DMA_TO_DEVICE); in iser_create_send_desc() 155 memset(&tx_desc->iser_header, 0, sizeof(struct iser_ctrl)); in iser_create_send_desc() 156 tx_desc->iser_header.flags = ISER_VER; in iser_create_send_desc() 157 tx_desc->num_sge = 1; in iser_create_send_desc() 351 struct iser_tx_desc *tx_desc = &iser_task->desc; in iser_send_command() local 356 iser_create_send_desc(iser_conn, tx_desc, ISCSI_TX_SCSI_COMMAND, in iser_send_command() 397 err = iser_post_send(&iser_conn->ib_conn, tx_desc); in iser_send_command() [all …]
|
| D | iser_memory.c | 253 struct iser_tx_desc *tx_desc = &iser_task->desc; in iser_reg_sig_mr() local 257 struct ib_reg_wr *wr = &tx_desc->reg_wr; in iser_reg_sig_mr() 268 iser_inv_rkey(&tx_desc->inv_wr, mr, cqe, &wr->wr); in iser_reg_sig_mr() 281 wr->wr.next = &tx_desc->send_wr; in iser_reg_sig_mr() 310 struct iser_tx_desc *tx_desc = &iser_task->desc; in iser_fast_reg_mr() local 313 struct ib_reg_wr *wr = &tx_desc->reg_wr; in iser_fast_reg_mr() 317 iser_inv_rkey(&tx_desc->inv_wr, mr, cqe, &wr->wr); in iser_fast_reg_mr() 328 wr->wr.next = &tx_desc->send_wr; in iser_fast_reg_mr()
|
| /kernel/linux/linux-5.10/drivers/net/ethernet/hisilicon/hns3/ |
| D | hns3_debugfs.c | 175 struct hns3_desc *rx_desc, *tx_desc; in hns3_dbg_bd_info() local 207 tx_desc = &ring->desc[tx_index]; in hns3_dbg_bd_info() 208 addr = le64_to_cpu(tx_desc->addr); in hns3_dbg_bd_info() 211 dev_info(dev, "(TX)vlan_tag: %u\n", le16_to_cpu(tx_desc->tx.vlan_tag)); in hns3_dbg_bd_info() 213 le16_to_cpu(tx_desc->tx.send_size)); in hns3_dbg_bd_info() 214 dev_info(dev, "(TX)vlan_tso: %u\n", tx_desc->tx.type_cs_vlan_tso); in hns3_dbg_bd_info() 215 dev_info(dev, "(TX)l2_len: %u\n", tx_desc->tx.l2_len); in hns3_dbg_bd_info() 216 dev_info(dev, "(TX)l3_len: %u\n", tx_desc->tx.l3_len); in hns3_dbg_bd_info() 217 dev_info(dev, "(TX)l4_len: %u\n", tx_desc->tx.l4_len); in hns3_dbg_bd_info() 219 le16_to_cpu(tx_desc->tx.outer_vlan_tag)); in hns3_dbg_bd_info() [all …]
|
| /kernel/linux/linux-5.10/drivers/net/ethernet/intel/ixgbe/ |
| D | ixgbe_xsk.c | 386 union ixgbe_adv_tx_desc *tx_desc = NULL; in ixgbe_xmit_zc() local 413 tx_desc = IXGBE_TX_DESC(xdp_ring, xdp_ring->next_to_use); in ixgbe_xmit_zc() 414 tx_desc->read.buffer_addr = cpu_to_le64(dma); in ixgbe_xmit_zc() 421 tx_desc->read.cmd_type_len = cpu_to_le32(cmd_type); in ixgbe_xmit_zc() 422 tx_desc->read.olinfo_status = in ixgbe_xmit_zc() 430 if (tx_desc) { in ixgbe_xmit_zc() 454 union ixgbe_adv_tx_desc *tx_desc; in ixgbe_clean_xdp_tx_irq() local 459 tx_desc = IXGBE_TX_DESC(tx_ring, ntc); in ixgbe_clean_xdp_tx_irq() 462 if (!(tx_desc->wb.status & cpu_to_le32(IXGBE_TXD_STAT_DD))) in ixgbe_clean_xdp_tx_irq() 476 tx_desc++; in ixgbe_clean_xdp_tx_irq() [all …]
|
| /kernel/linux/linux-6.6/drivers/net/ethernet/intel/ixgbe/ |
| D | ixgbe_xsk.c | 399 union ixgbe_adv_tx_desc *tx_desc = NULL; in ixgbe_xmit_zc() local 426 tx_desc = IXGBE_TX_DESC(xdp_ring, xdp_ring->next_to_use); in ixgbe_xmit_zc() 427 tx_desc->read.buffer_addr = cpu_to_le64(dma); in ixgbe_xmit_zc() 434 tx_desc->read.cmd_type_len = cpu_to_le32(cmd_type); in ixgbe_xmit_zc() 435 tx_desc->read.olinfo_status = in ixgbe_xmit_zc() 443 if (tx_desc) { in ixgbe_xmit_zc() 467 union ixgbe_adv_tx_desc *tx_desc; in ixgbe_clean_xdp_tx_irq() local 472 tx_desc = IXGBE_TX_DESC(tx_ring, ntc); in ixgbe_clean_xdp_tx_irq() 475 if (!(tx_desc->wb.status & cpu_to_le32(IXGBE_TXD_STAT_DD))) in ixgbe_clean_xdp_tx_irq() 489 tx_desc++; in ixgbe_clean_xdp_tx_irq() [all …]
|
| /kernel/linux/linux-6.6/drivers/dma/ptdma/ |
| D | ptdma-dmaengine.c | 81 struct dma_async_tx_descriptor *tx_desc; in pt_handle_active_desc() local 94 tx_desc = &desc->vd.tx; in pt_handle_active_desc() 97 tx_desc = NULL; in pt_handle_active_desc() 107 dma_cookie_complete(tx_desc); in pt_handle_active_desc() 108 dma_descriptor_unmap(tx_desc); in pt_handle_active_desc() 112 tx_desc = NULL; in pt_handle_active_desc() 120 if (tx_desc) { in pt_handle_active_desc() 121 dmaengine_desc_get_callback_invoke(tx_desc, NULL); in pt_handle_active_desc() 122 dma_run_dependencies(tx_desc); in pt_handle_active_desc()
|
| /kernel/linux/linux-6.6/drivers/net/ethernet/intel/ice/ |
| D | ice_txrx_lib.c | 258 struct ice_tx_desc *tx_desc; in ice_clean_xdp_irq() local 267 tx_desc = ICE_TX_DESC(xdp_ring, idx); in ice_clean_xdp_irq() 268 if (tx_desc->cmd_type_offset_bsz & in ice_clean_xdp_irq() 314 tx_desc->cmd_type_offset_bsz = 0; in ice_clean_xdp_irq() 335 struct ice_tx_desc *tx_desc; in __ice_xmit_xdp_ring() local 358 tx_desc = ICE_TX_DESC(xdp_ring, ntu); in __ice_xmit_xdp_ring() 380 tx_desc->buf_addr = cpu_to_le64(dma); in __ice_xmit_xdp_ring() 381 tx_desc->cmd_type_offset_bsz = ice_build_ctob(0, 0, size, 0); in __ice_xmit_xdp_ring() 390 tx_desc = ICE_TX_DESC(xdp_ring, ntu); in __ice_xmit_xdp_ring() 408 tx_desc->cmd_type_offset_bsz |= in __ice_xmit_xdp_ring()
|
| D | ice_xsk.c | 620 struct ice_tx_desc *tx_desc; in ice_clean_xdp_irq_zc() local 629 tx_desc = ICE_TX_DESC(xdp_ring, last_rs); in ice_clean_xdp_irq_zc() 630 if (tx_desc->cmd_type_offset_bsz & in ice_clean_xdp_irq_zc() 663 tx_desc->cmd_type_offset_bsz = 0; in ice_clean_xdp_irq_zc() 691 struct ice_tx_desc *tx_desc; in ice_xmit_xdp_tx_zc() local 712 tx_desc = ICE_TX_DESC(xdp_ring, ntu); in ice_xmit_xdp_tx_zc() 724 tx_desc->buf_addr = cpu_to_le64(dma); in ice_xmit_xdp_tx_zc() 725 tx_desc->cmd_type_offset_bsz = ice_build_ctob(0, 0, size, 0); in ice_xmit_xdp_tx_zc() 735 tx_desc = ICE_TX_DESC(xdp_ring, ntu); in ice_xmit_xdp_tx_zc() 745 tx_desc->cmd_type_offset_bsz |= in ice_xmit_xdp_tx_zc() [all …]
|
| D | ice_txrx.c | 38 struct ice_tx_desc *tx_desc; in ice_prgm_fdir_fltr() local 74 tx_desc = ICE_TX_DESC(tx_ring, i); in ice_prgm_fdir_fltr() 84 tx_desc->buf_addr = cpu_to_le64(dma); in ice_prgm_fdir_fltr() 91 tx_desc->cmd_type_offset_bsz = in ice_prgm_fdir_fltr() 100 first->next_to_watch = tx_desc; in ice_prgm_fdir_fltr() 223 struct ice_tx_desc *tx_desc; in ice_clean_tx_irq() local 230 tx_desc = ICE_TX_DESC(tx_ring, i); in ice_clean_tx_irq() 247 ice_trace(clean_tx_irq, tx_ring, tx_desc, tx_buf); in ice_clean_tx_irq() 274 while (tx_desc != eop_desc) { in ice_clean_tx_irq() 275 ice_trace(clean_tx_irq_unmap, tx_ring, tx_desc, tx_buf); in ice_clean_tx_irq() [all …]
|
| /kernel/linux/linux-6.6/drivers/net/wireless/realtek/rtw88/ |
| D | tx.c | 37 struct rtw_tx_desc *tx_desc = (struct rtw_tx_desc *)skb->data; in rtw_tx_fill_tx_desc() local 43 tx_desc->w0 = le32_encode_bits(pkt_info->tx_pkt_size, RTW_TX_DESC_W0_TXPKTSIZE) | in rtw_tx_fill_tx_desc() 49 tx_desc->w1 = le32_encode_bits(pkt_info->qsel, RTW_TX_DESC_W1_QSEL) | in rtw_tx_fill_tx_desc() 55 tx_desc->w2 = le32_encode_bits(pkt_info->ampdu_en, RTW_TX_DESC_W2_AGG_EN) | in rtw_tx_fill_tx_desc() 60 tx_desc->w3 = le32_encode_bits(pkt_info->hw_ssn_sel, RTW_TX_DESC_W3_HW_SSN_SEL) | in rtw_tx_fill_tx_desc() 67 tx_desc->w4 = le32_encode_bits(pkt_info->rate, RTW_TX_DESC_W4_DATARATE); in rtw_tx_fill_tx_desc() 69 tx_desc->w5 = le32_encode_bits(pkt_info->short_gi, RTW_TX_DESC_W5_DATA_SHORT) | in rtw_tx_fill_tx_desc() 74 tx_desc->w6 = le32_encode_bits(pkt_info->sn, RTW_TX_DESC_W6_SW_DEFINE); in rtw_tx_fill_tx_desc() 76 tx_desc->w8 = le32_encode_bits(pkt_info->en_hwseq, RTW_TX_DESC_W8_EN_HWSEQ); in rtw_tx_fill_tx_desc() 78 tx_desc->w9 = le32_encode_bits(pkt_info->seq, RTW_TX_DESC_W9_SW_SEQ); in rtw_tx_fill_tx_desc() [all …]
|
| /kernel/linux/linux-5.10/drivers/staging/mt7621-dma/ |
| D | hsdma-mt7621.c | 234 struct hsdma_desc *tx_desc; in hsdma_dump_desc() local 242 tx_desc = &chan->tx_ring[i]; in hsdma_dump_desc() 247 i, tx_desc->addr0, tx_desc->flags, in hsdma_dump_desc() 248 tx_desc->addr1, rx_desc->addr0, rx_desc->flags); in hsdma_dump_desc() 319 struct hsdma_desc *tx_desc, *rx_desc; in mtk_hsdma_start_transfer() local 331 tx_desc = &chan->tx_ring[chan->tx_idx]; in mtk_hsdma_start_transfer() 339 tx_desc->addr1 = src; in mtk_hsdma_start_transfer() 340 tx_desc->flags |= HSDMA_DESC_PLEN1(tlen); in mtk_hsdma_start_transfer() 342 tx_desc->addr0 = src; in mtk_hsdma_start_transfer() 343 tx_desc->flags = HSDMA_DESC_PLEN0(tlen); in mtk_hsdma_start_transfer() [all …]
|
| /kernel/linux/linux-5.10/drivers/net/ethernet/intel/fm10k/ |
| D | fm10k_main.c | 752 struct fm10k_tx_desc *tx_desc; in fm10k_tso() local 781 tx_desc = FM10K_TX_DESC(tx_ring, tx_ring->next_to_use); in fm10k_tso() 782 tx_desc->hdrlen = hdrlen; in fm10k_tso() 783 tx_desc->mss = cpu_to_le16(skb_shinfo(skb)->gso_size); in fm10k_tso() 799 struct fm10k_tx_desc *tx_desc; in fm10k_tx_csum() local 875 tx_desc = FM10K_TX_DESC(tx_ring, tx_ring->next_to_use); in fm10k_tx_csum() 876 tx_desc->hdrlen = 0; in fm10k_tx_csum() 877 tx_desc->mss = 0; in fm10k_tx_csum() 898 struct fm10k_tx_desc *tx_desc, u16 i, in fm10k_tx_desc_push() argument 906 tx_desc->buffer_addr = cpu_to_le64(dma); in fm10k_tx_desc_push() [all …]
|
| /kernel/linux/linux-6.6/drivers/net/ethernet/intel/fm10k/ |
| D | fm10k_main.c | 747 struct fm10k_tx_desc *tx_desc; in fm10k_tso() local 776 tx_desc = FM10K_TX_DESC(tx_ring, tx_ring->next_to_use); in fm10k_tso() 777 tx_desc->hdrlen = hdrlen; in fm10k_tso() 778 tx_desc->mss = cpu_to_le16(skb_shinfo(skb)->gso_size); in fm10k_tso() 794 struct fm10k_tx_desc *tx_desc; in fm10k_tx_csum() local 870 tx_desc = FM10K_TX_DESC(tx_ring, tx_ring->next_to_use); in fm10k_tx_csum() 871 tx_desc->hdrlen = 0; in fm10k_tx_csum() 872 tx_desc->mss = 0; in fm10k_tx_csum() 893 struct fm10k_tx_desc *tx_desc, u16 i, in fm10k_tx_desc_push() argument 901 tx_desc->buffer_addr = cpu_to_le64(dma); in fm10k_tx_desc_push() [all …]
|
| /kernel/linux/linux-6.6/drivers/net/wireless/ath/ath12k/ |
| D | dp_tx.c | 80 struct ath12k_tx_desc_info *tx_desc, in ath12k_dp_tx_release_txbuf() argument 84 list_move_tail(&tx_desc->list, &dp->tx_desc_free_list[pool_id]); in ath12k_dp_tx_release_txbuf() 134 struct ath12k_tx_desc_info *tx_desc; in ath12k_dp_tx() local 177 tx_desc = ath12k_dp_tx_assign_buffer(dp, pool_id); in ath12k_dp_tx() 178 if (!tx_desc) in ath12k_dp_tx() 252 tx_desc->skb = skb; in ath12k_dp_tx() 253 tx_desc->mac_id = ar->pdev_idx; in ath12k_dp_tx() 254 ti.desc_id = tx_desc->desc_id; in ath12k_dp_tx() 340 ath12k_dp_tx_release_txbuf(dp, tx_desc, pool_id); in ath12k_dp_tx() 538 struct ath12k_tx_desc_info *tx_desc = NULL; in ath12k_dp_tx_completion_handler() local [all …]
|
| /kernel/linux/linux-6.6/tools/testing/selftests/bpf/prog_tests/ |
| D | xdp_metadata.c | 92 printf("%p: tx_desc[%d] -> %lx\n", xsk, i, addr); in open_xsk() 141 struct xdp_desc *tx_desc; in generate_packet() local 153 tx_desc = xsk_ring_prod__tx_desc(&xsk->tx, idx); in generate_packet() 154 tx_desc->addr = idx % (UMEM_NUM / 2) * UMEM_FRAME_SIZE; in generate_packet() 155 printf("%p: tx_desc[%u]->addr=%llx\n", xsk, idx, tx_desc->addr); in generate_packet() 156 data = xsk_umem__get_data(xsk->umem_area, tx_desc->addr); in generate_packet() 185 tx_desc->len = sizeof(*eth) + sizeof(*iph) + sizeof(*udph) + UDP_PAYLOAD_BYTES; in generate_packet()
|
| /kernel/linux/linux-6.6/drivers/net/ethernet/seeq/ |
| D | sgiseeq.c | 96 struct sgiseeq_tx_desc *tx_desc; member 196 sp->tx_desc[i].tdma.cntinfo = TCNTINFO_INIT; in seeq_init_ring() 197 dma_sync_desc_dev(dev, &sp->tx_desc[i]); in seeq_init_ring() 230 if (sp->tx_desc[i].skb) { in seeq_purge_ring() 231 dev_kfree_skb(sp->tx_desc[i].skb); in seeq_purge_ring() 232 sp->tx_desc[i].skb = NULL; in seeq_purge_ring() 253 struct sgiseeq_tx_desc *t = gpriv->tx_desc; in sgiseeq_dump_rings() 312 hregs->tx_ndptr = VIRT_TO_DMA(sp, sp->tx_desc); in init_seeq() 443 td = &sp->tx_desc[i]; in kick_tx() 448 td = &sp->tx_desc[i]; in kick_tx() [all …]
|