/drivers/net/ethernet/mellanox/mlx5/core/ |
D | en_tx.c | 44 static void mlx5e_dma_unmap_wqe_err(struct mlx5e_txqsq *sq, u8 num_dma) in mlx5e_dma_unmap_wqe_err() 218 mlx5e_txwqe_build_eseg_csum(struct mlx5e_txqsq *sq, struct sk_buff *skb, in mlx5e_txwqe_build_eseg_csum() 245 mlx5e_tx_get_gso_ihs(struct mlx5e_txqsq *sq, struct sk_buff *skb) in mlx5e_tx_get_gso_ihs() 267 mlx5e_txwqe_build_dsegs(struct mlx5e_txqsq *sq, struct sk_buff *skb, in mlx5e_txwqe_build_dsegs() 332 mlx5e_tx_wqe_inline_mode(struct mlx5e_txqsq *sq, struct sk_buff *skb, in mlx5e_tx_wqe_inline_mode() 351 static void mlx5e_sq_xmit_prepare(struct mlx5e_txqsq *sq, struct sk_buff *skb, in mlx5e_sq_xmit_prepare() 424 static void mlx5e_tx_check_stop(struct mlx5e_txqsq *sq) in mlx5e_tx_check_stop() 432 static void mlx5e_tx_flush(struct mlx5e_txqsq *sq) in mlx5e_tx_flush() 453 mlx5e_txwqe_complete(struct mlx5e_txqsq *sq, struct sk_buff *skb, in mlx5e_txwqe_complete() 496 mlx5e_sq_xmit_wqe(struct mlx5e_txqsq *sq, struct sk_buff *skb, in mlx5e_sq_xmit_wqe() [all …]
|
D | en_txrx.c | 48 static void mlx5e_handle_tx_dim(struct mlx5e_txqsq *sq) in mlx5e_handle_tx_dim() 119 struct mlx5e_txqsq __rcu **qos_sqs; in mlx5e_napi_poll() 146 struct mlx5e_txqsq *sq = rcu_dereference(qos_sqs[i]); in mlx5e_napi_poll() 214 struct mlx5e_txqsq *sq = rcu_dereference(qos_sqs[i]); in mlx5e_napi_poll()
|
D | en.h | 368 struct mlx5e_txqsq { struct 695 struct mlx5e_txqsq sq[MLX5E_MAX_NUM_TC]; 697 struct mlx5e_txqsq __rcu * __rcu *qos_sqs; 825 struct mlx5e_txqsq *txq2sq[(MLX5E_MAX_NUM_CHANNELS + 1) * MLX5E_MAX_NUM_TC + 1015 struct mlx5e_txqsq *sq, int tc, u16 qos_queue_group_id, u16 qos_qid); 1016 void mlx5e_activate_txqsq(struct mlx5e_txqsq *sq); 1017 void mlx5e_deactivate_txqsq(struct mlx5e_txqsq *sq); 1018 void mlx5e_free_txqsq(struct mlx5e_txqsq *sq); 1020 int mlx5e_alloc_txqsq_db(struct mlx5e_txqsq *sq, int numa); 1021 void mlx5e_free_txqsq_db(struct mlx5e_txqsq *sq); [all …]
|
D | en_dim.c | 57 struct mlx5e_txqsq *sq = container_of(dim, struct mlx5e_txqsq, dim); in mlx5e_tx_dim_work()
|
D | en_main.c | 1087 void mlx5e_free_txqsq_db(struct mlx5e_txqsq *sq) in mlx5e_free_txqsq_db() 1094 int mlx5e_alloc_txqsq_db(struct mlx5e_txqsq *sq, int numa) in mlx5e_alloc_txqsq_db() 1126 struct mlx5e_txqsq *sq, in mlx5e_alloc_txqsq() 1177 void mlx5e_free_txqsq(struct mlx5e_txqsq *sq) in mlx5e_free_txqsq() 1304 struct mlx5e_txqsq *sq, u32 rate); 1308 struct mlx5e_txqsq *sq, int tc, u16 qos_queue_group_id, u16 qos_qid) in mlx5e_open_txqsq() 1347 void mlx5e_activate_txqsq(struct mlx5e_txqsq *sq) in mlx5e_activate_txqsq() 1362 void mlx5e_deactivate_txqsq(struct mlx5e_txqsq *sq) in mlx5e_deactivate_txqsq() 1385 void mlx5e_close_txqsq(struct mlx5e_txqsq *sq) in mlx5e_close_txqsq() 1403 struct mlx5e_txqsq *sq = container_of(recover_work, struct mlx5e_txqsq, in mlx5e_tx_err_cqe_work() [all …]
|
/drivers/net/ethernet/mellanox/mlx5/core/en/ |
D | reporter_tx.c | 8 static int mlx5e_wait_for_sq_flush(struct mlx5e_txqsq *sq) in mlx5e_wait_for_sq_flush() 27 static void mlx5e_reset_txqsq_cc_pc(struct mlx5e_txqsq *sq) in mlx5e_reset_txqsq_cc_pc() 41 struct mlx5e_txqsq *sq; in mlx5e_tx_reporter_err_cqe_recover() 89 struct mlx5e_txqsq *sq; 98 struct mlx5e_txqsq *sq; in mlx5e_tx_reporter_timeout_recover() 147 struct mlx5e_txqsq *sq, int tc) in mlx5e_tx_reporter_build_diagnose_output_sq_common() 195 struct mlx5e_txqsq *sq, int tc) in mlx5e_tx_reporter_build_diagnose_output() 257 struct mlx5e_txqsq *txqsq) in mlx5e_tx_reporter_diagnose_generic_txqsq() 312 struct mlx5e_txqsq *generic_sq = priv->txq2sq[0]; in mlx5e_tx_reporter_diagnose_common_config() 376 struct mlx5e_txqsq *sq = &c->sq[tc]; in mlx5e_tx_reporter_diagnose() [all …]
|
D | qos.c | 163 static struct mlx5e_txqsq *mlx5e_get_qos_sq(struct mlx5e_priv *priv, int qid) in mlx5e_get_qos_sq() 166 struct mlx5e_txqsq __rcu **qos_sqs; in mlx5e_get_qos_sq() 184 struct mlx5e_txqsq __rcu **qos_sqs; in mlx5e_open_qos_sq() 190 struct mlx5e_txqsq *sq; in mlx5e_open_qos_sq() 258 struct mlx5e_txqsq *sq; in mlx5e_activate_qos_sq() 276 struct mlx5e_txqsq *sq; in mlx5e_deactivate_qos_sq() 291 struct mlx5e_txqsq __rcu **qos_sqs; in mlx5e_close_qos_sq() 294 struct mlx5e_txqsq *sq; in mlx5e_close_qos_sq() 316 struct mlx5e_txqsq __rcu **qos_sqs; in mlx5e_qos_close_queues() 325 struct mlx5e_txqsq *sq; in mlx5e_qos_close_queues() [all …]
|
D | txrx.h | 74 void mlx5e_free_txqsq_descs(struct mlx5e_txqsq *sq); 147 static inline u16 mlx5e_txqsq_get_next_pi(struct mlx5e_txqsq *sq, u16 size) in mlx5e_txqsq_get_next_pi() 175 void mlx5e_txqsq_wake(struct mlx5e_txqsq *sq); 254 mlx5e_dma_get(struct mlx5e_txqsq *sq, u32 i) in mlx5e_dma_get() 260 mlx5e_dma_push(struct mlx5e_txqsq *sq, dma_addr_t addr, u32 size, in mlx5e_dma_push() 305 void mlx5e_sq_xmit_simple(struct mlx5e_txqsq *sq, struct sk_buff *skb, bool xmit_more); 306 void mlx5e_tx_mpwqe_ensure_complete(struct mlx5e_txqsq *sq);
|
D | health.h | 19 void mlx5e_reporter_tx_err_cqe(struct mlx5e_txqsq *sq); 20 int mlx5e_reporter_tx_timeout(struct mlx5e_txqsq *sq);
|
D | ptp.h | 15 struct mlx5e_txqsq txqsq; 71 static inline bool mlx5e_ptpsq_fifo_has_room(struct mlx5e_txqsq *sq) in mlx5e_ptpsq_fifo_has_room()
|
D | qos.h | 20 struct mlx5e_txqsq *mlx5e_get_sq(struct mlx5e_priv *priv, int qid);
|
D | ptp.c | 87 struct mlx5e_txqsq *sq = &ptpsq->txqsq; in mlx5e_ptp_handle_ts_cqe() 190 struct mlx5e_txqsq *sq, int tc, in mlx5e_ptp_alloc_txqsq() 280 struct mlx5e_txqsq *txqsq = &ptpsq->txqsq; in mlx5e_ptp_open_txqsq() 315 struct mlx5e_txqsq *sq = &ptpsq->txqsq; in mlx5e_ptp_close_txqsq()
|
D | reporter_rx.c | 473 struct mlx5e_txqsq *icosq = ctx; in mlx5e_rx_reporter_dump_icosq()
|
/drivers/net/ethernet/mellanox/mlx5/core/en_accel/ |
D | ktls_txrx.h | 19 bool mlx5e_ktls_handle_tx_skb(struct tls_context *tls_ctx, struct mlx5e_txqsq *sq, 29 void mlx5e_ktls_tx_handle_resync_dump_comp(struct mlx5e_txqsq *sq, 33 mlx5e_ktls_tx_try_handle_resync_dump_comp(struct mlx5e_txqsq *sq, in mlx5e_ktls_tx_try_handle_resync_dump_comp() 53 mlx5e_ktls_tx_try_handle_resync_dump_comp(struct mlx5e_txqsq *sq, in mlx5e_ktls_tx_try_handle_resync_dump_comp()
|
D | ktls_tx.c | 146 static void tx_fill_wi(struct mlx5e_txqsq *sq, in tx_fill_wi() 170 post_static_params(struct mlx5e_txqsq *sq, in post_static_params() 188 post_progress_params(struct mlx5e_txqsq *sq, in post_progress_params() 205 mlx5e_ktls_tx_post_param_wqes(struct mlx5e_txqsq *sq, in mlx5e_ktls_tx_post_param_wqes() 286 tx_post_resync_params(struct mlx5e_txqsq *sq, in tx_post_resync_params() 307 tx_post_resync_dump(struct mlx5e_txqsq *sq, skb_frag_t *frag, u32 tisn, bool first) in tx_post_resync_dump() 348 void mlx5e_ktls_tx_handle_resync_dump_comp(struct mlx5e_txqsq *sq, in mlx5e_ktls_tx_handle_resync_dump_comp() 364 static void tx_post_fence_nop(struct mlx5e_txqsq *sq) in tx_post_fence_nop() 376 struct mlx5e_txqsq *sq, in mlx5e_ktls_tx_handle_ooo() 450 bool mlx5e_ktls_handle_tx_skb(struct tls_context *tls_ctx, struct mlx5e_txqsq *sq, in mlx5e_ktls_handle_tx_skb()
|
D | en_accel.h | 118 struct mlx5e_txqsq *sq, in mlx5e_accel_tx_begin() 151 static inline unsigned int mlx5e_accel_tx_ids_len(struct mlx5e_txqsq *sq, in mlx5e_accel_tx_ids_len() 180 static inline void mlx5e_accel_tx_finish(struct mlx5e_txqsq *sq, in mlx5e_accel_tx_finish()
|
D | ipsec_rxtx.h | 131 mlx5e_ipsec_txwqe_build_eseg_csum(struct mlx5e_txqsq *sq, struct sk_buff *skb, in mlx5e_ipsec_txwqe_build_eseg_csum() 170 mlx5e_ipsec_txwqe_build_eseg_csum(struct mlx5e_txqsq *sq, struct sk_buff *skb, in mlx5e_ipsec_txwqe_build_eseg_csum()
|
D | tls_rxtx.h | 48 bool mlx5e_tls_handle_tx_skb(struct net_device *netdev, struct mlx5e_txqsq *sq,
|
D | tls_rxtx.c | 188 struct mlx5e_txqsq *sq, struct sk_buff *skb, in mlx5e_tls_handle_ooo() 256 bool mlx5e_tls_handle_tx_skb(struct net_device *netdev, struct mlx5e_txqsq *sq, in mlx5e_tls_handle_tx_skb()
|
/drivers/net/ethernet/mellanox/mlx5/core/ipoib/ |
D | ipoib.h | 116 void mlx5i_sq_xmit(struct mlx5e_txqsq *sq, struct sk_buff *skb,
|
D | ipoib.c | 633 struct mlx5e_txqsq *sq = epriv->txq2sq[skb_get_queue_mapping(skb)]; in mlx5i_xmit()
|