• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 /* SPDX-License-Identifier: (GPL-2.0 OR BSD-3-Clause) */
2 /*
3  * Copyright(c) 2020 Intel Corporation.
4  *
5  */
6 
7 #ifndef HFI1_NETDEV_H
8 #define HFI1_NETDEV_H
9 
10 #include "hfi.h"
11 
12 #include <linux/netdevice.h>
13 #include <linux/xarray.h>
14 
15 /**
16  * struct hfi1_netdev_rxq - Receive Queue for HFI
17  * dummy netdev. Both IPoIB and VNIC netdevices will be working on
18  * top of this device.
19  * @napi: napi object
20  * @priv: ptr to netdev_priv
21  * @rcd:  ptr to receive context data
22  */
23 struct hfi1_netdev_rxq {
24 	struct napi_struct napi;
25 	struct hfi1_netdev_priv *priv;
26 	struct hfi1_ctxtdata *rcd;
27 };
28 
29 /*
30  * Number of netdev contexts used. Ensure it is less than or equal to
31  * max queues supported by VNIC (HFI1_VNIC_MAX_QUEUE).
32  */
33 #define HFI1_MAX_NETDEV_CTXTS   8
34 
35 /* Number of NETDEV RSM entries */
36 #define NUM_NETDEV_MAP_ENTRIES HFI1_MAX_NETDEV_CTXTS
37 
38 /**
39  * struct hfi1_netdev_priv: data required to setup and run HFI netdev.
40  * @dd:		hfi1_devdata
41  * @rxq:	pointer to dummy netdev receive queues.
42  * @num_rx_q:	number of receive queues
43  * @rmt_index:	first free index in RMT Array
44  * @msix_start: first free MSI-X interrupt vector.
45  * @dev_tbl:	netdev table for unique identifier VNIC and IPoIb VLANs.
46  * @enabled:	atomic counter of netdevs enabling receive queues.
47  *		When 0 NAPI will be disabled.
48  * @netdevs:	atomic counter of netdevs using dummy netdev.
49  *		When 0 receive queues will be freed.
50  */
51 struct hfi1_netdev_priv {
52 	struct hfi1_devdata *dd;
53 	struct hfi1_netdev_rxq *rxq;
54 	int num_rx_q;
55 	int rmt_start;
56 	struct xarray dev_tbl;
57 	/* count of enabled napi polls */
58 	atomic_t enabled;
59 	/* count of netdevs on top */
60 	atomic_t netdevs;
61 };
62 
63 static inline
hfi1_netdev_priv(struct net_device * dev)64 struct hfi1_netdev_priv *hfi1_netdev_priv(struct net_device *dev)
65 {
66 	return (struct hfi1_netdev_priv *)&dev[1];
67 }
68 
69 static inline
hfi1_netdev_ctxt_count(struct hfi1_devdata * dd)70 int hfi1_netdev_ctxt_count(struct hfi1_devdata *dd)
71 {
72 	struct hfi1_netdev_priv *priv = hfi1_netdev_priv(dd->dummy_netdev);
73 
74 	return priv->num_rx_q;
75 }
76 
77 static inline
hfi1_netdev_get_ctxt(struct hfi1_devdata * dd,int ctxt)78 struct hfi1_ctxtdata *hfi1_netdev_get_ctxt(struct hfi1_devdata *dd, int ctxt)
79 {
80 	struct hfi1_netdev_priv *priv = hfi1_netdev_priv(dd->dummy_netdev);
81 
82 	return priv->rxq[ctxt].rcd;
83 }
84 
85 static inline
hfi1_netdev_get_free_rmt_idx(struct hfi1_devdata * dd)86 int hfi1_netdev_get_free_rmt_idx(struct hfi1_devdata *dd)
87 {
88 	struct hfi1_netdev_priv *priv = hfi1_netdev_priv(dd->dummy_netdev);
89 
90 	return priv->rmt_start;
91 }
92 
93 static inline
hfi1_netdev_set_free_rmt_idx(struct hfi1_devdata * dd,int rmt_idx)94 void hfi1_netdev_set_free_rmt_idx(struct hfi1_devdata *dd, int rmt_idx)
95 {
96 	struct hfi1_netdev_priv *priv = hfi1_netdev_priv(dd->dummy_netdev);
97 
98 	priv->rmt_start = rmt_idx;
99 }
100 
101 u32 hfi1_num_netdev_contexts(struct hfi1_devdata *dd, u32 available_contexts,
102 			     struct cpumask *cpu_mask);
103 
104 void hfi1_netdev_enable_queues(struct hfi1_devdata *dd);
105 void hfi1_netdev_disable_queues(struct hfi1_devdata *dd);
106 int hfi1_netdev_rx_init(struct hfi1_devdata *dd);
107 int hfi1_netdev_rx_destroy(struct hfi1_devdata *dd);
108 int hfi1_netdev_alloc(struct hfi1_devdata *dd);
109 void hfi1_netdev_free(struct hfi1_devdata *dd);
110 int hfi1_netdev_add_data(struct hfi1_devdata *dd, int id, void *data);
111 void *hfi1_netdev_remove_data(struct hfi1_devdata *dd, int id);
112 void *hfi1_netdev_get_data(struct hfi1_devdata *dd, int id);
113 void *hfi1_netdev_get_first_data(struct hfi1_devdata *dd, int *start_id);
114 
115 /* chip.c  */
116 int hfi1_netdev_rx_napi(struct napi_struct *napi, int budget);
117 
118 #endif /* HFI1_NETDEV_H */
119