1 // SPDX-License-Identifier: GPL-2.0 OR Linux-OpenIB
2 /*
3 * Copyright 2015-2020 Amazon.com, Inc. or its affiliates. All rights reserved.
4 */
5
6 #include <linux/ethtool.h>
7 #include <linux/pci.h>
8
9 #include "ena_netdev.h"
10
11 struct ena_stats {
12 char name[ETH_GSTRING_LEN];
13 int stat_offset;
14 };
15
16 #define ENA_STAT_ENA_COM_ENTRY(stat) { \
17 .name = #stat, \
18 .stat_offset = offsetof(struct ena_com_stats_admin, stat) / sizeof(u64) \
19 }
20
21 #define ENA_STAT_ENTRY(stat, stat_type) { \
22 .name = #stat, \
23 .stat_offset = offsetof(struct ena_stats_##stat_type, stat) / sizeof(u64) \
24 }
25
26 #define ENA_STAT_HW_ENTRY(stat, stat_type) { \
27 .name = #stat, \
28 .stat_offset = offsetof(struct ena_admin_##stat_type, stat) / sizeof(u64) \
29 }
30
31 #define ENA_STAT_RX_ENTRY(stat) \
32 ENA_STAT_ENTRY(stat, rx)
33
34 #define ENA_STAT_TX_ENTRY(stat) \
35 ENA_STAT_ENTRY(stat, tx)
36
37 #define ENA_STAT_GLOBAL_ENTRY(stat) \
38 ENA_STAT_ENTRY(stat, dev)
39
40 #define ENA_STAT_ENI_ENTRY(stat) \
41 ENA_STAT_HW_ENTRY(stat, eni_stats)
42
43 static const struct ena_stats ena_stats_global_strings[] = {
44 ENA_STAT_GLOBAL_ENTRY(tx_timeout),
45 ENA_STAT_GLOBAL_ENTRY(suspend),
46 ENA_STAT_GLOBAL_ENTRY(resume),
47 ENA_STAT_GLOBAL_ENTRY(wd_expired),
48 ENA_STAT_GLOBAL_ENTRY(interface_up),
49 ENA_STAT_GLOBAL_ENTRY(interface_down),
50 ENA_STAT_GLOBAL_ENTRY(admin_q_pause),
51 };
52
53 static const struct ena_stats ena_stats_eni_strings[] = {
54 ENA_STAT_ENI_ENTRY(bw_in_allowance_exceeded),
55 ENA_STAT_ENI_ENTRY(bw_out_allowance_exceeded),
56 ENA_STAT_ENI_ENTRY(pps_allowance_exceeded),
57 ENA_STAT_ENI_ENTRY(conntrack_allowance_exceeded),
58 ENA_STAT_ENI_ENTRY(linklocal_allowance_exceeded),
59 };
60
61 static const struct ena_stats ena_stats_tx_strings[] = {
62 ENA_STAT_TX_ENTRY(cnt),
63 ENA_STAT_TX_ENTRY(bytes),
64 ENA_STAT_TX_ENTRY(queue_stop),
65 ENA_STAT_TX_ENTRY(queue_wakeup),
66 ENA_STAT_TX_ENTRY(dma_mapping_err),
67 ENA_STAT_TX_ENTRY(linearize),
68 ENA_STAT_TX_ENTRY(linearize_failed),
69 ENA_STAT_TX_ENTRY(napi_comp),
70 ENA_STAT_TX_ENTRY(tx_poll),
71 ENA_STAT_TX_ENTRY(doorbells),
72 ENA_STAT_TX_ENTRY(prepare_ctx_err),
73 ENA_STAT_TX_ENTRY(bad_req_id),
74 ENA_STAT_TX_ENTRY(llq_buffer_copy),
75 ENA_STAT_TX_ENTRY(missed_tx),
76 ENA_STAT_TX_ENTRY(unmask_interrupt),
77 };
78
79 static const struct ena_stats ena_stats_rx_strings[] = {
80 ENA_STAT_RX_ENTRY(cnt),
81 ENA_STAT_RX_ENTRY(bytes),
82 ENA_STAT_RX_ENTRY(rx_copybreak_pkt),
83 ENA_STAT_RX_ENTRY(csum_good),
84 ENA_STAT_RX_ENTRY(refil_partial),
85 ENA_STAT_RX_ENTRY(bad_csum),
86 ENA_STAT_RX_ENTRY(page_alloc_fail),
87 ENA_STAT_RX_ENTRY(skb_alloc_fail),
88 ENA_STAT_RX_ENTRY(dma_mapping_err),
89 ENA_STAT_RX_ENTRY(bad_desc_num),
90 ENA_STAT_RX_ENTRY(bad_req_id),
91 ENA_STAT_RX_ENTRY(empty_rx_ring),
92 ENA_STAT_RX_ENTRY(csum_unchecked),
93 ENA_STAT_RX_ENTRY(xdp_aborted),
94 ENA_STAT_RX_ENTRY(xdp_drop),
95 ENA_STAT_RX_ENTRY(xdp_pass),
96 ENA_STAT_RX_ENTRY(xdp_tx),
97 ENA_STAT_RX_ENTRY(xdp_invalid),
98 ENA_STAT_RX_ENTRY(xdp_redirect),
99 };
100
101 static const struct ena_stats ena_stats_ena_com_strings[] = {
102 ENA_STAT_ENA_COM_ENTRY(aborted_cmd),
103 ENA_STAT_ENA_COM_ENTRY(submitted_cmd),
104 ENA_STAT_ENA_COM_ENTRY(completed_cmd),
105 ENA_STAT_ENA_COM_ENTRY(out_of_space),
106 ENA_STAT_ENA_COM_ENTRY(no_completion),
107 };
108
109 #define ENA_STATS_ARRAY_GLOBAL ARRAY_SIZE(ena_stats_global_strings)
110 #define ENA_STATS_ARRAY_TX ARRAY_SIZE(ena_stats_tx_strings)
111 #define ENA_STATS_ARRAY_RX ARRAY_SIZE(ena_stats_rx_strings)
112 #define ENA_STATS_ARRAY_ENA_COM ARRAY_SIZE(ena_stats_ena_com_strings)
113 #define ENA_STATS_ARRAY_ENI(adapter) \
114 (ARRAY_SIZE(ena_stats_eni_strings) * (adapter)->eni_stats_supported)
115
ena_safe_update_stat(u64 * src,u64 * dst,struct u64_stats_sync * syncp)116 static void ena_safe_update_stat(u64 *src, u64 *dst,
117 struct u64_stats_sync *syncp)
118 {
119 unsigned int start;
120
121 do {
122 start = u64_stats_fetch_begin_irq(syncp);
123 *(dst) = *src;
124 } while (u64_stats_fetch_retry_irq(syncp, start));
125 }
126
ena_queue_stats(struct ena_adapter * adapter,u64 ** data)127 static void ena_queue_stats(struct ena_adapter *adapter, u64 **data)
128 {
129 const struct ena_stats *ena_stats;
130 struct ena_ring *ring;
131
132 u64 *ptr;
133 int i, j;
134
135 for (i = 0; i < adapter->num_io_queues + adapter->xdp_num_queues; i++) {
136 /* Tx stats */
137 ring = &adapter->tx_ring[i];
138
139 for (j = 0; j < ENA_STATS_ARRAY_TX; j++) {
140 ena_stats = &ena_stats_tx_strings[j];
141
142 ptr = (u64 *)&ring->tx_stats + ena_stats->stat_offset;
143
144 ena_safe_update_stat(ptr, (*data)++, &ring->syncp);
145 }
146 /* XDP TX queues don't have a RX queue counterpart */
147 if (!ENA_IS_XDP_INDEX(adapter, i)) {
148 /* Rx stats */
149 ring = &adapter->rx_ring[i];
150
151 for (j = 0; j < ENA_STATS_ARRAY_RX; j++) {
152 ena_stats = &ena_stats_rx_strings[j];
153
154 ptr = (u64 *)&ring->rx_stats +
155 ena_stats->stat_offset;
156
157 ena_safe_update_stat(ptr, (*data)++, &ring->syncp);
158 }
159 }
160 }
161 }
162
ena_dev_admin_queue_stats(struct ena_adapter * adapter,u64 ** data)163 static void ena_dev_admin_queue_stats(struct ena_adapter *adapter, u64 **data)
164 {
165 const struct ena_stats *ena_stats;
166 u64 *ptr;
167 int i;
168
169 for (i = 0; i < ENA_STATS_ARRAY_ENA_COM; i++) {
170 ena_stats = &ena_stats_ena_com_strings[i];
171
172 ptr = (u64 *)&adapter->ena_dev->admin_queue.stats +
173 ena_stats->stat_offset;
174
175 *(*data)++ = *ptr;
176 }
177 }
178
ena_get_stats(struct ena_adapter * adapter,u64 * data,bool eni_stats_needed)179 static void ena_get_stats(struct ena_adapter *adapter,
180 u64 *data,
181 bool eni_stats_needed)
182 {
183 const struct ena_stats *ena_stats;
184 u64 *ptr;
185 int i;
186
187 for (i = 0; i < ENA_STATS_ARRAY_GLOBAL; i++) {
188 ena_stats = &ena_stats_global_strings[i];
189
190 ptr = (u64 *)&adapter->dev_stats + ena_stats->stat_offset;
191
192 ena_safe_update_stat(ptr, data++, &adapter->syncp);
193 }
194
195 if (eni_stats_needed) {
196 ena_update_hw_stats(adapter);
197 for (i = 0; i < ENA_STATS_ARRAY_ENI(adapter); i++) {
198 ena_stats = &ena_stats_eni_strings[i];
199
200 ptr = (u64 *)&adapter->eni_stats +
201 ena_stats->stat_offset;
202
203 ena_safe_update_stat(ptr, data++, &adapter->syncp);
204 }
205 }
206
207 ena_queue_stats(adapter, &data);
208 ena_dev_admin_queue_stats(adapter, &data);
209 }
210
ena_get_ethtool_stats(struct net_device * netdev,struct ethtool_stats * stats,u64 * data)211 static void ena_get_ethtool_stats(struct net_device *netdev,
212 struct ethtool_stats *stats,
213 u64 *data)
214 {
215 struct ena_adapter *adapter = netdev_priv(netdev);
216
217 ena_get_stats(adapter, data, adapter->eni_stats_supported);
218 }
219
ena_get_sw_stats_count(struct ena_adapter * adapter)220 static int ena_get_sw_stats_count(struct ena_adapter *adapter)
221 {
222 return adapter->num_io_queues * (ENA_STATS_ARRAY_TX + ENA_STATS_ARRAY_RX)
223 + adapter->xdp_num_queues * ENA_STATS_ARRAY_TX
224 + ENA_STATS_ARRAY_GLOBAL + ENA_STATS_ARRAY_ENA_COM;
225 }
226
ena_get_hw_stats_count(struct ena_adapter * adapter)227 static int ena_get_hw_stats_count(struct ena_adapter *adapter)
228 {
229 return ENA_STATS_ARRAY_ENI(adapter);
230 }
231
ena_get_sset_count(struct net_device * netdev,int sset)232 int ena_get_sset_count(struct net_device *netdev, int sset)
233 {
234 struct ena_adapter *adapter = netdev_priv(netdev);
235
236 switch (sset) {
237 case ETH_SS_STATS:
238 return ena_get_sw_stats_count(adapter) +
239 ena_get_hw_stats_count(adapter);
240 }
241
242 return -EOPNOTSUPP;
243 }
244
ena_queue_strings(struct ena_adapter * adapter,u8 ** data)245 static void ena_queue_strings(struct ena_adapter *adapter, u8 **data)
246 {
247 const struct ena_stats *ena_stats;
248 bool is_xdp;
249 int i, j;
250
251 for (i = 0; i < adapter->num_io_queues + adapter->xdp_num_queues; i++) {
252 is_xdp = ENA_IS_XDP_INDEX(adapter, i);
253 /* Tx stats */
254 for (j = 0; j < ENA_STATS_ARRAY_TX; j++) {
255 ena_stats = &ena_stats_tx_strings[j];
256
257 ethtool_sprintf(data,
258 "queue_%u_%s_%s", i,
259 is_xdp ? "xdp_tx" : "tx",
260 ena_stats->name);
261 }
262
263 if (!is_xdp) {
264 /* RX stats, in XDP there isn't a RX queue
265 * counterpart
266 */
267 for (j = 0; j < ENA_STATS_ARRAY_RX; j++) {
268 ena_stats = &ena_stats_rx_strings[j];
269
270 ethtool_sprintf(data,
271 "queue_%u_rx_%s", i,
272 ena_stats->name);
273 }
274 }
275 }
276 }
277
ena_com_dev_strings(u8 ** data)278 static void ena_com_dev_strings(u8 **data)
279 {
280 const struct ena_stats *ena_stats;
281 int i;
282
283 for (i = 0; i < ENA_STATS_ARRAY_ENA_COM; i++) {
284 ena_stats = &ena_stats_ena_com_strings[i];
285
286 ethtool_sprintf(data,
287 "ena_admin_q_%s", ena_stats->name);
288 }
289 }
290
ena_get_strings(struct ena_adapter * adapter,u8 * data,bool eni_stats_needed)291 static void ena_get_strings(struct ena_adapter *adapter,
292 u8 *data,
293 bool eni_stats_needed)
294 {
295 const struct ena_stats *ena_stats;
296 int i;
297
298 for (i = 0; i < ENA_STATS_ARRAY_GLOBAL; i++) {
299 ena_stats = &ena_stats_global_strings[i];
300 ethtool_sprintf(&data, ena_stats->name);
301 }
302
303 if (eni_stats_needed) {
304 for (i = 0; i < ENA_STATS_ARRAY_ENI(adapter); i++) {
305 ena_stats = &ena_stats_eni_strings[i];
306 ethtool_sprintf(&data, ena_stats->name);
307 }
308 }
309
310 ena_queue_strings(adapter, &data);
311 ena_com_dev_strings(&data);
312 }
313
ena_get_ethtool_strings(struct net_device * netdev,u32 sset,u8 * data)314 static void ena_get_ethtool_strings(struct net_device *netdev,
315 u32 sset,
316 u8 *data)
317 {
318 struct ena_adapter *adapter = netdev_priv(netdev);
319
320 switch (sset) {
321 case ETH_SS_STATS:
322 ena_get_strings(adapter, data, adapter->eni_stats_supported);
323 break;
324 }
325 }
326
ena_get_link_ksettings(struct net_device * netdev,struct ethtool_link_ksettings * link_ksettings)327 static int ena_get_link_ksettings(struct net_device *netdev,
328 struct ethtool_link_ksettings *link_ksettings)
329 {
330 struct ena_adapter *adapter = netdev_priv(netdev);
331 struct ena_com_dev *ena_dev = adapter->ena_dev;
332 struct ena_admin_get_feature_link_desc *link;
333 struct ena_admin_get_feat_resp feat_resp;
334 int rc;
335
336 rc = ena_com_get_link_params(ena_dev, &feat_resp);
337 if (rc)
338 return rc;
339
340 link = &feat_resp.u.link;
341 link_ksettings->base.speed = link->speed;
342
343 if (link->flags & ENA_ADMIN_GET_FEATURE_LINK_DESC_AUTONEG_MASK) {
344 ethtool_link_ksettings_add_link_mode(link_ksettings,
345 supported, Autoneg);
346 ethtool_link_ksettings_add_link_mode(link_ksettings,
347 supported, Autoneg);
348 }
349
350 link_ksettings->base.autoneg =
351 (link->flags & ENA_ADMIN_GET_FEATURE_LINK_DESC_AUTONEG_MASK) ?
352 AUTONEG_ENABLE : AUTONEG_DISABLE;
353
354 link_ksettings->base.duplex = DUPLEX_FULL;
355
356 return 0;
357 }
358
ena_get_coalesce(struct net_device * net_dev,struct ethtool_coalesce * coalesce,struct kernel_ethtool_coalesce * kernel_coal,struct netlink_ext_ack * extack)359 static int ena_get_coalesce(struct net_device *net_dev,
360 struct ethtool_coalesce *coalesce,
361 struct kernel_ethtool_coalesce *kernel_coal,
362 struct netlink_ext_ack *extack)
363 {
364 struct ena_adapter *adapter = netdev_priv(net_dev);
365 struct ena_com_dev *ena_dev = adapter->ena_dev;
366
367 if (!ena_com_interrupt_moderation_supported(ena_dev))
368 return -EOPNOTSUPP;
369
370 coalesce->tx_coalesce_usecs =
371 ena_com_get_nonadaptive_moderation_interval_tx(ena_dev) *
372 ena_dev->intr_delay_resolution;
373
374 coalesce->rx_coalesce_usecs =
375 ena_com_get_nonadaptive_moderation_interval_rx(ena_dev)
376 * ena_dev->intr_delay_resolution;
377
378 coalesce->use_adaptive_rx_coalesce =
379 ena_com_get_adaptive_moderation_enabled(ena_dev);
380
381 return 0;
382 }
383
ena_update_tx_rings_nonadaptive_intr_moderation(struct ena_adapter * adapter)384 static void ena_update_tx_rings_nonadaptive_intr_moderation(struct ena_adapter *adapter)
385 {
386 unsigned int val;
387 int i;
388
389 val = ena_com_get_nonadaptive_moderation_interval_tx(adapter->ena_dev);
390
391 for (i = 0; i < adapter->num_io_queues; i++)
392 adapter->tx_ring[i].smoothed_interval = val;
393 }
394
ena_update_rx_rings_nonadaptive_intr_moderation(struct ena_adapter * adapter)395 static void ena_update_rx_rings_nonadaptive_intr_moderation(struct ena_adapter *adapter)
396 {
397 unsigned int val;
398 int i;
399
400 val = ena_com_get_nonadaptive_moderation_interval_rx(adapter->ena_dev);
401
402 for (i = 0; i < adapter->num_io_queues; i++)
403 adapter->rx_ring[i].smoothed_interval = val;
404 }
405
ena_set_coalesce(struct net_device * net_dev,struct ethtool_coalesce * coalesce,struct kernel_ethtool_coalesce * kernel_coal,struct netlink_ext_ack * extack)406 static int ena_set_coalesce(struct net_device *net_dev,
407 struct ethtool_coalesce *coalesce,
408 struct kernel_ethtool_coalesce *kernel_coal,
409 struct netlink_ext_ack *extack)
410 {
411 struct ena_adapter *adapter = netdev_priv(net_dev);
412 struct ena_com_dev *ena_dev = adapter->ena_dev;
413 int rc;
414
415 if (!ena_com_interrupt_moderation_supported(ena_dev))
416 return -EOPNOTSUPP;
417
418 rc = ena_com_update_nonadaptive_moderation_interval_tx(ena_dev,
419 coalesce->tx_coalesce_usecs);
420 if (rc)
421 return rc;
422
423 ena_update_tx_rings_nonadaptive_intr_moderation(adapter);
424
425 rc = ena_com_update_nonadaptive_moderation_interval_rx(ena_dev,
426 coalesce->rx_coalesce_usecs);
427 if (rc)
428 return rc;
429
430 ena_update_rx_rings_nonadaptive_intr_moderation(adapter);
431
432 if (coalesce->use_adaptive_rx_coalesce &&
433 !ena_com_get_adaptive_moderation_enabled(ena_dev))
434 ena_com_enable_adaptive_moderation(ena_dev);
435
436 if (!coalesce->use_adaptive_rx_coalesce &&
437 ena_com_get_adaptive_moderation_enabled(ena_dev))
438 ena_com_disable_adaptive_moderation(ena_dev);
439
440 return 0;
441 }
442
ena_get_msglevel(struct net_device * netdev)443 static u32 ena_get_msglevel(struct net_device *netdev)
444 {
445 struct ena_adapter *adapter = netdev_priv(netdev);
446
447 return adapter->msg_enable;
448 }
449
ena_set_msglevel(struct net_device * netdev,u32 value)450 static void ena_set_msglevel(struct net_device *netdev, u32 value)
451 {
452 struct ena_adapter *adapter = netdev_priv(netdev);
453
454 adapter->msg_enable = value;
455 }
456
ena_get_drvinfo(struct net_device * dev,struct ethtool_drvinfo * info)457 static void ena_get_drvinfo(struct net_device *dev,
458 struct ethtool_drvinfo *info)
459 {
460 struct ena_adapter *adapter = netdev_priv(dev);
461
462 strlcpy(info->driver, DRV_MODULE_NAME, sizeof(info->driver));
463 strlcpy(info->bus_info, pci_name(adapter->pdev),
464 sizeof(info->bus_info));
465 }
466
ena_get_ringparam(struct net_device * netdev,struct ethtool_ringparam * ring)467 static void ena_get_ringparam(struct net_device *netdev,
468 struct ethtool_ringparam *ring)
469 {
470 struct ena_adapter *adapter = netdev_priv(netdev);
471
472 ring->tx_max_pending = adapter->max_tx_ring_size;
473 ring->rx_max_pending = adapter->max_rx_ring_size;
474 ring->tx_pending = adapter->tx_ring[0].ring_size;
475 ring->rx_pending = adapter->rx_ring[0].ring_size;
476 }
477
ena_set_ringparam(struct net_device * netdev,struct ethtool_ringparam * ring)478 static int ena_set_ringparam(struct net_device *netdev,
479 struct ethtool_ringparam *ring)
480 {
481 struct ena_adapter *adapter = netdev_priv(netdev);
482 u32 new_tx_size, new_rx_size;
483
484 new_tx_size = ring->tx_pending < ENA_MIN_RING_SIZE ?
485 ENA_MIN_RING_SIZE : ring->tx_pending;
486 new_tx_size = rounddown_pow_of_two(new_tx_size);
487
488 new_rx_size = ring->rx_pending < ENA_MIN_RING_SIZE ?
489 ENA_MIN_RING_SIZE : ring->rx_pending;
490 new_rx_size = rounddown_pow_of_two(new_rx_size);
491
492 if (new_tx_size == adapter->requested_tx_ring_size &&
493 new_rx_size == adapter->requested_rx_ring_size)
494 return 0;
495
496 return ena_update_queue_sizes(adapter, new_tx_size, new_rx_size);
497 }
498
ena_flow_hash_to_flow_type(u16 hash_fields)499 static u32 ena_flow_hash_to_flow_type(u16 hash_fields)
500 {
501 u32 data = 0;
502
503 if (hash_fields & ENA_ADMIN_RSS_L2_DA)
504 data |= RXH_L2DA;
505
506 if (hash_fields & ENA_ADMIN_RSS_L3_DA)
507 data |= RXH_IP_DST;
508
509 if (hash_fields & ENA_ADMIN_RSS_L3_SA)
510 data |= RXH_IP_SRC;
511
512 if (hash_fields & ENA_ADMIN_RSS_L4_DP)
513 data |= RXH_L4_B_2_3;
514
515 if (hash_fields & ENA_ADMIN_RSS_L4_SP)
516 data |= RXH_L4_B_0_1;
517
518 return data;
519 }
520
ena_flow_data_to_flow_hash(u32 hash_fields)521 static u16 ena_flow_data_to_flow_hash(u32 hash_fields)
522 {
523 u16 data = 0;
524
525 if (hash_fields & RXH_L2DA)
526 data |= ENA_ADMIN_RSS_L2_DA;
527
528 if (hash_fields & RXH_IP_DST)
529 data |= ENA_ADMIN_RSS_L3_DA;
530
531 if (hash_fields & RXH_IP_SRC)
532 data |= ENA_ADMIN_RSS_L3_SA;
533
534 if (hash_fields & RXH_L4_B_2_3)
535 data |= ENA_ADMIN_RSS_L4_DP;
536
537 if (hash_fields & RXH_L4_B_0_1)
538 data |= ENA_ADMIN_RSS_L4_SP;
539
540 return data;
541 }
542
ena_get_rss_hash(struct ena_com_dev * ena_dev,struct ethtool_rxnfc * cmd)543 static int ena_get_rss_hash(struct ena_com_dev *ena_dev,
544 struct ethtool_rxnfc *cmd)
545 {
546 enum ena_admin_flow_hash_proto proto;
547 u16 hash_fields;
548 int rc;
549
550 cmd->data = 0;
551
552 switch (cmd->flow_type) {
553 case TCP_V4_FLOW:
554 proto = ENA_ADMIN_RSS_TCP4;
555 break;
556 case UDP_V4_FLOW:
557 proto = ENA_ADMIN_RSS_UDP4;
558 break;
559 case TCP_V6_FLOW:
560 proto = ENA_ADMIN_RSS_TCP6;
561 break;
562 case UDP_V6_FLOW:
563 proto = ENA_ADMIN_RSS_UDP6;
564 break;
565 case IPV4_FLOW:
566 proto = ENA_ADMIN_RSS_IP4;
567 break;
568 case IPV6_FLOW:
569 proto = ENA_ADMIN_RSS_IP6;
570 break;
571 case ETHER_FLOW:
572 proto = ENA_ADMIN_RSS_NOT_IP;
573 break;
574 case AH_V4_FLOW:
575 case ESP_V4_FLOW:
576 case AH_V6_FLOW:
577 case ESP_V6_FLOW:
578 case SCTP_V4_FLOW:
579 case AH_ESP_V4_FLOW:
580 return -EOPNOTSUPP;
581 default:
582 return -EINVAL;
583 }
584
585 rc = ena_com_get_hash_ctrl(ena_dev, proto, &hash_fields);
586 if (rc)
587 return rc;
588
589 cmd->data = ena_flow_hash_to_flow_type(hash_fields);
590
591 return 0;
592 }
593
ena_set_rss_hash(struct ena_com_dev * ena_dev,struct ethtool_rxnfc * cmd)594 static int ena_set_rss_hash(struct ena_com_dev *ena_dev,
595 struct ethtool_rxnfc *cmd)
596 {
597 enum ena_admin_flow_hash_proto proto;
598 u16 hash_fields;
599
600 switch (cmd->flow_type) {
601 case TCP_V4_FLOW:
602 proto = ENA_ADMIN_RSS_TCP4;
603 break;
604 case UDP_V4_FLOW:
605 proto = ENA_ADMIN_RSS_UDP4;
606 break;
607 case TCP_V6_FLOW:
608 proto = ENA_ADMIN_RSS_TCP6;
609 break;
610 case UDP_V6_FLOW:
611 proto = ENA_ADMIN_RSS_UDP6;
612 break;
613 case IPV4_FLOW:
614 proto = ENA_ADMIN_RSS_IP4;
615 break;
616 case IPV6_FLOW:
617 proto = ENA_ADMIN_RSS_IP6;
618 break;
619 case ETHER_FLOW:
620 proto = ENA_ADMIN_RSS_NOT_IP;
621 break;
622 case AH_V4_FLOW:
623 case ESP_V4_FLOW:
624 case AH_V6_FLOW:
625 case ESP_V6_FLOW:
626 case SCTP_V4_FLOW:
627 case AH_ESP_V4_FLOW:
628 return -EOPNOTSUPP;
629 default:
630 return -EINVAL;
631 }
632
633 hash_fields = ena_flow_data_to_flow_hash(cmd->data);
634
635 return ena_com_fill_hash_ctrl(ena_dev, proto, hash_fields);
636 }
637
ena_set_rxnfc(struct net_device * netdev,struct ethtool_rxnfc * info)638 static int ena_set_rxnfc(struct net_device *netdev, struct ethtool_rxnfc *info)
639 {
640 struct ena_adapter *adapter = netdev_priv(netdev);
641 int rc = 0;
642
643 switch (info->cmd) {
644 case ETHTOOL_SRXFH:
645 rc = ena_set_rss_hash(adapter->ena_dev, info);
646 break;
647 case ETHTOOL_SRXCLSRLDEL:
648 case ETHTOOL_SRXCLSRLINS:
649 default:
650 netif_err(adapter, drv, netdev,
651 "Command parameter %d is not supported\n", info->cmd);
652 rc = -EOPNOTSUPP;
653 }
654
655 return rc;
656 }
657
ena_get_rxnfc(struct net_device * netdev,struct ethtool_rxnfc * info,u32 * rules)658 static int ena_get_rxnfc(struct net_device *netdev, struct ethtool_rxnfc *info,
659 u32 *rules)
660 {
661 struct ena_adapter *adapter = netdev_priv(netdev);
662 int rc = 0;
663
664 switch (info->cmd) {
665 case ETHTOOL_GRXRINGS:
666 info->data = adapter->num_io_queues;
667 rc = 0;
668 break;
669 case ETHTOOL_GRXFH:
670 rc = ena_get_rss_hash(adapter->ena_dev, info);
671 break;
672 case ETHTOOL_GRXCLSRLCNT:
673 case ETHTOOL_GRXCLSRULE:
674 case ETHTOOL_GRXCLSRLALL:
675 default:
676 netif_err(adapter, drv, netdev,
677 "Command parameter %d is not supported\n", info->cmd);
678 rc = -EOPNOTSUPP;
679 }
680
681 return rc;
682 }
683
ena_get_rxfh_indir_size(struct net_device * netdev)684 static u32 ena_get_rxfh_indir_size(struct net_device *netdev)
685 {
686 return ENA_RX_RSS_TABLE_SIZE;
687 }
688
ena_get_rxfh_key_size(struct net_device * netdev)689 static u32 ena_get_rxfh_key_size(struct net_device *netdev)
690 {
691 return ENA_HASH_KEY_SIZE;
692 }
693
ena_indirection_table_set(struct ena_adapter * adapter,const u32 * indir)694 static int ena_indirection_table_set(struct ena_adapter *adapter,
695 const u32 *indir)
696 {
697 struct ena_com_dev *ena_dev = adapter->ena_dev;
698 int i, rc;
699
700 for (i = 0; i < ENA_RX_RSS_TABLE_SIZE; i++) {
701 rc = ena_com_indirect_table_fill_entry(ena_dev,
702 i,
703 ENA_IO_RXQ_IDX(indir[i]));
704 if (unlikely(rc)) {
705 netif_err(adapter, drv, adapter->netdev,
706 "Cannot fill indirect table (index is too large)\n");
707 return rc;
708 }
709 }
710
711 rc = ena_com_indirect_table_set(ena_dev);
712 if (rc) {
713 netif_err(adapter, drv, adapter->netdev,
714 "Cannot set indirect table\n");
715 return rc == -EPERM ? -EOPNOTSUPP : rc;
716 }
717 return rc;
718 }
719
ena_indirection_table_get(struct ena_adapter * adapter,u32 * indir)720 static int ena_indirection_table_get(struct ena_adapter *adapter, u32 *indir)
721 {
722 struct ena_com_dev *ena_dev = adapter->ena_dev;
723 int i, rc;
724
725 if (!indir)
726 return 0;
727
728 rc = ena_com_indirect_table_get(ena_dev, indir);
729 if (rc)
730 return rc;
731
732 /* Our internal representation of the indices is: even indices
733 * for Tx and uneven indices for Rx. We need to convert the Rx
734 * indices to be consecutive
735 */
736 for (i = 0; i < ENA_RX_RSS_TABLE_SIZE; i++)
737 indir[i] = ENA_IO_RXQ_IDX_TO_COMBINED_IDX(indir[i]);
738
739 return rc;
740 }
741
ena_get_rxfh(struct net_device * netdev,u32 * indir,u8 * key,u8 * hfunc)742 static int ena_get_rxfh(struct net_device *netdev, u32 *indir, u8 *key,
743 u8 *hfunc)
744 {
745 struct ena_adapter *adapter = netdev_priv(netdev);
746 enum ena_admin_hash_functions ena_func;
747 u8 func;
748 int rc;
749
750 rc = ena_indirection_table_get(adapter, indir);
751 if (rc)
752 return rc;
753
754 /* We call this function in order to check if the device
755 * supports getting/setting the hash function.
756 */
757 rc = ena_com_get_hash_function(adapter->ena_dev, &ena_func);
758 if (rc) {
759 if (rc == -EOPNOTSUPP)
760 rc = 0;
761
762 return rc;
763 }
764
765 rc = ena_com_get_hash_key(adapter->ena_dev, key);
766 if (rc)
767 return rc;
768
769 switch (ena_func) {
770 case ENA_ADMIN_TOEPLITZ:
771 func = ETH_RSS_HASH_TOP;
772 break;
773 case ENA_ADMIN_CRC32:
774 func = ETH_RSS_HASH_CRC32;
775 break;
776 default:
777 netif_err(adapter, drv, netdev,
778 "Command parameter is not supported\n");
779 return -EOPNOTSUPP;
780 }
781
782 if (hfunc)
783 *hfunc = func;
784
785 return 0;
786 }
787
ena_set_rxfh(struct net_device * netdev,const u32 * indir,const u8 * key,const u8 hfunc)788 static int ena_set_rxfh(struct net_device *netdev, const u32 *indir,
789 const u8 *key, const u8 hfunc)
790 {
791 struct ena_adapter *adapter = netdev_priv(netdev);
792 struct ena_com_dev *ena_dev = adapter->ena_dev;
793 enum ena_admin_hash_functions func = 0;
794 int rc;
795
796 if (indir) {
797 rc = ena_indirection_table_set(adapter, indir);
798 if (rc)
799 return rc;
800 }
801
802 switch (hfunc) {
803 case ETH_RSS_HASH_NO_CHANGE:
804 func = ena_com_get_current_hash_function(ena_dev);
805 break;
806 case ETH_RSS_HASH_TOP:
807 func = ENA_ADMIN_TOEPLITZ;
808 break;
809 case ETH_RSS_HASH_CRC32:
810 func = ENA_ADMIN_CRC32;
811 break;
812 default:
813 netif_err(adapter, drv, netdev, "Unsupported hfunc %d\n",
814 hfunc);
815 return -EOPNOTSUPP;
816 }
817
818 if (key || func) {
819 rc = ena_com_fill_hash_function(ena_dev, func, key,
820 ENA_HASH_KEY_SIZE,
821 0xFFFFFFFF);
822 if (unlikely(rc)) {
823 netif_err(adapter, drv, netdev, "Cannot fill key\n");
824 return rc == -EPERM ? -EOPNOTSUPP : rc;
825 }
826 }
827
828 return 0;
829 }
830
ena_get_channels(struct net_device * netdev,struct ethtool_channels * channels)831 static void ena_get_channels(struct net_device *netdev,
832 struct ethtool_channels *channels)
833 {
834 struct ena_adapter *adapter = netdev_priv(netdev);
835
836 channels->max_combined = adapter->max_num_io_queues;
837 channels->combined_count = adapter->num_io_queues;
838 }
839
ena_set_channels(struct net_device * netdev,struct ethtool_channels * channels)840 static int ena_set_channels(struct net_device *netdev,
841 struct ethtool_channels *channels)
842 {
843 struct ena_adapter *adapter = netdev_priv(netdev);
844 u32 count = channels->combined_count;
845 /* The check for max value is already done in ethtool */
846 if (count < ENA_MIN_NUM_IO_QUEUES ||
847 (ena_xdp_present(adapter) &&
848 !ena_xdp_legal_queue_count(adapter, count)))
849 return -EINVAL;
850
851 return ena_update_queue_count(adapter, count);
852 }
853
ena_get_tunable(struct net_device * netdev,const struct ethtool_tunable * tuna,void * data)854 static int ena_get_tunable(struct net_device *netdev,
855 const struct ethtool_tunable *tuna, void *data)
856 {
857 struct ena_adapter *adapter = netdev_priv(netdev);
858 int ret = 0;
859
860 switch (tuna->id) {
861 case ETHTOOL_RX_COPYBREAK:
862 *(u32 *)data = adapter->rx_copybreak;
863 break;
864 default:
865 ret = -EINVAL;
866 break;
867 }
868
869 return ret;
870 }
871
ena_set_tunable(struct net_device * netdev,const struct ethtool_tunable * tuna,const void * data)872 static int ena_set_tunable(struct net_device *netdev,
873 const struct ethtool_tunable *tuna,
874 const void *data)
875 {
876 struct ena_adapter *adapter = netdev_priv(netdev);
877 int ret = 0;
878 u32 len;
879
880 switch (tuna->id) {
881 case ETHTOOL_RX_COPYBREAK:
882 len = *(u32 *)data;
883 ret = ena_set_rx_copybreak(adapter, len);
884 break;
885 default:
886 ret = -EINVAL;
887 break;
888 }
889
890 return ret;
891 }
892
893 static const struct ethtool_ops ena_ethtool_ops = {
894 .supported_coalesce_params = ETHTOOL_COALESCE_USECS |
895 ETHTOOL_COALESCE_USE_ADAPTIVE_RX,
896 .get_link_ksettings = ena_get_link_ksettings,
897 .get_drvinfo = ena_get_drvinfo,
898 .get_msglevel = ena_get_msglevel,
899 .set_msglevel = ena_set_msglevel,
900 .get_link = ethtool_op_get_link,
901 .get_coalesce = ena_get_coalesce,
902 .set_coalesce = ena_set_coalesce,
903 .get_ringparam = ena_get_ringparam,
904 .set_ringparam = ena_set_ringparam,
905 .get_sset_count = ena_get_sset_count,
906 .get_strings = ena_get_ethtool_strings,
907 .get_ethtool_stats = ena_get_ethtool_stats,
908 .get_rxnfc = ena_get_rxnfc,
909 .set_rxnfc = ena_set_rxnfc,
910 .get_rxfh_indir_size = ena_get_rxfh_indir_size,
911 .get_rxfh_key_size = ena_get_rxfh_key_size,
912 .get_rxfh = ena_get_rxfh,
913 .set_rxfh = ena_set_rxfh,
914 .get_channels = ena_get_channels,
915 .set_channels = ena_set_channels,
916 .get_tunable = ena_get_tunable,
917 .set_tunable = ena_set_tunable,
918 .get_ts_info = ethtool_op_get_ts_info,
919 };
920
ena_set_ethtool_ops(struct net_device * netdev)921 void ena_set_ethtool_ops(struct net_device *netdev)
922 {
923 netdev->ethtool_ops = &ena_ethtool_ops;
924 }
925
ena_dump_stats_ex(struct ena_adapter * adapter,u8 * buf)926 static void ena_dump_stats_ex(struct ena_adapter *adapter, u8 *buf)
927 {
928 struct net_device *netdev = adapter->netdev;
929 u8 *strings_buf;
930 u64 *data_buf;
931 int strings_num;
932 int i, rc;
933
934 strings_num = ena_get_sw_stats_count(adapter);
935 if (strings_num <= 0) {
936 netif_err(adapter, drv, netdev, "Can't get stats num\n");
937 return;
938 }
939
940 strings_buf = devm_kcalloc(&adapter->pdev->dev,
941 ETH_GSTRING_LEN, strings_num,
942 GFP_ATOMIC);
943 if (!strings_buf) {
944 netif_err(adapter, drv, netdev,
945 "Failed to allocate strings_buf\n");
946 return;
947 }
948
949 data_buf = devm_kcalloc(&adapter->pdev->dev,
950 strings_num, sizeof(u64),
951 GFP_ATOMIC);
952 if (!data_buf) {
953 netif_err(adapter, drv, netdev,
954 "Failed to allocate data buf\n");
955 devm_kfree(&adapter->pdev->dev, strings_buf);
956 return;
957 }
958
959 ena_get_strings(adapter, strings_buf, false);
960 ena_get_stats(adapter, data_buf, false);
961
962 /* If there is a buffer, dump stats, otherwise print them to dmesg */
963 if (buf)
964 for (i = 0; i < strings_num; i++) {
965 rc = snprintf(buf, ETH_GSTRING_LEN + sizeof(u64),
966 "%s %llu\n",
967 strings_buf + i * ETH_GSTRING_LEN,
968 data_buf[i]);
969 buf += rc;
970 }
971 else
972 for (i = 0; i < strings_num; i++)
973 netif_err(adapter, drv, netdev, "%s: %llu\n",
974 strings_buf + i * ETH_GSTRING_LEN,
975 data_buf[i]);
976
977 devm_kfree(&adapter->pdev->dev, strings_buf);
978 devm_kfree(&adapter->pdev->dev, data_buf);
979 }
980
ena_dump_stats_to_buf(struct ena_adapter * adapter,u8 * buf)981 void ena_dump_stats_to_buf(struct ena_adapter *adapter, u8 *buf)
982 {
983 if (!buf)
984 return;
985
986 ena_dump_stats_ex(adapter, buf);
987 }
988
ena_dump_stats_to_dmesg(struct ena_adapter * adapter)989 void ena_dump_stats_to_dmesg(struct ena_adapter *adapter)
990 {
991 ena_dump_stats_ex(adapter, NULL);
992 }
993