1 // SPDX-License-Identifier: GPL-2.0 OR Linux-OpenIB
3 * Copyright 2015-2020 Amazon.com, Inc. or its affiliates. All rights reserved.
6 #include <linux/ethtool.h>
9 #include "ena_netdev.h"
13 char name[ETH_GSTRING_LEN];
17 #define ENA_STAT_ENA_COM_ENTRY(stat) { \
19 .stat_offset = offsetof(struct ena_com_stats_admin, stat) / sizeof(u64) \
22 #define ENA_STAT_ENTRY(stat, stat_type) { \
24 .stat_offset = offsetof(struct ena_stats_##stat_type, stat) / sizeof(u64) \
27 #define ENA_STAT_HW_ENTRY(stat, stat_type) { \
29 .stat_offset = offsetof(struct ena_admin_##stat_type, stat) / sizeof(u64) \
32 #define ENA_STAT_RX_ENTRY(stat) \
33 ENA_STAT_ENTRY(stat, rx)
35 #define ENA_STAT_TX_ENTRY(stat) \
36 ENA_STAT_ENTRY(stat, tx)
38 #define ENA_STAT_GLOBAL_ENTRY(stat) \
39 ENA_STAT_ENTRY(stat, dev)
41 #define ENA_STAT_ENI_ENTRY(stat) \
42 ENA_STAT_HW_ENTRY(stat, eni_stats)
44 static const struct ena_stats ena_stats_global_strings[] = {
45 ENA_STAT_GLOBAL_ENTRY(tx_timeout),
46 ENA_STAT_GLOBAL_ENTRY(suspend),
47 ENA_STAT_GLOBAL_ENTRY(resume),
48 ENA_STAT_GLOBAL_ENTRY(wd_expired),
49 ENA_STAT_GLOBAL_ENTRY(interface_up),
50 ENA_STAT_GLOBAL_ENTRY(interface_down),
51 ENA_STAT_GLOBAL_ENTRY(admin_q_pause),
52 ENA_STAT_GLOBAL_ENTRY(reset_fail),
55 static const struct ena_stats ena_stats_eni_strings[] = {
56 ENA_STAT_ENI_ENTRY(bw_in_allowance_exceeded),
57 ENA_STAT_ENI_ENTRY(bw_out_allowance_exceeded),
58 ENA_STAT_ENI_ENTRY(pps_allowance_exceeded),
59 ENA_STAT_ENI_ENTRY(conntrack_allowance_exceeded),
60 ENA_STAT_ENI_ENTRY(linklocal_allowance_exceeded),
63 static const struct ena_stats ena_stats_tx_strings[] = {
64 ENA_STAT_TX_ENTRY(cnt),
65 ENA_STAT_TX_ENTRY(bytes),
66 ENA_STAT_TX_ENTRY(queue_stop),
67 ENA_STAT_TX_ENTRY(queue_wakeup),
68 ENA_STAT_TX_ENTRY(dma_mapping_err),
69 ENA_STAT_TX_ENTRY(linearize),
70 ENA_STAT_TX_ENTRY(linearize_failed),
71 ENA_STAT_TX_ENTRY(napi_comp),
72 ENA_STAT_TX_ENTRY(tx_poll),
73 ENA_STAT_TX_ENTRY(doorbells),
74 ENA_STAT_TX_ENTRY(prepare_ctx_err),
75 ENA_STAT_TX_ENTRY(bad_req_id),
76 ENA_STAT_TX_ENTRY(llq_buffer_copy),
77 ENA_STAT_TX_ENTRY(missed_tx),
78 ENA_STAT_TX_ENTRY(unmask_interrupt),
81 static const struct ena_stats ena_stats_rx_strings[] = {
82 ENA_STAT_RX_ENTRY(cnt),
83 ENA_STAT_RX_ENTRY(bytes),
84 ENA_STAT_RX_ENTRY(rx_copybreak_pkt),
85 ENA_STAT_RX_ENTRY(csum_good),
86 ENA_STAT_RX_ENTRY(refil_partial),
87 ENA_STAT_RX_ENTRY(csum_bad),
88 ENA_STAT_RX_ENTRY(page_alloc_fail),
89 ENA_STAT_RX_ENTRY(skb_alloc_fail),
90 ENA_STAT_RX_ENTRY(dma_mapping_err),
91 ENA_STAT_RX_ENTRY(bad_desc_num),
92 ENA_STAT_RX_ENTRY(bad_req_id),
93 ENA_STAT_RX_ENTRY(empty_rx_ring),
94 ENA_STAT_RX_ENTRY(csum_unchecked),
95 ENA_STAT_RX_ENTRY(xdp_aborted),
96 ENA_STAT_RX_ENTRY(xdp_drop),
97 ENA_STAT_RX_ENTRY(xdp_pass),
98 ENA_STAT_RX_ENTRY(xdp_tx),
99 ENA_STAT_RX_ENTRY(xdp_invalid),
100 ENA_STAT_RX_ENTRY(xdp_redirect),
103 static const struct ena_stats ena_stats_ena_com_strings[] = {
104 ENA_STAT_ENA_COM_ENTRY(aborted_cmd),
105 ENA_STAT_ENA_COM_ENTRY(submitted_cmd),
106 ENA_STAT_ENA_COM_ENTRY(completed_cmd),
107 ENA_STAT_ENA_COM_ENTRY(out_of_space),
108 ENA_STAT_ENA_COM_ENTRY(no_completion),
111 #define ENA_STATS_ARRAY_GLOBAL ARRAY_SIZE(ena_stats_global_strings)
112 #define ENA_STATS_ARRAY_TX ARRAY_SIZE(ena_stats_tx_strings)
113 #define ENA_STATS_ARRAY_RX ARRAY_SIZE(ena_stats_rx_strings)
114 #define ENA_STATS_ARRAY_ENA_COM ARRAY_SIZE(ena_stats_ena_com_strings)
115 #define ENA_STATS_ARRAY_ENI(adapter) ARRAY_SIZE(ena_stats_eni_strings)
117 static void ena_safe_update_stat(u64 *src, u64 *dst,
118 struct u64_stats_sync *syncp)
123 start = u64_stats_fetch_begin(syncp);
125 } while (u64_stats_fetch_retry(syncp, start));
128 static void ena_queue_stats(struct ena_adapter *adapter, u64 **data)
130 const struct ena_stats *ena_stats;
131 struct ena_ring *ring;
136 for (i = 0; i < adapter->num_io_queues + adapter->xdp_num_queues; i++) {
138 ring = &adapter->tx_ring[i];
140 for (j = 0; j < ENA_STATS_ARRAY_TX; j++) {
141 ena_stats = &ena_stats_tx_strings[j];
143 ptr = (u64 *)&ring->tx_stats + ena_stats->stat_offset;
145 ena_safe_update_stat(ptr, (*data)++, &ring->syncp);
147 /* XDP TX queues don't have a RX queue counterpart */
148 if (!ENA_IS_XDP_INDEX(adapter, i)) {
150 ring = &adapter->rx_ring[i];
152 for (j = 0; j < ENA_STATS_ARRAY_RX; j++) {
153 ena_stats = &ena_stats_rx_strings[j];
155 ptr = (u64 *)&ring->rx_stats +
156 ena_stats->stat_offset;
158 ena_safe_update_stat(ptr, (*data)++, &ring->syncp);
164 static void ena_dev_admin_queue_stats(struct ena_adapter *adapter, u64 **data)
166 const struct ena_stats *ena_stats;
170 for (i = 0; i < ENA_STATS_ARRAY_ENA_COM; i++) {
171 ena_stats = &ena_stats_ena_com_strings[i];
173 ptr = (u64 *)&adapter->ena_dev->admin_queue.stats +
174 ena_stats->stat_offset;
180 static void ena_get_stats(struct ena_adapter *adapter,
182 bool eni_stats_needed)
184 const struct ena_stats *ena_stats;
188 for (i = 0; i < ENA_STATS_ARRAY_GLOBAL; i++) {
189 ena_stats = &ena_stats_global_strings[i];
191 ptr = (u64 *)&adapter->dev_stats + ena_stats->stat_offset;
193 ena_safe_update_stat(ptr, data++, &adapter->syncp);
196 if (eni_stats_needed) {
197 ena_update_hw_stats(adapter);
198 for (i = 0; i < ENA_STATS_ARRAY_ENI(adapter); i++) {
199 ena_stats = &ena_stats_eni_strings[i];
201 ptr = (u64 *)&adapter->eni_stats +
202 ena_stats->stat_offset;
204 ena_safe_update_stat(ptr, data++, &adapter->syncp);
208 ena_queue_stats(adapter, &data);
209 ena_dev_admin_queue_stats(adapter, &data);
212 static void ena_get_ethtool_stats(struct net_device *netdev,
213 struct ethtool_stats *stats,
216 struct ena_adapter *adapter = netdev_priv(netdev);
217 struct ena_com_dev *dev = adapter->ena_dev;
219 ena_get_stats(adapter, data, ena_com_get_cap(dev, ENA_ADMIN_ENI_STATS));
222 static int ena_get_sw_stats_count(struct ena_adapter *adapter)
224 return adapter->num_io_queues * (ENA_STATS_ARRAY_TX + ENA_STATS_ARRAY_RX)
225 + adapter->xdp_num_queues * ENA_STATS_ARRAY_TX
226 + ENA_STATS_ARRAY_GLOBAL + ENA_STATS_ARRAY_ENA_COM;
229 static int ena_get_hw_stats_count(struct ena_adapter *adapter)
231 bool supported = ena_com_get_cap(adapter->ena_dev, ENA_ADMIN_ENI_STATS);
233 return ENA_STATS_ARRAY_ENI(adapter) * supported;
236 int ena_get_sset_count(struct net_device *netdev, int sset)
238 struct ena_adapter *adapter = netdev_priv(netdev);
242 return ena_get_sw_stats_count(adapter) +
243 ena_get_hw_stats_count(adapter);
249 static void ena_queue_strings(struct ena_adapter *adapter, u8 **data)
251 const struct ena_stats *ena_stats;
255 for (i = 0; i < adapter->num_io_queues + adapter->xdp_num_queues; i++) {
256 is_xdp = ENA_IS_XDP_INDEX(adapter, i);
258 for (j = 0; j < ENA_STATS_ARRAY_TX; j++) {
259 ena_stats = &ena_stats_tx_strings[j];
261 ethtool_sprintf(data,
263 is_xdp ? "xdp_tx" : "tx",
267 /* In XDP there isn't an RX queue counterpart */
271 for (j = 0; j < ENA_STATS_ARRAY_RX; j++) {
272 ena_stats = &ena_stats_rx_strings[j];
274 ethtool_sprintf(data, "queue_%u_rx_%s", i, ena_stats->name);
279 static void ena_com_dev_strings(u8 **data)
281 const struct ena_stats *ena_stats;
284 for (i = 0; i < ENA_STATS_ARRAY_ENA_COM; i++) {
285 ena_stats = &ena_stats_ena_com_strings[i];
287 ethtool_sprintf(data,
288 "ena_admin_q_%s", ena_stats->name);
292 static void ena_get_strings(struct ena_adapter *adapter,
294 bool eni_stats_needed)
296 const struct ena_stats *ena_stats;
299 for (i = 0; i < ENA_STATS_ARRAY_GLOBAL; i++) {
300 ena_stats = &ena_stats_global_strings[i];
301 ethtool_puts(&data, ena_stats->name);
304 if (eni_stats_needed) {
305 for (i = 0; i < ENA_STATS_ARRAY_ENI(adapter); i++) {
306 ena_stats = &ena_stats_eni_strings[i];
307 ethtool_puts(&data, ena_stats->name);
311 ena_queue_strings(adapter, &data);
312 ena_com_dev_strings(&data);
315 static void ena_get_ethtool_strings(struct net_device *netdev,
319 struct ena_adapter *adapter = netdev_priv(netdev);
320 struct ena_com_dev *dev = adapter->ena_dev;
324 ena_get_strings(adapter, data, ena_com_get_cap(dev, ENA_ADMIN_ENI_STATS));
329 static int ena_get_link_ksettings(struct net_device *netdev,
330 struct ethtool_link_ksettings *link_ksettings)
332 struct ena_adapter *adapter = netdev_priv(netdev);
333 struct ena_com_dev *ena_dev = adapter->ena_dev;
334 struct ena_admin_get_feature_link_desc *link;
335 struct ena_admin_get_feat_resp feat_resp;
338 rc = ena_com_get_link_params(ena_dev, &feat_resp);
342 link = &feat_resp.u.link;
343 link_ksettings->base.speed = link->speed;
345 if (link->flags & ENA_ADMIN_GET_FEATURE_LINK_DESC_AUTONEG_MASK) {
346 ethtool_link_ksettings_add_link_mode(link_ksettings,
348 ethtool_link_ksettings_add_link_mode(link_ksettings,
352 link_ksettings->base.autoneg =
353 (link->flags & ENA_ADMIN_GET_FEATURE_LINK_DESC_AUTONEG_MASK) ?
354 AUTONEG_ENABLE : AUTONEG_DISABLE;
356 link_ksettings->base.duplex = DUPLEX_FULL;
361 static int ena_get_coalesce(struct net_device *net_dev,
362 struct ethtool_coalesce *coalesce,
363 struct kernel_ethtool_coalesce *kernel_coal,
364 struct netlink_ext_ack *extack)
366 struct ena_adapter *adapter = netdev_priv(net_dev);
367 struct ena_com_dev *ena_dev = adapter->ena_dev;
369 if (!ena_com_interrupt_moderation_supported(ena_dev))
372 coalesce->tx_coalesce_usecs =
373 ena_com_get_nonadaptive_moderation_interval_tx(ena_dev) *
374 ena_dev->intr_delay_resolution;
376 coalesce->rx_coalesce_usecs =
377 ena_com_get_nonadaptive_moderation_interval_rx(ena_dev)
378 * ena_dev->intr_delay_resolution;
380 coalesce->use_adaptive_rx_coalesce =
381 ena_com_get_adaptive_moderation_enabled(ena_dev);
386 static void ena_update_tx_rings_nonadaptive_intr_moderation(struct ena_adapter *adapter)
391 val = ena_com_get_nonadaptive_moderation_interval_tx(adapter->ena_dev);
393 for (i = 0; i < adapter->num_io_queues; i++)
394 adapter->tx_ring[i].smoothed_interval = val;
397 static void ena_update_rx_rings_nonadaptive_intr_moderation(struct ena_adapter *adapter)
402 val = ena_com_get_nonadaptive_moderation_interval_rx(adapter->ena_dev);
404 for (i = 0; i < adapter->num_io_queues; i++)
405 adapter->rx_ring[i].smoothed_interval = val;
408 static int ena_set_coalesce(struct net_device *net_dev,
409 struct ethtool_coalesce *coalesce,
410 struct kernel_ethtool_coalesce *kernel_coal,
411 struct netlink_ext_ack *extack)
413 struct ena_adapter *adapter = netdev_priv(net_dev);
414 struct ena_com_dev *ena_dev = adapter->ena_dev;
417 if (!ena_com_interrupt_moderation_supported(ena_dev))
420 rc = ena_com_update_nonadaptive_moderation_interval_tx(ena_dev,
421 coalesce->tx_coalesce_usecs);
425 ena_update_tx_rings_nonadaptive_intr_moderation(adapter);
427 rc = ena_com_update_nonadaptive_moderation_interval_rx(ena_dev,
428 coalesce->rx_coalesce_usecs);
432 ena_update_rx_rings_nonadaptive_intr_moderation(adapter);
434 if (coalesce->use_adaptive_rx_coalesce &&
435 !ena_com_get_adaptive_moderation_enabled(ena_dev))
436 ena_com_enable_adaptive_moderation(ena_dev);
438 if (!coalesce->use_adaptive_rx_coalesce &&
439 ena_com_get_adaptive_moderation_enabled(ena_dev))
440 ena_com_disable_adaptive_moderation(ena_dev);
445 static u32 ena_get_msglevel(struct net_device *netdev)
447 struct ena_adapter *adapter = netdev_priv(netdev);
449 return adapter->msg_enable;
452 static void ena_set_msglevel(struct net_device *netdev, u32 value)
454 struct ena_adapter *adapter = netdev_priv(netdev);
456 adapter->msg_enable = value;
459 static void ena_get_drvinfo(struct net_device *dev,
460 struct ethtool_drvinfo *info)
462 struct ena_adapter *adapter = netdev_priv(dev);
465 ret = strscpy(info->driver, DRV_MODULE_NAME, sizeof(info->driver));
467 netif_dbg(adapter, drv, dev,
468 "module name will be truncated, status = %zd\n", ret);
470 ret = strscpy(info->bus_info, pci_name(adapter->pdev),
471 sizeof(info->bus_info));
473 netif_dbg(adapter, drv, dev,
474 "bus info will be truncated, status = %zd\n", ret);
477 static void ena_get_ringparam(struct net_device *netdev,
478 struct ethtool_ringparam *ring,
479 struct kernel_ethtool_ringparam *kernel_ring,
480 struct netlink_ext_ack *extack)
482 struct ena_adapter *adapter = netdev_priv(netdev);
484 ring->tx_max_pending = adapter->max_tx_ring_size;
485 ring->rx_max_pending = adapter->max_rx_ring_size;
486 if (adapter->ena_dev->tx_mem_queue_type == ENA_ADMIN_PLACEMENT_POLICY_DEV) {
487 bool large_llq_supported = adapter->large_llq_header_supported;
489 kernel_ring->tx_push = true;
490 kernel_ring->tx_push_buf_len = adapter->ena_dev->tx_max_header_size;
491 if (large_llq_supported)
492 kernel_ring->tx_push_buf_max_len = ENA_LLQ_LARGE_HEADER;
494 kernel_ring->tx_push_buf_max_len = ENA_LLQ_HEADER;
496 kernel_ring->tx_push = false;
497 kernel_ring->tx_push_buf_max_len = 0;
498 kernel_ring->tx_push_buf_len = 0;
501 ring->tx_pending = adapter->tx_ring[0].ring_size;
502 ring->rx_pending = adapter->rx_ring[0].ring_size;
505 static int ena_set_ringparam(struct net_device *netdev,
506 struct ethtool_ringparam *ring,
507 struct kernel_ethtool_ringparam *kernel_ring,
508 struct netlink_ext_ack *extack)
510 struct ena_adapter *adapter = netdev_priv(netdev);
511 u32 new_tx_size, new_rx_size, new_tx_push_buf_len;
512 bool changed = false;
514 new_tx_size = ring->tx_pending < ENA_MIN_RING_SIZE ?
515 ENA_MIN_RING_SIZE : ring->tx_pending;
516 new_tx_size = rounddown_pow_of_two(new_tx_size);
518 new_rx_size = ring->rx_pending < ENA_MIN_RING_SIZE ?
519 ENA_MIN_RING_SIZE : ring->rx_pending;
520 new_rx_size = rounddown_pow_of_two(new_rx_size);
522 changed |= new_tx_size != adapter->requested_tx_ring_size ||
523 new_rx_size != adapter->requested_rx_ring_size;
525 /* This value is ignored if LLQ is not supported */
526 new_tx_push_buf_len = adapter->ena_dev->tx_max_header_size;
528 if ((adapter->ena_dev->tx_mem_queue_type == ENA_ADMIN_PLACEMENT_POLICY_DEV) !=
529 kernel_ring->tx_push) {
530 NL_SET_ERR_MSG_MOD(extack, "Push mode state cannot be modified");
534 /* Validate that the push buffer is supported on the underlying device */
535 if (kernel_ring->tx_push_buf_len) {
536 enum ena_admin_placement_policy_type placement;
538 new_tx_push_buf_len = kernel_ring->tx_push_buf_len;
540 placement = adapter->ena_dev->tx_mem_queue_type;
541 if (placement == ENA_ADMIN_PLACEMENT_POLICY_HOST)
544 if (new_tx_push_buf_len != ENA_LLQ_HEADER &&
545 new_tx_push_buf_len != ENA_LLQ_LARGE_HEADER) {
546 bool large_llq_sup = adapter->large_llq_header_supported;
547 char large_llq_size_str[40];
549 snprintf(large_llq_size_str, 40, ", %lu", ENA_LLQ_LARGE_HEADER);
551 NL_SET_ERR_MSG_FMT_MOD(extack,
552 "Supported tx push buff values: [%lu%s]",
554 large_llq_sup ? large_llq_size_str : "");
559 changed |= new_tx_push_buf_len != adapter->ena_dev->tx_max_header_size;
565 return ena_update_queue_params(adapter, new_tx_size, new_rx_size,
566 new_tx_push_buf_len);
569 static u32 ena_flow_hash_to_flow_type(u16 hash_fields)
573 if (hash_fields & ENA_ADMIN_RSS_L2_DA)
576 if (hash_fields & ENA_ADMIN_RSS_L3_DA)
579 if (hash_fields & ENA_ADMIN_RSS_L3_SA)
582 if (hash_fields & ENA_ADMIN_RSS_L4_DP)
583 data |= RXH_L4_B_2_3;
585 if (hash_fields & ENA_ADMIN_RSS_L4_SP)
586 data |= RXH_L4_B_0_1;
591 static u16 ena_flow_data_to_flow_hash(u32 hash_fields)
595 if (hash_fields & RXH_L2DA)
596 data |= ENA_ADMIN_RSS_L2_DA;
598 if (hash_fields & RXH_IP_DST)
599 data |= ENA_ADMIN_RSS_L3_DA;
601 if (hash_fields & RXH_IP_SRC)
602 data |= ENA_ADMIN_RSS_L3_SA;
604 if (hash_fields & RXH_L4_B_2_3)
605 data |= ENA_ADMIN_RSS_L4_DP;
607 if (hash_fields & RXH_L4_B_0_1)
608 data |= ENA_ADMIN_RSS_L4_SP;
613 static int ena_get_rss_hash(struct ena_com_dev *ena_dev,
614 struct ethtool_rxnfc *cmd)
616 enum ena_admin_flow_hash_proto proto;
622 switch (cmd->flow_type) {
624 proto = ENA_ADMIN_RSS_TCP4;
627 proto = ENA_ADMIN_RSS_UDP4;
630 proto = ENA_ADMIN_RSS_TCP6;
633 proto = ENA_ADMIN_RSS_UDP6;
636 proto = ENA_ADMIN_RSS_IP4;
639 proto = ENA_ADMIN_RSS_IP6;
642 proto = ENA_ADMIN_RSS_NOT_IP;
655 rc = ena_com_get_hash_ctrl(ena_dev, proto, &hash_fields);
659 cmd->data = ena_flow_hash_to_flow_type(hash_fields);
664 static int ena_set_rss_hash(struct ena_com_dev *ena_dev,
665 struct ethtool_rxnfc *cmd)
667 enum ena_admin_flow_hash_proto proto;
670 switch (cmd->flow_type) {
672 proto = ENA_ADMIN_RSS_TCP4;
675 proto = ENA_ADMIN_RSS_UDP4;
678 proto = ENA_ADMIN_RSS_TCP6;
681 proto = ENA_ADMIN_RSS_UDP6;
684 proto = ENA_ADMIN_RSS_IP4;
687 proto = ENA_ADMIN_RSS_IP6;
690 proto = ENA_ADMIN_RSS_NOT_IP;
703 hash_fields = ena_flow_data_to_flow_hash(cmd->data);
705 return ena_com_fill_hash_ctrl(ena_dev, proto, hash_fields);
708 static int ena_set_rxnfc(struct net_device *netdev, struct ethtool_rxnfc *info)
710 struct ena_adapter *adapter = netdev_priv(netdev);
715 rc = ena_set_rss_hash(adapter->ena_dev, info);
717 case ETHTOOL_SRXCLSRLDEL:
718 case ETHTOOL_SRXCLSRLINS:
720 netif_err(adapter, drv, netdev,
721 "Command parameter %d is not supported\n", info->cmd);
728 static int ena_get_rxnfc(struct net_device *netdev, struct ethtool_rxnfc *info,
731 struct ena_adapter *adapter = netdev_priv(netdev);
735 case ETHTOOL_GRXRINGS:
736 info->data = adapter->num_io_queues;
740 rc = ena_get_rss_hash(adapter->ena_dev, info);
742 case ETHTOOL_GRXCLSRLCNT:
743 case ETHTOOL_GRXCLSRULE:
744 case ETHTOOL_GRXCLSRLALL:
746 netif_err(adapter, drv, netdev,
747 "Command parameter %d is not supported\n", info->cmd);
754 static u32 ena_get_rxfh_indir_size(struct net_device *netdev)
756 return ENA_RX_RSS_TABLE_SIZE;
759 static u32 ena_get_rxfh_key_size(struct net_device *netdev)
761 return ENA_HASH_KEY_SIZE;
764 static int ena_indirection_table_set(struct ena_adapter *adapter,
767 struct ena_com_dev *ena_dev = adapter->ena_dev;
770 for (i = 0; i < ENA_RX_RSS_TABLE_SIZE; i++) {
771 rc = ena_com_indirect_table_fill_entry(ena_dev,
773 ENA_IO_RXQ_IDX(indir[i]));
775 netif_err(adapter, drv, adapter->netdev,
776 "Cannot fill indirect table (index is too large)\n");
781 rc = ena_com_indirect_table_set(ena_dev);
783 netif_err(adapter, drv, adapter->netdev,
784 "Cannot set indirect table\n");
785 return rc == -EPERM ? -EOPNOTSUPP : rc;
790 static int ena_indirection_table_get(struct ena_adapter *adapter, u32 *indir)
792 struct ena_com_dev *ena_dev = adapter->ena_dev;
798 rc = ena_com_indirect_table_get(ena_dev, indir);
802 /* Our internal representation of the indices is: even indices
803 * for Tx and uneven indices for Rx. We need to convert the Rx
804 * indices to be consecutive
806 for (i = 0; i < ENA_RX_RSS_TABLE_SIZE; i++)
807 indir[i] = ENA_IO_RXQ_IDX_TO_COMBINED_IDX(indir[i]);
812 static int ena_get_rxfh(struct net_device *netdev,
813 struct ethtool_rxfh_param *rxfh)
815 struct ena_adapter *adapter = netdev_priv(netdev);
816 enum ena_admin_hash_functions ena_func;
820 rc = ena_indirection_table_get(adapter, rxfh->indir);
824 /* We call this function in order to check if the device
825 * supports getting/setting the hash function.
827 rc = ena_com_get_hash_function(adapter->ena_dev, &ena_func);
829 if (rc == -EOPNOTSUPP)
835 rc = ena_com_get_hash_key(adapter->ena_dev, rxfh->key);
840 case ENA_ADMIN_TOEPLITZ:
841 func = ETH_RSS_HASH_TOP;
843 case ENA_ADMIN_CRC32:
844 func = ETH_RSS_HASH_CRC32;
847 netif_err(adapter, drv, netdev,
848 "Command parameter is not supported\n");
857 static int ena_set_rxfh(struct net_device *netdev,
858 struct ethtool_rxfh_param *rxfh,
859 struct netlink_ext_ack *extack)
861 struct ena_adapter *adapter = netdev_priv(netdev);
862 struct ena_com_dev *ena_dev = adapter->ena_dev;
863 enum ena_admin_hash_functions func = 0;
867 rc = ena_indirection_table_set(adapter, rxfh->indir);
872 switch (rxfh->hfunc) {
873 case ETH_RSS_HASH_NO_CHANGE:
874 func = ena_com_get_current_hash_function(ena_dev);
876 case ETH_RSS_HASH_TOP:
877 func = ENA_ADMIN_TOEPLITZ;
879 case ETH_RSS_HASH_CRC32:
880 func = ENA_ADMIN_CRC32;
883 netif_err(adapter, drv, netdev, "Unsupported hfunc %d\n",
888 if (rxfh->key || func) {
889 rc = ena_com_fill_hash_function(ena_dev, func, rxfh->key,
893 netif_err(adapter, drv, netdev, "Cannot fill key\n");
894 return rc == -EPERM ? -EOPNOTSUPP : rc;
901 static void ena_get_channels(struct net_device *netdev,
902 struct ethtool_channels *channels)
904 struct ena_adapter *adapter = netdev_priv(netdev);
906 channels->max_combined = adapter->max_num_io_queues;
907 channels->combined_count = adapter->num_io_queues;
910 static int ena_set_channels(struct net_device *netdev,
911 struct ethtool_channels *channels)
913 struct ena_adapter *adapter = netdev_priv(netdev);
914 u32 count = channels->combined_count;
915 /* The check for max value is already done in ethtool */
916 if (count < ENA_MIN_NUM_IO_QUEUES)
919 if (!ena_xdp_legal_queue_count(adapter, count)) {
920 if (ena_xdp_present(adapter))
923 xdp_clear_features_flag(netdev);
925 xdp_set_features_flag(netdev,
926 NETDEV_XDP_ACT_BASIC |
927 NETDEV_XDP_ACT_REDIRECT);
930 return ena_update_queue_count(adapter, count);
933 static int ena_get_tunable(struct net_device *netdev,
934 const struct ethtool_tunable *tuna, void *data)
936 struct ena_adapter *adapter = netdev_priv(netdev);
940 case ETHTOOL_RX_COPYBREAK:
941 *(u32 *)data = adapter->rx_copybreak;
951 static int ena_set_tunable(struct net_device *netdev,
952 const struct ethtool_tunable *tuna,
955 struct ena_adapter *adapter = netdev_priv(netdev);
960 case ETHTOOL_RX_COPYBREAK:
962 ret = ena_set_rx_copybreak(adapter, len);
972 static const struct ethtool_ops ena_ethtool_ops = {
973 .supported_coalesce_params = ETHTOOL_COALESCE_USECS |
974 ETHTOOL_COALESCE_USE_ADAPTIVE_RX,
975 .supported_ring_params = ETHTOOL_RING_USE_TX_PUSH_BUF_LEN |
976 ETHTOOL_RING_USE_TX_PUSH,
977 .get_link_ksettings = ena_get_link_ksettings,
978 .get_drvinfo = ena_get_drvinfo,
979 .get_msglevel = ena_get_msglevel,
980 .set_msglevel = ena_set_msglevel,
981 .get_link = ethtool_op_get_link,
982 .get_coalesce = ena_get_coalesce,
983 .set_coalesce = ena_set_coalesce,
984 .get_ringparam = ena_get_ringparam,
985 .set_ringparam = ena_set_ringparam,
986 .get_sset_count = ena_get_sset_count,
987 .get_strings = ena_get_ethtool_strings,
988 .get_ethtool_stats = ena_get_ethtool_stats,
989 .get_rxnfc = ena_get_rxnfc,
990 .set_rxnfc = ena_set_rxnfc,
991 .get_rxfh_indir_size = ena_get_rxfh_indir_size,
992 .get_rxfh_key_size = ena_get_rxfh_key_size,
993 .get_rxfh = ena_get_rxfh,
994 .set_rxfh = ena_set_rxfh,
995 .get_channels = ena_get_channels,
996 .set_channels = ena_set_channels,
997 .get_tunable = ena_get_tunable,
998 .set_tunable = ena_set_tunable,
999 .get_ts_info = ethtool_op_get_ts_info,
1002 void ena_set_ethtool_ops(struct net_device *netdev)
1004 netdev->ethtool_ops = &ena_ethtool_ops;
1007 static void ena_dump_stats_ex(struct ena_adapter *adapter, u8 *buf)
1009 struct net_device *netdev = adapter->netdev;
1015 strings_num = ena_get_sw_stats_count(adapter);
1016 if (strings_num <= 0) {
1017 netif_err(adapter, drv, netdev, "Can't get stats num\n");
1021 strings_buf = devm_kcalloc(&adapter->pdev->dev,
1022 ETH_GSTRING_LEN, strings_num,
1025 netif_err(adapter, drv, netdev,
1026 "Failed to allocate strings_buf\n");
1030 data_buf = devm_kcalloc(&adapter->pdev->dev,
1031 strings_num, sizeof(u64),
1034 netif_err(adapter, drv, netdev,
1035 "Failed to allocate data buf\n");
1036 devm_kfree(&adapter->pdev->dev, strings_buf);
1040 ena_get_strings(adapter, strings_buf, false);
1041 ena_get_stats(adapter, data_buf, false);
1043 /* If there is a buffer, dump stats, otherwise print them to dmesg */
1045 for (i = 0; i < strings_num; i++) {
1046 rc = snprintf(buf, ETH_GSTRING_LEN + sizeof(u64),
1048 strings_buf + i * ETH_GSTRING_LEN,
1053 for (i = 0; i < strings_num; i++)
1054 netif_err(adapter, drv, netdev, "%s: %llu\n",
1055 strings_buf + i * ETH_GSTRING_LEN,
1058 devm_kfree(&adapter->pdev->dev, strings_buf);
1059 devm_kfree(&adapter->pdev->dev, data_buf);
1062 void ena_dump_stats_to_buf(struct ena_adapter *adapter, u8 *buf)
1067 ena_dump_stats_ex(adapter, buf);
1070 void ena_dump_stats_to_dmesg(struct ena_adapter *adapter)
1072 ena_dump_stats_ex(adapter, NULL);