1// SPDX-License-Identifier: GPL-2.0 OR BSD-3-Clause
2/* Copyright (c) 2021, Microsoft Corporation. */
3
4#include <linux/inetdevice.h>
5#include <linux/etherdevice.h>
6#include <linux/ethtool.h>
7
8#include <net/mana/mana.h>
9
10struct mana_stats_desc {
11 char name[ETH_GSTRING_LEN];
12 u16 offset;
13};
14
15static const struct mana_stats_desc mana_eth_stats[] = {
16 {"stop_queue", offsetof(struct mana_ethtool_stats, stop_queue)},
17 {"wake_queue", offsetof(struct mana_ethtool_stats, wake_queue)},
18 {"tx_cq_err", offsetof(struct mana_ethtool_stats, tx_cqe_err)},
19 {"tx_cqe_unknown_type", offsetof(struct mana_ethtool_stats,
20 tx_cqe_unknown_type)},
21 {"tx_linear_pkt_cnt", offsetof(struct mana_ethtool_stats,
22 tx_linear_pkt_cnt)},
23 {"rx_coalesced_err", offsetof(struct mana_ethtool_stats,
24 rx_coalesced_err)},
25 {"rx_cqe_unknown_type", offsetof(struct mana_ethtool_stats,
26 rx_cqe_unknown_type)},
27};
28
29static const struct mana_stats_desc mana_hc_stats[] = {
30 {"hc_rx_discards_no_wqe", offsetof(struct mana_ethtool_hc_stats,
31 hc_rx_discards_no_wqe)},
32 {"hc_rx_err_vport_disabled", offsetof(struct mana_ethtool_hc_stats,
33 hc_rx_err_vport_disabled)},
34 {"hc_rx_bytes", offsetof(struct mana_ethtool_hc_stats, hc_rx_bytes)},
35 {"hc_rx_ucast_pkts", offsetof(struct mana_ethtool_hc_stats,
36 hc_rx_ucast_pkts)},
37 {"hc_rx_ucast_bytes", offsetof(struct mana_ethtool_hc_stats,
38 hc_rx_ucast_bytes)},
39 {"hc_rx_bcast_pkts", offsetof(struct mana_ethtool_hc_stats,
40 hc_rx_bcast_pkts)},
41 {"hc_rx_bcast_bytes", offsetof(struct mana_ethtool_hc_stats,
42 hc_rx_bcast_bytes)},
43 {"hc_rx_mcast_pkts", offsetof(struct mana_ethtool_hc_stats,
44 hc_rx_mcast_pkts)},
45 {"hc_rx_mcast_bytes", offsetof(struct mana_ethtool_hc_stats,
46 hc_rx_mcast_bytes)},
47 {"hc_tx_err_gf_disabled", offsetof(struct mana_ethtool_hc_stats,
48 hc_tx_err_gf_disabled)},
49 {"hc_tx_err_vport_disabled", offsetof(struct mana_ethtool_hc_stats,
50 hc_tx_err_vport_disabled)},
51 {"hc_tx_err_inval_vportoffset_pkt",
52 offsetof(struct mana_ethtool_hc_stats,
53 hc_tx_err_inval_vportoffset_pkt)},
54 {"hc_tx_err_vlan_enforcement", offsetof(struct mana_ethtool_hc_stats,
55 hc_tx_err_vlan_enforcement)},
56 {"hc_tx_err_eth_type_enforcement",
57 offsetof(struct mana_ethtool_hc_stats, hc_tx_err_eth_type_enforcement)},
58 {"hc_tx_err_sa_enforcement", offsetof(struct mana_ethtool_hc_stats,
59 hc_tx_err_sa_enforcement)},
60 {"hc_tx_err_sqpdid_enforcement",
61 offsetof(struct mana_ethtool_hc_stats, hc_tx_err_sqpdid_enforcement)},
62 {"hc_tx_err_cqpdid_enforcement",
63 offsetof(struct mana_ethtool_hc_stats, hc_tx_err_cqpdid_enforcement)},
64 {"hc_tx_err_mtu_violation", offsetof(struct mana_ethtool_hc_stats,
65 hc_tx_err_mtu_violation)},
66 {"hc_tx_err_inval_oob", offsetof(struct mana_ethtool_hc_stats,
67 hc_tx_err_inval_oob)},
68 {"hc_tx_err_gdma", offsetof(struct mana_ethtool_hc_stats,
69 hc_tx_err_gdma)},
70 {"hc_tx_bytes", offsetof(struct mana_ethtool_hc_stats, hc_tx_bytes)},
71 {"hc_tx_ucast_pkts", offsetof(struct mana_ethtool_hc_stats,
72 hc_tx_ucast_pkts)},
73 {"hc_tx_ucast_bytes", offsetof(struct mana_ethtool_hc_stats,
74 hc_tx_ucast_bytes)},
75 {"hc_tx_bcast_pkts", offsetof(struct mana_ethtool_hc_stats,
76 hc_tx_bcast_pkts)},
77 {"hc_tx_bcast_bytes", offsetof(struct mana_ethtool_hc_stats,
78 hc_tx_bcast_bytes)},
79 {"hc_tx_mcast_pkts", offsetof(struct mana_ethtool_hc_stats,
80 hc_tx_mcast_pkts)},
81 {"hc_tx_mcast_bytes", offsetof(struct mana_ethtool_hc_stats,
82 hc_tx_mcast_bytes)},
83};
84
85static const struct mana_stats_desc mana_phy_stats[] = {
86 { "hc_rx_pkt_drop_phy", offsetof(struct mana_ethtool_phy_stats, rx_pkt_drop_phy) },
87 { "hc_tx_pkt_drop_phy", offsetof(struct mana_ethtool_phy_stats, tx_pkt_drop_phy) },
88 { "hc_tc0_rx_pkt_phy", offsetof(struct mana_ethtool_phy_stats, rx_pkt_tc0_phy) },
89 { "hc_tc0_rx_byte_phy", offsetof(struct mana_ethtool_phy_stats, rx_byte_tc0_phy) },
90 { "hc_tc0_tx_pkt_phy", offsetof(struct mana_ethtool_phy_stats, tx_pkt_tc0_phy) },
91 { "hc_tc0_tx_byte_phy", offsetof(struct mana_ethtool_phy_stats, tx_byte_tc0_phy) },
92 { "hc_tc1_rx_pkt_phy", offsetof(struct mana_ethtool_phy_stats, rx_pkt_tc1_phy) },
93 { "hc_tc1_rx_byte_phy", offsetof(struct mana_ethtool_phy_stats, rx_byte_tc1_phy) },
94 { "hc_tc1_tx_pkt_phy", offsetof(struct mana_ethtool_phy_stats, tx_pkt_tc1_phy) },
95 { "hc_tc1_tx_byte_phy", offsetof(struct mana_ethtool_phy_stats, tx_byte_tc1_phy) },
96 { "hc_tc2_rx_pkt_phy", offsetof(struct mana_ethtool_phy_stats, rx_pkt_tc2_phy) },
97 { "hc_tc2_rx_byte_phy", offsetof(struct mana_ethtool_phy_stats, rx_byte_tc2_phy) },
98 { "hc_tc2_tx_pkt_phy", offsetof(struct mana_ethtool_phy_stats, tx_pkt_tc2_phy) },
99 { "hc_tc2_tx_byte_phy", offsetof(struct mana_ethtool_phy_stats, tx_byte_tc2_phy) },
100 { "hc_tc3_rx_pkt_phy", offsetof(struct mana_ethtool_phy_stats, rx_pkt_tc3_phy) },
101 { "hc_tc3_rx_byte_phy", offsetof(struct mana_ethtool_phy_stats, rx_byte_tc3_phy) },
102 { "hc_tc3_tx_pkt_phy", offsetof(struct mana_ethtool_phy_stats, tx_pkt_tc3_phy) },
103 { "hc_tc3_tx_byte_phy", offsetof(struct mana_ethtool_phy_stats, tx_byte_tc3_phy) },
104 { "hc_tc4_rx_pkt_phy", offsetof(struct mana_ethtool_phy_stats, rx_pkt_tc4_phy) },
105 { "hc_tc4_rx_byte_phy", offsetof(struct mana_ethtool_phy_stats, rx_byte_tc4_phy) },
106 { "hc_tc4_tx_pkt_phy", offsetof(struct mana_ethtool_phy_stats, tx_pkt_tc4_phy) },
107 { "hc_tc4_tx_byte_phy", offsetof(struct mana_ethtool_phy_stats, tx_byte_tc4_phy) },
108 { "hc_tc5_rx_pkt_phy", offsetof(struct mana_ethtool_phy_stats, rx_pkt_tc5_phy) },
109 { "hc_tc5_rx_byte_phy", offsetof(struct mana_ethtool_phy_stats, rx_byte_tc5_phy) },
110 { "hc_tc5_tx_pkt_phy", offsetof(struct mana_ethtool_phy_stats, tx_pkt_tc5_phy) },
111 { "hc_tc5_tx_byte_phy", offsetof(struct mana_ethtool_phy_stats, tx_byte_tc5_phy) },
112 { "hc_tc6_rx_pkt_phy", offsetof(struct mana_ethtool_phy_stats, rx_pkt_tc6_phy) },
113 { "hc_tc6_rx_byte_phy", offsetof(struct mana_ethtool_phy_stats, rx_byte_tc6_phy) },
114 { "hc_tc6_tx_pkt_phy", offsetof(struct mana_ethtool_phy_stats, tx_pkt_tc6_phy) },
115 { "hc_tc6_tx_byte_phy", offsetof(struct mana_ethtool_phy_stats, tx_byte_tc6_phy) },
116 { "hc_tc7_rx_pkt_phy", offsetof(struct mana_ethtool_phy_stats, rx_pkt_tc7_phy) },
117 { "hc_tc7_rx_byte_phy", offsetof(struct mana_ethtool_phy_stats, rx_byte_tc7_phy) },
118 { "hc_tc7_tx_pkt_phy", offsetof(struct mana_ethtool_phy_stats, tx_pkt_tc7_phy) },
119 { "hc_tc7_tx_byte_phy", offsetof(struct mana_ethtool_phy_stats, tx_byte_tc7_phy) },
120 { "hc_tc0_rx_pause_phy", offsetof(struct mana_ethtool_phy_stats, rx_pause_tc0_phy) },
121 { "hc_tc0_tx_pause_phy", offsetof(struct mana_ethtool_phy_stats, tx_pause_tc0_phy) },
122 { "hc_tc1_rx_pause_phy", offsetof(struct mana_ethtool_phy_stats, rx_pause_tc1_phy) },
123 { "hc_tc1_tx_pause_phy", offsetof(struct mana_ethtool_phy_stats, tx_pause_tc1_phy) },
124 { "hc_tc2_rx_pause_phy", offsetof(struct mana_ethtool_phy_stats, rx_pause_tc2_phy) },
125 { "hc_tc2_tx_pause_phy", offsetof(struct mana_ethtool_phy_stats, tx_pause_tc2_phy) },
126 { "hc_tc3_rx_pause_phy", offsetof(struct mana_ethtool_phy_stats, rx_pause_tc3_phy) },
127 { "hc_tc3_tx_pause_phy", offsetof(struct mana_ethtool_phy_stats, tx_pause_tc3_phy) },
128 { "hc_tc4_rx_pause_phy", offsetof(struct mana_ethtool_phy_stats, rx_pause_tc4_phy) },
129 { "hc_tc4_tx_pause_phy", offsetof(struct mana_ethtool_phy_stats, tx_pause_tc4_phy) },
130 { "hc_tc5_rx_pause_phy", offsetof(struct mana_ethtool_phy_stats, rx_pause_tc5_phy) },
131 { "hc_tc5_tx_pause_phy", offsetof(struct mana_ethtool_phy_stats, tx_pause_tc5_phy) },
132 { "hc_tc6_rx_pause_phy", offsetof(struct mana_ethtool_phy_stats, rx_pause_tc6_phy) },
133 { "hc_tc6_tx_pause_phy", offsetof(struct mana_ethtool_phy_stats, tx_pause_tc6_phy) },
134 { "hc_tc7_rx_pause_phy", offsetof(struct mana_ethtool_phy_stats, rx_pause_tc7_phy) },
135 { "hc_tc7_tx_pause_phy", offsetof(struct mana_ethtool_phy_stats, tx_pause_tc7_phy) },
136};
137
138static int mana_get_sset_count(struct net_device *ndev, int stringset)
139{
140 struct mana_port_context *apc = netdev_priv(dev: ndev);
141 unsigned int num_queues = apc->num_queues;
142
143 if (stringset != ETH_SS_STATS)
144 return -EINVAL;
145
146 return ARRAY_SIZE(mana_eth_stats) + ARRAY_SIZE(mana_phy_stats) + ARRAY_SIZE(mana_hc_stats) +
147 num_queues * (MANA_STATS_RX_COUNT + MANA_STATS_TX_COUNT);
148}
149
150static void mana_get_strings(struct net_device *ndev, u32 stringset, u8 *data)
151{
152 struct mana_port_context *apc = netdev_priv(dev: ndev);
153 unsigned int num_queues = apc->num_queues;
154 int i;
155
156 if (stringset != ETH_SS_STATS)
157 return;
158 for (i = 0; i < ARRAY_SIZE(mana_eth_stats); i++)
159 ethtool_puts(data: &data, str: mana_eth_stats[i].name);
160
161 for (i = 0; i < ARRAY_SIZE(mana_hc_stats); i++)
162 ethtool_puts(data: &data, str: mana_hc_stats[i].name);
163
164 for (i = 0; i < ARRAY_SIZE(mana_phy_stats); i++)
165 ethtool_puts(data: &data, str: mana_phy_stats[i].name);
166
167 for (i = 0; i < num_queues; i++) {
168 ethtool_sprintf(data: &data, fmt: "rx_%d_packets", i);
169 ethtool_sprintf(data: &data, fmt: "rx_%d_bytes", i);
170 ethtool_sprintf(data: &data, fmt: "rx_%d_xdp_drop", i);
171 ethtool_sprintf(data: &data, fmt: "rx_%d_xdp_tx", i);
172 ethtool_sprintf(data: &data, fmt: "rx_%d_xdp_redirect", i);
173 }
174
175 for (i = 0; i < num_queues; i++) {
176 ethtool_sprintf(data: &data, fmt: "tx_%d_packets", i);
177 ethtool_sprintf(data: &data, fmt: "tx_%d_bytes", i);
178 ethtool_sprintf(data: &data, fmt: "tx_%d_xdp_xmit", i);
179 ethtool_sprintf(data: &data, fmt: "tx_%d_tso_packets", i);
180 ethtool_sprintf(data: &data, fmt: "tx_%d_tso_bytes", i);
181 ethtool_sprintf(data: &data, fmt: "tx_%d_tso_inner_packets", i);
182 ethtool_sprintf(data: &data, fmt: "tx_%d_tso_inner_bytes", i);
183 ethtool_sprintf(data: &data, fmt: "tx_%d_long_pkt_fmt", i);
184 ethtool_sprintf(data: &data, fmt: "tx_%d_short_pkt_fmt", i);
185 ethtool_sprintf(data: &data, fmt: "tx_%d_csum_partial", i);
186 ethtool_sprintf(data: &data, fmt: "tx_%d_mana_map_err", i);
187 }
188}
189
190static void mana_get_ethtool_stats(struct net_device *ndev,
191 struct ethtool_stats *e_stats, u64 *data)
192{
193 struct mana_port_context *apc = netdev_priv(dev: ndev);
194 unsigned int num_queues = apc->num_queues;
195 void *eth_stats = &apc->eth_stats;
196 void *hc_stats = &apc->ac->hc_stats;
197 void *phy_stats = &apc->phy_stats;
198 struct mana_stats_rx *rx_stats;
199 struct mana_stats_tx *tx_stats;
200 unsigned int start;
201 u64 packets, bytes;
202 u64 xdp_redirect;
203 u64 xdp_xmit;
204 u64 xdp_drop;
205 u64 xdp_tx;
206 u64 tso_packets;
207 u64 tso_bytes;
208 u64 tso_inner_packets;
209 u64 tso_inner_bytes;
210 u64 long_pkt_fmt;
211 u64 short_pkt_fmt;
212 u64 csum_partial;
213 u64 mana_map_err;
214 int q, i = 0;
215
216 if (!apc->port_is_up)
217 return;
218
219 /* We call this mana function to get the phy stats from GDMA and includes
220 * aggregate tx/rx drop counters, Per-TC(Traffic Channel) tx/rx and pause
221 * counters.
222 */
223 mana_query_phy_stats(apc);
224
225 for (q = 0; q < ARRAY_SIZE(mana_eth_stats); q++)
226 data[i++] = *(u64 *)(eth_stats + mana_eth_stats[q].offset);
227
228 for (q = 0; q < ARRAY_SIZE(mana_hc_stats); q++)
229 data[i++] = *(u64 *)(hc_stats + mana_hc_stats[q].offset);
230
231 for (q = 0; q < ARRAY_SIZE(mana_phy_stats); q++)
232 data[i++] = *(u64 *)(phy_stats + mana_phy_stats[q].offset);
233
234 for (q = 0; q < num_queues; q++) {
235 rx_stats = &apc->rxqs[q]->stats;
236
237 do {
238 start = u64_stats_fetch_begin(syncp: &rx_stats->syncp);
239 packets = rx_stats->packets;
240 bytes = rx_stats->bytes;
241 xdp_drop = rx_stats->xdp_drop;
242 xdp_tx = rx_stats->xdp_tx;
243 xdp_redirect = rx_stats->xdp_redirect;
244 } while (u64_stats_fetch_retry(syncp: &rx_stats->syncp, start));
245
246 data[i++] = packets;
247 data[i++] = bytes;
248 data[i++] = xdp_drop;
249 data[i++] = xdp_tx;
250 data[i++] = xdp_redirect;
251 }
252
253 for (q = 0; q < num_queues; q++) {
254 tx_stats = &apc->tx_qp[q].txq.stats;
255
256 do {
257 start = u64_stats_fetch_begin(syncp: &tx_stats->syncp);
258 packets = tx_stats->packets;
259 bytes = tx_stats->bytes;
260 xdp_xmit = tx_stats->xdp_xmit;
261 tso_packets = tx_stats->tso_packets;
262 tso_bytes = tx_stats->tso_bytes;
263 tso_inner_packets = tx_stats->tso_inner_packets;
264 tso_inner_bytes = tx_stats->tso_inner_bytes;
265 long_pkt_fmt = tx_stats->long_pkt_fmt;
266 short_pkt_fmt = tx_stats->short_pkt_fmt;
267 csum_partial = tx_stats->csum_partial;
268 mana_map_err = tx_stats->mana_map_err;
269 } while (u64_stats_fetch_retry(syncp: &tx_stats->syncp, start));
270
271 data[i++] = packets;
272 data[i++] = bytes;
273 data[i++] = xdp_xmit;
274 data[i++] = tso_packets;
275 data[i++] = tso_bytes;
276 data[i++] = tso_inner_packets;
277 data[i++] = tso_inner_bytes;
278 data[i++] = long_pkt_fmt;
279 data[i++] = short_pkt_fmt;
280 data[i++] = csum_partial;
281 data[i++] = mana_map_err;
282 }
283}
284
285static int mana_get_rxnfc(struct net_device *ndev, struct ethtool_rxnfc *cmd,
286 u32 *rules)
287{
288 struct mana_port_context *apc = netdev_priv(dev: ndev);
289
290 switch (cmd->cmd) {
291 case ETHTOOL_GRXRINGS:
292 cmd->data = apc->num_queues;
293 return 0;
294 }
295
296 return -EOPNOTSUPP;
297}
298
299static u32 mana_get_rxfh_key_size(struct net_device *ndev)
300{
301 return MANA_HASH_KEY_SIZE;
302}
303
304static u32 mana_rss_indir_size(struct net_device *ndev)
305{
306 struct mana_port_context *apc = netdev_priv(dev: ndev);
307
308 return apc->indir_table_sz;
309}
310
311static int mana_get_rxfh(struct net_device *ndev,
312 struct ethtool_rxfh_param *rxfh)
313{
314 struct mana_port_context *apc = netdev_priv(dev: ndev);
315 int i;
316
317 rxfh->hfunc = ETH_RSS_HASH_TOP; /* Toeplitz */
318
319 if (rxfh->indir) {
320 for (i = 0; i < apc->indir_table_sz; i++)
321 rxfh->indir[i] = apc->indir_table[i];
322 }
323
324 if (rxfh->key)
325 memcpy(rxfh->key, apc->hashkey, MANA_HASH_KEY_SIZE);
326
327 return 0;
328}
329
330static int mana_set_rxfh(struct net_device *ndev,
331 struct ethtool_rxfh_param *rxfh,
332 struct netlink_ext_ack *extack)
333{
334 struct mana_port_context *apc = netdev_priv(dev: ndev);
335 bool update_hash = false, update_table = false;
336 u8 save_key[MANA_HASH_KEY_SIZE];
337 u32 *save_table;
338 int i, err;
339
340 if (!apc->port_is_up)
341 return -EOPNOTSUPP;
342
343 if (rxfh->hfunc != ETH_RSS_HASH_NO_CHANGE &&
344 rxfh->hfunc != ETH_RSS_HASH_TOP)
345 return -EOPNOTSUPP;
346
347 save_table = kcalloc(apc->indir_table_sz, sizeof(u32), GFP_KERNEL);
348 if (!save_table)
349 return -ENOMEM;
350
351 if (rxfh->indir) {
352 for (i = 0; i < apc->indir_table_sz; i++)
353 if (rxfh->indir[i] >= apc->num_queues) {
354 err = -EINVAL;
355 goto cleanup;
356 }
357
358 update_table = true;
359 for (i = 0; i < apc->indir_table_sz; i++) {
360 save_table[i] = apc->indir_table[i];
361 apc->indir_table[i] = rxfh->indir[i];
362 }
363 }
364
365 if (rxfh->key) {
366 update_hash = true;
367 memcpy(save_key, apc->hashkey, MANA_HASH_KEY_SIZE);
368 memcpy(apc->hashkey, rxfh->key, MANA_HASH_KEY_SIZE);
369 }
370
371 err = mana_config_rss(ac: apc, rx: TRI_STATE_TRUE, update_hash, update_tab: update_table);
372
373 if (err) { /* recover to original values */
374 if (update_table) {
375 for (i = 0; i < apc->indir_table_sz; i++)
376 apc->indir_table[i] = save_table[i];
377 }
378
379 if (update_hash)
380 memcpy(apc->hashkey, save_key, MANA_HASH_KEY_SIZE);
381
382 mana_config_rss(ac: apc, rx: TRI_STATE_TRUE, update_hash, update_tab: update_table);
383 }
384
385cleanup:
386 kfree(objp: save_table);
387
388 return err;
389}
390
391static void mana_get_channels(struct net_device *ndev,
392 struct ethtool_channels *channel)
393{
394 struct mana_port_context *apc = netdev_priv(dev: ndev);
395
396 channel->max_combined = apc->max_queues;
397 channel->combined_count = apc->num_queues;
398}
399
400static int mana_set_channels(struct net_device *ndev,
401 struct ethtool_channels *channels)
402{
403 struct mana_port_context *apc = netdev_priv(dev: ndev);
404 unsigned int new_count = channels->combined_count;
405 unsigned int old_count = apc->num_queues;
406 int err;
407
408 err = mana_pre_alloc_rxbufs(apc, mtu: ndev->mtu, num_queues: new_count);
409 if (err) {
410 netdev_err(dev: ndev, format: "Insufficient memory for new allocations");
411 return err;
412 }
413
414 err = mana_detach(ndev, from_close: false);
415 if (err) {
416 netdev_err(dev: ndev, format: "mana_detach failed: %d\n", err);
417 goto out;
418 }
419
420 apc->num_queues = new_count;
421 err = mana_attach(ndev);
422 if (err) {
423 apc->num_queues = old_count;
424 netdev_err(dev: ndev, format: "mana_attach failed: %d\n", err);
425 }
426
427out:
428 mana_pre_dealloc_rxbufs(apc);
429 return err;
430}
431
432static void mana_get_ringparam(struct net_device *ndev,
433 struct ethtool_ringparam *ring,
434 struct kernel_ethtool_ringparam *kernel_ring,
435 struct netlink_ext_ack *extack)
436{
437 struct mana_port_context *apc = netdev_priv(dev: ndev);
438
439 ring->rx_pending = apc->rx_queue_size;
440 ring->tx_pending = apc->tx_queue_size;
441 ring->rx_max_pending = MAX_RX_BUFFERS_PER_QUEUE;
442 ring->tx_max_pending = MAX_TX_BUFFERS_PER_QUEUE;
443}
444
445static int mana_set_ringparam(struct net_device *ndev,
446 struct ethtool_ringparam *ring,
447 struct kernel_ethtool_ringparam *kernel_ring,
448 struct netlink_ext_ack *extack)
449{
450 struct mana_port_context *apc = netdev_priv(dev: ndev);
451 u32 new_tx, new_rx;
452 u32 old_tx, old_rx;
453 int err;
454
455 old_tx = apc->tx_queue_size;
456 old_rx = apc->rx_queue_size;
457
458 if (ring->tx_pending < MIN_TX_BUFFERS_PER_QUEUE) {
459 NL_SET_ERR_MSG_FMT(extack, "tx:%d less than the min:%d", ring->tx_pending,
460 MIN_TX_BUFFERS_PER_QUEUE);
461 return -EINVAL;
462 }
463
464 if (ring->rx_pending < MIN_RX_BUFFERS_PER_QUEUE) {
465 NL_SET_ERR_MSG_FMT(extack, "rx:%d less than the min:%d", ring->rx_pending,
466 MIN_RX_BUFFERS_PER_QUEUE);
467 return -EINVAL;
468 }
469
470 new_rx = roundup_pow_of_two(ring->rx_pending);
471 new_tx = roundup_pow_of_two(ring->tx_pending);
472 netdev_info(dev: ndev, format: "Using nearest power of 2 values for Txq:%d Rxq:%d\n",
473 new_tx, new_rx);
474
475 /* pre-allocating new buffers to prevent failures in mana_attach() later */
476 apc->rx_queue_size = new_rx;
477 err = mana_pre_alloc_rxbufs(apc, mtu: ndev->mtu, num_queues: apc->num_queues);
478 apc->rx_queue_size = old_rx;
479 if (err) {
480 netdev_err(dev: ndev, format: "Insufficient memory for new allocations\n");
481 return err;
482 }
483
484 err = mana_detach(ndev, from_close: false);
485 if (err) {
486 netdev_err(dev: ndev, format: "mana_detach failed: %d\n", err);
487 goto out;
488 }
489
490 apc->tx_queue_size = new_tx;
491 apc->rx_queue_size = new_rx;
492
493 err = mana_attach(ndev);
494 if (err) {
495 netdev_err(dev: ndev, format: "mana_attach failed: %d\n", err);
496 apc->tx_queue_size = old_tx;
497 apc->rx_queue_size = old_rx;
498 }
499out:
500 mana_pre_dealloc_rxbufs(apc);
501 return err;
502}
503
504static int mana_get_link_ksettings(struct net_device *ndev,
505 struct ethtool_link_ksettings *cmd)
506{
507 struct mana_port_context *apc = netdev_priv(dev: ndev);
508 int err;
509
510 err = mana_query_link_cfg(apc);
511 cmd->base.speed = (err) ? SPEED_UNKNOWN : apc->max_speed;
512
513 cmd->base.duplex = DUPLEX_FULL;
514 cmd->base.port = PORT_OTHER;
515
516 return 0;
517}
518
519const struct ethtool_ops mana_ethtool_ops = {
520 .get_ethtool_stats = mana_get_ethtool_stats,
521 .get_sset_count = mana_get_sset_count,
522 .get_strings = mana_get_strings,
523 .get_rxnfc = mana_get_rxnfc,
524 .get_rxfh_key_size = mana_get_rxfh_key_size,
525 .get_rxfh_indir_size = mana_rss_indir_size,
526 .get_rxfh = mana_get_rxfh,
527 .set_rxfh = mana_set_rxfh,
528 .get_channels = mana_get_channels,
529 .set_channels = mana_set_channels,
530 .get_ringparam = mana_get_ringparam,
531 .set_ringparam = mana_set_ringparam,
532 .get_link_ksettings = mana_get_link_ksettings,
533 .get_link = ethtool_op_get_link,
534};
535

source code of linux/drivers/net/ethernet/microsoft/mana/mana_ethtool.c