be2net: allow register dump only for PFs
[linux-2.6/btrfs-unstable.git] / drivers / net / benet / be_ethtool.c
blob22523b92b92a1ca22ad2643c964921a241874ba0
1 /*
2 * Copyright (C) 2005 - 2011 Emulex
3 * All rights reserved.
5 * This program is free software; you can redistribute it and/or
6 * modify it under the terms of the GNU General Public License version 2
7 * as published by the Free Software Foundation. The full GNU General
8 * Public License is included in this distribution in the file called COPYING.
10 * Contact Information:
11 * linux-drivers@emulex.com
13 * Emulex
14 * 3333 Susan Street
15 * Costa Mesa, CA 92626
18 #include "be.h"
19 #include "be_cmds.h"
20 #include <linux/ethtool.h>
22 struct be_ethtool_stat {
23 char desc[ETH_GSTRING_LEN];
24 int type;
25 int size;
26 int offset;
29 enum {NETSTAT, PORTSTAT, MISCSTAT, DRVSTAT_TX, DRVSTAT_RX, ERXSTAT,
30 PMEMSTAT, DRVSTAT};
31 #define FIELDINFO(_struct, field) FIELD_SIZEOF(_struct, field), \
32 offsetof(_struct, field)
33 #define NETSTAT_INFO(field) #field, NETSTAT,\
34 FIELDINFO(struct net_device_stats,\
35 field)
36 #define DRVSTAT_TX_INFO(field) #field, DRVSTAT_TX,\
37 FIELDINFO(struct be_tx_stats, field)
38 #define DRVSTAT_RX_INFO(field) #field, DRVSTAT_RX,\
39 FIELDINFO(struct be_rx_stats, field)
40 #define MISCSTAT_INFO(field) #field, MISCSTAT,\
41 FIELDINFO(struct be_rxf_stats, field)
42 #define PORTSTAT_INFO(field) #field, PORTSTAT,\
43 FIELDINFO(struct be_port_rxf_stats, \
44 field)
45 #define ERXSTAT_INFO(field) #field, ERXSTAT,\
46 FIELDINFO(struct be_erx_stats, field)
47 #define PMEMSTAT_INFO(field) #field, PMEMSTAT,\
48 FIELDINFO(struct be_pmem_stats, field)
49 #define DRVSTAT_INFO(field) #field, DRVSTAT,\
50 FIELDINFO(struct be_drv_stats, \
51 field)
53 static const struct be_ethtool_stat et_stats[] = {
54 {NETSTAT_INFO(rx_packets)},
55 {NETSTAT_INFO(tx_packets)},
56 {NETSTAT_INFO(rx_bytes)},
57 {NETSTAT_INFO(tx_bytes)},
58 {NETSTAT_INFO(rx_errors)},
59 {NETSTAT_INFO(tx_errors)},
60 {NETSTAT_INFO(rx_dropped)},
61 {NETSTAT_INFO(tx_dropped)},
62 {DRVSTAT_TX_INFO(be_tx_rate)},
63 {DRVSTAT_TX_INFO(be_tx_reqs)},
64 {DRVSTAT_TX_INFO(be_tx_wrbs)},
65 {DRVSTAT_TX_INFO(be_tx_stops)},
66 {DRVSTAT_TX_INFO(be_tx_events)},
67 {DRVSTAT_TX_INFO(be_tx_compl)},
68 {PORTSTAT_INFO(rx_unicast_frames)},
69 {PORTSTAT_INFO(rx_multicast_frames)},
70 {PORTSTAT_INFO(rx_broadcast_frames)},
71 {PORTSTAT_INFO(rx_crc_errors)},
72 {PORTSTAT_INFO(rx_alignment_symbol_errors)},
73 {PORTSTAT_INFO(rx_pause_frames)},
74 {PORTSTAT_INFO(rx_control_frames)},
75 {PORTSTAT_INFO(rx_in_range_errors)},
76 {PORTSTAT_INFO(rx_out_range_errors)},
77 {PORTSTAT_INFO(rx_frame_too_long)},
78 {PORTSTAT_INFO(rx_address_match_errors)},
79 {PORTSTAT_INFO(rx_vlan_mismatch)},
80 {PORTSTAT_INFO(rx_dropped_too_small)},
81 {PORTSTAT_INFO(rx_dropped_too_short)},
82 {PORTSTAT_INFO(rx_dropped_header_too_small)},
83 {PORTSTAT_INFO(rx_dropped_tcp_length)},
84 {PORTSTAT_INFO(rx_dropped_runt)},
85 {PORTSTAT_INFO(rx_fifo_overflow)},
86 {PORTSTAT_INFO(rx_input_fifo_overflow)},
87 {PORTSTAT_INFO(rx_ip_checksum_errs)},
88 {PORTSTAT_INFO(rx_tcp_checksum_errs)},
89 {PORTSTAT_INFO(rx_udp_checksum_errs)},
90 {PORTSTAT_INFO(rx_non_rss_packets)},
91 {PORTSTAT_INFO(rx_ipv4_packets)},
92 {PORTSTAT_INFO(rx_ipv6_packets)},
93 {PORTSTAT_INFO(rx_switched_unicast_packets)},
94 {PORTSTAT_INFO(rx_switched_multicast_packets)},
95 {PORTSTAT_INFO(rx_switched_broadcast_packets)},
96 {PORTSTAT_INFO(tx_unicastframes)},
97 {PORTSTAT_INFO(tx_multicastframes)},
98 {PORTSTAT_INFO(tx_broadcastframes)},
99 {PORTSTAT_INFO(tx_pauseframes)},
100 {PORTSTAT_INFO(tx_controlframes)},
101 {MISCSTAT_INFO(rx_drops_no_pbuf)},
102 {MISCSTAT_INFO(rx_drops_no_txpb)},
103 {MISCSTAT_INFO(rx_drops_no_erx_descr)},
104 {MISCSTAT_INFO(rx_drops_no_tpre_descr)},
105 {MISCSTAT_INFO(rx_drops_too_many_frags)},
106 {MISCSTAT_INFO(rx_drops_invalid_ring)},
107 {MISCSTAT_INFO(forwarded_packets)},
108 {MISCSTAT_INFO(rx_drops_mtu)},
109 {MISCSTAT_INFO(port0_jabber_events)},
110 {MISCSTAT_INFO(port1_jabber_events)},
111 {PMEMSTAT_INFO(eth_red_drops)},
112 {DRVSTAT_INFO(be_on_die_temperature)}
114 #define ETHTOOL_STATS_NUM ARRAY_SIZE(et_stats)
116 /* Stats related to multi RX queues */
117 static const struct be_ethtool_stat et_rx_stats[] = {
118 {DRVSTAT_RX_INFO(rx_bytes)},
119 {DRVSTAT_RX_INFO(rx_pkts)},
120 {DRVSTAT_RX_INFO(rx_rate)},
121 {DRVSTAT_RX_INFO(rx_polls)},
122 {DRVSTAT_RX_INFO(rx_events)},
123 {DRVSTAT_RX_INFO(rx_compl)},
124 {DRVSTAT_RX_INFO(rx_mcast_pkts)},
125 {DRVSTAT_RX_INFO(rx_post_fail)},
126 {ERXSTAT_INFO(rx_drops_no_fragments)}
128 #define ETHTOOL_RXSTATS_NUM (ARRAY_SIZE(et_rx_stats))
130 static const char et_self_tests[][ETH_GSTRING_LEN] = {
131 "MAC Loopback test",
132 "PHY Loopback test",
133 "External Loopback test",
134 "DDR DMA test",
135 "Link test"
138 #define ETHTOOL_TESTS_NUM ARRAY_SIZE(et_self_tests)
139 #define BE_MAC_LOOPBACK 0x0
140 #define BE_PHY_LOOPBACK 0x1
141 #define BE_ONE_PORT_EXT_LOOPBACK 0x2
142 #define BE_NO_LOOPBACK 0xff
144 static void
145 be_get_drvinfo(struct net_device *netdev, struct ethtool_drvinfo *drvinfo)
147 struct be_adapter *adapter = netdev_priv(netdev);
149 strcpy(drvinfo->driver, DRV_NAME);
150 strcpy(drvinfo->version, DRV_VER);
151 strncpy(drvinfo->fw_version, adapter->fw_ver, FW_VER_LEN);
152 strcpy(drvinfo->bus_info, pci_name(adapter->pdev));
153 drvinfo->testinfo_len = 0;
154 drvinfo->regdump_len = 0;
155 drvinfo->eedump_len = 0;
158 static int
159 be_get_reg_len(struct net_device *netdev)
161 struct be_adapter *adapter = netdev_priv(netdev);
162 u32 log_size = 0;
164 if (be_physfn(adapter))
165 be_cmd_get_reg_len(adapter, &log_size);
167 return log_size;
170 static void
171 be_get_regs(struct net_device *netdev, struct ethtool_regs *regs, void *buf)
173 struct be_adapter *adapter = netdev_priv(netdev);
175 if (be_physfn(adapter)) {
176 memset(buf, 0, regs->len);
177 be_cmd_get_regs(adapter, regs->len, buf);
181 static int
182 be_get_coalesce(struct net_device *netdev, struct ethtool_coalesce *coalesce)
184 struct be_adapter *adapter = netdev_priv(netdev);
185 struct be_eq_obj *rx_eq = &adapter->rx_obj[0].rx_eq;
186 struct be_eq_obj *tx_eq = &adapter->tx_eq;
188 coalesce->rx_coalesce_usecs = rx_eq->cur_eqd;
189 coalesce->rx_coalesce_usecs_high = rx_eq->max_eqd;
190 coalesce->rx_coalesce_usecs_low = rx_eq->min_eqd;
192 coalesce->tx_coalesce_usecs = tx_eq->cur_eqd;
193 coalesce->tx_coalesce_usecs_high = tx_eq->max_eqd;
194 coalesce->tx_coalesce_usecs_low = tx_eq->min_eqd;
196 coalesce->use_adaptive_rx_coalesce = rx_eq->enable_aic;
197 coalesce->use_adaptive_tx_coalesce = tx_eq->enable_aic;
199 return 0;
203 * This routine is used to set interrup coalescing delay
205 static int
206 be_set_coalesce(struct net_device *netdev, struct ethtool_coalesce *coalesce)
208 struct be_adapter *adapter = netdev_priv(netdev);
209 struct be_rx_obj *rxo;
210 struct be_eq_obj *rx_eq;
211 struct be_eq_obj *tx_eq = &adapter->tx_eq;
212 u32 rx_max, rx_min, rx_cur;
213 int status = 0, i;
214 u32 tx_cur;
216 if (coalesce->use_adaptive_tx_coalesce == 1)
217 return -EINVAL;
219 for_all_rx_queues(adapter, rxo, i) {
220 rx_eq = &rxo->rx_eq;
222 if (!rx_eq->enable_aic && coalesce->use_adaptive_rx_coalesce)
223 rx_eq->cur_eqd = 0;
224 rx_eq->enable_aic = coalesce->use_adaptive_rx_coalesce;
226 rx_max = coalesce->rx_coalesce_usecs_high;
227 rx_min = coalesce->rx_coalesce_usecs_low;
228 rx_cur = coalesce->rx_coalesce_usecs;
230 if (rx_eq->enable_aic) {
231 if (rx_max > BE_MAX_EQD)
232 rx_max = BE_MAX_EQD;
233 if (rx_min > rx_max)
234 rx_min = rx_max;
235 rx_eq->max_eqd = rx_max;
236 rx_eq->min_eqd = rx_min;
237 if (rx_eq->cur_eqd > rx_max)
238 rx_eq->cur_eqd = rx_max;
239 if (rx_eq->cur_eqd < rx_min)
240 rx_eq->cur_eqd = rx_min;
241 } else {
242 if (rx_cur > BE_MAX_EQD)
243 rx_cur = BE_MAX_EQD;
244 if (rx_eq->cur_eqd != rx_cur) {
245 status = be_cmd_modify_eqd(adapter, rx_eq->q.id,
246 rx_cur);
247 if (!status)
248 rx_eq->cur_eqd = rx_cur;
253 tx_cur = coalesce->tx_coalesce_usecs;
255 if (tx_cur > BE_MAX_EQD)
256 tx_cur = BE_MAX_EQD;
257 if (tx_eq->cur_eqd != tx_cur) {
258 status = be_cmd_modify_eqd(adapter, tx_eq->q.id, tx_cur);
259 if (!status)
260 tx_eq->cur_eqd = tx_cur;
263 return 0;
266 static void
267 be_get_ethtool_stats(struct net_device *netdev,
268 struct ethtool_stats *stats, uint64_t *data)
270 struct be_adapter *adapter = netdev_priv(netdev);
271 struct be_hw_stats *hw_stats = hw_stats_from_cmd(adapter->stats_cmd.va);
272 struct be_erx_stats *erx_stats = &hw_stats->erx;
273 struct be_rx_obj *rxo;
274 void *p = NULL;
275 int i, j;
277 for (i = 0; i < ETHTOOL_STATS_NUM; i++) {
278 switch (et_stats[i].type) {
279 case NETSTAT:
280 p = &netdev->stats;
281 break;
282 case DRVSTAT_TX:
283 p = &adapter->tx_stats;
284 break;
285 case PORTSTAT:
286 p = &hw_stats->rxf.port[adapter->port_num];
287 break;
288 case MISCSTAT:
289 p = &hw_stats->rxf;
290 break;
291 case PMEMSTAT:
292 p = &hw_stats->pmem;
293 break;
294 case DRVSTAT:
295 p = &adapter->drv_stats;
296 break;
299 p = (u8 *)p + et_stats[i].offset;
300 data[i] = (et_stats[i].size == sizeof(u64)) ?
301 *(u64 *)p: *(u32 *)p;
304 for_all_rx_queues(adapter, rxo, j) {
305 for (i = 0; i < ETHTOOL_RXSTATS_NUM; i++) {
306 switch (et_rx_stats[i].type) {
307 case DRVSTAT_RX:
308 p = (u8 *)&rxo->stats + et_rx_stats[i].offset;
309 break;
310 case ERXSTAT:
311 p = (u32 *)erx_stats + rxo->q.id;
312 break;
314 data[ETHTOOL_STATS_NUM + j * ETHTOOL_RXSTATS_NUM + i] =
315 (et_rx_stats[i].size == sizeof(u64)) ?
316 *(u64 *)p: *(u32 *)p;
321 static void
322 be_get_stat_strings(struct net_device *netdev, uint32_t stringset,
323 uint8_t *data)
325 struct be_adapter *adapter = netdev_priv(netdev);
326 int i, j;
328 switch (stringset) {
329 case ETH_SS_STATS:
330 for (i = 0; i < ETHTOOL_STATS_NUM; i++) {
331 memcpy(data, et_stats[i].desc, ETH_GSTRING_LEN);
332 data += ETH_GSTRING_LEN;
334 for (i = 0; i < adapter->num_rx_qs; i++) {
335 for (j = 0; j < ETHTOOL_RXSTATS_NUM; j++) {
336 sprintf(data, "rxq%d: %s", i,
337 et_rx_stats[j].desc);
338 data += ETH_GSTRING_LEN;
341 break;
342 case ETH_SS_TEST:
343 for (i = 0; i < ETHTOOL_TESTS_NUM; i++) {
344 memcpy(data, et_self_tests[i], ETH_GSTRING_LEN);
345 data += ETH_GSTRING_LEN;
347 break;
351 static int be_get_sset_count(struct net_device *netdev, int stringset)
353 struct be_adapter *adapter = netdev_priv(netdev);
355 switch (stringset) {
356 case ETH_SS_TEST:
357 return ETHTOOL_TESTS_NUM;
358 case ETH_SS_STATS:
359 return ETHTOOL_STATS_NUM +
360 adapter->num_rx_qs * ETHTOOL_RXSTATS_NUM;
361 default:
362 return -EINVAL;
366 static int be_get_settings(struct net_device *netdev, struct ethtool_cmd *ecmd)
368 struct be_adapter *adapter = netdev_priv(netdev);
369 struct be_dma_mem phy_cmd;
370 struct be_cmd_resp_get_phy_info *resp;
371 u8 mac_speed = 0;
372 u16 link_speed = 0;
373 bool link_up = false;
374 int status;
375 u16 intf_type;
377 if ((adapter->link_speed < 0) || (!(netdev->flags & IFF_UP))) {
378 status = be_cmd_link_status_query(adapter, &link_up,
379 &mac_speed, &link_speed);
381 be_link_status_update(adapter, link_up);
382 /* link_speed is in units of 10 Mbps */
383 if (link_speed) {
384 ecmd->speed = link_speed*10;
385 } else {
386 switch (mac_speed) {
387 case PHY_LINK_SPEED_1GBPS:
388 ecmd->speed = SPEED_1000;
389 break;
390 case PHY_LINK_SPEED_10GBPS:
391 ecmd->speed = SPEED_10000;
392 break;
396 phy_cmd.size = sizeof(struct be_cmd_req_get_phy_info);
397 phy_cmd.va = dma_alloc_coherent(&adapter->pdev->dev,
398 phy_cmd.size, &phy_cmd.dma,
399 GFP_KERNEL);
400 if (!phy_cmd.va) {
401 dev_err(&adapter->pdev->dev, "Memory alloc failure\n");
402 return -ENOMEM;
404 status = be_cmd_get_phy_info(adapter, &phy_cmd);
405 if (!status) {
406 resp = (struct be_cmd_resp_get_phy_info *) phy_cmd.va;
407 intf_type = le16_to_cpu(resp->interface_type);
409 switch (intf_type) {
410 case PHY_TYPE_XFP_10GB:
411 case PHY_TYPE_SFP_1GB:
412 case PHY_TYPE_SFP_PLUS_10GB:
413 ecmd->port = PORT_FIBRE;
414 break;
415 default:
416 ecmd->port = PORT_TP;
417 break;
420 switch (intf_type) {
421 case PHY_TYPE_KR_10GB:
422 case PHY_TYPE_KX4_10GB:
423 ecmd->autoneg = AUTONEG_ENABLE;
424 ecmd->transceiver = XCVR_INTERNAL;
425 break;
426 default:
427 ecmd->autoneg = AUTONEG_DISABLE;
428 ecmd->transceiver = XCVR_EXTERNAL;
429 break;
433 /* Save for future use */
434 adapter->link_speed = ecmd->speed;
435 adapter->port_type = ecmd->port;
436 adapter->transceiver = ecmd->transceiver;
437 adapter->autoneg = ecmd->autoneg;
438 dma_free_coherent(&adapter->pdev->dev, phy_cmd.size, phy_cmd.va,
439 phy_cmd.dma);
440 } else {
441 ecmd->speed = adapter->link_speed;
442 ecmd->port = adapter->port_type;
443 ecmd->transceiver = adapter->transceiver;
444 ecmd->autoneg = adapter->autoneg;
447 ecmd->duplex = DUPLEX_FULL;
448 ecmd->phy_address = adapter->port_num;
449 switch (ecmd->port) {
450 case PORT_FIBRE:
451 ecmd->supported = (SUPPORTED_10000baseT_Full | SUPPORTED_FIBRE);
452 break;
453 case PORT_TP:
454 ecmd->supported = (SUPPORTED_10000baseT_Full | SUPPORTED_TP);
455 break;
456 case PORT_AUI:
457 ecmd->supported = (SUPPORTED_10000baseT_Full | SUPPORTED_AUI);
458 break;
461 if (ecmd->autoneg) {
462 ecmd->supported |= SUPPORTED_1000baseT_Full;
463 ecmd->supported |= SUPPORTED_Autoneg;
464 ecmd->advertising |= (ADVERTISED_10000baseT_Full |
465 ADVERTISED_1000baseT_Full);
468 return 0;
471 static void
472 be_get_ringparam(struct net_device *netdev, struct ethtool_ringparam *ring)
474 struct be_adapter *adapter = netdev_priv(netdev);
476 ring->rx_max_pending = adapter->rx_obj[0].q.len;
477 ring->tx_max_pending = adapter->tx_obj.q.len;
479 ring->rx_pending = atomic_read(&adapter->rx_obj[0].q.used);
480 ring->tx_pending = atomic_read(&adapter->tx_obj.q.used);
483 static void
484 be_get_pauseparam(struct net_device *netdev, struct ethtool_pauseparam *ecmd)
486 struct be_adapter *adapter = netdev_priv(netdev);
488 be_cmd_get_flow_control(adapter, &ecmd->tx_pause, &ecmd->rx_pause);
489 ecmd->autoneg = 0;
492 static int
493 be_set_pauseparam(struct net_device *netdev, struct ethtool_pauseparam *ecmd)
495 struct be_adapter *adapter = netdev_priv(netdev);
496 int status;
498 if (ecmd->autoneg != 0)
499 return -EINVAL;
500 adapter->tx_fc = ecmd->tx_pause;
501 adapter->rx_fc = ecmd->rx_pause;
503 status = be_cmd_set_flow_control(adapter,
504 adapter->tx_fc, adapter->rx_fc);
505 if (status)
506 dev_warn(&adapter->pdev->dev, "Pause param set failed.\n");
508 return status;
511 static int
512 be_set_phys_id(struct net_device *netdev,
513 enum ethtool_phys_id_state state)
515 struct be_adapter *adapter = netdev_priv(netdev);
517 switch (state) {
518 case ETHTOOL_ID_ACTIVE:
519 be_cmd_get_beacon_state(adapter, adapter->hba_port_num,
520 &adapter->beacon_state);
521 return 1; /* cycle on/off once per second */
523 case ETHTOOL_ID_ON:
524 be_cmd_set_beacon_state(adapter, adapter->hba_port_num, 0, 0,
525 BEACON_STATE_ENABLED);
526 break;
528 case ETHTOOL_ID_OFF:
529 be_cmd_set_beacon_state(adapter, adapter->hba_port_num, 0, 0,
530 BEACON_STATE_DISABLED);
531 break;
533 case ETHTOOL_ID_INACTIVE:
534 be_cmd_set_beacon_state(adapter, adapter->hba_port_num, 0, 0,
535 adapter->beacon_state);
538 return 0;
541 static bool
542 be_is_wol_supported(struct be_adapter *adapter)
544 if (!be_physfn(adapter))
545 return false;
546 else
547 return true;
550 static void
551 be_get_wol(struct net_device *netdev, struct ethtool_wolinfo *wol)
553 struct be_adapter *adapter = netdev_priv(netdev);
555 if (be_is_wol_supported(adapter))
556 wol->supported = WAKE_MAGIC;
558 if (adapter->wol)
559 wol->wolopts = WAKE_MAGIC;
560 else
561 wol->wolopts = 0;
562 memset(&wol->sopass, 0, sizeof(wol->sopass));
565 static int
566 be_set_wol(struct net_device *netdev, struct ethtool_wolinfo *wol)
568 struct be_adapter *adapter = netdev_priv(netdev);
570 if (wol->wolopts & ~WAKE_MAGIC)
571 return -EINVAL;
573 if ((wol->wolopts & WAKE_MAGIC) && be_is_wol_supported(adapter))
574 adapter->wol = true;
575 else
576 adapter->wol = false;
578 return 0;
581 static int
582 be_test_ddr_dma(struct be_adapter *adapter)
584 int ret, i;
585 struct be_dma_mem ddrdma_cmd;
586 static const u64 pattern[2] = {
587 0x5a5a5a5a5a5a5a5aULL, 0xa5a5a5a5a5a5a5a5ULL
590 ddrdma_cmd.size = sizeof(struct be_cmd_req_ddrdma_test);
591 ddrdma_cmd.va = dma_alloc_coherent(&adapter->pdev->dev, ddrdma_cmd.size,
592 &ddrdma_cmd.dma, GFP_KERNEL);
593 if (!ddrdma_cmd.va) {
594 dev_err(&adapter->pdev->dev, "Memory allocation failure\n");
595 return -ENOMEM;
598 for (i = 0; i < 2; i++) {
599 ret = be_cmd_ddr_dma_test(adapter, pattern[i],
600 4096, &ddrdma_cmd);
601 if (ret != 0)
602 goto err;
605 err:
606 dma_free_coherent(&adapter->pdev->dev, ddrdma_cmd.size, ddrdma_cmd.va,
607 ddrdma_cmd.dma);
608 return ret;
611 static u64 be_loopback_test(struct be_adapter *adapter, u8 loopback_type,
612 u64 *status)
614 be_cmd_set_loopback(adapter, adapter->hba_port_num,
615 loopback_type, 1);
616 *status = be_cmd_loopback_test(adapter, adapter->hba_port_num,
617 loopback_type, 1500,
618 2, 0xabc);
619 be_cmd_set_loopback(adapter, adapter->hba_port_num,
620 BE_NO_LOOPBACK, 1);
621 return *status;
624 static void
625 be_self_test(struct net_device *netdev, struct ethtool_test *test, u64 *data)
627 struct be_adapter *adapter = netdev_priv(netdev);
628 bool link_up;
629 u8 mac_speed = 0;
630 u16 qos_link_speed = 0;
632 memset(data, 0, sizeof(u64) * ETHTOOL_TESTS_NUM);
634 if (test->flags & ETH_TEST_FL_OFFLINE) {
635 if (be_loopback_test(adapter, BE_MAC_LOOPBACK,
636 &data[0]) != 0) {
637 test->flags |= ETH_TEST_FL_FAILED;
639 if (be_loopback_test(adapter, BE_PHY_LOOPBACK,
640 &data[1]) != 0) {
641 test->flags |= ETH_TEST_FL_FAILED;
643 if (be_loopback_test(adapter, BE_ONE_PORT_EXT_LOOPBACK,
644 &data[2]) != 0) {
645 test->flags |= ETH_TEST_FL_FAILED;
649 if (be_test_ddr_dma(adapter) != 0) {
650 data[3] = 1;
651 test->flags |= ETH_TEST_FL_FAILED;
654 if (be_cmd_link_status_query(adapter, &link_up, &mac_speed,
655 &qos_link_speed) != 0) {
656 test->flags |= ETH_TEST_FL_FAILED;
657 data[4] = -1;
658 } else if (!mac_speed) {
659 test->flags |= ETH_TEST_FL_FAILED;
660 data[4] = 1;
664 static int
665 be_do_flash(struct net_device *netdev, struct ethtool_flash *efl)
667 struct be_adapter *adapter = netdev_priv(netdev);
668 char file_name[ETHTOOL_FLASH_MAX_FILENAME];
670 file_name[ETHTOOL_FLASH_MAX_FILENAME - 1] = 0;
671 strcpy(file_name, efl->data);
673 return be_load_fw(adapter, file_name);
676 static int
677 be_get_eeprom_len(struct net_device *netdev)
679 return BE_READ_SEEPROM_LEN;
682 static int
683 be_read_eeprom(struct net_device *netdev, struct ethtool_eeprom *eeprom,
684 uint8_t *data)
686 struct be_adapter *adapter = netdev_priv(netdev);
687 struct be_dma_mem eeprom_cmd;
688 struct be_cmd_resp_seeprom_read *resp;
689 int status;
691 if (!eeprom->len)
692 return -EINVAL;
694 eeprom->magic = BE_VENDOR_ID | (adapter->pdev->device<<16);
696 memset(&eeprom_cmd, 0, sizeof(struct be_dma_mem));
697 eeprom_cmd.size = sizeof(struct be_cmd_req_seeprom_read);
698 eeprom_cmd.va = dma_alloc_coherent(&adapter->pdev->dev, eeprom_cmd.size,
699 &eeprom_cmd.dma, GFP_KERNEL);
701 if (!eeprom_cmd.va) {
702 dev_err(&adapter->pdev->dev,
703 "Memory allocation failure. Could not read eeprom\n");
704 return -ENOMEM;
707 status = be_cmd_get_seeprom_data(adapter, &eeprom_cmd);
709 if (!status) {
710 resp = (struct be_cmd_resp_seeprom_read *) eeprom_cmd.va;
711 memcpy(data, resp->seeprom_data + eeprom->offset, eeprom->len);
713 dma_free_coherent(&adapter->pdev->dev, eeprom_cmd.size, eeprom_cmd.va,
714 eeprom_cmd.dma);
716 return status;
719 const struct ethtool_ops be_ethtool_ops = {
720 .get_settings = be_get_settings,
721 .get_drvinfo = be_get_drvinfo,
722 .get_wol = be_get_wol,
723 .set_wol = be_set_wol,
724 .get_link = ethtool_op_get_link,
725 .get_eeprom_len = be_get_eeprom_len,
726 .get_eeprom = be_read_eeprom,
727 .get_coalesce = be_get_coalesce,
728 .set_coalesce = be_set_coalesce,
729 .get_ringparam = be_get_ringparam,
730 .get_pauseparam = be_get_pauseparam,
731 .set_pauseparam = be_set_pauseparam,
732 .get_strings = be_get_stat_strings,
733 .set_phys_id = be_set_phys_id,
734 .get_sset_count = be_get_sset_count,
735 .get_ethtool_stats = be_get_ethtool_stats,
736 .get_regs_len = be_get_reg_len,
737 .get_regs = be_get_regs,
738 .flash_device = be_do_flash,
739 .self_test = be_self_test,