2 * Copyright (C) 2005 - 2011 Emulex
5 * This program is free software; you can redistribute it and/or
6 * modify it under the terms of the GNU General Public License version 2
7 * as published by the Free Software Foundation. The full GNU General
8 * Public License is included in this distribution in the file called COPYING.
10 * Contact Information:
11 * linux-drivers@emulex.com
15 * Costa Mesa, CA 92626
20 #include <linux/ethtool.h>
22 struct be_ethtool_stat {
23 char desc[ETH_GSTRING_LEN];
29 enum {DRVSTAT_TX, DRVSTAT_RX, DRVSTAT};
30 #define FIELDINFO(_struct, field) FIELD_SIZEOF(_struct, field), \
31 offsetof(_struct, field)
32 #define DRVSTAT_TX_INFO(field) #field, DRVSTAT_TX,\
33 FIELDINFO(struct be_tx_stats, field)
34 #define DRVSTAT_RX_INFO(field) #field, DRVSTAT_RX,\
35 FIELDINFO(struct be_rx_stats, field)
36 #define DRVSTAT_INFO(field) #field, DRVSTAT,\
37 FIELDINFO(struct be_drv_stats, field)
39 static const struct be_ethtool_stat et_stats[] = {
40 {DRVSTAT_INFO(rx_crc_errors)},
41 {DRVSTAT_INFO(rx_alignment_symbol_errors)},
42 {DRVSTAT_INFO(rx_pause_frames)},
43 {DRVSTAT_INFO(rx_control_frames)},
44 /* Received packets dropped when the Ethernet length field
45 * is not equal to the actual Ethernet data length.
47 {DRVSTAT_INFO(rx_in_range_errors)},
48 /* Received packets dropped when their length field is >= 1501 bytes
51 {DRVSTAT_INFO(rx_out_range_errors)},
52 /* Received packets dropped when they are longer than 9216 bytes */
53 {DRVSTAT_INFO(rx_frame_too_long)},
54 /* Received packets dropped when they don't pass the unicast or
55 * multicast address filtering.
57 {DRVSTAT_INFO(rx_address_mismatch_drops)},
58 /* Received packets dropped when IP packet length field is less than
59 * the IP header length field.
61 {DRVSTAT_INFO(rx_dropped_too_small)},
62 /* Received packets dropped when IP length field is greater than
63 * the actual packet length.
65 {DRVSTAT_INFO(rx_dropped_too_short)},
66 /* Received packets dropped when the IP header length field is less
69 {DRVSTAT_INFO(rx_dropped_header_too_small)},
70 /* Received packets dropped when the TCP header length field is less
71 * than 5 or the TCP header length + IP header length is more
72 * than IP packet length.
74 {DRVSTAT_INFO(rx_dropped_tcp_length)},
75 {DRVSTAT_INFO(rx_dropped_runt)},
76 /* Number of received packets dropped when a fifo for descriptors going
77 * into the packet demux block overflows. In normal operation, this
78 * fifo must never overflow.
80 {DRVSTAT_INFO(rxpp_fifo_overflow_drop)},
81 {DRVSTAT_INFO(rx_input_fifo_overflow_drop)},
82 {DRVSTAT_INFO(rx_ip_checksum_errs)},
83 {DRVSTAT_INFO(rx_tcp_checksum_errs)},
84 {DRVSTAT_INFO(rx_udp_checksum_errs)},
85 {DRVSTAT_INFO(tx_pauseframes)},
86 {DRVSTAT_INFO(tx_controlframes)},
87 {DRVSTAT_INFO(rx_priority_pause_frames)},
88 /* Received packets dropped when an internal fifo going into
89 * main packet buffer tank (PMEM) overflows.
91 {DRVSTAT_INFO(pmem_fifo_overflow_drop)},
92 {DRVSTAT_INFO(jabber_events)},
93 /* Received packets dropped due to lack of available HW packet buffers
94 * used to temporarily hold the received packets.
96 {DRVSTAT_INFO(rx_drops_no_pbuf)},
97 /* Received packets dropped due to input receive buffer
98 * descriptor fifo overflowing.
100 {DRVSTAT_INFO(rx_drops_no_erx_descr)},
101 /* Packets dropped because the internal FIFO to the offloaded TCP
102 * receive processing block is full. This could happen only for
103 * offloaded iSCSI or FCoE trarffic.
105 {DRVSTAT_INFO(rx_drops_no_tpre_descr)},
106 /* Received packets dropped when they need more than 8
107 * receive buffers. This cannot happen as the driver configures
108 * 2048 byte receive buffers.
110 {DRVSTAT_INFO(rx_drops_too_many_frags)},
111 {DRVSTAT_INFO(forwarded_packets)},
112 /* Received packets dropped when the frame length
113 * is more than 9018 bytes
115 {DRVSTAT_INFO(rx_drops_mtu)},
116 /* Number of packets dropped due to random early drop function */
117 {DRVSTAT_INFO(eth_red_drops)},
118 {DRVSTAT_INFO(be_on_die_temperature)}
120 #define ETHTOOL_STATS_NUM ARRAY_SIZE(et_stats)
122 /* Stats related to multi RX queues: get_stats routine assumes bytes, pkts
123 * are first and second members respectively.
125 static const struct be_ethtool_stat et_rx_stats[] = {
126 {DRVSTAT_RX_INFO(rx_bytes)},/* If moving this member see above note */
127 {DRVSTAT_RX_INFO(rx_pkts)}, /* If moving this member see above note */
128 {DRVSTAT_RX_INFO(rx_compl)},
129 {DRVSTAT_RX_INFO(rx_mcast_pkts)},
130 /* Number of page allocation failures while posting receive buffers
133 {DRVSTAT_RX_INFO(rx_post_fail)},
134 /* Recevied packets dropped due to skb allocation failure */
135 {DRVSTAT_RX_INFO(rx_drops_no_skbs)},
136 /* Received packets dropped due to lack of available fetched buffers
137 * posted by the driver.
139 {DRVSTAT_RX_INFO(rx_drops_no_frags)}
141 #define ETHTOOL_RXSTATS_NUM (ARRAY_SIZE(et_rx_stats))
143 /* Stats related to multi TX queues: get_stats routine assumes compl is the
146 static const struct be_ethtool_stat et_tx_stats[] = {
147 {DRVSTAT_TX_INFO(tx_compl)}, /* If moving this member see above note */
148 {DRVSTAT_TX_INFO(tx_bytes)},
149 {DRVSTAT_TX_INFO(tx_pkts)},
150 /* Number of skbs queued for trasmission by the driver */
151 {DRVSTAT_TX_INFO(tx_reqs)},
152 /* Number of TX work request blocks DMAed to HW */
153 {DRVSTAT_TX_INFO(tx_wrbs)},
154 /* Number of times the TX queue was stopped due to lack
155 * of spaces in the TXQ.
157 {DRVSTAT_TX_INFO(tx_stops)}
159 #define ETHTOOL_TXSTATS_NUM (ARRAY_SIZE(et_tx_stats))
161 static const char et_self_tests[][ETH_GSTRING_LEN] = {
164 "External Loopback test",
169 #define ETHTOOL_TESTS_NUM ARRAY_SIZE(et_self_tests)
170 #define BE_MAC_LOOPBACK 0x0
171 #define BE_PHY_LOOPBACK 0x1
172 #define BE_ONE_PORT_EXT_LOOPBACK 0x2
173 #define BE_NO_LOOPBACK 0xff
175 static void be_get_drvinfo(struct net_device *netdev,
176 struct ethtool_drvinfo *drvinfo)
178 struct be_adapter *adapter = netdev_priv(netdev);
179 char fw_on_flash[FW_VER_LEN];
181 memset(fw_on_flash, 0 , sizeof(fw_on_flash));
182 be_cmd_get_fw_ver(adapter, adapter->fw_ver, fw_on_flash);
184 strlcpy(drvinfo->driver, DRV_NAME, sizeof(drvinfo->driver));
185 strlcpy(drvinfo->version, DRV_VER, sizeof(drvinfo->version));
186 strncpy(drvinfo->fw_version, adapter->fw_ver, FW_VER_LEN);
187 if (memcmp(adapter->fw_ver, fw_on_flash, FW_VER_LEN) != 0) {
188 strcat(drvinfo->fw_version, " [");
189 strcat(drvinfo->fw_version, fw_on_flash);
190 strcat(drvinfo->fw_version, "]");
193 strlcpy(drvinfo->bus_info, pci_name(adapter->pdev),
194 sizeof(drvinfo->bus_info));
195 drvinfo->testinfo_len = 0;
196 drvinfo->regdump_len = 0;
197 drvinfo->eedump_len = 0;
201 lancer_cmd_get_file_len(struct be_adapter *adapter, u8 *file_name)
203 u32 data_read = 0, eof;
205 struct be_dma_mem data_len_cmd;
208 memset(&data_len_cmd, 0, sizeof(data_len_cmd));
209 /* data_offset and data_size should be 0 to get reg len */
210 status = lancer_cmd_read_object(adapter, &data_len_cmd, 0, 0,
211 file_name, &data_read, &eof, &addn_status);
217 lancer_cmd_read_file(struct be_adapter *adapter, u8 *file_name,
218 u32 buf_len, void *buf)
220 struct be_dma_mem read_cmd;
221 u32 read_len = 0, total_read_len = 0, chunk_size;
226 read_cmd.size = LANCER_READ_FILE_CHUNK;
227 read_cmd.va = pci_alloc_consistent(adapter->pdev, read_cmd.size,
231 dev_err(&adapter->pdev->dev,
232 "Memory allocation failure while reading dump\n");
236 while ((total_read_len < buf_len) && !eof) {
237 chunk_size = min_t(u32, (buf_len - total_read_len),
238 LANCER_READ_FILE_CHUNK);
239 chunk_size = ALIGN(chunk_size, 4);
240 status = lancer_cmd_read_object(adapter, &read_cmd, chunk_size,
241 total_read_len, file_name, &read_len,
244 memcpy(buf + total_read_len, read_cmd.va, read_len);
245 total_read_len += read_len;
246 eof &= LANCER_READ_FILE_EOF_MASK;
252 pci_free_consistent(adapter->pdev, read_cmd.size, read_cmd.va,
259 be_get_reg_len(struct net_device *netdev)
261 struct be_adapter *adapter = netdev_priv(netdev);
264 if (be_physfn(adapter)) {
265 if (lancer_chip(adapter))
266 log_size = lancer_cmd_get_file_len(adapter,
267 LANCER_FW_DUMP_FILE);
269 be_cmd_get_reg_len(adapter, &log_size);
275 be_get_regs(struct net_device *netdev, struct ethtool_regs *regs, void *buf)
277 struct be_adapter *adapter = netdev_priv(netdev);
279 if (be_physfn(adapter)) {
280 memset(buf, 0, regs->len);
281 if (lancer_chip(adapter))
282 lancer_cmd_read_file(adapter, LANCER_FW_DUMP_FILE,
285 be_cmd_get_regs(adapter, regs->len, buf);
289 static int be_get_coalesce(struct net_device *netdev,
290 struct ethtool_coalesce *et)
292 struct be_adapter *adapter = netdev_priv(netdev);
293 struct be_eq_obj *eqo = &adapter->eq_obj[0];
296 et->rx_coalesce_usecs = eqo->cur_eqd;
297 et->rx_coalesce_usecs_high = eqo->max_eqd;
298 et->rx_coalesce_usecs_low = eqo->min_eqd;
300 et->tx_coalesce_usecs = eqo->cur_eqd;
301 et->tx_coalesce_usecs_high = eqo->max_eqd;
302 et->tx_coalesce_usecs_low = eqo->min_eqd;
304 et->use_adaptive_rx_coalesce = eqo->enable_aic;
305 et->use_adaptive_tx_coalesce = eqo->enable_aic;
310 /* TX attributes are ignored. Only RX attributes are considered
311 * eqd cmd is issued in the worker thread.
313 static int be_set_coalesce(struct net_device *netdev,
314 struct ethtool_coalesce *et)
316 struct be_adapter *adapter = netdev_priv(netdev);
317 struct be_eq_obj *eqo;
320 for_all_evt_queues(adapter, eqo, i) {
321 eqo->enable_aic = et->use_adaptive_rx_coalesce;
322 eqo->max_eqd = min(et->rx_coalesce_usecs_high, BE_MAX_EQD);
323 eqo->min_eqd = min(et->rx_coalesce_usecs_low, eqo->max_eqd);
324 eqo->eqd = et->rx_coalesce_usecs;
331 be_get_ethtool_stats(struct net_device *netdev,
332 struct ethtool_stats *stats, uint64_t *data)
334 struct be_adapter *adapter = netdev_priv(netdev);
335 struct be_rx_obj *rxo;
336 struct be_tx_obj *txo;
338 unsigned int i, j, base = 0, start;
340 for (i = 0; i < ETHTOOL_STATS_NUM; i++) {
341 p = (u8 *)&adapter->drv_stats + et_stats[i].offset;
344 base += ETHTOOL_STATS_NUM;
346 for_all_rx_queues(adapter, rxo, j) {
347 struct be_rx_stats *stats = rx_stats(rxo);
350 start = u64_stats_fetch_begin_bh(&stats->sync);
351 data[base] = stats->rx_bytes;
352 data[base + 1] = stats->rx_pkts;
353 } while (u64_stats_fetch_retry_bh(&stats->sync, start));
355 for (i = 2; i < ETHTOOL_RXSTATS_NUM; i++) {
356 p = (u8 *)stats + et_rx_stats[i].offset;
357 data[base + i] = *(u32 *)p;
359 base += ETHTOOL_RXSTATS_NUM;
362 for_all_tx_queues(adapter, txo, j) {
363 struct be_tx_stats *stats = tx_stats(txo);
366 start = u64_stats_fetch_begin_bh(&stats->sync_compl);
367 data[base] = stats->tx_compl;
368 } while (u64_stats_fetch_retry_bh(&stats->sync_compl, start));
371 start = u64_stats_fetch_begin_bh(&stats->sync);
372 for (i = 1; i < ETHTOOL_TXSTATS_NUM; i++) {
373 p = (u8 *)stats + et_tx_stats[i].offset;
375 (et_tx_stats[i].size == sizeof(u64)) ?
376 *(u64 *)p : *(u32 *)p;
378 } while (u64_stats_fetch_retry_bh(&stats->sync, start));
379 base += ETHTOOL_TXSTATS_NUM;
384 be_get_stat_strings(struct net_device *netdev, uint32_t stringset,
387 struct be_adapter *adapter = netdev_priv(netdev);
392 for (i = 0; i < ETHTOOL_STATS_NUM; i++) {
393 memcpy(data, et_stats[i].desc, ETH_GSTRING_LEN);
394 data += ETH_GSTRING_LEN;
396 for (i = 0; i < adapter->num_rx_qs; i++) {
397 for (j = 0; j < ETHTOOL_RXSTATS_NUM; j++) {
398 sprintf(data, "rxq%d: %s", i,
399 et_rx_stats[j].desc);
400 data += ETH_GSTRING_LEN;
403 for (i = 0; i < adapter->num_tx_qs; i++) {
404 for (j = 0; j < ETHTOOL_TXSTATS_NUM; j++) {
405 sprintf(data, "txq%d: %s", i,
406 et_tx_stats[j].desc);
407 data += ETH_GSTRING_LEN;
412 for (i = 0; i < ETHTOOL_TESTS_NUM; i++) {
413 memcpy(data, et_self_tests[i], ETH_GSTRING_LEN);
414 data += ETH_GSTRING_LEN;
420 static int be_get_sset_count(struct net_device *netdev, int stringset)
422 struct be_adapter *adapter = netdev_priv(netdev);
426 return ETHTOOL_TESTS_NUM;
428 return ETHTOOL_STATS_NUM +
429 adapter->num_rx_qs * ETHTOOL_RXSTATS_NUM +
430 adapter->num_tx_qs * ETHTOOL_TXSTATS_NUM;
436 static u32 be_get_port_type(u32 phy_type, u32 dac_cable_len)
441 case PHY_TYPE_BASET_1GB:
442 case PHY_TYPE_BASEX_1GB:
446 case PHY_TYPE_SFP_PLUS_10GB:
447 port = dac_cable_len ? PORT_DA : PORT_FIBRE;
449 case PHY_TYPE_XFP_10GB:
450 case PHY_TYPE_SFP_1GB:
453 case PHY_TYPE_BASET_10GB:
463 static u32 convert_to_et_setting(u32 if_type, u32 if_speeds)
468 case PHY_TYPE_BASET_1GB:
469 case PHY_TYPE_BASEX_1GB:
472 if (if_speeds & BE_SUPPORTED_SPEED_1GBPS)
473 val |= SUPPORTED_1000baseT_Full;
474 if (if_speeds & BE_SUPPORTED_SPEED_100MBPS)
475 val |= SUPPORTED_100baseT_Full;
476 if (if_speeds & BE_SUPPORTED_SPEED_10MBPS)
477 val |= SUPPORTED_10baseT_Full;
479 case PHY_TYPE_KX4_10GB:
480 val |= SUPPORTED_Backplane;
481 if (if_speeds & BE_SUPPORTED_SPEED_1GBPS)
482 val |= SUPPORTED_1000baseKX_Full;
483 if (if_speeds & BE_SUPPORTED_SPEED_10GBPS)
484 val |= SUPPORTED_10000baseKX4_Full;
486 case PHY_TYPE_KR_10GB:
487 val |= SUPPORTED_Backplane |
488 SUPPORTED_10000baseKR_Full;
490 case PHY_TYPE_SFP_PLUS_10GB:
491 case PHY_TYPE_XFP_10GB:
492 case PHY_TYPE_SFP_1GB:
493 val |= SUPPORTED_FIBRE;
494 if (if_speeds & BE_SUPPORTED_SPEED_10GBPS)
495 val |= SUPPORTED_10000baseT_Full;
496 if (if_speeds & BE_SUPPORTED_SPEED_1GBPS)
497 val |= SUPPORTED_1000baseT_Full;
499 case PHY_TYPE_BASET_10GB:
501 if (if_speeds & BE_SUPPORTED_SPEED_10GBPS)
502 val |= SUPPORTED_10000baseT_Full;
503 if (if_speeds & BE_SUPPORTED_SPEED_1GBPS)
504 val |= SUPPORTED_1000baseT_Full;
505 if (if_speeds & BE_SUPPORTED_SPEED_100MBPS)
506 val |= SUPPORTED_100baseT_Full;
515 static int convert_to_et_speed(u32 be_speed)
517 int et_speed = SPEED_10000;
520 case PHY_LINK_SPEED_10MBPS:
523 case PHY_LINK_SPEED_100MBPS:
524 et_speed = SPEED_100;
526 case PHY_LINK_SPEED_1GBPS:
527 et_speed = SPEED_1000;
529 case PHY_LINK_SPEED_10GBPS:
530 et_speed = SPEED_10000;
537 bool be_pause_supported(struct be_adapter *adapter)
539 return (adapter->phy.interface_type == PHY_TYPE_SFP_PLUS_10GB ||
540 adapter->phy.interface_type == PHY_TYPE_XFP_10GB) ?
544 static int be_get_settings(struct net_device *netdev, struct ethtool_cmd *ecmd)
546 struct be_adapter *adapter = netdev_priv(netdev);
553 if (adapter->phy.link_speed < 0 || !(netdev->flags & IFF_UP)) {
554 if (adapter->phy.forced_port_speed < 0) {
555 status = be_cmd_link_status_query(adapter, &port_speed,
556 &link_speed, &link_status, 0);
558 be_link_status_update(adapter, link_status);
560 et_speed = link_speed * 10;
561 else if (link_status)
562 et_speed = convert_to_et_speed(port_speed);
564 et_speed = adapter->phy.forced_port_speed;
567 ethtool_cmd_speed_set(ecmd, et_speed);
569 status = be_cmd_get_phy_info(adapter);
574 convert_to_et_setting(adapter->phy.interface_type,
575 adapter->phy.auto_speeds_supported |
576 adapter->phy.fixed_speeds_supported);
578 convert_to_et_setting(adapter->phy.interface_type,
579 adapter->phy.auto_speeds_supported);
581 ecmd->port = be_get_port_type(adapter->phy.interface_type,
582 adapter->phy.dac_cable_len);
584 if (adapter->phy.auto_speeds_supported) {
585 ecmd->supported |= SUPPORTED_Autoneg;
586 ecmd->autoneg = AUTONEG_ENABLE;
587 ecmd->advertising |= ADVERTISED_Autoneg;
590 if (be_pause_supported(adapter)) {
591 ecmd->supported |= SUPPORTED_Pause;
592 ecmd->advertising |= ADVERTISED_Pause;
595 switch (adapter->phy.interface_type) {
596 case PHY_TYPE_KR_10GB:
597 case PHY_TYPE_KX4_10GB:
598 ecmd->transceiver = XCVR_INTERNAL;
601 ecmd->transceiver = XCVR_EXTERNAL;
605 /* Save for future use */
606 adapter->phy.link_speed = ethtool_cmd_speed(ecmd);
607 adapter->phy.port_type = ecmd->port;
608 adapter->phy.transceiver = ecmd->transceiver;
609 adapter->phy.autoneg = ecmd->autoneg;
610 adapter->phy.advertising = ecmd->advertising;
611 adapter->phy.supported = ecmd->supported;
613 ethtool_cmd_speed_set(ecmd, adapter->phy.link_speed);
614 ecmd->port = adapter->phy.port_type;
615 ecmd->transceiver = adapter->phy.transceiver;
616 ecmd->autoneg = adapter->phy.autoneg;
617 ecmd->advertising = adapter->phy.advertising;
618 ecmd->supported = adapter->phy.supported;
621 ecmd->duplex = netif_carrier_ok(netdev) ? DUPLEX_FULL : DUPLEX_UNKNOWN;
622 ecmd->phy_address = adapter->port_num;
627 static void be_get_ringparam(struct net_device *netdev,
628 struct ethtool_ringparam *ring)
630 struct be_adapter *adapter = netdev_priv(netdev);
632 ring->rx_max_pending = ring->rx_pending = adapter->rx_obj[0].q.len;
633 ring->tx_max_pending = ring->tx_pending = adapter->tx_obj[0].q.len;
637 be_get_pauseparam(struct net_device *netdev, struct ethtool_pauseparam *ecmd)
639 struct be_adapter *adapter = netdev_priv(netdev);
641 be_cmd_get_flow_control(adapter, &ecmd->tx_pause, &ecmd->rx_pause);
642 ecmd->autoneg = adapter->phy.fc_autoneg;
646 be_set_pauseparam(struct net_device *netdev, struct ethtool_pauseparam *ecmd)
648 struct be_adapter *adapter = netdev_priv(netdev);
651 if (ecmd->autoneg != adapter->phy.fc_autoneg)
653 adapter->tx_fc = ecmd->tx_pause;
654 adapter->rx_fc = ecmd->rx_pause;
656 status = be_cmd_set_flow_control(adapter,
657 adapter->tx_fc, adapter->rx_fc);
659 dev_warn(&adapter->pdev->dev, "Pause param set failed.\n");
665 be_set_phys_id(struct net_device *netdev,
666 enum ethtool_phys_id_state state)
668 struct be_adapter *adapter = netdev_priv(netdev);
671 case ETHTOOL_ID_ACTIVE:
672 be_cmd_get_beacon_state(adapter, adapter->hba_port_num,
673 &adapter->beacon_state);
674 return 1; /* cycle on/off once per second */
677 be_cmd_set_beacon_state(adapter, adapter->hba_port_num, 0, 0,
678 BEACON_STATE_ENABLED);
682 be_cmd_set_beacon_state(adapter, adapter->hba_port_num, 0, 0,
683 BEACON_STATE_DISABLED);
686 case ETHTOOL_ID_INACTIVE:
687 be_cmd_set_beacon_state(adapter, adapter->hba_port_num, 0, 0,
688 adapter->beacon_state);
696 be_get_wol(struct net_device *netdev, struct ethtool_wolinfo *wol)
698 struct be_adapter *adapter = netdev_priv(netdev);
700 if (be_is_wol_supported(adapter)) {
701 wol->supported |= WAKE_MAGIC;
702 wol->wolopts |= WAKE_MAGIC;
705 memset(&wol->sopass, 0, sizeof(wol->sopass));
709 be_set_wol(struct net_device *netdev, struct ethtool_wolinfo *wol)
711 struct be_adapter *adapter = netdev_priv(netdev);
713 if (wol->wolopts & ~WAKE_MAGIC)
716 if (!be_is_wol_supported(adapter)) {
717 dev_warn(&adapter->pdev->dev, "WOL not supported\n");
721 if (wol->wolopts & WAKE_MAGIC)
724 adapter->wol = false;
730 be_test_ddr_dma(struct be_adapter *adapter)
733 struct be_dma_mem ddrdma_cmd;
734 static const u64 pattern[2] = {
735 0x5a5a5a5a5a5a5a5aULL, 0xa5a5a5a5a5a5a5a5ULL
738 ddrdma_cmd.size = sizeof(struct be_cmd_req_ddrdma_test);
739 ddrdma_cmd.va = dma_alloc_coherent(&adapter->pdev->dev, ddrdma_cmd.size,
740 &ddrdma_cmd.dma, GFP_KERNEL);
741 if (!ddrdma_cmd.va) {
742 dev_err(&adapter->pdev->dev, "Memory allocation failure\n");
746 for (i = 0; i < 2; i++) {
747 ret = be_cmd_ddr_dma_test(adapter, pattern[i],
754 dma_free_coherent(&adapter->pdev->dev, ddrdma_cmd.size, ddrdma_cmd.va,
759 static u64 be_loopback_test(struct be_adapter *adapter, u8 loopback_type,
762 be_cmd_set_loopback(adapter, adapter->hba_port_num,
764 *status = be_cmd_loopback_test(adapter, adapter->hba_port_num,
767 be_cmd_set_loopback(adapter, adapter->hba_port_num,
773 be_self_test(struct net_device *netdev, struct ethtool_test *test, u64 *data)
775 struct be_adapter *adapter = netdev_priv(netdev);
777 u16 qos_link_speed = 0;
779 memset(data, 0, sizeof(u64) * ETHTOOL_TESTS_NUM);
781 if (test->flags & ETH_TEST_FL_OFFLINE) {
782 if (be_loopback_test(adapter, BE_MAC_LOOPBACK,
784 test->flags |= ETH_TEST_FL_FAILED;
786 if (be_loopback_test(adapter, BE_PHY_LOOPBACK,
788 test->flags |= ETH_TEST_FL_FAILED;
790 if (be_loopback_test(adapter, BE_ONE_PORT_EXT_LOOPBACK,
792 test->flags |= ETH_TEST_FL_FAILED;
796 if (!lancer_chip(adapter) && be_test_ddr_dma(adapter) != 0) {
798 test->flags |= ETH_TEST_FL_FAILED;
801 if (be_cmd_link_status_query(adapter, &mac_speed,
802 &qos_link_speed, NULL, 0) != 0) {
803 test->flags |= ETH_TEST_FL_FAILED;
805 } else if (!mac_speed) {
806 test->flags |= ETH_TEST_FL_FAILED;
812 be_do_flash(struct net_device *netdev, struct ethtool_flash *efl)
814 struct be_adapter *adapter = netdev_priv(netdev);
816 return be_load_fw(adapter, efl->data);
820 be_get_eeprom_len(struct net_device *netdev)
822 struct be_adapter *adapter = netdev_priv(netdev);
823 if (lancer_chip(adapter)) {
824 if (be_physfn(adapter))
825 return lancer_cmd_get_file_len(adapter,
828 return lancer_cmd_get_file_len(adapter,
831 return BE_READ_SEEPROM_LEN;
836 be_read_eeprom(struct net_device *netdev, struct ethtool_eeprom *eeprom,
839 struct be_adapter *adapter = netdev_priv(netdev);
840 struct be_dma_mem eeprom_cmd;
841 struct be_cmd_resp_seeprom_read *resp;
847 if (lancer_chip(adapter)) {
848 if (be_physfn(adapter))
849 return lancer_cmd_read_file(adapter, LANCER_VPD_PF_FILE,
852 return lancer_cmd_read_file(adapter, LANCER_VPD_VF_FILE,
856 eeprom->magic = BE_VENDOR_ID | (adapter->pdev->device<<16);
858 memset(&eeprom_cmd, 0, sizeof(struct be_dma_mem));
859 eeprom_cmd.size = sizeof(struct be_cmd_req_seeprom_read);
860 eeprom_cmd.va = dma_alloc_coherent(&adapter->pdev->dev, eeprom_cmd.size,
861 &eeprom_cmd.dma, GFP_KERNEL);
863 if (!eeprom_cmd.va) {
864 dev_err(&adapter->pdev->dev,
865 "Memory allocation failure. Could not read eeprom\n");
869 status = be_cmd_get_seeprom_data(adapter, &eeprom_cmd);
872 resp = eeprom_cmd.va;
873 memcpy(data, resp->seeprom_data + eeprom->offset, eeprom->len);
875 dma_free_coherent(&adapter->pdev->dev, eeprom_cmd.size, eeprom_cmd.va,
881 static u32 be_get_msg_level(struct net_device *netdev)
883 struct be_adapter *adapter = netdev_priv(netdev);
885 if (lancer_chip(adapter)) {
886 dev_err(&adapter->pdev->dev, "Operation not supported\n");
890 return adapter->msg_enable;
893 static void be_set_fw_log_level(struct be_adapter *adapter, u32 level)
895 struct be_dma_mem extfat_cmd;
896 struct be_fat_conf_params *cfgs;
900 memset(&extfat_cmd, 0, sizeof(struct be_dma_mem));
901 extfat_cmd.size = sizeof(struct be_cmd_resp_get_ext_fat_caps);
902 extfat_cmd.va = pci_alloc_consistent(adapter->pdev, extfat_cmd.size,
904 if (!extfat_cmd.va) {
905 dev_err(&adapter->pdev->dev, "%s: Memory allocation failure\n",
909 status = be_cmd_get_ext_fat_capabilites(adapter, &extfat_cmd);
911 cfgs = (struct be_fat_conf_params *)(extfat_cmd.va +
912 sizeof(struct be_cmd_resp_hdr));
913 for (i = 0; i < le32_to_cpu(cfgs->num_modules); i++) {
914 u32 num_modes = le32_to_cpu(cfgs->module[i].num_modes);
915 for (j = 0; j < num_modes; j++) {
916 if (cfgs->module[i].trace_lvl[j].mode ==
918 cfgs->module[i].trace_lvl[j].dbg_lvl =
922 status = be_cmd_set_ext_fat_capabilites(adapter, &extfat_cmd,
925 dev_err(&adapter->pdev->dev,
926 "Message level set failed\n");
928 dev_err(&adapter->pdev->dev, "Message level get failed\n");
931 pci_free_consistent(adapter->pdev, extfat_cmd.size, extfat_cmd.va,
937 static void be_set_msg_level(struct net_device *netdev, u32 level)
939 struct be_adapter *adapter = netdev_priv(netdev);
941 if (lancer_chip(adapter)) {
942 dev_err(&adapter->pdev->dev, "Operation not supported\n");
946 if (adapter->msg_enable == level)
949 if ((level & NETIF_MSG_HW) != (adapter->msg_enable & NETIF_MSG_HW))
950 be_set_fw_log_level(adapter, level & NETIF_MSG_HW ?
951 FW_LOG_LEVEL_DEFAULT : FW_LOG_LEVEL_FATAL);
952 adapter->msg_enable = level;
957 const struct ethtool_ops be_ethtool_ops = {
958 .get_settings = be_get_settings,
959 .get_drvinfo = be_get_drvinfo,
960 .get_wol = be_get_wol,
961 .set_wol = be_set_wol,
962 .get_link = ethtool_op_get_link,
963 .get_eeprom_len = be_get_eeprom_len,
964 .get_eeprom = be_read_eeprom,
965 .get_coalesce = be_get_coalesce,
966 .set_coalesce = be_set_coalesce,
967 .get_ringparam = be_get_ringparam,
968 .get_pauseparam = be_get_pauseparam,
969 .set_pauseparam = be_set_pauseparam,
970 .get_strings = be_get_stat_strings,
971 .set_phys_id = be_set_phys_id,
972 .get_msglevel = be_get_msg_level,
973 .set_msglevel = be_set_msg_level,
974 .get_sset_count = be_get_sset_count,
975 .get_ethtool_stats = be_get_ethtool_stats,
976 .get_regs_len = be_get_reg_len,
977 .get_regs = be_get_regs,
978 .flash_device = be_do_flash,
979 .self_test = be_self_test,