M7350/kernel/drivers/net/ethernet/emulex/benet/be_ethtool.c

1317 lines
36 KiB
C
Raw Permalink Normal View History

2024-09-09 08:52:07 +00:00
/*
2024-09-09 08:57:42 +00:00
* Copyright (C) 2005 - 2014 Emulex
2024-09-09 08:52:07 +00:00
* All rights reserved.
*
* This program is free software; you can redistribute it and/or
* modify it under the terms of the GNU General Public License version 2
* as published by the Free Software Foundation. The full GNU General
* Public License is included in this distribution in the file called COPYING.
*
* Contact Information:
* linux-drivers@emulex.com
*
* Emulex
* 3333 Susan Street
* Costa Mesa, CA 92626
*/
#include "be.h"
#include "be_cmds.h"
#include <linux/ethtool.h>
struct be_ethtool_stat {
char desc[ETH_GSTRING_LEN];
int type;
int size;
int offset;
};
enum {DRVSTAT_TX, DRVSTAT_RX, DRVSTAT};
#define FIELDINFO(_struct, field) FIELD_SIZEOF(_struct, field), \
offsetof(_struct, field)
#define DRVSTAT_TX_INFO(field) #field, DRVSTAT_TX,\
FIELDINFO(struct be_tx_stats, field)
#define DRVSTAT_RX_INFO(field) #field, DRVSTAT_RX,\
FIELDINFO(struct be_rx_stats, field)
#define DRVSTAT_INFO(field) #field, DRVSTAT,\
FIELDINFO(struct be_drv_stats, field)
static const struct be_ethtool_stat et_stats[] = {
{DRVSTAT_INFO(rx_crc_errors)},
{DRVSTAT_INFO(rx_alignment_symbol_errors)},
{DRVSTAT_INFO(rx_pause_frames)},
{DRVSTAT_INFO(rx_control_frames)},
/* Received packets dropped when the Ethernet length field
* is not equal to the actual Ethernet data length.
*/
{DRVSTAT_INFO(rx_in_range_errors)},
/* Received packets dropped when their length field is >= 1501 bytes
* and <= 1535 bytes.
*/
{DRVSTAT_INFO(rx_out_range_errors)},
/* Received packets dropped when they are longer than 9216 bytes */
{DRVSTAT_INFO(rx_frame_too_long)},
/* Received packets dropped when they don't pass the unicast or
* multicast address filtering.
*/
2024-09-09 08:57:42 +00:00
{DRVSTAT_INFO(rx_address_filtered)},
2024-09-09 08:52:07 +00:00
/* Received packets dropped when IP packet length field is less than
* the IP header length field.
*/
{DRVSTAT_INFO(rx_dropped_too_small)},
/* Received packets dropped when IP length field is greater than
* the actual packet length.
*/
{DRVSTAT_INFO(rx_dropped_too_short)},
/* Received packets dropped when the IP header length field is less
* than 5.
*/
{DRVSTAT_INFO(rx_dropped_header_too_small)},
/* Received packets dropped when the TCP header length field is less
* than 5 or the TCP header length + IP header length is more
* than IP packet length.
*/
{DRVSTAT_INFO(rx_dropped_tcp_length)},
{DRVSTAT_INFO(rx_dropped_runt)},
/* Number of received packets dropped when a fifo for descriptors going
* into the packet demux block overflows. In normal operation, this
* fifo must never overflow.
*/
{DRVSTAT_INFO(rxpp_fifo_overflow_drop)},
2024-09-09 08:57:42 +00:00
/* Received packets dropped when the RX block runs out of space in
* one of its input FIFOs. This could happen due a long burst of
* minimum-sized (64b) frames in the receive path.
* This counter may also be erroneously incremented rarely.
*/
2024-09-09 08:52:07 +00:00
{DRVSTAT_INFO(rx_input_fifo_overflow_drop)},
{DRVSTAT_INFO(rx_ip_checksum_errs)},
{DRVSTAT_INFO(rx_tcp_checksum_errs)},
{DRVSTAT_INFO(rx_udp_checksum_errs)},
{DRVSTAT_INFO(tx_pauseframes)},
{DRVSTAT_INFO(tx_controlframes)},
{DRVSTAT_INFO(rx_priority_pause_frames)},
2024-09-09 08:57:42 +00:00
{DRVSTAT_INFO(tx_priority_pauseframes)},
2024-09-09 08:52:07 +00:00
/* Received packets dropped when an internal fifo going into
* main packet buffer tank (PMEM) overflows.
*/
{DRVSTAT_INFO(pmem_fifo_overflow_drop)},
{DRVSTAT_INFO(jabber_events)},
/* Received packets dropped due to lack of available HW packet buffers
* used to temporarily hold the received packets.
*/
{DRVSTAT_INFO(rx_drops_no_pbuf)},
/* Received packets dropped due to input receive buffer
* descriptor fifo overflowing.
*/
{DRVSTAT_INFO(rx_drops_no_erx_descr)},
/* Packets dropped because the internal FIFO to the offloaded TCP
* receive processing block is full. This could happen only for
* offloaded iSCSI or FCoE trarffic.
*/
{DRVSTAT_INFO(rx_drops_no_tpre_descr)},
/* Received packets dropped when they need more than 8
* receive buffers. This cannot happen as the driver configures
* 2048 byte receive buffers.
*/
{DRVSTAT_INFO(rx_drops_too_many_frags)},
{DRVSTAT_INFO(forwarded_packets)},
/* Received packets dropped when the frame length
* is more than 9018 bytes
*/
{DRVSTAT_INFO(rx_drops_mtu)},
2024-09-09 08:57:42 +00:00
/* Number of dma mapping errors */
{DRVSTAT_INFO(dma_map_errors)},
2024-09-09 08:52:07 +00:00
/* Number of packets dropped due to random early drop function */
{DRVSTAT_INFO(eth_red_drops)},
2024-09-09 08:57:42 +00:00
{DRVSTAT_INFO(be_on_die_temperature)},
{DRVSTAT_INFO(rx_roce_bytes_lsd)},
{DRVSTAT_INFO(rx_roce_bytes_msd)},
{DRVSTAT_INFO(rx_roce_frames)},
{DRVSTAT_INFO(roce_drops_payload_len)},
{DRVSTAT_INFO(roce_drops_crc)}
2024-09-09 08:52:07 +00:00
};
2024-09-09 08:57:42 +00:00
2024-09-09 08:52:07 +00:00
#define ETHTOOL_STATS_NUM ARRAY_SIZE(et_stats)
/* Stats related to multi RX queues: get_stats routine assumes bytes, pkts
* are first and second members respectively.
*/
static const struct be_ethtool_stat et_rx_stats[] = {
{DRVSTAT_RX_INFO(rx_bytes)},/* If moving this member see above note */
{DRVSTAT_RX_INFO(rx_pkts)}, /* If moving this member see above note */
{DRVSTAT_RX_INFO(rx_compl)},
2024-09-09 08:57:42 +00:00
{DRVSTAT_RX_INFO(rx_compl_err)},
2024-09-09 08:52:07 +00:00
{DRVSTAT_RX_INFO(rx_mcast_pkts)},
/* Number of page allocation failures while posting receive buffers
* to HW.
*/
{DRVSTAT_RX_INFO(rx_post_fail)},
/* Recevied packets dropped due to skb allocation failure */
{DRVSTAT_RX_INFO(rx_drops_no_skbs)},
/* Received packets dropped due to lack of available fetched buffers
* posted by the driver.
*/
{DRVSTAT_RX_INFO(rx_drops_no_frags)}
};
2024-09-09 08:57:42 +00:00
2024-09-09 08:52:07 +00:00
#define ETHTOOL_RXSTATS_NUM (ARRAY_SIZE(et_rx_stats))
/* Stats related to multi TX queues: get_stats routine assumes compl is the
* first member
*/
static const struct be_ethtool_stat et_tx_stats[] = {
{DRVSTAT_TX_INFO(tx_compl)}, /* If moving this member see above note */
2024-09-09 08:57:42 +00:00
/* This counter is incremented when the HW encounters an error while
* parsing the packet header of an outgoing TX request. This counter is
* applicable only for BE2, BE3 and Skyhawk based adapters.
*/
{DRVSTAT_TX_INFO(tx_hdr_parse_err)},
/* This counter is incremented when an error occurs in the DMA
* operation associated with the TX request from the host to the device.
*/
{DRVSTAT_TX_INFO(tx_dma_err)},
/* This counter is incremented when MAC or VLAN spoof checking is
* enabled on the interface and the TX request fails the spoof check
* in HW.
*/
{DRVSTAT_TX_INFO(tx_spoof_check_err)},
/* This counter is incremented when the HW encounters an error while
* performing TSO offload. This counter is applicable only for Lancer
* adapters.
*/
{DRVSTAT_TX_INFO(tx_tso_err)},
/* This counter is incremented when the HW detects Q-in-Q style VLAN
* tagging in a packet and such tagging is not expected on the outgoing
* interface. This counter is applicable only for Lancer adapters.
*/
{DRVSTAT_TX_INFO(tx_qinq_err)},
/* This counter is incremented when the HW detects parity errors in the
* packet data. This counter is applicable only for Lancer adapters.
*/
{DRVSTAT_TX_INFO(tx_internal_parity_err)},
2024-09-09 08:52:07 +00:00
{DRVSTAT_TX_INFO(tx_bytes)},
{DRVSTAT_TX_INFO(tx_pkts)},
/* Number of skbs queued for trasmission by the driver */
{DRVSTAT_TX_INFO(tx_reqs)},
/* Number of TX work request blocks DMAed to HW */
{DRVSTAT_TX_INFO(tx_wrbs)},
/* Number of times the TX queue was stopped due to lack
* of spaces in the TXQ.
*/
2024-09-09 08:57:42 +00:00
{DRVSTAT_TX_INFO(tx_stops)},
/* Pkts dropped in the driver's transmit path */
{DRVSTAT_TX_INFO(tx_drv_drops)}
2024-09-09 08:52:07 +00:00
};
2024-09-09 08:57:42 +00:00
2024-09-09 08:52:07 +00:00
#define ETHTOOL_TXSTATS_NUM (ARRAY_SIZE(et_tx_stats))
static const char et_self_tests[][ETH_GSTRING_LEN] = {
"MAC Loopback test",
"PHY Loopback test",
"External Loopback test",
"DDR DMA test",
"Link test"
};
#define ETHTOOL_TESTS_NUM ARRAY_SIZE(et_self_tests)
#define BE_MAC_LOOPBACK 0x0
#define BE_PHY_LOOPBACK 0x1
#define BE_ONE_PORT_EXT_LOOPBACK 0x2
#define BE_NO_LOOPBACK 0xff
static void be_get_drvinfo(struct net_device *netdev,
2024-09-09 08:57:42 +00:00
struct ethtool_drvinfo *drvinfo)
2024-09-09 08:52:07 +00:00
{
struct be_adapter *adapter = netdev_priv(netdev);
strlcpy(drvinfo->driver, DRV_NAME, sizeof(drvinfo->driver));
strlcpy(drvinfo->version, DRV_VER, sizeof(drvinfo->version));
2024-09-09 08:57:42 +00:00
if (!memcmp(adapter->fw_ver, adapter->fw_on_flash, FW_VER_LEN))
strlcpy(drvinfo->fw_version, adapter->fw_ver,
sizeof(drvinfo->fw_version));
else
snprintf(drvinfo->fw_version, sizeof(drvinfo->fw_version),
"%s [%s]", adapter->fw_ver, adapter->fw_on_flash);
2024-09-09 08:52:07 +00:00
strlcpy(drvinfo->bus_info, pci_name(adapter->pdev),
sizeof(drvinfo->bus_info));
drvinfo->testinfo_len = 0;
drvinfo->regdump_len = 0;
drvinfo->eedump_len = 0;
}
2024-09-09 08:57:42 +00:00
static u32 lancer_cmd_get_file_len(struct be_adapter *adapter, u8 *file_name)
2024-09-09 08:52:07 +00:00
{
u32 data_read = 0, eof;
u8 addn_status;
struct be_dma_mem data_len_cmd;
int status;
memset(&data_len_cmd, 0, sizeof(data_len_cmd));
/* data_offset and data_size should be 0 to get reg len */
status = lancer_cmd_read_object(adapter, &data_len_cmd, 0, 0,
2024-09-09 08:57:42 +00:00
file_name, &data_read, &eof,
&addn_status);
2024-09-09 08:52:07 +00:00
return data_read;
}
2024-09-09 08:57:42 +00:00
static int lancer_cmd_read_file(struct be_adapter *adapter, u8 *file_name,
u32 buf_len, void *buf)
2024-09-09 08:52:07 +00:00
{
struct be_dma_mem read_cmd;
u32 read_len = 0, total_read_len = 0, chunk_size;
u32 eof = 0;
u8 addn_status;
int status = 0;
read_cmd.size = LANCER_READ_FILE_CHUNK;
2024-09-09 08:57:42 +00:00
read_cmd.va = dma_zalloc_coherent(&adapter->pdev->dev, read_cmd.size,
&read_cmd.dma, GFP_ATOMIC);
2024-09-09 08:52:07 +00:00
if (!read_cmd.va) {
dev_err(&adapter->pdev->dev,
2024-09-09 08:57:42 +00:00
"Memory allocation failure while reading dump\n");
2024-09-09 08:52:07 +00:00
return -ENOMEM;
}
while ((total_read_len < buf_len) && !eof) {
chunk_size = min_t(u32, (buf_len - total_read_len),
2024-09-09 08:57:42 +00:00
LANCER_READ_FILE_CHUNK);
2024-09-09 08:52:07 +00:00
chunk_size = ALIGN(chunk_size, 4);
status = lancer_cmd_read_object(adapter, &read_cmd, chunk_size,
2024-09-09 08:57:42 +00:00
total_read_len, file_name,
&read_len, &eof, &addn_status);
2024-09-09 08:52:07 +00:00
if (!status) {
memcpy(buf + total_read_len, read_cmd.va, read_len);
total_read_len += read_len;
eof &= LANCER_READ_FILE_EOF_MASK;
} else {
status = -EIO;
break;
}
}
2024-09-09 08:57:42 +00:00
dma_free_coherent(&adapter->pdev->dev, read_cmd.size, read_cmd.va,
read_cmd.dma);
2024-09-09 08:52:07 +00:00
return status;
}
2024-09-09 08:57:42 +00:00
static int be_get_reg_len(struct net_device *netdev)
2024-09-09 08:52:07 +00:00
{
struct be_adapter *adapter = netdev_priv(netdev);
u32 log_size = 0;
2024-09-09 08:57:42 +00:00
if (!check_privilege(adapter, MAX_PRIVILEGES))
return 0;
2024-09-09 08:52:07 +00:00
if (be_physfn(adapter)) {
if (lancer_chip(adapter))
log_size = lancer_cmd_get_file_len(adapter,
2024-09-09 08:57:42 +00:00
LANCER_FW_DUMP_FILE);
2024-09-09 08:52:07 +00:00
else
be_cmd_get_reg_len(adapter, &log_size);
}
return log_size;
}
static void
be_get_regs(struct net_device *netdev, struct ethtool_regs *regs, void *buf)
{
struct be_adapter *adapter = netdev_priv(netdev);
if (be_physfn(adapter)) {
memset(buf, 0, regs->len);
if (lancer_chip(adapter))
lancer_cmd_read_file(adapter, LANCER_FW_DUMP_FILE,
2024-09-09 08:57:42 +00:00
regs->len, buf);
2024-09-09 08:52:07 +00:00
else
be_cmd_get_regs(adapter, regs->len, buf);
}
}
static int be_get_coalesce(struct net_device *netdev,
struct ethtool_coalesce *et)
{
struct be_adapter *adapter = netdev_priv(netdev);
2024-09-09 08:57:42 +00:00
struct be_aic_obj *aic = &adapter->aic_obj[0];
2024-09-09 08:52:07 +00:00
2024-09-09 08:57:42 +00:00
et->rx_coalesce_usecs = aic->prev_eqd;
et->rx_coalesce_usecs_high = aic->max_eqd;
et->rx_coalesce_usecs_low = aic->min_eqd;
2024-09-09 08:52:07 +00:00
2024-09-09 08:57:42 +00:00
et->tx_coalesce_usecs = aic->prev_eqd;
et->tx_coalesce_usecs_high = aic->max_eqd;
et->tx_coalesce_usecs_low = aic->min_eqd;
2024-09-09 08:52:07 +00:00
2024-09-09 08:57:42 +00:00
et->use_adaptive_rx_coalesce = aic->enable;
et->use_adaptive_tx_coalesce = aic->enable;
2024-09-09 08:52:07 +00:00
return 0;
}
/* TX attributes are ignored. Only RX attributes are considered
* eqd cmd is issued in the worker thread.
*/
static int be_set_coalesce(struct net_device *netdev,
struct ethtool_coalesce *et)
{
struct be_adapter *adapter = netdev_priv(netdev);
2024-09-09 08:57:42 +00:00
struct be_aic_obj *aic = &adapter->aic_obj[0];
2024-09-09 08:52:07 +00:00
struct be_eq_obj *eqo;
int i;
for_all_evt_queues(adapter, eqo, i) {
2024-09-09 08:57:42 +00:00
aic->enable = et->use_adaptive_rx_coalesce;
aic->max_eqd = min(et->rx_coalesce_usecs_high, BE_MAX_EQD);
aic->min_eqd = min(et->rx_coalesce_usecs_low, aic->max_eqd);
aic->et_eqd = min(et->rx_coalesce_usecs, aic->max_eqd);
aic->et_eqd = max(aic->et_eqd, aic->min_eqd);
aic++;
2024-09-09 08:52:07 +00:00
}
return 0;
}
2024-09-09 08:57:42 +00:00
static void be_get_ethtool_stats(struct net_device *netdev,
struct ethtool_stats *stats, uint64_t *data)
2024-09-09 08:52:07 +00:00
{
struct be_adapter *adapter = netdev_priv(netdev);
struct be_rx_obj *rxo;
struct be_tx_obj *txo;
void *p;
unsigned int i, j, base = 0, start;
for (i = 0; i < ETHTOOL_STATS_NUM; i++) {
p = (u8 *)&adapter->drv_stats + et_stats[i].offset;
data[i] = *(u32 *)p;
}
base += ETHTOOL_STATS_NUM;
for_all_rx_queues(adapter, rxo, j) {
struct be_rx_stats *stats = rx_stats(rxo);
do {
2024-09-09 08:57:42 +00:00
start = u64_stats_fetch_begin_irq(&stats->sync);
2024-09-09 08:52:07 +00:00
data[base] = stats->rx_bytes;
data[base + 1] = stats->rx_pkts;
2024-09-09 08:57:42 +00:00
} while (u64_stats_fetch_retry_irq(&stats->sync, start));
2024-09-09 08:52:07 +00:00
for (i = 2; i < ETHTOOL_RXSTATS_NUM; i++) {
p = (u8 *)stats + et_rx_stats[i].offset;
data[base + i] = *(u32 *)p;
}
base += ETHTOOL_RXSTATS_NUM;
}
for_all_tx_queues(adapter, txo, j) {
struct be_tx_stats *stats = tx_stats(txo);
do {
2024-09-09 08:57:42 +00:00
start = u64_stats_fetch_begin_irq(&stats->sync_compl);
2024-09-09 08:52:07 +00:00
data[base] = stats->tx_compl;
2024-09-09 08:57:42 +00:00
} while (u64_stats_fetch_retry_irq(&stats->sync_compl, start));
2024-09-09 08:52:07 +00:00
do {
2024-09-09 08:57:42 +00:00
start = u64_stats_fetch_begin_irq(&stats->sync);
2024-09-09 08:52:07 +00:00
for (i = 1; i < ETHTOOL_TXSTATS_NUM; i++) {
p = (u8 *)stats + et_tx_stats[i].offset;
data[base + i] =
(et_tx_stats[i].size == sizeof(u64)) ?
*(u64 *)p : *(u32 *)p;
}
2024-09-09 08:57:42 +00:00
} while (u64_stats_fetch_retry_irq(&stats->sync, start));
2024-09-09 08:52:07 +00:00
base += ETHTOOL_TXSTATS_NUM;
}
}
2024-09-09 08:57:42 +00:00
static void be_get_stat_strings(struct net_device *netdev, uint32_t stringset,
uint8_t *data)
2024-09-09 08:52:07 +00:00
{
struct be_adapter *adapter = netdev_priv(netdev);
int i, j;
switch (stringset) {
case ETH_SS_STATS:
for (i = 0; i < ETHTOOL_STATS_NUM; i++) {
memcpy(data, et_stats[i].desc, ETH_GSTRING_LEN);
data += ETH_GSTRING_LEN;
}
for (i = 0; i < adapter->num_rx_qs; i++) {
for (j = 0; j < ETHTOOL_RXSTATS_NUM; j++) {
sprintf(data, "rxq%d: %s", i,
et_rx_stats[j].desc);
data += ETH_GSTRING_LEN;
}
}
for (i = 0; i < adapter->num_tx_qs; i++) {
for (j = 0; j < ETHTOOL_TXSTATS_NUM; j++) {
sprintf(data, "txq%d: %s", i,
et_tx_stats[j].desc);
data += ETH_GSTRING_LEN;
}
}
break;
case ETH_SS_TEST:
for (i = 0; i < ETHTOOL_TESTS_NUM; i++) {
memcpy(data, et_self_tests[i], ETH_GSTRING_LEN);
data += ETH_GSTRING_LEN;
}
break;
}
}
static int be_get_sset_count(struct net_device *netdev, int stringset)
{
struct be_adapter *adapter = netdev_priv(netdev);
switch (stringset) {
case ETH_SS_TEST:
return ETHTOOL_TESTS_NUM;
case ETH_SS_STATS:
return ETHTOOL_STATS_NUM +
adapter->num_rx_qs * ETHTOOL_RXSTATS_NUM +
adapter->num_tx_qs * ETHTOOL_TXSTATS_NUM;
default:
return -EINVAL;
}
}
2024-09-09 08:57:42 +00:00
static u32 be_get_port_type(struct be_adapter *adapter)
2024-09-09 08:52:07 +00:00
{
2024-09-09 08:57:42 +00:00
u32 port;
2024-09-09 08:52:07 +00:00
2024-09-09 08:57:42 +00:00
switch (adapter->phy.interface_type) {
case PHY_TYPE_BASET_1GB:
case PHY_TYPE_BASEX_1GB:
case PHY_TYPE_SGMII:
port = PORT_TP;
break;
case PHY_TYPE_SFP_PLUS_10GB:
if (adapter->phy.cable_type & SFP_PLUS_COPPER_CABLE)
port = PORT_DA;
else
port = PORT_FIBRE;
break;
case PHY_TYPE_QSFP:
if (adapter->phy.cable_type & QSFP_PLUS_CR4_CABLE)
port = PORT_DA;
else
port = PORT_FIBRE;
break;
case PHY_TYPE_XFP_10GB:
case PHY_TYPE_SFP_1GB:
port = PORT_FIBRE;
break;
case PHY_TYPE_BASET_10GB:
port = PORT_TP;
break;
default:
port = PORT_OTHER;
}
2024-09-09 08:52:07 +00:00
2024-09-09 08:57:42 +00:00
return port;
}
static u32 convert_to_et_setting(struct be_adapter *adapter, u32 if_speeds)
{
u32 val = 0;
switch (adapter->phy.interface_type) {
case PHY_TYPE_BASET_1GB:
case PHY_TYPE_BASEX_1GB:
case PHY_TYPE_SGMII:
val |= SUPPORTED_TP;
if (if_speeds & BE_SUPPORTED_SPEED_1GBPS)
val |= SUPPORTED_1000baseT_Full;
if (if_speeds & BE_SUPPORTED_SPEED_100MBPS)
val |= SUPPORTED_100baseT_Full;
if (if_speeds & BE_SUPPORTED_SPEED_10MBPS)
val |= SUPPORTED_10baseT_Full;
break;
case PHY_TYPE_KX4_10GB:
val |= SUPPORTED_Backplane;
if (if_speeds & BE_SUPPORTED_SPEED_1GBPS)
val |= SUPPORTED_1000baseKX_Full;
if (if_speeds & BE_SUPPORTED_SPEED_10GBPS)
val |= SUPPORTED_10000baseKX4_Full;
break;
case PHY_TYPE_KR2_20GB:
val |= SUPPORTED_Backplane;
if (if_speeds & BE_SUPPORTED_SPEED_10GBPS)
val |= SUPPORTED_10000baseKR_Full;
if (if_speeds & BE_SUPPORTED_SPEED_20GBPS)
val |= SUPPORTED_20000baseKR2_Full;
break;
case PHY_TYPE_KR_10GB:
val |= SUPPORTED_Backplane |
SUPPORTED_10000baseKR_Full;
break;
case PHY_TYPE_KR4_40GB:
val |= SUPPORTED_Backplane;
if (if_speeds & BE_SUPPORTED_SPEED_10GBPS)
val |= SUPPORTED_10000baseKR_Full;
if (if_speeds & BE_SUPPORTED_SPEED_40GBPS)
val |= SUPPORTED_40000baseKR4_Full;
break;
case PHY_TYPE_QSFP:
if (if_speeds & BE_SUPPORTED_SPEED_40GBPS) {
switch (adapter->phy.cable_type) {
case QSFP_PLUS_CR4_CABLE:
val |= SUPPORTED_40000baseCR4_Full;
2024-09-09 08:52:07 +00:00
break;
2024-09-09 08:57:42 +00:00
case QSFP_PLUS_LR4_CABLE:
val |= SUPPORTED_40000baseLR4_Full;
2024-09-09 08:52:07 +00:00
break;
2024-09-09 08:57:42 +00:00
default:
val |= SUPPORTED_40000baseSR4_Full;
2024-09-09 08:52:07 +00:00
break;
}
}
2024-09-09 08:57:42 +00:00
case PHY_TYPE_SFP_PLUS_10GB:
case PHY_TYPE_XFP_10GB:
case PHY_TYPE_SFP_1GB:
val |= SUPPORTED_FIBRE;
if (if_speeds & BE_SUPPORTED_SPEED_10GBPS)
val |= SUPPORTED_10000baseT_Full;
if (if_speeds & BE_SUPPORTED_SPEED_1GBPS)
val |= SUPPORTED_1000baseT_Full;
break;
case PHY_TYPE_BASET_10GB:
val |= SUPPORTED_TP;
if (if_speeds & BE_SUPPORTED_SPEED_10GBPS)
val |= SUPPORTED_10000baseT_Full;
if (if_speeds & BE_SUPPORTED_SPEED_1GBPS)
val |= SUPPORTED_1000baseT_Full;
if (if_speeds & BE_SUPPORTED_SPEED_100MBPS)
val |= SUPPORTED_100baseT_Full;
break;
default:
val |= SUPPORTED_TP;
}
return val;
}
bool be_pause_supported(struct be_adapter *adapter)
{
return (adapter->phy.interface_type == PHY_TYPE_SFP_PLUS_10GB ||
adapter->phy.interface_type == PHY_TYPE_XFP_10GB) ?
false : true;
}
static int be_get_settings(struct net_device *netdev, struct ethtool_cmd *ecmd)
{
struct be_adapter *adapter = netdev_priv(netdev);
u8 link_status;
u16 link_speed = 0;
int status;
u32 auto_speeds;
u32 fixed_speeds;
if (adapter->phy.link_speed < 0) {
status = be_cmd_link_status_query(adapter, &link_speed,
&link_status, 0);
if (!status)
be_link_status_update(adapter, link_status);
ethtool_cmd_speed_set(ecmd, link_speed);
2024-09-09 08:52:07 +00:00
2024-09-09 08:57:42 +00:00
status = be_cmd_get_phy_info(adapter);
2024-09-09 08:52:07 +00:00
if (!status) {
2024-09-09 08:57:42 +00:00
auto_speeds = adapter->phy.auto_speeds_supported;
fixed_speeds = adapter->phy.fixed_speeds_supported;
be_cmd_query_cable_type(adapter);
ecmd->supported =
convert_to_et_setting(adapter,
auto_speeds |
fixed_speeds);
ecmd->advertising =
convert_to_et_setting(adapter, auto_speeds);
ecmd->port = be_get_port_type(adapter);
if (adapter->phy.auto_speeds_supported) {
ecmd->supported |= SUPPORTED_Autoneg;
ecmd->autoneg = AUTONEG_ENABLE;
ecmd->advertising |= ADVERTISED_Autoneg;
2024-09-09 08:52:07 +00:00
}
2024-09-09 08:57:42 +00:00
ecmd->supported |= SUPPORTED_Pause;
if (be_pause_supported(adapter))
ecmd->advertising |= ADVERTISED_Pause;
switch (adapter->phy.interface_type) {
2024-09-09 08:52:07 +00:00
case PHY_TYPE_KR_10GB:
case PHY_TYPE_KX4_10GB:
2024-09-09 08:57:42 +00:00
ecmd->transceiver = XCVR_INTERNAL;
2024-09-09 08:52:07 +00:00
break;
default:
ecmd->transceiver = XCVR_EXTERNAL;
break;
}
2024-09-09 08:57:42 +00:00
} else {
ecmd->port = PORT_OTHER;
ecmd->autoneg = AUTONEG_DISABLE;
ecmd->transceiver = XCVR_DUMMY1;
2024-09-09 08:52:07 +00:00
}
/* Save for future use */
2024-09-09 08:57:42 +00:00
adapter->phy.link_speed = ethtool_cmd_speed(ecmd);
adapter->phy.port_type = ecmd->port;
adapter->phy.transceiver = ecmd->transceiver;
adapter->phy.autoneg = ecmd->autoneg;
adapter->phy.advertising = ecmd->advertising;
adapter->phy.supported = ecmd->supported;
2024-09-09 08:52:07 +00:00
} else {
2024-09-09 08:57:42 +00:00
ethtool_cmd_speed_set(ecmd, adapter->phy.link_speed);
ecmd->port = adapter->phy.port_type;
ecmd->transceiver = adapter->phy.transceiver;
ecmd->autoneg = adapter->phy.autoneg;
ecmd->advertising = adapter->phy.advertising;
ecmd->supported = adapter->phy.supported;
2024-09-09 08:52:07 +00:00
}
2024-09-09 08:57:42 +00:00
ecmd->duplex = netif_carrier_ok(netdev) ? DUPLEX_FULL : DUPLEX_UNKNOWN;
2024-09-09 08:52:07 +00:00
ecmd->phy_address = adapter->port_num;
return 0;
}
static void be_get_ringparam(struct net_device *netdev,
struct ethtool_ringparam *ring)
{
struct be_adapter *adapter = netdev_priv(netdev);
2024-09-09 08:57:42 +00:00
ring->rx_max_pending = adapter->rx_obj[0].q.len;
ring->rx_pending = adapter->rx_obj[0].q.len;
ring->tx_max_pending = adapter->tx_obj[0].q.len;
ring->tx_pending = adapter->tx_obj[0].q.len;
2024-09-09 08:52:07 +00:00
}
static void
be_get_pauseparam(struct net_device *netdev, struct ethtool_pauseparam *ecmd)
{
struct be_adapter *adapter = netdev_priv(netdev);
be_cmd_get_flow_control(adapter, &ecmd->tx_pause, &ecmd->rx_pause);
2024-09-09 08:57:42 +00:00
ecmd->autoneg = adapter->phy.fc_autoneg;
2024-09-09 08:52:07 +00:00
}
static int
be_set_pauseparam(struct net_device *netdev, struct ethtool_pauseparam *ecmd)
{
struct be_adapter *adapter = netdev_priv(netdev);
int status;
2024-09-09 08:57:42 +00:00
if (ecmd->autoneg != adapter->phy.fc_autoneg)
2024-09-09 08:52:07 +00:00
return -EINVAL;
adapter->tx_fc = ecmd->tx_pause;
adapter->rx_fc = ecmd->rx_pause;
status = be_cmd_set_flow_control(adapter,
2024-09-09 08:57:42 +00:00
adapter->tx_fc, adapter->rx_fc);
2024-09-09 08:52:07 +00:00
if (status)
2024-09-09 08:57:42 +00:00
dev_warn(&adapter->pdev->dev, "Pause param set failed\n");
2024-09-09 08:52:07 +00:00
2024-09-09 08:57:42 +00:00
return be_cmd_status(status);
2024-09-09 08:52:07 +00:00
}
2024-09-09 08:57:42 +00:00
static int be_set_phys_id(struct net_device *netdev,
enum ethtool_phys_id_state state)
2024-09-09 08:52:07 +00:00
{
struct be_adapter *adapter = netdev_priv(netdev);
switch (state) {
case ETHTOOL_ID_ACTIVE:
be_cmd_get_beacon_state(adapter, adapter->hba_port_num,
&adapter->beacon_state);
return 1; /* cycle on/off once per second */
case ETHTOOL_ID_ON:
be_cmd_set_beacon_state(adapter, adapter->hba_port_num, 0, 0,
BEACON_STATE_ENABLED);
break;
case ETHTOOL_ID_OFF:
be_cmd_set_beacon_state(adapter, adapter->hba_port_num, 0, 0,
BEACON_STATE_DISABLED);
break;
case ETHTOOL_ID_INACTIVE:
be_cmd_set_beacon_state(adapter, adapter->hba_port_num, 0, 0,
adapter->beacon_state);
}
return 0;
}
2024-09-09 08:57:42 +00:00
static int be_set_dump(struct net_device *netdev, struct ethtool_dump *dump)
{
struct be_adapter *adapter = netdev_priv(netdev);
struct device *dev = &adapter->pdev->dev;
int status;
2024-09-09 08:52:07 +00:00
2024-09-09 08:57:42 +00:00
if (!lancer_chip(adapter) ||
!check_privilege(adapter, MAX_PRIVILEGES))
return -EOPNOTSUPP;
switch (dump->flag) {
case LANCER_INITIATE_FW_DUMP:
status = lancer_initiate_dump(adapter);
if (!status)
dev_info(dev, "FW dump initiated successfully\n");
break;
case LANCER_DELETE_FW_DUMP:
status = lancer_delete_dump(adapter);
if (!status)
dev_info(dev, "FW dump deleted successfully\n");
break;
default:
dev_err(dev, "Invalid dump level: 0x%x\n", dump->flag);
return -EINVAL;
}
return status;
}
static void be_get_wol(struct net_device *netdev, struct ethtool_wolinfo *wol)
2024-09-09 08:52:07 +00:00
{
struct be_adapter *adapter = netdev_priv(netdev);
2024-09-09 08:57:42 +00:00
if (adapter->wol_cap & BE_WOL_CAP) {
2024-09-09 08:52:07 +00:00
wol->supported |= WAKE_MAGIC;
2024-09-09 08:57:42 +00:00
if (adapter->wol_en)
wol->wolopts |= WAKE_MAGIC;
} else {
2024-09-09 08:52:07 +00:00
wol->wolopts = 0;
2024-09-09 08:57:42 +00:00
}
2024-09-09 08:52:07 +00:00
memset(&wol->sopass, 0, sizeof(wol->sopass));
}
2024-09-09 08:57:42 +00:00
static int be_set_wol(struct net_device *netdev, struct ethtool_wolinfo *wol)
2024-09-09 08:52:07 +00:00
{
struct be_adapter *adapter = netdev_priv(netdev);
if (wol->wolopts & ~WAKE_MAGIC)
return -EOPNOTSUPP;
2024-09-09 08:57:42 +00:00
if (!(adapter->wol_cap & BE_WOL_CAP)) {
2024-09-09 08:52:07 +00:00
dev_warn(&adapter->pdev->dev, "WOL not supported\n");
return -EOPNOTSUPP;
}
if (wol->wolopts & WAKE_MAGIC)
2024-09-09 08:57:42 +00:00
adapter->wol_en = true;
2024-09-09 08:52:07 +00:00
else
2024-09-09 08:57:42 +00:00
adapter->wol_en = false;
2024-09-09 08:52:07 +00:00
return 0;
}
2024-09-09 08:57:42 +00:00
static int be_test_ddr_dma(struct be_adapter *adapter)
2024-09-09 08:52:07 +00:00
{
int ret, i;
struct be_dma_mem ddrdma_cmd;
static const u64 pattern[2] = {
0x5a5a5a5a5a5a5a5aULL, 0xa5a5a5a5a5a5a5a5ULL
};
ddrdma_cmd.size = sizeof(struct be_cmd_req_ddrdma_test);
2024-09-09 08:57:42 +00:00
ddrdma_cmd.va = dma_zalloc_coherent(&adapter->pdev->dev,
ddrdma_cmd.size, &ddrdma_cmd.dma,
GFP_KERNEL);
if (!ddrdma_cmd.va)
2024-09-09 08:52:07 +00:00
return -ENOMEM;
for (i = 0; i < 2; i++) {
ret = be_cmd_ddr_dma_test(adapter, pattern[i],
2024-09-09 08:57:42 +00:00
4096, &ddrdma_cmd);
2024-09-09 08:52:07 +00:00
if (ret != 0)
goto err;
}
err:
dma_free_coherent(&adapter->pdev->dev, ddrdma_cmd.size, ddrdma_cmd.va,
ddrdma_cmd.dma);
2024-09-09 08:57:42 +00:00
return be_cmd_status(ret);
2024-09-09 08:52:07 +00:00
}
static u64 be_loopback_test(struct be_adapter *adapter, u8 loopback_type,
2024-09-09 08:57:42 +00:00
u64 *status)
2024-09-09 08:52:07 +00:00
{
2024-09-09 08:57:42 +00:00
be_cmd_set_loopback(adapter, adapter->hba_port_num, loopback_type, 1);
2024-09-09 08:52:07 +00:00
*status = be_cmd_loopback_test(adapter, adapter->hba_port_num,
2024-09-09 08:57:42 +00:00
loopback_type, 1500, 2, 0xabc);
be_cmd_set_loopback(adapter, adapter->hba_port_num, BE_NO_LOOPBACK, 1);
2024-09-09 08:52:07 +00:00
return *status;
}
2024-09-09 08:57:42 +00:00
static void be_self_test(struct net_device *netdev, struct ethtool_test *test,
u64 *data)
2024-09-09 08:52:07 +00:00
{
struct be_adapter *adapter = netdev_priv(netdev);
2024-09-09 08:57:42 +00:00
int status;
u8 link_status = 0;
if (adapter->function_caps & BE_FUNCTION_CAPS_SUPER_NIC) {
dev_err(&adapter->pdev->dev, "Self test not supported\n");
test->flags |= ETH_TEST_FL_FAILED;
return;
}
2024-09-09 08:52:07 +00:00
memset(data, 0, sizeof(u64) * ETHTOOL_TESTS_NUM);
if (test->flags & ETH_TEST_FL_OFFLINE) {
2024-09-09 08:57:42 +00:00
if (be_loopback_test(adapter, BE_MAC_LOOPBACK, &data[0]) != 0)
2024-09-09 08:52:07 +00:00
test->flags |= ETH_TEST_FL_FAILED;
2024-09-09 08:57:42 +00:00
if (be_loopback_test(adapter, BE_PHY_LOOPBACK, &data[1]) != 0)
2024-09-09 08:52:07 +00:00
test->flags |= ETH_TEST_FL_FAILED;
2024-09-09 08:57:42 +00:00
if (test->flags & ETH_TEST_FL_EXTERNAL_LB) {
if (be_loopback_test(adapter, BE_ONE_PORT_EXT_LOOPBACK,
&data[2]) != 0)
test->flags |= ETH_TEST_FL_FAILED;
test->flags |= ETH_TEST_FL_EXTERNAL_LB_DONE;
2024-09-09 08:52:07 +00:00
}
}
2024-09-09 08:57:42 +00:00
if (!lancer_chip(adapter) && be_test_ddr_dma(adapter) != 0) {
2024-09-09 08:52:07 +00:00
data[3] = 1;
test->flags |= ETH_TEST_FL_FAILED;
}
2024-09-09 08:57:42 +00:00
status = be_cmd_link_status_query(adapter, NULL, &link_status, 0);
if (status) {
2024-09-09 08:52:07 +00:00
test->flags |= ETH_TEST_FL_FAILED;
data[4] = -1;
2024-09-09 08:57:42 +00:00
} else if (!link_status) {
2024-09-09 08:52:07 +00:00
test->flags |= ETH_TEST_FL_FAILED;
data[4] = 1;
}
}
2024-09-09 08:57:42 +00:00
static int be_do_flash(struct net_device *netdev, struct ethtool_flash *efl)
2024-09-09 08:52:07 +00:00
{
struct be_adapter *adapter = netdev_priv(netdev);
return be_load_fw(adapter, efl->data);
}
2024-09-09 08:57:42 +00:00
static int be_get_eeprom_len(struct net_device *netdev)
2024-09-09 08:52:07 +00:00
{
struct be_adapter *adapter = netdev_priv(netdev);
2024-09-09 08:57:42 +00:00
if (!check_privilege(adapter, MAX_PRIVILEGES))
return 0;
2024-09-09 08:52:07 +00:00
if (lancer_chip(adapter)) {
if (be_physfn(adapter))
return lancer_cmd_get_file_len(adapter,
2024-09-09 08:57:42 +00:00
LANCER_VPD_PF_FILE);
2024-09-09 08:52:07 +00:00
else
return lancer_cmd_get_file_len(adapter,
2024-09-09 08:57:42 +00:00
LANCER_VPD_VF_FILE);
2024-09-09 08:52:07 +00:00
} else {
return BE_READ_SEEPROM_LEN;
}
}
2024-09-09 08:57:42 +00:00
static int be_read_eeprom(struct net_device *netdev,
struct ethtool_eeprom *eeprom, uint8_t *data)
2024-09-09 08:52:07 +00:00
{
struct be_adapter *adapter = netdev_priv(netdev);
struct be_dma_mem eeprom_cmd;
struct be_cmd_resp_seeprom_read *resp;
int status;
if (!eeprom->len)
return -EINVAL;
if (lancer_chip(adapter)) {
if (be_physfn(adapter))
return lancer_cmd_read_file(adapter, LANCER_VPD_PF_FILE,
2024-09-09 08:57:42 +00:00
eeprom->len, data);
2024-09-09 08:52:07 +00:00
else
return lancer_cmd_read_file(adapter, LANCER_VPD_VF_FILE,
2024-09-09 08:57:42 +00:00
eeprom->len, data);
2024-09-09 08:52:07 +00:00
}
eeprom->magic = BE_VENDOR_ID | (adapter->pdev->device<<16);
memset(&eeprom_cmd, 0, sizeof(struct be_dma_mem));
eeprom_cmd.size = sizeof(struct be_cmd_req_seeprom_read);
2024-09-09 08:57:42 +00:00
eeprom_cmd.va = dma_zalloc_coherent(&adapter->pdev->dev,
eeprom_cmd.size, &eeprom_cmd.dma,
GFP_KERNEL);
2024-09-09 08:52:07 +00:00
2024-09-09 08:57:42 +00:00
if (!eeprom_cmd.va)
2024-09-09 08:52:07 +00:00
return -ENOMEM;
status = be_cmd_get_seeprom_data(adapter, &eeprom_cmd);
if (!status) {
resp = eeprom_cmd.va;
memcpy(data, resp->seeprom_data + eeprom->offset, eeprom->len);
}
dma_free_coherent(&adapter->pdev->dev, eeprom_cmd.size, eeprom_cmd.va,
eeprom_cmd.dma);
2024-09-09 08:57:42 +00:00
return be_cmd_status(status);
}
static u32 be_get_msg_level(struct net_device *netdev)
{
struct be_adapter *adapter = netdev_priv(netdev);
return adapter->msg_enable;
}
static void be_set_msg_level(struct net_device *netdev, u32 level)
{
struct be_adapter *adapter = netdev_priv(netdev);
if (adapter->msg_enable == level)
return;
if ((level & NETIF_MSG_HW) != (adapter->msg_enable & NETIF_MSG_HW))
if (BEx_chip(adapter))
be_cmd_set_fw_log_level(adapter, level & NETIF_MSG_HW ?
FW_LOG_LEVEL_DEFAULT :
FW_LOG_LEVEL_FATAL);
adapter->msg_enable = level;
}
static u64 be_get_rss_hash_opts(struct be_adapter *adapter, u64 flow_type)
{
u64 data = 0;
switch (flow_type) {
case TCP_V4_FLOW:
if (adapter->rss_info.rss_flags & RSS_ENABLE_IPV4)
data |= RXH_IP_DST | RXH_IP_SRC;
if (adapter->rss_info.rss_flags & RSS_ENABLE_TCP_IPV4)
data |= RXH_L4_B_0_1 | RXH_L4_B_2_3;
break;
case UDP_V4_FLOW:
if (adapter->rss_info.rss_flags & RSS_ENABLE_IPV4)
data |= RXH_IP_DST | RXH_IP_SRC;
if (adapter->rss_info.rss_flags & RSS_ENABLE_UDP_IPV4)
data |= RXH_L4_B_0_1 | RXH_L4_B_2_3;
break;
case TCP_V6_FLOW:
if (adapter->rss_info.rss_flags & RSS_ENABLE_IPV6)
data |= RXH_IP_DST | RXH_IP_SRC;
if (adapter->rss_info.rss_flags & RSS_ENABLE_TCP_IPV6)
data |= RXH_L4_B_0_1 | RXH_L4_B_2_3;
break;
case UDP_V6_FLOW:
if (adapter->rss_info.rss_flags & RSS_ENABLE_IPV6)
data |= RXH_IP_DST | RXH_IP_SRC;
if (adapter->rss_info.rss_flags & RSS_ENABLE_UDP_IPV6)
data |= RXH_L4_B_0_1 | RXH_L4_B_2_3;
break;
}
return data;
}
static int be_get_rxnfc(struct net_device *netdev, struct ethtool_rxnfc *cmd,
u32 *rule_locs)
{
struct be_adapter *adapter = netdev_priv(netdev);
if (!be_multi_rxq(adapter)) {
dev_info(&adapter->pdev->dev,
"ethtool::get_rxnfc: RX flow hashing is disabled\n");
return -EINVAL;
}
switch (cmd->cmd) {
case ETHTOOL_GRXFH:
cmd->data = be_get_rss_hash_opts(adapter, cmd->flow_type);
break;
case ETHTOOL_GRXRINGS:
cmd->data = adapter->num_rx_qs - 1;
break;
default:
return -EINVAL;
}
return 0;
}
static int be_set_rss_hash_opts(struct be_adapter *adapter,
struct ethtool_rxnfc *cmd)
{
struct be_rx_obj *rxo;
int status = 0, i, j;
u8 rsstable[128];
u32 rss_flags = adapter->rss_info.rss_flags;
if (cmd->data != L3_RSS_FLAGS &&
cmd->data != (L3_RSS_FLAGS | L4_RSS_FLAGS))
return -EINVAL;
switch (cmd->flow_type) {
case TCP_V4_FLOW:
if (cmd->data == L3_RSS_FLAGS)
rss_flags &= ~RSS_ENABLE_TCP_IPV4;
else if (cmd->data == (L3_RSS_FLAGS | L4_RSS_FLAGS))
rss_flags |= RSS_ENABLE_IPV4 |
RSS_ENABLE_TCP_IPV4;
break;
case TCP_V6_FLOW:
if (cmd->data == L3_RSS_FLAGS)
rss_flags &= ~RSS_ENABLE_TCP_IPV6;
else if (cmd->data == (L3_RSS_FLAGS | L4_RSS_FLAGS))
rss_flags |= RSS_ENABLE_IPV6 |
RSS_ENABLE_TCP_IPV6;
break;
case UDP_V4_FLOW:
if ((cmd->data == (L3_RSS_FLAGS | L4_RSS_FLAGS)) &&
BEx_chip(adapter))
return -EINVAL;
if (cmd->data == L3_RSS_FLAGS)
rss_flags &= ~RSS_ENABLE_UDP_IPV4;
else if (cmd->data == (L3_RSS_FLAGS | L4_RSS_FLAGS))
rss_flags |= RSS_ENABLE_IPV4 |
RSS_ENABLE_UDP_IPV4;
break;
case UDP_V6_FLOW:
if ((cmd->data == (L3_RSS_FLAGS | L4_RSS_FLAGS)) &&
BEx_chip(adapter))
return -EINVAL;
if (cmd->data == L3_RSS_FLAGS)
rss_flags &= ~RSS_ENABLE_UDP_IPV6;
else if (cmd->data == (L3_RSS_FLAGS | L4_RSS_FLAGS))
rss_flags |= RSS_ENABLE_IPV6 |
RSS_ENABLE_UDP_IPV6;
break;
default:
return -EINVAL;
}
if (rss_flags == adapter->rss_info.rss_flags)
return status;
if (be_multi_rxq(adapter)) {
for (j = 0; j < 128; j += adapter->num_rx_qs - 1) {
for_all_rss_queues(adapter, rxo, i) {
if ((j + i) >= 128)
break;
rsstable[j + i] = rxo->rss_id;
}
}
}
status = be_cmd_rss_config(adapter, adapter->rss_info.rsstable,
rss_flags, 128, adapter->rss_info.rss_hkey);
if (!status)
adapter->rss_info.rss_flags = rss_flags;
return be_cmd_status(status);
}
static int be_set_rxnfc(struct net_device *netdev, struct ethtool_rxnfc *cmd)
{
struct be_adapter *adapter = netdev_priv(netdev);
int status = 0;
if (!be_multi_rxq(adapter)) {
dev_err(&adapter->pdev->dev,
"ethtool::set_rxnfc: RX flow hashing is disabled\n");
return -EINVAL;
}
switch (cmd->cmd) {
case ETHTOOL_SRXFH:
status = be_set_rss_hash_opts(adapter, cmd);
break;
default:
return -EINVAL;
}
2024-09-09 08:52:07 +00:00
return status;
}
2024-09-09 08:57:42 +00:00
static void be_get_channels(struct net_device *netdev,
struct ethtool_channels *ch)
{
struct be_adapter *adapter = netdev_priv(netdev);
ch->combined_count = adapter->num_evt_qs;
ch->max_combined = be_max_qs(adapter);
}
static int be_set_channels(struct net_device *netdev,
struct ethtool_channels *ch)
{
struct be_adapter *adapter = netdev_priv(netdev);
int status;
if (ch->rx_count || ch->tx_count || ch->other_count ||
!ch->combined_count || ch->combined_count > be_max_qs(adapter))
return -EINVAL;
adapter->cfg_num_qs = ch->combined_count;
status = be_update_queues(adapter);
return be_cmd_status(status);
}
static u32 be_get_rxfh_indir_size(struct net_device *netdev)
{
return RSS_INDIR_TABLE_LEN;
}
static u32 be_get_rxfh_key_size(struct net_device *netdev)
{
return RSS_HASH_KEY_LEN;
}
static int be_get_rxfh(struct net_device *netdev, u32 *indir, u8 *hkey)
{
struct be_adapter *adapter = netdev_priv(netdev);
int i;
struct rss_info *rss = &adapter->rss_info;
if (indir) {
for (i = 0; i < RSS_INDIR_TABLE_LEN; i++)
indir[i] = rss->rss_queue[i];
}
if (hkey)
memcpy(hkey, rss->rss_hkey, RSS_HASH_KEY_LEN);
return 0;
}
static int be_set_rxfh(struct net_device *netdev, const u32 *indir,
const u8 *hkey)
{
int rc = 0, i, j;
struct be_adapter *adapter = netdev_priv(netdev);
u8 rsstable[RSS_INDIR_TABLE_LEN];
if (indir) {
struct be_rx_obj *rxo;
for (i = 0; i < RSS_INDIR_TABLE_LEN; i++) {
j = indir[i];
rxo = &adapter->rx_obj[j];
rsstable[i] = rxo->rss_id;
adapter->rss_info.rss_queue[i] = j;
}
} else {
memcpy(rsstable, adapter->rss_info.rsstable,
RSS_INDIR_TABLE_LEN);
}
if (!hkey)
hkey = adapter->rss_info.rss_hkey;
rc = be_cmd_rss_config(adapter, rsstable,
adapter->rss_info.rss_flags,
RSS_INDIR_TABLE_LEN, hkey);
if (rc) {
adapter->rss_info.rss_flags = RSS_ENABLE_NONE;
return -EIO;
}
memcpy(adapter->rss_info.rss_hkey, hkey, RSS_HASH_KEY_LEN);
memcpy(adapter->rss_info.rsstable, rsstable,
RSS_INDIR_TABLE_LEN);
return 0;
}
static int be_get_module_info(struct net_device *netdev,
struct ethtool_modinfo *modinfo)
{
struct be_adapter *adapter = netdev_priv(netdev);
u8 page_data[PAGE_DATA_LEN];
int status;
if (!check_privilege(adapter, MAX_PRIVILEGES))
return -EOPNOTSUPP;
status = be_cmd_read_port_transceiver_data(adapter, TR_PAGE_A0,
page_data);
if (!status) {
if (!page_data[SFP_PLUS_SFF_8472_COMP]) {
modinfo->type = ETH_MODULE_SFF_8079;
modinfo->eeprom_len = PAGE_DATA_LEN;
} else {
modinfo->type = ETH_MODULE_SFF_8472;
modinfo->eeprom_len = 2 * PAGE_DATA_LEN;
}
}
return be_cmd_status(status);
}
static int be_get_module_eeprom(struct net_device *netdev,
struct ethtool_eeprom *eeprom, u8 *data)
{
struct be_adapter *adapter = netdev_priv(netdev);
int status;
if (!check_privilege(adapter, MAX_PRIVILEGES))
return -EOPNOTSUPP;
status = be_cmd_read_port_transceiver_data(adapter, TR_PAGE_A0,
data);
if (status)
goto err;
if (eeprom->offset + eeprom->len > PAGE_DATA_LEN) {
status = be_cmd_read_port_transceiver_data(adapter,
TR_PAGE_A2,
data +
PAGE_DATA_LEN);
if (status)
goto err;
}
if (eeprom->offset)
memcpy(data, data + eeprom->offset, eeprom->len);
err:
return be_cmd_status(status);
}
2024-09-09 08:52:07 +00:00
const struct ethtool_ops be_ethtool_ops = {
.get_settings = be_get_settings,
.get_drvinfo = be_get_drvinfo,
.get_wol = be_get_wol,
.set_wol = be_set_wol,
.get_link = ethtool_op_get_link,
.get_eeprom_len = be_get_eeprom_len,
.get_eeprom = be_read_eeprom,
.get_coalesce = be_get_coalesce,
.set_coalesce = be_set_coalesce,
.get_ringparam = be_get_ringparam,
.get_pauseparam = be_get_pauseparam,
.set_pauseparam = be_set_pauseparam,
.get_strings = be_get_stat_strings,
.set_phys_id = be_set_phys_id,
2024-09-09 08:57:42 +00:00
.set_dump = be_set_dump,
.get_msglevel = be_get_msg_level,
.set_msglevel = be_set_msg_level,
2024-09-09 08:52:07 +00:00
.get_sset_count = be_get_sset_count,
.get_ethtool_stats = be_get_ethtool_stats,
.get_regs_len = be_get_reg_len,
.get_regs = be_get_regs,
.flash_device = be_do_flash,
.self_test = be_self_test,
2024-09-09 08:57:42 +00:00
.get_rxnfc = be_get_rxnfc,
.set_rxnfc = be_set_rxnfc,
.get_rxfh_indir_size = be_get_rxfh_indir_size,
.get_rxfh_key_size = be_get_rxfh_key_size,
.get_rxfh = be_get_rxfh,
.set_rxfh = be_set_rxfh,
.get_channels = be_get_channels,
.set_channels = be_set_channels,
.get_module_info = be_get_module_info,
.get_module_eeprom = be_get_module_eeprom
2024-09-09 08:52:07 +00:00
};