|  | #include <linux/kernel.h> | 
|  | #include <linux/init.h> | 
|  | #include <linux/types.h> | 
|  | #include <linux/module.h> | 
|  | #include <linux/list.h> | 
|  | #include <linux/pci.h> | 
|  | #include <linux/dma-mapping.h> | 
|  | #include <linux/pagemap.h> | 
|  | #include <linux/sched.h> | 
|  | #include <linux/slab.h> | 
|  | #include <linux/dmapool.h> | 
|  | #include <linux/mempool.h> | 
|  | #include <linux/spinlock.h> | 
|  | #include <linux/kthread.h> | 
|  | #include <linux/interrupt.h> | 
|  | #include <linux/errno.h> | 
|  | #include <linux/ioport.h> | 
|  | #include <linux/in.h> | 
|  | #include <linux/ip.h> | 
|  | #include <linux/ipv6.h> | 
|  | #include <net/ipv6.h> | 
|  | #include <linux/tcp.h> | 
|  | #include <linux/udp.h> | 
|  | #include <linux/if_arp.h> | 
|  | #include <linux/if_ether.h> | 
|  | #include <linux/netdevice.h> | 
|  | #include <linux/etherdevice.h> | 
|  | #include <linux/ethtool.h> | 
|  | #include <linux/skbuff.h> | 
|  | #include <linux/rtnetlink.h> | 
|  | #include <linux/if_vlan.h> | 
|  | #include <linux/delay.h> | 
|  | #include <linux/mm.h> | 
|  | #include <linux/vmalloc.h> | 
|  |  | 
|  | #include <linux/version.h> | 
|  |  | 
|  | #include "qlge.h" | 
|  |  | 
|  | static int ql_update_ring_coalescing(struct ql_adapter *qdev) | 
|  | { | 
|  | int i, status = 0; | 
|  | struct rx_ring *rx_ring; | 
|  | struct cqicb *cqicb; | 
|  |  | 
|  | if (!netif_running(qdev->ndev)) | 
|  | return status; | 
|  |  | 
|  | spin_lock(&qdev->hw_lock); | 
|  | /* Skip the default queue, and update the outbound handler | 
|  | * queues if they changed. | 
|  | */ | 
|  | cqicb = (struct cqicb *)&qdev->rx_ring[1]; | 
|  | if (le16_to_cpu(cqicb->irq_delay) != qdev->tx_coalesce_usecs || | 
|  | le16_to_cpu(cqicb->pkt_delay) != qdev->tx_max_coalesced_frames) { | 
|  | for (i = 1; i < qdev->rss_ring_first_cq_id; i++, rx_ring++) { | 
|  | rx_ring = &qdev->rx_ring[i]; | 
|  | cqicb = (struct cqicb *)rx_ring; | 
|  | cqicb->irq_delay = cpu_to_le16(qdev->tx_coalesce_usecs); | 
|  | cqicb->pkt_delay = | 
|  | cpu_to_le16(qdev->tx_max_coalesced_frames); | 
|  | cqicb->flags = FLAGS_LI; | 
|  | status = ql_write_cfg(qdev, cqicb, sizeof(cqicb), | 
|  | CFG_LCQ, rx_ring->cq_id); | 
|  | if (status) { | 
|  | QPRINTK(qdev, IFUP, ERR, | 
|  | "Failed to load CQICB.\n"); | 
|  | goto exit; | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | /* Update the inbound (RSS) handler queues if they changed. */ | 
|  | cqicb = (struct cqicb *)&qdev->rx_ring[qdev->rss_ring_first_cq_id]; | 
|  | if (le16_to_cpu(cqicb->irq_delay) != qdev->rx_coalesce_usecs || | 
|  | le16_to_cpu(cqicb->pkt_delay) != qdev->rx_max_coalesced_frames) { | 
|  | for (i = qdev->rss_ring_first_cq_id; | 
|  | i <= qdev->rss_ring_first_cq_id + qdev->rss_ring_count; | 
|  | i++) { | 
|  | rx_ring = &qdev->rx_ring[i]; | 
|  | cqicb = (struct cqicb *)rx_ring; | 
|  | cqicb->irq_delay = cpu_to_le16(qdev->rx_coalesce_usecs); | 
|  | cqicb->pkt_delay = | 
|  | cpu_to_le16(qdev->rx_max_coalesced_frames); | 
|  | cqicb->flags = FLAGS_LI; | 
|  | status = ql_write_cfg(qdev, cqicb, sizeof(cqicb), | 
|  | CFG_LCQ, rx_ring->cq_id); | 
|  | if (status) { | 
|  | QPRINTK(qdev, IFUP, ERR, | 
|  | "Failed to load CQICB.\n"); | 
|  | goto exit; | 
|  | } | 
|  | } | 
|  | } | 
|  | exit: | 
|  | spin_unlock(&qdev->hw_lock); | 
|  | return status; | 
|  | } | 
|  |  | 
|  | static void ql_update_stats(struct ql_adapter *qdev) | 
|  | { | 
|  | u32 i; | 
|  | u64 data; | 
|  | u64 *iter = &qdev->nic_stats.tx_pkts; | 
|  |  | 
|  | spin_lock(&qdev->stats_lock); | 
|  | if (ql_sem_spinlock(qdev, qdev->xg_sem_mask)) { | 
|  | QPRINTK(qdev, DRV, ERR, | 
|  | "Couldn't get xgmac sem.\n"); | 
|  | goto quit; | 
|  | } | 
|  | /* | 
|  | * Get TX statistics. | 
|  | */ | 
|  | for (i = 0x200; i < 0x280; i += 8) { | 
|  | if (ql_read_xgmac_reg64(qdev, i, &data)) { | 
|  | QPRINTK(qdev, DRV, ERR, | 
|  | "Error reading status register 0x%.04x.\n", i); | 
|  | goto end; | 
|  | } else | 
|  | *iter = data; | 
|  | iter++; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Get RX statistics. | 
|  | */ | 
|  | for (i = 0x300; i < 0x3d0; i += 8) { | 
|  | if (ql_read_xgmac_reg64(qdev, i, &data)) { | 
|  | QPRINTK(qdev, DRV, ERR, | 
|  | "Error reading status register 0x%.04x.\n", i); | 
|  | goto end; | 
|  | } else | 
|  | *iter = data; | 
|  | iter++; | 
|  | } | 
|  |  | 
|  | end: | 
|  | ql_sem_unlock(qdev, qdev->xg_sem_mask); | 
|  | quit: | 
|  | spin_unlock(&qdev->stats_lock); | 
|  |  | 
|  | QL_DUMP_STAT(qdev); | 
|  |  | 
|  | return; | 
|  | } | 
|  |  | 
|  | static char ql_stats_str_arr[][ETH_GSTRING_LEN] = { | 
|  | {"tx_pkts"}, | 
|  | {"tx_bytes"}, | 
|  | {"tx_mcast_pkts"}, | 
|  | {"tx_bcast_pkts"}, | 
|  | {"tx_ucast_pkts"}, | 
|  | {"tx_ctl_pkts"}, | 
|  | {"tx_pause_pkts"}, | 
|  | {"tx_64_pkts"}, | 
|  | {"tx_65_to_127_pkts"}, | 
|  | {"tx_128_to_255_pkts"}, | 
|  | {"tx_256_511_pkts"}, | 
|  | {"tx_512_to_1023_pkts"}, | 
|  | {"tx_1024_to_1518_pkts"}, | 
|  | {"tx_1519_to_max_pkts"}, | 
|  | {"tx_undersize_pkts"}, | 
|  | {"tx_oversize_pkts"}, | 
|  | {"rx_bytes"}, | 
|  | {"rx_bytes_ok"}, | 
|  | {"rx_pkts"}, | 
|  | {"rx_pkts_ok"}, | 
|  | {"rx_bcast_pkts"}, | 
|  | {"rx_mcast_pkts"}, | 
|  | {"rx_ucast_pkts"}, | 
|  | {"rx_undersize_pkts"}, | 
|  | {"rx_oversize_pkts"}, | 
|  | {"rx_jabber_pkts"}, | 
|  | {"rx_undersize_fcerr_pkts"}, | 
|  | {"rx_drop_events"}, | 
|  | {"rx_fcerr_pkts"}, | 
|  | {"rx_align_err"}, | 
|  | {"rx_symbol_err"}, | 
|  | {"rx_mac_err"}, | 
|  | {"rx_ctl_pkts"}, | 
|  | {"rx_pause_pkts"}, | 
|  | {"rx_64_pkts"}, | 
|  | {"rx_65_to_127_pkts"}, | 
|  | {"rx_128_255_pkts"}, | 
|  | {"rx_256_511_pkts"}, | 
|  | {"rx_512_to_1023_pkts"}, | 
|  | {"rx_1024_to_1518_pkts"}, | 
|  | {"rx_1519_to_max_pkts"}, | 
|  | {"rx_len_err_pkts"}, | 
|  | }; | 
|  |  | 
|  | static void ql_get_strings(struct net_device *dev, u32 stringset, u8 *buf) | 
|  | { | 
|  | switch (stringset) { | 
|  | case ETH_SS_STATS: | 
|  | memcpy(buf, ql_stats_str_arr, sizeof(ql_stats_str_arr)); | 
|  | break; | 
|  | } | 
|  | } | 
|  |  | 
|  | static int ql_get_sset_count(struct net_device *dev, int sset) | 
|  | { | 
|  | switch (sset) { | 
|  | case ETH_SS_STATS: | 
|  | return ARRAY_SIZE(ql_stats_str_arr); | 
|  | default: | 
|  | return -EOPNOTSUPP; | 
|  | } | 
|  | } | 
|  |  | 
|  | static void | 
|  | ql_get_ethtool_stats(struct net_device *ndev, | 
|  | struct ethtool_stats *stats, u64 *data) | 
|  | { | 
|  | struct ql_adapter *qdev = netdev_priv(ndev); | 
|  | struct nic_stats *s = &qdev->nic_stats; | 
|  |  | 
|  | ql_update_stats(qdev); | 
|  |  | 
|  | *data++ = s->tx_pkts; | 
|  | *data++ = s->tx_bytes; | 
|  | *data++ = s->tx_mcast_pkts; | 
|  | *data++ = s->tx_bcast_pkts; | 
|  | *data++ = s->tx_ucast_pkts; | 
|  | *data++ = s->tx_ctl_pkts; | 
|  | *data++ = s->tx_pause_pkts; | 
|  | *data++ = s->tx_64_pkt; | 
|  | *data++ = s->tx_65_to_127_pkt; | 
|  | *data++ = s->tx_128_to_255_pkt; | 
|  | *data++ = s->tx_256_511_pkt; | 
|  | *data++ = s->tx_512_to_1023_pkt; | 
|  | *data++ = s->tx_1024_to_1518_pkt; | 
|  | *data++ = s->tx_1519_to_max_pkt; | 
|  | *data++ = s->tx_undersize_pkt; | 
|  | *data++ = s->tx_oversize_pkt; | 
|  | *data++ = s->rx_bytes; | 
|  | *data++ = s->rx_bytes_ok; | 
|  | *data++ = s->rx_pkts; | 
|  | *data++ = s->rx_pkts_ok; | 
|  | *data++ = s->rx_bcast_pkts; | 
|  | *data++ = s->rx_mcast_pkts; | 
|  | *data++ = s->rx_ucast_pkts; | 
|  | *data++ = s->rx_undersize_pkts; | 
|  | *data++ = s->rx_oversize_pkts; | 
|  | *data++ = s->rx_jabber_pkts; | 
|  | *data++ = s->rx_undersize_fcerr_pkts; | 
|  | *data++ = s->rx_drop_events; | 
|  | *data++ = s->rx_fcerr_pkts; | 
|  | *data++ = s->rx_align_err; | 
|  | *data++ = s->rx_symbol_err; | 
|  | *data++ = s->rx_mac_err; | 
|  | *data++ = s->rx_ctl_pkts; | 
|  | *data++ = s->rx_pause_pkts; | 
|  | *data++ = s->rx_64_pkts; | 
|  | *data++ = s->rx_65_to_127_pkts; | 
|  | *data++ = s->rx_128_255_pkts; | 
|  | *data++ = s->rx_256_511_pkts; | 
|  | *data++ = s->rx_512_to_1023_pkts; | 
|  | *data++ = s->rx_1024_to_1518_pkts; | 
|  | *data++ = s->rx_1519_to_max_pkts; | 
|  | *data++ = s->rx_len_err_pkts; | 
|  | } | 
|  |  | 
|  | static int ql_get_settings(struct net_device *ndev, | 
|  | struct ethtool_cmd *ecmd) | 
|  | { | 
|  | struct ql_adapter *qdev = netdev_priv(ndev); | 
|  |  | 
|  | ecmd->supported = SUPPORTED_10000baseT_Full; | 
|  | ecmd->advertising = ADVERTISED_10000baseT_Full; | 
|  | ecmd->autoneg = AUTONEG_ENABLE; | 
|  | ecmd->transceiver = XCVR_EXTERNAL; | 
|  | if ((qdev->link_status & STS_LINK_TYPE_MASK) == | 
|  | STS_LINK_TYPE_10GBASET) { | 
|  | ecmd->supported |= (SUPPORTED_TP | SUPPORTED_Autoneg); | 
|  | ecmd->advertising |= (ADVERTISED_TP | ADVERTISED_Autoneg); | 
|  | ecmd->port = PORT_TP; | 
|  | } else { | 
|  | ecmd->supported |= SUPPORTED_FIBRE; | 
|  | ecmd->advertising |= ADVERTISED_FIBRE; | 
|  | ecmd->port = PORT_FIBRE; | 
|  | } | 
|  |  | 
|  | ecmd->speed = SPEED_10000; | 
|  | ecmd->duplex = DUPLEX_FULL; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static void ql_get_drvinfo(struct net_device *ndev, | 
|  | struct ethtool_drvinfo *drvinfo) | 
|  | { | 
|  | struct ql_adapter *qdev = netdev_priv(ndev); | 
|  | strncpy(drvinfo->driver, qlge_driver_name, 32); | 
|  | strncpy(drvinfo->version, qlge_driver_version, 32); | 
|  | strncpy(drvinfo->fw_version, "N/A", 32); | 
|  | strncpy(drvinfo->bus_info, pci_name(qdev->pdev), 32); | 
|  | drvinfo->n_stats = 0; | 
|  | drvinfo->testinfo_len = 0; | 
|  | drvinfo->regdump_len = 0; | 
|  | drvinfo->eedump_len = 0; | 
|  | } | 
|  |  | 
|  | static int ql_get_coalesce(struct net_device *dev, struct ethtool_coalesce *c) | 
|  | { | 
|  | struct ql_adapter *qdev = netdev_priv(dev); | 
|  |  | 
|  | c->rx_coalesce_usecs = qdev->rx_coalesce_usecs; | 
|  | c->tx_coalesce_usecs = qdev->tx_coalesce_usecs; | 
|  |  | 
|  | /* This chip coalesces as follows: | 
|  | * If a packet arrives, hold off interrupts until | 
|  | * cqicb->int_delay expires, but if no other packets arrive don't | 
|  | * wait longer than cqicb->pkt_int_delay. But ethtool doesn't use a | 
|  | * timer to coalesce on a frame basis.  So, we have to take ethtool's | 
|  | * max_coalesced_frames value and convert it to a delay in microseconds. | 
|  | * We do this by using a basic thoughput of 1,000,000 frames per | 
|  | * second @ (1024 bytes).  This means one frame per usec. So it's a | 
|  | * simple one to one ratio. | 
|  | */ | 
|  | c->rx_max_coalesced_frames = qdev->rx_max_coalesced_frames; | 
|  | c->tx_max_coalesced_frames = qdev->tx_max_coalesced_frames; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int ql_set_coalesce(struct net_device *ndev, struct ethtool_coalesce *c) | 
|  | { | 
|  | struct ql_adapter *qdev = netdev_priv(ndev); | 
|  |  | 
|  | /* Validate user parameters. */ | 
|  | if (c->rx_coalesce_usecs > qdev->rx_ring_size / 2) | 
|  | return -EINVAL; | 
|  | /* Don't wait more than 10 usec. */ | 
|  | if (c->rx_max_coalesced_frames > MAX_INTER_FRAME_WAIT) | 
|  | return -EINVAL; | 
|  | if (c->tx_coalesce_usecs > qdev->tx_ring_size / 2) | 
|  | return -EINVAL; | 
|  | if (c->tx_max_coalesced_frames > MAX_INTER_FRAME_WAIT) | 
|  | return -EINVAL; | 
|  |  | 
|  | /* Verify a change took place before updating the hardware. */ | 
|  | if (qdev->rx_coalesce_usecs == c->rx_coalesce_usecs && | 
|  | qdev->tx_coalesce_usecs == c->tx_coalesce_usecs && | 
|  | qdev->rx_max_coalesced_frames == c->rx_max_coalesced_frames && | 
|  | qdev->tx_max_coalesced_frames == c->tx_max_coalesced_frames) | 
|  | return 0; | 
|  |  | 
|  | qdev->rx_coalesce_usecs = c->rx_coalesce_usecs; | 
|  | qdev->tx_coalesce_usecs = c->tx_coalesce_usecs; | 
|  | qdev->rx_max_coalesced_frames = c->rx_max_coalesced_frames; | 
|  | qdev->tx_max_coalesced_frames = c->tx_max_coalesced_frames; | 
|  |  | 
|  | return ql_update_ring_coalescing(qdev); | 
|  | } | 
|  |  | 
|  | static u32 ql_get_rx_csum(struct net_device *netdev) | 
|  | { | 
|  | struct ql_adapter *qdev = netdev_priv(netdev); | 
|  | return qdev->rx_csum; | 
|  | } | 
|  |  | 
|  | static int ql_set_rx_csum(struct net_device *netdev, uint32_t data) | 
|  | { | 
|  | struct ql_adapter *qdev = netdev_priv(netdev); | 
|  | qdev->rx_csum = data; | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int ql_set_tso(struct net_device *ndev, uint32_t data) | 
|  | { | 
|  |  | 
|  | if (data) { | 
|  | ndev->features |= NETIF_F_TSO; | 
|  | ndev->features |= NETIF_F_TSO6; | 
|  | } else { | 
|  | ndev->features &= ~NETIF_F_TSO; | 
|  | ndev->features &= ~NETIF_F_TSO6; | 
|  | } | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static u32 ql_get_msglevel(struct net_device *ndev) | 
|  | { | 
|  | struct ql_adapter *qdev = netdev_priv(ndev); | 
|  | return qdev->msg_enable; | 
|  | } | 
|  |  | 
|  | static void ql_set_msglevel(struct net_device *ndev, u32 value) | 
|  | { | 
|  | struct ql_adapter *qdev = netdev_priv(ndev); | 
|  | qdev->msg_enable = value; | 
|  | } | 
|  |  | 
|  | const struct ethtool_ops qlge_ethtool_ops = { | 
|  | .get_settings = ql_get_settings, | 
|  | .get_drvinfo = ql_get_drvinfo, | 
|  | .get_msglevel = ql_get_msglevel, | 
|  | .set_msglevel = ql_set_msglevel, | 
|  | .get_link = ethtool_op_get_link, | 
|  | .get_rx_csum = ql_get_rx_csum, | 
|  | .set_rx_csum = ql_set_rx_csum, | 
|  | .get_tx_csum = ethtool_op_get_tx_csum, | 
|  | .get_sg = ethtool_op_get_sg, | 
|  | .set_sg = ethtool_op_set_sg, | 
|  | .get_tso = ethtool_op_get_tso, | 
|  | .set_tso = ql_set_tso, | 
|  | .get_coalesce = ql_get_coalesce, | 
|  | .set_coalesce = ql_set_coalesce, | 
|  | .get_sset_count = ql_get_sset_count, | 
|  | .get_strings = ql_get_strings, | 
|  | .get_ethtool_stats = ql_get_ethtool_stats, | 
|  | }; | 
|  |  |