[EHEA]: Use LRO.
authorJan-Bernd Themann <themann@de.ibm.com>
Tue, 25 Sep 2007 23:16:34 +0000 (16:16 -0700)
committerDavid S. Miller <davem@sunset.davemloft.net>
Wed, 10 Oct 2007 23:47:47 +0000 (16:47 -0700)
Signed-off-by: Jan-Bernd Themann <themann@de.ibm.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
drivers/net/Kconfig
drivers/net/ehea/ehea.h
drivers/net/ehea/ehea_ethtool.c
drivers/net/ehea/ehea_main.c

index e0a9791..61dedfe 100644 (file)
@@ -2507,6 +2507,7 @@ config CHELSIO_T3
 config EHEA
        tristate "eHEA Ethernet support"
        depends on IBMEBUS
+       select INET_LRO
        ---help---
          This driver supports the IBM pSeries eHEA ethernet adapter.
 
index a154681..c0cbd94 100644 (file)
@@ -33,6 +33,7 @@
 #include <linux/ethtool.h>
 #include <linux/vmalloc.h>
 #include <linux/if_vlan.h>
+#include <linux/inet_lro.h>
 
 #include <asm/ibmebus.h>
 #include <asm/abs_addr.h>
@@ -58,6 +59,7 @@
 
 #define EHEA_SMALL_QUEUES
 #define EHEA_NUM_TX_QP 1
+#define EHEA_LRO_MAX_AGGR 64
 
 #ifdef EHEA_SMALL_QUEUES
 #define EHEA_MAX_CQE_COUNT      1023
@@ -84,6 +86,8 @@
 #define EHEA_RQ2_PKT_SIZE       1522
 #define EHEA_L_PKT_SIZE         256    /* low latency */
 
+#define MAX_LRO_DESCRIPTORS 8
+
 /* Send completion signaling */
 
 /* Protection Domain Identifier */
@@ -376,6 +380,8 @@ struct ehea_port_res {
        u64 tx_packets;
        u64 rx_packets;
        u32 poll_counter;
+       struct net_lro_mgr lro_mgr;
+       struct net_lro_desc lro_desc[MAX_LRO_DESCRIPTORS];
 };
 
 
@@ -429,6 +435,7 @@ struct ehea_port {
        u32 msg_enable;
        u32 sig_comp_iv;
        u32 state;
+       u32 lro_max_aggr;
        u8 phy_link;
        u8 full_duplex;
        u8 autoneg;
index decec8c..29ef7a9 100644 (file)
@@ -183,6 +183,9 @@ static char ehea_ethtool_stats_keys[][ETH_GSTRING_LEN] = {
        {"PR5 free_swqes"},
        {"PR6 free_swqes"},
        {"PR7 free_swqes"},
+       {"LRO aggregated"},
+       {"LRO flushed"},
+       {"LRO no_desc"},
 };
 
 static void ehea_get_strings(struct net_device *dev, u32 stringset, u8 *data)
@@ -239,6 +242,18 @@ static void ehea_get_ethtool_stats(struct net_device *dev,
        for (k = 0; k < 8; k++)
                data[i++] = atomic_read(&port->port_res[k].swqe_avail);
 
+       for (k = 0, tmp = 0; k < EHEA_MAX_PORT_RES; k++)
+               tmp |= port->port_res[k].lro_mgr.stats.aggregated;
+       data[i++] = tmp;
+
+       for (k = 0, tmp = 0; k < EHEA_MAX_PORT_RES; k++)
+               tmp |= port->port_res[k].lro_mgr.stats.flushed;
+       data[i++] = tmp;
+
+       for (k = 0, tmp = 0; k < EHEA_MAX_PORT_RES; k++)
+               tmp |= port->port_res[k].lro_mgr.stats.no_desc;
+       data[i++] = tmp;
+
 }
 
 const struct ethtool_ops ehea_ethtool_ops = {
index 5ebd545..b8e0039 100644 (file)
@@ -52,6 +52,8 @@ static int rq2_entries = EHEA_DEF_ENTRIES_RQ2;
 static int rq3_entries = EHEA_DEF_ENTRIES_RQ3;
 static int sq_entries = EHEA_DEF_ENTRIES_SQ;
 static int use_mcs = 0;
+static int use_lro = 0;
+static int lro_max_aggr = EHEA_LRO_MAX_AGGR;
 static int num_tx_qps = EHEA_NUM_TX_QP;
 static int prop_carrier_state = 0;
 
@@ -62,6 +64,8 @@ module_param(rq3_entries, int, 0);
 module_param(sq_entries, int, 0);
 module_param(prop_carrier_state, int, 0);
 module_param(use_mcs, int, 0);
+module_param(use_lro, int, 0);
+module_param(lro_max_aggr, int, 0);
 module_param(num_tx_qps, int, 0);
 
 MODULE_PARM_DESC(num_tx_qps, "Number of TX-QPS");
@@ -82,6 +86,11 @@ MODULE_PARM_DESC(sq_entries, " Number of entries for the Send Queue  "
                 __MODULE_STRING(EHEA_DEF_ENTRIES_SQ) ")");
 MODULE_PARM_DESC(use_mcs, " 0:NAPI, 1:Multiple receive queues, Default = 0 ");
 
+MODULE_PARM_DESC(lro_max_aggr, " LRO: Max packets to be aggregated. Default = "
+                __MODULE_STRING(EHEA_LRO_MAX_AGGR));
+MODULE_PARM_DESC(use_lro, " Large Receive Offload, 1: enable, 0: disable, "
+                "Default = 0");
+
 static int port_name_cnt = 0;
 static LIST_HEAD(adapter_list);
 u64 ehea_driver_flags = 0;
@@ -393,6 +402,60 @@ static int ehea_treat_poll_error(struct ehea_port_res *pr, int rq,
        return 0;
 }
 
+static int get_skb_hdr(struct sk_buff *skb, void **iphdr,
+                      void **tcph, u64 *hdr_flags, void *priv)
+{
+       struct ehea_cqe *cqe = priv;
+       unsigned int ip_len;
+       struct iphdr *iph;
+
+       /* non tcp/udp packets */
+       if (!cqe->header_length)
+               return -1;
+
+       /* non tcp packet */
+       skb_reset_network_header(skb);
+       iph = ip_hdr(skb);
+       if (iph->protocol != IPPROTO_TCP)
+               return -1;
+
+       ip_len = ip_hdrlen(skb);
+       skb_set_transport_header(skb, ip_len);
+       *tcph = tcp_hdr(skb);
+
+       /* check if ip header and tcp header are complete */
+       if (iph->tot_len < ip_len + tcp_hdrlen(skb))
+               return -1;
+
+       *hdr_flags = LRO_IPV4 | LRO_TCP;
+       *iphdr = iph;
+
+       return 0;
+}
+
+static void ehea_proc_skb(struct ehea_port_res *pr, struct ehea_cqe *cqe,
+                         struct sk_buff *skb)
+{
+       int vlan_extracted = (cqe->status & EHEA_CQE_VLAN_TAG_XTRACT)
+               && pr->port->vgrp;
+
+       if (use_lro) {
+               if (vlan_extracted)
+                       lro_vlan_hwaccel_receive_skb(&pr->lro_mgr, skb,
+                                                    pr->port->vgrp,
+                                                    cqe->vlan_tag,
+                                                    cqe);
+               else
+                       lro_receive_skb(&pr->lro_mgr, skb, cqe);
+       } else {
+               if (vlan_extracted)
+                       vlan_hwaccel_receive_skb(skb, pr->port->vgrp,
+                                                cqe->vlan_tag);
+               else
+                       netif_receive_skb(skb);
+       }
+}
+
 static int ehea_proc_rwqes(struct net_device *dev,
                           struct ehea_port_res *pr,
                           int budget)
@@ -462,13 +525,7 @@ static int ehea_proc_rwqes(struct net_device *dev,
                                processed_rq3++;
                        }
 
-                       if ((cqe->status & EHEA_CQE_VLAN_TAG_XTRACT)
-                           && port->vgrp)
-                               vlan_hwaccel_receive_skb(skb, port->vgrp,
-                                                        cqe->vlan_tag);
-                       else
-                               netif_receive_skb(skb);
-
+                       ehea_proc_skb(pr, cqe, skb);
                        dev->last_rx = jiffies;
                } else {
                        pr->p_stats.poll_receive_errors++;
@@ -480,6 +537,8 @@ static int ehea_proc_rwqes(struct net_device *dev,
                }
                cqe = ehea_poll_rq1(qp, &wqe_index);
        }
+       if (use_lro)
+               lro_flush_all(&pr->lro_mgr);
 
        pr->rx_packets += processed;
 
@@ -1231,6 +1290,15 @@ static int ehea_init_port_res(struct ehea_port *port, struct ehea_port_res *pr,
 
        netif_napi_add(pr->port->netdev, &pr->napi, ehea_poll, 64);
 
+       pr->lro_mgr.max_aggr = pr->port->lro_max_aggr;
+       pr->lro_mgr.max_desc = MAX_LRO_DESCRIPTORS;
+       pr->lro_mgr.lro_arr = pr->lro_desc;
+       pr->lro_mgr.get_skb_header = get_skb_hdr;
+       pr->lro_mgr.features = LRO_F_NAPI | LRO_F_EXTRACT_VLAN_ID;
+       pr->lro_mgr.dev = port->netdev;
+       pr->lro_mgr.ip_summed = CHECKSUM_UNNECESSARY;
+       pr->lro_mgr.ip_summed_aggr = CHECKSUM_UNNECESSARY;
+
        ret = 0;
        goto out;
 
@@ -2682,6 +2750,8 @@ struct ehea_port *ehea_setup_single_port(struct ehea_adapter *adapter,
                goto out_dereg_bc;
        }
 
+       port->lro_max_aggr = lro_max_aggr;
+
        ret = ehea_get_jumboframe_status(port, &jumbo);
        if (ret)
                ehea_error("failed determining jumbo frame status for %s",