]> err.no Git - linux-2.6/commitdiff
ehea: removing unused functionality
authorJan-Bernd Themann <ossthema@de.ibm.com>
Fri, 23 Mar 2007 16:18:53 +0000 (17:18 +0100)
committerJeff Garzik <jeff@garzik.org>
Sat, 28 Apr 2007 15:01:02 +0000 (11:01 -0400)
This patch includes:
- removal of unused fields in structs
- ethtool statistics cleanup
- removes unsed functionality from send path

Signed-off-by: Jan-Bernd Themann <themann@de.ibm.com>
Signed-off-by: Jeff Garzik <jeff@garzik.org>
drivers/net/ehea/ehea.h
drivers/net/ehea/ehea_ethtool.c
drivers/net/ehea/ehea_main.c
drivers/net/ehea/ehea_qmr.h

index f8899339baa0d0516f443bd8bf62cec5851f3b86..1405d0b0b7e7f9998b56254e11f4f8e2181ab2fd 100644 (file)
@@ -39,7 +39,7 @@
 #include <asm/io.h>
 
 #define DRV_NAME       "ehea"
-#define DRV_VERSION    "EHEA_0054"
+#define DRV_VERSION    "EHEA_0055"
 
 #define EHEA_MSG_DEFAULT (NETIF_MSG_LINK | NETIF_MSG_TIMER \
        | NETIF_MSG_RX_ERR | NETIF_MSG_TX_ERR)
@@ -79,7 +79,6 @@
 #define EHEA_L_PKT_SIZE         256    /* low latency */
 
 /* Send completion signaling */
-#define EHEA_SIG_IV_LONG           1
 
 /* Protection Domain Identifier */
 #define EHEA_PD_ID        0xaabcdeff
 #define EHEA_CACHE_LINE          128
 
 /* Memory Regions */
-#define EHEA_MR_MAX_TX_PAGES   20
-#define EHEA_MR_TX_DATA_PN      3
 #define EHEA_MR_ACC_CTRL       0x00800000
-#define EHEA_RWQES_PER_MR_RQ2  10
-#define EHEA_RWQES_PER_MR_RQ3  10
 
 #define EHEA_WATCH_DOG_TIMEOUT 10*HZ
 
@@ -318,17 +313,12 @@ struct ehea_mr {
 /*
  * Port state information
  */
-struct port_state {
-       int poll_max_processed;
+struct port_stats {
        int poll_receive_errors;
-       int ehea_poll;
        int queue_stopped;
-       int min_swqe_avail;
-       u64 sqc_stop_sum;
-       int pkt_send;
-       int pkt_xmit;
-       int send_tasklet;
-       int nwqe;
+       int err_tcp_cksum;
+       int err_ip_cksum;
+       int err_frame_crc;
 };
 
 #define EHEA_IRQ_NAME_SIZE 20
@@ -347,6 +337,7 @@ struct ehea_q_skb_arr {
  * Port resources
  */
 struct ehea_port_res {
+       struct port_stats p_stats;
        struct ehea_mr send_mr;         /* send memory region */
        struct ehea_mr recv_mr;         /* receive memory region */
        spinlock_t xmit_lock;
@@ -358,7 +349,6 @@ struct ehea_port_res {
        struct ehea_cq *recv_cq;
        struct ehea_eq *eq;
        struct net_device *d_netdev;
-       spinlock_t send_lock;
        struct ehea_q_skb_arr rq1_skba;
        struct ehea_q_skb_arr rq2_skba;
        struct ehea_q_skb_arr rq3_skba;
@@ -368,11 +358,8 @@ struct ehea_port_res {
        int swqe_refill_th;
        atomic_t swqe_avail;
        int swqe_ll_count;
-       int swqe_count;
        u32 swqe_id_counter;
        u64 tx_packets;
-       spinlock_t recv_lock;
-       struct port_state p_state;
        u64 rx_packets;
        u32 poll_counter;
 };
index 19950273ceb903105480c3a9fe28734c5a471b32..decec8cfe96b8c2789705971409f2344f8b5558d 100644 (file)
@@ -166,33 +166,23 @@ static u32 ehea_get_rx_csum(struct net_device *dev)
 }
 
 static char ehea_ethtool_stats_keys[][ETH_GSTRING_LEN] = {
-       {"poll_max_processed"},
-       {"queue_stopped"},
-       {"min_swqe_avail"},
-       {"poll_receive_err"},
-       {"pkt_send"},
-       {"pkt_xmit"},
-       {"send_tasklet"},
-       {"ehea_poll"},
-       {"nwqe"},
-       {"swqe_available_0"},
        {"sig_comp_iv"},
        {"swqe_refill_th"},
        {"port resets"},
-       {"rxo"},
-       {"rx64"},
-       {"rx65"},
-       {"rx128"},
-       {"rx256"},
-       {"rx512"},
-       {"rx1024"},
-       {"txo"},
-       {"tx64"},
-       {"tx65"},
-       {"tx128"},
-       {"tx256"},
-       {"tx512"},
-       {"tx1024"},
+       {"Receive errors"},
+       {"TCP cksum errors"},
+       {"IP cksum errors"},
+       {"Frame cksum errors"},
+       {"num SQ stopped"},
+       {"SQ stopped"},
+       {"PR0 free_swqes"},
+       {"PR1 free_swqes"},
+       {"PR2 free_swqes"},
+       {"PR3 free_swqes"},
+       {"PR4 free_swqes"},
+       {"PR5 free_swqes"},
+       {"PR6 free_swqes"},
+       {"PR7 free_swqes"},
 };
 
 static void ehea_get_strings(struct net_device *dev, u32 stringset, u8 *data)
@@ -211,63 +201,44 @@ static int ehea_get_stats_count(struct net_device *dev)
 static void ehea_get_ethtool_stats(struct net_device *dev,
                                     struct ethtool_stats *stats, u64 *data)
 {
-       u64 hret;
-       int i;
+       int i, k, tmp;
        struct ehea_port *port = netdev_priv(dev);
-       struct ehea_adapter *adapter = port->adapter;
-       struct ehea_port_res *pr = &port->port_res[0];
-       struct port_state *p_state = &pr->p_state;
-       struct hcp_ehea_port_cb6 *cb6;
 
        for (i = 0; i < ehea_get_stats_count(dev); i++)
                data[i] = 0;
-
        i = 0;
 
-       data[i++] = p_state->poll_max_processed;
-       data[i++] = p_state->queue_stopped;
-       data[i++] = p_state->min_swqe_avail;
-       data[i++] = p_state->poll_receive_errors;
-       data[i++] = p_state->pkt_send;
-       data[i++] = p_state->pkt_xmit;
-       data[i++] = p_state->send_tasklet;
-       data[i++] = p_state->ehea_poll;
-       data[i++] = p_state->nwqe;
-       data[i++] = atomic_read(&port->port_res[0].swqe_avail);
        data[i++] = port->sig_comp_iv;
        data[i++] = port->port_res[0].swqe_refill_th;
        data[i++] = port->resets;
 
-       cb6 = kzalloc(PAGE_SIZE, GFP_KERNEL);
-       if (!cb6) {
-               ehea_error("no mem for cb6");
-               return;
-       }
+       for (k = 0, tmp = 0; k < EHEA_MAX_PORT_RES; k++)
+               tmp += port->port_res[k].p_stats.poll_receive_errors;
+       data[i++] = tmp;
+
+       for (k = 0, tmp = 0; k < EHEA_MAX_PORT_RES; k++)
+               tmp += port->port_res[k].p_stats.err_tcp_cksum;
+       data[i++] = tmp;
+
+       for (k = 0, tmp = 0; k < EHEA_MAX_PORT_RES; k++)
+               tmp += port->port_res[k].p_stats.err_ip_cksum;
+       data[i++] = tmp;
+
+       for (k = 0, tmp = 0; k < EHEA_MAX_PORT_RES; k++)
+               tmp += port->port_res[k].p_stats.err_frame_crc;
+       data[i++] = tmp;
+
+       for (k = 0, tmp = 0; k < EHEA_MAX_PORT_RES; k++)
+               tmp += port->port_res[k].p_stats.queue_stopped;
+       data[i++] = tmp;
+
+       for (k = 0, tmp = 0; k < EHEA_MAX_PORT_RES; k++)
+               tmp |= port->port_res[k].queue_stopped;
+       data[i++] = tmp;
+
+       for (k = 0; k < 8; k++)
+               data[i++] = atomic_read(&port->port_res[k].swqe_avail);
 
-       hret = ehea_h_query_ehea_port(adapter->handle, port->logical_port_id,
-                                     H_PORT_CB6, H_PORT_CB6_ALL, cb6);
-       if (netif_msg_hw(port))
-               ehea_dump(cb6, sizeof(*cb6), "ehea_get_ethtool_stats");
-
-       if (hret == H_SUCCESS) {
-               data[i++] = cb6->rxo;
-               data[i++] = cb6->rx64;
-               data[i++] = cb6->rx65;
-               data[i++] = cb6->rx128;
-               data[i++] = cb6->rx256;
-               data[i++] = cb6->rx512;
-               data[i++] = cb6->rx1024;
-               data[i++] = cb6->txo;
-               data[i++] = cb6->tx64;
-               data[i++] = cb6->tx65;
-               data[i++] = cb6->tx128;
-               data[i++] = cb6->tx256;
-               data[i++] = cb6->tx512;
-               data[i++] = cb6->tx1024;
-       } else
-               ehea_error("query_ehea_port failed");
-
-       kfree(cb6);
 }
 
 const struct ethtool_ops ehea_ethtool_ops = {
index 8bceb4e6bb8243d31f04b60d7f40d62ca6291ba2..e6fe2cfbd999a194aa33594c03bad283e3f2789f 100644 (file)
@@ -327,6 +327,13 @@ static int ehea_treat_poll_error(struct ehea_port_res *pr, int rq,
 {
        struct sk_buff *skb;
 
+       if (cqe->status & EHEA_CQE_STAT_ERR_TCP)
+               pr->p_stats.err_tcp_cksum++;
+       if (cqe->status & EHEA_CQE_STAT_ERR_IP)
+               pr->p_stats.err_ip_cksum++;
+       if (cqe->status & EHEA_CQE_STAT_ERR_CRC)
+               pr->p_stats.err_frame_crc++;
+
        if (netif_msg_rx_err(pr->port)) {
                ehea_error("CQE Error for QP %d", pr->qp->init_attr.qp_nr);
                ehea_dump(cqe, sizeof(*cqe), "CQE");
@@ -428,7 +435,7 @@ static struct ehea_cqe *ehea_proc_rwqes(struct net_device *dev,
                        else
                                netif_receive_skb(skb);
                } else {
-                       pr->p_state.poll_receive_errors++;
+                       pr->p_stats.poll_receive_errors++;
                        port_reset = ehea_treat_poll_error(pr, rq, cqe,
                                                           &processed_rq2,
                                                           &processed_rq3);
@@ -449,34 +456,15 @@ static struct ehea_cqe *ehea_proc_rwqes(struct net_device *dev,
        return cqe;
 }
 
-static void ehea_free_sent_skbs(struct ehea_cqe *cqe, struct ehea_port_res *pr)
-{
-       struct sk_buff *skb;
-       int index, max_index_mask, i;
-
-       index = EHEA_BMASK_GET(EHEA_WR_ID_INDEX, cqe->wr_id);
-       max_index_mask = pr->sq_skba.len - 1;
-       for (i = 0; i < EHEA_BMASK_GET(EHEA_WR_ID_REFILL, cqe->wr_id); i++) {
-               skb = pr->sq_skba.arr[index];
-               if (likely(skb)) {
-                       dev_kfree_skb(skb);
-                       pr->sq_skba.arr[index] = NULL;
-               } else {
-                       ehea_error("skb=NULL, wr_id=%lX, loop=%d, index=%d",
-                                  cqe->wr_id, i, index);
-               }
-               index--;
-               index &= max_index_mask;
-       }
-}
-
 static struct ehea_cqe *ehea_proc_cqes(struct ehea_port_res *pr, int my_quota)
 {
+       struct sk_buff *skb;
        struct ehea_cq *send_cq = pr->send_cq;
        struct ehea_cqe *cqe;
        int quota = my_quota;
        int cqe_counter = 0;
        int swqe_av = 0;
+       int index;
        unsigned long flags;
 
        cqe = ehea_poll_cq(send_cq);
@@ -498,8 +486,13 @@ static struct ehea_cqe *ehea_proc_cqes(struct ehea_port_res *pr, int my_quota)
                        ehea_dump(cqe, sizeof(*cqe), "CQE");
 
                if (likely(EHEA_BMASK_GET(EHEA_WR_ID_TYPE, cqe->wr_id)
-                          == EHEA_SWQE2_TYPE))
-                       ehea_free_sent_skbs(cqe, pr);
+                          == EHEA_SWQE2_TYPE)) {
+
+                       index = EHEA_BMASK_GET(EHEA_WR_ID_INDEX, cqe->wr_id);
+                       skb = pr->sq_skba.arr[index];
+                       dev_kfree_skb(skb);
+                       pr->sq_skba.arr[index] = NULL;
+               }
 
                swqe_av += EHEA_BMASK_GET(EHEA_WR_ID_REFILL, cqe->wr_id);
                quota--;
@@ -1092,8 +1085,6 @@ static int ehea_init_port_res(struct ehea_port *port, struct ehea_port_res *pr,
        memset(pr, 0, sizeof(struct ehea_port_res));
 
        pr->port = port;
-       spin_lock_init(&pr->send_lock);
-       spin_lock_init(&pr->recv_lock);
        spin_lock_init(&pr->xmit_lock);
        spin_lock_init(&pr->netif_queue);
 
@@ -1811,7 +1802,6 @@ static int ehea_start_xmit(struct sk_buff *skb, struct net_device *dev)
 
        pr = &port->port_res[ehea_hash_skb(skb, port->num_tx_qps)];
 
-
        if (!spin_trylock(&pr->xmit_lock))
                return NETDEV_TX_BUSY;
 
@@ -1841,6 +1831,7 @@ static int ehea_start_xmit(struct sk_buff *skb, struct net_device *dev)
                swqe->wr_id =
                        EHEA_BMASK_SET(EHEA_WR_ID_TYPE, EHEA_SWQE2_TYPE)
                      | EHEA_BMASK_SET(EHEA_WR_ID_COUNT, pr->swqe_id_counter)
+                     | EHEA_BMASK_SET(EHEA_WR_ID_REFILL, 1)
                      | EHEA_BMASK_SET(EHEA_WR_ID_INDEX, pr->sq_skba.index);
                pr->sq_skba.arr[pr->sq_skba.index] = skb;
 
@@ -1849,14 +1840,7 @@ static int ehea_start_xmit(struct sk_buff *skb, struct net_device *dev)
 
                lkey = pr->send_mr.lkey;
                ehea_xmit2(skb, dev, swqe, lkey);
-
-               if (pr->swqe_count >= (EHEA_SIG_IV_LONG - 1)) {
-                       swqe->wr_id |= EHEA_BMASK_SET(EHEA_WR_ID_REFILL,
-                                                     EHEA_SIG_IV_LONG);
-                       swqe->tx_control |= EHEA_SWQE_SIGNALLED_COMPLETION;
-                       pr->swqe_count = 0;
-               } else
-                       pr->swqe_count += 1;
+               swqe->tx_control |= EHEA_SWQE_SIGNALLED_COMPLETION;
        }
        pr->swqe_id_counter += 1;
 
@@ -1876,6 +1860,7 @@ static int ehea_start_xmit(struct sk_buff *skb, struct net_device *dev)
        if (unlikely(atomic_read(&pr->swqe_avail) <= 1)) {
                spin_lock_irqsave(&pr->netif_queue, flags);
                if (unlikely(atomic_read(&pr->swqe_avail) <= 1)) {
+                       pr->p_stats.queue_stopped++;
                        netif_stop_queue(dev);
                        pr->queue_stopped = 1;
                }
index 24603312eb84f1500b74c6bca334f0c33afbb4df..c0eb3e03a1021ddd4273bd473b9715fab603403c 100644 (file)
@@ -142,6 +142,8 @@ struct ehea_rwqe {
 #define EHEA_CQE_STAT_ERR_MASK     0x721F
 #define EHEA_CQE_STAT_FAT_ERR_MASK 0x1F
 #define EHEA_CQE_STAT_ERR_TCP      0x4000
+#define EHEA_CQE_STAT_ERR_IP       0x2000
+#define EHEA_CQE_STAT_ERR_CRC      0x1000
 
 struct ehea_cqe {
        u64 wr_id;              /* work request ID from WQE */