{
        int retval;
 
-       spin_lock_bh(&xprt->sock_lock);
+       spin_lock_bh(&xprt->transport_lock);
        retval = __xprt_lock_write(xprt, task);
-       spin_unlock_bh(&xprt->sock_lock);
+       spin_unlock_bh(&xprt->transport_lock);
        return retval;
 }
 
 static inline void
 xprt_release_write(struct rpc_xprt *xprt, struct rpc_task *task)
 {
-       spin_lock_bh(&xprt->sock_lock);
+       spin_lock_bh(&xprt->transport_lock);
        __xprt_release_write(xprt, task);
-       spin_unlock_bh(&xprt->sock_lock);
+       spin_unlock_bh(&xprt->transport_lock);
 }
 
 /*
                req->rq_retries = 0;
                xprt_reset_majortimeo(req);
                /* Reset the RTT counters == "slow start" */
-               spin_lock_bh(&xprt->sock_lock);
+               spin_lock_bh(&xprt->transport_lock);
                rpc_init_rtt(req->rq_task->tk_client->cl_rtt, to->to_initval);
-               spin_unlock_bh(&xprt->sock_lock);
+               spin_unlock_bh(&xprt->transport_lock);
                pprintk("RPC: %lu timeout\n", jiffies);
                status = -ETIMEDOUT;
        }
 void xprt_disconnect(struct rpc_xprt *xprt)
 {
        dprintk("RPC:      disconnected transport %p\n", xprt);
-       spin_lock_bh(&xprt->sock_lock);
+       spin_lock_bh(&xprt->transport_lock);
        xprt_clear_connected(xprt);
        rpc_wake_up_status(&xprt->pending, -ENOTCONN);
-       spin_unlock_bh(&xprt->sock_lock);
+       spin_unlock_bh(&xprt->transport_lock);
 }
 
 static void
 {
        struct rpc_xprt *xprt = (struct rpc_xprt *)data;
 
-       spin_lock(&xprt->sock_lock);
+       spin_lock(&xprt->transport_lock);
        if (!list_empty(&xprt->recv) || xprt->shutdown)
                goto out_abort;
        if (test_and_set_bit(XPRT_LOCKED, &xprt->sockstate))
                goto out_abort;
-       spin_unlock(&xprt->sock_lock);
+       spin_unlock(&xprt->transport_lock);
        /* Let keventd close the socket */
        if (test_bit(XPRT_CONNECTING, &xprt->sockstate) != 0)
                xprt_release_write(xprt, NULL);
                schedule_work(&xprt->task_cleanup);
        return;
 out_abort:
-       spin_unlock(&xprt->sock_lock);
+       spin_unlock(&xprt->transport_lock);
 }
 
 /**
        struct rpc_rqst *req = task->tk_rqstp;
        struct rpc_xprt *xprt = req->rq_xprt;
 
-       spin_lock(&xprt->sock_lock);
+       spin_lock(&xprt->transport_lock);
        if (req->rq_received)
                goto out;
 
 out:
        task->tk_timeout = 0;
        rpc_wake_up_task(task);
-       spin_unlock(&xprt->sock_lock);
+       spin_unlock(&xprt->transport_lock);
 }
 
 /**
        if (xprt->shutdown)
                return -EIO;
 
-       spin_lock_bh(&xprt->sock_lock);
+       spin_lock_bh(&xprt->transport_lock);
        if (req->rq_received && !req->rq_bytes_sent) {
                err = req->rq_received;
                goto out_unlock;
                goto out_unlock;
        }
 out_unlock:
-       spin_unlock_bh(&xprt->sock_lock);
+       spin_unlock_bh(&xprt->transport_lock);
        return err;
 }
 
        smp_rmb();
        if (!req->rq_received) {
                if (list_empty(&req->rq_list)) {
-                       spin_lock_bh(&xprt->sock_lock);
+                       spin_lock_bh(&xprt->transport_lock);
                        /* Update the softirq receive buffer */
                        memcpy(&req->rq_private_buf, &req->rq_rcv_buf,
                                        sizeof(req->rq_private_buf));
                        /* Add request to the receive list */
                        list_add_tail(&req->rq_list, &xprt->recv);
-                       spin_unlock_bh(&xprt->sock_lock);
+                       spin_unlock_bh(&xprt->transport_lock);
                        xprt_reset_majortimeo(req);
                        /* Turn off autodisconnect */
                        del_singleshot_timer_sync(&xprt->timer);
  out_receive:
        dprintk("RPC: %4d xmit complete\n", task->tk_pid);
        /* Set the task's receive timeout value */
-       spin_lock_bh(&xprt->sock_lock);
+       spin_lock_bh(&xprt->transport_lock);
        if (!xprt->nocong) {
                int timer = task->tk_msg.rpc_proc->p_timer;
                task->tk_timeout = rpc_calc_rto(clnt->cl_rtt, timer);
        else if (!req->rq_received)
                rpc_sleep_on(&xprt->pending, task, NULL, xprt_timer);
        __xprt_release_write(xprt, task);
-       spin_unlock_bh(&xprt->sock_lock);
+       spin_unlock_bh(&xprt->transport_lock);
 }
 
 static inline void do_xprt_reserve(struct rpc_task *task)
 
        if (!(req = task->tk_rqstp))
                return;
-       spin_lock_bh(&xprt->sock_lock);
+       spin_lock_bh(&xprt->transport_lock);
        __xprt_release_write(xprt, task);
        __xprt_put_cong(xprt, req);
        if (!list_empty(&req->rq_list))
        if (list_empty(&xprt->recv) && !xprt->shutdown)
                mod_timer(&xprt->timer,
                                xprt->last_used + RPC_IDLE_DISCONNECT_TIMEOUT);
-       spin_unlock_bh(&xprt->sock_lock);
+       spin_unlock_bh(&xprt->transport_lock);
        task->tk_rqstp = NULL;
        memset(req, 0, sizeof(*req));   /* mark unused */
 
                return ERR_PTR(result);
        }
 
-       spin_lock_init(&xprt->sock_lock);
+       spin_lock_init(&xprt->transport_lock);
        spin_lock_init(&xprt->xprt_lock);
        init_waitqueue_head(&xprt->cong_wait);
 
 
        if (status == -EAGAIN) {
                if (test_bit(SOCK_ASYNC_NOSPACE, &xprt->sock->flags)) {
                        /* Protect against races with xs_write_space */
-                       spin_lock_bh(&xprt->sock_lock);
+                       spin_lock_bh(&xprt->transport_lock);
                        /* Don't race with disconnect */
                        if (!xprt_connected(xprt))
                                task->tk_status = -ENOTCONN;
                                task->tk_timeout = req->rq_timeout;
                                rpc_sleep_on(&xprt->pending, task, NULL, NULL);
                        }
-                       spin_unlock_bh(&xprt->sock_lock);
+                       spin_unlock_bh(&xprt->transport_lock);
                        return status;
                }
                /* Keep holding the socket if it is blocked */
                goto dropit;
 
        /* Look up and lock the request corresponding to the given XID */
-       spin_lock(&xprt->sock_lock);
+       spin_lock(&xprt->transport_lock);
        rovr = xprt_lookup_rqst(xprt, *xp);
        if (!rovr)
                goto out_unlock;
        xprt_complete_rqst(xprt, rovr, copied);
 
  out_unlock:
-       spin_unlock(&xprt->sock_lock);
+       spin_unlock(&xprt->transport_lock);
  dropit:
        skb_free_datagram(sk, skb);
  out:
        ssize_t r;
 
        /* Find and lock the request corresponding to this xid */
-       spin_lock(&xprt->sock_lock);
+       spin_lock(&xprt->transport_lock);
        req = xprt_lookup_rqst(xprt, xprt->tcp_xid);
        if (!req) {
                xprt->tcp_flags &= ~XPRT_COPY_DATA;
                dprintk("RPC:      XID %08x request not found!\n",
                                ntohl(xprt->tcp_xid));
-               spin_unlock(&xprt->sock_lock);
+               spin_unlock(&xprt->transport_lock);
                return;
        }
 
                                req->rq_task->tk_pid);
                xprt_complete_rqst(xprt, req, xprt->tcp_copied);
        }
-       spin_unlock(&xprt->sock_lock);
+       spin_unlock(&xprt->transport_lock);
        xs_tcp_check_recm(xprt);
 }
 
 
        switch (sk->sk_state) {
        case TCP_ESTABLISHED:
-               spin_lock_bh(&xprt->sock_lock);
+               spin_lock_bh(&xprt->transport_lock);
                if (!xprt_test_and_set_connected(xprt)) {
                        /* Reset TCP record info */
                        xprt->tcp_offset = 0;
                        xprt->tcp_flags = XPRT_COPY_RECM | XPRT_COPY_XID;
                        rpc_wake_up(&xprt->pending);
                }
-               spin_unlock_bh(&xprt->sock_lock);
+               spin_unlock_bh(&xprt->transport_lock);
                break;
        case TCP_SYN_SENT:
        case TCP_SYN_RECV:
        if (!test_and_clear_bit(SOCK_NOSPACE, &sock->flags))
                goto out;
 
-       spin_lock_bh(&xprt->sock_lock);
+       spin_lock_bh(&xprt->transport_lock);
        if (xprt->snd_task)
                rpc_wake_up_task(xprt->snd_task);
-       spin_unlock_bh(&xprt->sock_lock);
+       spin_unlock_bh(&xprt->transport_lock);
 out:
        read_unlock(&sk->sk_callback_lock);
 }