rpc_init_task(&data->task, &task_setup_data);
                NFS_PROTO(inode)->read_setup(data);
 
-               data->task.tk_cookie = (unsigned long) inode;
-
                rpc_execute(&data->task);
 
                dprintk("NFS: %5u initiated direct read call "
                rpc_init_task(&data->task, &task_setup_data);
                NFS_PROTO(inode)->write_setup(data, FLUSH_STABLE);
 
-               data->task.tk_priority = RPC_PRIORITY_NORMAL;
-               data->task.tk_cookie = (unsigned long) inode;
-
                /*
                 * We're called via an RPC callback, so BKL is already held.
                 */
        rpc_init_task(&data->task, &task_setup_data);
        NFS_PROTO(data->inode)->commit_setup(data, 0);
 
-       data->task.tk_priority = RPC_PRIORITY_NORMAL;
-       data->task.tk_cookie = (unsigned long)data->inode;
        /* Note: task.tk_ops->rpc_release will free dreq->commit_data */
        dreq->commit_data = NULL;
 
                rpc_init_task(&data->task, &task_setup_data);
                NFS_PROTO(inode)->write_setup(data, sync);
 
-               data->task.tk_priority = RPC_PRIORITY_NORMAL;
-               data->task.tk_cookie = (unsigned long) inode;
-
                rpc_execute(&data->task);
 
                dprintk("NFS: %5u initiated direct write call "
 
        rpc_init_task(&data->task, &task_setup_data);
        NFS_PROTO(inode)->read_setup(data);
 
-       data->task.tk_cookie = (unsigned long)inode;
-
        dprintk("NFS: %5u initiated read call (req %s/%Ld, %u bytes @ offset %Lu)\n",
                        data->task.tk_pid,
                        inode->i_sb->s_id,
 
        nfs_clear_page_tag_locked(req);
 }
 
-static inline int flush_task_priority(int how)
+static int flush_task_priority(int how)
 {
        switch (how & (FLUSH_HIGHPRI|FLUSH_LOWPRI)) {
                case FLUSH_HIGHPRI:
 {
        struct inode *inode = req->wb_context->path.dentry->d_inode;
        int flags = (how & FLUSH_SYNC) ? 0 : RPC_TASK_ASYNC;
+       int priority = flush_task_priority(how);
        struct rpc_task_setup task_setup_data = {
                .rpc_client = NFS_CLIENT(inode),
                .callback_ops = call_ops,
                .callback_data = data,
                .flags = flags,
+               .priority = priority,
        };
 
        /* Set up the RPC argument and reply structs
        rpc_init_task(&data->task, &task_setup_data);
        NFS_PROTO(inode)->write_setup(data, how);
 
-       data->task.tk_priority = flush_task_priority(how);
-       data->task.tk_cookie = (unsigned long)inode;
-
        dprintk("NFS: %5u initiated write call "
                "(req %s/%Ld, %u bytes @ offset %Lu)\n",
                data->task.tk_pid,
        struct nfs_page *first = nfs_list_entry(head->next);
        struct inode *inode = first->wb_context->path.dentry->d_inode;
        int flags = (how & FLUSH_SYNC) ? 0 : RPC_TASK_ASYNC;
+       int priority = flush_task_priority(how);
        struct rpc_task_setup task_setup_data = {
                .rpc_client = NFS_CLIENT(inode),
                .callback_ops = &nfs_commit_ops,
                .callback_data = data,
                .flags = flags,
+               .priority = priority,
        };
 
        /* Set up the RPC argument and reply structs
        rpc_init_task(&data->task, &task_setup_data);
        NFS_PROTO(inode)->commit_setup(data, how);
 
-       data->task.tk_priority = flush_task_priority(how);
-       data->task.tk_cookie = (unsigned long)inode;
-       
        dprintk("NFS: %5u initiated commit call\n", data->task.tk_pid);
 }
 
 
        __u8                    tk_garb_retry;
        __u8                    tk_cred_retry;
 
-       unsigned long           tk_cookie;      /* Cookie for batching tasks */
-
        /*
         * timeout_fn   to be executed by timer bottom half
         * callback     to be executed after waking up
        struct timer_list       tk_timer;       /* kernel timer */
        unsigned long           tk_timeout;     /* timeout for rpc_sleep() */
        unsigned short          tk_flags;       /* misc flags */
-       unsigned char           tk_priority : 2;/* Task priority */
        unsigned long           tk_runstate;    /* Task run status */
        struct workqueue_struct *tk_workqueue;  /* Normally rpciod, but could
                                                 * be any workqueue
        unsigned long           tk_start;       /* RPC task init timestamp */
        long                    tk_rtt;         /* round-trip time (jiffies) */
 
+       pid_t                   tk_owner;       /* Process id for batching tasks */
+       unsigned char           tk_priority : 2;/* Task priority */
+
 #ifdef RPC_DEBUG
        unsigned short          tk_pid;         /* debugging aid */
 #endif
        const struct rpc_call_ops *callback_ops;
        void *callback_data;
        unsigned short flags;
+       signed char priority;
 };
 
 /*
  * Note: if you change these, you must also change
  * the task initialization definitions below.
  */
-#define RPC_PRIORITY_LOW       0
-#define RPC_PRIORITY_NORMAL    1
-#define RPC_PRIORITY_HIGH      2
-#define RPC_NR_PRIORITY                (RPC_PRIORITY_HIGH+1)
+#define RPC_PRIORITY_LOW       (-1)
+#define RPC_PRIORITY_NORMAL    (0)
+#define RPC_PRIORITY_HIGH      (1)
+#define RPC_NR_PRIORITY                (1 + RPC_PRIORITY_HIGH - RPC_PRIORITY_LOW)
 
 /*
  * RPC synchronization objects
 struct rpc_wait_queue {
        spinlock_t              lock;
        struct list_head        tasks[RPC_NR_PRIORITY]; /* task queue for each priority level */
-       unsigned long           cookie;                 /* cookie of last task serviced */
+       pid_t                   owner;                  /* process id of last task serviced */
        unsigned char           maxpriority;            /* maximum priority (0 if queue is not a priority queue) */
        unsigned char           priority;               /* current priority */
        unsigned char           count;                  /* # task groups remaining serviced so far */
 
        if (unlikely(task->tk_priority > queue->maxpriority))
                q = &queue->tasks[queue->maxpriority];
        list_for_each_entry(t, q, u.tk_wait.list) {
-               if (t->tk_cookie == task->tk_cookie) {
+               if (t->tk_owner == task->tk_owner) {
                        list_add_tail(&task->u.tk_wait.list, &t->u.tk_wait.links);
                        return;
                }
        queue->count = 1 << (priority * 2);
 }
 
-static inline void rpc_set_waitqueue_cookie(struct rpc_wait_queue *queue, unsigned long cookie)
+static inline void rpc_set_waitqueue_owner(struct rpc_wait_queue *queue, pid_t pid)
 {
-       queue->cookie = cookie;
+       queue->owner = pid;
        queue->nr = RPC_BATCH_COUNT;
 }
 
 static inline void rpc_reset_waitqueue_priority(struct rpc_wait_queue *queue)
 {
        rpc_set_waitqueue_priority(queue, queue->maxpriority);
-       rpc_set_waitqueue_cookie(queue, 0);
+       rpc_set_waitqueue_owner(queue, 0);
 }
 
-static void __rpc_init_priority_wait_queue(struct rpc_wait_queue *queue, const char *qname, int maxprio)
+static void __rpc_init_priority_wait_queue(struct rpc_wait_queue *queue, const char *qname, unsigned char nr_queues)
 {
        int i;
 
        spin_lock_init(&queue->lock);
        for (i = 0; i < ARRAY_SIZE(queue->tasks); i++)
                INIT_LIST_HEAD(&queue->tasks[i]);
-       queue->maxpriority = maxprio;
+       queue->maxpriority = nr_queues - 1;
        rpc_reset_waitqueue_priority(queue);
 #ifdef RPC_DEBUG
        queue->name = qname;
 
 void rpc_init_priority_wait_queue(struct rpc_wait_queue *queue, const char *qname)
 {
-       __rpc_init_priority_wait_queue(queue, qname, RPC_PRIORITY_HIGH);
+       __rpc_init_priority_wait_queue(queue, qname, RPC_NR_PRIORITY);
 }
 
 void rpc_init_wait_queue(struct rpc_wait_queue *queue, const char *qname)
 {
-       __rpc_init_priority_wait_queue(queue, qname, 0);
+       __rpc_init_priority_wait_queue(queue, qname, 1);
 }
 EXPORT_SYMBOL_GPL(rpc_init_wait_queue);
 
        struct rpc_task *task;
 
        /*
-        * Service a batch of tasks from a single cookie.
+        * Service a batch of tasks from a single owner.
         */
        q = &queue->tasks[queue->priority];
        if (!list_empty(q)) {
                task = list_entry(q->next, struct rpc_task, u.tk_wait.list);
-               if (queue->cookie == task->tk_cookie) {
+               if (queue->owner == task->tk_owner) {
                        if (--queue->nr)
                                goto out;
                        list_move_tail(&task->u.tk_wait.list, q);
                 * Check if we need to switch queues.
                 */
                if (--queue->count)
-                       goto new_cookie;
+                       goto new_owner;
        }
 
        /*
 
 new_queue:
        rpc_set_waitqueue_priority(queue, (unsigned int)(q - &queue->tasks[0]));
-new_cookie:
-       rpc_set_waitqueue_cookie(queue, task->tk_cookie);
+new_owner:
+       rpc_set_waitqueue_owner(queue, task->tk_owner);
 out:
        __rpc_wake_up_task(task);
        return task;
        task->tk_garb_retry = 2;
        task->tk_cred_retry = 2;
 
-       task->tk_priority = RPC_PRIORITY_NORMAL;
-       task->tk_cookie = (unsigned long)current;
+       task->tk_priority = task_setup_data->priority - RPC_PRIORITY_LOW;
+       task->tk_owner = current->tgid;
 
        /* Initialize workqueue for async tasks */
        task->tk_workqueue = rpciod_workqueue;