get_file(file);
                        if (tmp->vm_flags & VM_DENYWRITE)
                                atomic_dec(&inode->i_writecount);
-      
+
                        /* insert tmp into the share list, just after mpnt */
                        spin_lock(&file->f_mapping->i_mmap_lock);
                        tmp->vm_truncate_count = mpnt->vm_truncate_count;
 #define mm_free_pgd(mm)
 #endif /* CONFIG_MMU */
 
- __cacheline_aligned_in_smp DEFINE_SPINLOCK(mmlist_lock);
+__cacheline_aligned_in_smp DEFINE_SPINLOCK(mmlist_lock);
 
 #define allocate_mm()  (kmem_cache_alloc(mm_cachep, GFP_KERNEL))
 #define free_mm(mm)    (kmem_cache_free(mm_cachep, (mm)))
        /* compute the remainder to be cleared */
        size = (new_fdt->max_fds - open_files) * sizeof(struct file *);
 
-       /* This is long word aligned thus could use a optimized version */ 
-       memset(new_fds, 0, size); 
+       /* This is long word aligned thus could use a optimized version */
+       memset(new_fds, 0, size);
 
        if (new_fdt->max_fds > open_files) {
                int left = (new_fdt->max_fds-open_files)/8;
        task_io_accounting_init(p);
        acct_clear_integrals(p);
 
-       p->it_virt_expires = cputime_zero;
+       p->it_virt_expires = cputime_zero;
        p->it_prof_expires = cputime_zero;
-       p->it_sched_expires = 0;
-       INIT_LIST_HEAD(&p->cpu_timers[0]);
-       INIT_LIST_HEAD(&p->cpu_timers[1]);
-       INIT_LIST_HEAD(&p->cpu_timers[2]);
+       p->it_sched_expires = 0;
+       INIT_LIST_HEAD(&p->cpu_timers[0]);
+       INIT_LIST_HEAD(&p->cpu_timers[1]);
+       INIT_LIST_HEAD(&p->cpu_timers[2]);
 
        p->lock_depth = -1;             /* -1 = no lock */
        do_posix_clock_monotonic_gettime(&p->start_time);
         * A fatal signal pending means that current will exit, so the new
         * thread can't slip out of an OOM kill (or normal SIGKILL).
         */
-       recalc_sigpending();
+       recalc_sigpending();
        if (signal_pending(current)) {
                spin_unlock(¤t->sighand->siglock);
                write_unlock_irq(&tasklist_lock);