]> pilppa.org Git - linux-2.6-omap-h63xx.git/blobdiff - init/main.c
generic: add irq_desc in function in parameter
[linux-2.6-omap-h63xx.git] / init / main.c
index 3820323c4c8467c07831a8844532fb85976e9a9f..0d2e60144f83511e11ba89c5b3479dbd4b5c8e93 100644 (file)
@@ -391,17 +391,23 @@ EXPORT_SYMBOL(__per_cpu_offset);
 
 static void __init setup_per_cpu_areas(void)
 {
-       unsigned long size, i;
+       unsigned long size, i, old_size;
        char *ptr;
        unsigned long nr_possible_cpus = num_possible_cpus();
+       unsigned long align = 1;
+       unsigned da_size;
 
        /* Copy section for each CPU (we discard the original) */
-       size = ALIGN(PERCPU_ENOUGH_ROOM, PAGE_SIZE);
+       old_size = PERCPU_ENOUGH_ROOM;
+       da_size = per_cpu_dyn_array_size(&align);
+       align = max_t(unsigned long, PAGE_SIZE, align);
+       size = ALIGN(old_size + da_size, align);
        ptr = alloc_bootmem_pages(size * nr_possible_cpus);
 
        for_each_possible_cpu(i) {
                __per_cpu_offset[i] = ptr - __per_cpu_start;
                memcpy(ptr, __per_cpu_start, __per_cpu_end - __per_cpu_start);
+               per_cpu_alloc_dyn_array(i, ptr + old_size);
                ptr += size;
        }
 }
@@ -536,6 +542,124 @@ void __init __weak thread_info_cache_init(void)
 {
 }
 
+void pre_alloc_dyn_array(void)
+{
+#ifdef CONFIG_HAVE_DYN_ARRAY
+       unsigned long total_size = 0, size, phys;
+       unsigned long max_align = 1;
+       struct dyn_array **daa;
+       char *ptr;
+
+       /* get the total size at first */
+       for (daa = __dyn_array_start ; daa < __dyn_array_end; daa++) {
+               struct dyn_array *da = *daa;
+
+               size = da->size * (*da->nr);
+               print_fn_descriptor_symbol("dyn_array %s ", da->name);
+               printk(KERN_CONT "size:%#lx nr:%d align:%#lx\n",
+                       da->size, *da->nr, da->align);
+               total_size += roundup(size, da->align);
+               if (da->align > max_align)
+                       max_align = da->align;
+       }
+       if (total_size)
+               printk(KERN_DEBUG "dyn_array total_size: %#lx\n",
+                        total_size);
+       else
+               return;
+
+       /* allocate them all together */
+       max_align = max_t(unsigned long, max_align, PAGE_SIZE);
+       ptr = __alloc_bootmem_nopanic(total_size, max_align, 0);
+       if (!ptr)
+               panic("Can not alloc dyn_alloc\n");
+
+       phys = virt_to_phys(ptr);
+       for (daa = __dyn_array_start ; daa < __dyn_array_end; daa++) {
+               struct dyn_array *da = *daa;
+
+               size = da->size * (*da->nr);
+               print_fn_descriptor_symbol("dyn_array %s ", da->name);
+
+               phys = roundup(phys, da->align);
+               *da->name = phys_to_virt(phys);
+               printk(KERN_CONT " ==> [%#lx - %#lx]\n", phys, phys + size);
+
+               phys += size;
+
+               if (da->init_work)
+                       da->init_work(da);
+       }
+#else
+#ifdef CONFIF_GENERIC_HARDIRQS
+       unsigned int i;
+
+       for (i = 0; i < NR_IRQS; i++)
+               irq_desc[i].irq = i;
+#endif
+#endif
+}
+
+unsigned long per_cpu_dyn_array_size(unsigned long *align)
+{
+       unsigned long total_size = 0;
+#ifdef CONFIG_HAVE_DYN_ARRAY
+       unsigned long size;
+       struct dyn_array **daa;
+       unsigned max_align = 1;
+
+       for (daa = __per_cpu_dyn_array_start ; daa < __per_cpu_dyn_array_end; daa++) {
+               struct dyn_array *da = *daa;
+
+               size = da->size * (*da->nr);
+               print_fn_descriptor_symbol("per_cpu_dyn_array %s ", da->name);
+               printk(KERN_CONT "size:%#lx nr:%d align:%#lx\n",
+                       da->size, *da->nr, da->align);
+               total_size += roundup(size, da->align);
+               if (da->align > max_align)
+                       max_align = da->align;
+       }
+       if (total_size) {
+               printk(KERN_DEBUG "per_cpu_dyn_array total_size: %#lx\n",
+                        total_size);
+               *align = max_align;
+       }
+#endif
+       return total_size;
+}
+
+void per_cpu_alloc_dyn_array(int cpu, char *ptr)
+{
+#ifdef CONFIG_HAVE_DYN_ARRAY
+       unsigned long size, phys;
+       struct dyn_array **daa;
+       unsigned long addr;
+       void **array;
+
+       phys = virt_to_phys(ptr);
+       for (daa = __per_cpu_dyn_array_start ; daa < __per_cpu_dyn_array_end; daa++) {
+               struct dyn_array *da = *daa;
+
+               size = da->size * (*da->nr);
+               print_fn_descriptor_symbol("per_cpu_dyn_array %s ", da->name);
+
+               phys = roundup(phys, da->align);
+               addr = (unsigned long)da->name;
+               addr += per_cpu_offset(cpu);
+               array = (void **)addr;
+               *array = phys_to_virt(phys);
+               *da->name = *array; /* so init_work could use it directly */
+               printk(KERN_CONT " ==> [%#lx - %#lx]\n", phys, phys + size);
+
+               phys += size;
+
+               if (da->init_work) {
+                       da->init_work(da);
+               }
+       }
+#endif
+}
+
 asmlinkage void __init start_kernel(void)
 {
        char * command_line;
@@ -567,6 +691,7 @@ asmlinkage void __init start_kernel(void)
        printk(KERN_NOTICE);
        printk(linux_banner);
        setup_arch(&command_line);
+       pre_alloc_dyn_array();
        mm_init_owner(&init_mm, &init_task);
        setup_command_line(command_line);
        unwind_setup();
@@ -708,7 +833,7 @@ int do_one_initcall(initcall_t fn)
        int result;
 
        if (initcall_debug) {
-               printk("calling  %pF\n", fn);
+               printk("calling  %pF @ %i\n", fn, task_pid_nr(current));
                t0 = ktime_get();
        }