[Patch 3/3] Allocate mca/init stacks as part of per_node reservation.

From: <holt_at_sgi.com>
Date: 2008-02-14 04:55:22
This patch will reserve space on the node at the same time as per_cpu
space is reserved.  By the time the for_each_possible_cpu() cpumask is
initialized, we will have already reserved and initialized the mca and
init stack areas for every possible cpu.

Signed-off-by: Robin Holt <holt@sgi.com>
Signed-off-by: Russ Anderson <rja@sgi.com>

---


Index: per_cpu_mca_v1/arch/ia64/mm/discontig.c
===================================================================
--- per_cpu_mca_v1.orig/arch/ia64/mm/discontig.c	2008-02-13 11:53:55.695414286 -0600
+++ per_cpu_mca_v1/arch/ia64/mm/discontig.c	2008-02-13 11:53:55.875436635 -0600
@@ -27,6 +27,7 @@
 #include <asm/meminit.h>
 #include <asm/numa.h>
 #include <asm/sections.h>
+#include <asm/mca.h>
 
 /*
  * Track per-node information needed to setup the boot memory allocator, the
@@ -125,11 +126,38 @@ static unsigned long __meminit compute_p
 	pernodesize += L1_CACHE_ALIGN(sizeof(pg_data_t));
 	pernodesize += L1_CACHE_ALIGN(sizeof(struct ia64_node_data));
 	pernodesize += L1_CACHE_ALIGN(sizeof(pg_data_t));
+	pernodesize = ALIGN(pernodesize, KERNEL_STACK_SIZE);
+	pernodesize += sizeof(struct ia64_mca_cpu) * cpus;
 	pernodesize = PAGE_ALIGN(pernodesize);
 	return pernodesize;
 }
 
 /**
+ * per_node_mca_setup - setup mca areas on each node for its cpus
+ * @mca_area: mca area on this node
+ * @node: node to setup
+ */
+static unsigned long per_node_mca_setup(unsigned long mca_area, int node)
+{
+#ifdef CONFIG_SMP
+	int cpu;
+
+	for_each_possible_early_cpu(cpu) {
+		if (node == node_cpuid[cpu].nid) {
+			__per_cpu_mca[cpu] = mca_area;
+			mca_area += sizeof(struct ia64_mca_cpu);
+		}
+	}
+#else
+	if (node == 0) {
+		__per_cpu_mca[0] = mca_area;
+		mca_area += sizeof(struct ia64_mca_cpu);
+	}
+#endif
+	return mca_area;
+}
+
+/**
  * per_cpu_node_setup - setup per-cpu areas on each node
  * @cpu_data: per-cpu area on this node
  * @node: node to setup
@@ -186,6 +214,10 @@ static void __init fill_pernode(int node
 	pgdat_list[node]->bdata = bdp;
 	pernode += L1_CACHE_ALIGN(sizeof(pg_data_t));
 
+	pernode = ALIGN(pernode, KERNEL_STACK_SIZE);
+
+	pernode = per_node_mca_setup(pernode, node);
+
 	cpu_data = per_cpu_node_setup(cpu_data, node);
 
 	return;
Index: per_cpu_mca_v1/arch/ia64/mm/contig.c
===================================================================
--- per_cpu_mca_v1.orig/arch/ia64/mm/contig.c	2008-02-13 11:53:32.424525355 -0600
+++ per_cpu_mca_v1/arch/ia64/mm/contig.c	2008-02-13 11:53:55.887438125 -0600
@@ -182,8 +182,26 @@ alloc_per_cpu_data(void)
 	cpu_data = __alloc_bootmem(PERCPU_PAGE_SIZE * NR_CPUS,
 				   PERCPU_PAGE_SIZE, __pa(MAX_DMA_ADDRESS));
 }
+
+static inline void
+alloc_mca_data(void)
+{
+	int cpu;
+
+	for_each_possible_cpu(cpu) {
+		__per_cpu_mca[cpu] = __pa(__alloc_bootmem(
+			sizeof(struct ia64_mca_cpu), KERNEL_STACK_SIZE, 0));
+	}
+}
 #else
 #define alloc_per_cpu_data() do { } while (0)
+
+static inline void
+alloc_mca_data(void)
+{
+	__per_cpu_mca[0] = __pa(__alloc_bootmem(
+			sizeof(struct ia64_mca_cpu), KERNEL_STACK_SIZE, 0));
+}
 #endif /* CONFIG_SMP */
 
 /**
@@ -223,6 +241,8 @@ find_memory (void)
 	find_initrd();
 
 	alloc_per_cpu_data();
+
+	alloc_mca_data();
 }
 
 static int
Index: per_cpu_mca_v1/arch/ia64/kernel/mca.c
===================================================================
--- per_cpu_mca_v1.orig/arch/ia64/kernel/mca.c	2008-02-13 11:53:32.424525355 -0600
+++ per_cpu_mca_v1/arch/ia64/kernel/mca.c	2008-02-13 11:53:55.907440608 -0600
@@ -1759,45 +1759,26 @@ format_mca_init_stack(void *mca_data, un
 	strncpy(p->comm, type, sizeof(p->comm)-1);
 }
 
-/* Caller prevents this from being called after init */
-static void * __init_refok mca_bootmem(void)
-{
-	return __alloc_bootmem(sizeof(struct ia64_mca_cpu),
-	                    KERNEL_STACK_SIZE, 0);
-}
-
 /* Do per-CPU MCA-related initialization.  */
 void __cpuinit
 ia64_mca_cpu_init(void *cpu_data)
 {
 	void *pal_vaddr;
 	void *data;
-	long sz = sizeof(struct ia64_mca_cpu);
 	int cpu = smp_processor_id();
-	static int first_time = 1;
 
 	/*
 	 * Structure will already be allocated if cpu has been online,
 	 * then offlined.
 	 */
-	if (__per_cpu_mca[cpu]) {
-		data = __va(__per_cpu_mca[cpu]);
-	} else {
-		if (first_time) {
-			data = mca_bootmem();
-			first_time = 0;
-		} else
-			data = page_address(alloc_pages_node(numa_node_id(),
-					GFP_KERNEL, get_order(sz)));
-		if (!data)
-			panic("Could not allocate MCA memory for cpu %d\n",
-					cpu);
-	}
+	BUG_ON(__per_cpu_mca[cpu] == 0);
+	data = __va(__per_cpu_mca[cpu]);
+
 	format_mca_init_stack(data, offsetof(struct ia64_mca_cpu, mca_stack),
 		"MCA", cpu);
 	format_mca_init_stack(data, offsetof(struct ia64_mca_cpu, init_stack),
 		"INIT", cpu);
-	__get_cpu_var(ia64_mca_data) = __per_cpu_mca[cpu] = __pa(data);
+	__get_cpu_var(ia64_mca_data) = __pa(data);
 
 	/*
 	 * Stash away a copy of the PTE needed to map the per-CPU page.

-- 
-
To unsubscribe from this list: send the line "unsubscribe linux-ia64" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Received on Thu Feb 14 04:56:27 2008

This archive was generated by hypermail 2.1.8 : 2008-02-14 05:00:13 EST