diff options
Diffstat (limited to 'arch/x86/kernel/setup_percpu.c')
| -rw-r--r-- | arch/x86/kernel/setup_percpu.c | 66 | 
1 files changed, 8 insertions, 58 deletions
| diff --git a/arch/x86/kernel/setup_percpu.c b/arch/x86/kernel/setup_percpu.c index 7b65275544b2..49325caa7307 100644 --- a/arch/x86/kernel/setup_percpu.c +++ b/arch/x86/kernel/setup_percpu.c @@ -84,60 +84,6 @@ static bool __init pcpu_need_numa(void)  }  #endif -/** - * pcpu_alloc_bootmem - NUMA friendly alloc_bootmem wrapper for percpu - * @cpu: cpu to allocate for - * @size: size allocation in bytes - * @align: alignment - * - * Allocate @size bytes aligned at @align for cpu @cpu.  This wrapper - * does the right thing for NUMA regardless of the current - * configuration. - * - * RETURNS: - * Pointer to the allocated area on success, NULL on failure. - */ -static void * __init pcpu_alloc_bootmem(unsigned int cpu, unsigned long size, -					unsigned long align) -{ -	const unsigned long goal = __pa(MAX_DMA_ADDRESS); -#ifdef CONFIG_NUMA -	int node = early_cpu_to_node(cpu); -	void *ptr; - -	if (!node_online(node) || !NODE_DATA(node)) { -		ptr = memblock_alloc_from(size, align, goal); -		pr_info("cpu %d has no node %d or node-local memory\n", -			cpu, node); -		pr_debug("per cpu data for cpu%d %lu bytes at %016lx\n", -			 cpu, size, __pa(ptr)); -	} else { -		ptr = memblock_alloc_try_nid(size, align, goal, -					     MEMBLOCK_ALLOC_ACCESSIBLE, -					     node); - -		pr_debug("per cpu data for cpu%d %lu bytes on node%d at %016lx\n", -			 cpu, size, node, __pa(ptr)); -	} -	return ptr; -#else -	return memblock_alloc_from(size, align, goal); -#endif -} - -/* - * Helpers for first chunk memory allocation - */ -static void * __init pcpu_fc_alloc(unsigned int cpu, size_t size, size_t align) -{ -	return pcpu_alloc_bootmem(cpu, size, align); -} - -static void __init pcpu_fc_free(void *ptr, size_t size) -{ -	memblock_free(ptr, size); -} -  static int __init pcpu_cpu_distance(unsigned int from, unsigned int to)  {  #ifdef CONFIG_NUMA @@ -150,7 +96,12 @@ static int __init pcpu_cpu_distance(unsigned int from, unsigned int to)  #endif  } -static void __init pcpup_populate_pte(unsigned long addr) +static int __init pcpu_cpu_to_node(int cpu) +{ +	return early_cpu_to_node(cpu); +} + +void __init pcpu_populate_pte(unsigned long addr)  {  	populate_extra_pte(addr);  } @@ -205,15 +156,14 @@ void __init setup_per_cpu_areas(void)  		rc = pcpu_embed_first_chunk(PERCPU_FIRST_CHUNK_RESERVE,  					    dyn_size, atom_size,  					    pcpu_cpu_distance, -					    pcpu_fc_alloc, pcpu_fc_free); +					    pcpu_cpu_to_node);  		if (rc < 0)  			pr_warn("%s allocator failed (%d), falling back to page size\n",  				pcpu_fc_names[pcpu_chosen_fc], rc);  	}  	if (rc < 0)  		rc = pcpu_page_first_chunk(PERCPU_FIRST_CHUNK_RESERVE, -					   pcpu_fc_alloc, pcpu_fc_free, -					   pcpup_populate_pte); +					   pcpu_cpu_to_node);  	if (rc < 0)  		panic("cannot initialize percpu area (err=%d)", rc); |