提交 37b07e41 编写于 作者: L Lee Schermerhorn 提交者: Linus Torvalds

memoryless nodes: fixup uses of node_online_map in generic code

Here's a cut at fixing up uses of the online node map in generic code.

mm/shmem.c:shmem_parse_mpol()

	Ensure nodelist is subset of nodes with memory.
	Use node_states[N_HIGH_MEMORY] as default for missing
	nodelist for interleave policy.

mm/shmem.c:shmem_fill_super()

	initialize policy_nodes to node_states[N_HIGH_MEMORY]

mm/page-writeback.c:highmem_dirtyable_memory()

	sum over nodes with memory

mm/page_alloc.c:zlc_setup()

	allowednodes - use nodes with memory.

mm/page_alloc.c:default_zonelist_order()

	average over nodes with memory.

mm/page_alloc.c:find_next_best_node()

	skip nodes w/o memory.
	N_HIGH_MEMORY state mask may not be initialized at this time,
	unless we want to depend on early_calculate_totalpages() [see
	below].  Will ZONE_MOVABLE ever be configurable?

mm/page_alloc.c:find_zone_movable_pfns_for_nodes()

	spread kernelcore over nodes with memory.

	This required calling early_calculate_totalpages()
	unconditionally, and populating N_HIGH_MEMORY node
	state therein from nodes in the early_node_map[].
	If we can depend on this, we can eliminate the
	population of N_HIGH_MEMORY mask from __build_all_zonelists()
	and use the N_HIGH_MEMORY mask in find_next_best_node().

mm/mempolicy.c:mpol_check_policy()

	Ensure nodes specified for policy are subset of
	nodes with memory.

[akpm@linux-foundation.org: fix warnings]
Signed-off-by: NLee Schermerhorn <lee.schermerhorn@hp.com>
Acked-by: NChristoph Lameter <clameter@sgi.com>
Cc: Shaohua Li <shaohua.li@intel.com>
Signed-off-by: NAndrew Morton <akpm@linux-foundation.org>
Signed-off-by: NLinus Torvalds <torvalds@linux-foundation.org>
上级 0e1e7c7a
...@@ -126,7 +126,7 @@ static int mpol_check_policy(int mode, nodemask_t *nodes) ...@@ -126,7 +126,7 @@ static int mpol_check_policy(int mode, nodemask_t *nodes)
return -EINVAL; return -EINVAL;
break; break;
} }
return nodes_subset(*nodes, node_online_map) ? 0 : -EINVAL; return nodes_subset(*nodes, node_states[N_HIGH_MEMORY]) ? 0 : -EINVAL;
} }
/* Generate a custom zonelist for the BIND policy. */ /* Generate a custom zonelist for the BIND policy. */
...@@ -965,7 +965,7 @@ asmlinkage long sys_migrate_pages(pid_t pid, unsigned long maxnode, ...@@ -965,7 +965,7 @@ asmlinkage long sys_migrate_pages(pid_t pid, unsigned long maxnode,
goto out; goto out;
} }
if (!nodes_subset(new, node_online_map)) { if (!nodes_subset(new, node_states[N_HIGH_MEMORY])) {
err = -EINVAL; err = -EINVAL;
goto out; goto out;
} }
......
...@@ -126,7 +126,7 @@ static unsigned long highmem_dirtyable_memory(unsigned long total) ...@@ -126,7 +126,7 @@ static unsigned long highmem_dirtyable_memory(unsigned long total)
int node; int node;
unsigned long x = 0; unsigned long x = 0;
for_each_online_node(node) { for_each_node_state(node, N_HIGH_MEMORY) {
struct zone *z = struct zone *z =
&NODE_DATA(node)->node_zones[ZONE_HIGHMEM]; &NODE_DATA(node)->node_zones[ZONE_HIGHMEM];
......
...@@ -1040,7 +1040,7 @@ int zone_watermark_ok(struct zone *z, int order, unsigned long mark, ...@@ -1040,7 +1040,7 @@ int zone_watermark_ok(struct zone *z, int order, unsigned long mark,
* *
* If the zonelist cache is present in the passed in zonelist, then * If the zonelist cache is present in the passed in zonelist, then
* returns a pointer to the allowed node mask (either the current * returns a pointer to the allowed node mask (either the current
* tasks mems_allowed, or node_online_map.) * tasks mems_allowed, or node_states[N_HIGH_MEMORY].)
* *
* If the zonelist cache is not available for this zonelist, does * If the zonelist cache is not available for this zonelist, does
* nothing and returns NULL. * nothing and returns NULL.
...@@ -1069,7 +1069,7 @@ static nodemask_t *zlc_setup(struct zonelist *zonelist, int alloc_flags) ...@@ -1069,7 +1069,7 @@ static nodemask_t *zlc_setup(struct zonelist *zonelist, int alloc_flags)
allowednodes = !in_interrupt() && (alloc_flags & ALLOC_CPUSET) ? allowednodes = !in_interrupt() && (alloc_flags & ALLOC_CPUSET) ?
&cpuset_current_mems_allowed : &cpuset_current_mems_allowed :
&node_online_map; &node_states[N_HIGH_MEMORY];
return allowednodes; return allowednodes;
} }
...@@ -1802,7 +1802,7 @@ static int find_next_best_node(int node, nodemask_t *used_node_mask) ...@@ -1802,7 +1802,7 @@ static int find_next_best_node(int node, nodemask_t *used_node_mask)
return node; return node;
} }
for_each_online_node(n) { for_each_node_state(n, N_HIGH_MEMORY) {
cpumask_t tmp; cpumask_t tmp;
/* Don't want a node to appear more than once */ /* Don't want a node to appear more than once */
...@@ -1939,7 +1939,8 @@ static int default_zonelist_order(void) ...@@ -1939,7 +1939,8 @@ static int default_zonelist_order(void)
* If there is a node whose DMA/DMA32 memory is very big area on * If there is a node whose DMA/DMA32 memory is very big area on
* local memory, NODE_ORDER may be suitable. * local memory, NODE_ORDER may be suitable.
*/ */
average_size = total_size / (num_online_nodes() + 1); average_size = total_size /
(nodes_weight(node_states[N_HIGH_MEMORY]) + 1);
for_each_online_node(nid) { for_each_online_node(nid) {
low_kmem_size = 0; low_kmem_size = 0;
total_size = 0; total_size = 0;
...@@ -2098,20 +2099,6 @@ static void build_zonelist_cache(pg_data_t *pgdat) ...@@ -2098,20 +2099,6 @@ static void build_zonelist_cache(pg_data_t *pgdat)
#endif /* CONFIG_NUMA */ #endif /* CONFIG_NUMA */
/* Any regular memory on that node ? */
static void check_for_regular_memory(pg_data_t *pgdat)
{
#ifdef CONFIG_HIGHMEM
enum zone_type zone_type;
for (zone_type = 0; zone_type <= ZONE_NORMAL; zone_type++) {
struct zone *zone = &pgdat->node_zones[zone_type];
if (zone->present_pages)
node_set_state(zone_to_nid(zone), N_NORMAL_MEMORY);
}
#endif
}
/* return values int ....just for stop_machine_run() */ /* return values int ....just for stop_machine_run() */
static int __build_all_zonelists(void *dummy) static int __build_all_zonelists(void *dummy)
{ {
...@@ -2122,11 +2109,6 @@ static int __build_all_zonelists(void *dummy) ...@@ -2122,11 +2109,6 @@ static int __build_all_zonelists(void *dummy)
build_zonelists(pgdat); build_zonelists(pgdat);
build_zonelist_cache(pgdat); build_zonelist_cache(pgdat);
/* Any memory on that node */
if (pgdat->node_present_pages)
node_set_state(nid, N_HIGH_MEMORY);
check_for_regular_memory(pgdat);
} }
return 0; return 0;
} }
...@@ -3282,16 +3264,24 @@ unsigned long __init find_max_pfn_with_active_regions(void) ...@@ -3282,16 +3264,24 @@ unsigned long __init find_max_pfn_with_active_regions(void)
return max_pfn; return max_pfn;
} }
/*
* early_calculate_totalpages()
* Sum pages in active regions for movable zone.
* Populate N_HIGH_MEMORY for calculating usable_nodes.
*/
unsigned long __init early_calculate_totalpages(void) unsigned long __init early_calculate_totalpages(void)
{ {
int i; int i;
unsigned long totalpages = 0; unsigned long totalpages = 0;
for (i = 0; i < nr_nodemap_entries; i++) for (i = 0; i < nr_nodemap_entries; i++) {
totalpages += early_node_map[i].end_pfn - unsigned long pages = early_node_map[i].end_pfn -
early_node_map[i].start_pfn; early_node_map[i].start_pfn;
totalpages += pages;
return totalpages; if (pages)
node_set_state(early_node_map[i].nid, N_HIGH_MEMORY);
}
return totalpages;
} }
/* /*
...@@ -3305,7 +3295,8 @@ void __init find_zone_movable_pfns_for_nodes(unsigned long *movable_pfn) ...@@ -3305,7 +3295,8 @@ void __init find_zone_movable_pfns_for_nodes(unsigned long *movable_pfn)
int i, nid; int i, nid;
unsigned long usable_startpfn; unsigned long usable_startpfn;
unsigned long kernelcore_node, kernelcore_remaining; unsigned long kernelcore_node, kernelcore_remaining;
int usable_nodes = num_online_nodes(); unsigned long totalpages = early_calculate_totalpages();
int usable_nodes = nodes_weight(node_states[N_HIGH_MEMORY]);
/* /*
* If movablecore was specified, calculate what size of * If movablecore was specified, calculate what size of
...@@ -3316,7 +3307,6 @@ void __init find_zone_movable_pfns_for_nodes(unsigned long *movable_pfn) ...@@ -3316,7 +3307,6 @@ void __init find_zone_movable_pfns_for_nodes(unsigned long *movable_pfn)
* what movablecore would have allowed. * what movablecore would have allowed.
*/ */
if (required_movablecore) { if (required_movablecore) {
unsigned long totalpages = early_calculate_totalpages();
unsigned long corepages; unsigned long corepages;
/* /*
...@@ -3341,7 +3331,7 @@ void __init find_zone_movable_pfns_for_nodes(unsigned long *movable_pfn) ...@@ -3341,7 +3331,7 @@ void __init find_zone_movable_pfns_for_nodes(unsigned long *movable_pfn)
restart: restart:
/* Spread kernelcore memory as evenly as possible throughout nodes */ /* Spread kernelcore memory as evenly as possible throughout nodes */
kernelcore_node = required_kernelcore / usable_nodes; kernelcore_node = required_kernelcore / usable_nodes;
for_each_online_node(nid) { for_each_node_state(nid, N_HIGH_MEMORY) {
/* /*
* Recalculate kernelcore_node if the division per node * Recalculate kernelcore_node if the division per node
* now exceeds what is necessary to satisfy the requested * now exceeds what is necessary to satisfy the requested
...@@ -3433,6 +3423,20 @@ void __init find_zone_movable_pfns_for_nodes(unsigned long *movable_pfn) ...@@ -3433,6 +3423,20 @@ void __init find_zone_movable_pfns_for_nodes(unsigned long *movable_pfn)
roundup(zone_movable_pfn[nid], MAX_ORDER_NR_PAGES); roundup(zone_movable_pfn[nid], MAX_ORDER_NR_PAGES);
} }
/* Any regular memory on that node ? */
static void check_for_regular_memory(pg_data_t *pgdat)
{
#ifdef CONFIG_HIGHMEM
enum zone_type zone_type;
for (zone_type = 0; zone_type <= ZONE_NORMAL; zone_type++) {
struct zone *zone = &pgdat->node_zones[zone_type];
if (zone->present_pages)
node_set_state(zone_to_nid(zone), N_NORMAL_MEMORY);
}
#endif
}
/** /**
* free_area_init_nodes - Initialise all pg_data_t and zone data * free_area_init_nodes - Initialise all pg_data_t and zone data
* @max_zone_pfn: an array of max PFNs for each zone * @max_zone_pfn: an array of max PFNs for each zone
...@@ -3507,6 +3511,11 @@ void __init free_area_init_nodes(unsigned long *max_zone_pfn) ...@@ -3507,6 +3511,11 @@ void __init free_area_init_nodes(unsigned long *max_zone_pfn)
pg_data_t *pgdat = NODE_DATA(nid); pg_data_t *pgdat = NODE_DATA(nid);
free_area_init_node(nid, pgdat, NULL, free_area_init_node(nid, pgdat, NULL,
find_min_pfn_for_node(nid), NULL); find_min_pfn_for_node(nid), NULL);
/* Any memory on that node */
if (pgdat->node_present_pages)
node_set_state(nid, N_HIGH_MEMORY);
check_for_regular_memory(pgdat);
} }
} }
......
...@@ -971,7 +971,7 @@ static inline int shmem_parse_mpol(char *value, int *policy, nodemask_t *policy_ ...@@ -971,7 +971,7 @@ static inline int shmem_parse_mpol(char *value, int *policy, nodemask_t *policy_
*nodelist++ = '\0'; *nodelist++ = '\0';
if (nodelist_parse(nodelist, *policy_nodes)) if (nodelist_parse(nodelist, *policy_nodes))
goto out; goto out;
if (!nodes_subset(*policy_nodes, node_online_map)) if (!nodes_subset(*policy_nodes, node_states[N_HIGH_MEMORY]))
goto out; goto out;
} }
if (!strcmp(value, "default")) { if (!strcmp(value, "default")) {
...@@ -996,9 +996,11 @@ static inline int shmem_parse_mpol(char *value, int *policy, nodemask_t *policy_ ...@@ -996,9 +996,11 @@ static inline int shmem_parse_mpol(char *value, int *policy, nodemask_t *policy_
err = 0; err = 0;
} else if (!strcmp(value, "interleave")) { } else if (!strcmp(value, "interleave")) {
*policy = MPOL_INTERLEAVE; *policy = MPOL_INTERLEAVE;
/* Default to nodes online if no nodelist */ /*
* Default to online nodes with memory if no nodelist
*/
if (!nodelist) if (!nodelist)
*policy_nodes = node_online_map; *policy_nodes = node_states[N_HIGH_MEMORY];
err = 0; err = 0;
} }
out: out:
...@@ -1060,7 +1062,8 @@ shmem_alloc_page(gfp_t gfp, struct shmem_inode_info *info, ...@@ -1060,7 +1062,8 @@ shmem_alloc_page(gfp_t gfp, struct shmem_inode_info *info,
return page; return page;
} }
#else #else
static inline int shmem_parse_mpol(char *value, int *policy, nodemask_t *policy_nodes) static inline int shmem_parse_mpol(char *value, int *policy,
nodemask_t *policy_nodes)
{ {
return 1; return 1;
} }
...@@ -2238,7 +2241,7 @@ static int shmem_fill_super(struct super_block *sb, ...@@ -2238,7 +2241,7 @@ static int shmem_fill_super(struct super_block *sb,
unsigned long blocks = 0; unsigned long blocks = 0;
unsigned long inodes = 0; unsigned long inodes = 0;
int policy = MPOL_DEFAULT; int policy = MPOL_DEFAULT;
nodemask_t policy_nodes = node_online_map; nodemask_t policy_nodes = node_states[N_HIGH_MEMORY];
#ifdef CONFIG_TMPFS #ifdef CONFIG_TMPFS
/* /*
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册