Commit 5dfe8660 authored by Tejun Heo's avatar Tejun Heo Committed by H. Peter Anvin

bootmem: Replace work_with_active_regions() with for_each_mem_pfn_range()

Callback based iteration is cumbersome and much less useful than
for_each_*() iterator.  This patch implements for_each_mem_pfn_range()
which replaces work_with_active_regions().  All the current users of
work_with_active_regions() are converted.

This simplifies walking over early_node_map and will allow converting
internal logics in page_alloc to use iterator instead of walking
early_node_map directly, which in turn will enable moving node
information to memblock.

powerpc change is only compile tested.
Signed-off-by: default avatarTejun Heo <tj@kernel.org>
Link: http://lkml.kernel.org/r/20110714074610.GD3455@htj.dyndns.org
Cc: Yinghai Lu <yinghai@kernel.org>
Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
Signed-off-by: default avatarH. Peter Anvin <hpa@linux.intel.com>
parent fc769a8e
...@@ -127,45 +127,25 @@ static int __cpuinit fake_numa_create_new_node(unsigned long end_pfn, ...@@ -127,45 +127,25 @@ static int __cpuinit fake_numa_create_new_node(unsigned long end_pfn,
} }
/* /*
* get_active_region_work_fn - A helper function for get_node_active_region * get_node_active_region - Return active region containing pfn
* Returns datax set to the start_pfn and end_pfn if they contain
* the initial value of datax->start_pfn between them
* @start_pfn: start page(inclusive) of region to check
* @end_pfn: end page(exclusive) of region to check
* @datax: comes in with ->start_pfn set to value to search for and
* goes out with active range if it contains it
* Returns 1 if search value is in range else 0
*/
static int __init get_active_region_work_fn(unsigned long start_pfn,
unsigned long end_pfn, void *datax)
{
struct node_active_region *data;
data = (struct node_active_region *)datax;
if (start_pfn <= data->start_pfn && end_pfn > data->start_pfn) {
data->start_pfn = start_pfn;
data->end_pfn = end_pfn;
return 1;
}
return 0;
}
/*
* get_node_active_region - Return active region containing start_pfn
* Active range returned is empty if none found. * Active range returned is empty if none found.
* @start_pfn: The page to return the region for. * @pfn: The page to return the region for
* @node_ar: Returned set to the active region containing start_pfn * @node_ar: Returned set to the active region containing @pfn
*/ */
static void __init get_node_active_region(unsigned long start_pfn, static void __init get_node_active_region(unsigned long pfn,
struct node_active_region *node_ar) struct node_active_region *node_ar)
{ {
int nid = early_pfn_to_nid(start_pfn); unsigned long start_pfn, end_pfn;
int i, nid;
node_ar->nid = nid; for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) {
node_ar->start_pfn = start_pfn; if (pfn >= start_pfn && pfn < end_pfn) {
node_ar->end_pfn = start_pfn; node_ar->nid = nid;
work_with_active_regions(nid, get_active_region_work_fn, node_ar); node_ar->start_pfn = start_pfn;
node_ar->end_pfn = end_pfn;
break;
}
}
} }
static void map_cpu_to_node(int cpu, int node) static void map_cpu_to_node(int cpu, int node)
......
...@@ -115,28 +115,13 @@ static void __init memblock_x86_subtract_reserved(struct range *range, int az) ...@@ -115,28 +115,13 @@ static void __init memblock_x86_subtract_reserved(struct range *range, int az)
memblock_reserve_reserved_regions(); memblock_reserve_reserved_regions();
} }
struct count_data {
int nr;
};
static int __init count_work_fn(unsigned long start_pfn,
unsigned long end_pfn, void *datax)
{
struct count_data *data = datax;
data->nr++;
return 0;
}
static int __init count_early_node_map(int nodeid) static int __init count_early_node_map(int nodeid)
{ {
struct count_data data; int i, cnt = 0;
data.nr = 0;
work_with_active_regions(nodeid, count_work_fn, &data);
return data.nr; for_each_mem_pfn_range(i, nodeid, NULL, NULL, NULL)
cnt++;
return cnt;
} }
int __init __get_free_all_memory_range(struct range **rangep, int nodeid, int __init __get_free_all_memory_range(struct range **rangep, int nodeid,
......
...@@ -2178,18 +2178,6 @@ static inline void iommu_prepare_isa(void) ...@@ -2178,18 +2178,6 @@ static inline void iommu_prepare_isa(void)
static int md_domain_init(struct dmar_domain *domain, int guest_width); static int md_domain_init(struct dmar_domain *domain, int guest_width);
static int __init si_domain_work_fn(unsigned long start_pfn,
unsigned long end_pfn, void *datax)
{
int *ret = datax;
*ret = iommu_domain_identity_map(si_domain,
(uint64_t)start_pfn << PAGE_SHIFT,
(uint64_t)end_pfn << PAGE_SHIFT);
return *ret;
}
static int __init si_domain_init(int hw) static int __init si_domain_init(int hw)
{ {
struct dmar_drhd_unit *drhd; struct dmar_drhd_unit *drhd;
...@@ -2221,9 +2209,15 @@ static int __init si_domain_init(int hw) ...@@ -2221,9 +2209,15 @@ static int __init si_domain_init(int hw)
return 0; return 0;
for_each_online_node(nid) { for_each_online_node(nid) {
work_with_active_regions(nid, si_domain_work_fn, &ret); unsigned long start_pfn, end_pfn;
if (ret) int i;
return ret;
for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, NULL) {
ret = iommu_domain_identity_map(si_domain,
PFN_PHYS(start_pfn), PFN_PHYS(end_pfn));
if (ret)
return ret;
}
} }
return 0; return 0;
......
...@@ -1327,9 +1327,27 @@ int add_from_early_node_map(struct range *range, int az, ...@@ -1327,9 +1327,27 @@ int add_from_early_node_map(struct range *range, int az,
int nr_range, int nid); int nr_range, int nid);
u64 __init find_memory_core_early(int nid, u64 size, u64 align, u64 __init find_memory_core_early(int nid, u64 size, u64 align,
u64 goal, u64 limit); u64 goal, u64 limit);
typedef int (*work_fn_t)(unsigned long, unsigned long, void *);
extern void work_with_active_regions(int nid, work_fn_t work_fn, void *data);
extern void sparse_memory_present_with_active_regions(int nid); extern void sparse_memory_present_with_active_regions(int nid);
extern void __next_mem_pfn_range(int *idx, int nid,
unsigned long *out_start_pfn,
unsigned long *out_end_pfn, int *out_nid);
/**
* for_each_mem_pfn_range - early memory pfn range iterator
* @i: an integer used as loop variable
* @nid: node selector, %MAX_NUMNODES for all nodes
* @p_start: ptr to ulong for start pfn of the range, can be %NULL
* @p_end: ptr to ulong for end pfn of the range, can be %NULL
* @p_nid: ptr to int for nid of the range, can be %NULL
*
* Walks over configured memory ranges. Available after early_node_map is
* populated.
*/
#define for_each_mem_pfn_range(i, nid, p_start, p_end, p_nid) \
for (i = -1, __next_mem_pfn_range(&i, nid, p_start, p_end, p_nid); \
i >= 0; __next_mem_pfn_range(&i, nid, p_start, p_end, p_nid))
#endif /* CONFIG_ARCH_POPULATES_NODE_MAP */ #endif /* CONFIG_ARCH_POPULATES_NODE_MAP */
#if !defined(CONFIG_ARCH_POPULATES_NODE_MAP) && \ #if !defined(CONFIG_ARCH_POPULATES_NODE_MAP) && \
......
...@@ -3903,18 +3903,6 @@ int __init add_from_early_node_map(struct range *range, int az, ...@@ -3903,18 +3903,6 @@ int __init add_from_early_node_map(struct range *range, int az,
return nr_range; return nr_range;
} }
void __init work_with_active_regions(int nid, work_fn_t work_fn, void *data)
{
int i;
int ret;
for_each_active_range_index_in_nid(i, nid) {
ret = work_fn(early_node_map[i].start_pfn,
early_node_map[i].end_pfn, data);
if (ret)
break;
}
}
/** /**
* sparse_memory_present_with_active_regions - Call memory_present for each active range * sparse_memory_present_with_active_regions - Call memory_present for each active range
* @nid: The node to call memory_present for. If MAX_NUMNODES, all nodes will be used. * @nid: The node to call memory_present for. If MAX_NUMNODES, all nodes will be used.
...@@ -4421,6 +4409,34 @@ static inline void setup_nr_node_ids(void) ...@@ -4421,6 +4409,34 @@ static inline void setup_nr_node_ids(void)
} }
#endif #endif
/*
* Common iterator interface used to define for_each_mem_pfn_range().
*/
void __meminit __next_mem_pfn_range(int *idx, int nid,
unsigned long *out_start_pfn,
unsigned long *out_end_pfn, int *out_nid)
{
struct node_active_region *r = NULL;
while (++*idx < nr_nodemap_entries) {
if (nid == MAX_NUMNODES || nid == early_node_map[*idx].nid) {
r = &early_node_map[*idx];
break;
}
}
if (!r) {
*idx = -1;
return;
}
if (out_start_pfn)
*out_start_pfn = r->start_pfn;
if (out_end_pfn)
*out_end_pfn = r->end_pfn;
if (out_nid)
*out_nid = r->nid;
}
/** /**
* add_active_range - Register a range of PFNs backed by physical memory * add_active_range - Register a range of PFNs backed by physical memory
* @nid: The node ID the range resides on * @nid: The node ID the range resides on
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment