memblock: Factor the lowest level alloc function

Signed-off-by: default avatarBenjamin Herrenschmidt <benh@kernel.crashing.org>
parent 35a1f0bd
...@@ -294,8 +294,8 @@ static u64 memblock_align_up(u64 addr, u64 size) ...@@ -294,8 +294,8 @@ static u64 memblock_align_up(u64 addr, u64 size)
return (addr + (size - 1)) & ~(size - 1); return (addr + (size - 1)) & ~(size - 1);
} }
static u64 __init memblock_alloc_nid_unreserved(u64 start, u64 end, static u64 __init memblock_alloc_region(u64 start, u64 end,
u64 size, u64 align) u64 size, u64 align)
{ {
u64 base, res_base; u64 base, res_base;
long j; long j;
...@@ -318,6 +318,13 @@ static u64 __init memblock_alloc_nid_unreserved(u64 start, u64 end, ...@@ -318,6 +318,13 @@ static u64 __init memblock_alloc_nid_unreserved(u64 start, u64 end,
return ~(u64)0; return ~(u64)0;
} }
u64 __weak __init memblock_nid_range(u64 start, u64 end, int *nid)
{
*nid = 0;
return end;
}
static u64 __init memblock_alloc_nid_region(struct memblock_region *mp, static u64 __init memblock_alloc_nid_region(struct memblock_region *mp,
u64 size, u64 align, int nid) u64 size, u64 align, int nid)
{ {
...@@ -333,8 +340,7 @@ static u64 __init memblock_alloc_nid_region(struct memblock_region *mp, ...@@ -333,8 +340,7 @@ static u64 __init memblock_alloc_nid_region(struct memblock_region *mp,
this_end = memblock_nid_range(start, end, &this_nid); this_end = memblock_nid_range(start, end, &this_nid);
if (this_nid == nid) { if (this_nid == nid) {
u64 ret = memblock_alloc_nid_unreserved(start, this_end, u64 ret = memblock_alloc_region(start, this_end, size, align);
size, align);
if (ret != ~(u64)0) if (ret != ~(u64)0)
return ret; return ret;
} }
...@@ -351,6 +357,10 @@ u64 __init memblock_alloc_nid(u64 size, u64 align, int nid) ...@@ -351,6 +357,10 @@ u64 __init memblock_alloc_nid(u64 size, u64 align, int nid)
BUG_ON(0 == size); BUG_ON(0 == size);
/* We do a bottom-up search for a region with the right
* nid since that's easier considering how memblock_nid_range()
* works
*/
size = memblock_align_up(size, align); size = memblock_align_up(size, align);
for (i = 0; i < mem->cnt; i++) { for (i = 0; i < mem->cnt; i++) {
...@@ -383,7 +393,7 @@ u64 __init memblock_alloc_base(u64 size, u64 align, u64 max_addr) ...@@ -383,7 +393,7 @@ u64 __init memblock_alloc_base(u64 size, u64 align, u64 max_addr)
u64 __init __memblock_alloc_base(u64 size, u64 align, u64 max_addr) u64 __init __memblock_alloc_base(u64 size, u64 align, u64 max_addr)
{ {
long i, j; long i;
u64 base = 0; u64 base = 0;
u64 res_base; u64 res_base;
...@@ -396,33 +406,24 @@ u64 __init __memblock_alloc_base(u64 size, u64 align, u64 max_addr) ...@@ -396,33 +406,24 @@ u64 __init __memblock_alloc_base(u64 size, u64 align, u64 max_addr)
if (max_addr == MEMBLOCK_ALLOC_ANYWHERE) if (max_addr == MEMBLOCK_ALLOC_ANYWHERE)
max_addr = MEMBLOCK_REAL_LIMIT; max_addr = MEMBLOCK_REAL_LIMIT;
/* Pump up max_addr */
if (max_addr == MEMBLOCK_ALLOC_ANYWHERE)
max_addr = ~(u64)0;
/* We do a top-down search, this tends to limit memory
* fragmentation by keeping early boot allocs near the
* top of memory
*/
for (i = memblock.memory.cnt - 1; i >= 0; i--) { for (i = memblock.memory.cnt - 1; i >= 0; i--) {
u64 memblockbase = memblock.memory.regions[i].base; u64 memblockbase = memblock.memory.regions[i].base;
u64 memblocksize = memblock.memory.regions[i].size; u64 memblocksize = memblock.memory.regions[i].size;
if (memblocksize < size) if (memblocksize < size)
continue; continue;
if (max_addr == MEMBLOCK_ALLOC_ANYWHERE) base = min(memblockbase + memblocksize, max_addr);
base = memblock_align_down(memblockbase + memblocksize - size, align); res_base = memblock_alloc_region(memblockbase, base, size, align);
else if (memblockbase < max_addr) { if (res_base != ~(u64)0)
base = min(memblockbase + memblocksize, max_addr); return res_base;
base = memblock_align_down(base - size, align);
} else
continue;
while (base && memblockbase <= base) {
j = memblock_overlaps_region(&memblock.reserved, base, size);
if (j < 0) {
/* this area isn't reserved, take it */
if (memblock_add_region(&memblock.reserved, base, size) < 0)
return 0;
return base;
}
res_base = memblock.reserved.regions[j].base;
if (res_base < size)
break;
base = memblock_align_down(res_base - size, align);
}
} }
return 0; return 0;
} }
...@@ -528,9 +529,3 @@ int memblock_is_region_reserved(u64 base, u64 size) ...@@ -528,9 +529,3 @@ int memblock_is_region_reserved(u64 base, u64 size)
return memblock_overlaps_region(&memblock.reserved, base, size) >= 0; return memblock_overlaps_region(&memblock.reserved, base, size) >= 0;
} }
u64 __weak memblock_nid_range(u64 start, u64 end, int *nid)
{
*nid = 0;
return end;
}
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment