Commit 56165c20 authored by unknown's avatar unknown

ndb - bug#34378

  Using more than 16g can cause record-pool ptr.i values to overflow
  Fix by splitting memory into 2 zones, lo(16g)/hi(rest)
  When record pools only use zone_lo, and datamemory, buffers etc...can use any


storage/ndb/src/kernel/blocks/lgman.cpp:
  adopt to changed interface for Ndbd_mem_manager
storage/ndb/src/kernel/vm/Pool.cpp:
  Always use ZONE_LO for record pools
    as they use ptr.i == 19 bit page id + 13 bit page index
storage/ndb/src/kernel/vm/ndbd_malloc_impl.cpp:
  Add zones to Ndbd_mem_manager
  ZONE_LO = lower 16g
  ZONE_HI = rest
storage/ndb/src/kernel/vm/ndbd_malloc_impl.hpp:
  Add zones to Ndbd_mem_manager
  ZONE_LO = lower 16g
  ZONE_HI = rest
parent 25bddc4d
...@@ -918,7 +918,7 @@ Lgman::alloc_logbuffer_memory(Ptr<Logfile_group> ptr, Uint32 bytes) ...@@ -918,7 +918,7 @@ Lgman::alloc_logbuffer_memory(Ptr<Logfile_group> ptr, Uint32 bytes)
{ {
Uint32 ptrI; Uint32 ptrI;
Uint32 cnt = pages > 64 ? 64 : pages; Uint32 cnt = pages > 64 ? 64 : pages;
m_ctx.m_mm.alloc(&ptrI, &cnt, 1); m_ctx.m_mm.alloc_pages(RG_DISK_OPERATIONS, &ptrI, &cnt, 1);
if (cnt) if (cnt)
{ {
Buffer_idx range; Buffer_idx range;
...@@ -1037,7 +1037,7 @@ Lgman::free_logbuffer_memory(Ptr<Logfile_group> ptr) ...@@ -1037,7 +1037,7 @@ Lgman::free_logbuffer_memory(Ptr<Logfile_group> ptr)
ndbrequire(map.next(it)); ndbrequire(map.next(it));
tmp[1] = *it.data; tmp[1] = *it.data;
m_ctx.m_mm.release(range.m_ptr_i, range.m_idx); m_ctx.m_mm.release_pages(RG_DISK_OPERATIONS, range.m_ptr_i, range.m_idx);
map.next(it); map.next(it);
} }
map.release(); map.release();
......
...@@ -20,7 +20,8 @@ ...@@ -20,7 +20,8 @@
void* void*
Pool_context::alloc_page(Uint32 type_id, Uint32 *i) Pool_context::alloc_page(Uint32 type_id, Uint32 *i)
{ {
return m_block->m_ctx.m_mm.alloc_page(type_id, i); return m_block->m_ctx.m_mm.alloc_page(type_id, i,
Ndbd_mem_manager::NDB_ZONE_LO);
} }
void void
......
...@@ -25,19 +25,22 @@ extern EventLogger g_eventLogger; ...@@ -25,19 +25,22 @@ extern EventLogger g_eventLogger;
extern EventLogger g_eventLogger; extern EventLogger g_eventLogger;
#endif #endif
#ifdef NDBD_MALLOC_METHOD static int f_method_idx = 0;
#if NDBD_MALLOC_METHOD == sbrk #ifdef NDBD_MALLOC_METHOD_SBRK
static const char * f_method = "sbrk"; static const char * f_method = "SMsm";
#else #else
static const char * f_method = "malloc"; static const char * f_method = "MSms";
#endif
#elif SIZEOF_CHARP == 8
static const char * f_method = "sbrk";
#else
static const char * f_method = "malloc";
#endif #endif
#define MAX_CHUNKS 10 #define MAX_CHUNKS 10
#define ZONE_LO 0
#define ZONE_HI 1
/**
* POOL_RECORD_BITS == 13 => 32 - 13 = 19 bits for page
*/
#define ZONE_LO_BOUND (1u << 19)
struct InitChunk struct InitChunk
{ {
Uint32 m_cnt; Uint32 m_cnt;
...@@ -54,28 +57,42 @@ do_malloc(Uint32 pages, InitChunk* chunk) ...@@ -54,28 +57,42 @@ do_malloc(Uint32 pages, InitChunk* chunk)
pages += 1; pages += 1;
void * ptr = 0; void * ptr = 0;
Uint32 sz = pages; Uint32 sz = pages;
if (strcmp(f_method, "sbrk") == 0)
retry:
char method = f_method[f_method_idx];
switch(method){
case 0:
return false;
case 'S':
case 's':
{ {
ptr = 0; ptr = 0;
while (ptr == 0) while (ptr == 0)
{ {
ptr = sbrk(sizeof(Alloc_page) * sz); ptr = sbrk(sizeof(Alloc_page) * sz);
if (ptr == (void*)-1) if (ptr == (void*)-1)
{ {
if (method == 'S')
{
f_method_idx++;
goto retry;
}
ptr = 0; ptr = 0;
sz = 1 + (9 * sz) / 10; sz = 1 + (9 * sz) / 10;
if (pages >= 32 && sz < 32) if (pages >= 32 && sz < 32)
{ {
sz = pages; sz = pages;
f_method = "malloc"; f_method_idx++;
g_eventLogger.info("sbrk(%lld) failed, trying malloc", goto retry;
(Uint64)(sizeof(Alloc_page) * sz));
break;
} }
} }
} }
break;
} }
if (strcmp(f_method, "malloc") == 0) case 'M':
case 'm':
{ {
ptr = 0; ptr = 0;
while (ptr == 0) while (ptr == 0)
...@@ -83,15 +100,26 @@ do_malloc(Uint32 pages, InitChunk* chunk) ...@@ -83,15 +100,26 @@ do_malloc(Uint32 pages, InitChunk* chunk)
ptr = malloc(sizeof(Alloc_page) * sz); ptr = malloc(sizeof(Alloc_page) * sz);
if (ptr == 0) if (ptr == 0)
{ {
if (method == 'M')
{
f_method_idx++;
goto retry;
}
sz = 1 + (9 * sz) / 10; sz = 1 + (9 * sz) / 10;
if (pages >= 32 && sz < 32) if (pages >= 32 && sz < 32)
{ {
return false; f_method_idx++;
goto retry;
} }
} }
} }
break;
} }
default:
return false;
}
chunk->m_cnt = sz; chunk->m_cnt = sz;
chunk->m_ptr = (Alloc_page*)ptr; chunk->m_ptr = (Alloc_page*)ptr;
const UintPtr align = sizeof(Alloc_page) - 1; const UintPtr align = sizeof(Alloc_page) - 1;
...@@ -151,6 +179,12 @@ Ndbd_mem_manager::Ndbd_mem_manager() ...@@ -151,6 +179,12 @@ Ndbd_mem_manager::Ndbd_mem_manager()
} }
} }
/**
* m_min = reserved
* m_curr = current
* m_max = max alloc, 0 = no limit
*/
void void
Ndbd_mem_manager::set_resource_limit(const Resource_limit& rl) Ndbd_mem_manager::set_resource_limit(const Resource_limit& rl)
{ {
...@@ -176,6 +210,40 @@ Ndbd_mem_manager::get_resource_limit(Uint32 id, Resource_limit& rl) const ...@@ -176,6 +210,40 @@ Ndbd_mem_manager::get_resource_limit(Uint32 id, Resource_limit& rl) const
return false; return false;
} }
static
inline
void
check_resource_limits(Resource_limit* rl)
{
#ifdef VM_TRACE
Uint32 curr = 0;
Uint32 res_alloc = 0;
Uint32 shared_alloc = 0;
Uint32 sumres = 0;
for (Uint32 i = 1; i<XX_RL_COUNT; i++)
{
curr += rl[i].m_curr;
sumres += rl[i].m_min;
assert(rl[i].m_max == 0 || rl[i].m_curr <= rl[i].m_max);
if (rl[i].m_curr > rl[i].m_min)
{
shared_alloc += rl[i].m_curr - rl[i].m_min;
res_alloc += rl[i].m_min;
}
else
{
res_alloc += rl[i].m_curr;
}
}
assert(curr == rl[0].m_curr);
assert(res_alloc + shared_alloc == curr);
assert(res_alloc <= sumres);
assert(sumres == res_alloc + rl[0].m_min);
assert(rl[0].m_curr <= rl[0].m_max);
#endif
}
bool bool
Ndbd_mem_manager::init(bool alloc_less_memory) Ndbd_mem_manager::init(bool alloc_less_memory)
{ {
...@@ -292,6 +360,8 @@ Ndbd_mem_manager::init(bool alloc_less_memory) ...@@ -292,6 +360,8 @@ Ndbd_mem_manager::init(bool alloc_less_memory)
grow(chunks[i].m_start, chunks[i].m_cnt); grow(chunks[i].m_start, chunks[i].m_cnt);
} }
check_resource_limits(m_resource_limit);
return true; return true;
} }
...@@ -321,35 +391,68 @@ Ndbd_mem_manager::grow(Uint32 start, Uint32 cnt) ...@@ -321,35 +391,68 @@ Ndbd_mem_manager::grow(Uint32 start, Uint32 cnt)
cnt--; // last page is always marked as empty cnt--; // last page is always marked as empty
} }
if (!m_used_bitmap_pages.get(start_bmp)) for (Uint32 i = 0; i<m_used_bitmap_pages.size(); i++)
{ if (m_used_bitmap_pages[i] == start_bmp)
if (start != (start_bmp << BPP_2LOG)) goto found;
{
ndbout_c("ndbd_malloc_impl.cpp:%d:grow(%d, %d) %d!=%d"
" - Unable to use due to bitmap pages missaligned!!",
__LINE__, start, cnt, start, (start_bmp << BPP_2LOG));
g_eventLogger.error("ndbd_malloc_impl.cpp:%d:grow(%d, %d)"
" - Unable to use due to bitmap pages missaligned!!",
__LINE__, start, cnt);
return;
}
if (start != (start_bmp << BPP_2LOG))
{
ndbout_c("ndbd_malloc_impl.cpp:%d:grow(%d, %d) %d!=%d not using %uMb"
" - Unable to use due to bitmap pages missaligned!!",
__LINE__, start, cnt, start, (start_bmp << BPP_2LOG),
(cnt >> (20 - 15)));
g_eventLogger.error("ndbd_malloc_impl.cpp:%d:grow(%d, %d) not using %uMb"
" - Unable to use due to bitmap pages missaligned!!",
__LINE__, start, cnt,
(cnt >> (20 - 15)));
dump();
return;
}
#ifdef UNIT_TEST #ifdef UNIT_TEST
ndbout_c("creating bitmap page %d", start_bmp); ndbout_c("creating bitmap page %d", start_bmp);
#endif #endif
{
Alloc_page* bmp = m_base_page + start; Alloc_page* bmp = m_base_page + start;
memset(bmp, 0, sizeof(Alloc_page)); memset(bmp, 0, sizeof(Alloc_page));
m_used_bitmap_pages.set(start_bmp);
cnt--; cnt--;
start++; start++;
} }
m_used_bitmap_pages.push_back(start_bmp);
found:
if (cnt) if (cnt)
{ {
m_resource_limit[0].m_curr += cnt; m_resource_limit[0].m_curr += cnt;
m_resource_limit[0].m_max += cnt; m_resource_limit[0].m_max += cnt;
release(start, cnt); if (start >= ZONE_LO_BOUND)
{
Uint64 mbytes = ((Uint64(cnt) * 32) + 1023) / 1024;
ndbout_c("Adding %uMb to ZONE_HI (%u,%u)", (Uint32)mbytes, start, cnt);
release(start, cnt);
}
else if (start + cnt <= ZONE_LO_BOUND)
{
Uint64 mbytes = ((Uint64(cnt)*32) + 1023) / 1024;
ndbout_c("Adding %uMb to ZONE_LO (%u,%u)", (Uint32)mbytes, start, cnt);
release(start, cnt);
}
else
{
Uint32 cnt0 = ZONE_LO_BOUND - start;
Uint32 cnt1 = start + cnt - ZONE_LO_BOUND;
Uint64 mbytes0 = ((Uint64(cnt0)*32) + 1023) / 1024;
Uint64 mbytes1 = ((Uint64(cnt1)*32) + 1023) / 1024;
ndbout_c("Adding %uMb to ZONE_LO (split %u,%u)", (Uint32)mbytes0,
start, cnt0);
ndbout_c("Adding %uMb to ZONE_HI (split %u,%u)", (Uint32)mbytes1,
ZONE_LO_BOUND, cnt1);
release(start, cnt0);
release(ZONE_LO_BOUND, cnt1);
}
} }
} }
...@@ -362,40 +465,58 @@ Ndbd_mem_manager::release(Uint32 start, Uint32 cnt) ...@@ -362,40 +465,58 @@ Ndbd_mem_manager::release(Uint32 start, Uint32 cnt)
set(start, start+cnt-1); set(start, start+cnt-1);
release_impl(start, cnt); Uint32 zone = start < ZONE_LO_BOUND ? 0 : 1;
release_impl(zone, start, cnt);
} }
void void
Ndbd_mem_manager::release_impl(Uint32 start, Uint32 cnt) Ndbd_mem_manager::release_impl(Uint32 zone, Uint32 start, Uint32 cnt)
{ {
assert(start); assert(start);
Uint32 test = check(start-1, start+cnt); Uint32 test = check(start-1, start+cnt);
if (test & 1) if (start != ZONE_LO_BOUND && test & 1)
{ {
Free_page_data *fd = get_free_page_data(m_base_page + start - 1, Free_page_data *fd = get_free_page_data(m_base_page + start - 1,
start - 1); start - 1);
Uint32 sz = fd->m_size; Uint32 sz = fd->m_size;
Uint32 left = start - sz; Uint32 left = start - sz;
remove_free_list(left, fd->m_list); remove_free_list(zone, left, fd->m_list);
cnt += sz; cnt += sz;
start = left; start = left;
} }
Uint32 right = start + cnt; Uint32 right = start + cnt;
if (test & 2) if (right != ZONE_LO_BOUND && test & 2)
{ {
Free_page_data *fd = get_free_page_data(m_base_page+right, right); Free_page_data *fd = get_free_page_data(m_base_page+right, right);
Uint32 sz = fd->m_size; Uint32 sz = fd->m_size;
remove_free_list(right, fd->m_list); remove_free_list(zone, right, fd->m_list);
cnt += sz; cnt += sz;
} }
insert_free_list(start, cnt); insert_free_list(zone, start, cnt);
} }
void void
Ndbd_mem_manager::alloc(Uint32* ret, Uint32 *pages, Uint32 min) Ndbd_mem_manager::alloc(AllocZone zone,
Uint32* ret, Uint32 *pages, Uint32 min)
{
if (zone == NDB_ZONE_ANY)
{
Uint32 save = * pages;
alloc_impl(ZONE_HI, ret, pages, min);
if (*pages)
return;
* pages = save;
}
alloc_impl(ZONE_LO, ret, pages, min);
}
void
Ndbd_mem_manager::alloc_impl(Uint32 zone,
Uint32* ret, Uint32 *pages, Uint32 min)
{ {
Int32 i; Int32 i;
Uint32 start; Uint32 start;
...@@ -407,19 +528,19 @@ Ndbd_mem_manager::alloc(Uint32* ret, Uint32 *pages, Uint32 min) ...@@ -407,19 +528,19 @@ Ndbd_mem_manager::alloc(Uint32* ret, Uint32 *pages, Uint32 min)
for (i = list; i < 16; i++) for (i = list; i < 16; i++)
{ {
if ((start = m_buddy_lists[i])) if ((start = m_buddy_lists[zone][i]))
{ {
/* ---------------------------------------------------------------- */ /* ---------------------------------------------------------------- */
/* PROPER AMOUNT OF PAGES WERE FOUND. NOW SPLIT THE FOUND */ /* PROPER AMOUNT OF PAGES WERE FOUND. NOW SPLIT THE FOUND */
/* AREA AND RETURN THE PART NOT NEEDED. */ /* AREA AND RETURN THE PART NOT NEEDED. */
/* ---------------------------------------------------------------- */ /* ---------------------------------------------------------------- */
Uint32 sz = remove_free_list(start, i); Uint32 sz = remove_free_list(zone, start, i);
Uint32 extra = sz - cnt; Uint32 extra = sz - cnt;
assert(sz >= cnt); assert(sz >= cnt);
if (extra) if (extra)
{ {
insert_free_list(start + cnt, extra); insert_free_list(zone, start + cnt, extra);
clear_and_set(start, start+cnt-1); clear_and_set(start, start+cnt-1);
} }
else else
...@@ -427,8 +548,7 @@ Ndbd_mem_manager::alloc(Uint32* ret, Uint32 *pages, Uint32 min) ...@@ -427,8 +548,7 @@ Ndbd_mem_manager::alloc(Uint32* ret, Uint32 *pages, Uint32 min)
clear(start, start+cnt-1); clear(start, start+cnt-1);
} }
* ret = start; * ret = start;
m_resource_limit[0].m_curr += cnt; assert(m_resource_limit[0].m_curr + cnt <= m_resource_limit[0].m_max);
assert(m_resource_limit[0].m_curr <= m_resource_limit[0].m_max);
return; return;
} }
} }
...@@ -442,13 +562,13 @@ Ndbd_mem_manager::alloc(Uint32* ret, Uint32 *pages, Uint32 min) ...@@ -442,13 +562,13 @@ Ndbd_mem_manager::alloc(Uint32* ret, Uint32 *pages, Uint32 min)
assert((Int32)list >= min_list); assert((Int32)list >= min_list);
for (i = list - 1; i >= min_list; i--) for (i = list - 1; i >= min_list; i--)
{ {
if ((start = m_buddy_lists[i])) if ((start = m_buddy_lists[zone][i]))
{ {
Uint32 sz = remove_free_list(start, i); Uint32 sz = remove_free_list(zone, start, i);
Uint32 extra = sz - cnt; Uint32 extra = sz - cnt;
if (sz > cnt) if (sz > cnt)
{ {
insert_free_list(start + cnt, extra); insert_free_list(zone, start + cnt, extra);
sz -= extra; sz -= extra;
clear_and_set(start, start+sz-1); clear_and_set(start, start+sz-1);
} }
...@@ -459,8 +579,7 @@ Ndbd_mem_manager::alloc(Uint32* ret, Uint32 *pages, Uint32 min) ...@@ -459,8 +579,7 @@ Ndbd_mem_manager::alloc(Uint32* ret, Uint32 *pages, Uint32 min)
* ret = start; * ret = start;
* pages = sz; * pages = sz;
m_resource_limit[0].m_curr += sz; assert(m_resource_limit[0].m_curr + sz <= m_resource_limit[0].m_max);
assert(m_resource_limit[0].m_curr <= m_resource_limit[0].m_max);
return; return;
} }
} }
...@@ -468,12 +587,12 @@ Ndbd_mem_manager::alloc(Uint32* ret, Uint32 *pages, Uint32 min) ...@@ -468,12 +587,12 @@ Ndbd_mem_manager::alloc(Uint32* ret, Uint32 *pages, Uint32 min)
} }
void void
Ndbd_mem_manager::insert_free_list(Uint32 start, Uint32 size) Ndbd_mem_manager::insert_free_list(Uint32 zone, Uint32 start, Uint32 size)
{ {
Uint32 list = ndb_log2(size) - 1; Uint32 list = ndb_log2(size) - 1;
Uint32 last = start + size - 1; Uint32 last = start + size - 1;
Uint32 head = m_buddy_lists[list]; Uint32 head = m_buddy_lists[zone][list];
Free_page_data* fd_first = get_free_page_data(m_base_page+start, Free_page_data* fd_first = get_free_page_data(m_base_page+start,
start); start);
fd_first->m_list = list; fd_first->m_list = list;
...@@ -495,11 +614,11 @@ Ndbd_mem_manager::insert_free_list(Uint32 start, Uint32 size) ...@@ -495,11 +614,11 @@ Ndbd_mem_manager::insert_free_list(Uint32 start, Uint32 size)
fd->m_prev = start; fd->m_prev = start;
} }
m_buddy_lists[list] = start; m_buddy_lists[zone][list] = start;
} }
Uint32 Uint32
Ndbd_mem_manager::remove_free_list(Uint32 start, Uint32 list) Ndbd_mem_manager::remove_free_list(Uint32 zone, Uint32 start, Uint32 list)
{ {
Free_page_data* fd = get_free_page_data(m_base_page+start, start); Free_page_data* fd = get_free_page_data(m_base_page+start, start);
Uint32 size = fd->m_size; Uint32 size = fd->m_size;
...@@ -509,7 +628,7 @@ Ndbd_mem_manager::remove_free_list(Uint32 start, Uint32 list) ...@@ -509,7 +628,7 @@ Ndbd_mem_manager::remove_free_list(Uint32 start, Uint32 list)
if (prev) if (prev)
{ {
assert(m_buddy_lists[list] != start); assert(m_buddy_lists[zone][list] != start);
fd = get_free_page_data(m_base_page+prev, prev); fd = get_free_page_data(m_base_page+prev, prev);
assert(fd->m_next == start); assert(fd->m_next == start);
assert(fd->m_list == list); assert(fd->m_list == list);
...@@ -517,8 +636,8 @@ Ndbd_mem_manager::remove_free_list(Uint32 start, Uint32 list) ...@@ -517,8 +636,8 @@ Ndbd_mem_manager::remove_free_list(Uint32 start, Uint32 list)
} }
else else
{ {
assert(m_buddy_lists[list] == start); assert(m_buddy_lists[zone][list] == start);
m_buddy_lists[list] = next; m_buddy_lists[zone][list] = next;
} }
if (next) if (next)
...@@ -535,42 +654,62 @@ Ndbd_mem_manager::remove_free_list(Uint32 start, Uint32 list) ...@@ -535,42 +654,62 @@ Ndbd_mem_manager::remove_free_list(Uint32 start, Uint32 list)
void void
Ndbd_mem_manager::dump() const Ndbd_mem_manager::dump() const
{ {
for(Uint32 i = 0; i<16; i++) for (Uint32 zone = 0; zone < 2; zone ++)
{ {
printf(" list: %d - ", i); for (Uint32 i = 0; i<16; i++)
Uint32 head = m_buddy_lists[i];
while(head)
{ {
Free_page_data* fd = get_free_page_data(m_base_page+head, head); printf(" list: %d - ", i);
printf("[ i: %d prev %d next %d list %d size %d ] ", Uint32 head = m_buddy_lists[zone][i];
head, fd->m_prev, fd->m_next, fd->m_list, fd->m_size); while(head)
head = fd->m_next; {
Free_page_data* fd = get_free_page_data(m_base_page+head, head);
printf("[ i: %d prev %d next %d list %d size %d ] ",
head, fd->m_prev, fd->m_next, fd->m_list, fd->m_size);
head = fd->m_next;
}
printf("EOL\n");
}
for (Uint32 i = 0; i<XX_RL_COUNT; i++)
{
printf("ri: %d min: %d curr: %d max: %d\n",
i,
m_resource_limit[i].m_min,
m_resource_limit[i].m_curr,
m_resource_limit[i].m_max);
} }
printf("EOL\n");
} }
} }
void* void*
Ndbd_mem_manager::alloc_page(Uint32 type, Uint32* i) Ndbd_mem_manager::alloc_page(Uint32 type, Uint32* i, AllocZone zone)
{ {
Uint32 idx = type & RG_MASK; Uint32 idx = type & RG_MASK;
assert(idx && idx < XX_RL_COUNT); assert(idx && idx < XX_RL_COUNT);
Resource_limit tot = m_resource_limit[0]; Resource_limit tot = m_resource_limit[0];
Resource_limit rl = m_resource_limit[idx]; Resource_limit rl = m_resource_limit[idx];
Uint32 add = (rl.m_curr < rl.m_min) ? 0 : 1; // Over min ? Uint32 cnt = 1;
Uint32 res0 = (rl.m_curr < rl.m_min) ? 1 : 0;
Uint32 limit = (rl.m_max == 0 || rl.m_curr < rl.m_max) ? 0 : 1; // Over limit Uint32 limit = (rl.m_max == 0 || rl.m_curr < rl.m_max) ? 0 : 1; // Over limit
Uint32 free = (tot.m_min + tot.m_curr < tot.m_max) ? 1 : 0; // Has free Uint32 free = (tot.m_min + tot.m_curr < tot.m_max) ? 1 : 0; // Has free
if (likely(add == 0 || (limit == 0 && free == 1))) assert(tot.m_min >= res0);
if (likely(res0 == 1 || (limit == 0 && free == 1)))
{ {
Uint32 cnt = 1; alloc(zone, i, &cnt, 1);
alloc(i, &cnt, 1); if (likely(cnt))
assert(cnt); {
m_resource_limit[0].m_curr = tot.m_curr + add; m_resource_limit[0].m_curr = tot.m_curr + cnt;
m_resource_limit[idx].m_curr = rl.m_curr + 1; m_resource_limit[0].m_min = tot.m_min - res0;
return m_base_page + *i; m_resource_limit[idx].m_curr = rl.m_curr + cnt;
check_resource_limits(m_resource_limit);
return m_base_page + *i;
}
} }
return 0; return 0;
} }
...@@ -582,10 +721,102 @@ Ndbd_mem_manager::release_page(Uint32 type, Uint32 i) ...@@ -582,10 +721,102 @@ Ndbd_mem_manager::release_page(Uint32 type, Uint32 i)
Resource_limit tot = m_resource_limit[0]; Resource_limit tot = m_resource_limit[0];
Resource_limit rl = m_resource_limit[idx]; Resource_limit rl = m_resource_limit[idx];
Uint32 sub = (rl.m_curr < rl.m_min) ? 0 : 1; // Over min ? Uint32 sub = (rl.m_curr <= rl.m_min) ? 1 : 0; // Over min ?
release(i, 1); release(i, 1);
m_resource_limit[0].m_curr = tot.m_curr - sub; m_resource_limit[0].m_curr = tot.m_curr - 1;
m_resource_limit[0].m_min = tot.m_min + sub;
m_resource_limit[idx].m_curr = rl.m_curr - 1; m_resource_limit[idx].m_curr = rl.m_curr - 1;
check_resource_limits(m_resource_limit);
}
void
Ndbd_mem_manager::alloc_pages(Uint32 type, Uint32* i, Uint32 *cnt, Uint32 min)
{
Uint32 idx = type & RG_MASK;
assert(idx && idx < XX_RL_COUNT);
Resource_limit tot = m_resource_limit[0];
Resource_limit rl = m_resource_limit[idx];
Uint32 req = *cnt;
Uint32 max = rl.m_max - rl.m_curr;
Uint32 res0 = rl.m_min - rl.m_curr;
Uint32 free_shared = tot.m_max - (tot.m_min + tot.m_curr);
Uint32 res1;
if (rl.m_curr + req <= rl.m_min)
{
// all is reserved...
res0 = req;
res1 = 0;
}
else
{
req = rl.m_max ? max : req;
res0 = (rl.m_curr > rl.m_min) ? 0 : res0;
res1 = req - res0;
if (unlikely(res1 > free_shared))
{
res1 = free_shared;
req = res0 + res1;
}
}
// req = pages to alloc
// res0 = portion that is reserved
// res1 = part that is over reserver
assert (res0 + res1 == req);
assert (tot.m_min >= res0);
if (likely(req))
{
// Hi order allocations can always use any zone
alloc(NDB_ZONE_ANY, i, &req, 1);
* cnt = req;
if (unlikely(req < res0)) // Got min than what was reserved :-(
{
res0 = req;
}
assert(tot.m_min >= res0);
assert(tot.m_curr + req <= tot.m_max);
m_resource_limit[0].m_curr = tot.m_curr + req;
m_resource_limit[0].m_min = tot.m_min - res0;
m_resource_limit[idx].m_curr = rl.m_curr + req;
check_resource_limits(m_resource_limit);
return ;
}
* cnt = req;
return;
}
void
Ndbd_mem_manager::release_pages(Uint32 type, Uint32 i, Uint32 cnt)
{
Uint32 idx = type & RG_MASK;
assert(idx && idx < XX_RL_COUNT);
Resource_limit tot = m_resource_limit[0];
Resource_limit rl = m_resource_limit[idx];
release(i, cnt);
Uint32 currnew = rl.m_curr - cnt;
if (rl.m_curr > rl.m_min)
{
if (currnew < rl.m_min)
{
m_resource_limit[0].m_min = tot.m_min + (rl.m_min - currnew);
}
}
else
{
m_resource_limit[0].m_min = tot.m_min + cnt;
}
m_resource_limit[0].m_curr = tot.m_curr - cnt;
m_resource_limit[idx].m_curr = currnew;
check_resource_limits(m_resource_limit);
} }
#ifdef UNIT_TEST #ifdef UNIT_TEST
...@@ -781,3 +1012,4 @@ main(int argc, char** argv) ...@@ -781,3 +1012,4 @@ main(int argc, char** argv)
template class Vector<Chunk>; template class Vector<Chunk>;
#endif #endif
template class Vector<Uint32>;
...@@ -20,6 +20,7 @@ ...@@ -20,6 +20,7 @@
#include <Bitmask.hpp> #include <Bitmask.hpp>
#include <assert.h> #include <assert.h>
#include "Pool.hpp" #include "Pool.hpp"
#include <Vector.hpp>
/** /**
* 13 -> 8192 words -> 32768 bytes * 13 -> 8192 words -> 32768 bytes
...@@ -59,16 +60,19 @@ public: ...@@ -59,16 +60,19 @@ public:
bool init(bool allow_alloc_less_than_requested = true); bool init(bool allow_alloc_less_than_requested = true);
void* get_memroot() const { return (void*)m_base_page;} void* get_memroot() const { return (void*)m_base_page;}
void alloc(Uint32* ret, Uint32 *pages, Uint32 min_requested);
void release(Uint32 start, Uint32 cnt);
void dump() const ; void dump() const ;
void* alloc_page(Uint32 type, Uint32* i); enum AllocZone
{
NDB_ZONE_LO = 0, // Only allocate with page_id < (1 << 13)
NDB_ZONE_ANY = 1 // Allocate with any page_id
};
void* alloc_page(Uint32 type, Uint32* i, enum AllocZone);
void release_page(Uint32 type, Uint32 i); void release_page(Uint32 type, Uint32 i);
void* alloc_pages(Uint32 type, Uint32* i, Uint32 *cnt, Uint32 min = 1); void alloc_pages(Uint32 type, Uint32* i, Uint32 *cnt, Uint32 min = 1);
void release_pages(Uint32 type, Uint32 i, void*p, Uint32 cnt); void release_pages(Uint32 type, Uint32 i, Uint32 cnt);
/** /**
* Compute 2log of size * Compute 2log of size
...@@ -80,25 +84,29 @@ public: ...@@ -80,25 +84,29 @@ public:
private: private:
void grow(Uint32 start, Uint32 cnt); void grow(Uint32 start, Uint32 cnt);
#define XX_RL_COUNT 3 #define XX_RL_COUNT 4
/** /**
* Return pointer to free page data on page * Return pointer to free page data on page
*/ */
static Free_page_data* get_free_page_data(Alloc_page*, Uint32 idx); static Free_page_data* get_free_page_data(Alloc_page*, Uint32 idx);
Bitmask<1> m_used_bitmap_pages; Vector<Uint32> m_used_bitmap_pages;
Uint32 m_buddy_lists[16]; Uint32 m_buddy_lists[2][16];
Resource_limit m_resource_limit[XX_RL_COUNT]; // RG_COUNT in record_types.hpp Resource_limit m_resource_limit[XX_RL_COUNT]; // RG_COUNT in record_types.hpp
Alloc_page * m_base_page; Alloc_page * m_base_page;
void release_impl(Uint32 start, Uint32 cnt); void release_impl(Uint32 zone, Uint32 start, Uint32 cnt);
void insert_free_list(Uint32 start, Uint32 cnt); void insert_free_list(Uint32 zone, Uint32 start, Uint32 cnt);
Uint32 remove_free_list(Uint32 start, Uint32 list); Uint32 remove_free_list(Uint32 zone, Uint32 start, Uint32 list);
void set(Uint32 first, Uint32 last); void set(Uint32 first, Uint32 last);
void clear(Uint32 first, Uint32 last); void clear(Uint32 first, Uint32 last);
void clear_and_set(Uint32 first, Uint32 last); void clear_and_set(Uint32 first, Uint32 last);
Uint32 check(Uint32 first, Uint32 last); Uint32 check(Uint32 first, Uint32 last);
void alloc(AllocZone, Uint32* ret, Uint32 *pages, Uint32 min_requested);
void alloc_impl(Uint32 zone, Uint32* ret, Uint32 *pages, Uint32 min);
void release(Uint32 start, Uint32 cnt);
}; };
inline inline
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment