reserve huge pages incrementally

This commit is contained in:
daan 2019-11-03 13:16:07 -08:00
parent e320488791
commit f36ec5d9d8
3 changed files with 96 additions and 48 deletions

View file

@ -27,7 +27,10 @@ with on-demand coalescing.
void* _mi_os_alloc_aligned(size_t size, size_t alignment, bool commit, bool* large, mi_os_tld_t* tld); void* _mi_os_alloc_aligned(size_t size, size_t alignment, bool commit, bool* large, mi_os_tld_t* tld);
//int _mi_os_alloc_huge_os_pages(size_t pages, double max_secs, void** pstart, size_t* pages_reserved, size_t* psize) mi_attr_noexcept; //int _mi_os_alloc_huge_os_pages(size_t pages, double max_secs, void** pstart, size_t* pages_reserved, size_t* psize) mi_attr_noexcept;
void _mi_os_free(void* p, size_t size, mi_stats_t* stats); void _mi_os_free(void* p, size_t size, mi_stats_t* stats);
void* _mi_os_alloc_huge_os_pages(size_t pages, int numa_node, size_t* psize);
void* _mi_os_alloc_huge_os_pages(size_t pages, int numa_node, double max_secs, size_t* pages_reserved, size_t* psize);
void _mi_os_free_huge_pages(void* p, size_t size, mi_stats_t* stats);
int _mi_os_numa_node_count(void); int _mi_os_numa_node_count(void);
/* ----------------------------------------------------------- /* -----------------------------------------------------------
@ -234,12 +237,12 @@ static void* mi_arena_alloc_from(mi_arena_t* arena, size_t arena_index, size_t n
void* p = mi_arena_alloc(arena, needed_bcount, is_zero, &block_index); void* p = mi_arena_alloc(arena, needed_bcount, is_zero, &block_index);
if (p != NULL) { if (p != NULL) {
mi_assert_internal(block_index != SIZE_MAX); mi_assert_internal(block_index != SIZE_MAX);
#if MI_DEBUG>=1 #if MI_DEBUG>=1
_Atomic(mi_block_info_t)* block = &arena->blocks[block_index]; _Atomic(mi_block_info_t)* block = &arena->blocks[block_index];
mi_block_info_t binfo = mi_atomic_read(block); mi_block_info_t binfo = mi_atomic_read(block);
mi_assert_internal(mi_block_is_in_use(binfo)); mi_assert_internal(mi_block_is_in_use(binfo));
mi_assert_internal(mi_block_count(binfo) >= needed_bcount); mi_assert_internal(mi_block_count(binfo) >= needed_bcount);
#endif #endif
*memid = mi_memid_create(arena_index, block_index); *memid = mi_memid_create(arena_index, block_index);
*commit = true; // TODO: support commit on demand? *commit = true; // TODO: support commit on demand?
*large = arena->is_large; *large = arena->is_large;
@ -382,18 +385,22 @@ static bool mi_arena_add(mi_arena_t* arena) {
// reserve at a specific numa node // reserve at a specific numa node
int mi_reserve_huge_os_pages_at(size_t pages, int numa_node) mi_attr_noexcept { int mi_reserve_huge_os_pages_at(size_t pages, int numa_node) mi_attr_noexcept {
size_t hsize = 0;
if (numa_node < -1) numa_node = -1; if (numa_node < -1) numa_node = -1;
if (numa_node >= 0) numa_node = numa_node % _mi_os_numa_node_count(); if (numa_node >= 0) numa_node = numa_node % _mi_os_numa_node_count();
void* p = _mi_os_alloc_huge_os_pages(pages, numa_node, &hsize); size_t hsize = 0;
if (p==NULL) return ENOMEM; size_t pages_reserved = 0;
_mi_verbose_message("reserved %zu huge (1GiB) pages\n", pages); void* p = _mi_os_alloc_huge_os_pages(pages, numa_node, (double)pages / 2.0, &pages_reserved, &hsize);
if (p==NULL || pages_reserved==0) {
_mi_warning_message("failed to reserve %zu gb huge pages\n", pages);
return ENOMEM;
}
_mi_verbose_message("reserved %zu gb huge pages\n", pages_reserved);
size_t bcount = hsize / MI_ARENA_BLOCK_SIZE; size_t bcount = hsize / MI_ARENA_BLOCK_SIZE;
size_t asize = sizeof(mi_arena_t) + (bcount*sizeof(mi_block_info_t)); // one too much size_t asize = sizeof(mi_arena_t) + (bcount*sizeof(mi_block_info_t)); // one too much
mi_arena_t* arena = (mi_arena_t*)_mi_os_alloc(asize, &_mi_stats_main); // TODO: can we avoid allocating from the OS? mi_arena_t* arena = (mi_arena_t*)_mi_os_alloc(asize, &_mi_stats_main); // TODO: can we avoid allocating from the OS?
if (arena == NULL) { if (arena == NULL) {
_mi_os_free(p, hsize, &_mi_stats_main); _mi_os_free_huge_pages(p, hsize, &_mi_stats_main);
return ENOMEM; return ENOMEM;
} }
arena->block_count = bcount; arena->block_count = bcount;

View file

@ -221,7 +221,6 @@ static void mi_add_stderr_output() {
// -------------------------------------------------------- // --------------------------------------------------------
// Messages, all end up calling `_mi_fputs`. // Messages, all end up calling `_mi_fputs`.
// -------------------------------------------------------- // --------------------------------------------------------
#define MAX_ERROR_COUNT (10)
static volatile _Atomic(uintptr_t) error_count; // = 0; // when MAX_ERROR_COUNT stop emitting errors and warnings static volatile _Atomic(uintptr_t) error_count; // = 0; // when MAX_ERROR_COUNT stop emitting errors and warnings
// When overriding malloc, we may recurse into mi_vfprintf if an allocation // When overriding malloc, we may recurse into mi_vfprintf if an allocation

120
src/os.c
View file

@ -339,7 +339,8 @@ static void* mi_unix_mmap(void* addr, size_t size, size_t try_alignment, int pro
lflags |= MAP_HUGETLB; lflags |= MAP_HUGETLB;
#endif #endif
#ifdef MAP_HUGE_1GB #ifdef MAP_HUGE_1GB
if ((size % GiB) == 0) { static bool mi_huge_pages_available = true;
if ((size % GiB) == 0 && mi_huge_pages_available) {
lflags |= MAP_HUGE_1GB; lflags |= MAP_HUGE_1GB;
} }
else else
@ -358,6 +359,7 @@ static void* mi_unix_mmap(void* addr, size_t size, size_t try_alignment, int pro
p = mi_unix_mmapx(addr, size, try_alignment, protect_flags, lflags, lfd); p = mi_unix_mmapx(addr, size, try_alignment, protect_flags, lflags, lfd);
#ifdef MAP_HUGE_1GB #ifdef MAP_HUGE_1GB
if (p == NULL && (lflags & MAP_HUGE_1GB) != 0) { if (p == NULL && (lflags & MAP_HUGE_1GB) != 0) {
mi_huge_pages_available = false; // don't try huge 1GiB pages again
_mi_warning_message("unable to allocate huge (1GiB) page, trying large (2MiB) pages instead (error %i)\n", errno); _mi_warning_message("unable to allocate huge (1GiB) page, trying large (2MiB) pages instead (error %i)\n", errno);
lflags = ((lflags & ~MAP_HUGE_1GB) | MAP_HUGE_2MB); lflags = ((lflags & ~MAP_HUGE_1GB) | MAP_HUGE_2MB);
p = mi_unix_mmapx(addr, size, try_alignment, protect_flags, lflags, lfd); p = mi_unix_mmapx(addr, size, try_alignment, protect_flags, lflags, lfd);
@ -799,11 +801,11 @@ static void* mi_os_alloc_huge_os_pagesx(void* addr, size_t size, int numa_node)
mi_win_enable_large_os_pages(); mi_win_enable_large_os_pages();
void* p = NULL;
#if defined(MEM_EXTENDED_PARAMETER_TYPE_BITS) #if defined(MEM_EXTENDED_PARAMETER_TYPE_BITS)
MEM_EXTENDED_PARAMETER params[3] = { {0,0},{0,0},{0,0} }; MEM_EXTENDED_PARAMETER params[3] = { {0,0},{0,0},{0,0} };
// on modern Windows try use NtAllocateVirtualMemoryEx for 1GiB huge pages // on modern Windows try use NtAllocateVirtualMemoryEx for 1GiB huge pages
if (pNtAllocateVirtualMemoryEx != NULL) { static bool mi_huge_pages_available = true;
if (pNtAllocateVirtualMemoryEx != NULL && mi_huge_pages_available) {
#ifndef MEM_EXTENDED_PARAMETER_NONPAGED_HUGE #ifndef MEM_EXTENDED_PARAMETER_NONPAGED_HUGE
#define MEM_EXTENDED_PARAMETER_NONPAGED_HUGE (0x10) #define MEM_EXTENDED_PARAMETER_NONPAGED_HUGE (0x10)
#endif #endif
@ -822,7 +824,8 @@ static void* mi_os_alloc_huge_os_pagesx(void* addr, size_t size, int numa_node)
return base; return base;
} }
else { else {
// fall back to regular huge pages // fall back to regular large pages
mi_huge_pages_available = false; // don't try further huge pages
_mi_warning_message("unable to allocate using huge (1GiB) pages, trying large (2MiB) pages instead (status 0x%lx)\n", err); _mi_warning_message("unable to allocate using huge (1GiB) pages, trying large (2MiB) pages instead (status 0x%lx)\n", err);
} }
} }
@ -830,20 +833,11 @@ static void* mi_os_alloc_huge_os_pagesx(void* addr, size_t size, int numa_node)
if (pVirtualAlloc2 != NULL && numa_node >= 0) { if (pVirtualAlloc2 != NULL && numa_node >= 0) {
params[0].Type = MemExtendedParameterNumaNode; params[0].Type = MemExtendedParameterNumaNode;
params[0].ULong = (unsigned)numa_node; params[0].ULong = (unsigned)numa_node;
p = (*pVirtualAlloc2)(GetCurrentProcess(), addr, size, flags, PAGE_READWRITE, params, 1); return (*pVirtualAlloc2)(GetCurrentProcess(), addr, size, flags, PAGE_READWRITE, params, 1);
} }
else
#endif #endif
// use regular virtual alloc on older windows // otherwise use regular virtual alloc on older windows
{ return VirtualAlloc(addr, size, flags, PAGE_READWRITE);
p = VirtualAlloc(addr, size, flags, PAGE_READWRITE);
}
if (p == NULL) {
DWORD winerr = GetLastError();
_mi_warning_message("failed to allocate huge OS pages (size %zu) (windows error %d%s)\n", size, winerr, (winerr==1450 ? " (insufficient resources)" : ""));
}
return p;
} }
#elif defined(MI_OS_USE_MMAP) && (MI_INTPTR_SIZE >= 8) #elif defined(MI_OS_USE_MMAP) && (MI_INTPTR_SIZE >= 8)
@ -880,44 +874,92 @@ static void* mi_os_alloc_huge_os_pagesx(void* addr, size_t size, int numa_node)
// To ensure proper alignment, use our own area for huge OS pages // To ensure proper alignment, use our own area for huge OS pages
static _Atomic(uintptr_t) mi_huge_start; // = 0 static _Atomic(uintptr_t) mi_huge_start; // = 0
// Allocate MI_SEGMENT_SIZE aligned huge pages // Claim an aligned address range for huge pages
void* _mi_os_alloc_huge_os_pages(size_t pages, int numa_node, size_t* psize) { static uint8_t* mi_os_claim_huge_pages(size_t pages, size_t* total_size) {
if (psize != NULL) *psize = 0; if (total_size != NULL) *total_size = 0;
const size_t size = pages * MI_HUGE_OS_PAGE_SIZE; const size_t size = pages * MI_HUGE_OS_PAGE_SIZE;
// Find a new aligned address for the huge pages
uintptr_t start = 0; uintptr_t start = 0;
uintptr_t end = 0; uintptr_t end = 0;
uintptr_t expected; uintptr_t expected;
do { do {
start = expected = mi_atomic_read_relaxed(&mi_huge_start); start = expected = mi_atomic_read_relaxed(&mi_huge_start);
if (start == 0) { if (start == 0) {
// Initialize the start address after the 32TiB area // Initialize the start address after the 32TiB area
start = ((uintptr_t)32 << 40); // 32TiB virtual start address start = ((uintptr_t)32 << 40); // 32TiB virtual start address
#if (MI_SECURE>0 || MI_DEBUG==0) // security: randomize start of huge pages unless in debug mode #if (MI_SECURE>0 || MI_DEBUG==0) // security: randomize start of huge pages unless in debug mode
uintptr_t r = _mi_random_init((uintptr_t)&_mi_os_alloc_huge_os_pages); uintptr_t r = _mi_random_init((uintptr_t)&mi_os_claim_huge_pages);
start = start + ((uintptr_t)MI_HUGE_OS_PAGE_SIZE * ((r>>17) & 0x3FF)); // (randomly 0-1024)*1GiB == 0 to 1TiB start = start + ((uintptr_t)MI_HUGE_OS_PAGE_SIZE * ((r>>17) & 0x3FF)); // (randomly 0-1024)*1GiB == 0 to 1TiB
#endif #endif
} }
end = start + size; end = start + size;
mi_assert_internal(end % MI_SEGMENT_SIZE == 0); mi_assert_internal(end % MI_SEGMENT_SIZE == 0);
} while (!mi_atomic_cas_strong(&mi_huge_start, end, expected)); } while (!mi_atomic_cas_strong(&mi_huge_start, end, expected));
// And allocate if (total_size != NULL) *total_size = size;
void* p = mi_os_alloc_huge_os_pagesx((void*)start, size, numa_node); return (uint8_t*)start;
if (p == NULL) { }
return NULL;
} // Allocate MI_SEGMENT_SIZE aligned huge pages
_mi_stat_increase(&_mi_stats_main.committed, size); void* _mi_os_alloc_huge_os_pages(size_t pages, int numa_node, double max_secs, size_t* pages_reserved, size_t* psize) {
_mi_stat_increase(&_mi_stats_main.reserved, size); if (psize != NULL) *psize = 0;
if ((uintptr_t)p % MI_SEGMENT_SIZE != 0) { // must be aligned if (pages_reserved != NULL) *pages_reserved = 0;
_mi_warning_message("huge page area was not aligned\n"); size_t size = 0;
_mi_os_free(p,size,&_mi_stats_main); uint8_t* start = mi_os_claim_huge_pages(pages, &size);
return NULL;
}
if (psize != NULL) *psize = size; // Allocate one page at the time but try to place them contiguously
return p; // We allocate one page at the time to be able to abort if it takes too long
// or to at least allocate as many as available on the system.
double start_t = _mi_clock_start();
size_t page;
for (page = 0; page < pages; page++) {
// allocate a page
bool is_large = true;
void* addr = start + (page * MI_HUGE_OS_PAGE_SIZE);
void* p = mi_os_alloc_huge_os_pagesx(addr, MI_HUGE_OS_PAGE_SIZE, numa_node);
// Did we succeed at a contiguous address?
if (p != addr) {
// no success, issue a warning and break
if (p != NULL) {
_mi_warning_message("could not allocate contiguous huge page %zu at 0x%p\n", page, addr);
_mi_os_free(p, MI_HUGE_OS_PAGE_SIZE, &_mi_stats_main);
}
break;
}
// success, record it
_mi_stat_increase(&_mi_stats_main.committed, MI_HUGE_OS_PAGE_SIZE);
_mi_stat_increase(&_mi_stats_main.reserved, MI_HUGE_OS_PAGE_SIZE);
// check for timeout
double elapsed = _mi_clock_end(start_t);
if (page >= 1) {
double estimate = ((elapsed / (double)(page+1)) * (double)pages);
if (estimate > 1.5*max_secs) { // seems like we are going to timeout, break
elapsed = max_secs + 1.0;
}
}
if (elapsed > max_secs) {
_mi_warning_message("huge page allocation timed out\n");
break;
}
}
mi_assert_internal(page*MI_HUGE_OS_PAGE_SIZE <= size);
if (pages_reserved != NULL) *pages_reserved = page;
if (psize != NULL) *psize = page * MI_HUGE_OS_PAGE_SIZE;
return (page == 0 ? NULL : start);
}
// free every huge page in a range individually (as we allocated per page)
// note: needed with VirtualAlloc but could potentially be done in one go on mmap'd systems.
void _mi_os_free_huge_pages(void* p, size_t size, mi_stats_t* stats) {
if (p==NULL || size==0) return;
uint8_t* base = (uint8_t*)p;
while (size >= MI_HUGE_OS_PAGE_SIZE) {
_mi_os_free(base, MI_HUGE_OS_PAGE_SIZE, stats);
size -= MI_HUGE_OS_PAGE_SIZE;
}
} }
/* ---------------------------------------------------------------------------- /* ----------------------------------------------------------------------------