mirror of
https://github.com/microsoft/mimalloc.git
synced 2025-05-06 15:29:31 +03:00
807 lines
30 KiB
C
807 lines
30 KiB
C
/* ----------------------------------------------------------------------------
|
|
Copyright (c) 2018-2023, Microsoft Research, Daan Leijen
|
|
This is free software; you can redistribute it and/or modify it under the
|
|
terms of the MIT license. A copy of the license can be found in the file
|
|
"LICENSE" at the root of this distribution.
|
|
-----------------------------------------------------------------------------*/
|
|
#include "mimalloc.h"
|
|
#include "mimalloc/internal.h"
|
|
#include "mimalloc/atomic.h"
|
|
#include "mimalloc/prim.h"
|
|
|
|
/* -----------------------------------------------------------
|
|
Initialization.
|
|
----------------------------------------------------------- */
|
|
#ifndef MI_DEFAULT_PHYSICAL_MEMORY_IN_KIB
|
|
#if MI_INTPTR_SIZE < 8
|
|
#define MI_DEFAULT_PHYSICAL_MEMORY_IN_KIB 4*MI_MiB // 4 GiB
|
|
#else
|
|
#define MI_DEFAULT_PHYSICAL_MEMORY_IN_KIB 32*MI_MiB // 32 GiB
|
|
#endif
|
|
#endif
|
|
|
|
static mi_os_mem_config_t mi_os_mem_config = {
|
|
4096, // page size
|
|
0, // large page size (usually 2MiB)
|
|
4096, // allocation granularity
|
|
MI_DEFAULT_PHYSICAL_MEMORY_IN_KIB,
|
|
MI_MAX_VABITS, // in `bits.h`
|
|
true, // has overcommit? (if true we use MAP_NORESERVE on mmap systems)
|
|
false, // can we partially free allocated blocks? (on mmap systems we can free anywhere in a mapped range, but on Windows we must free the entire span)
|
|
true // has virtual reserve? (if true we can reserve virtual address space without using commit or physical memory)
|
|
};
|
|
|
|
bool _mi_os_has_overcommit(void) {
|
|
return mi_os_mem_config.has_overcommit;
|
|
}
|
|
|
|
bool _mi_os_has_virtual_reserve(void) {
|
|
return mi_os_mem_config.has_virtual_reserve;
|
|
}
|
|
|
|
|
|
// OS (small) page size
|
|
size_t _mi_os_page_size(void) {
|
|
return mi_os_mem_config.page_size;
|
|
}
|
|
|
|
// if large OS pages are supported (2 or 4MiB), then return the size, otherwise return the small page size (4KiB)
|
|
size_t _mi_os_large_page_size(void) {
|
|
return (mi_os_mem_config.large_page_size != 0 ? mi_os_mem_config.large_page_size : _mi_os_page_size());
|
|
}
|
|
|
|
size_t _mi_os_guard_page_size(void) {
|
|
const size_t gsize = _mi_os_page_size();
|
|
mi_assert(gsize <= (MI_ARENA_SLICE_SIZE/8));
|
|
return gsize;
|
|
}
|
|
|
|
size_t _mi_os_virtual_address_bits(void) {
|
|
const size_t vbits = mi_os_mem_config.virtual_address_bits;
|
|
mi_assert(vbits <= MI_MAX_VABITS);
|
|
return vbits;
|
|
}
|
|
|
|
bool _mi_os_use_large_page(size_t size, size_t alignment) {
|
|
// if we have access, check the size and alignment requirements
|
|
if (mi_os_mem_config.large_page_size == 0 || !mi_option_is_enabled(mi_option_allow_large_os_pages)) return false;
|
|
return ((size % mi_os_mem_config.large_page_size) == 0 && (alignment % mi_os_mem_config.large_page_size) == 0);
|
|
}
|
|
|
|
// round to a good OS allocation size (bounded by max 12.5% waste)
|
|
size_t _mi_os_good_alloc_size(size_t size) {
|
|
size_t align_size;
|
|
if (size < 512*MI_KiB) align_size = _mi_os_page_size();
|
|
else if (size < 2*MI_MiB) align_size = 64*MI_KiB;
|
|
else if (size < 8*MI_MiB) align_size = 256*MI_KiB;
|
|
else if (size < 32*MI_MiB) align_size = 1*MI_MiB;
|
|
else align_size = 4*MI_MiB;
|
|
if mi_unlikely(size >= (SIZE_MAX - align_size)) return size; // possible overflow?
|
|
return _mi_align_up(size, align_size);
|
|
}
|
|
|
|
void _mi_os_init(void) {
|
|
_mi_prim_mem_init(&mi_os_mem_config);
|
|
}
|
|
|
|
|
|
/* -----------------------------------------------------------
|
|
Util
|
|
-------------------------------------------------------------- */
|
|
bool _mi_os_decommit(void* addr, size_t size);
|
|
bool _mi_os_commit(void* addr, size_t size, bool* is_zero);
|
|
|
|
void* _mi_os_get_aligned_hint(size_t try_alignment, size_t size) {
|
|
MI_UNUSED(try_alignment); MI_UNUSED(size);
|
|
return NULL;
|
|
}
|
|
|
|
|
|
/* -----------------------------------------------------------
|
|
Guard page allocation
|
|
----------------------------------------------------------- */
|
|
|
|
// In secure mode, return the size of a guard page, otherwise 0
|
|
size_t _mi_os_secure_guard_page_size(void) {
|
|
#if MI_SECURE > 0
|
|
return _mi_os_guard_page_size();
|
|
#else
|
|
return 0;
|
|
#endif
|
|
}
|
|
|
|
// In secure mode, try to decommit an area and output a warning if this fails.
|
|
bool _mi_os_secure_guard_page_set_at(void* addr, mi_memid_t memid) {
|
|
if (addr == NULL) return true;
|
|
#if MI_SECURE > 0
|
|
bool ok = false;
|
|
if (!memid.is_pinned) {
|
|
mi_arena_t* const arena = mi_memid_arena(memid);
|
|
if (arena != NULL && arena->commit_fun != NULL) {
|
|
ok = (*(arena->commit_fun))(false /* decommit */, addr, _mi_os_secure_guard_page_size(), NULL, arena->commit_fun_arg);
|
|
}
|
|
else {
|
|
ok = _mi_os_decommit(addr, _mi_os_secure_guard_page_size());
|
|
}
|
|
}
|
|
if (!ok) {
|
|
_mi_error_message(EINVAL, "secure level %d, but failed to commit guard page (at %p of size %zu)\n", MI_SECURE, addr, _mi_os_secure_guard_page_size());
|
|
}
|
|
return ok;
|
|
#else
|
|
MI_UNUSED(memid);
|
|
return true;
|
|
#endif
|
|
}
|
|
|
|
// In secure mode, try to decommit an area and output a warning if this fails.
|
|
bool _mi_os_secure_guard_page_set_before(void* addr, mi_memid_t memid) {
|
|
return _mi_os_secure_guard_page_set_at((uint8_t*)addr - _mi_os_secure_guard_page_size(), memid);
|
|
}
|
|
|
|
// In secure mode, try to recommit an area
|
|
bool _mi_os_secure_guard_page_reset_at(void* addr, mi_memid_t memid) {
|
|
if (addr == NULL) return true;
|
|
#if MI_SECURE > 0
|
|
if (!memid.is_pinned) {
|
|
mi_arena_t* const arena = mi_memid_arena(memid);
|
|
if (arena != NULL && arena->commit_fun != NULL) {
|
|
return (*(arena->commit_fun))(true, addr, _mi_os_secure_guard_page_size(), NULL, arena->commit_fun_arg);
|
|
}
|
|
else {
|
|
return _mi_os_commit(addr, _mi_os_secure_guard_page_size(), NULL);
|
|
}
|
|
}
|
|
#else
|
|
MI_UNUSED(memid);
|
|
#endif
|
|
return true;
|
|
}
|
|
|
|
// In secure mode, try to recommit an area
|
|
bool _mi_os_secure_guard_page_reset_before(void* addr, mi_memid_t memid) {
|
|
return _mi_os_secure_guard_page_reset_at((uint8_t*)addr - _mi_os_secure_guard_page_size(), memid);
|
|
}
|
|
|
|
|
|
/* -----------------------------------------------------------
|
|
Free memory
|
|
-------------------------------------------------------------- */
|
|
|
|
static void mi_os_free_huge_os_pages(void* p, size_t size);
|
|
|
|
static void mi_os_prim_free(void* addr, size_t size, size_t commit_size) {
|
|
mi_assert_internal((size % _mi_os_page_size()) == 0);
|
|
if (addr == NULL || size == 0) return; // || _mi_os_is_huge_reserved(addr)
|
|
int err = _mi_prim_free(addr, size);
|
|
if (err != 0) {
|
|
_mi_warning_message("unable to free OS memory (error: %d (0x%x), size: 0x%zx bytes, address: %p)\n", err, err, size, addr);
|
|
}
|
|
if (commit_size > 0) {
|
|
mi_os_stat_decrease(committed, commit_size);
|
|
}
|
|
mi_os_stat_decrease(reserved, size);
|
|
}
|
|
|
|
void _mi_os_free_ex(void* addr, size_t size, bool still_committed, mi_memid_t memid) {
|
|
if (mi_memkind_is_os(memid.memkind)) {
|
|
size_t csize = memid.mem.os.size;
|
|
if (csize==0) { _mi_os_good_alloc_size(size); }
|
|
size_t commit_size = (still_committed ? csize : 0);
|
|
void* base = addr;
|
|
// different base? (due to alignment)
|
|
if (memid.mem.os.base != base) {
|
|
mi_assert(memid.mem.os.base <= addr);
|
|
base = memid.mem.os.base;
|
|
const size_t diff = (uint8_t*)addr - (uint8_t*)memid.mem.os.base;
|
|
if (memid.mem.os.size==0) {
|
|
csize += diff;
|
|
}
|
|
if (still_committed) {
|
|
commit_size -= diff; // the (addr-base) part was already un-committed
|
|
}
|
|
}
|
|
// free it
|
|
if (memid.memkind == MI_MEM_OS_HUGE) {
|
|
mi_assert(memid.is_pinned);
|
|
mi_os_free_huge_os_pages(base, csize);
|
|
}
|
|
else {
|
|
mi_os_prim_free(base, csize, (still_committed ? commit_size : 0));
|
|
}
|
|
}
|
|
else {
|
|
// nothing to do
|
|
mi_assert(memid.memkind < MI_MEM_OS);
|
|
}
|
|
}
|
|
|
|
void _mi_os_free(void* p, size_t size, mi_memid_t memid) {
|
|
_mi_os_free_ex(p, size, true, memid);
|
|
}
|
|
|
|
|
|
/* -----------------------------------------------------------
|
|
Primitive allocation from the OS.
|
|
-------------------------------------------------------------- */
|
|
|
|
// Note: the `try_alignment` is just a hint and the returned pointer is not guaranteed to be aligned.
|
|
// Also `hint_addr` is a hint and may be ignored.
|
|
static void* mi_os_prim_alloc_at(void* hint_addr, size_t size, size_t try_alignment, bool commit, bool allow_large, bool* is_large, bool* is_zero) {
|
|
mi_assert_internal(size > 0 && (size % _mi_os_page_size()) == 0);
|
|
mi_assert_internal(is_zero != NULL);
|
|
mi_assert_internal(is_large != NULL);
|
|
if (size == 0) return NULL;
|
|
if (!commit) { allow_large = false; }
|
|
if (try_alignment == 0) { try_alignment = 1; } // avoid 0 to ensure there will be no divide by zero when aligning
|
|
*is_zero = false;
|
|
void* p = NULL;
|
|
int err = _mi_prim_alloc(hint_addr, size, try_alignment, commit, allow_large, is_large, is_zero, &p);
|
|
if (err != 0) {
|
|
_mi_warning_message("unable to allocate OS memory (error: %d (0x%x), addr: %p, size: 0x%zx bytes, align: 0x%zx, commit: %d, allow large: %d)\n", err, err, hint_addr, size, try_alignment, commit, allow_large);
|
|
}
|
|
|
|
mi_os_stat_counter_increase(mmap_calls, 1);
|
|
if (p != NULL) {
|
|
mi_os_stat_increase(reserved, size);
|
|
if (commit) {
|
|
mi_os_stat_increase(committed, size);
|
|
// seems needed for asan (or `mimalloc-test-api` fails)
|
|
#ifdef MI_TRACK_ASAN
|
|
if (*is_zero) { mi_track_mem_defined(p,size); }
|
|
else { mi_track_mem_undefined(p,size); }
|
|
#endif
|
|
}
|
|
}
|
|
return p;
|
|
}
|
|
|
|
static void* mi_os_prim_alloc(size_t size, size_t try_alignment, bool commit, bool allow_large, bool* is_large, bool* is_zero) {
|
|
return mi_os_prim_alloc_at(NULL, size, try_alignment, commit, allow_large, is_large, is_zero);
|
|
}
|
|
|
|
|
|
// Primitive aligned allocation from the OS.
|
|
// This function guarantees the allocated memory is aligned.
|
|
static void* mi_os_prim_alloc_aligned(size_t size, size_t alignment, bool commit, bool allow_large, bool* is_large, bool* is_zero, void** base) {
|
|
mi_assert_internal(alignment >= _mi_os_page_size() && ((alignment & (alignment - 1)) == 0));
|
|
mi_assert_internal(size > 0 && (size % _mi_os_page_size()) == 0);
|
|
mi_assert_internal(is_large != NULL);
|
|
mi_assert_internal(is_zero != NULL);
|
|
mi_assert_internal(base != NULL);
|
|
if (!commit) allow_large = false;
|
|
if (!(alignment >= _mi_os_page_size() && ((alignment & (alignment - 1)) == 0))) return NULL;
|
|
size = _mi_align_up(size, _mi_os_page_size());
|
|
|
|
// try a direct allocation if the alignment is below the default, or if larger than 1/8 fraction of the size.
|
|
const bool try_direct_alloc = (alignment <= mi_os_mem_config.alloc_granularity || alignment > size/8);
|
|
|
|
void* p = NULL;
|
|
if (try_direct_alloc) {
|
|
p = mi_os_prim_alloc(size, alignment, commit, allow_large, is_large, is_zero);
|
|
}
|
|
|
|
// aligned already?
|
|
if (p != NULL && ((uintptr_t)p % alignment) == 0) {
|
|
*base = p;
|
|
}
|
|
else {
|
|
// if not aligned, free it, overallocate, and unmap around it
|
|
#if !MI_TRACK_ASAN
|
|
if (try_direct_alloc) {
|
|
_mi_warning_message("unable to allocate aligned OS memory directly, fall back to over-allocation (size: 0x%zx bytes, address: %p, alignment: 0x%zx, commit: %d)\n", size, p, alignment, commit);
|
|
}
|
|
#endif
|
|
if (p != NULL) { mi_os_prim_free(p, size, (commit ? size : 0)); }
|
|
if (size >= (SIZE_MAX - alignment)) return NULL; // overflow
|
|
const size_t over_size = size + alignment;
|
|
|
|
if (!mi_os_mem_config.has_partial_free) { // win32 virtualAlloc cannot free parts of an allocated block
|
|
// over-allocate uncommitted (virtual) memory
|
|
p = mi_os_prim_alloc(over_size, 1 /*alignment*/, false /* commit? */, false /* allow_large */, is_large, is_zero);
|
|
if (p == NULL) return NULL;
|
|
|
|
// set p to the aligned part in the full region
|
|
// note: on Windows VirtualFree needs the actual base pointer
|
|
// this is handledby having the `base` field in the memid.
|
|
*base = p; // remember the base
|
|
p = _mi_align_up_ptr(p, alignment);
|
|
|
|
// explicitly commit only the aligned part
|
|
if (commit) {
|
|
_mi_os_commit(p, size, NULL);
|
|
}
|
|
}
|
|
else { // mmap can free inside an allocation
|
|
// overallocate...
|
|
p = mi_os_prim_alloc(over_size, 1, commit, false, is_large, is_zero);
|
|
if (p == NULL) return NULL;
|
|
|
|
// and selectively unmap parts around the over-allocated area.
|
|
void* aligned_p = _mi_align_up_ptr(p, alignment);
|
|
size_t pre_size = (uint8_t*)aligned_p - (uint8_t*)p;
|
|
size_t mid_size = _mi_align_up(size, _mi_os_page_size());
|
|
size_t post_size = over_size - pre_size - mid_size;
|
|
mi_assert_internal(pre_size < over_size&& post_size < over_size&& mid_size >= size);
|
|
if (pre_size > 0) { mi_os_prim_free(p, pre_size, (commit ? pre_size : 0)); }
|
|
if (post_size > 0) { mi_os_prim_free((uint8_t*)aligned_p + mid_size, post_size, (commit ? post_size : 0)); }
|
|
// we can return the aligned pointer on `mmap` systems
|
|
p = aligned_p;
|
|
*base = aligned_p; // since we freed the pre part, `*base == p`.
|
|
}
|
|
}
|
|
|
|
mi_assert_internal(p == NULL || (p != NULL && *base != NULL && ((uintptr_t)p % alignment) == 0));
|
|
return p;
|
|
}
|
|
|
|
|
|
/* -----------------------------------------------------------
|
|
OS API: alloc and alloc_aligned
|
|
----------------------------------------------------------- */
|
|
|
|
void* _mi_os_alloc(size_t size, mi_memid_t* memid) {
|
|
*memid = _mi_memid_none();
|
|
if (size == 0) return NULL;
|
|
size = _mi_os_good_alloc_size(size);
|
|
bool os_is_large = false;
|
|
bool os_is_zero = false;
|
|
void* p = mi_os_prim_alloc(size, 0, true, false, &os_is_large, &os_is_zero);
|
|
if (p != NULL) {
|
|
*memid = _mi_memid_create_os(p, size, true, os_is_zero, os_is_large);
|
|
}
|
|
return p;
|
|
}
|
|
|
|
void* _mi_os_alloc_aligned(size_t size, size_t alignment, bool commit, bool allow_large, mi_memid_t* memid)
|
|
{
|
|
MI_UNUSED(&_mi_os_get_aligned_hint); // suppress unused warnings
|
|
*memid = _mi_memid_none();
|
|
if (size == 0) return NULL;
|
|
size = _mi_os_good_alloc_size(size);
|
|
alignment = _mi_align_up(alignment, _mi_os_page_size());
|
|
|
|
bool os_is_large = false;
|
|
bool os_is_zero = false;
|
|
void* os_base = NULL;
|
|
void* p = mi_os_prim_alloc_aligned(size, alignment, commit, allow_large, &os_is_large, &os_is_zero, &os_base);
|
|
if (p != NULL) {
|
|
*memid = _mi_memid_create_os(p, size, commit, os_is_zero, os_is_large);
|
|
memid->mem.os.base = os_base;
|
|
// memid->mem.os.alignment = alignment;
|
|
memid->mem.os.size += ((uint8_t*)p - (uint8_t*)os_base); // todo: return from prim_alloc_aligned
|
|
}
|
|
return p;
|
|
}
|
|
|
|
void* _mi_os_zalloc(size_t size, mi_memid_t* memid) {
|
|
void* p = _mi_os_alloc(size, memid);
|
|
if (p == NULL) return NULL;
|
|
|
|
// zero the OS memory if needed
|
|
if (!memid->initially_zero) {
|
|
_mi_memzero_aligned(p, size);
|
|
memid->initially_zero = true;
|
|
}
|
|
return p;
|
|
}
|
|
|
|
/* -----------------------------------------------------------
|
|
OS aligned allocation with an offset. This is used
|
|
for large alignments > MI_BLOCK_ALIGNMENT_MAX. We use a large mimalloc
|
|
page where the object can be aligned at an offset from the start of the segment.
|
|
As we may need to overallocate, we need to free such pointers using `mi_free_aligned`
|
|
to use the actual start of the memory region.
|
|
----------------------------------------------------------- */
|
|
|
|
void* _mi_os_alloc_aligned_at_offset(size_t size, size_t alignment, size_t offset, bool commit, bool allow_large, mi_memid_t* memid) {
|
|
mi_assert(offset <= size);
|
|
mi_assert((alignment % _mi_os_page_size()) == 0);
|
|
*memid = _mi_memid_none();
|
|
if (offset == 0) {
|
|
// regular aligned allocation
|
|
return _mi_os_alloc_aligned(size, alignment, commit, allow_large, memid);
|
|
}
|
|
else {
|
|
// overallocate to align at an offset
|
|
const size_t extra = _mi_align_up(offset, alignment) - offset;
|
|
const size_t oversize = size + extra;
|
|
void* const start = _mi_os_alloc_aligned(oversize, alignment, commit, allow_large, memid);
|
|
if (start == NULL) return NULL;
|
|
|
|
void* const p = (uint8_t*)start + extra;
|
|
mi_assert(_mi_is_aligned((uint8_t*)p + offset, alignment));
|
|
// decommit the overallocation at the start
|
|
if (commit && extra > _mi_os_page_size()) {
|
|
_mi_os_decommit(start, extra);
|
|
}
|
|
return p;
|
|
}
|
|
}
|
|
|
|
/* -----------------------------------------------------------
|
|
OS memory API: reset, commit, decommit, protect, unprotect.
|
|
----------------------------------------------------------- */
|
|
|
|
// OS page align within a given area, either conservative (pages inside the area only),
|
|
// or not (straddling pages outside the area is possible)
|
|
static void* mi_os_page_align_areax(bool conservative, void* addr, size_t size, size_t* newsize) {
|
|
mi_assert(addr != NULL && size > 0);
|
|
if (newsize != NULL) *newsize = 0;
|
|
if (size == 0 || addr == NULL) return NULL;
|
|
|
|
// page align conservatively within the range, or liberally straddling pages outside the range
|
|
void* start = (conservative ? _mi_align_up_ptr(addr, _mi_os_page_size())
|
|
: mi_align_down_ptr(addr, _mi_os_page_size()));
|
|
void* end = (conservative ? mi_align_down_ptr((uint8_t*)addr + size, _mi_os_page_size())
|
|
: _mi_align_up_ptr((uint8_t*)addr + size, _mi_os_page_size()));
|
|
ptrdiff_t diff = (uint8_t*)end - (uint8_t*)start;
|
|
if (diff <= 0) return NULL;
|
|
|
|
mi_assert_internal((conservative && (size_t)diff <= size) || (!conservative && (size_t)diff >= size));
|
|
if (newsize != NULL) *newsize = (size_t)diff;
|
|
return start;
|
|
}
|
|
|
|
static void* mi_os_page_align_area_conservative(void* addr, size_t size, size_t* newsize) {
|
|
return mi_os_page_align_areax(true, addr, size, newsize);
|
|
}
|
|
|
|
bool _mi_os_commit_ex(void* addr, size_t size, bool* is_zero, size_t stat_size) {
|
|
if (is_zero != NULL) { *is_zero = false; }
|
|
mi_os_stat_increase(committed, stat_size); // use size for precise commit vs. decommit
|
|
mi_os_stat_counter_increase(commit_calls, 1);
|
|
|
|
// page align range
|
|
size_t csize;
|
|
void* start = mi_os_page_align_areax(false /* conservative? */, addr, size, &csize);
|
|
if (csize == 0) return true;
|
|
|
|
// commit
|
|
bool os_is_zero = false;
|
|
int err = _mi_prim_commit(start, csize, &os_is_zero);
|
|
if (err != 0) {
|
|
_mi_warning_message("cannot commit OS memory (error: %d (0x%x), address: %p, size: 0x%zx bytes)\n", err, err, start, csize);
|
|
return false;
|
|
}
|
|
if (os_is_zero && is_zero != NULL) {
|
|
*is_zero = true;
|
|
mi_assert_expensive(mi_mem_is_zero(start, csize));
|
|
}
|
|
// note: the following seems required for asan (otherwise `mimalloc-test-stress` fails)
|
|
#ifdef MI_TRACK_ASAN
|
|
if (os_is_zero) { mi_track_mem_defined(start,csize); }
|
|
else { mi_track_mem_undefined(start,csize); }
|
|
#endif
|
|
return true;
|
|
}
|
|
|
|
bool _mi_os_commit(void* addr, size_t size, bool* is_zero) {
|
|
return _mi_os_commit_ex(addr, size, is_zero, size);
|
|
}
|
|
|
|
static bool mi_os_decommit_ex(void* addr, size_t size, bool* needs_recommit, size_t stat_size) {
|
|
mi_assert_internal(needs_recommit!=NULL);
|
|
mi_os_stat_decrease(committed, stat_size);
|
|
|
|
// page align
|
|
size_t csize;
|
|
void* start = mi_os_page_align_area_conservative(addr, size, &csize);
|
|
if (csize == 0) return true;
|
|
|
|
// decommit
|
|
*needs_recommit = true;
|
|
int err = _mi_prim_decommit(start,csize,needs_recommit);
|
|
if (err != 0) {
|
|
_mi_warning_message("cannot decommit OS memory (error: %d (0x%x), address: %p, size: 0x%zx bytes)\n", err, err, start, csize);
|
|
}
|
|
mi_assert_internal(err == 0);
|
|
return (err == 0);
|
|
}
|
|
|
|
bool _mi_os_decommit(void* addr, size_t size) {
|
|
bool needs_recommit;
|
|
return mi_os_decommit_ex(addr, size, &needs_recommit, size);
|
|
}
|
|
|
|
|
|
// Signal to the OS that the address range is no longer in use
|
|
// but may be used later again. This will release physical memory
|
|
// pages and reduce swapping while keeping the memory committed.
|
|
// We page align to a conservative area inside the range to reset.
|
|
bool _mi_os_reset(void* addr, size_t size) {
|
|
// page align conservatively within the range
|
|
size_t csize;
|
|
void* start = mi_os_page_align_area_conservative(addr, size, &csize);
|
|
if (csize == 0) return true; // || _mi_os_is_huge_reserved(addr)
|
|
mi_os_stat_increase(reset, csize);
|
|
mi_os_stat_counter_increase(reset_calls, 1);
|
|
|
|
#if (MI_DEBUG>1) && !MI_SECURE && !MI_TRACK_ENABLED // && !MI_TSAN
|
|
memset(start, 0, csize); // pretend it is eagerly reset
|
|
#endif
|
|
|
|
int err = _mi_prim_reset(start, csize);
|
|
if (err != 0) {
|
|
_mi_warning_message("cannot reset OS memory (error: %d (0x%x), address: %p, size: 0x%zx bytes)\n", err, err, start, csize);
|
|
}
|
|
return (err == 0);
|
|
}
|
|
|
|
|
|
// either resets or decommits memory, returns true if the memory needs
|
|
// to be recommitted if it is to be re-used later on.
|
|
bool _mi_os_purge_ex(void* p, size_t size, bool allow_reset, size_t stat_size, mi_commit_fun_t* commit_fun, void* commit_fun_arg)
|
|
{
|
|
if (mi_option_get(mi_option_purge_delay) < 0) return false; // is purging allowed?
|
|
mi_os_stat_counter_increase(purge_calls, 1);
|
|
mi_os_stat_increase(purged, size);
|
|
|
|
if (commit_fun != NULL) {
|
|
bool decommitted = (*commit_fun)(false, p, size, NULL, commit_fun_arg);
|
|
return decommitted; // needs_recommit?
|
|
}
|
|
else if (mi_option_is_enabled(mi_option_purge_decommits) && // should decommit?
|
|
!_mi_preloading()) // don't decommit during preloading (unsafe)
|
|
{
|
|
bool needs_recommit = true;
|
|
mi_os_decommit_ex(p, size, &needs_recommit, stat_size);
|
|
return needs_recommit;
|
|
}
|
|
else {
|
|
if (allow_reset) { // this can sometimes be not allowed if the range is not fully committed (on Windows, we cannot reset uncommitted memory)
|
|
_mi_os_reset(p, size);
|
|
}
|
|
return false; // needs no recommit
|
|
}
|
|
}
|
|
|
|
// either resets or decommits memory, returns true if the memory needs
|
|
// to be recommitted if it is to be re-used later on.
|
|
bool _mi_os_purge(void* p, size_t size) {
|
|
return _mi_os_purge_ex(p, size, true, size, NULL, NULL);
|
|
}
|
|
|
|
|
|
// Protect a region in memory to be not accessible.
|
|
static bool mi_os_protectx(void* addr, size_t size, bool protect) {
|
|
// page align conservatively within the range
|
|
size_t csize = 0;
|
|
void* start = mi_os_page_align_area_conservative(addr, size, &csize);
|
|
if (csize == 0) return false;
|
|
/*
|
|
if (_mi_os_is_huge_reserved(addr)) {
|
|
_mi_warning_message("cannot mprotect memory allocated in huge OS pages\n");
|
|
}
|
|
*/
|
|
int err = _mi_prim_protect(start,csize,protect);
|
|
if (err != 0) {
|
|
_mi_warning_message("cannot %s OS memory (error: %d (0x%x), address: %p, size: 0x%zx bytes)\n", (protect ? "protect" : "unprotect"), err, err, start, csize);
|
|
}
|
|
return (err == 0);
|
|
}
|
|
|
|
bool _mi_os_protect(void* addr, size_t size) {
|
|
return mi_os_protectx(addr, size, true);
|
|
}
|
|
|
|
bool _mi_os_unprotect(void* addr, size_t size) {
|
|
return mi_os_protectx(addr, size, false);
|
|
}
|
|
|
|
|
|
|
|
/* ----------------------------------------------------------------------------
|
|
Support for allocating huge OS pages (1Gib) that are reserved up-front
|
|
and possibly associated with a specific NUMA node. (use `numa_node>=0`)
|
|
-----------------------------------------------------------------------------*/
|
|
#define MI_HUGE_OS_PAGE_SIZE (MI_GiB)
|
|
|
|
|
|
#if (MI_INTPTR_SIZE >= 8)
|
|
// To ensure proper alignment, use our own area for huge OS pages
|
|
static mi_decl_cache_align _Atomic(uintptr_t) mi_huge_start; // = 0
|
|
|
|
// Claim an aligned address range for huge pages
|
|
static uint8_t* mi_os_claim_huge_pages(size_t pages, size_t* total_size) {
|
|
if (total_size != NULL) *total_size = 0;
|
|
const size_t size = pages * MI_HUGE_OS_PAGE_SIZE;
|
|
|
|
uintptr_t start = 0;
|
|
uintptr_t end = 0;
|
|
uintptr_t huge_start = mi_atomic_load_relaxed(&mi_huge_start);
|
|
do {
|
|
start = huge_start;
|
|
if (start == 0) {
|
|
// Initialize the start address after the 32TiB area
|
|
start = ((uintptr_t)8 << 40); // 8TiB virtual start address
|
|
#if (MI_SECURE>0 || MI_DEBUG==0) // security: randomize start of huge pages unless in debug mode
|
|
uintptr_t r = _mi_heap_random_next(mi_prim_get_default_heap());
|
|
start = start + ((uintptr_t)MI_HUGE_OS_PAGE_SIZE * ((r>>17) & 0x0FFF)); // (randomly 12bits)*1GiB == between 0 to 4TiB
|
|
#endif
|
|
}
|
|
end = start + size;
|
|
} while (!mi_atomic_cas_weak_acq_rel(&mi_huge_start, &huge_start, end));
|
|
|
|
if (total_size != NULL) *total_size = size;
|
|
return (uint8_t*)start;
|
|
}
|
|
#else
|
|
static uint8_t* mi_os_claim_huge_pages(size_t pages, size_t* total_size) {
|
|
MI_UNUSED(pages);
|
|
if (total_size != NULL) *total_size = 0;
|
|
return NULL;
|
|
}
|
|
#endif
|
|
|
|
// Allocate MI_ARENA_SLICE_ALIGN aligned huge pages
|
|
void* _mi_os_alloc_huge_os_pages(size_t pages, int numa_node, mi_msecs_t max_msecs, size_t* pages_reserved, size_t* psize, mi_memid_t* memid) {
|
|
*memid = _mi_memid_none();
|
|
if (psize != NULL) *psize = 0;
|
|
if (pages_reserved != NULL) *pages_reserved = 0;
|
|
size_t size = 0;
|
|
uint8_t* start = mi_os_claim_huge_pages(pages, &size);
|
|
if (start == NULL) return NULL; // or 32-bit systems
|
|
|
|
// Allocate one page at the time but try to place them contiguously
|
|
// We allocate one page at the time to be able to abort if it takes too long
|
|
// or to at least allocate as many as available on the system.
|
|
mi_msecs_t start_t = _mi_clock_start();
|
|
size_t page = 0;
|
|
bool all_zero = true;
|
|
while (page < pages) {
|
|
// allocate a page
|
|
bool is_zero = false;
|
|
void* addr = start + (page * MI_HUGE_OS_PAGE_SIZE);
|
|
void* p = NULL;
|
|
int err = _mi_prim_alloc_huge_os_pages(addr, MI_HUGE_OS_PAGE_SIZE, numa_node, &is_zero, &p);
|
|
if (!is_zero) { all_zero = false; }
|
|
if (err != 0) {
|
|
_mi_warning_message("unable to allocate huge OS page (error: %d (0x%x), address: %p, size: %zx bytes)\n", err, err, addr, MI_HUGE_OS_PAGE_SIZE);
|
|
break;
|
|
}
|
|
|
|
// Did we succeed at a contiguous address?
|
|
if (p != addr) {
|
|
// no success, issue a warning and break
|
|
if (p != NULL) {
|
|
_mi_warning_message("could not allocate contiguous huge OS page %zu at %p\n", page, addr);
|
|
mi_os_prim_free(p, MI_HUGE_OS_PAGE_SIZE, MI_HUGE_OS_PAGE_SIZE);
|
|
}
|
|
break;
|
|
}
|
|
|
|
// success, record it
|
|
page++; // increase before timeout check (see issue #711)
|
|
mi_os_stat_increase(committed, MI_HUGE_OS_PAGE_SIZE);
|
|
mi_os_stat_increase(reserved, MI_HUGE_OS_PAGE_SIZE);
|
|
|
|
// check for timeout
|
|
if (max_msecs > 0) {
|
|
mi_msecs_t elapsed = _mi_clock_end(start_t);
|
|
if (page >= 1) {
|
|
mi_msecs_t estimate = ((elapsed / (page+1)) * pages);
|
|
if (estimate > 2*max_msecs) { // seems like we are going to timeout, break
|
|
elapsed = max_msecs + 1;
|
|
}
|
|
}
|
|
if (elapsed > max_msecs) {
|
|
_mi_warning_message("huge OS page allocation timed out (after allocating %zu page(s))\n", page);
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
mi_assert_internal(page*MI_HUGE_OS_PAGE_SIZE <= size);
|
|
if (pages_reserved != NULL) { *pages_reserved = page; }
|
|
if (psize != NULL) { *psize = page * MI_HUGE_OS_PAGE_SIZE; }
|
|
if (page != 0) {
|
|
mi_assert(start != NULL);
|
|
*memid = _mi_memid_create_os(start, *psize, true /* is committed */, all_zero, true /* is_large */);
|
|
memid->memkind = MI_MEM_OS_HUGE;
|
|
mi_assert(memid->is_pinned);
|
|
#ifdef MI_TRACK_ASAN
|
|
if (all_zero) { mi_track_mem_defined(start,size); }
|
|
#endif
|
|
}
|
|
return (page == 0 ? NULL : start);
|
|
}
|
|
|
|
// free every huge page in a range individually (as we allocated per page)
|
|
// note: needed with VirtualAlloc but could potentially be done in one go on mmap'd systems.
|
|
static void mi_os_free_huge_os_pages(void* p, size_t size) {
|
|
if (p==NULL || size==0) return;
|
|
uint8_t* base = (uint8_t*)p;
|
|
while (size >= MI_HUGE_OS_PAGE_SIZE) {
|
|
mi_os_prim_free(base, MI_HUGE_OS_PAGE_SIZE, MI_HUGE_OS_PAGE_SIZE);
|
|
size -= MI_HUGE_OS_PAGE_SIZE;
|
|
base += MI_HUGE_OS_PAGE_SIZE;
|
|
}
|
|
}
|
|
|
|
/* ----------------------------------------------------------------------------
|
|
Support NUMA aware allocation
|
|
-----------------------------------------------------------------------------*/
|
|
|
|
static _Atomic(int) _mi_numa_node_count; // = 0 // cache the node count
|
|
|
|
int _mi_os_numa_node_count(void) {
|
|
int count = mi_atomic_load_acquire(&_mi_numa_node_count);
|
|
if mi_unlikely(count <= 0) {
|
|
long ncount = mi_option_get(mi_option_use_numa_nodes); // given explicitly?
|
|
if (ncount > 0 && ncount < INT_MAX) {
|
|
count = (int)ncount;
|
|
}
|
|
else {
|
|
const size_t n = _mi_prim_numa_node_count(); // or detect dynamically
|
|
if (n == 0 || n > INT_MAX) { count = 1; }
|
|
else { count = (int)n; }
|
|
}
|
|
mi_atomic_store_release(&_mi_numa_node_count, count); // save it
|
|
_mi_verbose_message("using %zd numa regions\n", count);
|
|
}
|
|
return count;
|
|
}
|
|
|
|
|
|
static int mi_os_numa_node_get(void) {
|
|
int numa_count = _mi_os_numa_node_count();
|
|
if (numa_count<=1) return 0; // optimize on single numa node systems: always node 0
|
|
// never more than the node count and >= 0
|
|
const size_t n = _mi_prim_numa_node();
|
|
int numa_node = (n < INT_MAX ? (int)n : 0);
|
|
if (numa_node >= numa_count) { numa_node = numa_node % numa_count; }
|
|
return numa_node;
|
|
}
|
|
|
|
int _mi_os_numa_node(void) {
|
|
if mi_likely(mi_atomic_load_relaxed(&_mi_numa_node_count) == 1) { return 0; }
|
|
else return mi_os_numa_node_get();
|
|
}
|
|
|
|
|
|
|
|
|
|
/* ----------------------------------------------------------------------------
|
|
Public API
|
|
-----------------------------------------------------------------------------*/
|
|
#if 0
|
|
mi_decl_export void* mi_os_alloc(size_t size, bool commit, size_t* full_size) {
|
|
return mi_os_alloc_aligned(size, mi_os_mem_config.alloc_granularity, commit, NULL, full_size);
|
|
}
|
|
|
|
static void* mi_os_alloc_aligned_ex(size_t size, size_t alignment, bool commit, bool allow_large, bool* is_committed, bool* is_pinned, void** base, size_t* full_size) {
|
|
mi_memid_t memid = _mi_memid_none();
|
|
void* p = _mi_os_alloc_aligned(size, alignment, commit, allow_large, &memid);
|
|
if (p == NULL) return p;
|
|
if (is_committed != NULL) { *is_committed = memid.initially_committed; }
|
|
if (is_pinned != NULL) { *is_pinned = memid.is_pinned; }
|
|
if (base != NULL) { *base = memid.mem.os.base; }
|
|
if (full_size != NULL) { *full_size = memid.mem.os.size; }
|
|
if (!memid.initially_zero && memid.initially_committed) {
|
|
_mi_memzero_aligned(memid.mem.os.base, memid.mem.os.size);
|
|
}
|
|
return p;
|
|
}
|
|
|
|
mi_decl_export void* mi_os_alloc_aligned(size_t size, size_t alignment, bool commit, void** base, size_t* full_size) {
|
|
return mi_os_alloc_aligned_ex(size, alignment, commit, false, NULL, NULL, base, full_size);
|
|
}
|
|
|
|
mi_decl_export void* mi_os_alloc_aligned_allow_large(size_t size, size_t alignment, bool commit, bool* is_committed, bool* is_pinned, void** base, size_t* full_size) {
|
|
return mi_os_alloc_aligned_ex(size, alignment, commit, true, is_committed, is_pinned, base, full_size);
|
|
}
|
|
|
|
mi_decl_export void mi_os_free(void* p, size_t size) {
|
|
if (p==NULL || size == 0) return;
|
|
mi_memid_t memid = _mi_memid_create_os(p, size, true, false, false);
|
|
_mi_os_free(p, size, memid);
|
|
}
|
|
|
|
mi_decl_export void mi_os_commit(void* p, size_t size) {
|
|
_mi_os_commit(p, size, NULL);
|
|
}
|
|
|
|
mi_decl_export void mi_os_decommit(void* p, size_t size) {
|
|
_mi_os_decommit(p, size);
|
|
}
|
|
#endif
|