Path: blob/main/system/lib/mimalloc/src/segment.c
6175 views
/* ----------------------------------------------------------------------------1Copyright (c) 2018-2024, Microsoft Research, Daan Leijen2This is free software; you can redistribute it and/or modify it under the3terms of the MIT license. A copy of the license can be found in the file4"LICENSE" at the root of this distribution.5-----------------------------------------------------------------------------*/6#include "mimalloc.h"7#include "mimalloc/internal.h"8#include "mimalloc/atomic.h"910#include <string.h> // memset11#include <stdio.h>1213// -------------------------------------------------------------------14// Segments15// mimalloc pages reside in segments. See `mi_segment_valid` for invariants.16// -------------------------------------------------------------------171819static void mi_segment_try_purge(mi_segment_t* segment, bool force, mi_stats_t* stats);202122// -------------------------------------------------------------------23// commit mask24// -------------------------------------------------------------------2526static bool mi_commit_mask_all_set(const mi_commit_mask_t* commit, const mi_commit_mask_t* cm) {27for (size_t i = 0; i < MI_COMMIT_MASK_FIELD_COUNT; i++) {28if ((commit->mask[i] & cm->mask[i]) != cm->mask[i]) return false;29}30return true;31}3233static bool mi_commit_mask_any_set(const mi_commit_mask_t* commit, const mi_commit_mask_t* cm) {34for (size_t i = 0; i < MI_COMMIT_MASK_FIELD_COUNT; i++) {35if ((commit->mask[i] & cm->mask[i]) != 0) return true;36}37return false;38}3940static void mi_commit_mask_create_intersect(const mi_commit_mask_t* commit, const mi_commit_mask_t* cm, mi_commit_mask_t* res) {41for (size_t i = 0; i < MI_COMMIT_MASK_FIELD_COUNT; i++) {42res->mask[i] = (commit->mask[i] & cm->mask[i]);43}44}4546static void mi_commit_mask_clear(mi_commit_mask_t* res, const mi_commit_mask_t* cm) {47for (size_t i = 0; i < MI_COMMIT_MASK_FIELD_COUNT; i++) {48res->mask[i] &= ~(cm->mask[i]);49}50}5152static void mi_commit_mask_set(mi_commit_mask_t* res, const mi_commit_mask_t* cm) {53for (size_t i = 0; i < MI_COMMIT_MASK_FIELD_COUNT; i++) {54res->mask[i] |= cm->mask[i];55}56}5758static void mi_commit_mask_create(size_t bitidx, size_t bitcount, mi_commit_mask_t* cm) {59mi_assert_internal(bitidx < MI_COMMIT_MASK_BITS);60mi_assert_internal((bitidx + bitcount) <= MI_COMMIT_MASK_BITS);61if (bitcount == MI_COMMIT_MASK_BITS) {62mi_assert_internal(bitidx==0);63mi_commit_mask_create_full(cm);64}65else if (bitcount == 0) {66mi_commit_mask_create_empty(cm);67}68else {69mi_commit_mask_create_empty(cm);70size_t i = bitidx / MI_COMMIT_MASK_FIELD_BITS;71size_t ofs = bitidx % MI_COMMIT_MASK_FIELD_BITS;72while (bitcount > 0) {73mi_assert_internal(i < MI_COMMIT_MASK_FIELD_COUNT);74size_t avail = MI_COMMIT_MASK_FIELD_BITS - ofs;75size_t count = (bitcount > avail ? avail : bitcount);76size_t mask = (count >= MI_COMMIT_MASK_FIELD_BITS ? ~((size_t)0) : (((size_t)1 << count) - 1) << ofs);77cm->mask[i] = mask;78bitcount -= count;79ofs = 0;80i++;81}82}83}8485size_t _mi_commit_mask_committed_size(const mi_commit_mask_t* cm, size_t total) {86mi_assert_internal((total%MI_COMMIT_MASK_BITS)==0);87size_t count = 0;88for (size_t i = 0; i < MI_COMMIT_MASK_FIELD_COUNT; i++) {89size_t mask = cm->mask[i];90if (~mask == 0) {91count += MI_COMMIT_MASK_FIELD_BITS;92}93else {94for (; mask != 0; mask >>= 1) { // todo: use popcount95if ((mask&1)!=0) count++;96}97}98}99// we use total since for huge segments each commit bit may represent a larger size100return ((total / MI_COMMIT_MASK_BITS) * count);101}102103104size_t _mi_commit_mask_next_run(const mi_commit_mask_t* cm, size_t* idx) {105size_t i = (*idx) / MI_COMMIT_MASK_FIELD_BITS;106size_t ofs = (*idx) % MI_COMMIT_MASK_FIELD_BITS;107size_t mask = 0;108// find first ones109while (i < MI_COMMIT_MASK_FIELD_COUNT) {110mask = cm->mask[i];111mask >>= ofs;112if (mask != 0) {113while ((mask&1) == 0) {114mask >>= 1;115ofs++;116}117break;118}119i++;120ofs = 0;121}122if (i >= MI_COMMIT_MASK_FIELD_COUNT) {123// not found124*idx = MI_COMMIT_MASK_BITS;125return 0;126}127else {128// found, count ones129size_t count = 0;130*idx = (i*MI_COMMIT_MASK_FIELD_BITS) + ofs;131do {132mi_assert_internal(ofs < MI_COMMIT_MASK_FIELD_BITS && (mask&1) == 1);133do {134count++;135mask >>= 1;136} while ((mask&1) == 1);137if ((((*idx + count) % MI_COMMIT_MASK_FIELD_BITS) == 0)) {138i++;139if (i >= MI_COMMIT_MASK_FIELD_COUNT) break;140mask = cm->mask[i];141ofs = 0;142}143} while ((mask&1) == 1);144mi_assert_internal(count > 0);145return count;146}147}148149150/* --------------------------------------------------------------------------------151Segment allocation152-------------------------------------------------------------------------------- */153154155/* -----------------------------------------------------------156Slices157----------------------------------------------------------- */158159160static const mi_slice_t* mi_segment_slices_end(const mi_segment_t* segment) {161return &segment->slices[segment->slice_entries];162}163164static uint8_t* mi_slice_start(const mi_slice_t* slice) {165mi_segment_t* segment = _mi_ptr_segment(slice);166mi_assert_internal(slice >= segment->slices && slice < mi_segment_slices_end(segment));167return ((uint8_t*)segment + ((slice - segment->slices)*MI_SEGMENT_SLICE_SIZE));168}169170171/* -----------------------------------------------------------172Bins173----------------------------------------------------------- */174// Use bit scan forward to quickly find the first zero bit if it is available175176static inline size_t mi_slice_bin8(size_t slice_count) {177if (slice_count<=1) return slice_count;178mi_assert_internal(slice_count <= MI_SLICES_PER_SEGMENT);179slice_count--;180size_t s = mi_bsr(slice_count); // slice_count > 1181if (s <= 2) return slice_count + 1;182size_t bin = ((s << 2) | ((slice_count >> (s - 2))&0x03)) - 4;183return bin;184}185186static inline size_t mi_slice_bin(size_t slice_count) {187mi_assert_internal(slice_count*MI_SEGMENT_SLICE_SIZE <= MI_SEGMENT_SIZE);188mi_assert_internal(mi_slice_bin8(MI_SLICES_PER_SEGMENT) <= MI_SEGMENT_BIN_MAX);189size_t bin = mi_slice_bin8(slice_count);190mi_assert_internal(bin <= MI_SEGMENT_BIN_MAX);191return bin;192}193194static inline size_t mi_slice_index(const mi_slice_t* slice) {195mi_segment_t* segment = _mi_ptr_segment(slice);196ptrdiff_t index = slice - segment->slices;197mi_assert_internal(index >= 0 && index < (ptrdiff_t)segment->slice_entries);198return index;199}200201202/* -----------------------------------------------------------203Slice span queues204----------------------------------------------------------- */205206static void mi_span_queue_push(mi_span_queue_t* sq, mi_slice_t* slice) {207// todo: or push to the end?208mi_assert_internal(slice->prev == NULL && slice->next==NULL);209slice->prev = NULL; // paranoia210slice->next = sq->first;211sq->first = slice;212if (slice->next != NULL) slice->next->prev = slice;213else sq->last = slice;214slice->block_size = 0; // free215}216217static mi_span_queue_t* mi_span_queue_for(size_t slice_count, mi_segments_tld_t* tld) {218size_t bin = mi_slice_bin(slice_count);219mi_span_queue_t* sq = &tld->spans[bin];220mi_assert_internal(sq->slice_count >= slice_count);221return sq;222}223224static void mi_span_queue_delete(mi_span_queue_t* sq, mi_slice_t* slice) {225mi_assert_internal(slice->block_size==0 && slice->slice_count>0 && slice->slice_offset==0);226// should work too if the queue does not contain slice (which can happen during reclaim)227if (slice->prev != NULL) slice->prev->next = slice->next;228if (slice == sq->first) sq->first = slice->next;229if (slice->next != NULL) slice->next->prev = slice->prev;230if (slice == sq->last) sq->last = slice->prev;231slice->prev = NULL;232slice->next = NULL;233slice->block_size = 1; // no more free234}235236237/* -----------------------------------------------------------238Invariant checking239----------------------------------------------------------- */240241static bool mi_slice_is_used(const mi_slice_t* slice) {242return (slice->block_size > 0);243}244245246#if (MI_DEBUG>=3)247static bool mi_span_queue_contains(mi_span_queue_t* sq, mi_slice_t* slice) {248for (mi_slice_t* s = sq->first; s != NULL; s = s->next) {249if (s==slice) return true;250}251return false;252}253254static bool mi_segment_is_valid(mi_segment_t* segment, mi_segments_tld_t* tld) {255mi_assert_internal(segment != NULL);256mi_assert_internal(_mi_ptr_cookie(segment) == segment->cookie);257mi_assert_internal(segment->abandoned <= segment->used);258mi_assert_internal(segment->thread_id == 0 || segment->thread_id == _mi_thread_id());259mi_assert_internal(mi_commit_mask_all_set(&segment->commit_mask, &segment->purge_mask)); // can only decommit committed blocks260//mi_assert_internal(segment->segment_info_size % MI_SEGMENT_SLICE_SIZE == 0);261mi_slice_t* slice = &segment->slices[0];262const mi_slice_t* end = mi_segment_slices_end(segment);263size_t used_count = 0;264mi_span_queue_t* sq;265while(slice < end) {266mi_assert_internal(slice->slice_count > 0);267mi_assert_internal(slice->slice_offset == 0);268size_t index = mi_slice_index(slice);269size_t maxindex = (index + slice->slice_count >= segment->slice_entries ? segment->slice_entries : index + slice->slice_count) - 1;270if (mi_slice_is_used(slice)) { // a page in use, we need at least MAX_SLICE_OFFSET_COUNT valid back offsets271used_count++;272mi_assert_internal(slice->is_huge == (segment->kind == MI_SEGMENT_HUGE));273for (size_t i = 0; i <= MI_MAX_SLICE_OFFSET_COUNT && index + i <= maxindex; i++) {274mi_assert_internal(segment->slices[index + i].slice_offset == i*sizeof(mi_slice_t));275mi_assert_internal(i==0 || segment->slices[index + i].slice_count == 0);276mi_assert_internal(i==0 || segment->slices[index + i].block_size == 1);277}278// and the last entry as well (for coalescing)279const mi_slice_t* last = slice + slice->slice_count - 1;280if (last > slice && last < mi_segment_slices_end(segment)) {281mi_assert_internal(last->slice_offset == (slice->slice_count-1)*sizeof(mi_slice_t));282mi_assert_internal(last->slice_count == 0);283mi_assert_internal(last->block_size == 1);284}285}286else { // free range of slices; only last slice needs a valid back offset287mi_slice_t* last = &segment->slices[maxindex];288if (segment->kind != MI_SEGMENT_HUGE || slice->slice_count <= (segment->slice_entries - segment->segment_info_slices)) {289mi_assert_internal((uint8_t*)slice == (uint8_t*)last - last->slice_offset);290}291mi_assert_internal(slice == last || last->slice_count == 0 );292mi_assert_internal(last->block_size == 0 || (segment->kind==MI_SEGMENT_HUGE && last->block_size==1));293if (segment->kind != MI_SEGMENT_HUGE && segment->thread_id != 0) { // segment is not huge or abandoned294sq = mi_span_queue_for(slice->slice_count,tld);295mi_assert_internal(mi_span_queue_contains(sq,slice));296}297}298slice = &segment->slices[maxindex+1];299}300mi_assert_internal(slice == end);301mi_assert_internal(used_count == segment->used + 1);302return true;303}304#endif305306/* -----------------------------------------------------------307Segment size calculations308----------------------------------------------------------- */309310static size_t mi_segment_info_size(mi_segment_t* segment) {311return segment->segment_info_slices * MI_SEGMENT_SLICE_SIZE;312}313314static uint8_t* _mi_segment_page_start_from_slice(const mi_segment_t* segment, const mi_slice_t* slice, size_t block_size, size_t* page_size)315{316const ptrdiff_t idx = slice - segment->slices;317const size_t psize = (size_t)slice->slice_count * MI_SEGMENT_SLICE_SIZE;318uint8_t* const pstart = (uint8_t*)segment + (idx*MI_SEGMENT_SLICE_SIZE);319// make the start not OS page aligned for smaller blocks to avoid page/cache effects320// note: the offset must always be a block_size multiple since we assume small allocations321// are aligned (see `mi_heap_malloc_aligned`).322size_t start_offset = 0;323if (block_size > 0 && block_size <= MI_MAX_ALIGN_GUARANTEE) {324// for small objects, ensure the page start is aligned with the block size (PR#66 by kickunderscore)325const size_t adjust = block_size - ((uintptr_t)pstart % block_size);326if (adjust < block_size && psize >= block_size + adjust) {327start_offset += adjust;328}329}330if (block_size >= MI_INTPTR_SIZE) {331if (block_size <= 64) { start_offset += 3*block_size; }332else if (block_size <= 512) { start_offset += block_size; }333}334if (page_size != NULL) { *page_size = psize - start_offset; }335return (pstart + start_offset);336}337338// Start of the page available memory; can be used on uninitialized pages339uint8_t* _mi_segment_page_start(const mi_segment_t* segment, const mi_page_t* page, size_t* page_size)340{341const mi_slice_t* slice = mi_page_to_slice((mi_page_t*)page);342uint8_t* p = _mi_segment_page_start_from_slice(segment, slice, mi_page_block_size(page), page_size);343mi_assert_internal(mi_page_block_size(page) > 0 || _mi_ptr_page(p) == page);344mi_assert_internal(_mi_ptr_segment(p) == segment);345return p;346}347348349static size_t mi_segment_calculate_slices(size_t required, size_t* info_slices) {350size_t page_size = _mi_os_page_size();351size_t isize = _mi_align_up(sizeof(mi_segment_t), page_size);352size_t guardsize = 0;353354if (MI_SECURE>0) {355// in secure mode, we set up a protected page in between the segment info356// and the page data (and one at the end of the segment)357guardsize = page_size;358if (required > 0) {359required = _mi_align_up(required, MI_SEGMENT_SLICE_SIZE) + page_size;360}361}362363isize = _mi_align_up(isize + guardsize, MI_SEGMENT_SLICE_SIZE);364if (info_slices != NULL) *info_slices = isize / MI_SEGMENT_SLICE_SIZE;365size_t segment_size = (required==0 ? MI_SEGMENT_SIZE : _mi_align_up( required + isize + guardsize, MI_SEGMENT_SLICE_SIZE) );366mi_assert_internal(segment_size % MI_SEGMENT_SLICE_SIZE == 0);367return (segment_size / MI_SEGMENT_SLICE_SIZE);368}369370371/* ----------------------------------------------------------------------------372Segment caches373We keep a small segment cache per thread to increase local374reuse and avoid setting/clearing guard pages in secure mode.375------------------------------------------------------------------------------- */376377static void mi_segments_track_size(long segment_size, mi_segments_tld_t* tld) {378if (segment_size>=0) _mi_stat_increase(&tld->stats->segments,1);379else _mi_stat_decrease(&tld->stats->segments,1);380tld->count += (segment_size >= 0 ? 1 : -1);381if (tld->count > tld->peak_count) tld->peak_count = tld->count;382tld->current_size += segment_size;383if (tld->current_size > tld->peak_size) tld->peak_size = tld->current_size;384}385386static void mi_segment_os_free(mi_segment_t* segment, mi_segments_tld_t* tld) {387segment->thread_id = 0;388_mi_segment_map_freed_at(segment);389mi_segments_track_size(-((long)mi_segment_size(segment)),tld);390if (segment->was_reclaimed) {391tld->reclaim_count--;392segment->was_reclaimed = false;393}394if (MI_SECURE>0) {395// _mi_os_unprotect(segment, mi_segment_size(segment)); // ensure no more guard pages are set396// unprotect the guard pages; we cannot just unprotect the whole segment size as part may be decommitted397size_t os_pagesize = _mi_os_page_size();398_mi_os_unprotect((uint8_t*)segment + mi_segment_info_size(segment) - os_pagesize, os_pagesize);399uint8_t* end = (uint8_t*)segment + mi_segment_size(segment) - os_pagesize;400_mi_os_unprotect(end, os_pagesize);401}402403// purge delayed decommits now? (no, leave it to the arena)404// mi_segment_try_purge(segment,true,tld->stats);405406const size_t size = mi_segment_size(segment);407const size_t csize = _mi_commit_mask_committed_size(&segment->commit_mask, size);408409_mi_abandoned_await_readers(); // wait until safe to free410_mi_arena_free(segment, mi_segment_size(segment), csize, segment->memid, tld->stats);411}412413/* -----------------------------------------------------------414Commit/Decommit ranges415----------------------------------------------------------- */416417static void mi_segment_commit_mask(mi_segment_t* segment, bool conservative, uint8_t* p, size_t size, uint8_t** start_p, size_t* full_size, mi_commit_mask_t* cm) {418mi_assert_internal(_mi_ptr_segment(p + 1) == segment);419mi_assert_internal(segment->kind != MI_SEGMENT_HUGE);420mi_commit_mask_create_empty(cm);421if (size == 0 || size > MI_SEGMENT_SIZE || segment->kind == MI_SEGMENT_HUGE) return;422const size_t segstart = mi_segment_info_size(segment);423const size_t segsize = mi_segment_size(segment);424if (p >= (uint8_t*)segment + segsize) return;425426size_t pstart = (p - (uint8_t*)segment);427mi_assert_internal(pstart + size <= segsize);428429size_t start;430size_t end;431if (conservative) {432// decommit conservative433start = _mi_align_up(pstart, MI_COMMIT_SIZE);434end = _mi_align_down(pstart + size, MI_COMMIT_SIZE);435mi_assert_internal(start >= segstart);436mi_assert_internal(end <= segsize);437}438else {439// commit liberal440start = _mi_align_down(pstart, MI_MINIMAL_COMMIT_SIZE);441end = _mi_align_up(pstart + size, MI_MINIMAL_COMMIT_SIZE);442}443if (pstart >= segstart && start < segstart) { // note: the mask is also calculated for an initial commit of the info area444start = segstart;445}446if (end > segsize) {447end = segsize;448}449450mi_assert_internal(start <= pstart && (pstart + size) <= end);451mi_assert_internal(start % MI_COMMIT_SIZE==0 && end % MI_COMMIT_SIZE == 0);452*start_p = (uint8_t*)segment + start;453*full_size = (end > start ? end - start : 0);454if (*full_size == 0) return;455456size_t bitidx = start / MI_COMMIT_SIZE;457mi_assert_internal(bitidx < MI_COMMIT_MASK_BITS);458459size_t bitcount = *full_size / MI_COMMIT_SIZE; // can be 0460if (bitidx + bitcount > MI_COMMIT_MASK_BITS) {461_mi_warning_message("commit mask overflow: idx=%zu count=%zu start=%zx end=%zx p=0x%p size=%zu fullsize=%zu\n", bitidx, bitcount, start, end, p, size, *full_size);462}463mi_assert_internal((bitidx + bitcount) <= MI_COMMIT_MASK_BITS);464mi_commit_mask_create(bitidx, bitcount, cm);465}466467static bool mi_segment_commit(mi_segment_t* segment, uint8_t* p, size_t size, mi_stats_t* stats) {468mi_assert_internal(mi_commit_mask_all_set(&segment->commit_mask, &segment->purge_mask));469470// commit liberal471uint8_t* start = NULL;472size_t full_size = 0;473mi_commit_mask_t mask;474mi_segment_commit_mask(segment, false /* conservative? */, p, size, &start, &full_size, &mask);475if (mi_commit_mask_is_empty(&mask) || full_size == 0) return true;476477if (!mi_commit_mask_all_set(&segment->commit_mask, &mask)) {478// committing479bool is_zero = false;480mi_commit_mask_t cmask;481mi_commit_mask_create_intersect(&segment->commit_mask, &mask, &cmask);482_mi_stat_decrease(&_mi_stats_main.committed, _mi_commit_mask_committed_size(&cmask, MI_SEGMENT_SIZE)); // adjust for overlap483if (!_mi_os_commit(start, full_size, &is_zero, stats)) return false;484mi_commit_mask_set(&segment->commit_mask, &mask);485}486487// increase purge expiration when using part of delayed purges -- we assume more allocations are coming soon.488if (mi_commit_mask_any_set(&segment->purge_mask, &mask)) {489segment->purge_expire = _mi_clock_now() + mi_option_get(mi_option_purge_delay);490}491492// always clear any delayed purges in our range (as they are either committed now)493mi_commit_mask_clear(&segment->purge_mask, &mask);494return true;495}496497static bool mi_segment_ensure_committed(mi_segment_t* segment, uint8_t* p, size_t size, mi_stats_t* stats) {498mi_assert_internal(mi_commit_mask_all_set(&segment->commit_mask, &segment->purge_mask));499// note: assumes commit_mask is always full for huge segments as otherwise the commit mask bits can overflow500if (mi_commit_mask_is_full(&segment->commit_mask) && mi_commit_mask_is_empty(&segment->purge_mask)) return true; // fully committed501mi_assert_internal(segment->kind != MI_SEGMENT_HUGE);502return mi_segment_commit(segment, p, size, stats);503}504505static bool mi_segment_purge(mi_segment_t* segment, uint8_t* p, size_t size, mi_stats_t* stats) {506mi_assert_internal(mi_commit_mask_all_set(&segment->commit_mask, &segment->purge_mask));507if (!segment->allow_purge) return true;508509// purge conservative510uint8_t* start = NULL;511size_t full_size = 0;512mi_commit_mask_t mask;513mi_segment_commit_mask(segment, true /* conservative? */, p, size, &start, &full_size, &mask);514if (mi_commit_mask_is_empty(&mask) || full_size==0) return true;515516if (mi_commit_mask_any_set(&segment->commit_mask, &mask)) {517// purging518mi_assert_internal((void*)start != (void*)segment);519mi_assert_internal(segment->allow_decommit);520const bool decommitted = _mi_os_purge(start, full_size, stats); // reset or decommit521if (decommitted) {522mi_commit_mask_t cmask;523mi_commit_mask_create_intersect(&segment->commit_mask, &mask, &cmask);524_mi_stat_increase(&_mi_stats_main.committed, full_size - _mi_commit_mask_committed_size(&cmask, MI_SEGMENT_SIZE)); // adjust for double counting525mi_commit_mask_clear(&segment->commit_mask, &mask);526}527}528529// always clear any scheduled purges in our range530mi_commit_mask_clear(&segment->purge_mask, &mask);531return true;532}533534static void mi_segment_schedule_purge(mi_segment_t* segment, uint8_t* p, size_t size, mi_stats_t* stats) {535if (!segment->allow_purge) return;536537if (mi_option_get(mi_option_purge_delay) == 0) {538mi_segment_purge(segment, p, size, stats);539}540else {541// register for future purge in the purge mask542uint8_t* start = NULL;543size_t full_size = 0;544mi_commit_mask_t mask;545mi_segment_commit_mask(segment, true /*conservative*/, p, size, &start, &full_size, &mask);546if (mi_commit_mask_is_empty(&mask) || full_size==0) return;547548// update delayed commit549mi_assert_internal(segment->purge_expire > 0 || mi_commit_mask_is_empty(&segment->purge_mask));550mi_commit_mask_t cmask;551mi_commit_mask_create_intersect(&segment->commit_mask, &mask, &cmask); // only purge what is committed; span_free may try to decommit more552mi_commit_mask_set(&segment->purge_mask, &cmask);553mi_msecs_t now = _mi_clock_now();554if (segment->purge_expire == 0) {555// no previous purgess, initialize now556segment->purge_expire = now + mi_option_get(mi_option_purge_delay);557}558else if (segment->purge_expire <= now) {559// previous purge mask already expired560if (segment->purge_expire + mi_option_get(mi_option_purge_extend_delay) <= now) {561mi_segment_try_purge(segment, true, stats);562}563else {564segment->purge_expire = now + mi_option_get(mi_option_purge_extend_delay); // (mi_option_get(mi_option_purge_delay) / 8); // wait a tiny bit longer in case there is a series of free's565}566}567else {568// previous purge mask is not yet expired, increase the expiration by a bit.569segment->purge_expire += mi_option_get(mi_option_purge_extend_delay);570}571}572}573574static void mi_segment_try_purge(mi_segment_t* segment, bool force, mi_stats_t* stats) {575if (!segment->allow_purge || segment->purge_expire == 0 || mi_commit_mask_is_empty(&segment->purge_mask)) return;576mi_msecs_t now = _mi_clock_now();577if (!force && now < segment->purge_expire) return;578579mi_commit_mask_t mask = segment->purge_mask;580segment->purge_expire = 0;581mi_commit_mask_create_empty(&segment->purge_mask);582583size_t idx;584size_t count;585mi_commit_mask_foreach(&mask, idx, count) {586// if found, decommit that sequence587if (count > 0) {588uint8_t* p = (uint8_t*)segment + (idx*MI_COMMIT_SIZE);589size_t size = count * MI_COMMIT_SIZE;590mi_segment_purge(segment, p, size, stats);591}592}593mi_commit_mask_foreach_end()594mi_assert_internal(mi_commit_mask_is_empty(&segment->purge_mask));595}596597// called from `mi_heap_collect_ex`598// this can be called per-page so it is important that try_purge has fast exit path599void _mi_segment_collect(mi_segment_t* segment, bool force, mi_segments_tld_t* tld) {600mi_segment_try_purge(segment, force, tld->stats);601}602603/* -----------------------------------------------------------604Span free605----------------------------------------------------------- */606607static bool mi_segment_is_abandoned(mi_segment_t* segment) {608return (mi_atomic_load_relaxed(&segment->thread_id) == 0);609}610611// note: can be called on abandoned segments612static void mi_segment_span_free(mi_segment_t* segment, size_t slice_index, size_t slice_count, bool allow_purge, mi_segments_tld_t* tld) {613mi_assert_internal(slice_index < segment->slice_entries);614mi_span_queue_t* sq = (segment->kind == MI_SEGMENT_HUGE || mi_segment_is_abandoned(segment)615? NULL : mi_span_queue_for(slice_count,tld));616if (slice_count==0) slice_count = 1;617mi_assert_internal(slice_index + slice_count - 1 < segment->slice_entries);618619// set first and last slice (the intermediates can be undetermined)620mi_slice_t* slice = &segment->slices[slice_index];621slice->slice_count = (uint32_t)slice_count;622mi_assert_internal(slice->slice_count == slice_count); // no overflow?623slice->slice_offset = 0;624if (slice_count > 1) {625mi_slice_t* last = slice + slice_count - 1;626mi_slice_t* end = (mi_slice_t*)mi_segment_slices_end(segment);627if (last > end) { last = end; }628last->slice_count = 0;629last->slice_offset = (uint32_t)(sizeof(mi_page_t)*(slice_count - 1));630last->block_size = 0;631}632633// perhaps decommit634if (allow_purge) {635mi_segment_schedule_purge(segment, mi_slice_start(slice), slice_count * MI_SEGMENT_SLICE_SIZE, tld->stats);636}637638// and push it on the free page queue (if it was not a huge page)639if (sq != NULL) mi_span_queue_push( sq, slice );640else slice->block_size = 0; // mark huge page as free anyways641}642643/*644// called from reclaim to add existing free spans645static void mi_segment_span_add_free(mi_slice_t* slice, mi_segments_tld_t* tld) {646mi_segment_t* segment = _mi_ptr_segment(slice);647mi_assert_internal(slice->xblock_size==0 && slice->slice_count>0 && slice->slice_offset==0);648size_t slice_index = mi_slice_index(slice);649mi_segment_span_free(segment,slice_index,slice->slice_count,tld);650}651*/652653static void mi_segment_span_remove_from_queue(mi_slice_t* slice, mi_segments_tld_t* tld) {654mi_assert_internal(slice->slice_count > 0 && slice->slice_offset==0 && slice->block_size==0);655mi_assert_internal(_mi_ptr_segment(slice)->kind != MI_SEGMENT_HUGE);656mi_span_queue_t* sq = mi_span_queue_for(slice->slice_count, tld);657mi_span_queue_delete(sq, slice);658}659660// note: can be called on abandoned segments661static mi_slice_t* mi_segment_span_free_coalesce(mi_slice_t* slice, mi_segments_tld_t* tld) {662mi_assert_internal(slice != NULL && slice->slice_count > 0 && slice->slice_offset == 0);663mi_segment_t* const segment = _mi_ptr_segment(slice);664const bool is_abandoned = (segment->thread_id == 0); // mi_segment_is_abandoned(segment);665666// for huge pages, just mark as free but don't add to the queues667if (segment->kind == MI_SEGMENT_HUGE) {668// issue #691: segment->used can be 0 if the huge page block was freed while abandoned (reclaim will get here in that case)669mi_assert_internal((segment->used==0 && slice->block_size==0) || segment->used == 1); // decreased right after this call in `mi_segment_page_clear`670slice->block_size = 0; // mark as free anyways671// we should mark the last slice `xblock_size=0` now to maintain invariants but we skip it to672// avoid a possible cache miss (and the segment is about to be freed)673return slice;674}675676// otherwise coalesce the span and add to the free span queues677size_t slice_count = slice->slice_count;678mi_slice_t* next = slice + slice->slice_count;679mi_assert_internal(next <= mi_segment_slices_end(segment));680if (next < mi_segment_slices_end(segment) && next->block_size==0) {681// free next block -- remove it from free and merge682mi_assert_internal(next->slice_count > 0 && next->slice_offset==0);683slice_count += next->slice_count; // extend684if (!is_abandoned) { mi_segment_span_remove_from_queue(next, tld); }685}686if (slice > segment->slices) {687mi_slice_t* prev = mi_slice_first(slice - 1);688mi_assert_internal(prev >= segment->slices);689if (prev->block_size==0) {690// free previous slice -- remove it from free and merge691mi_assert_internal(prev->slice_count > 0 && prev->slice_offset==0);692slice_count += prev->slice_count;693if (!is_abandoned) { mi_segment_span_remove_from_queue(prev, tld); }694slice = prev;695}696}697698// and add the new free page699mi_segment_span_free(segment, mi_slice_index(slice), slice_count, true, tld);700return slice;701}702703704705/* -----------------------------------------------------------706Page allocation707----------------------------------------------------------- */708709// Note: may still return NULL if committing the memory failed710static mi_page_t* mi_segment_span_allocate(mi_segment_t* segment, size_t slice_index, size_t slice_count, mi_segments_tld_t* tld) {711mi_assert_internal(slice_index < segment->slice_entries);712mi_slice_t* const slice = &segment->slices[slice_index];713mi_assert_internal(slice->block_size==0 || slice->block_size==1);714715// commit before changing the slice data716if (!mi_segment_ensure_committed(segment, _mi_segment_page_start_from_slice(segment, slice, 0, NULL), slice_count * MI_SEGMENT_SLICE_SIZE, tld->stats)) {717return NULL; // commit failed!718}719720// convert the slices to a page721slice->slice_offset = 0;722slice->slice_count = (uint32_t)slice_count;723mi_assert_internal(slice->slice_count == slice_count);724const size_t bsize = slice_count * MI_SEGMENT_SLICE_SIZE;725slice->block_size = bsize;726mi_page_t* page = mi_slice_to_page(slice);727mi_assert_internal(mi_page_block_size(page) == bsize);728729// set slice back pointers for the first MI_MAX_SLICE_OFFSET_COUNT entries730size_t extra = slice_count-1;731if (extra > MI_MAX_SLICE_OFFSET_COUNT) extra = MI_MAX_SLICE_OFFSET_COUNT;732if (slice_index + extra >= segment->slice_entries) extra = segment->slice_entries - slice_index - 1; // huge objects may have more slices than avaiable entries in the segment->slices733734mi_slice_t* slice_next = slice + 1;735for (size_t i = 1; i <= extra; i++, slice_next++) {736slice_next->slice_offset = (uint32_t)(sizeof(mi_slice_t)*i);737slice_next->slice_count = 0;738slice_next->block_size = 1;739}740741// and also for the last one (if not set already) (the last one is needed for coalescing and for large alignments)742// note: the cast is needed for ubsan since the index can be larger than MI_SLICES_PER_SEGMENT for huge allocations (see #543)743mi_slice_t* last = slice + slice_count - 1;744mi_slice_t* end = (mi_slice_t*)mi_segment_slices_end(segment);745if (last > end) last = end;746if (last > slice) {747last->slice_offset = (uint32_t)(sizeof(mi_slice_t) * (last - slice));748last->slice_count = 0;749last->block_size = 1;750}751752// and initialize the page753page->is_committed = true;754page->is_huge = (segment->kind == MI_SEGMENT_HUGE);755segment->used++;756return page;757}758759static void mi_segment_slice_split(mi_segment_t* segment, mi_slice_t* slice, size_t slice_count, mi_segments_tld_t* tld) {760mi_assert_internal(_mi_ptr_segment(slice) == segment);761mi_assert_internal(slice->slice_count >= slice_count);762mi_assert_internal(slice->block_size > 0); // no more in free queue763if (slice->slice_count <= slice_count) return;764mi_assert_internal(segment->kind != MI_SEGMENT_HUGE);765size_t next_index = mi_slice_index(slice) + slice_count;766size_t next_count = slice->slice_count - slice_count;767mi_segment_span_free(segment, next_index, next_count, false /* don't purge left-over part */, tld);768slice->slice_count = (uint32_t)slice_count;769}770771static mi_page_t* mi_segments_page_find_and_allocate(size_t slice_count, mi_arena_id_t req_arena_id, mi_segments_tld_t* tld) {772mi_assert_internal(slice_count*MI_SEGMENT_SLICE_SIZE <= MI_LARGE_OBJ_SIZE_MAX);773// search from best fit up774mi_span_queue_t* sq = mi_span_queue_for(slice_count, tld);775if (slice_count == 0) slice_count = 1;776while (sq <= &tld->spans[MI_SEGMENT_BIN_MAX]) {777for (mi_slice_t* slice = sq->first; slice != NULL; slice = slice->next) {778if (slice->slice_count >= slice_count) {779// found one780mi_segment_t* segment = _mi_ptr_segment(slice);781if (_mi_arena_memid_is_suitable(segment->memid, req_arena_id)) {782// found a suitable page span783mi_span_queue_delete(sq, slice);784785if (slice->slice_count > slice_count) {786mi_segment_slice_split(segment, slice, slice_count, tld);787}788mi_assert_internal(slice != NULL && slice->slice_count == slice_count && slice->block_size > 0);789mi_page_t* page = mi_segment_span_allocate(segment, mi_slice_index(slice), slice->slice_count, tld);790if (page == NULL) {791// commit failed; return NULL but first restore the slice792mi_segment_span_free_coalesce(slice, tld);793return NULL;794}795return page;796}797}798}799sq++;800}801// could not find a page..802return NULL;803}804805806/* -----------------------------------------------------------807Segment allocation808----------------------------------------------------------- */809810static mi_segment_t* mi_segment_os_alloc( size_t required, size_t page_alignment, bool eager_delayed, mi_arena_id_t req_arena_id,811size_t* psegment_slices, size_t* pinfo_slices,812bool commit, mi_segments_tld_t* tld, mi_os_tld_t* os_tld)813814{815mi_memid_t memid;816bool allow_large = (!eager_delayed && (MI_SECURE == 0)); // only allow large OS pages once we are no longer lazy817size_t align_offset = 0;818size_t alignment = MI_SEGMENT_ALIGN;819820if (page_alignment > 0) {821// mi_assert_internal(huge_page != NULL);822mi_assert_internal(page_alignment >= MI_SEGMENT_ALIGN);823alignment = page_alignment;824const size_t info_size = (*pinfo_slices) * MI_SEGMENT_SLICE_SIZE;825align_offset = _mi_align_up( info_size, MI_SEGMENT_ALIGN );826const size_t extra = align_offset - info_size;827// recalculate due to potential guard pages828*psegment_slices = mi_segment_calculate_slices(required + extra, pinfo_slices);829mi_assert_internal(*psegment_slices > 0 && *psegment_slices <= UINT32_MAX);830}831832const size_t segment_size = (*psegment_slices) * MI_SEGMENT_SLICE_SIZE;833mi_segment_t* segment = (mi_segment_t*)_mi_arena_alloc_aligned(segment_size, alignment, align_offset, commit, allow_large, req_arena_id, &memid, os_tld);834if (segment == NULL) {835return NULL; // failed to allocate836}837838// ensure metadata part of the segment is committed839mi_commit_mask_t commit_mask;840if (memid.initially_committed) {841mi_commit_mask_create_full(&commit_mask);842}843else {844// at least commit the info slices845const size_t commit_needed = _mi_divide_up((*pinfo_slices)*MI_SEGMENT_SLICE_SIZE, MI_COMMIT_SIZE);846mi_assert_internal(commit_needed>0);847mi_commit_mask_create(0, commit_needed, &commit_mask);848mi_assert_internal(commit_needed*MI_COMMIT_SIZE >= (*pinfo_slices)*MI_SEGMENT_SLICE_SIZE);849if (!_mi_os_commit(segment, commit_needed*MI_COMMIT_SIZE, NULL, tld->stats)) {850_mi_arena_free(segment,segment_size,0,memid,tld->stats);851return NULL;852}853}854mi_assert_internal(segment != NULL && (uintptr_t)segment % MI_SEGMENT_SIZE == 0);855856segment->memid = memid;857segment->allow_decommit = !memid.is_pinned;858segment->allow_purge = segment->allow_decommit && (mi_option_get(mi_option_purge_delay) >= 0);859segment->segment_size = segment_size;860segment->commit_mask = commit_mask;861segment->purge_expire = 0;862mi_commit_mask_create_empty(&segment->purge_mask);863864mi_segments_track_size((long)(segment_size), tld);865_mi_segment_map_allocated_at(segment);866return segment;867}868869870// Allocate a segment from the OS aligned to `MI_SEGMENT_SIZE` .871static mi_segment_t* mi_segment_alloc(size_t required, size_t page_alignment, mi_arena_id_t req_arena_id, mi_segments_tld_t* tld, mi_os_tld_t* os_tld, mi_page_t** huge_page)872{873mi_assert_internal((required==0 && huge_page==NULL) || (required>0 && huge_page != NULL));874875// calculate needed sizes first876size_t info_slices;877size_t segment_slices = mi_segment_calculate_slices(required, &info_slices);878mi_assert_internal(segment_slices > 0 && segment_slices <= UINT32_MAX);879880// Commit eagerly only if not the first N lazy segments (to reduce impact of many threads that allocate just a little)881const bool eager_delay = (// !_mi_os_has_overcommit() && // never delay on overcommit systems882_mi_current_thread_count() > 1 && // do not delay for the first N threads883tld->count < (size_t)mi_option_get(mi_option_eager_commit_delay));884const bool eager = !eager_delay && mi_option_is_enabled(mi_option_eager_commit);885bool commit = eager || (required > 0);886887// Allocate the segment from the OS888mi_segment_t* segment = mi_segment_os_alloc(required, page_alignment, eager_delay, req_arena_id,889&segment_slices, &info_slices, commit, tld, os_tld);890if (segment == NULL) return NULL;891892// zero the segment info? -- not always needed as it may be zero initialized from the OS893if (!segment->memid.initially_zero) {894ptrdiff_t ofs = offsetof(mi_segment_t, next);895size_t prefix = offsetof(mi_segment_t, slices) - ofs;896size_t zsize = prefix + (sizeof(mi_slice_t) * (segment_slices + 1)); // one more897_mi_memzero((uint8_t*)segment + ofs, zsize);898}899900// initialize the rest of the segment info901const size_t slice_entries = (segment_slices > MI_SLICES_PER_SEGMENT ? MI_SLICES_PER_SEGMENT : segment_slices);902segment->segment_slices = segment_slices;903segment->segment_info_slices = info_slices;904segment->thread_id = _mi_thread_id();905segment->cookie = _mi_ptr_cookie(segment);906segment->slice_entries = slice_entries;907segment->kind = (required == 0 ? MI_SEGMENT_NORMAL : MI_SEGMENT_HUGE);908909// _mi_memzero(segment->slices, sizeof(mi_slice_t)*(info_slices+1));910_mi_stat_increase(&tld->stats->page_committed, mi_segment_info_size(segment));911912// set up guard pages913size_t guard_slices = 0;914if (MI_SECURE>0) {915// in secure mode, we set up a protected page in between the segment info916// and the page data, and at the end of the segment.917size_t os_pagesize = _mi_os_page_size();918_mi_os_protect((uint8_t*)segment + mi_segment_info_size(segment) - os_pagesize, os_pagesize);919uint8_t* end = (uint8_t*)segment + mi_segment_size(segment) - os_pagesize;920mi_segment_ensure_committed(segment, end, os_pagesize, tld->stats);921_mi_os_protect(end, os_pagesize);922if (slice_entries == segment_slices) segment->slice_entries--; // don't use the last slice :-(923guard_slices = 1;924}925926// reserve first slices for segment info927mi_page_t* page0 = mi_segment_span_allocate(segment, 0, info_slices, tld);928mi_assert_internal(page0!=NULL); if (page0==NULL) return NULL; // cannot fail as we always commit in advance929mi_assert_internal(segment->used == 1);930segment->used = 0; // don't count our internal slices towards usage931932// initialize initial free pages933if (segment->kind == MI_SEGMENT_NORMAL) { // not a huge page934mi_assert_internal(huge_page==NULL);935mi_segment_span_free(segment, info_slices, segment->slice_entries - info_slices, false /* don't purge */, tld);936}937else {938mi_assert_internal(huge_page!=NULL);939mi_assert_internal(mi_commit_mask_is_empty(&segment->purge_mask));940mi_assert_internal(mi_commit_mask_is_full(&segment->commit_mask));941*huge_page = mi_segment_span_allocate(segment, info_slices, segment_slices - info_slices - guard_slices, tld);942mi_assert_internal(*huge_page != NULL); // cannot fail as we commit in advance943}944945mi_assert_expensive(mi_segment_is_valid(segment,tld));946return segment;947}948949950static void mi_segment_free(mi_segment_t* segment, bool force, mi_segments_tld_t* tld) {951MI_UNUSED(force);952mi_assert_internal(segment != NULL);953mi_assert_internal(segment->next == NULL);954mi_assert_internal(segment->used == 0);955956// Remove the free pages957mi_slice_t* slice = &segment->slices[0];958const mi_slice_t* end = mi_segment_slices_end(segment);959#if MI_DEBUG>1960size_t page_count = 0;961#endif962while (slice < end) {963mi_assert_internal(slice->slice_count > 0);964mi_assert_internal(slice->slice_offset == 0);965mi_assert_internal(mi_slice_index(slice)==0 || slice->block_size == 0); // no more used pages ..966if (slice->block_size == 0 && segment->kind != MI_SEGMENT_HUGE) {967mi_segment_span_remove_from_queue(slice, tld);968}969#if MI_DEBUG>1970page_count++;971#endif972slice = slice + slice->slice_count;973}974mi_assert_internal(page_count == 2); // first page is allocated by the segment itself975976// stats977_mi_stat_decrease(&tld->stats->page_committed, mi_segment_info_size(segment));978979// return it to the OS980mi_segment_os_free(segment, tld);981}982983984/* -----------------------------------------------------------985Page Free986----------------------------------------------------------- */987988static void mi_segment_abandon(mi_segment_t* segment, mi_segments_tld_t* tld);989990// note: can be called on abandoned pages991static mi_slice_t* mi_segment_page_clear(mi_page_t* page, mi_segments_tld_t* tld) {992mi_assert_internal(page->block_size > 0);993mi_assert_internal(mi_page_all_free(page));994mi_segment_t* segment = _mi_ptr_segment(page);995mi_assert_internal(segment->used > 0);996997size_t inuse = page->capacity * mi_page_block_size(page);998_mi_stat_decrease(&tld->stats->page_committed, inuse);999_mi_stat_decrease(&tld->stats->pages, 1);10001001// reset the page memory to reduce memory pressure?1002if (segment->allow_decommit && mi_option_is_enabled(mi_option_deprecated_page_reset)) {1003size_t psize;1004uint8_t* start = _mi_segment_page_start(segment, page, &psize);1005_mi_os_reset(start, psize, tld->stats);1006}10071008// zero the page data, but not the segment fields and heap tag1009page->is_zero_init = false;1010uint8_t heap_tag = page->heap_tag;1011ptrdiff_t ofs = offsetof(mi_page_t, capacity);1012_mi_memzero((uint8_t*)page + ofs, sizeof(*page) - ofs);1013page->block_size = 1;1014page->heap_tag = heap_tag;10151016// and free it1017mi_slice_t* slice = mi_segment_span_free_coalesce(mi_page_to_slice(page), tld);1018segment->used--;1019// cannot assert segment valid as it is called during reclaim1020// mi_assert_expensive(mi_segment_is_valid(segment, tld));1021return slice;1022}10231024void _mi_segment_page_free(mi_page_t* page, bool force, mi_segments_tld_t* tld)1025{1026mi_assert(page != NULL);10271028mi_segment_t* segment = _mi_page_segment(page);1029mi_assert_expensive(mi_segment_is_valid(segment,tld));10301031// mark it as free now1032mi_segment_page_clear(page, tld);1033mi_assert_expensive(mi_segment_is_valid(segment, tld));10341035if (segment->used == 0) {1036// no more used pages; remove from the free list and free the segment1037mi_segment_free(segment, force, tld);1038}1039else if (segment->used == segment->abandoned) {1040// only abandoned pages; remove from free list and abandon1041mi_segment_abandon(segment,tld);1042}1043else {1044// perform delayed purges1045mi_segment_try_purge(segment, false /* force? */, tld->stats);1046}1047}104810491050/* -----------------------------------------------------------1051Abandonment10521053When threads terminate, they can leave segments with1054live blocks (reachable through other threads). Such segments1055are "abandoned" and will be reclaimed by other threads to1056reuse their pages and/or free them eventually. The1057`thread_id` of such segments is 0.10581059When a block is freed in an abandoned segment, the segment1060is reclaimed into that thread.10611062Moreover, if threads are looking for a fresh segment, they1063will first consider abondoned segments -- these can be found1064by scanning the arena memory1065(segments outside arena memoryare only reclaimed by a free).1066----------------------------------------------------------- */10671068// legacy: Wait until there are no more pending reads on segments that used to be in the abandoned list1069void _mi_abandoned_await_readers(void) {1070// nothing needed1071}10721073/* -----------------------------------------------------------1074Abandon segment/page1075----------------------------------------------------------- */10761077static void mi_segment_abandon(mi_segment_t* segment, mi_segments_tld_t* tld) {1078mi_assert_internal(segment->used == segment->abandoned);1079mi_assert_internal(segment->used > 0);1080mi_assert_internal(segment->abandoned_visits == 0);1081mi_assert_expensive(mi_segment_is_valid(segment,tld));10821083// remove the free pages from the free page queues1084mi_slice_t* slice = &segment->slices[0];1085const mi_slice_t* end = mi_segment_slices_end(segment);1086while (slice < end) {1087mi_assert_internal(slice->slice_count > 0);1088mi_assert_internal(slice->slice_offset == 0);1089if (slice->block_size == 0) { // a free page1090mi_segment_span_remove_from_queue(slice,tld);1091slice->block_size = 0; // but keep it free1092}1093slice = slice + slice->slice_count;1094}10951096// perform delayed decommits (forcing is much slower on mstress)1097// Only abandoned segments in arena memory can be reclaimed without a free1098// so if a segment is not from an arena we force purge here to be conservative.1099const bool force_purge = (segment->memid.memkind != MI_MEM_ARENA) || mi_option_is_enabled(mi_option_abandoned_page_purge);1100mi_segment_try_purge(segment, force_purge, tld->stats);11011102// all pages in the segment are abandoned; add it to the abandoned list1103_mi_stat_increase(&tld->stats->segments_abandoned, 1);1104mi_segments_track_size(-((long)mi_segment_size(segment)), tld);1105segment->thread_id = 0;1106segment->abandoned_visits = 1; // from 0 to 1 to signify it is abandoned1107if (segment->was_reclaimed) {1108tld->reclaim_count--;1109segment->was_reclaimed = false;1110}1111_mi_arena_segment_mark_abandoned(segment);1112}11131114void _mi_segment_page_abandon(mi_page_t* page, mi_segments_tld_t* tld) {1115mi_assert(page != NULL);1116mi_assert_internal(mi_page_thread_free_flag(page)==MI_NEVER_DELAYED_FREE);1117mi_assert_internal(mi_page_heap(page) == NULL);1118mi_segment_t* segment = _mi_page_segment(page);11191120mi_assert_expensive(mi_segment_is_valid(segment,tld));1121segment->abandoned++;11221123_mi_stat_increase(&tld->stats->pages_abandoned, 1);1124mi_assert_internal(segment->abandoned <= segment->used);1125if (segment->used == segment->abandoned) {1126// all pages are abandoned, abandon the entire segment1127mi_segment_abandon(segment, tld);1128}1129}11301131/* -----------------------------------------------------------1132Reclaim abandoned pages1133----------------------------------------------------------- */11341135static mi_slice_t* mi_slices_start_iterate(mi_segment_t* segment, const mi_slice_t** end) {1136mi_slice_t* slice = &segment->slices[0];1137*end = mi_segment_slices_end(segment);1138mi_assert_internal(slice->slice_count>0 && slice->block_size>0); // segment allocated page1139slice = slice + slice->slice_count; // skip the first segment allocated page1140return slice;1141}11421143// Possibly free pages and check if free space is available1144static bool mi_segment_check_free(mi_segment_t* segment, size_t slices_needed, size_t block_size, mi_segments_tld_t* tld)1145{1146mi_assert_internal(mi_segment_is_abandoned(segment));1147bool has_page = false;11481149// for all slices1150const mi_slice_t* end;1151mi_slice_t* slice = mi_slices_start_iterate(segment, &end);1152while (slice < end) {1153mi_assert_internal(slice->slice_count > 0);1154mi_assert_internal(slice->slice_offset == 0);1155if (mi_slice_is_used(slice)) { // used page1156// ensure used count is up to date and collect potential concurrent frees1157mi_page_t* const page = mi_slice_to_page(slice);1158_mi_page_free_collect(page, false);1159if (mi_page_all_free(page)) {1160// if this page is all free now, free it without adding to any queues (yet)1161mi_assert_internal(page->next == NULL && page->prev==NULL);1162_mi_stat_decrease(&tld->stats->pages_abandoned, 1);1163segment->abandoned--;1164slice = mi_segment_page_clear(page, tld); // re-assign slice due to coalesce!1165mi_assert_internal(!mi_slice_is_used(slice));1166if (slice->slice_count >= slices_needed) {1167has_page = true;1168}1169}1170else if (mi_page_block_size(page) == block_size && mi_page_has_any_available(page)) {1171// a page has available free blocks of the right size1172has_page = true;1173}1174}1175else {1176// empty span1177if (slice->slice_count >= slices_needed) {1178has_page = true;1179}1180}1181slice = slice + slice->slice_count;1182}1183return has_page;1184}11851186// Reclaim an abandoned segment; returns NULL if the segment was freed1187// set `right_page_reclaimed` to `true` if it reclaimed a page of the right `block_size` that was not full.1188static mi_segment_t* mi_segment_reclaim(mi_segment_t* segment, mi_heap_t* heap, size_t requested_block_size, bool* right_page_reclaimed, mi_segments_tld_t* tld) {1189if (right_page_reclaimed != NULL) { *right_page_reclaimed = false; }1190// can be 0 still with abandoned_next, or already a thread id for segments outside an arena that are reclaimed on a free.1191mi_assert_internal(mi_atomic_load_relaxed(&segment->thread_id) == 0 || mi_atomic_load_relaxed(&segment->thread_id) == _mi_thread_id());1192mi_atomic_store_release(&segment->thread_id, _mi_thread_id());1193segment->abandoned_visits = 0;1194segment->was_reclaimed = true;1195tld->reclaim_count++;1196mi_segments_track_size((long)mi_segment_size(segment), tld);1197mi_assert_internal(segment->next == NULL);1198_mi_stat_decrease(&tld->stats->segments_abandoned, 1);11991200// for all slices1201const mi_slice_t* end;1202mi_slice_t* slice = mi_slices_start_iterate(segment, &end);1203while (slice < end) {1204mi_assert_internal(slice->slice_count > 0);1205mi_assert_internal(slice->slice_offset == 0);1206if (mi_slice_is_used(slice)) {1207// in use: reclaim the page in our heap1208mi_page_t* page = mi_slice_to_page(slice);1209mi_assert_internal(page->is_committed);1210mi_assert_internal(mi_page_thread_free_flag(page)==MI_NEVER_DELAYED_FREE);1211mi_assert_internal(mi_page_heap(page) == NULL);1212mi_assert_internal(page->next == NULL && page->prev==NULL);1213_mi_stat_decrease(&tld->stats->pages_abandoned, 1);1214segment->abandoned--;1215// set the heap again and allow heap thread delayed free again.1216mi_heap_t* target_heap = _mi_heap_by_tag(heap, page->heap_tag); // allow custom heaps to separate objects1217if (target_heap == NULL) {1218target_heap = heap;1219_mi_error_message(EINVAL, "page with tag %u cannot be reclaimed by a heap with the same tag (using %u instead)\n", page->heap_tag, heap->tag );1220}1221mi_page_set_heap(page, target_heap);1222_mi_page_use_delayed_free(page, MI_USE_DELAYED_FREE, true); // override never (after heap is set)1223_mi_page_free_collect(page, false); // ensure used count is up to date1224if (mi_page_all_free(page)) {1225// if everything free by now, free the page1226slice = mi_segment_page_clear(page, tld); // set slice again due to coalesceing1227}1228else {1229// otherwise reclaim it into the heap1230_mi_page_reclaim(target_heap, page);1231if (requested_block_size == mi_page_block_size(page) && mi_page_has_any_available(page) && heap == target_heap) {1232if (right_page_reclaimed != NULL) { *right_page_reclaimed = true; }1233}1234}1235}1236else {1237// the span is free, add it to our page queues1238slice = mi_segment_span_free_coalesce(slice, tld); // set slice again due to coalesceing1239}1240mi_assert_internal(slice->slice_count>0 && slice->slice_offset==0);1241slice = slice + slice->slice_count;1242}12431244mi_assert(segment->abandoned == 0);1245mi_assert_expensive(mi_segment_is_valid(segment, tld));1246if (segment->used == 0) { // due to page_clear1247mi_assert_internal(right_page_reclaimed == NULL || !(*right_page_reclaimed));1248mi_segment_free(segment, false, tld);1249return NULL;1250}1251else {1252return segment;1253}1254}12551256// attempt to reclaim a particular segment (called from multi threaded free `alloc.c:mi_free_block_mt`)1257bool _mi_segment_attempt_reclaim(mi_heap_t* heap, mi_segment_t* segment) {1258if (mi_atomic_load_relaxed(&segment->thread_id) != 0) return false; // it is not abandoned1259// don't reclaim more from a free than half the current segments1260// this is to prevent a pure free-ing thread to start owning too many segments1261if (heap->tld->segments.reclaim_count * 2 > heap->tld->segments.count) return false;1262if (_mi_arena_segment_clear_abandoned(segment)) { // atomically unabandon1263mi_segment_t* res = mi_segment_reclaim(segment, heap, 0, NULL, &heap->tld->segments);1264mi_assert_internal(res == segment);1265return (res != NULL);1266}1267return false;1268}12691270void _mi_abandoned_reclaim_all(mi_heap_t* heap, mi_segments_tld_t* tld) {1271mi_segment_t* segment;1272mi_arena_field_cursor_t current; _mi_arena_field_cursor_init(heap, ¤t);1273while ((segment = _mi_arena_segment_clear_abandoned_next(¤t)) != NULL) {1274mi_segment_reclaim(segment, heap, 0, NULL, tld);1275}1276}12771278static long mi_segment_get_reclaim_tries(void) {1279// limit the tries to 10% (default) of the abandoned segments with at least 8 and at most 1024 tries.1280const size_t perc = (size_t)mi_option_get_clamp(mi_option_max_segment_reclaim, 0, 100);1281if (perc <= 0) return 0;1282const size_t total_count = _mi_arena_segment_abandoned_count();1283if (total_count == 0) return 0;1284const size_t relative_count = (total_count > 10000 ? (total_count / 100) * perc : (total_count * perc) / 100); // avoid overflow1285long max_tries = (long)(relative_count <= 1 ? 1 : (relative_count > 1024 ? 1024 : relative_count));1286if (max_tries < 8 && total_count > 8) { max_tries = 8; }1287return max_tries;1288}12891290static mi_segment_t* mi_segment_try_reclaim(mi_heap_t* heap, size_t needed_slices, size_t block_size, bool* reclaimed, mi_segments_tld_t* tld)1291{1292*reclaimed = false;1293long max_tries = mi_segment_get_reclaim_tries();1294if (max_tries <= 0) return NULL;12951296mi_segment_t* segment;1297mi_arena_field_cursor_t current; _mi_arena_field_cursor_init(heap, ¤t);1298while ((max_tries-- > 0) && ((segment = _mi_arena_segment_clear_abandoned_next(¤t)) != NULL))1299{1300segment->abandoned_visits++;1301// todo: should we respect numa affinity for abondoned reclaim? perhaps only for the first visit?1302// todo: an arena exclusive heap will potentially visit many abandoned unsuitable segments and use many tries1303// Perhaps we can skip non-suitable ones in a better way?1304bool is_suitable = _mi_heap_memid_is_suitable(heap, segment->memid);1305bool has_page = mi_segment_check_free(segment,needed_slices,block_size,tld); // try to free up pages (due to concurrent frees)1306if (segment->used == 0) {1307// free the segment (by forced reclaim) to make it available to other threads.1308// note1: we prefer to free a segment as that might lead to reclaiming another1309// segment that is still partially used.1310// note2: we could in principle optimize this by skipping reclaim and directly1311// freeing but that would violate some invariants temporarily)1312mi_segment_reclaim(segment, heap, 0, NULL, tld);1313}1314else if (has_page && is_suitable) {1315// found a large enough free span, or a page of the right block_size with free space1316// we return the result of reclaim (which is usually `segment`) as it might free1317// the segment due to concurrent frees (in which case `NULL` is returned).1318return mi_segment_reclaim(segment, heap, block_size, reclaimed, tld);1319}1320else if (segment->abandoned_visits > 3 && is_suitable) {1321// always reclaim on 3rd visit to limit the abandoned queue length.1322mi_segment_reclaim(segment, heap, 0, NULL, tld);1323}1324else {1325// otherwise, push on the visited list so it gets not looked at too quickly again1326mi_segment_try_purge(segment, false /* true force? */, tld->stats); // force purge if needed as we may not visit soon again1327_mi_arena_segment_mark_abandoned(segment);1328}1329}1330return NULL;1331}133213331334void _mi_abandoned_collect(mi_heap_t* heap, bool force, mi_segments_tld_t* tld)1335{1336mi_segment_t* segment;1337mi_arena_field_cursor_t current; _mi_arena_field_cursor_init(heap, ¤t);1338long max_tries = (force ? (long)_mi_arena_segment_abandoned_count() : 1024); // limit latency1339while ((max_tries-- > 0) && ((segment = _mi_arena_segment_clear_abandoned_next(¤t)) != NULL)) {1340mi_segment_check_free(segment,0,0,tld); // try to free up pages (due to concurrent frees)1341if (segment->used == 0) {1342// free the segment (by forced reclaim) to make it available to other threads.1343// note: we could in principle optimize this by skipping reclaim and directly1344// freeing but that would violate some invariants temporarily)1345mi_segment_reclaim(segment, heap, 0, NULL, tld);1346}1347else {1348// otherwise, purge if needed and push on the visited list1349// note: forced purge can be expensive if many threads are destroyed/created as in mstress.1350mi_segment_try_purge(segment, force, tld->stats);1351_mi_arena_segment_mark_abandoned(segment);1352}1353}1354}13551356/* -----------------------------------------------------------1357Reclaim or allocate1358----------------------------------------------------------- */13591360static mi_segment_t* mi_segment_reclaim_or_alloc(mi_heap_t* heap, size_t needed_slices, size_t block_size, mi_segments_tld_t* tld, mi_os_tld_t* os_tld)1361{1362mi_assert_internal(block_size <= MI_LARGE_OBJ_SIZE_MAX);13631364// 1. try to reclaim an abandoned segment1365bool reclaimed;1366mi_segment_t* segment = mi_segment_try_reclaim(heap, needed_slices, block_size, &reclaimed, tld);1367if (reclaimed) {1368// reclaimed the right page right into the heap1369mi_assert_internal(segment != NULL);1370return NULL; // pretend out-of-memory as the page will be in the page queue of the heap with available blocks1371}1372else if (segment != NULL) {1373// reclaimed a segment with a large enough empty span in it1374return segment;1375}1376// 2. otherwise allocate a fresh segment1377return mi_segment_alloc(0, 0, heap->arena_id, tld, os_tld, NULL);1378}137913801381/* -----------------------------------------------------------1382Page allocation1383----------------------------------------------------------- */13841385static mi_page_t* mi_segments_page_alloc(mi_heap_t* heap, mi_page_kind_t page_kind, size_t required, size_t block_size, mi_segments_tld_t* tld, mi_os_tld_t* os_tld)1386{1387mi_assert_internal(required <= MI_LARGE_OBJ_SIZE_MAX && page_kind <= MI_PAGE_LARGE);13881389// find a free page1390size_t page_size = _mi_align_up(required, (required > MI_MEDIUM_PAGE_SIZE ? MI_MEDIUM_PAGE_SIZE : MI_SEGMENT_SLICE_SIZE));1391size_t slices_needed = page_size / MI_SEGMENT_SLICE_SIZE;1392mi_assert_internal(slices_needed * MI_SEGMENT_SLICE_SIZE == page_size);1393mi_page_t* page = mi_segments_page_find_and_allocate(slices_needed, heap->arena_id, tld); //(required <= MI_SMALL_SIZE_MAX ? 0 : slices_needed), tld);1394if (page==NULL) {1395// no free page, allocate a new segment and try again1396if (mi_segment_reclaim_or_alloc(heap, slices_needed, block_size, tld, os_tld) == NULL) {1397// OOM or reclaimed a good page in the heap1398return NULL;1399}1400else {1401// otherwise try again1402return mi_segments_page_alloc(heap, page_kind, required, block_size, tld, os_tld);1403}1404}1405mi_assert_internal(page != NULL && page->slice_count*MI_SEGMENT_SLICE_SIZE == page_size);1406mi_assert_internal(_mi_ptr_segment(page)->thread_id == _mi_thread_id());1407mi_segment_try_purge(_mi_ptr_segment(page), false, tld->stats);1408return page;1409}1410141114121413/* -----------------------------------------------------------1414Huge page allocation1415----------------------------------------------------------- */14161417static mi_page_t* mi_segment_huge_page_alloc(size_t size, size_t page_alignment, mi_arena_id_t req_arena_id, mi_segments_tld_t* tld, mi_os_tld_t* os_tld)1418{1419mi_page_t* page = NULL;1420mi_segment_t* segment = mi_segment_alloc(size,page_alignment,req_arena_id,tld,os_tld,&page);1421if (segment == NULL || page==NULL) return NULL;1422mi_assert_internal(segment->used==1);1423mi_assert_internal(mi_page_block_size(page) >= size);1424#if MI_HUGE_PAGE_ABANDON1425segment->thread_id = 0; // huge segments are immediately abandoned1426#endif14271428// for huge pages we initialize the block_size as we may1429// overallocate to accommodate large alignments.1430size_t psize;1431uint8_t* start = _mi_segment_page_start(segment, page, &psize);1432page->block_size = psize;1433mi_assert_internal(page->is_huge);14341435// decommit the part of the prefix of a page that will not be used; this can be quite large (close to MI_SEGMENT_SIZE)1436if (page_alignment > 0 && segment->allow_decommit) {1437uint8_t* aligned_p = (uint8_t*)_mi_align_up((uintptr_t)start, page_alignment);1438mi_assert_internal(_mi_is_aligned(aligned_p, page_alignment));1439mi_assert_internal(psize - (aligned_p - start) >= size);1440uint8_t* decommit_start = start + sizeof(mi_block_t); // for the free list1441ptrdiff_t decommit_size = aligned_p - decommit_start;1442_mi_os_reset(decommit_start, decommit_size, &_mi_stats_main); // note: cannot use segment_decommit on huge segments1443}14441445return page;1446}14471448#if MI_HUGE_PAGE_ABANDON1449// free huge block from another thread1450void _mi_segment_huge_page_free(mi_segment_t* segment, mi_page_t* page, mi_block_t* block) {1451// huge page segments are always abandoned and can be freed immediately by any thread1452mi_assert_internal(segment->kind==MI_SEGMENT_HUGE);1453mi_assert_internal(segment == _mi_page_segment(page));1454mi_assert_internal(mi_atomic_load_relaxed(&segment->thread_id)==0);14551456// claim it and free1457mi_heap_t* heap = mi_heap_get_default(); // issue #221; don't use the internal get_default_heap as we need to ensure the thread is initialized.1458// paranoia: if this it the last reference, the cas should always succeed1459size_t expected_tid = 0;1460if (mi_atomic_cas_strong_acq_rel(&segment->thread_id, &expected_tid, heap->thread_id)) {1461mi_block_set_next(page, block, page->free);1462page->free = block;1463page->used--;1464page->is_zero_init = false;1465mi_assert(page->used == 0);1466mi_tld_t* tld = heap->tld;1467_mi_segment_page_free(page, true, &tld->segments);1468}1469#if (MI_DEBUG!=0)1470else {1471mi_assert_internal(false);1472}1473#endif1474}14751476#else1477// reset memory of a huge block from another thread1478void _mi_segment_huge_page_reset(mi_segment_t* segment, mi_page_t* page, mi_block_t* block) {1479MI_UNUSED(page);1480mi_assert_internal(segment->kind == MI_SEGMENT_HUGE);1481mi_assert_internal(segment == _mi_page_segment(page));1482mi_assert_internal(page->used == 1); // this is called just before the free1483mi_assert_internal(page->free == NULL);1484if (segment->allow_decommit) {1485size_t csize = mi_usable_size(block);1486if (csize > sizeof(mi_block_t)) {1487csize = csize - sizeof(mi_block_t);1488uint8_t* p = (uint8_t*)block + sizeof(mi_block_t);1489_mi_os_reset(p, csize, &_mi_stats_main); // note: cannot use segment_decommit on huge segments1490}1491}1492}1493#endif14941495/* -----------------------------------------------------------1496Page allocation and free1497----------------------------------------------------------- */1498mi_page_t* _mi_segment_page_alloc(mi_heap_t* heap, size_t block_size, size_t page_alignment, mi_segments_tld_t* tld, mi_os_tld_t* os_tld) {1499mi_page_t* page;1500if mi_unlikely(page_alignment > MI_BLOCK_ALIGNMENT_MAX) {1501mi_assert_internal(_mi_is_power_of_two(page_alignment));1502mi_assert_internal(page_alignment >= MI_SEGMENT_SIZE);1503if (page_alignment < MI_SEGMENT_SIZE) { page_alignment = MI_SEGMENT_SIZE; }1504page = mi_segment_huge_page_alloc(block_size,page_alignment,heap->arena_id,tld,os_tld);1505}1506else if (block_size <= MI_SMALL_OBJ_SIZE_MAX) {1507page = mi_segments_page_alloc(heap,MI_PAGE_SMALL,block_size,block_size,tld,os_tld);1508}1509else if (block_size <= MI_MEDIUM_OBJ_SIZE_MAX) {1510page = mi_segments_page_alloc(heap,MI_PAGE_MEDIUM,MI_MEDIUM_PAGE_SIZE,block_size,tld, os_tld);1511}1512else if (block_size <= MI_LARGE_OBJ_SIZE_MAX) {1513page = mi_segments_page_alloc(heap,MI_PAGE_LARGE,block_size,block_size,tld, os_tld);1514}1515else {1516page = mi_segment_huge_page_alloc(block_size,page_alignment,heap->arena_id,tld,os_tld);1517}1518mi_assert_internal(page == NULL || _mi_heap_memid_is_suitable(heap, _mi_page_segment(page)->memid));1519mi_assert_expensive(page == NULL || mi_segment_is_valid(_mi_page_segment(page),tld));1520return page;1521}15221523152415251526