mirror of
https://github.com/jemalloc/jemalloc.git
synced 2026-04-14 22:51:50 +03:00
398 lines
13 KiB
C
398 lines
13 KiB
C
#include "jemalloc/internal/jemalloc_preamble.h"
|
|
#include "jemalloc/internal/jemalloc_internal_includes.h"
|
|
|
|
#include "jemalloc/internal/eset.h"
|
|
|
|
#define ESET_NPSIZES (SC_NPSIZES + 1)
|
|
|
|
static void
|
|
eset_bin_init(eset_bin_t *bin) {
|
|
edata_heap_new(&bin->heap);
|
|
/*
|
|
* heap_min doesn't need initialization; it gets filled in when the bin
|
|
* goes from non-empty to empty.
|
|
*/
|
|
}
|
|
|
|
static void
|
|
eset_bin_stats_init(eset_bin_stats_t *bin_stats) {
|
|
atomic_store_zu(&bin_stats->nextents, 0, ATOMIC_RELAXED);
|
|
atomic_store_zu(&bin_stats->nbytes, 0, ATOMIC_RELAXED);
|
|
}
|
|
|
|
void
|
|
eset_init(eset_t *eset, extent_state_t state) {
|
|
for (unsigned i = 0; i < ESET_NPSIZES; i++) {
|
|
eset_bin_init(&eset->bins[i]);
|
|
eset_bin_stats_init(&eset->bin_stats[i]);
|
|
}
|
|
fb_init(eset->bitmap, ESET_NPSIZES);
|
|
edata_list_inactive_init(&eset->lru);
|
|
eset->state = state;
|
|
}
|
|
|
|
size_t
|
|
eset_npages_get(eset_t *eset) {
|
|
return atomic_load_zu(&eset->npages, ATOMIC_RELAXED);
|
|
}
|
|
|
|
size_t
|
|
eset_nextents_get(eset_t *eset, pszind_t pind) {
|
|
return atomic_load_zu(&eset->bin_stats[pind].nextents, ATOMIC_RELAXED);
|
|
}
|
|
|
|
size_t
|
|
eset_nbytes_get(eset_t *eset, pszind_t pind) {
|
|
return atomic_load_zu(&eset->bin_stats[pind].nbytes, ATOMIC_RELAXED);
|
|
}
|
|
|
|
static void
|
|
eset_stats_add(eset_t *eset, pszind_t pind, size_t sz) {
|
|
size_t cur = atomic_load_zu(
|
|
&eset->bin_stats[pind].nextents, ATOMIC_RELAXED);
|
|
atomic_store_zu(
|
|
&eset->bin_stats[pind].nextents, cur + 1, ATOMIC_RELAXED);
|
|
cur = atomic_load_zu(&eset->bin_stats[pind].nbytes, ATOMIC_RELAXED);
|
|
atomic_store_zu(
|
|
&eset->bin_stats[pind].nbytes, cur + sz, ATOMIC_RELAXED);
|
|
}
|
|
|
|
static void
|
|
eset_stats_sub(eset_t *eset, pszind_t pind, size_t sz) {
|
|
size_t cur = atomic_load_zu(
|
|
&eset->bin_stats[pind].nextents, ATOMIC_RELAXED);
|
|
atomic_store_zu(
|
|
&eset->bin_stats[pind].nextents, cur - 1, ATOMIC_RELAXED);
|
|
cur = atomic_load_zu(&eset->bin_stats[pind].nbytes, ATOMIC_RELAXED);
|
|
atomic_store_zu(
|
|
&eset->bin_stats[pind].nbytes, cur - sz, ATOMIC_RELAXED);
|
|
}
|
|
|
|
void
|
|
eset_insert(eset_t *eset, edata_t *edata) {
|
|
assert(edata_state_get(edata) == eset->state);
|
|
|
|
size_t size = edata_size_get(edata);
|
|
size_t psz = sz_psz_quantize_floor(size);
|
|
pszind_t pind = sz_psz2ind(psz);
|
|
|
|
edata_cmp_summary_t edata_cmp_summary = edata_cmp_summary_get(edata);
|
|
if (edata_heap_empty(&eset->bins[pind].heap)) {
|
|
fb_set(eset->bitmap, ESET_NPSIZES, (size_t)pind);
|
|
/* Only element is automatically the min element. */
|
|
eset->bins[pind].heap_min = edata_cmp_summary;
|
|
} else {
|
|
/*
|
|
* There's already a min element; update the summary if we're
|
|
* about to insert a lower one.
|
|
*/
|
|
if (edata_cmp_summary_comp(
|
|
edata_cmp_summary, eset->bins[pind].heap_min)
|
|
< 0) {
|
|
eset->bins[pind].heap_min = edata_cmp_summary;
|
|
}
|
|
}
|
|
edata_heap_insert(&eset->bins[pind].heap, edata);
|
|
|
|
if (config_stats) {
|
|
eset_stats_add(eset, pind, size);
|
|
}
|
|
|
|
edata_list_inactive_append(&eset->lru, edata);
|
|
size_t npages = size >> LG_PAGE;
|
|
/*
|
|
* All modifications to npages hold the mutex (as asserted above), so we
|
|
* don't need an atomic fetch-add; we can get by with a load followed by
|
|
* a store.
|
|
*/
|
|
size_t cur_eset_npages = atomic_load_zu(&eset->npages, ATOMIC_RELAXED);
|
|
atomic_store_zu(
|
|
&eset->npages, cur_eset_npages + npages, ATOMIC_RELAXED);
|
|
}
|
|
|
|
void
|
|
eset_remove(eset_t *eset, edata_t *edata) {
|
|
assert(edata_state_get(edata) == eset->state
|
|
|| edata_state_in_transition(edata_state_get(edata)));
|
|
|
|
size_t size = edata_size_get(edata);
|
|
size_t psz = sz_psz_quantize_floor(size);
|
|
pszind_t pind = sz_psz2ind(psz);
|
|
if (config_stats) {
|
|
eset_stats_sub(eset, pind, size);
|
|
}
|
|
|
|
edata_cmp_summary_t edata_cmp_summary = edata_cmp_summary_get(edata);
|
|
edata_heap_remove(&eset->bins[pind].heap, edata);
|
|
if (edata_heap_empty(&eset->bins[pind].heap)) {
|
|
fb_unset(eset->bitmap, ESET_NPSIZES, (size_t)pind);
|
|
} else {
|
|
/*
|
|
* This is a little weird; we compare if the summaries are
|
|
* equal, rather than if the edata we removed was the heap
|
|
* minimum. The reason why is that getting the heap minimum
|
|
* can cause a pairing heap merge operation. We can avoid this
|
|
* if we only update the min if it's changed, in which case the
|
|
* summaries of the removed element and the min element should
|
|
* compare equal.
|
|
*/
|
|
if (edata_cmp_summary_comp(
|
|
edata_cmp_summary, eset->bins[pind].heap_min)
|
|
== 0) {
|
|
eset->bins[pind].heap_min = edata_cmp_summary_get(
|
|
edata_heap_first(&eset->bins[pind].heap));
|
|
}
|
|
}
|
|
edata_list_inactive_remove(&eset->lru, edata);
|
|
size_t npages = size >> LG_PAGE;
|
|
/*
|
|
* As in eset_insert, we hold eset->mtx and so don't need atomic
|
|
* operations for updating eset->npages.
|
|
*/
|
|
size_t cur_extents_npages = atomic_load_zu(
|
|
&eset->npages, ATOMIC_RELAXED);
|
|
assert(cur_extents_npages >= npages);
|
|
atomic_store_zu(&eset->npages, cur_extents_npages - (size >> LG_PAGE),
|
|
ATOMIC_RELAXED);
|
|
}
|
|
|
|
static edata_t *
|
|
eset_enumerate_alignment_search(
|
|
eset_t *eset, size_t size, pszind_t bin_ind, size_t alignment) {
|
|
if (edata_heap_empty(&eset->bins[bin_ind].heap)) {
|
|
return NULL;
|
|
}
|
|
|
|
edata_t *edata = NULL;
|
|
edata_heap_enumerate_helper_t helper;
|
|
edata_heap_enumerate_prepare(&eset->bins[bin_ind].heap, &helper,
|
|
ESET_ENUMERATE_MAX_NUM, sizeof(helper.bfs_queue) / sizeof(void *));
|
|
while ((edata = edata_heap_enumerate_next(
|
|
&eset->bins[bin_ind].heap, &helper))
|
|
!= NULL) {
|
|
uintptr_t base = (uintptr_t)edata_base_get(edata);
|
|
size_t candidate_size = edata_size_get(edata);
|
|
if (candidate_size < size) {
|
|
continue;
|
|
}
|
|
|
|
uintptr_t next_align = ALIGNMENT_CEILING(
|
|
(uintptr_t)base, PAGE_CEILING(alignment));
|
|
if (base > next_align || base + candidate_size <= next_align) {
|
|
/* Overflow or not crossing the next alignment. */
|
|
continue;
|
|
}
|
|
|
|
size_t leadsize = next_align - base;
|
|
if (candidate_size - leadsize >= size) {
|
|
return edata;
|
|
}
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static edata_t *
|
|
eset_enumerate_search(eset_t *eset, size_t size, pszind_t bin_ind,
|
|
bool exact_only, edata_cmp_summary_t *ret_summ) {
|
|
if (edata_heap_empty(&eset->bins[bin_ind].heap)) {
|
|
return NULL;
|
|
}
|
|
|
|
edata_t *ret = NULL, *edata = NULL;
|
|
edata_heap_enumerate_helper_t helper;
|
|
edata_heap_enumerate_prepare(&eset->bins[bin_ind].heap, &helper,
|
|
ESET_ENUMERATE_MAX_NUM, sizeof(helper.bfs_queue) / sizeof(void *));
|
|
while ((edata = edata_heap_enumerate_next(
|
|
&eset->bins[bin_ind].heap, &helper))
|
|
!= NULL) {
|
|
if ((!exact_only && edata_size_get(edata) >= size)
|
|
|| (exact_only && edata_size_get(edata) == size)) {
|
|
edata_cmp_summary_t temp_summ = edata_cmp_summary_get(
|
|
edata);
|
|
if (ret == NULL
|
|
|| edata_cmp_summary_comp(temp_summ, *ret_summ)
|
|
< 0) {
|
|
ret = edata;
|
|
*ret_summ = temp_summ;
|
|
}
|
|
}
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
/*
|
|
* Find an extent with size [min_size, max_size) to satisfy the alignment
|
|
* requirement. For each size, try only the first extent in the heap.
|
|
*/
|
|
static edata_t *
|
|
eset_fit_alignment(
|
|
eset_t *eset, size_t min_size, size_t max_size, size_t alignment) {
|
|
pszind_t pind = sz_psz2ind(sz_psz_quantize_ceil(min_size));
|
|
pszind_t pind_max = sz_psz2ind(sz_psz_quantize_ceil(max_size));
|
|
|
|
/* See comments in eset_first_fit for why we enumerate search below. */
|
|
pszind_t pind_prev = sz_psz2ind(sz_psz_quantize_floor(min_size));
|
|
if (sz_large_size_classes_disabled() && pind != pind_prev) {
|
|
edata_t *ret = NULL;
|
|
ret = eset_enumerate_alignment_search(
|
|
eset, min_size, pind_prev, alignment);
|
|
if (ret != NULL) {
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
for (pszind_t i =
|
|
(pszind_t)fb_ffs(eset->bitmap, ESET_NPSIZES, (size_t)pind);
|
|
i < pind_max;
|
|
i = (pszind_t)fb_ffs(eset->bitmap, ESET_NPSIZES, (size_t)i + 1)) {
|
|
assert(i < SC_NPSIZES);
|
|
assert(!edata_heap_empty(&eset->bins[i].heap));
|
|
edata_t *edata = edata_heap_first(&eset->bins[i].heap);
|
|
uintptr_t base = (uintptr_t)edata_base_get(edata);
|
|
size_t candidate_size = edata_size_get(edata);
|
|
assert(candidate_size >= min_size);
|
|
|
|
uintptr_t next_align = ALIGNMENT_CEILING(
|
|
(uintptr_t)base, PAGE_CEILING(alignment));
|
|
if (base > next_align || base + candidate_size <= next_align) {
|
|
/* Overflow or not crossing the next alignment. */
|
|
continue;
|
|
}
|
|
|
|
size_t leadsize = next_align - base;
|
|
if (candidate_size - leadsize >= min_size) {
|
|
return edata;
|
|
}
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
/*
|
|
* Do first-fit extent selection, i.e. select the oldest/lowest extent that is
|
|
* large enough.
|
|
*
|
|
* lg_max_fit is the (log of the) maximum ratio between the requested size and
|
|
* the returned size that we'll allow. This can reduce fragmentation by
|
|
* avoiding reusing and splitting large extents for smaller sizes. In practice,
|
|
* it's set to opt_lg_extent_max_active_fit for the dirty eset and SC_PTR_BITS
|
|
* for others.
|
|
*/
|
|
static edata_t *
|
|
eset_first_fit(
|
|
eset_t *eset, size_t size, bool exact_only, unsigned lg_max_fit) {
|
|
edata_t *ret = NULL;
|
|
edata_cmp_summary_t ret_summ JEMALLOC_CC_SILENCE_INIT({0});
|
|
|
|
pszind_t pind = sz_psz2ind(sz_psz_quantize_ceil(size));
|
|
|
|
if (exact_only) {
|
|
if (sz_large_size_classes_disabled()) {
|
|
pszind_t pind_prev = sz_psz2ind(
|
|
sz_psz_quantize_floor(size));
|
|
return eset_enumerate_search(eset, size, pind_prev,
|
|
/* exact_only */ true, &ret_summ);
|
|
} else {
|
|
return edata_heap_empty(&eset->bins[pind].heap)
|
|
? NULL
|
|
: edata_heap_first(&eset->bins[pind].heap);
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Each element in the eset->bins is a heap corresponding to a size
|
|
* class. When sz_large_size_classes_disabled() is false, all heaps after
|
|
* pind (including pind itself) will surely satisfy the rquests while
|
|
* heaps before pind cannot satisfy the request because usize is
|
|
* calculated based on size classes then. However, when
|
|
* sz_large_size_classes_disabled() is true, usize is calculated by
|
|
* ceiling user requested size to the closest multiple of PAGE. This
|
|
* means in the heap before pind, i.e., pind_prev, there may exist
|
|
* extents able to satisfy the request and we should enumerate the heap
|
|
* when pind_prev != pind.
|
|
*
|
|
* For example, when PAGE=4KB and the user requested size is 1MB + 4KB,
|
|
* usize would be 1.25MB when sz_large_size_classes_disabled() is false.
|
|
* pind points to the heap containing extents ranging in
|
|
* [1.25MB, 1.5MB). Thus, searching starting from pind will not miss
|
|
* any candidates. When sz_large_size_classes_disabled() is true, the
|
|
* usize would be 1MB + 4KB and pind still points to the same heap.
|
|
* In this case, the heap pind_prev points to, which contains extents
|
|
* in the range [1MB, 1.25MB), may contain candidates satisfying the
|
|
* usize and thus should be enumerated.
|
|
*/
|
|
pszind_t pind_prev = sz_psz2ind(sz_psz_quantize_floor(size));
|
|
if (sz_large_size_classes_disabled() && pind != pind_prev) {
|
|
ret = eset_enumerate_search(eset, size, pind_prev,
|
|
/* exact_only */ false, &ret_summ);
|
|
}
|
|
|
|
for (pszind_t i =
|
|
(pszind_t)fb_ffs(eset->bitmap, ESET_NPSIZES, (size_t)pind);
|
|
i < ESET_NPSIZES;
|
|
i = (pszind_t)fb_ffs(eset->bitmap, ESET_NPSIZES, (size_t)i + 1)) {
|
|
assert(!edata_heap_empty(&eset->bins[i].heap));
|
|
if (lg_max_fit == SC_PTR_BITS) {
|
|
/*
|
|
* We'll shift by this below, and shifting out all the
|
|
* bits is undefined. Decreasing is safe, since the
|
|
* page size is larger than 1 byte.
|
|
*/
|
|
lg_max_fit = SC_PTR_BITS - 1;
|
|
}
|
|
if ((sz_pind2sz(i) >> lg_max_fit) > size) {
|
|
break;
|
|
}
|
|
if (ret == NULL
|
|
|| edata_cmp_summary_comp(eset->bins[i].heap_min, ret_summ)
|
|
< 0) {
|
|
/*
|
|
* We grab the edata as early as possible, even though
|
|
* we might change it later. Practically, a large
|
|
* portion of eset_fit calls succeed at the first valid
|
|
* index, so this doesn't cost much, and we get the
|
|
* effect of prefetching the edata as early as possible.
|
|
*/
|
|
edata_t *edata = edata_heap_first(&eset->bins[i].heap);
|
|
assert(edata_size_get(edata) >= size);
|
|
assert(ret == NULL || edata_snad_comp(edata, ret) < 0);
|
|
assert(ret == NULL
|
|
|| edata_cmp_summary_comp(eset->bins[i].heap_min,
|
|
edata_cmp_summary_get(edata))
|
|
== 0);
|
|
ret = edata;
|
|
ret_summ = eset->bins[i].heap_min;
|
|
}
|
|
if (i == SC_NPSIZES) {
|
|
break;
|
|
}
|
|
assert(i < SC_NPSIZES);
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
edata_t *
|
|
eset_fit(eset_t *eset, size_t esize, size_t alignment, bool exact_only,
|
|
unsigned lg_max_fit) {
|
|
size_t max_size = esize + PAGE_CEILING(alignment) - PAGE;
|
|
/* Beware size_t wrap-around. */
|
|
if (max_size < esize) {
|
|
return NULL;
|
|
}
|
|
|
|
edata_t *edata = eset_first_fit(eset, max_size, exact_only, lg_max_fit);
|
|
|
|
if (alignment > PAGE && edata == NULL) {
|
|
/*
|
|
* max_size guarantees the alignment requirement but is rather
|
|
* pessimistic. Next we try to satisfy the aligned allocation
|
|
* with sizes in [esize, max_size).
|
|
*/
|
|
edata = eset_fit_alignment(eset, esize, max_size, alignment);
|
|
}
|
|
|
|
return edata;
|
|
}
|