2012-04-16 22:30:26 +08:00
|
|
|
#ifndef JEMALLOC_INTERNAL_H
|
2013-12-09 14:28:27 +08:00
|
|
|
#define JEMALLOC_INTERNAL_H
|
2010-01-17 01:53:50 +08:00
|
|
|
|
Refactor to support more varied testing.
Refactor the test harness to support three types of tests:
- unit: White box unit tests. These tests have full access to all
internal jemalloc library symbols. Though in actuality all symbols
are prefixed by jet_, macro-based name mangling abstracts this away
from test code.
- integration: Black box integration tests. These tests link with
the installable shared jemalloc library, and with the exception of
some utility code and configure-generated macro definitions, they have
no access to jemalloc internals.
- stress: Black box stress tests. These tests link with the installable
shared jemalloc library, as well as with an internal allocator with
symbols prefixed by jet_ (same as for unit tests) that can be used to
allocate data structures that are internal to the test code.
Move existing tests into test/{unit,integration}/ as appropriate.
Split out internal parts of jemalloc_defs.h.in and put them in
jemalloc_internal_defs.h.in. This reduces internals exposure to
applications that #include <jemalloc/jemalloc.h>.
Refactor jemalloc.h header generation so that a single header file
results, and the prototypes can be used to generate jet_ prototypes for
tests. Split jemalloc.h.in into multiple parts (jemalloc_defs.h.in,
jemalloc_macros.h.in, jemalloc_protos.h.in, jemalloc_mangle.h.in) and
use a shell script to generate a unified jemalloc.h at configure time.
Change the default private namespace prefix from "" to "je_".
Add missing private namespace mangling.
Remove hard-coded private_namespace.h. Instead generate it and
private_unnamespace.h from private_symbols.txt. Use similar logic for
public symbols, which aids in name mangling for jet_ symbols.
Add test_warn() and test_fail(). Replace existing exit(1) calls with
test_fail() calls.
2013-12-01 07:25:42 +08:00
|
|
|
#include "jemalloc_internal_defs.h"
|
2014-05-28 11:39:13 +08:00
|
|
|
#include "jemalloc/internal/jemalloc_internal_decls.h"
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2012-04-06 04:36:17 +08:00
|
|
|
#ifdef JEMALLOC_UTRACE
|
|
|
|
#include <sys/ktrace.h>
|
|
|
|
#endif
|
|
|
|
|
Refactor to support more varied testing.
Refactor the test harness to support three types of tests:
- unit: White box unit tests. These tests have full access to all
internal jemalloc library symbols. Though in actuality all symbols
are prefixed by jet_, macro-based name mangling abstracts this away
from test code.
- integration: Black box integration tests. These tests link with
the installable shared jemalloc library, and with the exception of
some utility code and configure-generated macro definitions, they have
no access to jemalloc internals.
- stress: Black box stress tests. These tests link with the installable
shared jemalloc library, as well as with an internal allocator with
symbols prefixed by jet_ (same as for unit tests) that can be used to
allocate data structures that are internal to the test code.
Move existing tests into test/{unit,integration}/ as appropriate.
Split out internal parts of jemalloc_defs.h.in and put them in
jemalloc_internal_defs.h.in. This reduces internals exposure to
applications that #include <jemalloc/jemalloc.h>.
Refactor jemalloc.h header generation so that a single header file
results, and the prototypes can be used to generate jet_ prototypes for
tests. Split jemalloc.h.in into multiple parts (jemalloc_defs.h.in,
jemalloc_macros.h.in, jemalloc_protos.h.in, jemalloc_mangle.h.in) and
use a shell script to generate a unified jemalloc.h at configure time.
Change the default private namespace prefix from "" to "je_".
Add missing private namespace mangling.
Remove hard-coded private_namespace.h. Instead generate it and
private_unnamespace.h from private_symbols.txt. Use similar logic for
public symbols, which aids in name mangling for jet_ symbols.
Add test_warn() and test_fail(). Replace existing exit(1) calls with
test_fail() calls.
2013-12-01 07:25:42 +08:00
|
|
|
#define JEMALLOC_NO_DEMANGLE
|
|
|
|
#ifdef JEMALLOC_JET
|
|
|
|
# define JEMALLOC_N(n) jet_##n
|
|
|
|
# include "jemalloc/internal/public_namespace.h"
|
|
|
|
# define JEMALLOC_NO_RENAME
|
|
|
|
# include "../jemalloc@install_suffix@.h"
|
2014-01-17 09:38:01 +08:00
|
|
|
# undef JEMALLOC_NO_RENAME
|
Refactor to support more varied testing.
Refactor the test harness to support three types of tests:
- unit: White box unit tests. These tests have full access to all
internal jemalloc library symbols. Though in actuality all symbols
are prefixed by jet_, macro-based name mangling abstracts this away
from test code.
- integration: Black box integration tests. These tests link with
the installable shared jemalloc library, and with the exception of
some utility code and configure-generated macro definitions, they have
no access to jemalloc internals.
- stress: Black box stress tests. These tests link with the installable
shared jemalloc library, as well as with an internal allocator with
symbols prefixed by jet_ (same as for unit tests) that can be used to
allocate data structures that are internal to the test code.
Move existing tests into test/{unit,integration}/ as appropriate.
Split out internal parts of jemalloc_defs.h.in and put them in
jemalloc_internal_defs.h.in. This reduces internals exposure to
applications that #include <jemalloc/jemalloc.h>.
Refactor jemalloc.h header generation so that a single header file
results, and the prototypes can be used to generate jet_ prototypes for
tests. Split jemalloc.h.in into multiple parts (jemalloc_defs.h.in,
jemalloc_macros.h.in, jemalloc_protos.h.in, jemalloc_mangle.h.in) and
use a shell script to generate a unified jemalloc.h at configure time.
Change the default private namespace prefix from "" to "je_".
Add missing private namespace mangling.
Remove hard-coded private_namespace.h. Instead generate it and
private_unnamespace.h from private_symbols.txt. Use similar logic for
public symbols, which aids in name mangling for jet_ symbols.
Add test_warn() and test_fail(). Replace existing exit(1) calls with
test_fail() calls.
2013-12-01 07:25:42 +08:00
|
|
|
#else
|
|
|
|
# define JEMALLOC_N(n) @private_namespace@##n
|
|
|
|
# include "../jemalloc@install_suffix@.h"
|
|
|
|
#endif
|
2011-07-31 08:58:07 +08:00
|
|
|
#include "jemalloc/internal/private_namespace.h"
|
2011-07-31 07:40:52 +08:00
|
|
|
|
2012-02-11 12:22:09 +08:00
|
|
|
static const bool config_debug =
|
|
|
|
#ifdef JEMALLOC_DEBUG
|
|
|
|
true
|
|
|
|
#else
|
|
|
|
false
|
|
|
|
#endif
|
|
|
|
;
|
2014-04-16 03:09:48 +08:00
|
|
|
static const bool have_dss =
|
2012-02-11 12:22:09 +08:00
|
|
|
#ifdef JEMALLOC_DSS
|
|
|
|
true
|
|
|
|
#else
|
|
|
|
false
|
|
|
|
#endif
|
|
|
|
;
|
|
|
|
static const bool config_fill =
|
|
|
|
#ifdef JEMALLOC_FILL
|
|
|
|
true
|
|
|
|
#else
|
|
|
|
false
|
|
|
|
#endif
|
|
|
|
;
|
|
|
|
static const bool config_lazy_lock =
|
|
|
|
#ifdef JEMALLOC_LAZY_LOCK
|
|
|
|
true
|
|
|
|
#else
|
|
|
|
false
|
|
|
|
#endif
|
|
|
|
;
|
2016-02-08 06:23:22 +08:00
|
|
|
static const char * const config_malloc_conf = JEMALLOC_CONFIG_MALLOC_CONF;
|
2012-02-11 12:22:09 +08:00
|
|
|
static const bool config_prof =
|
|
|
|
#ifdef JEMALLOC_PROF
|
|
|
|
true
|
|
|
|
#else
|
|
|
|
false
|
|
|
|
#endif
|
|
|
|
;
|
|
|
|
static const bool config_prof_libgcc =
|
|
|
|
#ifdef JEMALLOC_PROF_LIBGCC
|
|
|
|
true
|
|
|
|
#else
|
|
|
|
false
|
|
|
|
#endif
|
|
|
|
;
|
|
|
|
static const bool config_prof_libunwind =
|
|
|
|
#ifdef JEMALLOC_PROF_LIBUNWIND
|
|
|
|
true
|
|
|
|
#else
|
|
|
|
false
|
|
|
|
#endif
|
|
|
|
;
|
2015-07-25 09:21:42 +08:00
|
|
|
static const bool maps_coalesce =
|
|
|
|
#ifdef JEMALLOC_MAPS_COALESCE
|
|
|
|
true
|
|
|
|
#else
|
|
|
|
false
|
|
|
|
#endif
|
|
|
|
;
|
2012-04-13 11:20:58 +08:00
|
|
|
static const bool config_munmap =
|
|
|
|
#ifdef JEMALLOC_MUNMAP
|
|
|
|
true
|
|
|
|
#else
|
|
|
|
false
|
|
|
|
#endif
|
|
|
|
;
|
2012-02-11 12:22:09 +08:00
|
|
|
static const bool config_stats =
|
|
|
|
#ifdef JEMALLOC_STATS
|
|
|
|
true
|
|
|
|
#else
|
|
|
|
false
|
|
|
|
#endif
|
|
|
|
;
|
|
|
|
static const bool config_tcache =
|
|
|
|
#ifdef JEMALLOC_TCACHE
|
|
|
|
true
|
|
|
|
#else
|
|
|
|
false
|
|
|
|
#endif
|
|
|
|
;
|
|
|
|
static const bool config_tls =
|
|
|
|
#ifdef JEMALLOC_TLS
|
|
|
|
true
|
|
|
|
#else
|
|
|
|
false
|
|
|
|
#endif
|
|
|
|
;
|
2012-04-06 04:36:17 +08:00
|
|
|
static const bool config_utrace =
|
|
|
|
#ifdef JEMALLOC_UTRACE
|
|
|
|
true
|
|
|
|
#else
|
|
|
|
false
|
|
|
|
#endif
|
|
|
|
;
|
2012-04-06 15:35:09 +08:00
|
|
|
static const bool config_valgrind =
|
|
|
|
#ifdef JEMALLOC_VALGRIND
|
|
|
|
true
|
|
|
|
#else
|
|
|
|
false
|
|
|
|
#endif
|
|
|
|
;
|
2012-02-11 12:22:09 +08:00
|
|
|
static const bool config_xmalloc =
|
|
|
|
#ifdef JEMALLOC_XMALLOC
|
|
|
|
true
|
|
|
|
#else
|
|
|
|
false
|
|
|
|
#endif
|
|
|
|
;
|
2015-03-19 12:06:58 +08:00
|
|
|
static const bool config_ivsalloc =
|
|
|
|
#ifdef JEMALLOC_IVSALLOC
|
|
|
|
true
|
|
|
|
#else
|
|
|
|
false
|
|
|
|
#endif
|
|
|
|
;
|
2015-05-05 00:58:36 +08:00
|
|
|
static const bool config_cache_oblivious =
|
|
|
|
#ifdef JEMALLOC_CACHE_OBLIVIOUS
|
|
|
|
true
|
|
|
|
#else
|
|
|
|
false
|
|
|
|
#endif
|
|
|
|
;
|
2012-02-11 12:22:09 +08:00
|
|
|
|
2014-12-06 09:42:41 +08:00
|
|
|
#ifdef JEMALLOC_C11ATOMICS
|
|
|
|
#include <stdatomic.h>
|
|
|
|
#endif
|
|
|
|
|
2012-04-18 04:17:54 +08:00
|
|
|
#ifdef JEMALLOC_ATOMIC9
|
|
|
|
#include <machine/atomic.h>
|
|
|
|
#endif
|
|
|
|
|
2011-03-19 10:30:18 +08:00
|
|
|
#if (defined(JEMALLOC_OSATOMIC) || defined(JEMALLOC_OSSPIN))
|
2011-03-19 10:10:31 +08:00
|
|
|
#include <libkern/OSAtomic.h>
|
|
|
|
#endif
|
|
|
|
|
2010-09-06 01:35:13 +08:00
|
|
|
#ifdef JEMALLOC_ZONE
|
|
|
|
#include <mach/mach_error.h>
|
|
|
|
#include <mach/mach_init.h>
|
|
|
|
#include <mach/vm_map.h>
|
|
|
|
#include <malloc/malloc.h>
|
|
|
|
#endif
|
|
|
|
|
2010-03-01 07:00:18 +08:00
|
|
|
#define RB_COMPACT
|
2010-02-12 06:45:59 +08:00
|
|
|
#include "jemalloc/internal/rb.h"
|
|
|
|
#include "jemalloc/internal/qr.h"
|
|
|
|
#include "jemalloc/internal/ql.h"
|
2010-01-17 01:53:50 +08:00
|
|
|
|
|
|
|
/*
|
2010-02-12 05:38:12 +08:00
|
|
|
* jemalloc can conceptually be broken into components (arena, tcache, etc.),
|
|
|
|
* but there are circular dependencies that cannot be broken without
|
2010-01-17 01:53:50 +08:00
|
|
|
* substantial performance degradation. In order to reduce the effect on
|
|
|
|
* visual code flow, read the header files in multiple passes, with one of the
|
|
|
|
* following cpp variables defined during each pass:
|
|
|
|
*
|
|
|
|
* JEMALLOC_H_TYPES : Preprocessor-defined constants and psuedo-opaque data
|
|
|
|
* types.
|
|
|
|
* JEMALLOC_H_STRUCTS : Data structures.
|
|
|
|
* JEMALLOC_H_EXTERNS : Extern data declarations and function prototypes.
|
|
|
|
* JEMALLOC_H_INLINES : Inline functions.
|
|
|
|
*/
|
|
|
|
/******************************************************************************/
|
2013-12-09 14:28:27 +08:00
|
|
|
#define JEMALLOC_H_TYPES
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2013-12-06 13:43:46 +08:00
|
|
|
#include "jemalloc/internal/jemalloc_internal_macros.h"
|
Add {,r,s,d}allocm().
Add allocm(), rallocm(), sallocm(), and dallocm(), which are a
functional superset of malloc(), calloc(), posix_memalign(),
malloc_usable_size(), and free().
2010-09-18 06:46:18 +08:00
|
|
|
|
2014-10-06 08:54:10 +08:00
|
|
|
/* Size class index type. */
|
2015-08-20 06:21:32 +08:00
|
|
|
typedef unsigned szind_t;
|
2014-10-06 08:54:10 +08:00
|
|
|
|
2015-01-30 07:30:47 +08:00
|
|
|
/*
|
|
|
|
* Flags bits:
|
|
|
|
*
|
|
|
|
* a: arena
|
|
|
|
* t: tcache
|
|
|
|
* 0: unused
|
|
|
|
* z: zero
|
|
|
|
* n: alignment
|
|
|
|
*
|
|
|
|
* aaaaaaaa aaaatttt tttttttt 0znnnnnn
|
|
|
|
*/
|
|
|
|
#define MALLOCX_ARENA_MASK ((int)~0xfffff)
|
|
|
|
#define MALLOCX_ARENA_MAX 0xffe
|
|
|
|
#define MALLOCX_TCACHE_MASK ((int)~0xfff000ffU)
|
|
|
|
#define MALLOCX_TCACHE_MAX 0xffd
|
Implement the *allocx() API.
Implement the *allocx() API, which is a successor to the *allocm() API.
The *allocx() functions are slightly simpler to use because they have
fewer parameters, they directly return the results of primary interest,
and mallocx()/rallocx() avoid the strict aliasing pitfall that
allocm()/rallocx() share with posix_memalign(). The following code
violates strict aliasing rules:
foo_t *foo;
allocm((void **)&foo, NULL, 42, 0);
whereas the following is safe:
foo_t *foo;
void *p;
allocm(&p, NULL, 42, 0);
foo = (foo_t *)p;
mallocx() does not have this problem:
foo_t *foo = (foo_t *)mallocx(42, 0);
2013-12-13 14:35:52 +08:00
|
|
|
#define MALLOCX_LG_ALIGN_MASK ((int)0x3f)
|
2014-09-08 05:40:19 +08:00
|
|
|
/* Use MALLOCX_ALIGN_GET() if alignment may not be specified in flags. */
|
|
|
|
#define MALLOCX_ALIGN_GET_SPECIFIED(flags) \
|
|
|
|
(ZU(1) << (flags & MALLOCX_LG_ALIGN_MASK))
|
|
|
|
#define MALLOCX_ALIGN_GET(flags) \
|
|
|
|
(MALLOCX_ALIGN_GET_SPECIFIED(flags) & (SIZE_T_MAX-1))
|
|
|
|
#define MALLOCX_ZERO_GET(flags) \
|
|
|
|
((bool)(flags & MALLOCX_ZERO))
|
2015-01-30 07:30:47 +08:00
|
|
|
|
|
|
|
#define MALLOCX_TCACHE_GET(flags) \
|
|
|
|
(((unsigned)((flags & MALLOCX_TCACHE_MASK) >> 8)) - 2)
|
2014-09-08 05:40:19 +08:00
|
|
|
#define MALLOCX_ARENA_GET(flags) \
|
2015-01-30 07:30:47 +08:00
|
|
|
(((unsigned)(((unsigned)flags) >> 20)) - 1)
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2012-02-29 08:50:47 +08:00
|
|
|
/* Smallest size class to support. */
|
|
|
|
#define TINY_MIN (1U << LG_TINY_MIN)
|
|
|
|
|
|
|
|
/*
|
2014-10-10 08:54:06 +08:00
|
|
|
* Minimum allocation alignment is 2^LG_QUANTUM bytes (ignoring tiny size
|
2012-02-29 08:50:47 +08:00
|
|
|
* classes).
|
|
|
|
*/
|
|
|
|
#ifndef LG_QUANTUM
|
2012-04-30 18:38:31 +08:00
|
|
|
# if (defined(__i386__) || defined(_M_IX86))
|
2012-02-29 08:50:47 +08:00
|
|
|
# define LG_QUANTUM 4
|
|
|
|
# endif
|
|
|
|
# ifdef __ia64__
|
|
|
|
# define LG_QUANTUM 4
|
|
|
|
# endif
|
|
|
|
# ifdef __alpha__
|
|
|
|
# define LG_QUANTUM 4
|
|
|
|
# endif
|
2015-09-16 01:42:36 +08:00
|
|
|
# if (defined(__sparc64__) || defined(__sparcv9))
|
2012-02-29 08:50:47 +08:00
|
|
|
# define LG_QUANTUM 4
|
|
|
|
# endif
|
2012-04-30 18:38:31 +08:00
|
|
|
# if (defined(__amd64__) || defined(__x86_64__) || defined(_M_X64))
|
2012-02-29 08:50:47 +08:00
|
|
|
# define LG_QUANTUM 4
|
|
|
|
# endif
|
|
|
|
# ifdef __arm__
|
|
|
|
# define LG_QUANTUM 3
|
|
|
|
# endif
|
2013-03-18 22:40:20 +08:00
|
|
|
# ifdef __aarch64__
|
|
|
|
# define LG_QUANTUM 4
|
|
|
|
# endif
|
2012-10-09 06:41:06 +08:00
|
|
|
# ifdef __hppa__
|
|
|
|
# define LG_QUANTUM 4
|
|
|
|
# endif
|
2012-02-29 08:50:47 +08:00
|
|
|
# ifdef __mips__
|
|
|
|
# define LG_QUANTUM 3
|
|
|
|
# endif
|
2014-07-30 06:11:26 +08:00
|
|
|
# ifdef __or1k__
|
|
|
|
# define LG_QUANTUM 3
|
|
|
|
# endif
|
2012-02-29 08:50:47 +08:00
|
|
|
# ifdef __powerpc__
|
|
|
|
# define LG_QUANTUM 4
|
|
|
|
# endif
|
2013-01-29 04:19:34 +08:00
|
|
|
# ifdef __s390__
|
2012-02-29 08:50:47 +08:00
|
|
|
# define LG_QUANTUM 4
|
|
|
|
# endif
|
2012-03-06 04:16:57 +08:00
|
|
|
# ifdef __SH4__
|
|
|
|
# define LG_QUANTUM 4
|
|
|
|
# endif
|
2012-02-29 08:50:47 +08:00
|
|
|
# ifdef __tile__
|
|
|
|
# define LG_QUANTUM 4
|
|
|
|
# endif
|
2014-05-29 10:37:02 +08:00
|
|
|
# ifdef __le32__
|
|
|
|
# define LG_QUANTUM 4
|
|
|
|
# endif
|
2012-02-29 08:50:47 +08:00
|
|
|
# ifndef LG_QUANTUM
|
2014-10-10 08:54:06 +08:00
|
|
|
# error "Unknown minimum alignment for architecture; specify via "
|
|
|
|
"--with-lg-quantum"
|
2012-02-29 08:50:47 +08:00
|
|
|
# endif
|
2010-01-17 01:53:50 +08:00
|
|
|
#endif
|
|
|
|
|
|
|
|
#define QUANTUM ((size_t)(1U << LG_QUANTUM))
|
|
|
|
#define QUANTUM_MASK (QUANTUM - 1)
|
|
|
|
|
|
|
|
/* Return the smallest quantum multiple that is >= a. */
|
|
|
|
#define QUANTUM_CEILING(a) \
|
|
|
|
(((a) + QUANTUM_MASK) & ~QUANTUM_MASK)
|
|
|
|
|
Use bitmaps to track small regions.
The previous free list implementation, which embedded singly linked
lists in available regions, had the unfortunate side effect of causing
many cache misses during thread cache fills. Fix this in two places:
- arena_run_t: Use a new bitmap implementation to track which regions
are available. Furthermore, revert to preferring the
lowest available region (as jemalloc did with its old
bitmap-based approach).
- tcache_t: Move read-only tcache_bin_t metadata into
tcache_bin_info_t, and add a contiguous array of pointers
to tcache_t in order to track cached objects. This
substantially increases the size of tcache_t, but results
in much higher data locality for common tcache operations.
As a side benefit, it is again possible to efficiently
flush the least recently used cached objects, so this
change changes flushing from MRU to LRU.
The new bitmap implementation uses a multi-level summary approach to
make finding the lowest available region very fast. In practice,
bitmaps only have one or two levels, though the implementation is
general enough to handle extremely large bitmaps, mainly so that large
page sizes can still be entertained.
Fix tcache_bin_flush_large() to always flush statistics, in the same way
that tcache_bin_flush_small() was recently fixed.
Use JEMALLOC_DEBUG rather than NDEBUG.
Add dassert(), and use it for debug-only asserts.
2011-03-17 01:30:13 +08:00
|
|
|
#define LONG ((size_t)(1U << LG_SIZEOF_LONG))
|
|
|
|
#define LONG_MASK (LONG - 1)
|
|
|
|
|
|
|
|
/* Return the smallest long multiple that is >= a. */
|
2012-03-14 02:09:23 +08:00
|
|
|
#define LONG_CEILING(a) \
|
Use bitmaps to track small regions.
The previous free list implementation, which embedded singly linked
lists in available regions, had the unfortunate side effect of causing
many cache misses during thread cache fills. Fix this in two places:
- arena_run_t: Use a new bitmap implementation to track which regions
are available. Furthermore, revert to preferring the
lowest available region (as jemalloc did with its old
bitmap-based approach).
- tcache_t: Move read-only tcache_bin_t metadata into
tcache_bin_info_t, and add a contiguous array of pointers
to tcache_t in order to track cached objects. This
substantially increases the size of tcache_t, but results
in much higher data locality for common tcache operations.
As a side benefit, it is again possible to efficiently
flush the least recently used cached objects, so this
change changes flushing from MRU to LRU.
The new bitmap implementation uses a multi-level summary approach to
make finding the lowest available region very fast. In practice,
bitmaps only have one or two levels, though the implementation is
general enough to handle extremely large bitmaps, mainly so that large
page sizes can still be entertained.
Fix tcache_bin_flush_large() to always flush statistics, in the same way
that tcache_bin_flush_small() was recently fixed.
Use JEMALLOC_DEBUG rather than NDEBUG.
Add dassert(), and use it for debug-only asserts.
2011-03-17 01:30:13 +08:00
|
|
|
(((a) + LONG_MASK) & ~LONG_MASK)
|
|
|
|
|
2010-01-17 01:53:50 +08:00
|
|
|
#define SIZEOF_PTR (1U << LG_SIZEOF_PTR)
|
Use bitmaps to track small regions.
The previous free list implementation, which embedded singly linked
lists in available regions, had the unfortunate side effect of causing
many cache misses during thread cache fills. Fix this in two places:
- arena_run_t: Use a new bitmap implementation to track which regions
are available. Furthermore, revert to preferring the
lowest available region (as jemalloc did with its old
bitmap-based approach).
- tcache_t: Move read-only tcache_bin_t metadata into
tcache_bin_info_t, and add a contiguous array of pointers
to tcache_t in order to track cached objects. This
substantially increases the size of tcache_t, but results
in much higher data locality for common tcache operations.
As a side benefit, it is again possible to efficiently
flush the least recently used cached objects, so this
change changes flushing from MRU to LRU.
The new bitmap implementation uses a multi-level summary approach to
make finding the lowest available region very fast. In practice,
bitmaps only have one or two levels, though the implementation is
general enough to handle extremely large bitmaps, mainly so that large
page sizes can still be entertained.
Fix tcache_bin_flush_large() to always flush statistics, in the same way
that tcache_bin_flush_small() was recently fixed.
Use JEMALLOC_DEBUG rather than NDEBUG.
Add dassert(), and use it for debug-only asserts.
2011-03-17 01:30:13 +08:00
|
|
|
#define PTR_MASK (SIZEOF_PTR - 1)
|
|
|
|
|
|
|
|
/* Return the smallest (void *) multiple that is >= a. */
|
2012-03-14 02:09:23 +08:00
|
|
|
#define PTR_CEILING(a) \
|
Use bitmaps to track small regions.
The previous free list implementation, which embedded singly linked
lists in available regions, had the unfortunate side effect of causing
many cache misses during thread cache fills. Fix this in two places:
- arena_run_t: Use a new bitmap implementation to track which regions
are available. Furthermore, revert to preferring the
lowest available region (as jemalloc did with its old
bitmap-based approach).
- tcache_t: Move read-only tcache_bin_t metadata into
tcache_bin_info_t, and add a contiguous array of pointers
to tcache_t in order to track cached objects. This
substantially increases the size of tcache_t, but results
in much higher data locality for common tcache operations.
As a side benefit, it is again possible to efficiently
flush the least recently used cached objects, so this
change changes flushing from MRU to LRU.
The new bitmap implementation uses a multi-level summary approach to
make finding the lowest available region very fast. In practice,
bitmaps only have one or two levels, though the implementation is
general enough to handle extremely large bitmaps, mainly so that large
page sizes can still be entertained.
Fix tcache_bin_flush_large() to always flush statistics, in the same way
that tcache_bin_flush_small() was recently fixed.
Use JEMALLOC_DEBUG rather than NDEBUG.
Add dassert(), and use it for debug-only asserts.
2011-03-17 01:30:13 +08:00
|
|
|
(((a) + PTR_MASK) & ~PTR_MASK)
|
2010-01-17 01:53:50 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Maximum size of L1 cache line. This is used to avoid cache line aliasing.
|
|
|
|
* In addition, this controls the spacing of cacheline-spaced size classes.
|
2012-05-02 16:22:16 +08:00
|
|
|
*
|
|
|
|
* CACHELINE cannot be based on LG_CACHELINE because __declspec(align()) can
|
|
|
|
* only handle raw constants.
|
2010-01-17 01:53:50 +08:00
|
|
|
*/
|
|
|
|
#define LG_CACHELINE 6
|
2012-05-02 16:22:16 +08:00
|
|
|
#define CACHELINE 64
|
2010-01-17 01:53:50 +08:00
|
|
|
#define CACHELINE_MASK (CACHELINE - 1)
|
|
|
|
|
|
|
|
/* Return the smallest cacheline multiple that is >= s. */
|
|
|
|
#define CACHELINE_CEILING(s) \
|
|
|
|
(((s) + CACHELINE_MASK) & ~CACHELINE_MASK)
|
|
|
|
|
2014-10-10 08:54:06 +08:00
|
|
|
/* Page size. LG_PAGE is determined by the configure script. */
|
2010-09-06 01:35:13 +08:00
|
|
|
#ifdef PAGE_MASK
|
|
|
|
# undef PAGE_MASK
|
|
|
|
#endif
|
2014-10-10 08:54:06 +08:00
|
|
|
#define PAGE ((size_t)(1U << LG_PAGE))
|
2012-04-02 22:04:34 +08:00
|
|
|
#define PAGE_MASK ((size_t)(PAGE - 1))
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2015-09-25 13:21:55 +08:00
|
|
|
/* Return the page base address for the page containing address a. */
|
|
|
|
#define PAGE_ADDR2BASE(a) \
|
|
|
|
((void *)((uintptr_t)(a) & ~PAGE_MASK))
|
|
|
|
|
2010-01-17 01:53:50 +08:00
|
|
|
/* Return the smallest pagesize multiple that is >= s. */
|
|
|
|
#define PAGE_CEILING(s) \
|
|
|
|
(((s) + PAGE_MASK) & ~PAGE_MASK)
|
|
|
|
|
2012-04-12 09:13:45 +08:00
|
|
|
/* Return the nearest aligned address at or below a. */
|
|
|
|
#define ALIGNMENT_ADDR2BASE(a, alignment) \
|
|
|
|
((void *)((uintptr_t)(a) & (-(alignment))))
|
|
|
|
|
|
|
|
/* Return the offset between a and the nearest aligned address at or below a. */
|
|
|
|
#define ALIGNMENT_ADDR2OFFSET(a, alignment) \
|
|
|
|
((size_t)((uintptr_t)(a) & (alignment - 1)))
|
|
|
|
|
|
|
|
/* Return the smallest alignment multiple that is >= s. */
|
|
|
|
#define ALIGNMENT_CEILING(s, alignment) \
|
|
|
|
(((s) + (alignment - 1)) & (-(alignment)))
|
|
|
|
|
2014-12-09 06:40:14 +08:00
|
|
|
/* Declare a variable-length array. */
|
2012-04-25 05:22:02 +08:00
|
|
|
#if __STDC_VERSION__ < 199901L
|
|
|
|
# ifdef _MSC_VER
|
|
|
|
# include <malloc.h>
|
|
|
|
# define alloca _alloca
|
|
|
|
# else
|
2012-12-03 09:58:40 +08:00
|
|
|
# ifdef JEMALLOC_HAS_ALLOCA_H
|
|
|
|
# include <alloca.h>
|
|
|
|
# else
|
|
|
|
# include <stdlib.h>
|
|
|
|
# endif
|
2012-04-25 05:22:02 +08:00
|
|
|
# endif
|
|
|
|
# define VARIABLE_ARRAY(type, name, count) \
|
2014-04-22 11:52:35 +08:00
|
|
|
type *name = alloca(sizeof(type) * (count))
|
2012-04-25 05:22:02 +08:00
|
|
|
#else
|
2014-04-22 11:52:35 +08:00
|
|
|
# define VARIABLE_ARRAY(type, name, count) type name[(count)]
|
2012-04-25 05:22:02 +08:00
|
|
|
#endif
|
|
|
|
|
2016-02-22 03:25:02 +08:00
|
|
|
#include "jemalloc/internal/nstime.h"
|
2014-04-16 07:35:08 +08:00
|
|
|
#include "jemalloc/internal/valgrind.h"
|
2012-03-07 06:57:45 +08:00
|
|
|
#include "jemalloc/internal/util.h"
|
2011-03-19 08:56:14 +08:00
|
|
|
#include "jemalloc/internal/atomic.h"
|
2012-03-03 07:59:45 +08:00
|
|
|
#include "jemalloc/internal/prng.h"
|
2016-02-03 12:27:54 +08:00
|
|
|
#include "jemalloc/internal/ticker.h"
|
2010-02-12 06:45:59 +08:00
|
|
|
#include "jemalloc/internal/ckh.h"
|
2012-02-29 08:50:47 +08:00
|
|
|
#include "jemalloc/internal/size_classes.h"
|
2016-02-06 16:46:19 +08:00
|
|
|
#include "jemalloc/internal/smoothstep.h"
|
2010-02-12 06:45:59 +08:00
|
|
|
#include "jemalloc/internal/stats.h"
|
|
|
|
#include "jemalloc/internal/ctl.h"
|
|
|
|
#include "jemalloc/internal/mutex.h"
|
2012-03-22 09:33:03 +08:00
|
|
|
#include "jemalloc/internal/tsd.h"
|
2010-02-12 07:56:23 +08:00
|
|
|
#include "jemalloc/internal/mb.h"
|
2010-02-12 06:45:59 +08:00
|
|
|
#include "jemalloc/internal/extent.h"
|
2016-03-01 03:22:52 +08:00
|
|
|
#include "jemalloc/internal/ph.h"
|
2010-02-12 06:45:59 +08:00
|
|
|
#include "jemalloc/internal/arena.h"
|
2011-03-23 00:00:56 +08:00
|
|
|
#include "jemalloc/internal/bitmap.h"
|
2010-02-12 06:45:59 +08:00
|
|
|
#include "jemalloc/internal/base.h"
|
Move centralized chunk management into arenas.
Migrate all centralized data structures related to huge allocations and
recyclable chunks into arena_t, so that each arena can manage huge
allocations and recyclable virtual memory completely independently of
other arenas.
Add chunk node caching to arenas, in order to avoid contention on the
base allocator.
Use chunks_rtree to look up huge allocations rather than a red-black
tree. Maintain a per arena unsorted list of huge allocations (which
will be needed to enumerate huge allocations during arena reset).
Remove the --enable-ivsalloc option, make ivsalloc() always available,
and use it for size queries if --enable-debug is enabled. The only
practical implications to this removal are that 1) ivsalloc() is now
always available during live debugging (and the underlying radix tree is
available during core-based debugging), and 2) size query validation can
no longer be enabled independent of --enable-debug.
Remove the stats.chunks.{current,total,high} mallctls, and replace their
underlying statistics with simpler atomically updated counters used
exclusively for gdump triggering. These statistics are no longer very
useful because each arena manages chunks independently, and per arena
statistics provide similar information.
Simplify chunk synchronization code, now that base chunk allocation
cannot cause recursive lock acquisition.
2015-02-12 04:24:27 +08:00
|
|
|
#include "jemalloc/internal/rtree.h"
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
#include "jemalloc/internal/pages.h"
|
2010-02-12 06:45:59 +08:00
|
|
|
#include "jemalloc/internal/chunk.h"
|
|
|
|
#include "jemalloc/internal/huge.h"
|
|
|
|
#include "jemalloc/internal/tcache.h"
|
|
|
|
#include "jemalloc/internal/hash.h"
|
2012-04-06 15:35:09 +08:00
|
|
|
#include "jemalloc/internal/quarantine.h"
|
2010-10-21 10:05:59 +08:00
|
|
|
#include "jemalloc/internal/prof.h"
|
2010-01-17 01:53:50 +08:00
|
|
|
|
|
|
|
#undef JEMALLOC_H_TYPES
|
|
|
|
/******************************************************************************/
|
2013-12-09 14:28:27 +08:00
|
|
|
#define JEMALLOC_H_STRUCTS
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2016-02-22 03:25:02 +08:00
|
|
|
#include "jemalloc/internal/nstime.h"
|
2014-04-16 07:35:08 +08:00
|
|
|
#include "jemalloc/internal/valgrind.h"
|
2012-03-07 06:57:45 +08:00
|
|
|
#include "jemalloc/internal/util.h"
|
2011-03-19 08:56:14 +08:00
|
|
|
#include "jemalloc/internal/atomic.h"
|
2012-03-03 07:59:45 +08:00
|
|
|
#include "jemalloc/internal/prng.h"
|
2016-02-03 12:27:54 +08:00
|
|
|
#include "jemalloc/internal/ticker.h"
|
2010-02-12 06:45:59 +08:00
|
|
|
#include "jemalloc/internal/ckh.h"
|
2012-02-29 08:50:47 +08:00
|
|
|
#include "jemalloc/internal/size_classes.h"
|
2016-02-06 16:46:19 +08:00
|
|
|
#include "jemalloc/internal/smoothstep.h"
|
2010-02-12 06:45:59 +08:00
|
|
|
#include "jemalloc/internal/stats.h"
|
|
|
|
#include "jemalloc/internal/ctl.h"
|
|
|
|
#include "jemalloc/internal/mutex.h"
|
2010-02-12 07:56:23 +08:00
|
|
|
#include "jemalloc/internal/mb.h"
|
Use bitmaps to track small regions.
The previous free list implementation, which embedded singly linked
lists in available regions, had the unfortunate side effect of causing
many cache misses during thread cache fills. Fix this in two places:
- arena_run_t: Use a new bitmap implementation to track which regions
are available. Furthermore, revert to preferring the
lowest available region (as jemalloc did with its old
bitmap-based approach).
- tcache_t: Move read-only tcache_bin_t metadata into
tcache_bin_info_t, and add a contiguous array of pointers
to tcache_t in order to track cached objects. This
substantially increases the size of tcache_t, but results
in much higher data locality for common tcache operations.
As a side benefit, it is again possible to efficiently
flush the least recently used cached objects, so this
change changes flushing from MRU to LRU.
The new bitmap implementation uses a multi-level summary approach to
make finding the lowest available region very fast. In practice,
bitmaps only have one or two levels, though the implementation is
general enough to handle extremely large bitmaps, mainly so that large
page sizes can still be entertained.
Fix tcache_bin_flush_large() to always flush statistics, in the same way
that tcache_bin_flush_small() was recently fixed.
Use JEMALLOC_DEBUG rather than NDEBUG.
Add dassert(), and use it for debug-only asserts.
2011-03-17 01:30:13 +08:00
|
|
|
#include "jemalloc/internal/bitmap.h"
|
2016-03-01 03:22:52 +08:00
|
|
|
#include "jemalloc/internal/ph.h"
|
2015-02-16 10:04:46 +08:00
|
|
|
#define JEMALLOC_ARENA_STRUCTS_A
|
|
|
|
#include "jemalloc/internal/arena.h"
|
|
|
|
#undef JEMALLOC_ARENA_STRUCTS_A
|
2010-02-12 06:45:59 +08:00
|
|
|
#include "jemalloc/internal/extent.h"
|
2015-02-16 10:04:46 +08:00
|
|
|
#define JEMALLOC_ARENA_STRUCTS_B
|
2010-02-12 06:45:59 +08:00
|
|
|
#include "jemalloc/internal/arena.h"
|
2015-02-16 10:04:46 +08:00
|
|
|
#undef JEMALLOC_ARENA_STRUCTS_B
|
2010-02-12 06:45:59 +08:00
|
|
|
#include "jemalloc/internal/base.h"
|
Move centralized chunk management into arenas.
Migrate all centralized data structures related to huge allocations and
recyclable chunks into arena_t, so that each arena can manage huge
allocations and recyclable virtual memory completely independently of
other arenas.
Add chunk node caching to arenas, in order to avoid contention on the
base allocator.
Use chunks_rtree to look up huge allocations rather than a red-black
tree. Maintain a per arena unsorted list of huge allocations (which
will be needed to enumerate huge allocations during arena reset).
Remove the --enable-ivsalloc option, make ivsalloc() always available,
and use it for size queries if --enable-debug is enabled. The only
practical implications to this removal are that 1) ivsalloc() is now
always available during live debugging (and the underlying radix tree is
available during core-based debugging), and 2) size query validation can
no longer be enabled independent of --enable-debug.
Remove the stats.chunks.{current,total,high} mallctls, and replace their
underlying statistics with simpler atomically updated counters used
exclusively for gdump triggering. These statistics are no longer very
useful because each arena manages chunks independently, and per arena
statistics provide similar information.
Simplify chunk synchronization code, now that base chunk allocation
cannot cause recursive lock acquisition.
2015-02-12 04:24:27 +08:00
|
|
|
#include "jemalloc/internal/rtree.h"
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
#include "jemalloc/internal/pages.h"
|
2010-02-12 06:45:59 +08:00
|
|
|
#include "jemalloc/internal/chunk.h"
|
|
|
|
#include "jemalloc/internal/huge.h"
|
|
|
|
#include "jemalloc/internal/tcache.h"
|
|
|
|
#include "jemalloc/internal/hash.h"
|
2012-04-06 15:35:09 +08:00
|
|
|
#include "jemalloc/internal/quarantine.h"
|
2010-10-21 10:05:59 +08:00
|
|
|
#include "jemalloc/internal/prof.h"
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2014-09-23 12:09:23 +08:00
|
|
|
#include "jemalloc/internal/tsd.h"
|
2011-02-14 10:11:54 +08:00
|
|
|
|
2010-01-17 01:53:50 +08:00
|
|
|
#undef JEMALLOC_H_STRUCTS
|
|
|
|
/******************************************************************************/
|
2013-12-09 14:28:27 +08:00
|
|
|
#define JEMALLOC_H_EXTERNS
|
2010-01-17 01:53:50 +08:00
|
|
|
|
|
|
|
extern bool opt_abort;
|
2014-12-09 05:12:41 +08:00
|
|
|
extern const char *opt_junk;
|
|
|
|
extern bool opt_junk_alloc;
|
|
|
|
extern bool opt_junk_free;
|
2012-04-06 15:35:09 +08:00
|
|
|
extern size_t opt_quarantine;
|
|
|
|
extern bool opt_redzone;
|
2012-04-06 04:36:17 +08:00
|
|
|
extern bool opt_utrace;
|
2010-01-17 01:53:50 +08:00
|
|
|
extern bool opt_xmalloc;
|
|
|
|
extern bool opt_zero;
|
2016-02-25 03:03:40 +08:00
|
|
|
extern unsigned opt_narenas;
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2014-04-16 05:33:50 +08:00
|
|
|
extern bool in_valgrind;
|
|
|
|
|
2010-01-17 01:53:50 +08:00
|
|
|
/* Number of CPUs. */
|
2016-02-25 15:58:10 +08:00
|
|
|
extern unsigned ncpus;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Arenas that are used to service external requests. Not all elements of the
|
|
|
|
* arenas array are necessarily used; arenas are created lazily as needed.
|
|
|
|
*/
|
|
|
|
extern arena_t **arenas;
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2014-10-06 08:54:10 +08:00
|
|
|
/*
|
|
|
|
* index2size_tab encodes the same information as could be computed (at
|
|
|
|
* unacceptable cost in some code paths) by index2size_compute().
|
|
|
|
*/
|
2015-10-28 06:12:10 +08:00
|
|
|
extern size_t const index2size_tab[NSIZES+1];
|
2014-10-06 08:54:10 +08:00
|
|
|
/*
|
|
|
|
* size2index_tab is a compact lookup table that rounds request sizes up to
|
|
|
|
* size classes. In order to reduce cache footprint, the table is compressed,
|
|
|
|
* and all accesses are via size2index().
|
|
|
|
*/
|
|
|
|
extern uint8_t const size2index_tab[];
|
|
|
|
|
2014-11-28 03:22:36 +08:00
|
|
|
void *a0malloc(size_t size);
|
2015-01-21 07:37:51 +08:00
|
|
|
void a0dalloc(void *ptr);
|
2015-02-04 04:39:55 +08:00
|
|
|
void *bootstrap_malloc(size_t size);
|
|
|
|
void *bootstrap_calloc(size_t num, size_t size);
|
|
|
|
void bootstrap_free(void *ptr);
|
2010-01-17 01:53:50 +08:00
|
|
|
arena_t *arenas_extend(unsigned ind);
|
Refactor/fix arenas manipulation.
Abstract arenas access to use arena_get() (or a0get() where appropriate)
rather than directly reading e.g. arenas[ind]. Prior to the addition of
the arenas.extend mallctl, the worst possible outcome of directly
accessing arenas was a stale read, but arenas.extend may allocate and
assign a new array to arenas.
Add a tsd-based arenas_cache, which amortizes arenas reads. This
introduces some subtle bootstrapping issues, with tsd_boot() now being
split into tsd_boot[01]() to support tsd wrapper allocation
bootstrapping, as well as an arenas_cache_bypass tsd variable which
dynamically terminates allocation of arenas_cache itself.
Promote a0malloc(), a0calloc(), and a0free() to be generally useful for
internal allocation, and use them in several places (more may be
appropriate).
Abstract arena->nthreads management and fix a missing decrement during
thread destruction (recent tsd refactoring left arenas_cleanup()
unused).
Change arena_choose() to propagate OOM, and handle OOM in all callers.
This is important for providing consistent allocation behavior when the
MALLOCX_ARENA() flag is being used. Prior to this fix, it was possible
for an OOM to result in allocation silently allocating from a different
arena than the one specified.
2014-10-08 14:14:57 +08:00
|
|
|
unsigned narenas_total_get(void);
|
2016-02-25 15:58:10 +08:00
|
|
|
arena_t *arena_init(unsigned ind);
|
2016-02-20 11:37:10 +08:00
|
|
|
arena_tdata_t *arena_tdata_get_hard(tsd_t *tsd, unsigned ind);
|
Refactor/fix arenas manipulation.
Abstract arenas access to use arena_get() (or a0get() where appropriate)
rather than directly reading e.g. arenas[ind]. Prior to the addition of
the arenas.extend mallctl, the worst possible outcome of directly
accessing arenas was a stale read, but arenas.extend may allocate and
assign a new array to arenas.
Add a tsd-based arenas_cache, which amortizes arenas reads. This
introduces some subtle bootstrapping issues, with tsd_boot() now being
split into tsd_boot[01]() to support tsd wrapper allocation
bootstrapping, as well as an arenas_cache_bypass tsd variable which
dynamically terminates allocation of arenas_cache itself.
Promote a0malloc(), a0calloc(), and a0free() to be generally useful for
internal allocation, and use them in several places (more may be
appropriate).
Abstract arena->nthreads management and fix a missing decrement during
thread destruction (recent tsd refactoring left arenas_cleanup()
unused).
Change arena_choose() to propagate OOM, and handle OOM in all callers.
This is important for providing consistent allocation behavior when the
MALLOCX_ARENA() flag is being used. Prior to this fix, it was possible
for an OOM to result in allocation silently allocating from a different
arena than the one specified.
2014-10-08 14:14:57 +08:00
|
|
|
arena_t *arena_choose_hard(tsd_t *tsd);
|
|
|
|
void arena_migrate(tsd_t *tsd, unsigned oldind, unsigned newind);
|
2014-09-23 12:09:23 +08:00
|
|
|
void thread_allocated_cleanup(tsd_t *tsd);
|
|
|
|
void thread_deallocated_cleanup(tsd_t *tsd);
|
|
|
|
void arena_cleanup(tsd_t *tsd);
|
2016-02-20 11:37:10 +08:00
|
|
|
void arenas_tdata_cleanup(tsd_t *tsd);
|
|
|
|
void narenas_tdata_cleanup(tsd_t *tsd);
|
|
|
|
void arenas_tdata_bypass_cleanup(tsd_t *tsd);
|
2010-09-06 01:35:13 +08:00
|
|
|
void jemalloc_prefork(void);
|
2012-03-14 07:31:41 +08:00
|
|
|
void jemalloc_postfork_parent(void);
|
|
|
|
void jemalloc_postfork_child(void);
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2016-02-22 03:25:02 +08:00
|
|
|
#include "jemalloc/internal/nstime.h"
|
2014-04-16 07:35:08 +08:00
|
|
|
#include "jemalloc/internal/valgrind.h"
|
2012-03-07 06:57:45 +08:00
|
|
|
#include "jemalloc/internal/util.h"
|
2011-03-19 08:56:14 +08:00
|
|
|
#include "jemalloc/internal/atomic.h"
|
2012-03-03 07:59:45 +08:00
|
|
|
#include "jemalloc/internal/prng.h"
|
2016-02-03 12:27:54 +08:00
|
|
|
#include "jemalloc/internal/ticker.h"
|
2010-02-12 06:45:59 +08:00
|
|
|
#include "jemalloc/internal/ckh.h"
|
2012-02-29 08:50:47 +08:00
|
|
|
#include "jemalloc/internal/size_classes.h"
|
2016-02-06 16:46:19 +08:00
|
|
|
#include "jemalloc/internal/smoothstep.h"
|
2010-02-12 06:45:59 +08:00
|
|
|
#include "jemalloc/internal/stats.h"
|
|
|
|
#include "jemalloc/internal/ctl.h"
|
|
|
|
#include "jemalloc/internal/mutex.h"
|
2010-02-12 07:56:23 +08:00
|
|
|
#include "jemalloc/internal/mb.h"
|
Use bitmaps to track small regions.
The previous free list implementation, which embedded singly linked
lists in available regions, had the unfortunate side effect of causing
many cache misses during thread cache fills. Fix this in two places:
- arena_run_t: Use a new bitmap implementation to track which regions
are available. Furthermore, revert to preferring the
lowest available region (as jemalloc did with its old
bitmap-based approach).
- tcache_t: Move read-only tcache_bin_t metadata into
tcache_bin_info_t, and add a contiguous array of pointers
to tcache_t in order to track cached objects. This
substantially increases the size of tcache_t, but results
in much higher data locality for common tcache operations.
As a side benefit, it is again possible to efficiently
flush the least recently used cached objects, so this
change changes flushing from MRU to LRU.
The new bitmap implementation uses a multi-level summary approach to
make finding the lowest available region very fast. In practice,
bitmaps only have one or two levels, though the implementation is
general enough to handle extremely large bitmaps, mainly so that large
page sizes can still be entertained.
Fix tcache_bin_flush_large() to always flush statistics, in the same way
that tcache_bin_flush_small() was recently fixed.
Use JEMALLOC_DEBUG rather than NDEBUG.
Add dassert(), and use it for debug-only asserts.
2011-03-17 01:30:13 +08:00
|
|
|
#include "jemalloc/internal/bitmap.h"
|
2010-02-12 06:45:59 +08:00
|
|
|
#include "jemalloc/internal/extent.h"
|
2016-03-01 03:22:52 +08:00
|
|
|
#include "jemalloc/internal/ph.h"
|
2010-02-12 06:45:59 +08:00
|
|
|
#include "jemalloc/internal/arena.h"
|
|
|
|
#include "jemalloc/internal/base.h"
|
Move centralized chunk management into arenas.
Migrate all centralized data structures related to huge allocations and
recyclable chunks into arena_t, so that each arena can manage huge
allocations and recyclable virtual memory completely independently of
other arenas.
Add chunk node caching to arenas, in order to avoid contention on the
base allocator.
Use chunks_rtree to look up huge allocations rather than a red-black
tree. Maintain a per arena unsorted list of huge allocations (which
will be needed to enumerate huge allocations during arena reset).
Remove the --enable-ivsalloc option, make ivsalloc() always available,
and use it for size queries if --enable-debug is enabled. The only
practical implications to this removal are that 1) ivsalloc() is now
always available during live debugging (and the underlying radix tree is
available during core-based debugging), and 2) size query validation can
no longer be enabled independent of --enable-debug.
Remove the stats.chunks.{current,total,high} mallctls, and replace their
underlying statistics with simpler atomically updated counters used
exclusively for gdump triggering. These statistics are no longer very
useful because each arena manages chunks independently, and per arena
statistics provide similar information.
Simplify chunk synchronization code, now that base chunk allocation
cannot cause recursive lock acquisition.
2015-02-12 04:24:27 +08:00
|
|
|
#include "jemalloc/internal/rtree.h"
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
#include "jemalloc/internal/pages.h"
|
2010-02-12 06:45:59 +08:00
|
|
|
#include "jemalloc/internal/chunk.h"
|
|
|
|
#include "jemalloc/internal/huge.h"
|
|
|
|
#include "jemalloc/internal/tcache.h"
|
|
|
|
#include "jemalloc/internal/hash.h"
|
2012-04-06 15:35:09 +08:00
|
|
|
#include "jemalloc/internal/quarantine.h"
|
2010-10-21 10:05:59 +08:00
|
|
|
#include "jemalloc/internal/prof.h"
|
2014-09-23 12:09:23 +08:00
|
|
|
#include "jemalloc/internal/tsd.h"
|
2010-01-17 01:53:50 +08:00
|
|
|
|
|
|
|
#undef JEMALLOC_H_EXTERNS
|
|
|
|
/******************************************************************************/
|
2013-12-09 14:28:27 +08:00
|
|
|
#define JEMALLOC_H_INLINES
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2016-02-22 03:25:02 +08:00
|
|
|
#include "jemalloc/internal/nstime.h"
|
2014-04-16 07:35:08 +08:00
|
|
|
#include "jemalloc/internal/valgrind.h"
|
2012-03-07 06:57:45 +08:00
|
|
|
#include "jemalloc/internal/util.h"
|
2011-03-19 08:56:14 +08:00
|
|
|
#include "jemalloc/internal/atomic.h"
|
2012-03-03 07:59:45 +08:00
|
|
|
#include "jemalloc/internal/prng.h"
|
2016-02-03 12:27:54 +08:00
|
|
|
#include "jemalloc/internal/ticker.h"
|
2010-02-12 06:45:59 +08:00
|
|
|
#include "jemalloc/internal/ckh.h"
|
2012-02-29 08:50:47 +08:00
|
|
|
#include "jemalloc/internal/size_classes.h"
|
2016-02-06 16:46:19 +08:00
|
|
|
#include "jemalloc/internal/smoothstep.h"
|
2010-02-12 06:45:59 +08:00
|
|
|
#include "jemalloc/internal/stats.h"
|
|
|
|
#include "jemalloc/internal/ctl.h"
|
|
|
|
#include "jemalloc/internal/mutex.h"
|
2012-03-22 09:33:03 +08:00
|
|
|
#include "jemalloc/internal/tsd.h"
|
2010-02-12 07:56:23 +08:00
|
|
|
#include "jemalloc/internal/mb.h"
|
2010-02-12 06:45:59 +08:00
|
|
|
#include "jemalloc/internal/extent.h"
|
2016-03-01 03:22:52 +08:00
|
|
|
#include "jemalloc/internal/ph.h"
|
2010-02-12 06:45:59 +08:00
|
|
|
#include "jemalloc/internal/base.h"
|
Move centralized chunk management into arenas.
Migrate all centralized data structures related to huge allocations and
recyclable chunks into arena_t, so that each arena can manage huge
allocations and recyclable virtual memory completely independently of
other arenas.
Add chunk node caching to arenas, in order to avoid contention on the
base allocator.
Use chunks_rtree to look up huge allocations rather than a red-black
tree. Maintain a per arena unsorted list of huge allocations (which
will be needed to enumerate huge allocations during arena reset).
Remove the --enable-ivsalloc option, make ivsalloc() always available,
and use it for size queries if --enable-debug is enabled. The only
practical implications to this removal are that 1) ivsalloc() is now
always available during live debugging (and the underlying radix tree is
available during core-based debugging), and 2) size query validation can
no longer be enabled independent of --enable-debug.
Remove the stats.chunks.{current,total,high} mallctls, and replace their
underlying statistics with simpler atomically updated counters used
exclusively for gdump triggering. These statistics are no longer very
useful because each arena manages chunks independently, and per arena
statistics provide similar information.
Simplify chunk synchronization code, now that base chunk allocation
cannot cause recursive lock acquisition.
2015-02-12 04:24:27 +08:00
|
|
|
#include "jemalloc/internal/rtree.h"
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
#include "jemalloc/internal/pages.h"
|
2010-02-12 06:45:59 +08:00
|
|
|
#include "jemalloc/internal/chunk.h"
|
|
|
|
#include "jemalloc/internal/huge.h"
|
2010-01-17 01:53:50 +08:00
|
|
|
|
|
|
|
#ifndef JEMALLOC_ENABLE_INLINE
|
2015-08-20 06:21:32 +08:00
|
|
|
szind_t size2index_compute(size_t size);
|
|
|
|
szind_t size2index_lookup(size_t size);
|
|
|
|
szind_t size2index(size_t size);
|
|
|
|
size_t index2size_compute(szind_t index);
|
|
|
|
size_t index2size_lookup(szind_t index);
|
|
|
|
size_t index2size(szind_t index);
|
2014-10-06 08:54:10 +08:00
|
|
|
size_t s2u_compute(size_t size);
|
|
|
|
size_t s2u_lookup(size_t size);
|
2010-10-21 08:39:18 +08:00
|
|
|
size_t s2u(size_t size);
|
2012-04-12 09:13:45 +08:00
|
|
|
size_t sa2u(size_t size, size_t alignment);
|
Refactor/fix arenas manipulation.
Abstract arenas access to use arena_get() (or a0get() where appropriate)
rather than directly reading e.g. arenas[ind]. Prior to the addition of
the arenas.extend mallctl, the worst possible outcome of directly
accessing arenas was a stale read, but arenas.extend may allocate and
assign a new array to arenas.
Add a tsd-based arenas_cache, which amortizes arenas reads. This
introduces some subtle bootstrapping issues, with tsd_boot() now being
split into tsd_boot[01]() to support tsd wrapper allocation
bootstrapping, as well as an arenas_cache_bypass tsd variable which
dynamically terminates allocation of arenas_cache itself.
Promote a0malloc(), a0calloc(), and a0free() to be generally useful for
internal allocation, and use them in several places (more may be
appropriate).
Abstract arena->nthreads management and fix a missing decrement during
thread destruction (recent tsd refactoring left arenas_cleanup()
unused).
Change arena_choose() to propagate OOM, and handle OOM in all callers.
This is important for providing consistent allocation behavior when the
MALLOCX_ARENA() flag is being used. Prior to this fix, it was possible
for an OOM to result in allocation silently allocating from a different
arena than the one specified.
2014-10-08 14:14:57 +08:00
|
|
|
arena_t *arena_choose(tsd_t *tsd, arena_t *arena);
|
2016-02-20 11:37:10 +08:00
|
|
|
arena_tdata_t *arena_tdata_get(tsd_t *tsd, unsigned ind,
|
|
|
|
bool refresh_if_missing);
|
2016-02-25 15:58:10 +08:00
|
|
|
arena_t *arena_get(unsigned ind, bool init_if_missing);
|
2016-02-20 12:09:31 +08:00
|
|
|
ticker_t *decay_ticker_get(tsd_t *tsd, unsigned ind);
|
2010-01-17 01:53:50 +08:00
|
|
|
#endif
|
|
|
|
|
|
|
|
#if (defined(JEMALLOC_ENABLE_INLINE) || defined(JEMALLOC_C_))
|
2015-08-20 06:21:32 +08:00
|
|
|
JEMALLOC_INLINE szind_t
|
2014-10-06 08:54:10 +08:00
|
|
|
size2index_compute(size_t size)
|
|
|
|
{
|
|
|
|
|
|
|
|
#if (NTBINS != 0)
|
|
|
|
if (size <= (ZU(1) << LG_TINY_MAXCLASS)) {
|
2016-02-25 03:04:51 +08:00
|
|
|
szind_t lg_tmin = LG_TINY_MAXCLASS - NTBINS + 1;
|
|
|
|
szind_t lg_ceil = lg_floor(pow2_ceil_zu(size));
|
2014-10-06 08:54:10 +08:00
|
|
|
return (lg_ceil < lg_tmin ? 0 : lg_ceil - lg_tmin);
|
2015-06-24 09:47:07 +08:00
|
|
|
}
|
2014-10-06 08:54:10 +08:00
|
|
|
#endif
|
|
|
|
{
|
2016-02-25 03:04:51 +08:00
|
|
|
szind_t x = unlikely(ZI(size) < 0) ? ((size<<1) ?
|
2015-07-10 12:36:33 +08:00
|
|
|
(ZU(1)<<(LG_SIZEOF_PTR+3)) : ((ZU(1)<<(LG_SIZEOF_PTR+3))-1))
|
|
|
|
: lg_floor((size<<1)-1);
|
2016-02-25 03:04:51 +08:00
|
|
|
szind_t shift = (x < LG_SIZE_CLASS_GROUP + LG_QUANTUM) ? 0 :
|
2014-10-06 08:54:10 +08:00
|
|
|
x - (LG_SIZE_CLASS_GROUP + LG_QUANTUM);
|
2016-02-25 03:04:51 +08:00
|
|
|
szind_t grp = shift << LG_SIZE_CLASS_GROUP;
|
2014-10-06 08:54:10 +08:00
|
|
|
|
2016-02-25 03:04:51 +08:00
|
|
|
szind_t lg_delta = (x < LG_SIZE_CLASS_GROUP + LG_QUANTUM + 1)
|
2014-10-06 08:54:10 +08:00
|
|
|
? LG_QUANTUM : x - LG_SIZE_CLASS_GROUP - 1;
|
|
|
|
|
|
|
|
size_t delta_inverse_mask = ZI(-1) << lg_delta;
|
2016-02-25 03:04:51 +08:00
|
|
|
szind_t mod = ((((size-1) & delta_inverse_mask) >> lg_delta)) &
|
2014-10-06 08:54:10 +08:00
|
|
|
((ZU(1) << LG_SIZE_CLASS_GROUP) - 1);
|
|
|
|
|
2016-02-25 03:04:51 +08:00
|
|
|
szind_t index = NTBINS + grp + mod;
|
2014-10-06 08:54:10 +08:00
|
|
|
return (index);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2015-08-20 06:21:32 +08:00
|
|
|
JEMALLOC_ALWAYS_INLINE szind_t
|
2014-10-06 08:54:10 +08:00
|
|
|
size2index_lookup(size_t size)
|
|
|
|
{
|
|
|
|
|
|
|
|
assert(size <= LOOKUP_MAXCLASS);
|
|
|
|
{
|
2016-02-25 03:04:51 +08:00
|
|
|
szind_t ret = (size2index_tab[(size-1) >> LG_TINY_MIN]);
|
2014-10-06 08:54:10 +08:00
|
|
|
assert(ret == size2index_compute(size));
|
|
|
|
return (ret);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2015-08-20 06:21:32 +08:00
|
|
|
JEMALLOC_ALWAYS_INLINE szind_t
|
2014-10-06 08:54:10 +08:00
|
|
|
size2index(size_t size)
|
|
|
|
{
|
|
|
|
|
|
|
|
assert(size > 0);
|
|
|
|
if (likely(size <= LOOKUP_MAXCLASS))
|
|
|
|
return (size2index_lookup(size));
|
2015-06-24 09:47:07 +08:00
|
|
|
return (size2index_compute(size));
|
2014-10-06 08:54:10 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
JEMALLOC_INLINE size_t
|
2015-08-20 06:21:32 +08:00
|
|
|
index2size_compute(szind_t index)
|
2014-10-06 08:54:10 +08:00
|
|
|
{
|
|
|
|
|
|
|
|
#if (NTBINS > 0)
|
|
|
|
if (index < NTBINS)
|
|
|
|
return (ZU(1) << (LG_TINY_MAXCLASS - NTBINS + 1 + index));
|
|
|
|
#endif
|
|
|
|
{
|
|
|
|
size_t reduced_index = index - NTBINS;
|
|
|
|
size_t grp = reduced_index >> LG_SIZE_CLASS_GROUP;
|
|
|
|
size_t mod = reduced_index & ((ZU(1) << LG_SIZE_CLASS_GROUP) -
|
|
|
|
1);
|
|
|
|
|
|
|
|
size_t grp_size_mask = ~((!!grp)-1);
|
|
|
|
size_t grp_size = ((ZU(1) << (LG_QUANTUM +
|
|
|
|
(LG_SIZE_CLASS_GROUP-1))) << grp) & grp_size_mask;
|
|
|
|
|
|
|
|
size_t shift = (grp == 0) ? 1 : grp;
|
|
|
|
size_t lg_delta = shift + (LG_QUANTUM-1);
|
|
|
|
size_t mod_size = (mod+1) << lg_delta;
|
|
|
|
|
|
|
|
size_t usize = grp_size + mod_size;
|
|
|
|
return (usize);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
JEMALLOC_ALWAYS_INLINE size_t
|
2015-08-20 06:21:32 +08:00
|
|
|
index2size_lookup(szind_t index)
|
2014-10-06 08:54:10 +08:00
|
|
|
{
|
|
|
|
size_t ret = (size_t)index2size_tab[index];
|
|
|
|
assert(ret == index2size_compute(index));
|
|
|
|
return (ret);
|
|
|
|
}
|
|
|
|
|
|
|
|
JEMALLOC_ALWAYS_INLINE size_t
|
2015-08-20 06:21:32 +08:00
|
|
|
index2size(szind_t index)
|
2014-10-06 08:54:10 +08:00
|
|
|
{
|
|
|
|
|
2016-02-26 07:29:49 +08:00
|
|
|
assert(index < NSIZES);
|
2014-10-06 08:54:10 +08:00
|
|
|
return (index2size_lookup(index));
|
|
|
|
}
|
|
|
|
|
|
|
|
JEMALLOC_ALWAYS_INLINE size_t
|
|
|
|
s2u_compute(size_t size)
|
|
|
|
{
|
|
|
|
|
|
|
|
#if (NTBINS > 0)
|
|
|
|
if (size <= (ZU(1) << LG_TINY_MAXCLASS)) {
|
|
|
|
size_t lg_tmin = LG_TINY_MAXCLASS - NTBINS + 1;
|
2016-02-10 08:28:40 +08:00
|
|
|
size_t lg_ceil = lg_floor(pow2_ceil_zu(size));
|
2014-10-06 08:54:10 +08:00
|
|
|
return (lg_ceil < lg_tmin ? (ZU(1) << lg_tmin) :
|
|
|
|
(ZU(1) << lg_ceil));
|
2015-06-24 09:47:07 +08:00
|
|
|
}
|
2014-10-06 08:54:10 +08:00
|
|
|
#endif
|
|
|
|
{
|
2015-07-10 12:36:33 +08:00
|
|
|
size_t x = unlikely(ZI(size) < 0) ? ((size<<1) ?
|
|
|
|
(ZU(1)<<(LG_SIZEOF_PTR+3)) : ((ZU(1)<<(LG_SIZEOF_PTR+3))-1))
|
|
|
|
: lg_floor((size<<1)-1);
|
2014-10-06 08:54:10 +08:00
|
|
|
size_t lg_delta = (x < LG_SIZE_CLASS_GROUP + LG_QUANTUM + 1)
|
|
|
|
? LG_QUANTUM : x - LG_SIZE_CLASS_GROUP - 1;
|
|
|
|
size_t delta = ZU(1) << lg_delta;
|
|
|
|
size_t delta_mask = delta - 1;
|
|
|
|
size_t usize = (size + delta_mask) & ~delta_mask;
|
|
|
|
return (usize);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
JEMALLOC_ALWAYS_INLINE size_t
|
|
|
|
s2u_lookup(size_t size)
|
|
|
|
{
|
|
|
|
size_t ret = index2size_lookup(size2index_lookup(size));
|
|
|
|
|
|
|
|
assert(ret == s2u_compute(size));
|
|
|
|
return (ret);
|
|
|
|
}
|
|
|
|
|
2010-10-21 08:39:18 +08:00
|
|
|
/*
|
|
|
|
* Compute usable size that would result from allocating an object with the
|
|
|
|
* specified size.
|
|
|
|
*/
|
2013-01-23 00:45:43 +08:00
|
|
|
JEMALLOC_ALWAYS_INLINE size_t
|
2010-10-21 08:39:18 +08:00
|
|
|
s2u(size_t size)
|
|
|
|
{
|
|
|
|
|
2014-10-06 08:54:10 +08:00
|
|
|
assert(size > 0);
|
|
|
|
if (likely(size <= LOOKUP_MAXCLASS))
|
|
|
|
return (s2u_lookup(size));
|
2015-06-24 09:47:07 +08:00
|
|
|
return (s2u_compute(size));
|
2010-10-21 08:39:18 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Compute usable size that would result from allocating an object with the
|
|
|
|
* specified size and alignment.
|
|
|
|
*/
|
2013-01-23 00:45:43 +08:00
|
|
|
JEMALLOC_ALWAYS_INLINE size_t
|
2012-04-12 09:13:45 +08:00
|
|
|
sa2u(size_t size, size_t alignment)
|
2010-10-21 08:39:18 +08:00
|
|
|
{
|
|
|
|
size_t usize;
|
|
|
|
|
2012-04-12 09:13:45 +08:00
|
|
|
assert(alignment != 0 && ((alignment - 1) & alignment) == 0);
|
|
|
|
|
2014-10-06 08:54:10 +08:00
|
|
|
/* Try for a small size class. */
|
|
|
|
if (size <= SMALL_MAXCLASS && alignment < PAGE) {
|
|
|
|
/*
|
|
|
|
* Round size up to the nearest multiple of alignment.
|
|
|
|
*
|
|
|
|
* This done, we can take advantage of the fact that for each
|
|
|
|
* small size class, every object is aligned at the smallest
|
|
|
|
* power of two that is non-zero in the base two representation
|
|
|
|
* of the size. For example:
|
|
|
|
*
|
|
|
|
* Size | Base 2 | Minimum alignment
|
|
|
|
* -----+----------+------------------
|
|
|
|
* 96 | 1100000 | 32
|
|
|
|
* 144 | 10100000 | 32
|
|
|
|
* 192 | 11000000 | 64
|
|
|
|
*/
|
|
|
|
usize = s2u(ALIGNMENT_CEILING(size, alignment));
|
|
|
|
if (usize < LARGE_MINCLASS)
|
|
|
|
return (usize);
|
2010-10-21 08:39:18 +08:00
|
|
|
}
|
|
|
|
|
2014-10-06 08:54:10 +08:00
|
|
|
/* Try for a large size class. */
|
2015-09-12 11:50:20 +08:00
|
|
|
if (likely(size <= large_maxclass) && likely(alignment < chunksize)) {
|
2010-10-21 08:39:18 +08:00
|
|
|
/*
|
|
|
|
* We can't achieve subpage alignment, so round up alignment
|
2014-10-06 08:54:10 +08:00
|
|
|
* to the minimum that can actually be supported.
|
2010-10-21 08:39:18 +08:00
|
|
|
*/
|
|
|
|
alignment = PAGE_CEILING(alignment);
|
2014-10-06 08:54:10 +08:00
|
|
|
|
|
|
|
/* Make sure result is a large size class. */
|
|
|
|
usize = (size <= LARGE_MINCLASS) ? LARGE_MINCLASS : s2u(size);
|
|
|
|
|
2010-10-21 08:39:18 +08:00
|
|
|
/*
|
2014-10-06 08:54:10 +08:00
|
|
|
* Calculate the size of the over-size run that arena_palloc()
|
|
|
|
* would need to allocate in order to guarantee the alignment.
|
2010-10-21 08:39:18 +08:00
|
|
|
*/
|
2015-07-24 08:14:11 +08:00
|
|
|
if (usize + large_pad + alignment - PAGE <= arena_maxrun)
|
2014-10-06 08:54:10 +08:00
|
|
|
return (usize);
|
|
|
|
}
|
|
|
|
|
2016-02-26 07:29:49 +08:00
|
|
|
/* Huge size class. Beware of overflow. */
|
|
|
|
|
|
|
|
if (unlikely(alignment > HUGE_MAXCLASS))
|
|
|
|
return (0);
|
2014-10-06 08:54:10 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* We can't achieve subchunk alignment, so round up alignment to the
|
|
|
|
* minimum that can actually be supported.
|
|
|
|
*/
|
|
|
|
alignment = CHUNK_CEILING(alignment);
|
|
|
|
|
|
|
|
/* Make sure result is a huge size class. */
|
|
|
|
if (size <= chunksize)
|
|
|
|
usize = chunksize;
|
|
|
|
else {
|
|
|
|
usize = s2u(size);
|
|
|
|
if (usize < size) {
|
2010-10-21 08:39:18 +08:00
|
|
|
/* size_t overflow. */
|
|
|
|
return (0);
|
|
|
|
}
|
2014-10-06 08:54:10 +08:00
|
|
|
}
|
2010-10-21 08:39:18 +08:00
|
|
|
|
2014-10-06 08:54:10 +08:00
|
|
|
/*
|
|
|
|
* Calculate the multi-chunk mapping that huge_palloc() would need in
|
|
|
|
* order to guarantee the alignment.
|
|
|
|
*/
|
|
|
|
if (usize + alignment - PAGE < usize) {
|
|
|
|
/* size_t overflow. */
|
|
|
|
return (0);
|
2010-10-21 08:39:18 +08:00
|
|
|
}
|
2014-10-06 08:54:10 +08:00
|
|
|
return (usize);
|
2010-10-21 08:39:18 +08:00
|
|
|
}
|
|
|
|
|
2012-03-14 02:09:23 +08:00
|
|
|
/* Choose an arena based on a per-thread value. */
|
2010-01-17 01:53:50 +08:00
|
|
|
JEMALLOC_INLINE arena_t *
|
Refactor/fix arenas manipulation.
Abstract arenas access to use arena_get() (or a0get() where appropriate)
rather than directly reading e.g. arenas[ind]. Prior to the addition of
the arenas.extend mallctl, the worst possible outcome of directly
accessing arenas was a stale read, but arenas.extend may allocate and
assign a new array to arenas.
Add a tsd-based arenas_cache, which amortizes arenas reads. This
introduces some subtle bootstrapping issues, with tsd_boot() now being
split into tsd_boot[01]() to support tsd wrapper allocation
bootstrapping, as well as an arenas_cache_bypass tsd variable which
dynamically terminates allocation of arenas_cache itself.
Promote a0malloc(), a0calloc(), and a0free() to be generally useful for
internal allocation, and use them in several places (more may be
appropriate).
Abstract arena->nthreads management and fix a missing decrement during
thread destruction (recent tsd refactoring left arenas_cleanup()
unused).
Change arena_choose() to propagate OOM, and handle OOM in all callers.
This is important for providing consistent allocation behavior when the
MALLOCX_ARENA() flag is being used. Prior to this fix, it was possible
for an OOM to result in allocation silently allocating from a different
arena than the one specified.
2014-10-08 14:14:57 +08:00
|
|
|
arena_choose(tsd_t *tsd, arena_t *arena)
|
2010-01-17 01:53:50 +08:00
|
|
|
{
|
|
|
|
arena_t *ret;
|
|
|
|
|
2012-04-04 00:28:00 +08:00
|
|
|
if (arena != NULL)
|
|
|
|
return (arena);
|
|
|
|
|
Refactor/fix arenas manipulation.
Abstract arenas access to use arena_get() (or a0get() where appropriate)
rather than directly reading e.g. arenas[ind]. Prior to the addition of
the arenas.extend mallctl, the worst possible outcome of directly
accessing arenas was a stale read, but arenas.extend may allocate and
assign a new array to arenas.
Add a tsd-based arenas_cache, which amortizes arenas reads. This
introduces some subtle bootstrapping issues, with tsd_boot() now being
split into tsd_boot[01]() to support tsd wrapper allocation
bootstrapping, as well as an arenas_cache_bypass tsd variable which
dynamically terminates allocation of arenas_cache itself.
Promote a0malloc(), a0calloc(), and a0free() to be generally useful for
internal allocation, and use them in several places (more may be
appropriate).
Abstract arena->nthreads management and fix a missing decrement during
thread destruction (recent tsd refactoring left arenas_cleanup()
unused).
Change arena_choose() to propagate OOM, and handle OOM in all callers.
This is important for providing consistent allocation behavior when the
MALLOCX_ARENA() flag is being used. Prior to this fix, it was possible
for an OOM to result in allocation silently allocating from a different
arena than the one specified.
2014-10-08 14:14:57 +08:00
|
|
|
if (unlikely((ret = tsd_arena_get(tsd)) == NULL))
|
|
|
|
ret = arena_choose_hard(tsd);
|
2010-01-17 01:53:50 +08:00
|
|
|
|
|
|
|
return (ret);
|
|
|
|
}
|
Refactor/fix arenas manipulation.
Abstract arenas access to use arena_get() (or a0get() where appropriate)
rather than directly reading e.g. arenas[ind]. Prior to the addition of
the arenas.extend mallctl, the worst possible outcome of directly
accessing arenas was a stale read, but arenas.extend may allocate and
assign a new array to arenas.
Add a tsd-based arenas_cache, which amortizes arenas reads. This
introduces some subtle bootstrapping issues, with tsd_boot() now being
split into tsd_boot[01]() to support tsd wrapper allocation
bootstrapping, as well as an arenas_cache_bypass tsd variable which
dynamically terminates allocation of arenas_cache itself.
Promote a0malloc(), a0calloc(), and a0free() to be generally useful for
internal allocation, and use them in several places (more may be
appropriate).
Abstract arena->nthreads management and fix a missing decrement during
thread destruction (recent tsd refactoring left arenas_cleanup()
unused).
Change arena_choose() to propagate OOM, and handle OOM in all callers.
This is important for providing consistent allocation behavior when the
MALLOCX_ARENA() flag is being used. Prior to this fix, it was possible
for an OOM to result in allocation silently allocating from a different
arena than the one specified.
2014-10-08 14:14:57 +08:00
|
|
|
|
2016-02-20 11:37:10 +08:00
|
|
|
JEMALLOC_INLINE arena_tdata_t *
|
|
|
|
arena_tdata_get(tsd_t *tsd, unsigned ind, bool refresh_if_missing)
|
|
|
|
{
|
|
|
|
arena_tdata_t *tdata;
|
|
|
|
arena_tdata_t *arenas_tdata = tsd_arenas_tdata_get(tsd);
|
|
|
|
|
|
|
|
if (unlikely(arenas_tdata == NULL)) {
|
|
|
|
/* arenas_tdata hasn't been initialized yet. */
|
|
|
|
return (arena_tdata_get_hard(tsd, ind));
|
|
|
|
}
|
|
|
|
if (unlikely(ind >= tsd_narenas_tdata_get(tsd))) {
|
|
|
|
/*
|
|
|
|
* ind is invalid, cache is old (too small), or tdata to be
|
|
|
|
* initialized.
|
|
|
|
*/
|
|
|
|
return (refresh_if_missing ? arena_tdata_get_hard(tsd, ind) :
|
|
|
|
NULL);
|
|
|
|
}
|
|
|
|
|
|
|
|
tdata = &arenas_tdata[ind];
|
|
|
|
if (likely(tdata != NULL) || !refresh_if_missing)
|
|
|
|
return (tdata);
|
|
|
|
return (arena_tdata_get_hard(tsd, ind));
|
|
|
|
}
|
|
|
|
|
Refactor/fix arenas manipulation.
Abstract arenas access to use arena_get() (or a0get() where appropriate)
rather than directly reading e.g. arenas[ind]. Prior to the addition of
the arenas.extend mallctl, the worst possible outcome of directly
accessing arenas was a stale read, but arenas.extend may allocate and
assign a new array to arenas.
Add a tsd-based arenas_cache, which amortizes arenas reads. This
introduces some subtle bootstrapping issues, with tsd_boot() now being
split into tsd_boot[01]() to support tsd wrapper allocation
bootstrapping, as well as an arenas_cache_bypass tsd variable which
dynamically terminates allocation of arenas_cache itself.
Promote a0malloc(), a0calloc(), and a0free() to be generally useful for
internal allocation, and use them in several places (more may be
appropriate).
Abstract arena->nthreads management and fix a missing decrement during
thread destruction (recent tsd refactoring left arenas_cleanup()
unused).
Change arena_choose() to propagate OOM, and handle OOM in all callers.
This is important for providing consistent allocation behavior when the
MALLOCX_ARENA() flag is being used. Prior to this fix, it was possible
for an OOM to result in allocation silently allocating from a different
arena than the one specified.
2014-10-08 14:14:57 +08:00
|
|
|
JEMALLOC_INLINE arena_t *
|
2016-02-25 15:58:10 +08:00
|
|
|
arena_get(unsigned ind, bool init_if_missing)
|
Refactor/fix arenas manipulation.
Abstract arenas access to use arena_get() (or a0get() where appropriate)
rather than directly reading e.g. arenas[ind]. Prior to the addition of
the arenas.extend mallctl, the worst possible outcome of directly
accessing arenas was a stale read, but arenas.extend may allocate and
assign a new array to arenas.
Add a tsd-based arenas_cache, which amortizes arenas reads. This
introduces some subtle bootstrapping issues, with tsd_boot() now being
split into tsd_boot[01]() to support tsd wrapper allocation
bootstrapping, as well as an arenas_cache_bypass tsd variable which
dynamically terminates allocation of arenas_cache itself.
Promote a0malloc(), a0calloc(), and a0free() to be generally useful for
internal allocation, and use them in several places (more may be
appropriate).
Abstract arena->nthreads management and fix a missing decrement during
thread destruction (recent tsd refactoring left arenas_cleanup()
unused).
Change arena_choose() to propagate OOM, and handle OOM in all callers.
This is important for providing consistent allocation behavior when the
MALLOCX_ARENA() flag is being used. Prior to this fix, it was possible
for an OOM to result in allocation silently allocating from a different
arena than the one specified.
2014-10-08 14:14:57 +08:00
|
|
|
{
|
2016-02-25 15:58:10 +08:00
|
|
|
arena_t *ret;
|
Refactor/fix arenas manipulation.
Abstract arenas access to use arena_get() (or a0get() where appropriate)
rather than directly reading e.g. arenas[ind]. Prior to the addition of
the arenas.extend mallctl, the worst possible outcome of directly
accessing arenas was a stale read, but arenas.extend may allocate and
assign a new array to arenas.
Add a tsd-based arenas_cache, which amortizes arenas reads. This
introduces some subtle bootstrapping issues, with tsd_boot() now being
split into tsd_boot[01]() to support tsd wrapper allocation
bootstrapping, as well as an arenas_cache_bypass tsd variable which
dynamically terminates allocation of arenas_cache itself.
Promote a0malloc(), a0calloc(), and a0free() to be generally useful for
internal allocation, and use them in several places (more may be
appropriate).
Abstract arena->nthreads management and fix a missing decrement during
thread destruction (recent tsd refactoring left arenas_cleanup()
unused).
Change arena_choose() to propagate OOM, and handle OOM in all callers.
This is important for providing consistent allocation behavior when the
MALLOCX_ARENA() flag is being used. Prior to this fix, it was possible
for an OOM to result in allocation silently allocating from a different
arena than the one specified.
2014-10-08 14:14:57 +08:00
|
|
|
|
2016-02-25 15:58:10 +08:00
|
|
|
assert(ind <= MALLOCX_ARENA_MAX);
|
2016-02-20 11:37:10 +08:00
|
|
|
|
2016-02-25 15:58:10 +08:00
|
|
|
ret = arenas[ind];
|
|
|
|
if (unlikely(ret == NULL)) {
|
|
|
|
ret = atomic_read_p((void *)&arenas[ind]);
|
|
|
|
if (init_if_missing && unlikely(ret == NULL))
|
|
|
|
ret = arena_init(ind);
|
|
|
|
}
|
|
|
|
return (ret);
|
Refactor/fix arenas manipulation.
Abstract arenas access to use arena_get() (or a0get() where appropriate)
rather than directly reading e.g. arenas[ind]. Prior to the addition of
the arenas.extend mallctl, the worst possible outcome of directly
accessing arenas was a stale read, but arenas.extend may allocate and
assign a new array to arenas.
Add a tsd-based arenas_cache, which amortizes arenas reads. This
introduces some subtle bootstrapping issues, with tsd_boot() now being
split into tsd_boot[01]() to support tsd wrapper allocation
bootstrapping, as well as an arenas_cache_bypass tsd variable which
dynamically terminates allocation of arenas_cache itself.
Promote a0malloc(), a0calloc(), and a0free() to be generally useful for
internal allocation, and use them in several places (more may be
appropriate).
Abstract arena->nthreads management and fix a missing decrement during
thread destruction (recent tsd refactoring left arenas_cleanup()
unused).
Change arena_choose() to propagate OOM, and handle OOM in all callers.
This is important for providing consistent allocation behavior when the
MALLOCX_ARENA() flag is being used. Prior to this fix, it was possible
for an OOM to result in allocation silently allocating from a different
arena than the one specified.
2014-10-08 14:14:57 +08:00
|
|
|
}
|
2016-02-20 12:09:31 +08:00
|
|
|
|
|
|
|
JEMALLOC_INLINE ticker_t *
|
|
|
|
decay_ticker_get(tsd_t *tsd, unsigned ind)
|
|
|
|
{
|
|
|
|
arena_tdata_t *tdata;
|
|
|
|
|
|
|
|
tdata = arena_tdata_get(tsd, ind, true);
|
|
|
|
if (unlikely(tdata == NULL))
|
|
|
|
return (NULL);
|
|
|
|
return (&tdata->decay_ticker);
|
|
|
|
}
|
2011-02-14 10:11:54 +08:00
|
|
|
#endif
|
2010-01-17 01:53:50 +08:00
|
|
|
|
Use bitmaps to track small regions.
The previous free list implementation, which embedded singly linked
lists in available regions, had the unfortunate side effect of causing
many cache misses during thread cache fills. Fix this in two places:
- arena_run_t: Use a new bitmap implementation to track which regions
are available. Furthermore, revert to preferring the
lowest available region (as jemalloc did with its old
bitmap-based approach).
- tcache_t: Move read-only tcache_bin_t metadata into
tcache_bin_info_t, and add a contiguous array of pointers
to tcache_t in order to track cached objects. This
substantially increases the size of tcache_t, but results
in much higher data locality for common tcache operations.
As a side benefit, it is again possible to efficiently
flush the least recently used cached objects, so this
change changes flushing from MRU to LRU.
The new bitmap implementation uses a multi-level summary approach to
make finding the lowest available region very fast. In practice,
bitmaps only have one or two levels, though the implementation is
general enough to handle extremely large bitmaps, mainly so that large
page sizes can still be entertained.
Fix tcache_bin_flush_large() to always flush statistics, in the same way
that tcache_bin_flush_small() was recently fixed.
Use JEMALLOC_DEBUG rather than NDEBUG.
Add dassert(), and use it for debug-only asserts.
2011-03-17 01:30:13 +08:00
|
|
|
#include "jemalloc/internal/bitmap.h"
|
2012-05-02 15:30:36 +08:00
|
|
|
/*
|
2014-10-06 08:54:10 +08:00
|
|
|
* Include portions of arena.h interleaved with tcache.h in order to resolve
|
|
|
|
* circular dependencies.
|
2012-05-02 15:30:36 +08:00
|
|
|
*/
|
2014-10-06 08:54:10 +08:00
|
|
|
#define JEMALLOC_ARENA_INLINE_A
|
2010-02-12 06:45:59 +08:00
|
|
|
#include "jemalloc/internal/arena.h"
|
2014-10-06 08:54:10 +08:00
|
|
|
#undef JEMALLOC_ARENA_INLINE_A
|
2014-04-17 08:14:33 +08:00
|
|
|
#include "jemalloc/internal/tcache.h"
|
2014-10-06 08:54:10 +08:00
|
|
|
#define JEMALLOC_ARENA_INLINE_B
|
2014-04-17 08:14:33 +08:00
|
|
|
#include "jemalloc/internal/arena.h"
|
2014-10-06 08:54:10 +08:00
|
|
|
#undef JEMALLOC_ARENA_INLINE_B
|
2010-02-12 06:45:59 +08:00
|
|
|
#include "jemalloc/internal/hash.h"
|
2012-04-06 15:35:09 +08:00
|
|
|
#include "jemalloc/internal/quarantine.h"
|
2010-01-17 01:53:50 +08:00
|
|
|
|
|
|
|
#ifndef JEMALLOC_ENABLE_INLINE
|
2014-11-28 03:22:36 +08:00
|
|
|
arena_t *iaalloc(const void *ptr);
|
|
|
|
size_t isalloc(const void *ptr, bool demote);
|
2015-10-28 06:12:10 +08:00
|
|
|
void *iallocztm(tsd_t *tsd, size_t size, szind_t ind, bool zero,
|
|
|
|
tcache_t *tcache, bool is_metadata, arena_t *arena, bool slow_path);
|
|
|
|
void *imalloct(tsd_t *tsd, size_t size, szind_t ind, tcache_t *tcache,
|
|
|
|
arena_t *arena);
|
|
|
|
void *imalloc(tsd_t *tsd, size_t size, szind_t ind, bool slow_path);
|
|
|
|
void *icalloct(tsd_t *tsd, size_t size, szind_t ind, tcache_t *tcache,
|
|
|
|
arena_t *arena);
|
|
|
|
void *icalloc(tsd_t *tsd, size_t size, szind_t ind);
|
2014-11-28 03:22:36 +08:00
|
|
|
void *ipallocztm(tsd_t *tsd, size_t usize, size_t alignment, bool zero,
|
2015-01-30 07:30:47 +08:00
|
|
|
tcache_t *tcache, bool is_metadata, arena_t *arena);
|
2014-09-23 12:09:23 +08:00
|
|
|
void *ipalloct(tsd_t *tsd, size_t usize, size_t alignment, bool zero,
|
2015-01-30 07:30:47 +08:00
|
|
|
tcache_t *tcache, arena_t *arena);
|
2014-09-23 12:09:23 +08:00
|
|
|
void *ipalloc(tsd_t *tsd, size_t usize, size_t alignment, bool zero);
|
2012-04-06 15:35:09 +08:00
|
|
|
size_t ivsalloc(const void *ptr, bool demote);
|
|
|
|
size_t u2rz(size_t usize);
|
|
|
|
size_t p2rz(const void *ptr);
|
2015-10-28 06:12:10 +08:00
|
|
|
void idalloctm(tsd_t *tsd, void *ptr, tcache_t *tcache, bool is_metadata,
|
|
|
|
bool slow_path);
|
2015-01-30 07:30:47 +08:00
|
|
|
void idalloct(tsd_t *tsd, void *ptr, tcache_t *tcache);
|
2014-09-23 12:09:23 +08:00
|
|
|
void idalloc(tsd_t *tsd, void *ptr);
|
2015-10-28 06:12:10 +08:00
|
|
|
void iqalloc(tsd_t *tsd, void *ptr, tcache_t *tcache, bool slow_path);
|
2015-01-30 07:30:47 +08:00
|
|
|
void isdalloct(tsd_t *tsd, void *ptr, size_t size, tcache_t *tcache);
|
|
|
|
void isqalloc(tsd_t *tsd, void *ptr, size_t size, tcache_t *tcache);
|
2014-09-23 12:09:23 +08:00
|
|
|
void *iralloct_realign(tsd_t *tsd, void *ptr, size_t oldsize, size_t size,
|
2015-01-30 07:30:47 +08:00
|
|
|
size_t extra, size_t alignment, bool zero, tcache_t *tcache,
|
2014-10-25 01:18:57 +08:00
|
|
|
arena_t *arena);
|
2015-01-30 07:30:47 +08:00
|
|
|
void *iralloct(tsd_t *tsd, void *ptr, size_t oldsize, size_t size,
|
|
|
|
size_t alignment, bool zero, tcache_t *tcache, arena_t *arena);
|
2014-10-25 01:18:57 +08:00
|
|
|
void *iralloc(tsd_t *tsd, void *ptr, size_t oldsize, size_t size,
|
|
|
|
size_t alignment, bool zero);
|
2016-02-20 12:09:31 +08:00
|
|
|
bool ixalloc(tsd_t *tsd, void *ptr, size_t oldsize, size_t size,
|
|
|
|
size_t extra, size_t alignment, bool zero);
|
2010-01-17 01:53:50 +08:00
|
|
|
#endif
|
|
|
|
|
|
|
|
#if (defined(JEMALLOC_ENABLE_INLINE) || defined(JEMALLOC_C_))
|
2014-11-28 03:22:36 +08:00
|
|
|
JEMALLOC_ALWAYS_INLINE arena_t *
|
|
|
|
iaalloc(const void *ptr)
|
|
|
|
{
|
|
|
|
|
|
|
|
assert(ptr != NULL);
|
|
|
|
|
2015-02-13 06:06:37 +08:00
|
|
|
return (arena_aalloc(ptr));
|
2014-11-28 03:22:36 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Typical usage:
|
|
|
|
* void *ptr = [...]
|
|
|
|
* size_t sz = isalloc(ptr, config_prof);
|
|
|
|
*/
|
|
|
|
JEMALLOC_ALWAYS_INLINE size_t
|
|
|
|
isalloc(const void *ptr, bool demote)
|
|
|
|
{
|
|
|
|
|
|
|
|
assert(ptr != NULL);
|
|
|
|
/* Demotion only makes sense if config_prof is true. */
|
|
|
|
assert(config_prof || !demote);
|
|
|
|
|
2015-02-13 06:06:37 +08:00
|
|
|
return (arena_salloc(ptr, demote));
|
2014-11-28 03:22:36 +08:00
|
|
|
}
|
|
|
|
|
2013-01-23 00:45:43 +08:00
|
|
|
JEMALLOC_ALWAYS_INLINE void *
|
2015-10-28 06:12:10 +08:00
|
|
|
iallocztm(tsd_t *tsd, size_t size, szind_t ind, bool zero, tcache_t *tcache,
|
|
|
|
bool is_metadata, arena_t *arena, bool slow_path)
|
2010-01-17 01:53:50 +08:00
|
|
|
{
|
2014-11-28 03:22:36 +08:00
|
|
|
void *ret;
|
2010-01-17 01:53:50 +08:00
|
|
|
|
|
|
|
assert(size != 0);
|
|
|
|
|
2015-10-28 06:12:10 +08:00
|
|
|
ret = arena_malloc(tsd, arena, size, ind, zero, tcache, slow_path);
|
2014-11-28 03:22:36 +08:00
|
|
|
if (config_stats && is_metadata && likely(ret != NULL)) {
|
|
|
|
arena_metadata_allocated_add(iaalloc(ret), isalloc(ret,
|
|
|
|
config_prof));
|
|
|
|
}
|
|
|
|
return (ret);
|
|
|
|
}
|
|
|
|
|
|
|
|
JEMALLOC_ALWAYS_INLINE void *
|
2015-10-28 06:12:10 +08:00
|
|
|
imalloct(tsd_t *tsd, size_t size, szind_t ind, tcache_t *tcache, arena_t *arena)
|
2014-11-28 03:22:36 +08:00
|
|
|
{
|
|
|
|
|
2015-10-28 06:12:10 +08:00
|
|
|
return (iallocztm(tsd, size, ind, false, tcache, false, arena, true));
|
2010-01-17 01:53:50 +08:00
|
|
|
}
|
|
|
|
|
2013-01-23 00:45:43 +08:00
|
|
|
JEMALLOC_ALWAYS_INLINE void *
|
2015-10-28 06:12:10 +08:00
|
|
|
imalloc(tsd_t *tsd, size_t size, szind_t ind, bool slow_path)
|
2012-10-12 04:53:15 +08:00
|
|
|
{
|
|
|
|
|
2015-10-28 06:12:10 +08:00
|
|
|
return (iallocztm(tsd, size, ind, false, tcache_get(tsd, true), false,
|
|
|
|
NULL, slow_path));
|
2012-10-12 04:53:15 +08:00
|
|
|
}
|
|
|
|
|
2013-01-23 00:45:43 +08:00
|
|
|
JEMALLOC_ALWAYS_INLINE void *
|
2015-10-28 06:12:10 +08:00
|
|
|
icalloct(tsd_t *tsd, size_t size, szind_t ind, tcache_t *tcache, arena_t *arena)
|
2010-01-17 01:53:50 +08:00
|
|
|
{
|
|
|
|
|
2015-10-28 06:12:10 +08:00
|
|
|
return (iallocztm(tsd, size, ind, true, tcache, false, arena, true));
|
2010-01-17 01:53:50 +08:00
|
|
|
}
|
|
|
|
|
2013-01-23 00:45:43 +08:00
|
|
|
JEMALLOC_ALWAYS_INLINE void *
|
2015-10-28 06:12:10 +08:00
|
|
|
icalloc(tsd_t *tsd, size_t size, szind_t ind)
|
2012-10-12 04:53:15 +08:00
|
|
|
{
|
|
|
|
|
2015-10-28 06:12:10 +08:00
|
|
|
return (iallocztm(tsd, size, ind, true, tcache_get(tsd, true), false,
|
|
|
|
NULL, true));
|
2012-10-12 04:53:15 +08:00
|
|
|
}
|
|
|
|
|
2013-01-23 00:45:43 +08:00
|
|
|
JEMALLOC_ALWAYS_INLINE void *
|
2014-11-28 03:22:36 +08:00
|
|
|
ipallocztm(tsd_t *tsd, size_t usize, size_t alignment, bool zero,
|
2015-01-30 07:30:47 +08:00
|
|
|
tcache_t *tcache, bool is_metadata, arena_t *arena)
|
2010-01-17 01:53:50 +08:00
|
|
|
{
|
2010-02-11 02:37:56 +08:00
|
|
|
void *ret;
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2011-03-23 15:37:29 +08:00
|
|
|
assert(usize != 0);
|
2012-04-12 09:13:45 +08:00
|
|
|
assert(usize == sa2u(usize, alignment));
|
2011-03-23 15:37:29 +08:00
|
|
|
|
2015-02-13 06:06:37 +08:00
|
|
|
ret = arena_palloc(tsd, arena, usize, alignment, zero, tcache);
|
2012-04-12 09:13:45 +08:00
|
|
|
assert(ALIGNMENT_ADDR2BASE(ret, alignment) == ret);
|
2014-11-28 03:22:36 +08:00
|
|
|
if (config_stats && is_metadata && likely(ret != NULL)) {
|
|
|
|
arena_metadata_allocated_add(iaalloc(ret), isalloc(ret,
|
|
|
|
config_prof));
|
|
|
|
}
|
2010-02-11 02:37:56 +08:00
|
|
|
return (ret);
|
2010-01-17 01:53:50 +08:00
|
|
|
}
|
|
|
|
|
2013-01-23 00:45:43 +08:00
|
|
|
JEMALLOC_ALWAYS_INLINE void *
|
2015-01-30 07:30:47 +08:00
|
|
|
ipalloct(tsd_t *tsd, size_t usize, size_t alignment, bool zero,
|
|
|
|
tcache_t *tcache, arena_t *arena)
|
2012-10-12 04:53:15 +08:00
|
|
|
{
|
|
|
|
|
2015-01-30 07:30:47 +08:00
|
|
|
return (ipallocztm(tsd, usize, alignment, zero, tcache, false, arena));
|
2012-10-12 04:53:15 +08:00
|
|
|
}
|
|
|
|
|
2014-11-28 03:22:36 +08:00
|
|
|
JEMALLOC_ALWAYS_INLINE void *
|
|
|
|
ipalloc(tsd_t *tsd, size_t usize, size_t alignment, bool zero)
|
2010-01-17 01:53:50 +08:00
|
|
|
{
|
|
|
|
|
2016-02-02 18:52:07 +08:00
|
|
|
return (ipallocztm(tsd, usize, alignment, zero, tcache_get(tsd, true),
|
|
|
|
false, NULL));
|
2010-01-17 01:53:50 +08:00
|
|
|
}
|
2010-02-11 02:37:56 +08:00
|
|
|
|
2013-01-23 00:45:43 +08:00
|
|
|
JEMALLOC_ALWAYS_INLINE size_t
|
2012-04-06 15:35:09 +08:00
|
|
|
ivsalloc(const void *ptr, bool demote)
|
2010-09-06 01:35:13 +08:00
|
|
|
{
|
Move centralized chunk management into arenas.
Migrate all centralized data structures related to huge allocations and
recyclable chunks into arena_t, so that each arena can manage huge
allocations and recyclable virtual memory completely independently of
other arenas.
Add chunk node caching to arenas, in order to avoid contention on the
base allocator.
Use chunks_rtree to look up huge allocations rather than a red-black
tree. Maintain a per arena unsorted list of huge allocations (which
will be needed to enumerate huge allocations during arena reset).
Remove the --enable-ivsalloc option, make ivsalloc() always available,
and use it for size queries if --enable-debug is enabled. The only
practical implications to this removal are that 1) ivsalloc() is now
always available during live debugging (and the underlying radix tree is
available during core-based debugging), and 2) size query validation can
no longer be enabled independent of --enable-debug.
Remove the stats.chunks.{current,total,high} mallctls, and replace their
underlying statistics with simpler atomically updated counters used
exclusively for gdump triggering. These statistics are no longer very
useful because each arena manages chunks independently, and per arena
statistics provide similar information.
Simplify chunk synchronization code, now that base chunk allocation
cannot cause recursive lock acquisition.
2015-02-12 04:24:27 +08:00
|
|
|
extent_node_t *node;
|
2010-09-06 01:35:13 +08:00
|
|
|
|
|
|
|
/* Return 0 if ptr is not within a chunk managed by jemalloc. */
|
2015-05-16 08:02:30 +08:00
|
|
|
node = chunk_lookup(ptr, false);
|
Move centralized chunk management into arenas.
Migrate all centralized data structures related to huge allocations and
recyclable chunks into arena_t, so that each arena can manage huge
allocations and recyclable virtual memory completely independently of
other arenas.
Add chunk node caching to arenas, in order to avoid contention on the
base allocator.
Use chunks_rtree to look up huge allocations rather than a red-black
tree. Maintain a per arena unsorted list of huge allocations (which
will be needed to enumerate huge allocations during arena reset).
Remove the --enable-ivsalloc option, make ivsalloc() always available,
and use it for size queries if --enable-debug is enabled. The only
practical implications to this removal are that 1) ivsalloc() is now
always available during live debugging (and the underlying radix tree is
available during core-based debugging), and 2) size query validation can
no longer be enabled independent of --enable-debug.
Remove the stats.chunks.{current,total,high} mallctls, and replace their
underlying statistics with simpler atomically updated counters used
exclusively for gdump triggering. These statistics are no longer very
useful because each arena manages chunks independently, and per arena
statistics provide similar information.
Simplify chunk synchronization code, now that base chunk allocation
cannot cause recursive lock acquisition.
2015-02-12 04:24:27 +08:00
|
|
|
if (node == NULL)
|
2010-09-06 01:35:13 +08:00
|
|
|
return (0);
|
Move centralized chunk management into arenas.
Migrate all centralized data structures related to huge allocations and
recyclable chunks into arena_t, so that each arena can manage huge
allocations and recyclable virtual memory completely independently of
other arenas.
Add chunk node caching to arenas, in order to avoid contention on the
base allocator.
Use chunks_rtree to look up huge allocations rather than a red-black
tree. Maintain a per arena unsorted list of huge allocations (which
will be needed to enumerate huge allocations during arena reset).
Remove the --enable-ivsalloc option, make ivsalloc() always available,
and use it for size queries if --enable-debug is enabled. The only
practical implications to this removal are that 1) ivsalloc() is now
always available during live debugging (and the underlying radix tree is
available during core-based debugging), and 2) size query validation can
no longer be enabled independent of --enable-debug.
Remove the stats.chunks.{current,total,high} mallctls, and replace their
underlying statistics with simpler atomically updated counters used
exclusively for gdump triggering. These statistics are no longer very
useful because each arena manages chunks independently, and per arena
statistics provide similar information.
Simplify chunk synchronization code, now that base chunk allocation
cannot cause recursive lock acquisition.
2015-02-12 04:24:27 +08:00
|
|
|
/* Only arena chunks should be looked up via interior pointers. */
|
2015-02-16 10:04:46 +08:00
|
|
|
assert(extent_node_addr_get(node) == ptr ||
|
|
|
|
extent_node_achunk_get(node));
|
2010-09-06 01:35:13 +08:00
|
|
|
|
2012-04-06 15:35:09 +08:00
|
|
|
return (isalloc(ptr, demote));
|
|
|
|
}
|
|
|
|
|
|
|
|
JEMALLOC_INLINE size_t
|
|
|
|
u2rz(size_t usize)
|
|
|
|
{
|
|
|
|
size_t ret;
|
|
|
|
|
|
|
|
if (usize <= SMALL_MAXCLASS) {
|
2015-08-20 06:21:32 +08:00
|
|
|
szind_t binind = size2index(usize);
|
2012-04-06 15:35:09 +08:00
|
|
|
ret = arena_bin_info[binind].redzone_size;
|
|
|
|
} else
|
|
|
|
ret = 0;
|
|
|
|
|
|
|
|
return (ret);
|
|
|
|
}
|
|
|
|
|
|
|
|
JEMALLOC_INLINE size_t
|
|
|
|
p2rz(const void *ptr)
|
|
|
|
{
|
|
|
|
size_t usize = isalloc(ptr, false);
|
|
|
|
|
|
|
|
return (u2rz(usize));
|
2010-09-06 01:35:13 +08:00
|
|
|
}
|
|
|
|
|
2013-01-23 00:45:43 +08:00
|
|
|
JEMALLOC_ALWAYS_INLINE void
|
2015-10-28 06:12:10 +08:00
|
|
|
idalloctm(tsd_t *tsd, void *ptr, tcache_t *tcache, bool is_metadata,
|
|
|
|
bool slow_path)
|
2010-09-21 10:20:48 +08:00
|
|
|
{
|
|
|
|
|
2012-04-03 06:18:24 +08:00
|
|
|
assert(ptr != NULL);
|
2014-11-28 03:22:36 +08:00
|
|
|
if (config_stats && is_metadata) {
|
|
|
|
arena_metadata_allocated_sub(iaalloc(ptr), isalloc(ptr,
|
|
|
|
config_prof));
|
|
|
|
}
|
2012-04-03 06:18:24 +08:00
|
|
|
|
2015-10-28 06:12:10 +08:00
|
|
|
arena_dalloc(tsd, ptr, tcache, slow_path);
|
2010-09-21 10:20:48 +08:00
|
|
|
}
|
|
|
|
|
2014-08-29 03:41:48 +08:00
|
|
|
JEMALLOC_ALWAYS_INLINE void
|
2015-01-30 07:30:47 +08:00
|
|
|
idalloct(tsd_t *tsd, void *ptr, tcache_t *tcache)
|
2014-08-29 03:41:48 +08:00
|
|
|
{
|
|
|
|
|
2015-10-28 06:12:10 +08:00
|
|
|
idalloctm(tsd, ptr, tcache, false, true);
|
2014-08-29 03:41:48 +08:00
|
|
|
}
|
|
|
|
|
2013-01-23 00:45:43 +08:00
|
|
|
JEMALLOC_ALWAYS_INLINE void
|
2014-09-23 12:09:23 +08:00
|
|
|
idalloc(tsd_t *tsd, void *ptr)
|
2012-10-12 04:53:15 +08:00
|
|
|
{
|
|
|
|
|
2015-10-28 06:12:10 +08:00
|
|
|
idalloctm(tsd, ptr, tcache_get(tsd, false), false, true);
|
2012-10-12 04:53:15 +08:00
|
|
|
}
|
|
|
|
|
2013-01-23 00:45:43 +08:00
|
|
|
JEMALLOC_ALWAYS_INLINE void
|
2015-10-28 06:12:10 +08:00
|
|
|
iqalloc(tsd_t *tsd, void *ptr, tcache_t *tcache, bool slow_path)
|
2012-04-06 15:35:09 +08:00
|
|
|
{
|
|
|
|
|
2015-10-28 06:12:10 +08:00
|
|
|
if (slow_path && config_fill && unlikely(opt_quarantine))
|
2014-09-23 12:09:23 +08:00
|
|
|
quarantine(tsd, ptr);
|
2012-04-06 15:35:09 +08:00
|
|
|
else
|
2015-10-28 06:12:10 +08:00
|
|
|
idalloctm(tsd, ptr, tcache, false, slow_path);
|
2014-11-28 03:22:36 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
JEMALLOC_ALWAYS_INLINE void
|
2015-01-30 07:30:47 +08:00
|
|
|
isdalloct(tsd_t *tsd, void *ptr, size_t size, tcache_t *tcache)
|
2014-11-28 03:22:36 +08:00
|
|
|
{
|
|
|
|
|
2015-02-13 06:06:37 +08:00
|
|
|
arena_sdalloc(tsd, ptr, size, tcache);
|
2012-10-12 04:53:15 +08:00
|
|
|
}
|
|
|
|
|
2014-08-29 03:41:48 +08:00
|
|
|
JEMALLOC_ALWAYS_INLINE void
|
2015-01-30 07:30:47 +08:00
|
|
|
isqalloc(tsd_t *tsd, void *ptr, size_t size, tcache_t *tcache)
|
2014-08-29 03:41:48 +08:00
|
|
|
{
|
|
|
|
|
2014-09-12 07:20:44 +08:00
|
|
|
if (config_fill && unlikely(opt_quarantine))
|
2014-09-23 12:09:23 +08:00
|
|
|
quarantine(tsd, ptr);
|
2014-08-29 03:41:48 +08:00
|
|
|
else
|
2015-01-30 07:30:47 +08:00
|
|
|
isdalloct(tsd, ptr, size, tcache);
|
2014-08-29 03:41:48 +08:00
|
|
|
}
|
|
|
|
|
2014-01-13 07:05:44 +08:00
|
|
|
JEMALLOC_ALWAYS_INLINE void *
|
2014-09-23 12:09:23 +08:00
|
|
|
iralloct_realign(tsd_t *tsd, void *ptr, size_t oldsize, size_t size,
|
2015-01-30 07:30:47 +08:00
|
|
|
size_t extra, size_t alignment, bool zero, tcache_t *tcache, arena_t *arena)
|
2014-01-13 07:05:44 +08:00
|
|
|
{
|
|
|
|
void *p;
|
|
|
|
size_t usize, copysize;
|
|
|
|
|
|
|
|
usize = sa2u(size + extra, alignment);
|
2016-02-26 07:29:49 +08:00
|
|
|
if (unlikely(usize == 0 || usize > HUGE_MAXCLASS))
|
2014-01-13 07:05:44 +08:00
|
|
|
return (NULL);
|
2015-01-30 07:30:47 +08:00
|
|
|
p = ipalloct(tsd, usize, alignment, zero, tcache, arena);
|
2014-01-13 07:05:44 +08:00
|
|
|
if (p == NULL) {
|
|
|
|
if (extra == 0)
|
|
|
|
return (NULL);
|
|
|
|
/* Try again, without extra this time. */
|
|
|
|
usize = sa2u(size, alignment);
|
2016-02-26 07:29:49 +08:00
|
|
|
if (unlikely(usize == 0 || usize > HUGE_MAXCLASS))
|
2014-01-13 07:05:44 +08:00
|
|
|
return (NULL);
|
2015-01-30 07:30:47 +08:00
|
|
|
p = ipalloct(tsd, usize, alignment, zero, tcache, arena);
|
2014-01-13 07:05:44 +08:00
|
|
|
if (p == NULL)
|
|
|
|
return (NULL);
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* Copy at most size bytes (not size+extra), since the caller has no
|
|
|
|
* expectation that the extra bytes will be reliably preserved.
|
|
|
|
*/
|
|
|
|
copysize = (size < oldsize) ? size : oldsize;
|
|
|
|
memcpy(p, ptr, copysize);
|
2015-01-30 07:30:47 +08:00
|
|
|
isqalloc(tsd, ptr, oldsize, tcache);
|
2014-01-13 07:05:44 +08:00
|
|
|
return (p);
|
|
|
|
}
|
|
|
|
|
2013-01-23 00:45:43 +08:00
|
|
|
JEMALLOC_ALWAYS_INLINE void *
|
2014-10-25 01:18:57 +08:00
|
|
|
iralloct(tsd_t *tsd, void *ptr, size_t oldsize, size_t size, size_t alignment,
|
2015-01-30 07:30:47 +08:00
|
|
|
bool zero, tcache_t *tcache, arena_t *arena)
|
2010-02-11 02:37:56 +08:00
|
|
|
{
|
|
|
|
|
|
|
|
assert(ptr != NULL);
|
|
|
|
assert(size != 0);
|
|
|
|
|
Add {,r,s,d}allocm().
Add allocm(), rallocm(), sallocm(), and dallocm(), which are a
functional superset of malloc(), calloc(), posix_memalign(),
malloc_usable_size(), and free().
2010-09-18 06:46:18 +08:00
|
|
|
if (alignment != 0 && ((uintptr_t)ptr & ((uintptr_t)alignment-1))
|
|
|
|
!= 0) {
|
|
|
|
/*
|
2012-04-06 15:35:09 +08:00
|
|
|
* Existing object alignment is inadequate; allocate new space
|
Add {,r,s,d}allocm().
Add allocm(), rallocm(), sallocm(), and dallocm(), which are a
functional superset of malloc(), calloc(), posix_memalign(),
malloc_usable_size(), and free().
2010-09-18 06:46:18 +08:00
|
|
|
* and copy.
|
|
|
|
*/
|
2014-09-23 12:09:23 +08:00
|
|
|
return (iralloct_realign(tsd, ptr, oldsize, size, 0, alignment,
|
2015-01-30 07:30:47 +08:00
|
|
|
zero, tcache, arena));
|
Add {,r,s,d}allocm().
Add allocm(), rallocm(), sallocm(), and dallocm(), which are a
functional superset of malloc(), calloc(), posix_memalign(),
malloc_usable_size(), and free().
2010-09-18 06:46:18 +08:00
|
|
|
}
|
|
|
|
|
2015-09-12 07:18:53 +08:00
|
|
|
return (arena_ralloc(tsd, arena, ptr, oldsize, size, alignment, zero,
|
2015-02-13 06:06:37 +08:00
|
|
|
tcache));
|
2010-02-11 02:37:56 +08:00
|
|
|
}
|
2012-03-22 09:33:03 +08:00
|
|
|
|
2013-01-23 00:45:43 +08:00
|
|
|
JEMALLOC_ALWAYS_INLINE void *
|
2014-10-25 01:18:57 +08:00
|
|
|
iralloc(tsd_t *tsd, void *ptr, size_t oldsize, size_t size, size_t alignment,
|
|
|
|
bool zero)
|
2014-01-13 07:05:44 +08:00
|
|
|
{
|
|
|
|
|
2015-01-30 07:30:47 +08:00
|
|
|
return (iralloct(tsd, ptr, oldsize, size, alignment, zero,
|
|
|
|
tcache_get(tsd, true), NULL));
|
2014-01-13 07:05:44 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
JEMALLOC_ALWAYS_INLINE bool
|
2016-02-20 12:09:31 +08:00
|
|
|
ixalloc(tsd_t *tsd, void *ptr, size_t oldsize, size_t size, size_t extra,
|
|
|
|
size_t alignment, bool zero)
|
2012-10-12 04:53:15 +08:00
|
|
|
{
|
|
|
|
|
2014-01-13 07:05:44 +08:00
|
|
|
assert(ptr != NULL);
|
|
|
|
assert(size != 0);
|
|
|
|
|
|
|
|
if (alignment != 0 && ((uintptr_t)ptr & ((uintptr_t)alignment-1))
|
|
|
|
!= 0) {
|
|
|
|
/* Existing object alignment is inadequate. */
|
|
|
|
return (true);
|
|
|
|
}
|
|
|
|
|
2016-02-20 12:09:31 +08:00
|
|
|
return (arena_ralloc_no_move(tsd, ptr, oldsize, size, extra, zero));
|
2012-10-12 04:53:15 +08:00
|
|
|
}
|
2010-01-17 01:53:50 +08:00
|
|
|
#endif
|
|
|
|
|
2010-10-21 10:05:59 +08:00
|
|
|
#include "jemalloc/internal/prof.h"
|
|
|
|
|
2010-01-17 01:53:50 +08:00
|
|
|
#undef JEMALLOC_H_INLINES
|
|
|
|
/******************************************************************************/
|
2012-04-16 22:30:26 +08:00
|
|
|
#endif /* JEMALLOC_INTERNAL_H */
|