0721b895ff
This avoids a gcc diagnostic note:
note: The ABI for passing parameters with 64-byte alignment has
changed in GCC 4.6
This note related to the cacheline alignment of rtree_ctx_t, which was
introduced by 4a346f5593
(Replace rtree
path cache with LRU cache.).
127 lines
2.9 KiB
C
127 lines
2.9 KiB
C
#ifndef JEMALLOC_INTERNAL_TSD_INLINES_H
|
|
#define JEMALLOC_INTERNAL_TSD_INLINES_H
|
|
|
|
#ifndef JEMALLOC_ENABLE_INLINE
|
|
malloc_tsd_protos(JEMALLOC_ATTR(unused), , tsd_t)
|
|
|
|
tsd_t *tsd_fetch_impl(bool init);
|
|
tsd_t *tsd_fetch(void);
|
|
tsdn_t *tsd_tsdn(tsd_t *tsd);
|
|
bool tsd_nominal(tsd_t *tsd);
|
|
#define O(n, t, gs, c) \
|
|
t *tsd_##n##p_get(tsd_t *tsd); \
|
|
t tsd_##n##_get(tsd_t *tsd); \
|
|
void tsd_##n##_set(tsd_t *tsd, t n);
|
|
MALLOC_TSD
|
|
#undef O
|
|
tsdn_t *tsdn_fetch(void);
|
|
bool tsdn_null(const tsdn_t *tsdn);
|
|
tsd_t *tsdn_tsd(tsdn_t *tsdn);
|
|
rtree_ctx_t *tsdn_rtree_ctx(tsdn_t *tsdn, rtree_ctx_t *fallback);
|
|
#endif
|
|
|
|
#if (defined(JEMALLOC_ENABLE_INLINE) || defined(JEMALLOC_TSD_C_))
|
|
malloc_tsd_externs(, tsd_t)
|
|
malloc_tsd_funcs(JEMALLOC_ALWAYS_INLINE, , tsd_t, tsd_initializer, tsd_cleanup)
|
|
|
|
JEMALLOC_ALWAYS_INLINE tsd_t *
|
|
tsd_fetch_impl(bool init) {
|
|
tsd_t *tsd = tsd_get(init);
|
|
|
|
if (!init && tsd_get_allocates() && tsd == NULL) {
|
|
return NULL;
|
|
}
|
|
assert(tsd != NULL);
|
|
|
|
if (unlikely(tsd->state != tsd_state_nominal)) {
|
|
if (tsd->state == tsd_state_uninitialized) {
|
|
tsd->state = tsd_state_nominal;
|
|
/* Trigger cleanup handler registration. */
|
|
tsd_set(tsd);
|
|
} else if (tsd->state == tsd_state_purgatory) {
|
|
tsd->state = tsd_state_reincarnated;
|
|
tsd_set(tsd);
|
|
} else {
|
|
assert(tsd->state == tsd_state_reincarnated);
|
|
}
|
|
}
|
|
|
|
return tsd;
|
|
}
|
|
|
|
JEMALLOC_ALWAYS_INLINE tsd_t *
|
|
tsd_fetch(void) {
|
|
return tsd_fetch_impl(true);
|
|
}
|
|
|
|
JEMALLOC_ALWAYS_INLINE tsdn_t *
|
|
tsd_tsdn(tsd_t *tsd) {
|
|
return (tsdn_t *)tsd;
|
|
}
|
|
|
|
JEMALLOC_INLINE bool
|
|
tsd_nominal(tsd_t *tsd) {
|
|
return (tsd->state == tsd_state_nominal);
|
|
}
|
|
|
|
#define MALLOC_TSD_getset_yes(n, t) \
|
|
JEMALLOC_ALWAYS_INLINE t \
|
|
tsd_##n##_get(tsd_t *tsd) { \
|
|
return *tsd_##n##p_get(tsd); \
|
|
} \
|
|
JEMALLOC_ALWAYS_INLINE void \
|
|
tsd_##n##_set(tsd_t *tsd, t n) { \
|
|
assert(tsd->state == tsd_state_nominal); \
|
|
tsd->n = n; \
|
|
}
|
|
#define MALLOC_TSD_getset_no(n, t)
|
|
#define O(n, t, gs, c) \
|
|
JEMALLOC_ALWAYS_INLINE t * \
|
|
tsd_##n##p_get(tsd_t *tsd) { \
|
|
return &tsd->n; \
|
|
} \
|
|
\
|
|
MALLOC_TSD_getset_##gs(n, t)
|
|
MALLOC_TSD
|
|
#undef MALLOC_TSD_getset_yes
|
|
#undef MALLOC_TSD_getset_no
|
|
#undef O
|
|
|
|
JEMALLOC_ALWAYS_INLINE tsdn_t *
|
|
tsdn_fetch(void) {
|
|
if (!tsd_booted_get()) {
|
|
return NULL;
|
|
}
|
|
|
|
return tsd_tsdn(tsd_fetch_impl(false));
|
|
}
|
|
|
|
JEMALLOC_ALWAYS_INLINE bool
|
|
tsdn_null(const tsdn_t *tsdn) {
|
|
return tsdn == NULL;
|
|
}
|
|
|
|
JEMALLOC_ALWAYS_INLINE tsd_t *
|
|
tsdn_tsd(tsdn_t *tsdn) {
|
|
assert(!tsdn_null(tsdn));
|
|
|
|
return &tsdn->tsd;
|
|
}
|
|
|
|
JEMALLOC_ALWAYS_INLINE rtree_ctx_t *
|
|
tsdn_rtree_ctx(tsdn_t *tsdn, rtree_ctx_t *fallback) {
|
|
/*
|
|
* If tsd cannot be accessed, initialize the fallback rtree_ctx and
|
|
* return a pointer to it.
|
|
*/
|
|
if (unlikely(tsdn_null(tsdn))) {
|
|
static const rtree_ctx_t rtree_ctx = RTREE_CTX_INITIALIZER;
|
|
memcpy(fallback, &rtree_ctx, sizeof(rtree_ctx_t));
|
|
return fallback;
|
|
}
|
|
return tsd_rtree_ctxp_get(tsdn_tsd(tsdn));
|
|
}
|
|
#endif
|
|
|
|
#endif /* JEMALLOC_INTERNAL_TSD_INLINES_H */
|