You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
149 lines
4.6 KiB
149 lines
4.6 KiB
//===-- tsd_exclusive.h -----------------------------------------*- C++ -*-===//
|
|
//
|
|
// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
|
|
// See https://llvm.org/LICENSE.txt for license information.
|
|
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
|
|
//
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
#ifndef SCUDO_TSD_EXCLUSIVE_H_
|
|
#define SCUDO_TSD_EXCLUSIVE_H_
|
|
|
|
#include "tsd.h"
|
|
|
|
namespace scudo {
|
|
|
|
struct ThreadState {
|
|
bool DisableMemInit : 1;
|
|
enum {
|
|
NotInitialized = 0,
|
|
Initialized,
|
|
TornDown,
|
|
} InitState : 2;
|
|
};
|
|
|
|
template <class Allocator> void teardownThread(void *Ptr);
|
|
|
|
template <class Allocator> struct TSDRegistryExT {
|
|
void initLinkerInitialized(Allocator *Instance) {
|
|
Instance->initLinkerInitialized();
|
|
CHECK_EQ(pthread_key_create(&PThreadKey, teardownThread<Allocator>), 0);
|
|
FallbackTSD.initLinkerInitialized(Instance);
|
|
Initialized = true;
|
|
}
|
|
void init(Allocator *Instance) {
|
|
memset(this, 0, sizeof(*this));
|
|
initLinkerInitialized(Instance);
|
|
}
|
|
|
|
void initOnceMaybe(Allocator *Instance) {
|
|
ScopedLock L(Mutex);
|
|
if (LIKELY(Initialized))
|
|
return;
|
|
initLinkerInitialized(Instance); // Sets Initialized.
|
|
}
|
|
|
|
void unmapTestOnly() {
|
|
Allocator *Instance =
|
|
reinterpret_cast<Allocator *>(pthread_getspecific(PThreadKey));
|
|
if (!Instance)
|
|
return;
|
|
ThreadTSD.commitBack(Instance);
|
|
State = {};
|
|
}
|
|
|
|
ALWAYS_INLINE void initThreadMaybe(Allocator *Instance, bool MinimalInit) {
|
|
if (LIKELY(State.InitState != ThreadState::NotInitialized))
|
|
return;
|
|
initThread(Instance, MinimalInit);
|
|
}
|
|
|
|
ALWAYS_INLINE TSD<Allocator> *getTSDAndLock(bool *UnlockRequired) {
|
|
if (LIKELY(State.InitState == ThreadState::Initialized &&
|
|
!atomic_load(&Disabled, memory_order_acquire))) {
|
|
*UnlockRequired = false;
|
|
return &ThreadTSD;
|
|
}
|
|
FallbackTSD.lock();
|
|
*UnlockRequired = true;
|
|
return &FallbackTSD;
|
|
}
|
|
|
|
// To disable the exclusive TSD registry, we effectively lock the fallback TSD
|
|
// and force all threads to attempt to use it instead of their local one.
|
|
void disable() {
|
|
Mutex.lock();
|
|
FallbackTSD.lock();
|
|
atomic_store(&Disabled, 1U, memory_order_release);
|
|
}
|
|
|
|
void enable() {
|
|
atomic_store(&Disabled, 0U, memory_order_release);
|
|
FallbackTSD.unlock();
|
|
Mutex.unlock();
|
|
}
|
|
|
|
bool setOption(Option O, UNUSED sptr Value) {
|
|
if (O == Option::ThreadDisableMemInit)
|
|
State.DisableMemInit = Value;
|
|
if (O == Option::MaxTSDsCount)
|
|
return false;
|
|
return true;
|
|
}
|
|
|
|
bool getDisableMemInit() { return State.DisableMemInit; }
|
|
|
|
private:
|
|
// Using minimal initialization allows for global initialization while keeping
|
|
// the thread specific structure untouched. The fallback structure will be
|
|
// used instead.
|
|
NOINLINE void initThread(Allocator *Instance, bool MinimalInit) {
|
|
initOnceMaybe(Instance);
|
|
if (UNLIKELY(MinimalInit))
|
|
return;
|
|
CHECK_EQ(
|
|
pthread_setspecific(PThreadKey, reinterpret_cast<void *>(Instance)), 0);
|
|
ThreadTSD.initLinkerInitialized(Instance);
|
|
State.InitState = ThreadState::Initialized;
|
|
Instance->callPostInitCallback();
|
|
}
|
|
|
|
pthread_key_t PThreadKey = {};
|
|
bool Initialized = false;
|
|
atomic_u8 Disabled = {};
|
|
TSD<Allocator> FallbackTSD;
|
|
HybridMutex Mutex;
|
|
static thread_local ThreadState State;
|
|
static thread_local TSD<Allocator> ThreadTSD;
|
|
|
|
friend void teardownThread<Allocator>(void *Ptr);
|
|
};
|
|
|
|
template <class Allocator>
|
|
thread_local TSD<Allocator> TSDRegistryExT<Allocator>::ThreadTSD;
|
|
template <class Allocator>
|
|
thread_local ThreadState TSDRegistryExT<Allocator>::State;
|
|
|
|
template <class Allocator> void teardownThread(void *Ptr) {
|
|
typedef TSDRegistryExT<Allocator> TSDRegistryT;
|
|
Allocator *Instance = reinterpret_cast<Allocator *>(Ptr);
|
|
// The glibc POSIX thread-local-storage deallocation routine calls user
|
|
// provided destructors in a loop of PTHREAD_DESTRUCTOR_ITERATIONS.
|
|
// We want to be called last since other destructors might call free and the
|
|
// like, so we wait until PTHREAD_DESTRUCTOR_ITERATIONS before draining the
|
|
// quarantine and swallowing the cache.
|
|
if (TSDRegistryT::ThreadTSD.DestructorIterations > 1) {
|
|
TSDRegistryT::ThreadTSD.DestructorIterations--;
|
|
// If pthread_setspecific fails, we will go ahead with the teardown.
|
|
if (LIKELY(pthread_setspecific(Instance->getTSDRegistry()->PThreadKey,
|
|
Ptr) == 0))
|
|
return;
|
|
}
|
|
TSDRegistryT::ThreadTSD.commitBack(Instance);
|
|
TSDRegistryT::State.InitState = ThreadState::TornDown;
|
|
}
|
|
|
|
} // namespace scudo
|
|
|
|
#endif // SCUDO_TSD_EXCLUSIVE_H_
|