You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

489 lines
21 KiB

/*
* Copyright (C) 2008 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#ifndef ART_RUNTIME_MONITOR_H_
#define ART_RUNTIME_MONITOR_H_
#include <pthread.h>
#include <stdint.h>
#include <stdlib.h>
#include <atomic>
#include <iosfwd>
#include <list>
#include <vector>
#include "base/allocator.h"
#include "base/atomic.h"
#include "base/mutex.h"
#include "gc_root.h"
#include "lock_word.h"
#include "obj_ptr.h"
#include "read_barrier_option.h"
#include "runtime_callbacks.h"
#include "thread_state.h"
namespace art {
class ArtMethod;
class IsMarkedVisitor;
class LockWord;
template<class T> class Handle;
class StackVisitor;
class Thread;
typedef uint32_t MonitorId;
namespace mirror {
class Object;
} // namespace mirror
enum class LockReason {
kForWait,
kForLock,
};
class Monitor {
public:
// The default number of spins that are done before thread suspension is used to forcibly inflate
// a lock word. See Runtime::max_spins_before_thin_lock_inflation_.
constexpr static size_t kDefaultMaxSpinsBeforeThinLockInflation = 50;
static constexpr int kDefaultMonitorTimeoutMs = 500;
static constexpr int kMonitorTimeoutMinMs = 200;
static constexpr int kMonitorTimeoutMaxMs = 1000; // 1 second
~Monitor();
static void Init(uint32_t lock_profiling_threshold, uint32_t stack_dump_lock_profiling_threshold);
// Return the thread id of the lock owner or 0 when there is no owner.
static uint32_t GetLockOwnerThreadId(ObjPtr<mirror::Object> obj)
NO_THREAD_SAFETY_ANALYSIS; // TODO: Reading lock owner without holding lock is racy.
// NO_THREAD_SAFETY_ANALYSIS for mon->Lock.
static ObjPtr<mirror::Object> MonitorEnter(Thread* thread,
ObjPtr<mirror::Object> obj,
bool trylock)
EXCLUSIVE_LOCK_FUNCTION(obj.Ptr())
NO_THREAD_SAFETY_ANALYSIS
REQUIRES(!Roles::uninterruptible_)
REQUIRES_SHARED(Locks::mutator_lock_);
// NO_THREAD_SAFETY_ANALYSIS for mon->Unlock.
static bool MonitorExit(Thread* thread, ObjPtr<mirror::Object> obj)
NO_THREAD_SAFETY_ANALYSIS
REQUIRES(!Roles::uninterruptible_)
REQUIRES_SHARED(Locks::mutator_lock_)
UNLOCK_FUNCTION(obj.Ptr());
static void Notify(Thread* self, ObjPtr<mirror::Object> obj)
REQUIRES_SHARED(Locks::mutator_lock_) {
DoNotify(self, obj, false);
}
static void NotifyAll(Thread* self, ObjPtr<mirror::Object> obj)
REQUIRES_SHARED(Locks::mutator_lock_) {
DoNotify(self, obj, true);
}
// Object.wait(). Also called for class init.
// NO_THREAD_SAFETY_ANALYSIS for mon->Wait.
static void Wait(Thread* self,
ObjPtr<mirror::Object> obj,
int64_t ms,
int32_t ns,
bool interruptShouldThrow, ThreadState why)
REQUIRES_SHARED(Locks::mutator_lock_) NO_THREAD_SAFETY_ANALYSIS;
static ThreadState FetchState(const Thread* thread,
/* out */ ObjPtr<mirror::Object>* monitor_object,
/* out */ uint32_t* lock_owner_tid)
REQUIRES(!Locks::thread_suspend_count_lock_)
REQUIRES_SHARED(Locks::mutator_lock_);
// Used to implement JDWP's ThreadReference.CurrentContendedMonitor.
static ObjPtr<mirror::Object> GetContendedMonitor(Thread* thread)
REQUIRES_SHARED(Locks::mutator_lock_);
// Calls 'callback' once for each lock held in the single stack frame represented by
// the current state of 'stack_visitor'.
// The abort_on_failure flag allows to not die when the state of the runtime is unorderly. This
// is necessary when we have already aborted but want to dump the stack as much as we can.
static void VisitLocks(StackVisitor* stack_visitor,
void (*callback)(ObjPtr<mirror::Object>, void*),
void* callback_context,
bool abort_on_failure = true)
REQUIRES_SHARED(Locks::mutator_lock_);
static bool IsValidLockWord(LockWord lock_word);
template<ReadBarrierOption kReadBarrierOption = kWithReadBarrier>
ObjPtr<mirror::Object> GetObject() REQUIRES_SHARED(Locks::mutator_lock_);
void SetObject(ObjPtr<mirror::Object> object);
// Provides no memory ordering guarantees.
Thread* GetOwner() const {
return owner_.load(std::memory_order_relaxed);
}
int32_t GetHashCode();
// Is the monitor currently locked? Debug only, provides no memory ordering guarantees.
bool IsLocked() REQUIRES_SHARED(Locks::mutator_lock_) REQUIRES(!monitor_lock_);
bool HasHashCode() const {
return hash_code_.load(std::memory_order_relaxed) != 0;
}
MonitorId GetMonitorId() const {
return monitor_id_;
}
// Inflate the lock on obj. May fail to inflate for spurious reasons, always re-check.
static void InflateThinLocked(Thread* self, Handle<mirror::Object> obj, LockWord lock_word,
uint32_t hash_code) REQUIRES_SHARED(Locks::mutator_lock_);
// Not exclusive because ImageWriter calls this during a Heap::VisitObjects() that
// does not allow a thread suspension in the middle. TODO: maybe make this exclusive.
// NO_THREAD_SAFETY_ANALYSIS for monitor->monitor_lock_.
static bool Deflate(Thread* self, ObjPtr<mirror::Object> obj)
REQUIRES_SHARED(Locks::mutator_lock_) NO_THREAD_SAFETY_ANALYSIS;
#ifndef __LP64__
void* operator new(size_t size) {
// Align Monitor* as per the monitor ID field size in the lock word.
void* result;
int error = posix_memalign(&result, LockWord::kMonitorIdAlignment, size);
CHECK_EQ(error, 0) << strerror(error);
return result;
}
void operator delete(void* ptr) {
free(ptr);
}
#endif
private:
Monitor(Thread* self, Thread* owner, ObjPtr<mirror::Object> obj, int32_t hash_code)
REQUIRES_SHARED(Locks::mutator_lock_);
Monitor(Thread* self, Thread* owner, ObjPtr<mirror::Object> obj, int32_t hash_code, MonitorId id)
REQUIRES_SHARED(Locks::mutator_lock_);
// Install the monitor into its object, may fail if another thread installs a different monitor
// first. Monitor remains in the same logical state as before, i.e. held the same # of times.
bool Install(Thread* self)
REQUIRES(!monitor_lock_)
REQUIRES_SHARED(Locks::mutator_lock_);
// Links a thread into a monitor's wait set. The monitor lock must be held by the caller of this
// routine.
void AppendToWaitSet(Thread* thread) REQUIRES(monitor_lock_);
// Unlinks a thread from a monitor's wait set. The monitor lock must be held by the caller of
// this routine.
void RemoveFromWaitSet(Thread* thread) REQUIRES(monitor_lock_);
// Release the monitor lock and signal a waiting thread that has been notified and now needs the
// lock. Assumes the monitor lock is held exactly once, and the owner_ field has been reset to
// null. Caller may be suspended (Wait) or runnable (MonitorExit).
void SignalWaiterAndReleaseMonitorLock(Thread* self) RELEASE(monitor_lock_);
// Changes the shape of a monitor from thin to fat, preserving the internal lock state. The
// calling thread must own the lock or the owner must be suspended. There's a race with other
// threads inflating the lock, installing hash codes and spurious failures. The caller should
// re-read the lock word following the call.
static void Inflate(Thread* self, Thread* owner, ObjPtr<mirror::Object> obj, int32_t hash_code)
REQUIRES_SHARED(Locks::mutator_lock_)
NO_THREAD_SAFETY_ANALYSIS; // For m->Install(self)
void LogContentionEvent(Thread* self,
uint32_t wait_ms,
uint32_t sample_percent,
ArtMethod* owner_method,
uint32_t owner_dex_pc)
REQUIRES_SHARED(Locks::mutator_lock_);
static void FailedUnlock(ObjPtr<mirror::Object> obj,
uint32_t expected_owner_thread_id,
uint32_t found_owner_thread_id,
Monitor* mon)
REQUIRES(!Locks::thread_list_lock_)
REQUIRES_SHARED(Locks::mutator_lock_);
// Try to lock without blocking, returns true if we acquired the lock.
// If spin is true, then we spin for a short period before failing.
bool TryLock(Thread* self, bool spin = false)
TRY_ACQUIRE(true, monitor_lock_)
REQUIRES_SHARED(Locks::mutator_lock_);
template<LockReason reason = LockReason::kForLock>
void Lock(Thread* self)
ACQUIRE(monitor_lock_)
REQUIRES_SHARED(Locks::mutator_lock_);
bool Unlock(Thread* thread)
RELEASE(monitor_lock_)
REQUIRES_SHARED(Locks::mutator_lock_);
static void DoNotify(Thread* self, ObjPtr<mirror::Object> obj, bool notify_all)
REQUIRES_SHARED(Locks::mutator_lock_) NO_THREAD_SAFETY_ANALYSIS; // For mon->Notify.
void Notify(Thread* self)
REQUIRES(monitor_lock_)
REQUIRES_SHARED(Locks::mutator_lock_);
void NotifyAll(Thread* self)
REQUIRES(monitor_lock_)
REQUIRES_SHARED(Locks::mutator_lock_);
static std::string PrettyContentionInfo(const std::string& owner_name,
pid_t owner_tid,
ArtMethod* owners_method,
uint32_t owners_dex_pc,
size_t num_waiters)
REQUIRES_SHARED(Locks::mutator_lock_);
// Wait on a monitor until timeout, interrupt, or notification. Used for Object.wait() and
// (somewhat indirectly) Thread.sleep() and Thread.join().
//
// If another thread calls Thread.interrupt(), we throw InterruptedException and return
// immediately if one of the following are true:
// - blocked in wait(), wait(long), or wait(long, int) methods of Object
// - blocked in join(), join(long), or join(long, int) methods of Thread
// - blocked in sleep(long), or sleep(long, int) methods of Thread
// Otherwise, we set the "interrupted" flag.
//
// Checks to make sure that "ns" is in the range 0-999999 (i.e. fractions of a millisecond) and
// throws the appropriate exception if it isn't.
//
// The spec allows "spurious wakeups", and recommends that all code using Object.wait() do so in
// a loop. This appears to derive from concerns about pthread_cond_wait() on multiprocessor
// systems. Some commentary on the web casts doubt on whether these can/should occur.
//
// Since we're allowed to wake up "early", we clamp extremely long durations to return at the end
// of the 32-bit time epoch.
void Wait(Thread* self, int64_t msec, int32_t nsec, bool interruptShouldThrow, ThreadState why)
REQUIRES(monitor_lock_)
REQUIRES_SHARED(Locks::mutator_lock_);
// Translates the provided method and pc into its declaring class' source file and line number.
static void TranslateLocation(ArtMethod* method, uint32_t pc,
const char** source_file,
int32_t* line_number)
REQUIRES_SHARED(Locks::mutator_lock_);
// Provides no memory ordering guarantees.
uint32_t GetOwnerThreadId() REQUIRES(!monitor_lock_);
// Set locking_method_ and locking_dex_pc_ corresponding to owner's current stack.
// owner is either self or suspended.
void SetLockingMethod(Thread* owner) REQUIRES(monitor_lock_)
REQUIRES_SHARED(Locks::mutator_lock_);
// The same, but without checking for a proxy method. Currently requires owner == self.
void SetLockingMethodNoProxy(Thread* owner) REQUIRES(monitor_lock_)
REQUIRES_SHARED(Locks::mutator_lock_);
// Support for systrace output of monitor operations.
ALWAYS_INLINE static void AtraceMonitorLock(Thread* self,
ObjPtr<mirror::Object> obj,
bool is_wait)
REQUIRES_SHARED(Locks::mutator_lock_);
static void AtraceMonitorLockImpl(Thread* self,
ObjPtr<mirror::Object> obj,
bool is_wait)
REQUIRES_SHARED(Locks::mutator_lock_);
ALWAYS_INLINE static void AtraceMonitorUnlock();
static uint32_t lock_profiling_threshold_;
static uint32_t stack_dump_lock_profiling_threshold_;
static bool capture_method_eagerly_;
// Holding the monitor N times is represented by holding monitor_lock_ N times.
Mutex monitor_lock_ DEFAULT_MUTEX_ACQUIRED_AFTER;
// Pretend to unlock monitor lock.
void FakeUnlockMonitorLock() RELEASE(monitor_lock_) NO_THREAD_SAFETY_ANALYSIS {}
// Number of threads either waiting on the condition or waiting on a contended
// monitor acquisition. Prevents deflation.
std::atomic<size_t> num_waiters_;
// Which thread currently owns the lock? monitor_lock_ only keeps the tid.
// Only set while holding monitor_lock_. Non-locking readers only use it to
// compare to self or for debugging.
std::atomic<Thread*> owner_;
// Owner's recursive lock depth. Owner_ non-null, and lock_count_ == 0 ==> held once.
unsigned int lock_count_ GUARDED_BY(monitor_lock_);
// Owner's recursive lock depth is given by monitor_lock_.GetDepth().
// What object are we part of. This is a weak root. Do not access
// this directly, use GetObject() to read it so it will be guarded
// by a read barrier.
GcRoot<mirror::Object> obj_;
// Threads currently waiting on this monitor.
Thread* wait_set_ GUARDED_BY(monitor_lock_);
// Threads that were waiting on this monitor, but are now contending on it.
Thread* wake_set_ GUARDED_BY(monitor_lock_);
// Stored object hash code, generated lazily by GetHashCode.
AtomicInteger hash_code_;
// Data structure used to remember the method and dex pc of a recent holder of the
// lock. Used for tracing and contention reporting. Setting these is expensive, since it
// involves a partial stack walk. We set them only as follows, to minimize the cost:
// - If tracing is enabled, they are needed immediately when we first notice contention, so we
// set them unconditionally when a monitor is acquired.
// - If contention reporting is enabled, we use the lock_owner_request_ field to have the
// contending thread request them. The current owner then sets them when releasing the monitor,
// making them available when the contending thread acquires the monitor.
// - If tracing and contention reporting are enabled, we do both. This usually prevents us from
// switching between reporting the end and beginning of critical sections for contention logging
// when tracing is enabled. We expect that tracing overhead is normally much higher than for
// contention logging, so the added cost should be small. It also minimizes glitches when
// enabling and disabling traces.
// We're tolerant of missing information. E.g. when tracing is initially turned on, we may
// not have the lock holder information if the holder acquired the lock with tracing off.
//
// We make this data unconditionally atomic; for contention logging all accesses are in fact
// protected by the monitor, but for tracing, reads are not. Writes are always
// protected by the monitor.
//
// The fields are always accessed without memory ordering. We store a checksum, and reread if
// the checksum doesn't correspond to the values. This results in values that are correct with
// very high probability, but not certainty.
//
// If we need lock_owner information for a certain thread for contenion logging, we store its
// tid in lock_owner_request_. To satisfy the request, we store lock_owner_tid_,
// lock_owner_method_, and lock_owner_dex_pc_ and the corresponding checksum while holding the
// monitor.
//
// At all times, either lock_owner_ is zero, the checksum is valid, or a thread is actively
// in the process of establishing one of those states. Only one thread at a time can be actively
// establishing such a state, since writes are protected by the monitor.
std::atomic<Thread*> lock_owner_; // *lock_owner_ may no longer exist!
std::atomic<ArtMethod*> lock_owner_method_;
std::atomic<uint32_t> lock_owner_dex_pc_;
std::atomic<uintptr_t> lock_owner_sum_;
// Request lock owner save method and dex_pc. Written asynchronously.
std::atomic<Thread*> lock_owner_request_;
// Compute method, dex pc, and tid "checksum".
uintptr_t LockOwnerInfoChecksum(ArtMethod* m, uint32_t dex_pc, Thread* t);
// Set owning method, dex pc, and tid. owner_ field is set and points to current thread.
void SetLockOwnerInfo(ArtMethod* method, uint32_t dex_pc, Thread* t)
REQUIRES(monitor_lock_);
// Get owning method and dex pc for the given thread, if available.
void GetLockOwnerInfo(/*out*/ArtMethod** method, /*out*/uint32_t* dex_pc, Thread* t);
// Do the same, while holding the monitor. There are no concurrent updates.
void GetLockOwnerInfoLocked(/*out*/ArtMethod** method, /*out*/uint32_t* dex_pc,
uint32_t thread_id)
REQUIRES(monitor_lock_);
// We never clear lock_owner method and dex pc. Since it often reflects
// ownership when we last detected contention, it may be inconsistent with owner_
// and not 100% reliable. For lock contention monitoring, in the absence of tracing,
// there is a small risk that the current owner may finish before noticing the request,
// or the information will be overwritten by another intervening request and monitor
// release, so it's also not 100% reliable. But if we report information at all, it
// should generally (modulo accidental checksum matches) pertain to to an acquisition of the
// right monitor by the right thread, so it's extremely unlikely to be seriously misleading.
// Since we track threads by a pointer to the Thread structure, there is a small chance we may
// confuse threads allocated at the same exact address, if a contending thread dies before
// we inquire about it.
// Check for and act on a pending lock_owner_request_
void CheckLockOwnerRequest(Thread* self)
REQUIRES(monitor_lock_) REQUIRES_SHARED(Locks::mutator_lock_);
void MaybeEnableTimeout() REQUIRES(Locks::mutator_lock_);
// The denser encoded version of this monitor as stored in the lock word.
MonitorId monitor_id_;
#ifdef __LP64__
// Free list for monitor pool.
Monitor* next_free_ GUARDED_BY(Locks::allocated_monitor_ids_lock_);
#endif
friend class MonitorInfo;
friend class MonitorList;
friend class MonitorPool;
friend class mirror::Object;
DISALLOW_COPY_AND_ASSIGN(Monitor);
};
class MonitorList {
public:
MonitorList();
~MonitorList();
void Add(Monitor* m) REQUIRES_SHARED(Locks::mutator_lock_) REQUIRES(!monitor_list_lock_);
void SweepMonitorList(IsMarkedVisitor* visitor)
REQUIRES(!monitor_list_lock_) REQUIRES_SHARED(Locks::mutator_lock_);
void DisallowNewMonitors() REQUIRES(!monitor_list_lock_);
void AllowNewMonitors() REQUIRES(!monitor_list_lock_);
void BroadcastForNewMonitors() REQUIRES(!monitor_list_lock_);
// Returns how many monitors were deflated.
size_t DeflateMonitors() REQUIRES(!monitor_list_lock_) REQUIRES(Locks::mutator_lock_);
size_t Size() REQUIRES(!monitor_list_lock_);
typedef std::list<Monitor*, TrackingAllocator<Monitor*, kAllocatorTagMonitorList>> Monitors;
private:
// During sweeping we may free an object and on a separate thread have an object created using
// the newly freed memory. That object may then have its lock-word inflated and a monitor created.
// If we allow new monitor registration during sweeping this monitor may be incorrectly freed as
// the object wasn't marked when sweeping began.
bool allow_new_monitors_ GUARDED_BY(monitor_list_lock_);
Mutex monitor_list_lock_ DEFAULT_MUTEX_ACQUIRED_AFTER;
ConditionVariable monitor_add_condition_ GUARDED_BY(monitor_list_lock_);
Monitors list_ GUARDED_BY(monitor_list_lock_);
friend class Monitor;
DISALLOW_COPY_AND_ASSIGN(MonitorList);
};
// Collects information about the current state of an object's monitor.
// This is very unsafe, and must only be called when all threads are suspended.
// For use only by the JDWP implementation.
class MonitorInfo {
public:
MonitorInfo() : owner_(nullptr), entry_count_(0) {}
MonitorInfo(const MonitorInfo&) = default;
MonitorInfo& operator=(const MonitorInfo&) = default;
explicit MonitorInfo(ObjPtr<mirror::Object> o) REQUIRES(Locks::mutator_lock_);
Thread* owner_;
size_t entry_count_;
std::vector<Thread*> waiters_;
};
} // namespace art
#endif // ART_RUNTIME_MONITOR_H_