blob: 5b28e458f82bb1c3d20fbce9a609e340fac9bdfb [file] [log] [blame]
/* Copyright (C) 2017 The Android Open Source Project
* DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
*
* This file implements interfaces from the file jvmti.h. This implementation
* is licensed under the same terms as the file jvmti.h. The
* copyright and license information for the file jvmti.h follows.
*
* Copyright (c) 2003, 2011, Oracle and/or its affiliates. All rights reserved.
* DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
*
* This code is free software; you can redistribute it and/or modify it
* under the terms of the GNU General Public License version 2 only, as
* published by the Free Software Foundation. Oracle designates this
* particular file as subject to the "Classpath" exception as provided
* by Oracle in the LICENSE file that accompanied this code.
*
* This code is distributed in the hope that it will be useful, but WITHOUT
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
* version 2 for more details (a copy is included in the LICENSE file that
* accompanied this code).
*
* You should have received a copy of the GNU General Public License version
* 2 along with this work; if not, write to the Free Software Foundation,
* Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
*
* Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
* or visit www.oracle.com if you need additional information or have any
* questions.
*/
#ifndef ART_OPENJDKJVMTI_JVMTI_WEAK_TABLE_INL_H_
#define ART_OPENJDKJVMTI_JVMTI_WEAK_TABLE_INL_H_
#include "jvmti_weak_table.h"
#include <limits>
#include <android-base/logging.h>
#include "art_jvmti.h"
#include "gc/allocation_listener.h"
#include "instrumentation.h"
#include "jni/jni_env_ext-inl.h"
#include "jvmti_allocator.h"
#include "mirror/class.h"
#include "mirror/object.h"
#include "nativehelper/scoped_local_ref.h"
#include "runtime.h"
namespace openjdkjvmti {
template <typename T>
void JvmtiWeakTable<T>::Lock() {
allow_disallow_lock_.ExclusiveLock(art::Thread::Current());
}
template <typename T>
void JvmtiWeakTable<T>::Unlock() {
allow_disallow_lock_.ExclusiveUnlock(art::Thread::Current());
}
template <typename T>
void JvmtiWeakTable<T>::AssertLocked() {
allow_disallow_lock_.AssertHeld(art::Thread::Current());
}
template <typename T>
void JvmtiWeakTable<T>::UpdateTableWithReadBarrier() {
update_since_last_sweep_ = true;
auto WithReadBarrierUpdater = [&](const art::GcRoot<art::mirror::Object>& original_root,
art::mirror::Object* original_obj ATTRIBUTE_UNUSED)
REQUIRES_SHARED(art::Locks::mutator_lock_) {
return original_root.Read<art::kWithReadBarrier>();
};
UpdateTableWith<decltype(WithReadBarrierUpdater), kIgnoreNull>(WithReadBarrierUpdater);
}
template <typename T>
bool JvmtiWeakTable<T>::GetTagSlowPath(art::Thread* self, art::ObjPtr<art::mirror::Object> obj, T* result) {
// Under concurrent GC, there is a window between moving objects and sweeping of system
// weaks in which mutators are active. We may receive a to-space object pointer in obj,
// but still have from-space pointers in the table. Explicitly update the table once.
// Note: this will keep *all* objects in the table live, but should be a rare occurrence.
UpdateTableWithReadBarrier();
return GetTagLocked(self, obj, result);
}
template <typename T>
bool JvmtiWeakTable<T>::Remove(art::ObjPtr<art::mirror::Object> obj, /* out */ T* tag) {
art::Thread* self = art::Thread::Current();
art::MutexLock mu(self, allow_disallow_lock_);
Wait(self);
return RemoveLocked(self, obj, tag);
}
template <typename T>
bool JvmtiWeakTable<T>::RemoveLocked(art::ObjPtr<art::mirror::Object> obj, T* tag) {
art::Thread* self = art::Thread::Current();
allow_disallow_lock_.AssertHeld(self);
Wait(self);
return RemoveLocked(self, obj, tag);
}
template <typename T>
bool JvmtiWeakTable<T>::RemoveLocked(art::Thread* self, art::ObjPtr<art::mirror::Object> obj, T* tag) {
auto it = tagged_objects_.find(art::GcRoot<art::mirror::Object>(obj));
if (it != tagged_objects_.end()) {
if (tag != nullptr) {
*tag = it->second;
}
tagged_objects_.erase(it);
return true;
}
if (art::kUseReadBarrier && self->GetIsGcMarking() && !update_since_last_sweep_) {
// Under concurrent GC, there is a window between moving objects and sweeping of system
// weaks in which mutators are active. We may receive a to-space object pointer in obj,
// but still have from-space pointers in the table. Explicitly update the table once.
// Note: this will keep *all* objects in the table live, but should be a rare occurrence.
// Update the table.
UpdateTableWithReadBarrier();
// And try again.
return RemoveLocked(self, obj, tag);
}
// Not in here.
return false;
}
template <typename T>
bool JvmtiWeakTable<T>::Set(art::ObjPtr<art::mirror::Object> obj, T new_tag) {
art::Thread* self = art::Thread::Current();
art::MutexLock mu(self, allow_disallow_lock_);
Wait(self);
return SetLocked(self, obj, new_tag);
}
template <typename T>
bool JvmtiWeakTable<T>::SetLocked(art::ObjPtr<art::mirror::Object> obj, T new_tag) {
art::Thread* self = art::Thread::Current();
allow_disallow_lock_.AssertHeld(self);
Wait(self);
return SetLocked(self, obj, new_tag);
}
template <typename T>
bool JvmtiWeakTable<T>::SetLocked(art::Thread* self, art::ObjPtr<art::mirror::Object> obj, T new_tag) {
auto it = tagged_objects_.find(art::GcRoot<art::mirror::Object>(obj));
if (it != tagged_objects_.end()) {
it->second = new_tag;
return true;
}
if (art::kUseReadBarrier && self->GetIsGcMarking() && !update_since_last_sweep_) {
// Under concurrent GC, there is a window between moving objects and sweeping of system
// weaks in which mutators are active. We may receive a to-space object pointer in obj,
// but still have from-space pointers in the table. Explicitly update the table once.
// Note: this will keep *all* objects in the table live, but should be a rare occurrence.
// Update the table.
UpdateTableWithReadBarrier();
// And try again.
return SetLocked(self, obj, new_tag);
}
// New element.
auto insert_it = tagged_objects_.emplace(art::GcRoot<art::mirror::Object>(obj), new_tag);
DCHECK(insert_it.second);
return false;
}
template <typename T>
void JvmtiWeakTable<T>::Sweep(art::IsMarkedVisitor* visitor) {
if (DoesHandleNullOnSweep()) {
SweepImpl<true>(visitor);
} else {
SweepImpl<false>(visitor);
}
// Under concurrent GC, there is a window between moving objects and sweeping of system
// weaks in which mutators are active. We may receive a to-space object pointer in obj,
// but still have from-space pointers in the table. We explicitly update the table then
// to ensure we compare against to-space pointers. But we want to do this only once. Once
// sweeping is done, we know all objects are to-space pointers until the next GC cycle,
// so we re-enable the explicit update for the next marking.
update_since_last_sweep_ = false;
}
template <typename T>
template <bool kHandleNull>
void JvmtiWeakTable<T>::SweepImpl(art::IsMarkedVisitor* visitor) {
art::Thread* self = art::Thread::Current();
art::MutexLock mu(self, allow_disallow_lock_);
auto IsMarkedUpdater = [&](const art::GcRoot<art::mirror::Object>& original_root ATTRIBUTE_UNUSED,
art::mirror::Object* original_obj) {
return visitor->IsMarked(original_obj);
};
UpdateTableWith<decltype(IsMarkedUpdater),
kHandleNull ? kCallHandleNull : kRemoveNull>(IsMarkedUpdater);
}
template <typename T>
template <typename Updater, typename JvmtiWeakTable<T>::TableUpdateNullTarget kTargetNull>
ALWAYS_INLINE inline void JvmtiWeakTable<T>::UpdateTableWith(Updater& updater) {
// We optimistically hope that elements will still be well-distributed when re-inserting them.
// So play with the map mechanics, and postpone rehashing. This avoids the need of a side
// vector and two passes.
float original_max_load_factor = tagged_objects_.max_load_factor();
tagged_objects_.max_load_factor(std::numeric_limits<float>::max());
// For checking that a max load-factor actually does what we expect.
size_t original_bucket_count = tagged_objects_.bucket_count();
for (auto it = tagged_objects_.begin(); it != tagged_objects_.end();) {
DCHECK(!it->first.IsNull());
art::mirror::Object* original_obj = it->first.template Read<art::kWithoutReadBarrier>();
art::mirror::Object* target_obj = updater(it->first, original_obj);
if (original_obj != target_obj) {
if (kTargetNull == kIgnoreNull && target_obj == nullptr) {
// Ignore null target, don't do anything.
} else {
T tag = it->second;
it = tagged_objects_.erase(it);
if (target_obj != nullptr) {
tagged_objects_.emplace(art::GcRoot<art::mirror::Object>(target_obj), tag);
DCHECK_EQ(original_bucket_count, tagged_objects_.bucket_count());
} else if (kTargetNull == kCallHandleNull) {
HandleNullSweep(tag);
}
continue; // Iterator was implicitly updated by erase.
}
}
it++;
}
tagged_objects_.max_load_factor(original_max_load_factor);
// TODO: consider rehash here.
}
template <typename T>
template <typename Storage, class Allocator>
struct JvmtiWeakTable<T>::ReleasableContainer {
using allocator_type = Allocator;
explicit ReleasableContainer(const allocator_type& alloc, size_t reserve = 10)
: allocator(alloc),
data(reserve > 0 ? allocator.allocate(reserve) : nullptr),
size(0),
capacity(reserve) {
}
~ReleasableContainer() {
if (data != nullptr) {
allocator.deallocate(data, capacity);
capacity = 0;
size = 0;
}
}
Storage* Release() {
Storage* tmp = data;
data = nullptr;
size = 0;
capacity = 0;
return tmp;
}
void Resize(size_t new_capacity) {
CHECK_GT(new_capacity, capacity);
Storage* tmp = allocator.allocate(new_capacity);
DCHECK(tmp != nullptr);
if (data != nullptr) {
memcpy(tmp, data, sizeof(Storage) * size);
}
Storage* old = data;
data = tmp;
allocator.deallocate(old, capacity);
capacity = new_capacity;
}
void Pushback(const Storage& elem) {
if (size == capacity) {
size_t new_capacity = 2 * capacity + 1;
Resize(new_capacity);
}
data[size++] = elem;
}
Allocator allocator;
Storage* data;
size_t size;
size_t capacity;
};
template <typename T>
jvmtiError JvmtiWeakTable<T>::GetTaggedObjects(jvmtiEnv* jvmti_env,
jint tag_count,
const T* tags,
jint* count_ptr,
jobject** object_result_ptr,
T** tag_result_ptr) {
if (tag_count < 0) {
return ERR(ILLEGAL_ARGUMENT);
}
if (tag_count > 0) {
for (size_t i = 0; i != static_cast<size_t>(tag_count); ++i) {
if (tags[i] == 0) {
return ERR(ILLEGAL_ARGUMENT);
}
}
}
if (tags == nullptr) {
return ERR(NULL_POINTER);
}
if (count_ptr == nullptr) {
return ERR(NULL_POINTER);
}
art::Thread* self = art::Thread::Current();
art::MutexLock mu(self, allow_disallow_lock_);
Wait(self);
art::JNIEnvExt* jni_env = self->GetJniEnv();
constexpr size_t kDefaultSize = 10;
size_t initial_object_size;
size_t initial_tag_size;
if (tag_count == 0) {
initial_object_size = (object_result_ptr != nullptr) ? tagged_objects_.size() : 0;
initial_tag_size = (tag_result_ptr != nullptr) ? tagged_objects_.size() : 0;
} else {
initial_object_size = initial_tag_size = kDefaultSize;
}
JvmtiAllocator<void> allocator(jvmti_env);
ReleasableContainer<jobject, JvmtiAllocator<jobject>> selected_objects(allocator,
initial_object_size);
ReleasableContainer<T, JvmtiAllocator<T>> selected_tags(allocator, initial_tag_size);
size_t count = 0;
for (auto& pair : tagged_objects_) {
bool select;
if (tag_count > 0) {
select = false;
for (size_t i = 0; i != static_cast<size_t>(tag_count); ++i) {
if (tags[i] == pair.second) {
select = true;
break;
}
}
} else {
select = true;
}
if (select) {
art::ObjPtr<art::mirror::Object> obj = pair.first.template Read<art::kWithReadBarrier>();
if (obj != nullptr) {
count++;
if (object_result_ptr != nullptr) {
selected_objects.Pushback(jni_env->AddLocalReference<jobject>(obj));
}
if (tag_result_ptr != nullptr) {
selected_tags.Pushback(pair.second);
}
}
}
}
if (object_result_ptr != nullptr) {
*object_result_ptr = selected_objects.Release();
}
if (tag_result_ptr != nullptr) {
*tag_result_ptr = selected_tags.Release();
}
*count_ptr = static_cast<jint>(count);
return ERR(NONE);
}
template <typename T>
art::ObjPtr<art::mirror::Object> JvmtiWeakTable<T>::Find(T tag) {
art::Thread* self = art::Thread::Current();
art::MutexLock mu(self, allow_disallow_lock_);
Wait(self);
for (auto& pair : tagged_objects_) {
if (tag == pair.second) {
art::ObjPtr<art::mirror::Object> obj = pair.first.template Read<art::kWithReadBarrier>();
if (obj != nullptr) {
return obj;
}
}
}
return nullptr;
}
} // namespace openjdkjvmti
#endif // ART_OPENJDKJVMTI_JVMTI_WEAK_TABLE_INL_H_