| /* |
| * Copyright (c) 2015, 2021, Red Hat, Inc. All rights reserved. |
| * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. |
| * |
| * This code is free software; you can redistribute it and/or modify it |
| * under the terms of the GNU General Public License version 2 only, as |
| * published by the Free Software Foundation. |
| * |
| * This code is distributed in the hope that it will be useful, but WITHOUT |
| * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
| * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License |
| * version 2 for more details (a copy is included in the LICENSE file that |
| * accompanied this code). |
| * |
| * You should have received a copy of the GNU General Public License version |
| * 2 along with this work; if not, write to the Free Software Foundation, |
| * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. |
| * |
| * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA |
| * or visit www.oracle.com if you need additional information or have any |
| * questions. |
| * |
| */ |
| |
| #ifndef SHARE_GC_SHENANDOAH_SHENANDOAHBARRIERSET_INLINE_HPP |
| #define SHARE_GC_SHENANDOAH_SHENANDOAHBARRIERSET_INLINE_HPP |
| |
| #include "gc/shenandoah/shenandoahBarrierSet.hpp" |
| |
| #include "gc/shared/accessBarrierSupport.inline.hpp" |
| #include "gc/shenandoah/shenandoahAsserts.hpp" |
| #include "gc/shenandoah/shenandoahCollectionSet.inline.hpp" |
| #include "gc/shenandoah/shenandoahEvacOOMHandler.inline.hpp" |
| #include "gc/shenandoah/shenandoahForwarding.inline.hpp" |
| #include "gc/shenandoah/shenandoahHeap.inline.hpp" |
| #include "gc/shenandoah/shenandoahHeapRegion.hpp" |
| #include "gc/shenandoah/shenandoahMarkingContext.inline.hpp" |
| #include "gc/shenandoah/shenandoahThreadLocalData.hpp" |
| #include "oops/oop.inline.hpp" |
| |
| inline oop ShenandoahBarrierSet::resolve_forwarded_not_null(oop p) { |
| return ShenandoahForwarding::get_forwardee(p); |
| } |
| |
| inline oop ShenandoahBarrierSet::resolve_forwarded(oop p) { |
| if (p != NULL) { |
| return resolve_forwarded_not_null(p); |
| } else { |
| return p; |
| } |
| } |
| |
| inline oop ShenandoahBarrierSet::resolve_forwarded_not_null_mutator(oop p) { |
| return ShenandoahForwarding::get_forwardee_mutator(p); |
| } |
| |
| template <class T> |
| inline oop ShenandoahBarrierSet::load_reference_barrier_mutator(oop obj, T* load_addr) { |
| assert(ShenandoahLoadRefBarrier, "should be enabled"); |
| shenandoah_assert_in_cset(load_addr, obj); |
| |
| oop fwd = resolve_forwarded_not_null_mutator(obj); |
| if (obj == fwd) { |
| assert(_heap->is_evacuation_in_progress(), |
| "evac should be in progress"); |
| Thread* const t = Thread::current(); |
| ShenandoahEvacOOMScope scope(t); |
| fwd = _heap->evacuate_object(obj, t); |
| } |
| |
| if (load_addr != NULL && fwd != obj) { |
| // Since we are here and we know the load address, update the reference. |
| ShenandoahHeap::cas_oop(fwd, load_addr, obj); |
| } |
| |
| return fwd; |
| } |
| |
| inline oop ShenandoahBarrierSet::load_reference_barrier(oop obj) { |
| if (!ShenandoahLoadRefBarrier) { |
| return obj; |
| } |
| if (_heap->has_forwarded_objects() && |
| _heap->in_collection_set(obj)) { // Subsumes NULL-check |
| assert(obj != NULL, "cset check must have subsumed NULL-check"); |
| oop fwd = resolve_forwarded_not_null(obj); |
| // TODO: It should not be necessary to check evac-in-progress here. |
| // We do it for mark-compact, which may have forwarded objects, |
| // and objects in cset and gets here via runtime barriers. |
| // We can probably fix this as soon as mark-compact has its own |
| // marking phase. |
| if (obj == fwd && _heap->is_evacuation_in_progress()) { |
| Thread* t = Thread::current(); |
| ShenandoahEvacOOMScope oom_evac_scope(t); |
| return _heap->evacuate_object(obj, t); |
| } |
| return fwd; |
| } |
| return obj; |
| } |
| |
| template <class T> |
| inline oop ShenandoahBarrierSet::load_reference_barrier(DecoratorSet decorators, oop obj, T* load_addr) { |
| if (obj == NULL) { |
| return NULL; |
| } |
| |
| // Prevent resurrection of unreachable phantom (i.e. weak-native) references. |
| if ((decorators & ON_PHANTOM_OOP_REF) != 0 && |
| _heap->is_concurrent_weak_root_in_progress() && |
| !_heap->marking_context()->is_marked(obj)) { |
| return NULL; |
| } |
| |
| // Prevent resurrection of unreachable weak references. |
| if ((decorators & ON_WEAK_OOP_REF) != 0 && |
| _heap->is_concurrent_weak_root_in_progress() && |
| !_heap->marking_context()->is_marked_strong(obj)) { |
| return NULL; |
| } |
| |
| // Prevent resurrection of unreachable objects that are visited during |
| // concurrent class-unloading. |
| if ((decorators & AS_NO_KEEPALIVE) != 0 && |
| _heap->is_evacuation_in_progress() && |
| !_heap->marking_context()->is_marked(obj)) { |
| return obj; |
| } |
| |
| oop fwd = load_reference_barrier(obj); |
| if (ShenandoahSelfFixing && load_addr != NULL && fwd != obj) { |
| // Since we are here and we know the load address, update the reference. |
| ShenandoahHeap::cas_oop(fwd, load_addr, obj); |
| } |
| |
| return fwd; |
| } |
| |
| inline void ShenandoahBarrierSet::enqueue(oop obj) { |
| assert(obj != NULL, "checked by caller"); |
| assert(_satb_mark_queue_set.is_active(), "only get here when SATB active"); |
| |
| // Filter marked objects before hitting the SATB queues. The same predicate would |
| // be used by SATBMQ::filter to eliminate already marked objects downstream, but |
| // filtering here helps to avoid wasteful SATB queueing work to begin with. |
| if (!_heap->requires_marking(obj)) return; |
| |
| SATBMarkQueue& queue = ShenandoahThreadLocalData::satb_mark_queue(Thread::current()); |
| _satb_mark_queue_set.enqueue_known_active(queue, obj); |
| } |
| |
| template <DecoratorSet decorators, typename T> |
| inline void ShenandoahBarrierSet::satb_barrier(T *field) { |
| if (HasDecorator<decorators, IS_DEST_UNINITIALIZED>::value || |
| HasDecorator<decorators, AS_NO_KEEPALIVE>::value) { |
| return; |
| } |
| if (ShenandoahSATBBarrier && _heap->is_concurrent_mark_in_progress()) { |
| T heap_oop = RawAccess<>::oop_load(field); |
| if (!CompressedOops::is_null(heap_oop)) { |
| enqueue(CompressedOops::decode(heap_oop)); |
| } |
| } |
| } |
| |
| inline void ShenandoahBarrierSet::satb_enqueue(oop value) { |
| if (value != NULL && ShenandoahSATBBarrier && _heap->is_concurrent_mark_in_progress()) { |
| enqueue(value); |
| } |
| } |
| |
| inline void ShenandoahBarrierSet::iu_barrier(oop obj) { |
| if (ShenandoahIUBarrier && obj != NULL && _heap->is_concurrent_mark_in_progress()) { |
| enqueue(obj); |
| } |
| } |
| |
| inline void ShenandoahBarrierSet::keep_alive_if_weak(DecoratorSet decorators, oop value) { |
| assert((decorators & ON_UNKNOWN_OOP_REF) == 0, "Reference strength must be known"); |
| const bool on_strong_oop_ref = (decorators & ON_STRONG_OOP_REF) != 0; |
| const bool peek = (decorators & AS_NO_KEEPALIVE) != 0; |
| if (!peek && !on_strong_oop_ref) { |
| satb_enqueue(value); |
| } |
| } |
| |
| template <typename T> |
| inline oop ShenandoahBarrierSet::oop_load(DecoratorSet decorators, T* addr) { |
| oop value = RawAccess<>::oop_load(addr); |
| value = load_reference_barrier(decorators, value, addr); |
| keep_alive_if_weak(decorators, value); |
| return value; |
| } |
| |
| template <typename T> |
| inline oop ShenandoahBarrierSet::oop_cmpxchg(DecoratorSet decorators, T* addr, oop compare_value, oop new_value) { |
| iu_barrier(new_value); |
| oop res; |
| oop expected = compare_value; |
| do { |
| compare_value = expected; |
| res = RawAccess<>::oop_atomic_cmpxchg(addr, compare_value, new_value); |
| expected = res; |
| } while ((compare_value != expected) && (resolve_forwarded(compare_value) == resolve_forwarded(expected))); |
| |
| // Note: We don't need a keep-alive-barrier here. We already enqueue any loaded reference for SATB anyway, |
| // because it must be the previous value. |
| res = load_reference_barrier(decorators, res, reinterpret_cast<T*>(NULL)); |
| satb_enqueue(res); |
| return res; |
| } |
| |
| template <typename T> |
| inline oop ShenandoahBarrierSet::oop_xchg(DecoratorSet decorators, T* addr, oop new_value) { |
| iu_barrier(new_value); |
| oop previous = RawAccess<>::oop_atomic_xchg(addr, new_value); |
| // Note: We don't need a keep-alive-barrier here. We already enqueue any loaded reference for SATB anyway, |
| // because it must be the previous value. |
| previous = load_reference_barrier<T>(decorators, previous, reinterpret_cast<T*>(NULL)); |
| satb_enqueue(previous); |
| return previous; |
| } |
| |
| template <DecoratorSet decorators, typename BarrierSetT> |
| template <typename T> |
| inline oop ShenandoahBarrierSet::AccessBarrier<decorators, BarrierSetT>::oop_load_not_in_heap(T* addr) { |
| assert((decorators & ON_UNKNOWN_OOP_REF) == 0, "must be absent"); |
| ShenandoahBarrierSet* const bs = ShenandoahBarrierSet::barrier_set(); |
| return bs->oop_load(decorators, addr); |
| } |
| |
| template <DecoratorSet decorators, typename BarrierSetT> |
| template <typename T> |
| inline oop ShenandoahBarrierSet::AccessBarrier<decorators, BarrierSetT>::oop_load_in_heap(T* addr) { |
| assert((decorators & ON_UNKNOWN_OOP_REF) == 0, "must be absent"); |
| ShenandoahBarrierSet* const bs = ShenandoahBarrierSet::barrier_set(); |
| return bs->oop_load(decorators, addr); |
| } |
| |
| template <DecoratorSet decorators, typename BarrierSetT> |
| inline oop ShenandoahBarrierSet::AccessBarrier<decorators, BarrierSetT>::oop_load_in_heap_at(oop base, ptrdiff_t offset) { |
| ShenandoahBarrierSet* const bs = ShenandoahBarrierSet::barrier_set(); |
| DecoratorSet resolved_decorators = AccessBarrierSupport::resolve_possibly_unknown_oop_ref_strength<decorators>(base, offset); |
| return bs->oop_load(resolved_decorators, AccessInternal::oop_field_addr<decorators>(base, offset)); |
| } |
| |
| template <DecoratorSet decorators, typename BarrierSetT> |
| template <typename T> |
| inline void ShenandoahBarrierSet::AccessBarrier<decorators, BarrierSetT>::oop_store_not_in_heap(T* addr, oop value) { |
| shenandoah_assert_marked_if(NULL, value, !CompressedOops::is_null(value) && ShenandoahHeap::heap()->is_evacuation_in_progress()); |
| shenandoah_assert_not_in_cset_if(addr, value, value != NULL && !ShenandoahHeap::heap()->cancelled_gc()); |
| ShenandoahBarrierSet* const bs = ShenandoahBarrierSet::barrier_set(); |
| bs->iu_barrier(value); |
| bs->satb_barrier<decorators>(addr); |
| Raw::oop_store(addr, value); |
| } |
| |
| template <DecoratorSet decorators, typename BarrierSetT> |
| template <typename T> |
| inline void ShenandoahBarrierSet::AccessBarrier<decorators, BarrierSetT>::oop_store_in_heap(T* addr, oop value) { |
| shenandoah_assert_not_in_cset_loc_except(addr, ShenandoahHeap::heap()->cancelled_gc()); |
| shenandoah_assert_not_forwarded_except (addr, value, value == NULL || ShenandoahHeap::heap()->cancelled_gc() || !ShenandoahHeap::heap()->is_concurrent_mark_in_progress()); |
| shenandoah_assert_not_in_cset_except (addr, value, value == NULL || ShenandoahHeap::heap()->cancelled_gc() || !ShenandoahHeap::heap()->is_concurrent_mark_in_progress()); |
| |
| oop_store_not_in_heap(addr, value); |
| } |
| |
| template <DecoratorSet decorators, typename BarrierSetT> |
| inline void ShenandoahBarrierSet::AccessBarrier<decorators, BarrierSetT>::oop_store_in_heap_at(oop base, ptrdiff_t offset, oop value) { |
| oop_store_in_heap(AccessInternal::oop_field_addr<decorators>(base, offset), value); |
| } |
| |
| template <DecoratorSet decorators, typename BarrierSetT> |
| template <typename T> |
| inline oop ShenandoahBarrierSet::AccessBarrier<decorators, BarrierSetT>::oop_atomic_cmpxchg_not_in_heap(T* addr, oop compare_value, oop new_value) { |
| assert((decorators & (AS_NO_KEEPALIVE | ON_UNKNOWN_OOP_REF)) == 0, "must be absent"); |
| ShenandoahBarrierSet* bs = ShenandoahBarrierSet::barrier_set(); |
| return bs->oop_cmpxchg(decorators, addr, compare_value, new_value); |
| } |
| |
| template <DecoratorSet decorators, typename BarrierSetT> |
| template <typename T> |
| inline oop ShenandoahBarrierSet::AccessBarrier<decorators, BarrierSetT>::oop_atomic_cmpxchg_in_heap(T* addr, oop compare_value, oop new_value) { |
| assert((decorators & (AS_NO_KEEPALIVE | ON_UNKNOWN_OOP_REF)) == 0, "must be absent"); |
| ShenandoahBarrierSet* bs = ShenandoahBarrierSet::barrier_set(); |
| return bs->oop_cmpxchg(decorators, addr, compare_value, new_value); |
| } |
| |
| template <DecoratorSet decorators, typename BarrierSetT> |
| inline oop ShenandoahBarrierSet::AccessBarrier<decorators, BarrierSetT>::oop_atomic_cmpxchg_in_heap_at(oop base, ptrdiff_t offset, oop compare_value, oop new_value) { |
| assert((decorators & AS_NO_KEEPALIVE) == 0, "must be absent"); |
| ShenandoahBarrierSet* bs = ShenandoahBarrierSet::barrier_set(); |
| DecoratorSet resolved_decorators = AccessBarrierSupport::resolve_possibly_unknown_oop_ref_strength<decorators>(base, offset); |
| return bs->oop_cmpxchg(resolved_decorators, AccessInternal::oop_field_addr<decorators>(base, offset), compare_value, new_value); |
| } |
| |
| template <DecoratorSet decorators, typename BarrierSetT> |
| template <typename T> |
| inline oop ShenandoahBarrierSet::AccessBarrier<decorators, BarrierSetT>::oop_atomic_xchg_not_in_heap(T* addr, oop new_value) { |
| assert((decorators & (AS_NO_KEEPALIVE | ON_UNKNOWN_OOP_REF)) == 0, "must be absent"); |
| ShenandoahBarrierSet* bs = ShenandoahBarrierSet::barrier_set(); |
| return bs->oop_xchg(decorators, addr, new_value); |
| } |
| |
| template <DecoratorSet decorators, typename BarrierSetT> |
| template <typename T> |
| inline oop ShenandoahBarrierSet::AccessBarrier<decorators, BarrierSetT>::oop_atomic_xchg_in_heap(T* addr, oop new_value) { |
| assert((decorators & (AS_NO_KEEPALIVE | ON_UNKNOWN_OOP_REF)) == 0, "must be absent"); |
| ShenandoahBarrierSet* bs = ShenandoahBarrierSet::barrier_set(); |
| return bs->oop_xchg(decorators, addr, new_value); |
| } |
| |
| template <DecoratorSet decorators, typename BarrierSetT> |
| inline oop ShenandoahBarrierSet::AccessBarrier<decorators, BarrierSetT>::oop_atomic_xchg_in_heap_at(oop base, ptrdiff_t offset, oop new_value) { |
| assert((decorators & AS_NO_KEEPALIVE) == 0, "must be absent"); |
| ShenandoahBarrierSet* bs = ShenandoahBarrierSet::barrier_set(); |
| DecoratorSet resolved_decorators = AccessBarrierSupport::resolve_possibly_unknown_oop_ref_strength<decorators>(base, offset); |
| return bs->oop_xchg(resolved_decorators, AccessInternal::oop_field_addr<decorators>(base, offset), new_value); |
| } |
| |
| // Clone barrier support |
| template <DecoratorSet decorators, typename BarrierSetT> |
| void ShenandoahBarrierSet::AccessBarrier<decorators, BarrierSetT>::clone_in_heap(oop src, oop dst, size_t size) { |
| if (ShenandoahCloneBarrier) { |
| ShenandoahBarrierSet::barrier_set()->clone_barrier_runtime(src); |
| } |
| Raw::clone(src, dst, size); |
| } |
| |
| template <DecoratorSet decorators, typename BarrierSetT> |
| template <typename T> |
| bool ShenandoahBarrierSet::AccessBarrier<decorators, BarrierSetT>::oop_arraycopy_in_heap(arrayOop src_obj, size_t src_offset_in_bytes, T* src_raw, |
| arrayOop dst_obj, size_t dst_offset_in_bytes, T* dst_raw, |
| size_t length) { |
| ShenandoahBarrierSet* bs = ShenandoahBarrierSet::barrier_set(); |
| bs->arraycopy_barrier(arrayOopDesc::obj_offset_to_raw(src_obj, src_offset_in_bytes, src_raw), |
| arrayOopDesc::obj_offset_to_raw(dst_obj, dst_offset_in_bytes, dst_raw), |
| length); |
| return Raw::oop_arraycopy_in_heap(src_obj, src_offset_in_bytes, src_raw, dst_obj, dst_offset_in_bytes, dst_raw, length); |
| } |
| |
| template <class T, bool HAS_FWD, bool EVAC, bool ENQUEUE> |
| void ShenandoahBarrierSet::arraycopy_work(T* src, size_t count) { |
| assert(HAS_FWD == _heap->has_forwarded_objects(), "Forwarded object status is sane"); |
| |
| Thread* thread = Thread::current(); |
| SATBMarkQueue& queue = ShenandoahThreadLocalData::satb_mark_queue(thread); |
| ShenandoahMarkingContext* ctx = _heap->marking_context(); |
| const ShenandoahCollectionSet* const cset = _heap->collection_set(); |
| T* end = src + count; |
| for (T* elem_ptr = src; elem_ptr < end; elem_ptr++) { |
| T o = RawAccess<>::oop_load(elem_ptr); |
| if (!CompressedOops::is_null(o)) { |
| oop obj = CompressedOops::decode_not_null(o); |
| if (HAS_FWD && cset->is_in(obj)) { |
| oop fwd = resolve_forwarded_not_null(obj); |
| if (EVAC && obj == fwd) { |
| fwd = _heap->evacuate_object(obj, thread); |
| } |
| assert(obj != fwd || _heap->cancelled_gc(), "must be forwarded"); |
| oop witness = ShenandoahHeap::cas_oop(fwd, elem_ptr, o); |
| obj = fwd; |
| } |
| if (ENQUEUE && !ctx->is_marked_strong(obj)) { |
| _satb_mark_queue_set.enqueue_known_active(queue, obj); |
| } |
| } |
| } |
| } |
| |
| template <class T> |
| void ShenandoahBarrierSet::arraycopy_barrier(T* src, T* dst, size_t count) { |
| if (count == 0) { |
| return; |
| } |
| int gc_state = _heap->gc_state(); |
| if ((gc_state & ShenandoahHeap::MARKING) != 0) { |
| arraycopy_marking(src, dst, count); |
| } else if ((gc_state & ShenandoahHeap::EVACUATION) != 0) { |
| arraycopy_evacuation(src, count); |
| } else if ((gc_state & ShenandoahHeap::UPDATEREFS) != 0) { |
| arraycopy_update(src, count); |
| } |
| } |
| |
| template <class T> |
| void ShenandoahBarrierSet::arraycopy_marking(T* src, T* dst, size_t count) { |
| assert(_heap->is_concurrent_mark_in_progress(), "only during marking"); |
| T* array = ShenandoahSATBBarrier ? dst : src; |
| if (!_heap->marking_context()->allocated_after_mark_start(reinterpret_cast<HeapWord*>(array))) { |
| arraycopy_work<T, false, false, true>(array, count); |
| } |
| } |
| |
| inline bool ShenandoahBarrierSet::need_bulk_update(HeapWord* ary) { |
| return ary < _heap->heap_region_containing(ary)->get_update_watermark(); |
| } |
| |
| template <class T> |
| void ShenandoahBarrierSet::arraycopy_evacuation(T* src, size_t count) { |
| assert(_heap->is_evacuation_in_progress(), "only during evacuation"); |
| if (need_bulk_update(reinterpret_cast<HeapWord*>(src))) { |
| ShenandoahEvacOOMScope oom_evac; |
| arraycopy_work<T, true, true, false>(src, count); |
| } |
| } |
| |
| template <class T> |
| void ShenandoahBarrierSet::arraycopy_update(T* src, size_t count) { |
| assert(_heap->is_update_refs_in_progress(), "only during update-refs"); |
| if (need_bulk_update(reinterpret_cast<HeapWord*>(src))) { |
| arraycopy_work<T, true, false, false>(src, count); |
| } |
| } |
| |
| #endif // SHARE_GC_SHENANDOAH_SHENANDOAHBARRIERSET_INLINE_HPP |