提交 df0c350a 编写于 作者: B brutisso

8043239: G1: Missing post barrier in processing of j.l.ref.Reference objects

Summary: Removed all write barriers during reference processing and added explicit write barriers when iterating through the discovered list.
Reviewed-by: pliden, jmasa, tschatzl
上级 7f5cdefb
...@@ -310,8 +310,7 @@ void CMSCollector::ref_processor_init() { ...@@ -310,8 +310,7 @@ void CMSCollector::ref_processor_init() {
_cmsGen->refs_discovery_is_mt(), // mt discovery _cmsGen->refs_discovery_is_mt(), // mt discovery
(int) MAX2(ConcGCThreads, ParallelGCThreads), // mt discovery degree (int) MAX2(ConcGCThreads, ParallelGCThreads), // mt discovery degree
_cmsGen->refs_discovery_is_atomic(), // discovery is not atomic _cmsGen->refs_discovery_is_atomic(), // discovery is not atomic
&_is_alive_closure, // closure for liveness info &_is_alive_closure); // closure for liveness info
false); // next field updates do not need write barrier
// Initialize the _ref_processor field of CMSGen // Initialize the _ref_processor field of CMSGen
_cmsGen->set_ref_processor(_ref_processor); _cmsGen->set_ref_processor(_ref_processor);
......
...@@ -2258,12 +2258,9 @@ void G1CollectedHeap::ref_processing_init() { ...@@ -2258,12 +2258,9 @@ void G1CollectedHeap::ref_processing_init() {
// degree of mt discovery // degree of mt discovery
false, false,
// Reference discovery is not atomic // Reference discovery is not atomic
&_is_alive_closure_cm, &_is_alive_closure_cm);
// is alive closure // is alive closure
// (for efficiency/performance) // (for efficiency/performance)
true);
// Setting next fields of discovered
// lists requires a barrier.
// STW ref processor // STW ref processor
_ref_processor_stw = _ref_processor_stw =
...@@ -2278,12 +2275,9 @@ void G1CollectedHeap::ref_processing_init() { ...@@ -2278,12 +2275,9 @@ void G1CollectedHeap::ref_processing_init() {
// degree of mt discovery // degree of mt discovery
true, true,
// Reference discovery is atomic // Reference discovery is atomic
&_is_alive_closure_stw, &_is_alive_closure_stw);
// is alive closure // is alive closure
// (for efficiency/performance) // (for efficiency/performance)
false);
// Setting next fields of discovered
// lists does not require a barrier.
} }
size_t G1CollectedHeap::capacity() const { size_t G1CollectedHeap::capacity() const {
......
...@@ -1638,8 +1638,7 @@ void ParNewGeneration::ref_processor_init() { ...@@ -1638,8 +1638,7 @@ void ParNewGeneration::ref_processor_init() {
refs_discovery_is_mt(), // mt discovery refs_discovery_is_mt(), // mt discovery
(int) ParallelGCThreads, // mt discovery degree (int) ParallelGCThreads, // mt discovery degree
refs_discovery_is_atomic(), // atomic_discovery refs_discovery_is_atomic(), // atomic_discovery
NULL, // is_alive_non_header NULL); // is_alive_non_header
false); // write barrier for next field updates
} }
} }
......
...@@ -853,8 +853,7 @@ void PSParallelCompact::post_initialize() { ...@@ -853,8 +853,7 @@ void PSParallelCompact::post_initialize() {
true, // mt discovery true, // mt discovery
(int) ParallelGCThreads, // mt discovery degree (int) ParallelGCThreads, // mt discovery degree
true, // atomic_discovery true, // atomic_discovery
&_is_alive_closure, // non-header is alive closure &_is_alive_closure); // non-header is alive closure
false); // write barrier for next field updates
_counters = new CollectorCounters("PSParallelCompact", 1); _counters = new CollectorCounters("PSParallelCompact", 1);
// Initialize static fields in ParCompactionManager. // Initialize static fields in ParCompactionManager.
......
...@@ -861,8 +861,7 @@ void PSScavenge::initialize() { ...@@ -861,8 +861,7 @@ void PSScavenge::initialize() {
true, // mt discovery true, // mt discovery
(int) ParallelGCThreads, // mt discovery degree (int) ParallelGCThreads, // mt discovery degree
true, // atomic_discovery true, // atomic_discovery
NULL, // header provides liveness info NULL); // header provides liveness info
false); // next field updates do not need write barrier
// Cache the cardtable // Cache the cardtable
BarrierSet* bs = Universe::heap()->barrier_set(); BarrierSet* bs = Universe::heap()->barrier_set();
......
...@@ -96,12 +96,10 @@ ReferenceProcessor::ReferenceProcessor(MemRegion span, ...@@ -96,12 +96,10 @@ ReferenceProcessor::ReferenceProcessor(MemRegion span,
bool mt_discovery, bool mt_discovery,
uint mt_discovery_degree, uint mt_discovery_degree,
bool atomic_discovery, bool atomic_discovery,
BoolObjectClosure* is_alive_non_header, BoolObjectClosure* is_alive_non_header) :
bool discovered_list_needs_post_barrier) :
_discovering_refs(false), _discovering_refs(false),
_enqueuing_is_done(false), _enqueuing_is_done(false),
_is_alive_non_header(is_alive_non_header), _is_alive_non_header(is_alive_non_header),
_discovered_list_needs_post_barrier(discovered_list_needs_post_barrier),
_processing_is_mt(mt_processing), _processing_is_mt(mt_processing),
_next_id(0) _next_id(0)
{ {
...@@ -340,10 +338,18 @@ void ReferenceProcessor::enqueue_discovered_reflist(DiscoveredList& refs_list, ...@@ -340,10 +338,18 @@ void ReferenceProcessor::enqueue_discovered_reflist(DiscoveredList& refs_list,
// (java.lang.ref.Reference.discovered), self-loop their "next" field // (java.lang.ref.Reference.discovered), self-loop their "next" field
// thus distinguishing them from active References, then // thus distinguishing them from active References, then
// prepend them to the pending list. // prepend them to the pending list.
//
// The Java threads will see the Reference objects linked together through
// the discovered field. Instead of trying to do the write barrier updates
// in all places in the reference processor where we manipulate the discovered
// field we make sure to do the barrier here where we anyway iterate through
// all linked Reference objects. Note that it is important to not dirty any
// cards during reference processing since this will cause card table
// verification to fail for G1.
//
// BKWRD COMPATIBILITY NOTE: For older JDKs (prior to the fix for 4956777), // BKWRD COMPATIBILITY NOTE: For older JDKs (prior to the fix for 4956777),
// the "next" field is used to chain the pending list, not the discovered // the "next" field is used to chain the pending list, not the discovered
// field. // field.
if (TraceReferenceGC && PrintGCDetails) { if (TraceReferenceGC && PrintGCDetails) {
gclog_or_tty->print_cr("ReferenceProcessor::enqueue_discovered_reflist list " gclog_or_tty->print_cr("ReferenceProcessor::enqueue_discovered_reflist list "
INTPTR_FORMAT, (address)refs_list.head()); INTPTR_FORMAT, (address)refs_list.head());
...@@ -351,7 +357,7 @@ void ReferenceProcessor::enqueue_discovered_reflist(DiscoveredList& refs_list, ...@@ -351,7 +357,7 @@ void ReferenceProcessor::enqueue_discovered_reflist(DiscoveredList& refs_list,
oop obj = NULL; oop obj = NULL;
oop next_d = refs_list.head(); oop next_d = refs_list.head();
if (pending_list_uses_discovered_field()) { // New behaviour if (pending_list_uses_discovered_field()) { // New behavior
// Walk down the list, self-looping the next field // Walk down the list, self-looping the next field
// so that the References are not considered active. // so that the References are not considered active.
while (obj != next_d) { while (obj != next_d) {
...@@ -365,15 +371,15 @@ void ReferenceProcessor::enqueue_discovered_reflist(DiscoveredList& refs_list, ...@@ -365,15 +371,15 @@ void ReferenceProcessor::enqueue_discovered_reflist(DiscoveredList& refs_list,
assert(java_lang_ref_Reference::next(obj) == NULL, assert(java_lang_ref_Reference::next(obj) == NULL,
"Reference not active; should not be discovered"); "Reference not active; should not be discovered");
// Self-loop next, so as to make Ref not active. // Self-loop next, so as to make Ref not active.
// Post-barrier not needed when looping to self.
java_lang_ref_Reference::set_next_raw(obj, obj); java_lang_ref_Reference::set_next_raw(obj, obj);
if (next_d == obj) { // obj is last if (next_d != obj) {
// Swap refs_list into pendling_list_addr and oopDesc::bs()->write_ref_field(java_lang_ref_Reference::discovered_addr(obj), next_d);
} else {
// This is the last object.
// Swap refs_list into pending_list_addr and
// set obj's discovered to what we read from pending_list_addr. // set obj's discovered to what we read from pending_list_addr.
oop old = oopDesc::atomic_exchange_oop(refs_list.head(), pending_list_addr); oop old = oopDesc::atomic_exchange_oop(refs_list.head(), pending_list_addr);
// Need post-barrier on pending_list_addr above; // Need post-barrier on pending_list_addr. See enqueue_discovered_ref_helper() above.
// see special post-barrier code at the end of
// enqueue_discovered_reflists() further below.
java_lang_ref_Reference::set_discovered_raw(obj, old); // old may be NULL java_lang_ref_Reference::set_discovered_raw(obj, old); // old may be NULL
oopDesc::bs()->write_ref_field(java_lang_ref_Reference::discovered_addr(obj), old); oopDesc::bs()->write_ref_field(java_lang_ref_Reference::discovered_addr(obj), old);
} }
...@@ -496,20 +502,15 @@ void DiscoveredListIterator::remove() { ...@@ -496,20 +502,15 @@ void DiscoveredListIterator::remove() {
// pre-barrier here because we know the Reference has already been found/marked, // pre-barrier here because we know the Reference has already been found/marked,
// that's how it ended up in the discovered list in the first place. // that's how it ended up in the discovered list in the first place.
oop_store_raw(_prev_next, new_next); oop_store_raw(_prev_next, new_next);
if (_discovered_list_needs_post_barrier && _prev_next != _refs_list.adr_head()) {
// Needs post-barrier and this is not the list head (which is not on the heap)
oopDesc::bs()->write_ref_field(_prev_next, new_next);
}
NOT_PRODUCT(_removed++); NOT_PRODUCT(_removed++);
_refs_list.dec_length(1); _refs_list.dec_length(1);
} }
// Make the Reference object active again. // Make the Reference object active again.
void DiscoveredListIterator::make_active() { void DiscoveredListIterator::make_active() {
// For G1 we don't want to use set_next - it // The pre barrier for G1 is probably just needed for the old
// will dirty the card for the next field of // reference processing behavior. Should we guard this with
// the reference object and will fail // ReferenceProcessor::pending_list_uses_discovered_field() ?
// CT verification.
if (UseG1GC) { if (UseG1GC) {
HeapWord* next_addr = java_lang_ref_Reference::next_addr(_ref); HeapWord* next_addr = java_lang_ref_Reference::next_addr(_ref);
if (UseCompressedOops) { if (UseCompressedOops) {
...@@ -517,10 +518,8 @@ void DiscoveredListIterator::make_active() { ...@@ -517,10 +518,8 @@ void DiscoveredListIterator::make_active() {
} else { } else {
oopDesc::bs()->write_ref_field_pre((oop*)next_addr, NULL); oopDesc::bs()->write_ref_field_pre((oop*)next_addr, NULL);
} }
java_lang_ref_Reference::set_next_raw(_ref, NULL);
} else {
java_lang_ref_Reference::set_next(_ref, NULL);
} }
java_lang_ref_Reference::set_next_raw(_ref, NULL);
} }
void DiscoveredListIterator::clear_referent() { void DiscoveredListIterator::clear_referent() {
...@@ -546,7 +545,7 @@ ReferenceProcessor::process_phase1(DiscoveredList& refs_list, ...@@ -546,7 +545,7 @@ ReferenceProcessor::process_phase1(DiscoveredList& refs_list,
OopClosure* keep_alive, OopClosure* keep_alive,
VoidClosure* complete_gc) { VoidClosure* complete_gc) {
assert(policy != NULL, "Must have a non-NULL policy"); assert(policy != NULL, "Must have a non-NULL policy");
DiscoveredListIterator iter(refs_list, keep_alive, is_alive, _discovered_list_needs_post_barrier); DiscoveredListIterator iter(refs_list, keep_alive, is_alive);
// Decide which softly reachable refs should be kept alive. // Decide which softly reachable refs should be kept alive.
while (iter.has_next()) { while (iter.has_next()) {
iter.load_ptrs(DEBUG_ONLY(!discovery_is_atomic() /* allow_null_referent */)); iter.load_ptrs(DEBUG_ONLY(!discovery_is_atomic() /* allow_null_referent */));
...@@ -586,7 +585,7 @@ ReferenceProcessor::pp2_work(DiscoveredList& refs_list, ...@@ -586,7 +585,7 @@ ReferenceProcessor::pp2_work(DiscoveredList& refs_list,
BoolObjectClosure* is_alive, BoolObjectClosure* is_alive,
OopClosure* keep_alive) { OopClosure* keep_alive) {
assert(discovery_is_atomic(), "Error"); assert(discovery_is_atomic(), "Error");
DiscoveredListIterator iter(refs_list, keep_alive, is_alive, _discovered_list_needs_post_barrier); DiscoveredListIterator iter(refs_list, keep_alive, is_alive);
while (iter.has_next()) { while (iter.has_next()) {
iter.load_ptrs(DEBUG_ONLY(false /* allow_null_referent */)); iter.load_ptrs(DEBUG_ONLY(false /* allow_null_referent */));
DEBUG_ONLY(oop next = java_lang_ref_Reference::next(iter.obj());) DEBUG_ONLY(oop next = java_lang_ref_Reference::next(iter.obj());)
...@@ -623,7 +622,7 @@ ReferenceProcessor::pp2_work_concurrent_discovery(DiscoveredList& refs_list, ...@@ -623,7 +622,7 @@ ReferenceProcessor::pp2_work_concurrent_discovery(DiscoveredList& refs_list,
OopClosure* keep_alive, OopClosure* keep_alive,
VoidClosure* complete_gc) { VoidClosure* complete_gc) {
assert(!discovery_is_atomic(), "Error"); assert(!discovery_is_atomic(), "Error");
DiscoveredListIterator iter(refs_list, keep_alive, is_alive, _discovered_list_needs_post_barrier); DiscoveredListIterator iter(refs_list, keep_alive, is_alive);
while (iter.has_next()) { while (iter.has_next()) {
iter.load_ptrs(DEBUG_ONLY(true /* allow_null_referent */)); iter.load_ptrs(DEBUG_ONLY(true /* allow_null_referent */));
HeapWord* next_addr = java_lang_ref_Reference::next_addr(iter.obj()); HeapWord* next_addr = java_lang_ref_Reference::next_addr(iter.obj());
...@@ -666,7 +665,7 @@ ReferenceProcessor::process_phase3(DiscoveredList& refs_list, ...@@ -666,7 +665,7 @@ ReferenceProcessor::process_phase3(DiscoveredList& refs_list,
OopClosure* keep_alive, OopClosure* keep_alive,
VoidClosure* complete_gc) { VoidClosure* complete_gc) {
ResourceMark rm; ResourceMark rm;
DiscoveredListIterator iter(refs_list, keep_alive, is_alive, _discovered_list_needs_post_barrier); DiscoveredListIterator iter(refs_list, keep_alive, is_alive);
while (iter.has_next()) { while (iter.has_next()) {
iter.update_discovered(); iter.update_discovered();
iter.load_ptrs(DEBUG_ONLY(false /* allow_null_referent */)); iter.load_ptrs(DEBUG_ONLY(false /* allow_null_referent */));
...@@ -782,13 +781,6 @@ private: ...@@ -782,13 +781,6 @@ private:
bool _clear_referent; bool _clear_referent;
}; };
void ReferenceProcessor::set_discovered(oop ref, oop value) {
java_lang_ref_Reference::set_discovered_raw(ref, value);
if (_discovered_list_needs_post_barrier) {
oopDesc::bs()->write_ref_field(java_lang_ref_Reference::discovered_addr(ref), value);
}
}
// Balances reference queues. // Balances reference queues.
// Move entries from all queues[0, 1, ..., _max_num_q-1] to // Move entries from all queues[0, 1, ..., _max_num_q-1] to
// queues[0, 1, ..., _num_q-1] because only the first _num_q // queues[0, 1, ..., _num_q-1] because only the first _num_q
...@@ -846,9 +838,9 @@ void ReferenceProcessor::balance_queues(DiscoveredList ref_lists[]) ...@@ -846,9 +838,9 @@ void ReferenceProcessor::balance_queues(DiscoveredList ref_lists[])
// Add the chain to the to list. // Add the chain to the to list.
if (ref_lists[to_idx].head() == NULL) { if (ref_lists[to_idx].head() == NULL) {
// to list is empty. Make a loop at the end. // to list is empty. Make a loop at the end.
set_discovered(move_tail, move_tail); java_lang_ref_Reference::set_discovered_raw(move_tail, move_tail);
} else { } else {
set_discovered(move_tail, ref_lists[to_idx].head()); java_lang_ref_Reference::set_discovered_raw(move_tail, ref_lists[to_idx].head());
} }
ref_lists[to_idx].set_head(move_head); ref_lists[to_idx].set_head(move_head);
ref_lists[to_idx].inc_length(refs_to_move); ref_lists[to_idx].inc_length(refs_to_move);
...@@ -982,7 +974,7 @@ void ReferenceProcessor::clean_up_discovered_references() { ...@@ -982,7 +974,7 @@ void ReferenceProcessor::clean_up_discovered_references() {
void ReferenceProcessor::clean_up_discovered_reflist(DiscoveredList& refs_list) { void ReferenceProcessor::clean_up_discovered_reflist(DiscoveredList& refs_list) {
assert(!discovery_is_atomic(), "Else why call this method?"); assert(!discovery_is_atomic(), "Else why call this method?");
DiscoveredListIterator iter(refs_list, NULL, NULL, _discovered_list_needs_post_barrier); DiscoveredListIterator iter(refs_list, NULL, NULL);
while (iter.has_next()) { while (iter.has_next()) {
iter.load_ptrs(DEBUG_ONLY(true /* allow_null_referent */)); iter.load_ptrs(DEBUG_ONLY(true /* allow_null_referent */));
oop next = java_lang_ref_Reference::next(iter.obj()); oop next = java_lang_ref_Reference::next(iter.obj());
...@@ -1071,16 +1063,6 @@ ReferenceProcessor::add_to_discovered_list_mt(DiscoveredList& refs_list, ...@@ -1071,16 +1063,6 @@ ReferenceProcessor::add_to_discovered_list_mt(DiscoveredList& refs_list,
// The last ref must have its discovered field pointing to itself. // The last ref must have its discovered field pointing to itself.
oop next_discovered = (current_head != NULL) ? current_head : obj; oop next_discovered = (current_head != NULL) ? current_head : obj;
// Note: In the case of G1, this specific pre-barrier is strictly
// not necessary because the only case we are interested in
// here is when *discovered_addr is NULL (see the CAS further below),
// so this will expand to nothing. As a result, we have manually
// elided this out for G1, but left in the test for some future
// collector that might have need for a pre-barrier here, e.g.:-
// oopDesc::bs()->write_ref_field_pre((oop* or narrowOop*)discovered_addr, next_discovered);
assert(!_discovered_list_needs_post_barrier || UseG1GC,
"Need to check non-G1 collector: "
"may need a pre-write-barrier for CAS from NULL below");
oop retest = oopDesc::atomic_compare_exchange_oop(next_discovered, discovered_addr, oop retest = oopDesc::atomic_compare_exchange_oop(next_discovered, discovered_addr,
NULL); NULL);
if (retest == NULL) { if (retest == NULL) {
...@@ -1089,9 +1071,6 @@ ReferenceProcessor::add_to_discovered_list_mt(DiscoveredList& refs_list, ...@@ -1089,9 +1071,6 @@ ReferenceProcessor::add_to_discovered_list_mt(DiscoveredList& refs_list,
// is necessary. // is necessary.
refs_list.set_head(obj); refs_list.set_head(obj);
refs_list.inc_length(1); refs_list.inc_length(1);
if (_discovered_list_needs_post_barrier) {
oopDesc::bs()->write_ref_field((void*)discovered_addr, next_discovered);
}
if (TraceReferenceGC) { if (TraceReferenceGC) {
gclog_or_tty->print_cr("Discovered reference (mt) (" INTPTR_FORMAT ": %s)", gclog_or_tty->print_cr("Discovered reference (mt) (" INTPTR_FORMAT ": %s)",
...@@ -1242,24 +1221,14 @@ bool ReferenceProcessor::discover_reference(oop obj, ReferenceType rt) { ...@@ -1242,24 +1221,14 @@ bool ReferenceProcessor::discover_reference(oop obj, ReferenceType rt) {
if (_discovery_is_mt) { if (_discovery_is_mt) {
add_to_discovered_list_mt(*list, obj, discovered_addr); add_to_discovered_list_mt(*list, obj, discovered_addr);
} else { } else {
// If "_discovered_list_needs_post_barrier", we do write barriers when // We do a raw store here: the field will be visited later when processing
// updating the discovered reference list. Otherwise, we do a raw store // the discovered references.
// here: the field will be visited later when processing the discovered
// references.
oop current_head = list->head(); oop current_head = list->head();
// The last ref must have its discovered field pointing to itself. // The last ref must have its discovered field pointing to itself.
oop next_discovered = (current_head != NULL) ? current_head : obj; oop next_discovered = (current_head != NULL) ? current_head : obj;
// As in the case further above, since we are over-writing a NULL
// pre-value, we can safely elide the pre-barrier here for the case of G1.
// e.g.:- oopDesc::bs()->write_ref_field_pre((oop* or narrowOop*)discovered_addr, next_discovered);
assert(discovered == NULL, "control point invariant"); assert(discovered == NULL, "control point invariant");
assert(!_discovered_list_needs_post_barrier || UseG1GC,
"For non-G1 collector, may need a pre-write-barrier for CAS from NULL below");
oop_store_raw(discovered_addr, next_discovered); oop_store_raw(discovered_addr, next_discovered);
if (_discovered_list_needs_post_barrier) {
oopDesc::bs()->write_ref_field((void*)discovered_addr, next_discovered);
}
list->set_head(obj); list->set_head(obj);
list->inc_length(1); list->inc_length(1);
...@@ -1353,7 +1322,7 @@ ReferenceProcessor::preclean_discovered_reflist(DiscoveredList& refs_list, ...@@ -1353,7 +1322,7 @@ ReferenceProcessor::preclean_discovered_reflist(DiscoveredList& refs_list,
OopClosure* keep_alive, OopClosure* keep_alive,
VoidClosure* complete_gc, VoidClosure* complete_gc,
YieldClosure* yield) { YieldClosure* yield) {
DiscoveredListIterator iter(refs_list, keep_alive, is_alive, _discovered_list_needs_post_barrier); DiscoveredListIterator iter(refs_list, keep_alive, is_alive);
while (iter.has_next()) { while (iter.has_next()) {
iter.load_ptrs(DEBUG_ONLY(true /* allow_null_referent */)); iter.load_ptrs(DEBUG_ONLY(true /* allow_null_referent */));
oop obj = iter.obj(); oop obj = iter.obj();
......
...@@ -99,7 +99,6 @@ private: ...@@ -99,7 +99,6 @@ private:
oop _referent; oop _referent;
OopClosure* _keep_alive; OopClosure* _keep_alive;
BoolObjectClosure* _is_alive; BoolObjectClosure* _is_alive;
bool _discovered_list_needs_post_barrier;
DEBUG_ONLY( DEBUG_ONLY(
oop _first_seen; // cyclic linked list check oop _first_seen; // cyclic linked list check
...@@ -113,8 +112,7 @@ private: ...@@ -113,8 +112,7 @@ private:
public: public:
inline DiscoveredListIterator(DiscoveredList& refs_list, inline DiscoveredListIterator(DiscoveredList& refs_list,
OopClosure* keep_alive, OopClosure* keep_alive,
BoolObjectClosure* is_alive, BoolObjectClosure* is_alive):
bool discovered_list_needs_post_barrier = false):
_refs_list(refs_list), _refs_list(refs_list),
_prev_next(refs_list.adr_head()), _prev_next(refs_list.adr_head()),
_prev(NULL), _prev(NULL),
...@@ -128,8 +126,7 @@ public: ...@@ -128,8 +126,7 @@ public:
#endif #endif
_next(NULL), _next(NULL),
_keep_alive(keep_alive), _keep_alive(keep_alive),
_is_alive(is_alive), _is_alive(is_alive)
_discovered_list_needs_post_barrier(discovered_list_needs_post_barrier)
{ } { }
// End Of List. // End Of List.
...@@ -230,14 +227,6 @@ class ReferenceProcessor : public CHeapObj<mtGC> { ...@@ -230,14 +227,6 @@ class ReferenceProcessor : public CHeapObj<mtGC> {
// other collectors in configuration // other collectors in configuration
bool _discovery_is_mt; // true if reference discovery is MT. bool _discovery_is_mt; // true if reference discovery is MT.
// If true, setting "next" field of a discovered refs list requires
// write post barrier. (Must be true if used in a collector in which
// elements of a discovered list may be moved during discovery: for
// example, a collector like Garbage-First that moves objects during a
// long-term concurrent marking phase that does weak reference
// discovery.)
bool _discovered_list_needs_post_barrier;
bool _enqueuing_is_done; // true if all weak references enqueued bool _enqueuing_is_done; // true if all weak references enqueued
bool _processing_is_mt; // true during phases when bool _processing_is_mt; // true during phases when
// reference processing is MT. // reference processing is MT.
...@@ -382,11 +371,6 @@ class ReferenceProcessor : public CHeapObj<mtGC> { ...@@ -382,11 +371,6 @@ class ReferenceProcessor : public CHeapObj<mtGC> {
void enqueue_discovered_reflists(HeapWord* pending_list_addr, AbstractRefProcTaskExecutor* task_executor); void enqueue_discovered_reflists(HeapWord* pending_list_addr, AbstractRefProcTaskExecutor* task_executor);
protected: protected:
// Set the 'discovered' field of the given reference to
// the given value - emitting post barriers depending upon
// the value of _discovered_list_needs_post_barrier.
void set_discovered(oop ref, oop value);
// "Preclean" the given discovered reference list // "Preclean" the given discovered reference list
// by removing references with strongly reachable referents. // by removing references with strongly reachable referents.
// Currently used in support of CMS only. // Currently used in support of CMS only.
...@@ -427,8 +411,7 @@ class ReferenceProcessor : public CHeapObj<mtGC> { ...@@ -427,8 +411,7 @@ class ReferenceProcessor : public CHeapObj<mtGC> {
bool mt_processing = false, uint mt_processing_degree = 1, bool mt_processing = false, uint mt_processing_degree = 1,
bool mt_discovery = false, uint mt_discovery_degree = 1, bool mt_discovery = false, uint mt_discovery_degree = 1,
bool atomic_discovery = true, bool atomic_discovery = true,
BoolObjectClosure* is_alive_non_header = NULL, BoolObjectClosure* is_alive_non_header = NULL);
bool discovered_list_needs_post_barrier = false);
// RefDiscoveryPolicy values // RefDiscoveryPolicy values
enum DiscoveryPolicy { enum DiscoveryPolicy {
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册