--- old/src/share/vm/gc/g1/g1CollectedHeap.cpp 2017-06-14 16:39:38.423497273 +0200 +++ new/src/share/vm/gc/g1/g1CollectedHeap.cpp 2017-06-14 16:39:38.151486536 +0200 @@ -51,7 +51,7 @@ #include "gc/g1/g1RemSet.inline.hpp" #include "gc/g1/g1RootClosures.hpp" #include "gc/g1/g1RootProcessor.hpp" -#include "gc/g1/g1SerialCollector.hpp" +#include "gc/g1/g1SerialFullCollector.hpp" #include "gc/g1/g1StringDedup.hpp" #include "gc/g1/g1YCTypes.hpp" #include "gc/g1/heapRegion.inline.hpp" @@ -1136,25 +1136,48 @@ collector_state()->set_gcs_are_young(true); } -void G1CollectedHeap::reset_card_cache_and_queue() { +void G1CollectedHeap::verify_before_full_collection(bool explicit_gc) { + assert(!GCCause::is_user_requested_gc(gc_cause()) || explicit_gc, "invariant"); + assert(used() == recalculate_used(), "Should be equal"); + _verifier->verify_region_sets_optional(); + _verifier->verify_before_gc(); + _verifier->check_bitmaps("Full GC Start"); +} + +void G1CollectedHeap::prepare_heap_for_mutators() { + // Delete metaspaces for unloaded class loaders and clean up loader_data graph + ClassLoaderDataGraph::purge(); + MetaspaceAux::verify_metrics(); + + // Prepare heap for normal collections. + assert(num_free_regions() == 0, "we should not have added any free regions"); + rebuild_region_sets(false /* free_list_only */); + abort_refinement(); + resize_if_necessary_after_full_collection(); + + // Rebuild the strong code root lists for each region + rebuild_strong_code_roots(); + + // Start a new incremental collection set for the next pause + start_new_collection_set(); + + _allocator->init_mutator_alloc_region(); + + // Post collection state updates. + MetaspaceGC::compute_new_size(); +} + +void G1CollectedHeap::abort_refinement() { if (_hot_card_cache->use_cache()) { _hot_card_cache->reset_card_counts(); _hot_card_cache->reset_hot_cache(); } - // Discard all stale remembered set updates. + // Discard all remembered set updates. JavaThread::dirty_card_queue_set().abandon_logs(); assert(dirty_card_queue_set().completed_buffers_num() == 0, "DCQS should be empty"); } -void G1CollectedHeap::verify_before_full_collection(bool explicit_gc) { - assert(!GCCause::is_user_requested_gc(gc_cause()) || explicit_gc, "invariant"); - assert(used() == recalculate_used(), "Should be equal"); - _verifier->verify_region_sets_optional(); - _verifier->verify_before_gc(); - _verifier->check_bitmaps("Full GC Start"); -} - void G1CollectedHeap::verify_after_full_collection() { check_gc_time_stamps(); _hrm.verify_optional(); @@ -1189,9 +1212,18 @@ ref_processor_cm()->verify_no_references_recorded(); } +void G1CollectedHeap::print_heap_after_full_collection(G1HeapTransition* heap_transition) { + print_hrm_post_compaction(); + heap_transition->print(); + print_heap_after_gc(); + print_heap_regions(); +#ifdef TRACESPINNING + ParallelTaskTerminator::print_termination_counts(); +#endif +} + void G1CollectedHeap::do_full_collection_inner(G1FullGCScope* scope) { GCTraceTime(Info, gc) tm("Pause Full", NULL, gc_cause(), true); - G1HeapTransition heap_transition(this); g1_policy()->record_full_collection_start(); print_heap_before_gc(); @@ -1203,35 +1235,15 @@ gc_prologue(true); prepare_heap_for_full_collection(); - G1SerialCollector serial(scope, ref_processor_stw()); + G1SerialFullCollector serial(scope, ref_processor_stw()); serial.prepare_collection(); serial.collect(); serial.complete_collection(); - assert(num_free_regions() == 0, "we should not have added any free regions"); - MemoryService::track_memory_usage(); + prepare_heap_for_mutators(); - // Delete metaspaces for unloaded class loaders and clean up loader_data graph - ClassLoaderDataGraph::purge(); - MetaspaceAux::verify_metrics(); - - // Prepare heap for normal collections. - rebuild_region_sets(false /* free_list_only */); - reset_card_cache_and_queue(); - resize_if_necessary_after_full_collection(); - - // Rebuild the strong code root lists for each region - rebuild_strong_code_roots(); - - // Start a new incremental collection set for the next pause - start_new_collection_set(); - - _allocator->init_mutator_alloc_region(); - - // Post collection state updates. - MetaspaceGC::compute_new_size(); - gc_epilogue(true); g1_policy()->record_full_collection_end(); + gc_epilogue(true); // Post collection verification. verify_after_full_collection(); @@ -1240,13 +1252,7 @@ // We should do this after we potentially resize the heap so // that all the COMMIT / UNCOMMIT events are generated before // the compaction events. - print_hrm_post_compaction(); - heap_transition.print(); - print_heap_after_gc(); - print_heap_regions(); -#ifdef TRACESPINNING - ParallelTaskTerminator::print_termination_counts(); -#endif + print_heap_after_full_collection(scope->heap_transition()); } bool G1CollectedHeap::do_full_collection(bool explicit_gc, @@ -2607,6 +2613,7 @@ allocation_context_stats().update(full); + MemoryService::track_memory_usage(); // We have just completed a GC. Update the soft reference // policy with the new heap occupancy Universe::update_heap_info_at_gc(); @@ -3157,8 +3164,6 @@ evacuation_info.set_collectionset_used_before(collection_set()->bytes_used_before()); evacuation_info.set_bytes_copied(g1_policy()->bytes_copied_during_gc()); - MemoryService::track_memory_usage(); - // In prepare_for_verify() below we'll need to scan the deferred // update buffers to bring the RSets up-to-date if // G1HRRSFlushLogBuffersOnVerify has been set. While scanning --- old/src/share/vm/gc/g1/g1CollectedHeap.hpp 2017-06-14 16:39:39.143525696 +0200 +++ new/src/share/vm/gc/g1/g1CollectedHeap.hpp 2017-06-14 16:39:39.003520169 +0200 @@ -34,6 +34,7 @@ #include "gc/g1/g1EdenRegions.hpp" #include "gc/g1/g1EvacFailure.hpp" #include "gc/g1/g1EvacStats.hpp" +#include "gc/g1/g1HeapTransition.hpp" #include "gc/g1/g1HeapVerifier.hpp" #include "gc/g1/g1HRPrinter.hpp" #include "gc/g1/g1InCSetState.hpp" @@ -515,10 +516,12 @@ // increase readability. void do_full_collection_inner(G1FullGCScope* scope); void abort_concurrent_cycle(); - void prepare_heap_for_full_collection(); - void reset_card_cache_and_queue(); void verify_before_full_collection(bool explicit_gc); + void prepare_heap_for_full_collection(); + void prepare_heap_for_mutators(); + void abort_refinement(); void verify_after_full_collection(); + void print_heap_after_full_collection(G1HeapTransition* heap_transition); // Helper method for satisfy_failed_allocation() HeapWord* satisfy_failed_allocation_helper(size_t word_size, --- old/src/share/vm/gc/g1/g1FullGCScope.cpp 2017-06-14 16:39:39.851553644 +0200 +++ new/src/share/vm/gc/g1/g1FullGCScope.cpp 2017-06-14 16:39:39.699547644 +0200 @@ -44,7 +44,8 @@ _cpu_time(), _soft_refs(clear_soft, _g1h->collector_policy()), _memory_stats(true, _g1h->gc_cause()), - _collector_stats(_g1h->g1mm()->full_collection_counters()) { + _collector_stats(_g1h->g1mm()->full_collection_counters()), + _heap_transition(_g1h) { assert(_instance == NULL, "Only one scope at a time"); _timer.register_gc_start(); _tracer.report_gc_start(_g1h->gc_cause(), _timer.gc_start()); @@ -81,3 +82,7 @@ SerialOldTracer* G1FullGCScope::tracer() { return &_tracer; } + +G1HeapTransition* G1FullGCScope::heap_transition() { + return &_heap_transition; +} --- old/src/share/vm/gc/g1/g1FullGCScope.hpp 2017-06-14 16:39:40.515579856 +0200 +++ new/src/share/vm/gc/g1/g1FullGCScope.hpp 2017-06-14 16:39:40.383574645 +0200 @@ -26,6 +26,7 @@ #define SHARE_VM_GC_G1_G1FULLGCSCOPE_HPP #include "gc/g1/g1CollectedHeap.hpp" +#include "gc/g1/g1HeapTransition.hpp" #include "gc/shared/collectorCounters.hpp" #include "gc/shared/gcId.hpp" #include "gc/shared/gcTrace.hpp" @@ -36,6 +37,7 @@ #include "memory/allocation.hpp" #include "services/memoryService.hpp" +// Class used to group scoped objects used in the Full GC together. class G1FullGCScope : public StackObj { ResourceMark _rm; bool _explicit_gc; @@ -49,6 +51,7 @@ ClearedAllSoftRefs _soft_refs; TraceCollectorStats _collector_stats; TraceMemoryManagerStats _memory_stats; + G1HeapTransition _heap_transition; // Singleton instance. static G1FullGCScope* _instance; @@ -63,6 +66,7 @@ STWGCTimer* timer(); SerialOldTracer* tracer(); + G1HeapTransition* heap_transition(); }; #endif //SHARE_VM_GC_G1_G1FULLGCSCOPE_HPP --- old/src/share/vm/gc/g1/g1SerialCollector.cpp 2017-06-14 16:39:41.323611752 +0200 +++ /dev/null 2017-06-07 10:58:23.444049309 +0200 @@ -1,141 +0,0 @@ -/* - * Copyright (c) 2017, Oracle and/or its affiliates. All rights reserved. - * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. - * - * This code is free software; you can redistribute it and/or modify it - * under the terms of the GNU General Public License version 2 only, as - * published by the Free Software Foundation. - * - * This code is distributed in the hope that it will be useful, but WITHOUT - * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or - * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License - * version 2 for more details (a copy is included in the LICENSE file that - * accompanied this code). - * - * You should have received a copy of the GNU General Public License version - * 2 along with this work; if not, write to the Free Software Foundation, - * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. - * - * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA - * or visit www.oracle.com if you need additional information or have any - * questions. - * - */ - -#include "precompiled.hpp" -#include "gc/g1/g1CollectedHeap.inline.hpp" -#include "gc/g1/g1FullGCScope.hpp" -#include "gc/g1/g1MarkSweep.hpp" -#include "gc/g1/g1RemSet.inline.hpp" -#include "gc/g1/g1SerialCollector.hpp" -#include "gc/g1/heapRegionRemSet.hpp" -#include "gc/shared/referenceProcessor.hpp" - -G1SerialCollector::G1SerialCollector(G1FullGCScope* scope, - ReferenceProcessor* reference_processor) : - _scope(scope), - _reference_processor(reference_processor), - _is_alive_mutator(_reference_processor, NULL), - _mt_discovery_mutator(_reference_processor, false) { - // Temporarily make discovery by the STW ref processor single threaded (non-MT) - // and clear the STW ref processor's _is_alive_non_header field. -} - -void G1SerialCollector::prepare_collection() { - _reference_processor->enable_discovery(); - _reference_processor->setup_policy(_scope->should_clear_soft_refs()); -} - -void G1SerialCollector::complete_collection() { - // Enqueue any discovered reference objects that have - // not been removed from the discovered lists. - _reference_processor->enqueue_discovered_references(); - - // Iterate the heap and rebuild the remembered sets. - rebuild_remembered_sets(); -} - -void G1SerialCollector::collect() { - // Do the actual collection work. - G1MarkSweep::invoke_at_safepoint(_reference_processor, _scope->should_clear_soft_refs()); -} - -class PostMCRemSetClearClosure: public HeapRegionClosure { - G1CollectedHeap* _g1h; - ModRefBarrierSet* _mr_bs; -public: - PostMCRemSetClearClosure(G1CollectedHeap* g1h, ModRefBarrierSet* mr_bs) : - _g1h(g1h), _mr_bs(mr_bs) {} - - bool doHeapRegion(HeapRegion* r) { - HeapRegionRemSet* hrrs = r->rem_set(); - - _g1h->reset_gc_time_stamps(r); - - if (r->is_continues_humongous()) { - // We'll assert that the strong code root list and RSet is empty - assert(hrrs->strong_code_roots_list_length() == 0, "sanity"); - assert(hrrs->occupied() == 0, "RSet should be empty"); - } else { - hrrs->clear(); - } - // You might think here that we could clear just the cards - // corresponding to the used region. But no: if we leave a dirty card - // in a region we might allocate into, then it would prevent that card - // from being enqueued, and cause it to be missed. - // Re: the performance cost: we shouldn't be doing full GC anyway! - _mr_bs->clear(MemRegion(r->bottom(), r->end())); - - return false; - } -}; - - -class RebuildRSOutOfRegionClosure: public HeapRegionClosure { - G1CollectedHeap* _g1h; - UpdateRSOopClosure _cl; -public: - RebuildRSOutOfRegionClosure(G1CollectedHeap* g1, uint worker_i = 0) : - _cl(g1->g1_rem_set(), worker_i), - _g1h(g1) - { } - - bool doHeapRegion(HeapRegion* r) { - if (!r->is_continues_humongous()) { - _cl.set_from(r); - r->oop_iterate(&_cl); - } - return false; - } -}; - -class ParRebuildRSTask: public AbstractGangTask { - G1CollectedHeap* _g1; - HeapRegionClaimer _hrclaimer; - -public: - ParRebuildRSTask(G1CollectedHeap* g1) : - AbstractGangTask("ParRebuildRSTask"), _g1(g1), _hrclaimer(g1->workers()->active_workers()) {} - - void work(uint worker_id) { - RebuildRSOutOfRegionClosure rebuild_rs(_g1, worker_id); - _g1->heap_region_par_iterate(&rebuild_rs, worker_id, &_hrclaimer); - } -}; - -void G1SerialCollector::rebuild_remembered_sets() { - G1CollectedHeap* g1h = G1CollectedHeap::heap(); - // First clear the stale remembered sets. - PostMCRemSetClearClosure rs_clear(g1h, g1h->g1_barrier_set()); - g1h->heap_region_iterate(&rs_clear); - - // Rebuild remembered sets of all regions. - uint n_workers = AdaptiveSizePolicy::calc_active_workers(g1h->workers()->total_workers(), - g1h->workers()->active_workers(), - Threads::number_of_non_daemon_threads()); - g1h->workers()->update_active_workers(n_workers); - log_info(gc,task)("Using %u workers of %u to rebuild remembered set", n_workers, g1h->workers()->total_workers()); - - ParRebuildRSTask rebuild_rs_task(g1h); - g1h->workers()->run_task(&rebuild_rs_task); -} --- /dev/null 2017-06-07 10:58:23.444049309 +0200 +++ new/src/share/vm/gc/g1/g1SerialFullCollector.cpp 2017-06-14 16:39:41.059601331 +0200 @@ -0,0 +1,141 @@ +/* + * Copyright (c) 2017, Oracle and/or its affiliates. All rights reserved. + * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. + * + * This code is free software; you can redistribute it and/or modify it + * under the terms of the GNU General Public License version 2 only, as + * published by the Free Software Foundation. + * + * This code is distributed in the hope that it will be useful, but WITHOUT + * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or + * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License + * version 2 for more details (a copy is included in the LICENSE file that + * accompanied this code). + * + * You should have received a copy of the GNU General Public License version + * 2 along with this work; if not, write to the Free Software Foundation, + * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. + * + * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA + * or visit www.oracle.com if you need additional information or have any + * questions. + * + */ + +#include "precompiled.hpp" +#include "gc/g1/g1CollectedHeap.inline.hpp" +#include "gc/g1/g1FullGCScope.hpp" +#include "gc/g1/g1MarkSweep.hpp" +#include "gc/g1/g1RemSet.inline.hpp" +#include "gc/g1/g1SerialFullCollector.hpp" +#include "gc/g1/heapRegionRemSet.hpp" +#include "gc/shared/referenceProcessor.hpp" + +G1SerialFullCollector::G1SerialFullCollector(G1FullGCScope* scope, + ReferenceProcessor* reference_processor) : + _scope(scope), + _reference_processor(reference_processor), + _is_alive_mutator(_reference_processor, NULL), + _mt_discovery_mutator(_reference_processor, false) { + // Temporarily make discovery by the STW ref processor single threaded (non-MT) + // and clear the STW ref processor's _is_alive_non_header field. +} + +void G1SerialFullCollector::prepare_collection() { + _reference_processor->enable_discovery(); + _reference_processor->setup_policy(_scope->should_clear_soft_refs()); +} + +void G1SerialFullCollector::complete_collection() { + // Enqueue any discovered reference objects that have + // not been removed from the discovered lists. + _reference_processor->enqueue_discovered_references(); + + // Iterate the heap and rebuild the remembered sets. + rebuild_remembered_sets(); +} + +void G1SerialFullCollector::collect() { + // Do the actual collection work. + G1MarkSweep::invoke_at_safepoint(_reference_processor, _scope->should_clear_soft_refs()); +} + +class PostMCRemSetClearClosure: public HeapRegionClosure { + G1CollectedHeap* _g1h; + ModRefBarrierSet* _mr_bs; +public: + PostMCRemSetClearClosure(G1CollectedHeap* g1h, ModRefBarrierSet* mr_bs) : + _g1h(g1h), _mr_bs(mr_bs) {} + + bool doHeapRegion(HeapRegion* r) { + HeapRegionRemSet* hrrs = r->rem_set(); + + _g1h->reset_gc_time_stamps(r); + + if (r->is_continues_humongous()) { + // We'll assert that the strong code root list and RSet is empty + assert(hrrs->strong_code_roots_list_length() == 0, "sanity"); + assert(hrrs->occupied() == 0, "RSet should be empty"); + } else { + hrrs->clear(); + } + // You might think here that we could clear just the cards + // corresponding to the used region. But no: if we leave a dirty card + // in a region we might allocate into, then it would prevent that card + // from being enqueued, and cause it to be missed. + // Re: the performance cost: we shouldn't be doing full GC anyway! + _mr_bs->clear(MemRegion(r->bottom(), r->end())); + + return false; + } +}; + + +class RebuildRSOutOfRegionClosure: public HeapRegionClosure { + G1CollectedHeap* _g1h; + UpdateRSOopClosure _cl; +public: + RebuildRSOutOfRegionClosure(G1CollectedHeap* g1, uint worker_i = 0) : + _cl(g1->g1_rem_set(), worker_i), + _g1h(g1) + { } + + bool doHeapRegion(HeapRegion* r) { + if (!r->is_continues_humongous()) { + _cl.set_from(r); + r->oop_iterate(&_cl); + } + return false; + } +}; + +class ParRebuildRSTask: public AbstractGangTask { + G1CollectedHeap* _g1; + HeapRegionClaimer _hrclaimer; + +public: + ParRebuildRSTask(G1CollectedHeap* g1) : + AbstractGangTask("ParRebuildRSTask"), _g1(g1), _hrclaimer(g1->workers()->active_workers()) {} + + void work(uint worker_id) { + RebuildRSOutOfRegionClosure rebuild_rs(_g1, worker_id); + _g1->heap_region_par_iterate(&rebuild_rs, worker_id, &_hrclaimer); + } +}; + +void G1SerialFullCollector::rebuild_remembered_sets() { + G1CollectedHeap* g1h = G1CollectedHeap::heap(); + // First clear the stale remembered sets. + PostMCRemSetClearClosure rs_clear(g1h, g1h->g1_barrier_set()); + g1h->heap_region_iterate(&rs_clear); + + // Rebuild remembered sets of all regions. + uint n_workers = AdaptiveSizePolicy::calc_active_workers(g1h->workers()->total_workers(), + g1h->workers()->active_workers(), + Threads::number_of_non_daemon_threads()); + g1h->workers()->update_active_workers(n_workers); + log_info(gc,task)("Using %u workers of %u to rebuild remembered set", n_workers, g1h->workers()->total_workers()); + + ParRebuildRSTask rebuild_rs_task(g1h); + g1h->workers()->run_task(&rebuild_rs_task); +} --- old/src/share/vm/gc/g1/g1SerialCollector.hpp 2017-06-14 16:39:42.159644754 +0200 +++ /dev/null 2017-06-07 10:58:23.444049309 +0200 @@ -1,49 +0,0 @@ -/* - * Copyright (c) 2017, Oracle and/or its affiliates. All rights reserved. - * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. - * - * This code is free software; you can redistribute it and/or modify it - * under the terms of the GNU General Public License version 2 only, as - * published by the Free Software Foundation. - * - * This code is distributed in the hope that it will be useful, but WITHOUT - * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or - * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License - * version 2 for more details (a copy is included in the LICENSE file that - * accompanied this code). - * - * You should have received a copy of the GNU General Public License version - * 2 along with this work; if not, write to the Free Software Foundation, - * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. - * - * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA - * or visit www.oracle.com if you need additional information or have any - * questions. - * - */ - -#ifndef SHARE_VM_GC_G1_G1SERIALCOLLECTOR_HPP -#define SHARE_VM_GC_G1_G1SERIALCOLLECTOR_HPP - -#include "memory/allocation.hpp" - -class G1FullGCScope; -class ReferenceProcessor; - -class G1SerialCollector : StackObj { - G1FullGCScope* _scope; - ReferenceProcessor* _reference_processor; - ReferenceProcessorIsAliveMutator _is_alive_mutator; - ReferenceProcessorMTDiscoveryMutator _mt_discovery_mutator; - - void rebuild_remembered_sets(); - -public: - G1SerialCollector(G1FullGCScope* scope, ReferenceProcessor* reference_processor); - - void prepare_collection(); - void collect(); - void complete_collection(); -}; - -#endif // SHARE_VM_GC_G1_G1SERIALCOLLECTOR_HPP --- /dev/null 2017-06-07 10:58:23.444049309 +0200 +++ new/src/share/vm/gc/g1/g1SerialFullCollector.hpp 2017-06-14 16:39:41.899634490 +0200 @@ -0,0 +1,49 @@ +/* + * Copyright (c) 2017, Oracle and/or its affiliates. All rights reserved. + * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. + * + * This code is free software; you can redistribute it and/or modify it + * under the terms of the GNU General Public License version 2 only, as + * published by the Free Software Foundation. + * + * This code is distributed in the hope that it will be useful, but WITHOUT + * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or + * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License + * version 2 for more details (a copy is included in the LICENSE file that + * accompanied this code). + * + * You should have received a copy of the GNU General Public License version + * 2 along with this work; if not, write to the Free Software Foundation, + * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. + * + * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA + * or visit www.oracle.com if you need additional information or have any + * questions. + * + */ + +#ifndef SHARE_VM_GC_G1_G1SERIALCOLLECTOR_HPP +#define SHARE_VM_GC_G1_G1SERIALCOLLECTOR_HPP + +#include "memory/allocation.hpp" + +class G1FullGCScope; +class ReferenceProcessor; + +class G1SerialFullCollector : StackObj { + G1FullGCScope* _scope; + ReferenceProcessor* _reference_processor; + ReferenceProcessorIsAliveMutator _is_alive_mutator; + ReferenceProcessorMTDiscoveryMutator _mt_discovery_mutator; + + void rebuild_remembered_sets(); + +public: + G1SerialFullCollector(G1FullGCScope* scope, ReferenceProcessor* reference_processor); + + void prepare_collection(); + void collect(); + void complete_collection(); +}; + +#endif // SHARE_VM_GC_G1_G1SERIALCOLLECTOR_HPP