1 /* 2 * Copyright (c) 2012, 2015, Oracle and/or its affiliates. All rights reserved. 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 4 * 5 * This code is free software; you can redistribute it and/or modify it 6 * under the terms of the GNU General Public License version 2 only, as 7 * published by the Free Software Foundation. 8 * 9 * This code is distributed in the hope that it will be useful, but WITHOUT 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 12 * version 2 for more details (a copy is included in the LICENSE file that 13 * accompanied this code). 14 * 15 * You should have received a copy of the GNU General Public License version 16 * 2 along with this work; if not, write to the Free Software Foundation, 17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 18 * 19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 20 * or visit www.oracle.com if you need additional information or have any 21 * questions. 22 * 23 */ 24 25 #include "precompiled.hpp" 26 #include "gc/shared/copyFailedInfo.hpp" 27 #include "gc/shared/gcHeapSummary.hpp" 28 #include "gc/shared/gcTimer.hpp" 29 #include "gc/shared/gcTrace.hpp" 30 #include "gc/shared/gcWhen.hpp" 31 #include "runtime/os.hpp" 32 #include "trace/traceBackend.hpp" 33 #include "trace/tracing.hpp" 34 #include "utilities/macros.hpp" 35 #if INCLUDE_ALL_GCS 36 #include "gc/g1/evacuationInfo.hpp" 37 #include "gc/g1/g1YCTypes.hpp" 38 #include "tracefiles/traceEventClasses.hpp" 39 #endif 40 41 // All GC dependencies against the trace framework is contained within this file. 42 43 typedef uintptr_t TraceAddress; 44 45 void GCTracer::send_garbage_collection_event() const { 46 EventGCGarbageCollection event(UNTIMED); 47 if (event.should_commit()) { 48 event.set_gcId(GCId::current()); 49 event.set_name(_shared_gc_info.name()); 50 event.set_cause((u2) _shared_gc_info.cause()); 51 event.set_sumOfPauses(_shared_gc_info.sum_of_pauses()); 52 event.set_longestPause(_shared_gc_info.longest_pause()); 53 event.set_starttime(_shared_gc_info.start_timestamp()); 54 event.set_endtime(_shared_gc_info.end_timestamp()); 55 event.commit(); 56 } 57 } 58 59 void GCTracer::send_reference_stats_event(ReferenceType type, size_t count) const { 60 EventGCReferenceStatistics e; 61 if (e.should_commit()) { 62 e.set_gcId(GCId::current()); 63 e.set_type((u1)type); 64 e.set_count(count); 65 e.commit(); 66 } 67 } 68 69 void GCTracer::send_metaspace_chunk_free_list_summary(GCWhen::Type when, Metaspace::MetadataType mdtype, 70 const MetaspaceChunkFreeListSummary& summary) const { 71 EventMetaspaceChunkFreeListSummary e; 72 if (e.should_commit()) { 73 e.set_gcId(GCId::current()); 74 e.set_when(when); 75 e.set_metadataType(mdtype); 76 77 e.set_specializedChunks(summary.num_specialized_chunks()); 78 e.set_specializedChunksTotalSize(summary.specialized_chunks_size_in_bytes()); 79 80 e.set_smallChunks(summary.num_small_chunks()); 81 e.set_smallChunksTotalSize(summary.small_chunks_size_in_bytes()); 82 83 e.set_mediumChunks(summary.num_medium_chunks()); 84 e.set_mediumChunksTotalSize(summary.medium_chunks_size_in_bytes()); 85 86 e.set_humongousChunks(summary.num_humongous_chunks()); 87 e.set_humongousChunksTotalSize(summary.humongous_chunks_size_in_bytes()); 88 89 e.commit(); 90 } 91 } 92 93 void ParallelOldTracer::send_parallel_old_event() const { 94 EventGCParallelOld e(UNTIMED); 95 if (e.should_commit()) { 96 e.set_gcId(GCId::current()); 97 e.set_densePrefix((TraceAddress)_parallel_old_gc_info.dense_prefix()); 98 e.set_starttime(_shared_gc_info.start_timestamp()); 99 e.set_endtime(_shared_gc_info.end_timestamp()); 100 e.commit(); 101 } 102 } 103 104 void YoungGCTracer::send_young_gc_event() const { 105 EventGCYoungGarbageCollection e(UNTIMED); 106 if (e.should_commit()) { 107 e.set_gcId(GCId::current()); 108 e.set_tenuringThreshold(_tenuring_threshold); 109 e.set_starttime(_shared_gc_info.start_timestamp()); 110 e.set_endtime(_shared_gc_info.end_timestamp()); 111 e.commit(); 112 } 113 } 114 115 bool YoungGCTracer::should_send_promotion_in_new_plab_event() const { 116 return EventPromoteObjectInNewPLAB::is_enabled(); 117 } 118 119 bool YoungGCTracer::should_send_promotion_outside_plab_event() const { 120 return EventPromoteObjectOutsidePLAB::is_enabled(); 121 } 122 123 void YoungGCTracer::send_promotion_in_new_plab_event(Klass* klass, size_t obj_size, 124 uint age, bool tenured, 125 size_t plab_size) const { 126 127 EventPromoteObjectInNewPLAB event; 128 if (event.should_commit()) { 129 event.set_gcId(GCId::current()); 130 event.set_class(klass); 131 event.set_objectSize(obj_size); 132 event.set_tenured(tenured); 133 event.set_tenuringAge(age); 134 event.set_plabSize(plab_size); 135 event.commit(); 136 } 137 } 138 139 void YoungGCTracer::send_promotion_outside_plab_event(Klass* klass, size_t obj_size, 140 uint age, bool tenured) const { 141 142 EventPromoteObjectOutsidePLAB event; 143 if (event.should_commit()) { 144 event.set_gcId(GCId::current()); 145 event.set_class(klass); 146 event.set_objectSize(obj_size); 147 event.set_tenured(tenured); 148 event.set_tenuringAge(age); 149 event.commit(); 150 } 151 } 152 153 void OldGCTracer::send_old_gc_event() const { 154 EventGCOldGarbageCollection e(UNTIMED); 155 if (e.should_commit()) { 156 e.set_gcId(GCId::current()); 157 e.set_starttime(_shared_gc_info.start_timestamp()); 158 e.set_endtime(_shared_gc_info.end_timestamp()); 159 e.commit(); 160 } 161 } 162 163 static TraceStructCopyFailed to_trace_struct(const CopyFailedInfo& cf_info) { 164 TraceStructCopyFailed failed_info; 165 failed_info.set_objectCount(cf_info.failed_count()); 166 failed_info.set_firstSize(cf_info.first_size()); 167 failed_info.set_smallestSize(cf_info.smallest_size()); 168 failed_info.set_totalSize(cf_info.total_size()); 169 return failed_info; 170 } 171 172 void YoungGCTracer::send_promotion_failed_event(const PromotionFailedInfo& pf_info) const { 173 EventPromotionFailed e; 174 if (e.should_commit()) { 175 e.set_gcId(GCId::current()); 176 e.set_data(to_trace_struct(pf_info)); 177 e.set_thread(pf_info.thread()->thread_id()); 178 e.commit(); 179 } 180 } 181 182 // Common to CMS and G1 183 void OldGCTracer::send_concurrent_mode_failure_event() { 184 EventConcurrentModeFailure e; 185 if (e.should_commit()) { 186 e.set_gcId(GCId::current()); 187 e.commit(); 188 } 189 } 190 191 #if INCLUDE_ALL_GCS 192 void G1NewTracer::send_g1_young_gc_event() { 193 EventGCG1GarbageCollection e(UNTIMED); 194 if (e.should_commit()) { 195 e.set_gcId(GCId::current()); 196 e.set_type(_g1_young_gc_info.type()); 197 e.set_starttime(_shared_gc_info.start_timestamp()); 198 e.set_endtime(_shared_gc_info.end_timestamp()); 199 e.commit(); 200 } 201 } 202 203 void G1MMUTracer::send_g1_mmu_event(double timeSlice, double gcTime, double maxTime) { 204 EventGCG1MMU e; 205 if (e.should_commit()) { 206 e.set_gcId(GCId::current()); 207 e.set_timeSlice(timeSlice); 208 e.set_gcTime(gcTime); 209 e.set_maxGcTime(maxTime); 210 e.commit(); 211 } 212 } 213 214 void G1NewTracer::send_evacuation_info_event(EvacuationInfo* info) { 215 EventEvacuationInfo e; 216 if (e.should_commit()) { 217 e.set_gcId(GCId::current()); 218 e.set_cSetRegions(info->collectionset_regions()); 219 e.set_cSetUsedBefore(info->collectionset_used_before()); 220 e.set_cSetUsedAfter(info->collectionset_used_after()); 221 e.set_allocationRegions(info->allocation_regions()); 222 e.set_allocRegionsUsedBefore(info->alloc_regions_used_before()); 223 e.set_allocRegionsUsedAfter(info->alloc_regions_used_before() + info->bytes_copied()); 224 e.set_bytesCopied(info->bytes_copied()); 225 e.set_regionsFreed(info->regions_freed()); 226 e.commit(); 227 } 228 } 229 230 void G1NewTracer::send_evacuation_failed_event(const EvacuationFailedInfo& ef_info) const { 231 EventEvacuationFailed e; 232 if (e.should_commit()) { 233 e.set_gcId(GCId::current()); 234 e.set_data(to_trace_struct(ef_info)); 235 e.commit(); 236 } 237 } 238 239 static TraceStructG1EvacStats create_g1_evacstats(unsigned gcid, const G1EvacSummary& summary) { 240 TraceStructG1EvacStats s; 241 s.set_gcId(gcid); 242 s.set_allocated(summary.allocated() * HeapWordSize); 243 s.set_wasted(summary.wasted() * HeapWordSize); 244 s.set_used(summary.used() * HeapWordSize); 245 s.set_undoWaste(summary.undo_wasted() * HeapWordSize); 246 s.set_regionEndWaste(summary.region_end_waste() * HeapWordSize); 247 s.set_regionsRefilled(summary.regions_filled()); 248 s.set_directAllocated(summary.direct_allocated() * HeapWordSize); 249 s.set_failureUsed(summary.failure_used() * HeapWordSize); 250 s.set_failureWaste(summary.failure_waste() * HeapWordSize); 251 return s; 252 } 253 254 void G1NewTracer::send_young_evacuation_statistics(const G1EvacSummary& summary) const { 255 EventGCG1EvacuationYoungStatistics surv_evt; 256 if (surv_evt.should_commit()) { 257 surv_evt.set_stats(create_g1_evacstats(GCId::current(), summary)); 258 surv_evt.commit(); 259 } 260 } 261 262 void G1NewTracer::send_old_evacuation_statistics(const G1EvacSummary& summary) const { 263 EventGCG1EvacuationOldStatistics old_evt; 264 if (old_evt.should_commit()) { 265 old_evt.set_stats(create_g1_evacstats(GCId::current(), summary)); 266 old_evt.commit(); 267 } 268 } 269 270 void G1NewTracer::send_basic_ihop_statistics(size_t threshold, 271 size_t target_occupancy, 272 size_t current_occupancy, 273 size_t last_allocation_ize, 274 double last_allocation_duration, 275 double last_marking_length) { 276 EventGCG1BasicIHOP evt; 277 if (evt.should_commit()) { 278 evt.set_gcId(GCId::current()); 279 evt.set_threshold(threshold); 280 evt.set_targetOccupancy(target_occupancy); 281 evt.set_thresholdPercentage(threshold * 100.0 / target_occupancy); 282 evt.set_currentOccupancy(current_occupancy); 283 evt.set_lastAllocationSize(last_allocation_ize); 284 evt.set_lastAllocationDuration(last_allocation_duration); 285 evt.set_lastAllocationRate(last_allocation_duration != 0.0 ? last_allocation_ize / last_allocation_duration : 0.0); 286 evt.set_lastMarkingLength(last_marking_length); 287 evt.commit(); 288 } 289 } 290 291 void G1NewTracer::send_adaptive_ihop_statistics(size_t additional_buffer_size, 292 double predicted_allocation_rate, 293 double predicted_marking_length, 294 bool prediction_active) { 295 EventGCG1AdaptiveIHOP evt; 296 if (evt.should_commit()) { 297 evt.set_gcId(GCId::current()); 298 evt.set_additionalBufferSize(additional_buffer_size); 299 evt.set_predictedAllocationRate(predicted_allocation_rate); 300 evt.set_predictedMarkingLength(predicted_marking_length); 301 evt.set_predictionActive(prediction_active); 302 evt.commit(); 303 } 304 } 305 306 #endif 307 308 static TraceStructVirtualSpace to_trace_struct(const VirtualSpaceSummary& summary) { 309 TraceStructVirtualSpace space; 310 space.set_start((TraceAddress)summary.start()); 311 space.set_committedEnd((TraceAddress)summary.committed_end()); 312 space.set_committedSize(summary.committed_size()); 313 space.set_reservedEnd((TraceAddress)summary.reserved_end()); 314 space.set_reservedSize(summary.reserved_size()); 315 return space; 316 } 317 318 static TraceStructObjectSpace to_trace_struct(const SpaceSummary& summary) { 319 TraceStructObjectSpace space; 320 space.set_start((TraceAddress)summary.start()); 321 space.set_end((TraceAddress)summary.end()); 322 space.set_used(summary.used()); 323 space.set_size(summary.size()); 324 return space; 325 } 326 327 class GCHeapSummaryEventSender : public GCHeapSummaryVisitor { 328 GCWhen::Type _when; 329 public: 330 GCHeapSummaryEventSender(GCWhen::Type when) : _when(when) {} 331 332 void visit(const GCHeapSummary* heap_summary) const { 333 const VirtualSpaceSummary& heap_space = heap_summary->heap(); 334 335 EventGCHeapSummary e; 336 if (e.should_commit()) { 337 e.set_gcId(GCId::current()); 338 e.set_when((u1)_when); 339 e.set_heapSpace(to_trace_struct(heap_space)); 340 e.set_heapUsed(heap_summary->used()); 341 e.commit(); 342 } 343 } 344 345 void visit(const G1HeapSummary* g1_heap_summary) const { 346 visit((GCHeapSummary*)g1_heap_summary); 347 348 EventG1HeapSummary e; 349 if (e.should_commit()) { 350 e.set_gcId(GCId::current()); 351 e.set_when((u1)_when); 352 e.set_edenUsedSize(g1_heap_summary->edenUsed()); 353 e.set_edenTotalSize(g1_heap_summary->edenCapacity()); 354 e.set_survivorUsedSize(g1_heap_summary->survivorUsed()); 355 e.commit(); 356 } 357 } 358 359 void visit(const PSHeapSummary* ps_heap_summary) const { 360 visit((GCHeapSummary*)ps_heap_summary); 361 362 const VirtualSpaceSummary& old_summary = ps_heap_summary->old(); 363 const SpaceSummary& old_space = ps_heap_summary->old_space(); 364 const VirtualSpaceSummary& young_summary = ps_heap_summary->young(); 365 const SpaceSummary& eden_space = ps_heap_summary->eden(); 366 const SpaceSummary& from_space = ps_heap_summary->from(); 367 const SpaceSummary& to_space = ps_heap_summary->to(); 368 369 EventPSHeapSummary e; 370 if (e.should_commit()) { 371 e.set_gcId(GCId::current()); 372 e.set_when((u1)_when); 373 374 e.set_oldSpace(to_trace_struct(ps_heap_summary->old())); 375 e.set_oldObjectSpace(to_trace_struct(ps_heap_summary->old_space())); 376 e.set_youngSpace(to_trace_struct(ps_heap_summary->young())); 377 e.set_edenSpace(to_trace_struct(ps_heap_summary->eden())); 378 e.set_fromSpace(to_trace_struct(ps_heap_summary->from())); 379 e.set_toSpace(to_trace_struct(ps_heap_summary->to())); 380 e.commit(); 381 } 382 } 383 }; 384 385 void GCTracer::send_gc_heap_summary_event(GCWhen::Type when, const GCHeapSummary& heap_summary) const { 386 GCHeapSummaryEventSender visitor(when); 387 heap_summary.accept(&visitor); 388 } 389 390 static TraceStructMetaspaceSizes to_trace_struct(const MetaspaceSizes& sizes) { 391 TraceStructMetaspaceSizes meta_sizes; 392 393 meta_sizes.set_committed(sizes.committed()); 394 meta_sizes.set_used(sizes.used()); 395 meta_sizes.set_reserved(sizes.reserved()); 396 397 return meta_sizes; 398 } 399 400 void GCTracer::send_meta_space_summary_event(GCWhen::Type when, const MetaspaceSummary& meta_space_summary) const { 401 EventMetaspaceSummary e; 402 if (e.should_commit()) { 403 e.set_gcId(GCId::current()); 404 e.set_when((u1) when); 405 e.set_gcThreshold(meta_space_summary.capacity_until_GC()); 406 e.set_metaspace(to_trace_struct(meta_space_summary.meta_space())); 407 e.set_dataSpace(to_trace_struct(meta_space_summary.data_space())); 408 e.set_classSpace(to_trace_struct(meta_space_summary.class_space())); 409 e.commit(); 410 } 411 } 412 413 class PhaseSender : public PhaseVisitor { 414 public: 415 template<typename T> 416 void send_phase(PausePhase* pause) { 417 T event(UNTIMED); 418 if (event.should_commit()) { 419 event.set_gcId(GCId::current()); 420 event.set_name(pause->name()); 421 event.set_starttime(pause->start()); 422 event.set_endtime(pause->end()); 423 event.commit(); 424 } 425 } 426 427 void visit(GCPhase* pause) { ShouldNotReachHere(); } 428 void visit(ConcurrentPhase* pause) { Unimplemented(); } 429 void visit(PausePhase* pause) { 430 assert(PhasesStack::PHASE_LEVELS == 5, "Need more event types"); 431 432 switch (pause->level()) { 433 case 0: send_phase<EventGCPhasePause>(pause); break; 434 case 1: send_phase<EventGCPhasePauseLevel1>(pause); break; 435 case 2: send_phase<EventGCPhasePauseLevel2>(pause); break; 436 case 3: send_phase<EventGCPhasePauseLevel3>(pause); break; 437 default: /* Ignore sending this phase */ break; 438 } 439 } 440 }; 441 442 void GCTracer::send_phase_events(TimePartitions* time_partitions) const { 443 PhaseSender phase_reporter; 444 445 TimePartitionPhasesIterator iter(time_partitions); 446 while (iter.has_next()) { 447 GCPhase* phase = iter.next(); 448 phase->accept(&phase_reporter); 449 } 450 }