Tweak compaction candidate selection to avoid keeping page with low occupancy around.
Increase slots buffer chain length to 15 to make compaction more aggressive and usefull. Pass gc and collector selection reasons to GCTracer to allow more meaningull --gc-trace. Print fragmentation of spaces that we do not compact. R=erik.corry@gmail.com Review URL: https://chromiumcodereview.appspot.com/9323007 git-svn-id: http://v8.googlecode.com/svn/branches/bleeding_edge@10601 ce2b1a6d-e550-0410-aec6-3dcde31c8c00
This commit is contained in:
parent
b34e202b20
commit
a7b0481b6d
@ -4077,7 +4077,7 @@ bool v8::V8::IdleNotification(int hint) {
|
|||||||
void v8::V8::LowMemoryNotification() {
|
void v8::V8::LowMemoryNotification() {
|
||||||
i::Isolate* isolate = i::Isolate::Current();
|
i::Isolate* isolate = i::Isolate::Current();
|
||||||
if (isolate == NULL || !isolate->IsInitialized()) return;
|
if (isolate == NULL || !isolate->IsInitialized()) return;
|
||||||
isolate->heap()->CollectAllAvailableGarbage();
|
isolate->heap()->CollectAllAvailableGarbage("low memory notification");
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
11
src/debug.cc
11
src/debug.cc
@ -1904,7 +1904,8 @@ void Debug::PrepareForBreakPoints() {
|
|||||||
{
|
{
|
||||||
// We are going to iterate heap to find all functions without
|
// We are going to iterate heap to find all functions without
|
||||||
// debug break slots.
|
// debug break slots.
|
||||||
isolate_->heap()->CollectAllGarbage(Heap::kMakeHeapIterableMask);
|
isolate_->heap()->CollectAllGarbage(Heap::kMakeHeapIterableMask,
|
||||||
|
"preparing for breakpoints");
|
||||||
|
|
||||||
// Ensure no GC in this scope as we are going to use gc_metadata
|
// Ensure no GC in this scope as we are going to use gc_metadata
|
||||||
// field in the Code object to mark active functions.
|
// field in the Code object to mark active functions.
|
||||||
@ -2230,8 +2231,9 @@ void Debug::CreateScriptCache() {
|
|||||||
// rid of all the cached script wrappers and the second gets rid of the
|
// rid of all the cached script wrappers and the second gets rid of the
|
||||||
// scripts which are no longer referenced. The second also sweeps precisely,
|
// scripts which are no longer referenced. The second also sweeps precisely,
|
||||||
// which saves us doing yet another GC to make the heap iterable.
|
// which saves us doing yet another GC to make the heap iterable.
|
||||||
heap->CollectAllGarbage(Heap::kNoGCFlags);
|
heap->CollectAllGarbage(Heap::kNoGCFlags, "Debug::CreateScriptCache");
|
||||||
heap->CollectAllGarbage(Heap::kMakeHeapIterableMask);
|
heap->CollectAllGarbage(Heap::kMakeHeapIterableMask,
|
||||||
|
"Debug::CreateScriptCache");
|
||||||
|
|
||||||
ASSERT(script_cache_ == NULL);
|
ASSERT(script_cache_ == NULL);
|
||||||
script_cache_ = new ScriptCache();
|
script_cache_ = new ScriptCache();
|
||||||
@ -2281,7 +2283,8 @@ Handle<FixedArray> Debug::GetLoadedScripts() {
|
|||||||
|
|
||||||
// Perform GC to get unreferenced scripts evicted from the cache before
|
// Perform GC to get unreferenced scripts evicted from the cache before
|
||||||
// returning the content.
|
// returning the content.
|
||||||
isolate_->heap()->CollectAllGarbage(Heap::kNoGCFlags);
|
isolate_->heap()->CollectAllGarbage(Heap::kNoGCFlags,
|
||||||
|
"Debug::GetLoadedScripts");
|
||||||
|
|
||||||
// Get the scripts from the cache.
|
// Get the scripts from the cache.
|
||||||
return script_cache_->GetScripts();
|
return script_cache_->GetScripts();
|
||||||
|
@ -877,7 +877,7 @@ MaybeObject* Execution::HandleStackGuardInterrupt() {
|
|||||||
StackGuard* stack_guard = isolate->stack_guard();
|
StackGuard* stack_guard = isolate->stack_guard();
|
||||||
|
|
||||||
if (stack_guard->IsGCRequest()) {
|
if (stack_guard->IsGCRequest()) {
|
||||||
isolate->heap()->CollectAllGarbage(false);
|
isolate->heap()->CollectAllGarbage(false, "StackGuard GC request");
|
||||||
stack_guard->Continue(GC_REQUEST);
|
stack_guard->Continue(GC_REQUEST);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -40,7 +40,7 @@ v8::Handle<v8::FunctionTemplate> GCExtension::GetNativeFunction(
|
|||||||
|
|
||||||
|
|
||||||
v8::Handle<v8::Value> GCExtension::GC(const v8::Arguments& args) {
|
v8::Handle<v8::Value> GCExtension::GC(const v8::Arguments& args) {
|
||||||
HEAP->CollectAllGarbage(Heap::kNoGCFlags);
|
HEAP->CollectAllGarbage(Heap::kNoGCFlags, "gc extension");
|
||||||
return v8::Undefined();
|
return v8::Undefined();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -438,8 +438,10 @@ void Heap::ScavengeObject(HeapObject** p, HeapObject* object) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
bool Heap::CollectGarbage(AllocationSpace space) {
|
bool Heap::CollectGarbage(AllocationSpace space, const char* gc_reason) {
|
||||||
return CollectGarbage(space, SelectGarbageCollector(space));
|
const char* collector_reason = NULL;
|
||||||
|
GarbageCollector collector = SelectGarbageCollector(space, &collector_reason);
|
||||||
|
return CollectGarbage(space, collector, gc_reason, collector_reason);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
@ -474,7 +476,7 @@ int Heap::AdjustAmountOfExternalAllocatedMemory(int change_in_bytes) {
|
|||||||
amount_of_external_allocated_memory_ -
|
amount_of_external_allocated_memory_ -
|
||||||
amount_of_external_allocated_memory_at_last_global_gc_;
|
amount_of_external_allocated_memory_at_last_global_gc_;
|
||||||
if (amount_since_last_global_gc > external_allocation_limit_) {
|
if (amount_since_last_global_gc > external_allocation_limit_) {
|
||||||
CollectAllGarbage(kNoGCFlags);
|
CollectAllGarbage(kNoGCFlags, "external memory allocation limit reached");
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
// Avoid underflow.
|
// Avoid underflow.
|
||||||
@ -523,7 +525,8 @@ Isolate* Heap::isolate() {
|
|||||||
} \
|
} \
|
||||||
if (!__maybe_object__->IsRetryAfterGC()) RETURN_EMPTY; \
|
if (!__maybe_object__->IsRetryAfterGC()) RETURN_EMPTY; \
|
||||||
ISOLATE->heap()->CollectGarbage(Failure::cast(__maybe_object__)-> \
|
ISOLATE->heap()->CollectGarbage(Failure::cast(__maybe_object__)-> \
|
||||||
allocation_space()); \
|
allocation_space(), \
|
||||||
|
"allocation failure"); \
|
||||||
__maybe_object__ = FUNCTION_CALL; \
|
__maybe_object__ = FUNCTION_CALL; \
|
||||||
if (__maybe_object__->ToObject(&__object__)) RETURN_VALUE; \
|
if (__maybe_object__->ToObject(&__object__)) RETURN_VALUE; \
|
||||||
if (__maybe_object__->IsOutOfMemory()) { \
|
if (__maybe_object__->IsOutOfMemory()) { \
|
||||||
@ -531,7 +534,7 @@ Isolate* Heap::isolate() {
|
|||||||
} \
|
} \
|
||||||
if (!__maybe_object__->IsRetryAfterGC()) RETURN_EMPTY; \
|
if (!__maybe_object__->IsRetryAfterGC()) RETURN_EMPTY; \
|
||||||
ISOLATE->counters()->gc_last_resort_from_handles()->Increment(); \
|
ISOLATE->counters()->gc_last_resort_from_handles()->Increment(); \
|
||||||
ISOLATE->heap()->CollectAllAvailableGarbage(); \
|
ISOLATE->heap()->CollectAllAvailableGarbage("last resort gc"); \
|
||||||
{ \
|
{ \
|
||||||
AlwaysAllocateScope __scope__; \
|
AlwaysAllocateScope __scope__; \
|
||||||
__maybe_object__ = FUNCTION_CALL; \
|
__maybe_object__ = FUNCTION_CALL; \
|
||||||
|
83
src/heap.cc
83
src/heap.cc
@ -236,16 +236,19 @@ int Heap::GcSafeSizeOfOldObject(HeapObject* object) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
GarbageCollector Heap::SelectGarbageCollector(AllocationSpace space) {
|
GarbageCollector Heap::SelectGarbageCollector(AllocationSpace space,
|
||||||
|
const char** reason) {
|
||||||
// Is global GC requested?
|
// Is global GC requested?
|
||||||
if (space != NEW_SPACE || FLAG_gc_global) {
|
if (space != NEW_SPACE || FLAG_gc_global) {
|
||||||
isolate_->counters()->gc_compactor_caused_by_request()->Increment();
|
isolate_->counters()->gc_compactor_caused_by_request()->Increment();
|
||||||
|
*reason = "GC in old space requested";
|
||||||
return MARK_COMPACTOR;
|
return MARK_COMPACTOR;
|
||||||
}
|
}
|
||||||
|
|
||||||
// Is enough data promoted to justify a global GC?
|
// Is enough data promoted to justify a global GC?
|
||||||
if (OldGenerationPromotionLimitReached()) {
|
if (OldGenerationPromotionLimitReached()) {
|
||||||
isolate_->counters()->gc_compactor_caused_by_promoted_data()->Increment();
|
isolate_->counters()->gc_compactor_caused_by_promoted_data()->Increment();
|
||||||
|
*reason = "promotion limit reached";
|
||||||
return MARK_COMPACTOR;
|
return MARK_COMPACTOR;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -253,6 +256,7 @@ GarbageCollector Heap::SelectGarbageCollector(AllocationSpace space) {
|
|||||||
if (old_gen_exhausted_) {
|
if (old_gen_exhausted_) {
|
||||||
isolate_->counters()->
|
isolate_->counters()->
|
||||||
gc_compactor_caused_by_oldspace_exhaustion()->Increment();
|
gc_compactor_caused_by_oldspace_exhaustion()->Increment();
|
||||||
|
*reason = "old generations exhausted";
|
||||||
return MARK_COMPACTOR;
|
return MARK_COMPACTOR;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -268,10 +272,12 @@ GarbageCollector Heap::SelectGarbageCollector(AllocationSpace space) {
|
|||||||
if (isolate_->memory_allocator()->MaxAvailable() <= new_space_.Size()) {
|
if (isolate_->memory_allocator()->MaxAvailable() <= new_space_.Size()) {
|
||||||
isolate_->counters()->
|
isolate_->counters()->
|
||||||
gc_compactor_caused_by_oldspace_exhaustion()->Increment();
|
gc_compactor_caused_by_oldspace_exhaustion()->Increment();
|
||||||
|
*reason = "scavenge might not succeed";
|
||||||
return MARK_COMPACTOR;
|
return MARK_COMPACTOR;
|
||||||
}
|
}
|
||||||
|
|
||||||
// Default
|
// Default
|
||||||
|
*reason = NULL;
|
||||||
return SCAVENGER;
|
return SCAVENGER;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -431,17 +437,17 @@ void Heap::GarbageCollectionEpilogue() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
void Heap::CollectAllGarbage(int flags) {
|
void Heap::CollectAllGarbage(int flags, const char* gc_reason) {
|
||||||
// Since we are ignoring the return value, the exact choice of space does
|
// Since we are ignoring the return value, the exact choice of space does
|
||||||
// not matter, so long as we do not specify NEW_SPACE, which would not
|
// not matter, so long as we do not specify NEW_SPACE, which would not
|
||||||
// cause a full GC.
|
// cause a full GC.
|
||||||
mark_compact_collector_.SetFlags(flags);
|
mark_compact_collector_.SetFlags(flags);
|
||||||
CollectGarbage(OLD_POINTER_SPACE);
|
CollectGarbage(OLD_POINTER_SPACE, gc_reason);
|
||||||
mark_compact_collector_.SetFlags(kNoGCFlags);
|
mark_compact_collector_.SetFlags(kNoGCFlags);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
void Heap::CollectAllAvailableGarbage() {
|
void Heap::CollectAllAvailableGarbage(const char* gc_reason) {
|
||||||
// Since we are ignoring the return value, the exact choice of space does
|
// Since we are ignoring the return value, the exact choice of space does
|
||||||
// not matter, so long as we do not specify NEW_SPACE, which would not
|
// not matter, so long as we do not specify NEW_SPACE, which would not
|
||||||
// cause a full GC.
|
// cause a full GC.
|
||||||
@ -453,11 +459,12 @@ void Heap::CollectAllAvailableGarbage() {
|
|||||||
// Note: as weak callbacks can execute arbitrary code, we cannot
|
// Note: as weak callbacks can execute arbitrary code, we cannot
|
||||||
// hope that eventually there will be no weak callbacks invocations.
|
// hope that eventually there will be no weak callbacks invocations.
|
||||||
// Therefore stop recollecting after several attempts.
|
// Therefore stop recollecting after several attempts.
|
||||||
mark_compact_collector()->SetFlags(kMakeHeapIterableMask);
|
mark_compact_collector()->SetFlags(kMakeHeapIterableMask |
|
||||||
|
kReduceMemoryFootprintMask);
|
||||||
isolate_->compilation_cache()->Clear();
|
isolate_->compilation_cache()->Clear();
|
||||||
const int kMaxNumberOfAttempts = 7;
|
const int kMaxNumberOfAttempts = 7;
|
||||||
for (int attempt = 0; attempt < kMaxNumberOfAttempts; attempt++) {
|
for (int attempt = 0; attempt < kMaxNumberOfAttempts; attempt++) {
|
||||||
if (!CollectGarbage(OLD_POINTER_SPACE, MARK_COMPACTOR)) {
|
if (!CollectGarbage(OLD_POINTER_SPACE, MARK_COMPACTOR, gc_reason, NULL)) {
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -469,7 +476,10 @@ void Heap::CollectAllAvailableGarbage() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
bool Heap::CollectGarbage(AllocationSpace space, GarbageCollector collector) {
|
bool Heap::CollectGarbage(AllocationSpace space,
|
||||||
|
GarbageCollector collector,
|
||||||
|
const char* gc_reason,
|
||||||
|
const char* collector_reason) {
|
||||||
// The VM is in the GC state until exiting this function.
|
// The VM is in the GC state until exiting this function.
|
||||||
VMState state(isolate_, GC);
|
VMState state(isolate_, GC);
|
||||||
|
|
||||||
@ -497,11 +507,12 @@ bool Heap::CollectGarbage(AllocationSpace space, GarbageCollector collector) {
|
|||||||
PrintF("[IncrementalMarking] Delaying MarkSweep.\n");
|
PrintF("[IncrementalMarking] Delaying MarkSweep.\n");
|
||||||
}
|
}
|
||||||
collector = SCAVENGER;
|
collector = SCAVENGER;
|
||||||
|
collector_reason = "incremental marking delaying mark-sweep";
|
||||||
}
|
}
|
||||||
|
|
||||||
bool next_gc_likely_to_collect_more = false;
|
bool next_gc_likely_to_collect_more = false;
|
||||||
|
|
||||||
{ GCTracer tracer(this);
|
{ GCTracer tracer(this, gc_reason, collector_reason);
|
||||||
GarbageCollectionPrologue();
|
GarbageCollectionPrologue();
|
||||||
// The GC count was incremented in the prologue. Tell the tracer about
|
// The GC count was incremented in the prologue. Tell the tracer about
|
||||||
// it.
|
// it.
|
||||||
@ -533,7 +544,7 @@ bool Heap::CollectGarbage(AllocationSpace space, GarbageCollector collector) {
|
|||||||
|
|
||||||
|
|
||||||
void Heap::PerformScavenge() {
|
void Heap::PerformScavenge() {
|
||||||
GCTracer tracer(this);
|
GCTracer tracer(this, NULL, NULL);
|
||||||
if (incremental_marking()->IsStopped()) {
|
if (incremental_marking()->IsStopped()) {
|
||||||
PerformGarbageCollection(SCAVENGER, &tracer);
|
PerformGarbageCollection(SCAVENGER, &tracer);
|
||||||
} else {
|
} else {
|
||||||
@ -588,27 +599,33 @@ void Heap::ReserveSpace(
|
|||||||
while (gc_performed && counter++ < kThreshold) {
|
while (gc_performed && counter++ < kThreshold) {
|
||||||
gc_performed = false;
|
gc_performed = false;
|
||||||
if (!new_space->ReserveSpace(new_space_size)) {
|
if (!new_space->ReserveSpace(new_space_size)) {
|
||||||
Heap::CollectGarbage(NEW_SPACE);
|
Heap::CollectGarbage(NEW_SPACE,
|
||||||
|
"failed to reserve space in the new space");
|
||||||
gc_performed = true;
|
gc_performed = true;
|
||||||
}
|
}
|
||||||
if (!old_pointer_space->ReserveSpace(pointer_space_size)) {
|
if (!old_pointer_space->ReserveSpace(pointer_space_size)) {
|
||||||
Heap::CollectGarbage(OLD_POINTER_SPACE);
|
Heap::CollectGarbage(OLD_POINTER_SPACE,
|
||||||
|
"failed to reserve space in the old pointer space");
|
||||||
gc_performed = true;
|
gc_performed = true;
|
||||||
}
|
}
|
||||||
if (!(old_data_space->ReserveSpace(data_space_size))) {
|
if (!(old_data_space->ReserveSpace(data_space_size))) {
|
||||||
Heap::CollectGarbage(OLD_DATA_SPACE);
|
Heap::CollectGarbage(OLD_DATA_SPACE,
|
||||||
|
"failed to reserve space in the old data space");
|
||||||
gc_performed = true;
|
gc_performed = true;
|
||||||
}
|
}
|
||||||
if (!(code_space->ReserveSpace(code_space_size))) {
|
if (!(code_space->ReserveSpace(code_space_size))) {
|
||||||
Heap::CollectGarbage(CODE_SPACE);
|
Heap::CollectGarbage(CODE_SPACE,
|
||||||
|
"failed to reserve space in the code space");
|
||||||
gc_performed = true;
|
gc_performed = true;
|
||||||
}
|
}
|
||||||
if (!(map_space->ReserveSpace(map_space_size))) {
|
if (!(map_space->ReserveSpace(map_space_size))) {
|
||||||
Heap::CollectGarbage(MAP_SPACE);
|
Heap::CollectGarbage(MAP_SPACE,
|
||||||
|
"failed to reserve space in the map space");
|
||||||
gc_performed = true;
|
gc_performed = true;
|
||||||
}
|
}
|
||||||
if (!(cell_space->ReserveSpace(cell_space_size))) {
|
if (!(cell_space->ReserveSpace(cell_space_size))) {
|
||||||
Heap::CollectGarbage(CELL_SPACE);
|
Heap::CollectGarbage(CELL_SPACE,
|
||||||
|
"failed to reserve space in the cell space");
|
||||||
gc_performed = true;
|
gc_performed = true;
|
||||||
}
|
}
|
||||||
// We add a slack-factor of 2 in order to have space for a series of
|
// We add a slack-factor of 2 in order to have space for a series of
|
||||||
@ -620,7 +637,8 @@ void Heap::ReserveSpace(
|
|||||||
large_object_size += cell_space_size + map_space_size + code_space_size +
|
large_object_size += cell_space_size + map_space_size + code_space_size +
|
||||||
data_space_size + pointer_space_size;
|
data_space_size + pointer_space_size;
|
||||||
if (!(lo_space->ReserveSpace(large_object_size))) {
|
if (!(lo_space->ReserveSpace(large_object_size))) {
|
||||||
Heap::CollectGarbage(LO_SPACE);
|
Heap::CollectGarbage(LO_SPACE,
|
||||||
|
"failed to reserve space in the large object space");
|
||||||
gc_performed = true;
|
gc_performed = true;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -4742,7 +4760,7 @@ bool Heap::IsHeapIterable() {
|
|||||||
void Heap::EnsureHeapIsIterable() {
|
void Heap::EnsureHeapIsIterable() {
|
||||||
ASSERT(IsAllocationAllowed());
|
ASSERT(IsAllocationAllowed());
|
||||||
if (!IsHeapIterable()) {
|
if (!IsHeapIterable()) {
|
||||||
CollectAllGarbage(kMakeHeapIterableMask);
|
CollectAllGarbage(kMakeHeapIterableMask, "Heap::EnsureHeapIsIterable");
|
||||||
}
|
}
|
||||||
ASSERT(IsHeapIterable());
|
ASSERT(IsHeapIterable());
|
||||||
}
|
}
|
||||||
@ -4812,7 +4830,7 @@ bool Heap::IdleNotification(int hint) {
|
|||||||
isolate_->compilation_cache()->Clear();
|
isolate_->compilation_cache()->Clear();
|
||||||
uncommit = true;
|
uncommit = true;
|
||||||
}
|
}
|
||||||
CollectAllGarbage(kNoGCFlags);
|
CollectAllGarbage(kNoGCFlags, "idle notification: finalize incremental");
|
||||||
gc_count_at_last_idle_gc_ = gc_count_;
|
gc_count_at_last_idle_gc_ = gc_count_;
|
||||||
if (uncommit) {
|
if (uncommit) {
|
||||||
new_space_.Shrink();
|
new_space_.Shrink();
|
||||||
@ -4853,9 +4871,10 @@ bool Heap::IdleGlobalGC() {
|
|||||||
if (number_idle_notifications_ == kIdlesBeforeScavenge) {
|
if (number_idle_notifications_ == kIdlesBeforeScavenge) {
|
||||||
if (contexts_disposed_ > 0) {
|
if (contexts_disposed_ > 0) {
|
||||||
HistogramTimerScope scope(isolate_->counters()->gc_context());
|
HistogramTimerScope scope(isolate_->counters()->gc_context());
|
||||||
CollectAllGarbage(kNoGCFlags);
|
CollectAllGarbage(kReduceMemoryFootprintMask,
|
||||||
|
"idle notification: contexts disposed");
|
||||||
} else {
|
} else {
|
||||||
CollectGarbage(NEW_SPACE);
|
CollectGarbage(NEW_SPACE, "idle notification");
|
||||||
}
|
}
|
||||||
new_space_.Shrink();
|
new_space_.Shrink();
|
||||||
last_idle_notification_gc_count_ = gc_count_;
|
last_idle_notification_gc_count_ = gc_count_;
|
||||||
@ -4865,12 +4884,12 @@ bool Heap::IdleGlobalGC() {
|
|||||||
// generated code for cached functions.
|
// generated code for cached functions.
|
||||||
isolate_->compilation_cache()->Clear();
|
isolate_->compilation_cache()->Clear();
|
||||||
|
|
||||||
CollectAllGarbage(kNoGCFlags);
|
CollectAllGarbage(kReduceMemoryFootprintMask, "idle notification");
|
||||||
new_space_.Shrink();
|
new_space_.Shrink();
|
||||||
last_idle_notification_gc_count_ = gc_count_;
|
last_idle_notification_gc_count_ = gc_count_;
|
||||||
|
|
||||||
} else if (number_idle_notifications_ == kIdlesBeforeMarkCompact) {
|
} else if (number_idle_notifications_ == kIdlesBeforeMarkCompact) {
|
||||||
CollectAllGarbage(kNoGCFlags);
|
CollectAllGarbage(kReduceMemoryFootprintMask, "idle notification");
|
||||||
new_space_.Shrink();
|
new_space_.Shrink();
|
||||||
last_idle_notification_gc_count_ = gc_count_;
|
last_idle_notification_gc_count_ = gc_count_;
|
||||||
number_idle_notifications_ = 0;
|
number_idle_notifications_ = 0;
|
||||||
@ -4880,7 +4899,8 @@ bool Heap::IdleGlobalGC() {
|
|||||||
contexts_disposed_ = 0;
|
contexts_disposed_ = 0;
|
||||||
} else {
|
} else {
|
||||||
HistogramTimerScope scope(isolate_->counters()->gc_context());
|
HistogramTimerScope scope(isolate_->counters()->gc_context());
|
||||||
CollectAllGarbage(kNoGCFlags);
|
CollectAllGarbage(kReduceMemoryFootprintMask,
|
||||||
|
"idle notification: contexts disposed");
|
||||||
last_idle_notification_gc_count_ = gc_count_;
|
last_idle_notification_gc_count_ = gc_count_;
|
||||||
}
|
}
|
||||||
// If this is the first idle notification, we reset the
|
// If this is the first idle notification, we reset the
|
||||||
@ -6511,7 +6531,9 @@ static intptr_t CountTotalHolesSize() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
GCTracer::GCTracer(Heap* heap)
|
GCTracer::GCTracer(Heap* heap,
|
||||||
|
const char* gc_reason,
|
||||||
|
const char* collector_reason)
|
||||||
: start_time_(0.0),
|
: start_time_(0.0),
|
||||||
start_object_size_(0),
|
start_object_size_(0),
|
||||||
start_memory_size_(0),
|
start_memory_size_(0),
|
||||||
@ -6520,7 +6542,9 @@ GCTracer::GCTracer(Heap* heap)
|
|||||||
allocated_since_last_gc_(0),
|
allocated_since_last_gc_(0),
|
||||||
spent_in_mutator_(0),
|
spent_in_mutator_(0),
|
||||||
promoted_objects_size_(0),
|
promoted_objects_size_(0),
|
||||||
heap_(heap) {
|
heap_(heap),
|
||||||
|
gc_reason_(gc_reason),
|
||||||
|
collector_reason_(collector_reason) {
|
||||||
if (!FLAG_trace_gc && !FLAG_print_cumulative_gc_stat) return;
|
if (!FLAG_trace_gc && !FLAG_print_cumulative_gc_stat) return;
|
||||||
start_time_ = OS::TimeCurrentMillis();
|
start_time_ = OS::TimeCurrentMillis();
|
||||||
start_object_size_ = heap_->SizeOfObjects();
|
start_object_size_ = heap_->SizeOfObjects();
|
||||||
@ -6599,6 +6623,15 @@ GCTracer::~GCTracer() {
|
|||||||
longest_step_);
|
longest_step_);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if (gc_reason_ != NULL) {
|
||||||
|
PrintF(" [%s]", gc_reason_);
|
||||||
|
}
|
||||||
|
|
||||||
|
if (collector_reason_ != NULL) {
|
||||||
|
PrintF(" [%s]", collector_reason_);
|
||||||
|
}
|
||||||
|
|
||||||
PrintF(".\n");
|
PrintF(".\n");
|
||||||
} else {
|
} else {
|
||||||
PrintF("pause=%d ", time);
|
PrintF("pause=%d ", time);
|
||||||
|
23
src/heap.h
23
src/heap.h
@ -1025,23 +1025,28 @@ class Heap {
|
|||||||
// Performs garbage collection operation.
|
// Performs garbage collection operation.
|
||||||
// Returns whether there is a chance that another major GC could
|
// Returns whether there is a chance that another major GC could
|
||||||
// collect more garbage.
|
// collect more garbage.
|
||||||
bool CollectGarbage(AllocationSpace space, GarbageCollector collector);
|
bool CollectGarbage(AllocationSpace space,
|
||||||
|
GarbageCollector collector,
|
||||||
|
const char* gc_reason,
|
||||||
|
const char* collector_reason);
|
||||||
|
|
||||||
// Performs garbage collection operation.
|
// Performs garbage collection operation.
|
||||||
// Returns whether there is a chance that another major GC could
|
// Returns whether there is a chance that another major GC could
|
||||||
// collect more garbage.
|
// collect more garbage.
|
||||||
inline bool CollectGarbage(AllocationSpace space);
|
inline bool CollectGarbage(AllocationSpace space,
|
||||||
|
const char* gc_reason = NULL);
|
||||||
|
|
||||||
static const int kNoGCFlags = 0;
|
static const int kNoGCFlags = 0;
|
||||||
static const int kMakeHeapIterableMask = 1;
|
static const int kMakeHeapIterableMask = 1;
|
||||||
|
static const int kReduceMemoryFootprintMask = 2;
|
||||||
|
|
||||||
// Performs a full garbage collection. If (flags & kMakeHeapIterableMask) is
|
// Performs a full garbage collection. If (flags & kMakeHeapIterableMask) is
|
||||||
// non-zero, then the slower precise sweeper is used, which leaves the heap
|
// non-zero, then the slower precise sweeper is used, which leaves the heap
|
||||||
// in a state where we can iterate over the heap visiting all objects.
|
// in a state where we can iterate over the heap visiting all objects.
|
||||||
void CollectAllGarbage(int flags);
|
void CollectAllGarbage(int flags, const char* gc_reason = NULL);
|
||||||
|
|
||||||
// Last hope GC, should try to squeeze as much as possible.
|
// Last hope GC, should try to squeeze as much as possible.
|
||||||
void CollectAllAvailableGarbage();
|
void CollectAllAvailableGarbage(const char* gc_reason = NULL);
|
||||||
|
|
||||||
// Check whether the heap is currently iterable.
|
// Check whether the heap is currently iterable.
|
||||||
bool IsHeapIterable();
|
bool IsHeapIterable();
|
||||||
@ -1741,7 +1746,8 @@ class Heap {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Checks whether a global GC is necessary
|
// Checks whether a global GC is necessary
|
||||||
GarbageCollector SelectGarbageCollector(AllocationSpace space);
|
GarbageCollector SelectGarbageCollector(AllocationSpace space,
|
||||||
|
const char** reason);
|
||||||
|
|
||||||
// Performs garbage collection
|
// Performs garbage collection
|
||||||
// Returns whether there is a chance another major GC could
|
// Returns whether there is a chance another major GC could
|
||||||
@ -2364,7 +2370,9 @@ class GCTracer BASE_EMBEDDED {
|
|||||||
double start_time_;
|
double start_time_;
|
||||||
};
|
};
|
||||||
|
|
||||||
explicit GCTracer(Heap* heap);
|
explicit GCTracer(Heap* heap,
|
||||||
|
const char* gc_reason,
|
||||||
|
const char* collector_reason);
|
||||||
~GCTracer();
|
~GCTracer();
|
||||||
|
|
||||||
// Sets the collector.
|
// Sets the collector.
|
||||||
@ -2432,6 +2440,9 @@ class GCTracer BASE_EMBEDDED {
|
|||||||
double steps_took_since_last_gc_;
|
double steps_took_since_last_gc_;
|
||||||
|
|
||||||
Heap* heap_;
|
Heap* heap_;
|
||||||
|
|
||||||
|
const char* gc_reason_;
|
||||||
|
const char* collector_reason_;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
||||||
|
@ -1521,7 +1521,8 @@ void Logger::LowLevelLogWriteBytes(const char* bytes, int size) {
|
|||||||
|
|
||||||
|
|
||||||
void Logger::LogCodeObjects() {
|
void Logger::LogCodeObjects() {
|
||||||
HEAP->CollectAllGarbage(Heap::kMakeHeapIterableMask);
|
HEAP->CollectAllGarbage(Heap::kMakeHeapIterableMask,
|
||||||
|
"Logger::LogCodeObjects");
|
||||||
HeapIterator iterator;
|
HeapIterator iterator;
|
||||||
AssertNoAllocation no_alloc;
|
AssertNoAllocation no_alloc;
|
||||||
for (HeapObject* obj = iterator.next(); obj != NULL; obj = iterator.next()) {
|
for (HeapObject* obj = iterator.next(); obj != NULL; obj = iterator.next()) {
|
||||||
@ -1576,7 +1577,8 @@ void Logger::LogExistingFunction(Handle<SharedFunctionInfo> shared,
|
|||||||
|
|
||||||
|
|
||||||
void Logger::LogCompiledFunctions() {
|
void Logger::LogCompiledFunctions() {
|
||||||
HEAP->CollectAllGarbage(Heap::kMakeHeapIterableMask);
|
HEAP->CollectAllGarbage(Heap::kMakeHeapIterableMask,
|
||||||
|
"Logger::LogCompiledFunctions");
|
||||||
HandleScope scope;
|
HandleScope scope;
|
||||||
const int compiled_funcs_count = EnumerateCompiledFunctions(NULL, NULL);
|
const int compiled_funcs_count = EnumerateCompiledFunctions(NULL, NULL);
|
||||||
ScopedVector< Handle<SharedFunctionInfo> > sfis(compiled_funcs_count);
|
ScopedVector< Handle<SharedFunctionInfo> > sfis(compiled_funcs_count);
|
||||||
@ -1595,7 +1597,8 @@ void Logger::LogCompiledFunctions() {
|
|||||||
|
|
||||||
|
|
||||||
void Logger::LogAccessorCallbacks() {
|
void Logger::LogAccessorCallbacks() {
|
||||||
HEAP->CollectAllGarbage(Heap::kMakeHeapIterableMask);
|
HEAP->CollectAllGarbage(Heap::kMakeHeapIterableMask,
|
||||||
|
"Logger::LogAccessorCallbacks");
|
||||||
HeapIterator iterator;
|
HeapIterator iterator;
|
||||||
AssertNoAllocation no_alloc;
|
AssertNoAllocation no_alloc;
|
||||||
for (HeapObject* obj = iterator.next(); obj != NULL; obj = iterator.next()) {
|
for (HeapObject* obj = iterator.next(); obj != NULL; obj = iterator.next()) {
|
||||||
|
@ -46,6 +46,7 @@ MarkBit Marking::MarkBitFrom(Address addr) {
|
|||||||
|
|
||||||
void MarkCompactCollector::SetFlags(int flags) {
|
void MarkCompactCollector::SetFlags(int flags) {
|
||||||
sweep_precisely_ = ((flags & Heap::kMakeHeapIterableMask) != 0);
|
sweep_precisely_ = ((flags & Heap::kMakeHeapIterableMask) != 0);
|
||||||
|
reduce_memory_footprint_ = ((flags & Heap::kReduceMemoryFootprintMask) != 0);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
@ -230,6 +230,18 @@ void MarkCompactCollector::AddEvacuationCandidate(Page* p) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
static void TraceFragmentation(PagedSpace* space) {
|
||||||
|
int number_of_pages = space->CountTotalPages();
|
||||||
|
intptr_t reserved = (number_of_pages * Page::kObjectAreaSize);
|
||||||
|
intptr_t free = reserved - space->SizeOfObjects();
|
||||||
|
PrintF("[%s]: %d pages, %d (%.1f%%) free\n",
|
||||||
|
AllocationSpaceName(space->identity()),
|
||||||
|
number_of_pages,
|
||||||
|
static_cast<int>(free),
|
||||||
|
static_cast<double>(free) * 100 / reserved);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
bool MarkCompactCollector::StartCompaction() {
|
bool MarkCompactCollector::StartCompaction() {
|
||||||
if (!compacting_) {
|
if (!compacting_) {
|
||||||
ASSERT(evacuation_candidates_.length() == 0);
|
ASSERT(evacuation_candidates_.length() == 0);
|
||||||
@ -239,6 +251,13 @@ bool MarkCompactCollector::StartCompaction() {
|
|||||||
|
|
||||||
if (FLAG_compact_code_space) {
|
if (FLAG_compact_code_space) {
|
||||||
CollectEvacuationCandidates(heap()->code_space());
|
CollectEvacuationCandidates(heap()->code_space());
|
||||||
|
} else if (FLAG_trace_fragmentation) {
|
||||||
|
TraceFragmentation(heap()->code_space());
|
||||||
|
}
|
||||||
|
|
||||||
|
if (FLAG_trace_fragmentation) {
|
||||||
|
TraceFragmentation(heap()->map_space());
|
||||||
|
TraceFragmentation(heap()->cell_space());
|
||||||
}
|
}
|
||||||
|
|
||||||
heap()->old_pointer_space()->EvictEvacuationCandidatesFromFreeLists();
|
heap()->old_pointer_space()->EvictEvacuationCandidatesFromFreeLists();
|
||||||
@ -414,6 +433,65 @@ const char* AllocationSpaceName(AllocationSpace space) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
// Returns zero for pages that have so little fragmentation that it is not
|
||||||
|
// worth defragmenting them. Otherwise a positive integer that gives an
|
||||||
|
// estimate of fragmentation on an arbitrary scale.
|
||||||
|
static int FreeListFragmentation(PagedSpace* space, Page* p) {
|
||||||
|
// If page was not swept then there are no free list items on it.
|
||||||
|
if (!p->WasSwept()) {
|
||||||
|
if (FLAG_trace_fragmentation) {
|
||||||
|
PrintF("%p [%s]: %d bytes live (unswept)\n",
|
||||||
|
reinterpret_cast<void*>(p),
|
||||||
|
AllocationSpaceName(space->identity()),
|
||||||
|
p->LiveBytes());
|
||||||
|
}
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
FreeList::SizeStats sizes;
|
||||||
|
space->CountFreeListItems(p, &sizes);
|
||||||
|
|
||||||
|
intptr_t ratio;
|
||||||
|
intptr_t ratio_threshold;
|
||||||
|
if (space->identity() == CODE_SPACE) {
|
||||||
|
ratio = (sizes.medium_size_ * 10 + sizes.large_size_ * 2) * 100 /
|
||||||
|
Page::kObjectAreaSize;
|
||||||
|
ratio_threshold = 10;
|
||||||
|
} else {
|
||||||
|
ratio = (sizes.small_size_ * 5 + sizes.medium_size_) * 100 /
|
||||||
|
Page::kObjectAreaSize;
|
||||||
|
ratio_threshold = 15;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (FLAG_trace_fragmentation) {
|
||||||
|
PrintF("%p [%s]: %d (%.2f%%) %d (%.2f%%) %d (%.2f%%) %d (%.2f%%) %s\n",
|
||||||
|
reinterpret_cast<void*>(p),
|
||||||
|
AllocationSpaceName(space->identity()),
|
||||||
|
static_cast<int>(sizes.small_size_),
|
||||||
|
static_cast<double>(sizes.small_size_ * 100) /
|
||||||
|
Page::kObjectAreaSize,
|
||||||
|
static_cast<int>(sizes.medium_size_),
|
||||||
|
static_cast<double>(sizes.medium_size_ * 100) /
|
||||||
|
Page::kObjectAreaSize,
|
||||||
|
static_cast<int>(sizes.large_size_),
|
||||||
|
static_cast<double>(sizes.large_size_ * 100) /
|
||||||
|
Page::kObjectAreaSize,
|
||||||
|
static_cast<int>(sizes.huge_size_),
|
||||||
|
static_cast<double>(sizes.huge_size_ * 100) /
|
||||||
|
Page::kObjectAreaSize,
|
||||||
|
(ratio > ratio_threshold) ? "[fragmented]" : "");
|
||||||
|
}
|
||||||
|
|
||||||
|
if (FLAG_always_compact && sizes.Total() != Page::kObjectAreaSize) {
|
||||||
|
return 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (ratio <= ratio_threshold) return 0; // Not fragmented.
|
||||||
|
|
||||||
|
return static_cast<int>(ratio - ratio_threshold);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
void MarkCompactCollector::CollectEvacuationCandidates(PagedSpace* space) {
|
void MarkCompactCollector::CollectEvacuationCandidates(PagedSpace* space) {
|
||||||
ASSERT(space->identity() == OLD_POINTER_SPACE ||
|
ASSERT(space->identity() == OLD_POINTER_SPACE ||
|
||||||
space->identity() == OLD_DATA_SPACE ||
|
space->identity() == OLD_DATA_SPACE ||
|
||||||
@ -421,7 +499,6 @@ void MarkCompactCollector::CollectEvacuationCandidates(PagedSpace* space) {
|
|||||||
|
|
||||||
int number_of_pages = space->CountTotalPages();
|
int number_of_pages = space->CountTotalPages();
|
||||||
|
|
||||||
PageIterator it(space);
|
|
||||||
const int kMaxMaxEvacuationCandidates = 1000;
|
const int kMaxMaxEvacuationCandidates = 1000;
|
||||||
int max_evacuation_candidates = Min(
|
int max_evacuation_candidates = Min(
|
||||||
kMaxMaxEvacuationCandidates,
|
kMaxMaxEvacuationCandidates,
|
||||||
@ -444,22 +521,89 @@ void MarkCompactCollector::CollectEvacuationCandidates(PagedSpace* space) {
|
|||||||
Page* page_;
|
Page* page_;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
enum CompactionMode {
|
||||||
|
COMPACT_FREE_LISTS,
|
||||||
|
REDUCE_MEMORY_FOOTPRINT
|
||||||
|
};
|
||||||
|
|
||||||
|
CompactionMode mode = COMPACT_FREE_LISTS;
|
||||||
|
|
||||||
|
intptr_t reserved = number_of_pages * Page::kObjectAreaSize;
|
||||||
|
intptr_t over_reserved = reserved - space->SizeOfObjects();
|
||||||
|
static const intptr_t kFreenessThreshold = 50;
|
||||||
|
|
||||||
|
if (over_reserved >= 2 * Page::kObjectAreaSize &&
|
||||||
|
reduce_memory_footprint_) {
|
||||||
|
mode = REDUCE_MEMORY_FOOTPRINT;
|
||||||
|
|
||||||
|
// We expect that empty pages are easier to compact so slightly bump the
|
||||||
|
// limit.
|
||||||
|
max_evacuation_candidates += 2;
|
||||||
|
|
||||||
|
if (FLAG_trace_fragmentation) {
|
||||||
|
PrintF("Estimated over reserved memory: %.1f MB (setting threshold %d)\n",
|
||||||
|
static_cast<double>(over_reserved) / MB,
|
||||||
|
kFreenessThreshold);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
intptr_t estimated_release = 0;
|
||||||
|
|
||||||
Candidate candidates[kMaxMaxEvacuationCandidates];
|
Candidate candidates[kMaxMaxEvacuationCandidates];
|
||||||
|
|
||||||
int count = 0;
|
int count = 0;
|
||||||
if (it.has_next()) it.next(); // Never compact the first page.
|
|
||||||
int fragmentation = 0;
|
int fragmentation = 0;
|
||||||
Candidate* least = NULL;
|
Candidate* least = NULL;
|
||||||
|
|
||||||
|
PageIterator it(space);
|
||||||
|
if (it.has_next()) it.next(); // Never compact the first page.
|
||||||
|
|
||||||
while (it.has_next()) {
|
while (it.has_next()) {
|
||||||
Page* p = it.next();
|
Page* p = it.next();
|
||||||
p->ClearEvacuationCandidate();
|
p->ClearEvacuationCandidate();
|
||||||
|
|
||||||
if (FLAG_stress_compaction) {
|
if (FLAG_stress_compaction) {
|
||||||
int counter = space->heap()->ms_count();
|
int counter = space->heap()->ms_count();
|
||||||
uintptr_t page_number = reinterpret_cast<uintptr_t>(p) >> kPageSizeBits;
|
uintptr_t page_number = reinterpret_cast<uintptr_t>(p) >> kPageSizeBits;
|
||||||
if ((counter & 1) == (page_number & 1)) fragmentation = 1;
|
if ((counter & 1) == (page_number & 1)) fragmentation = 1;
|
||||||
|
} else if (mode == REDUCE_MEMORY_FOOTPRINT) {
|
||||||
|
// Don't try to release too many pages.
|
||||||
|
if (estimated_release >= ((over_reserved * 3) / 4)) {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
intptr_t free_bytes = 0;
|
||||||
|
|
||||||
|
if (!p->WasSwept()) {
|
||||||
|
free_bytes = (Page::kObjectAreaSize - p->LiveBytes());
|
||||||
|
} else {
|
||||||
|
FreeList::SizeStats sizes;
|
||||||
|
space->CountFreeListItems(p, &sizes);
|
||||||
|
free_bytes = sizes.Total();
|
||||||
|
}
|
||||||
|
|
||||||
|
int free_pct = static_cast<int>(free_bytes * 100 / Page::kObjectAreaSize);
|
||||||
|
|
||||||
|
if (free_pct >= kFreenessThreshold) {
|
||||||
|
estimated_release += Page::kObjectAreaSize +
|
||||||
|
(Page::kObjectAreaSize - free_bytes);
|
||||||
|
fragmentation = free_pct;
|
||||||
|
} else {
|
||||||
|
fragmentation = 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (FLAG_trace_fragmentation) {
|
||||||
|
PrintF("%p [%s]: %d (%.2f%%) free %s\n",
|
||||||
|
reinterpret_cast<void*>(p),
|
||||||
|
AllocationSpaceName(space->identity()),
|
||||||
|
free_bytes,
|
||||||
|
static_cast<double>(free_bytes * 100) / Page::kObjectAreaSize,
|
||||||
|
(fragmentation > 0) ? "[fragmented]" : "");
|
||||||
|
}
|
||||||
} else {
|
} else {
|
||||||
fragmentation = space->Fragmentation(p);
|
fragmentation = FreeListFragmentation(space, p);
|
||||||
}
|
}
|
||||||
|
|
||||||
if (fragmentation != 0) {
|
if (fragmentation != 0) {
|
||||||
if (count < max_evacuation_candidates) {
|
if (count < max_evacuation_candidates) {
|
||||||
candidates[count++] = Candidate(fragmentation, p);
|
candidates[count++] = Candidate(fragmentation, p);
|
||||||
@ -479,6 +623,7 @@ void MarkCompactCollector::CollectEvacuationCandidates(PagedSpace* space) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
for (int i = 0; i < count; i++) {
|
for (int i = 0; i < count; i++) {
|
||||||
AddEvacuationCandidate(candidates[i].page());
|
AddEvacuationCandidate(candidates[i].page());
|
||||||
}
|
}
|
||||||
@ -3242,6 +3387,8 @@ void MarkCompactCollector::EvacuateNewSpaceAndCandidates() {
|
|||||||
p->set_scan_on_scavenge(false);
|
p->set_scan_on_scavenge(false);
|
||||||
slots_buffer_allocator_.DeallocateChain(p->slots_buffer_address());
|
slots_buffer_allocator_.DeallocateChain(p->slots_buffer_address());
|
||||||
p->ClearEvacuationCandidate();
|
p->ClearEvacuationCandidate();
|
||||||
|
p->ResetLiveBytes();
|
||||||
|
space->ReleasePage(p);
|
||||||
}
|
}
|
||||||
evacuation_candidates_.Rewind(0);
|
evacuation_candidates_.Rewind(0);
|
||||||
compacting_ = false;
|
compacting_ = false;
|
||||||
|
@ -374,7 +374,7 @@ class SlotsBuffer {
|
|||||||
static const int kNumberOfElements = 1021;
|
static const int kNumberOfElements = 1021;
|
||||||
|
|
||||||
private:
|
private:
|
||||||
static const int kChainLengthThreshold = 6;
|
static const int kChainLengthThreshold = 15;
|
||||||
|
|
||||||
intptr_t idx_;
|
intptr_t idx_;
|
||||||
intptr_t chain_length_;
|
intptr_t chain_length_;
|
||||||
@ -572,6 +572,8 @@ class MarkCompactCollector {
|
|||||||
// heap.
|
// heap.
|
||||||
bool sweep_precisely_;
|
bool sweep_precisely_;
|
||||||
|
|
||||||
|
bool reduce_memory_footprint_;
|
||||||
|
|
||||||
// True if we are collecting slots to perform evacuation from evacuation
|
// True if we are collecting slots to perform evacuation from evacuation
|
||||||
// candidates.
|
// candidates.
|
||||||
bool compacting_;
|
bool compacting_;
|
||||||
|
@ -312,7 +312,7 @@ int main(int argc, char** argv) {
|
|||||||
}
|
}
|
||||||
// If we don't do this then we end up with a stray root pointing at the
|
// If we don't do this then we end up with a stray root pointing at the
|
||||||
// context even after we have disposed of the context.
|
// context even after we have disposed of the context.
|
||||||
HEAP->CollectAllGarbage(i::Heap::kNoGCFlags);
|
HEAP->CollectAllGarbage(i::Heap::kNoGCFlags, "mksnapshot");
|
||||||
i::Object* raw_context = *(v8::Utils::OpenHandle(*context));
|
i::Object* raw_context = *(v8::Utils::OpenHandle(*context));
|
||||||
context.Dispose();
|
context.Dispose();
|
||||||
CppByteSink sink(argv[1]);
|
CppByteSink sink(argv[1]);
|
||||||
|
@ -1562,7 +1562,8 @@ void HeapSnapshotsCollection::RemoveSnapshot(HeapSnapshot* snapshot) {
|
|||||||
|
|
||||||
Handle<HeapObject> HeapSnapshotsCollection::FindHeapObjectById(uint64_t id) {
|
Handle<HeapObject> HeapSnapshotsCollection::FindHeapObjectById(uint64_t id) {
|
||||||
// First perform a full GC in order to avoid dead objects.
|
// First perform a full GC in order to avoid dead objects.
|
||||||
HEAP->CollectAllGarbage(Heap::kMakeHeapIterableMask);
|
HEAP->CollectAllGarbage(Heap::kMakeHeapIterableMask,
|
||||||
|
"HeapSnapshotsCollection::FindHeapObjectById");
|
||||||
AssertNoAllocation no_allocation;
|
AssertNoAllocation no_allocation;
|
||||||
HeapObject* object = NULL;
|
HeapObject* object = NULL;
|
||||||
HeapIterator iterator(HeapIterator::kFilterUnreachable);
|
HeapIterator iterator(HeapIterator::kFilterUnreachable);
|
||||||
@ -3056,8 +3057,12 @@ bool HeapSnapshotGenerator::GenerateSnapshot() {
|
|||||||
// full GC is reachable from the root when computing dominators.
|
// full GC is reachable from the root when computing dominators.
|
||||||
// This is not true for weakly reachable objects.
|
// This is not true for weakly reachable objects.
|
||||||
// As a temporary solution we call GC twice.
|
// As a temporary solution we call GC twice.
|
||||||
Isolate::Current()->heap()->CollectAllGarbage(Heap::kMakeHeapIterableMask);
|
Isolate::Current()->heap()->CollectAllGarbage(
|
||||||
Isolate::Current()->heap()->CollectAllGarbage(Heap::kMakeHeapIterableMask);
|
Heap::kMakeHeapIterableMask,
|
||||||
|
"HeapSnapshotGenerator::GenerateSnapshot");
|
||||||
|
Isolate::Current()->heap()->CollectAllGarbage(
|
||||||
|
Heap::kMakeHeapIterableMask,
|
||||||
|
"HeapSnapshotGenerator::GenerateSnapshot");
|
||||||
|
|
||||||
#ifdef DEBUG
|
#ifdef DEBUG
|
||||||
Heap* debug_heap = Isolate::Current()->heap();
|
Heap* debug_heap = Isolate::Current()->heap();
|
||||||
|
@ -12445,7 +12445,8 @@ RUNTIME_FUNCTION(MaybeObject*, Runtime_DebugReferencedBy) {
|
|||||||
ASSERT(args.length() == 3);
|
ASSERT(args.length() == 3);
|
||||||
|
|
||||||
// First perform a full GC in order to avoid references from dead objects.
|
// First perform a full GC in order to avoid references from dead objects.
|
||||||
isolate->heap()->CollectAllGarbage(Heap::kMakeHeapIterableMask);
|
isolate->heap()->CollectAllGarbage(Heap::kMakeHeapIterableMask,
|
||||||
|
"%DebugReferencedBy");
|
||||||
// The heap iterator reserves the right to do a GC to make the heap iterable.
|
// The heap iterator reserves the right to do a GC to make the heap iterable.
|
||||||
// Due to the GC above we know it won't need to do that, but it seems cleaner
|
// Due to the GC above we know it won't need to do that, but it seems cleaner
|
||||||
// to get the heap iterator constructed before we start having unprotected
|
// to get the heap iterator constructed before we start having unprotected
|
||||||
@ -12536,7 +12537,8 @@ RUNTIME_FUNCTION(MaybeObject*, Runtime_DebugConstructedBy) {
|
|||||||
ASSERT(args.length() == 2);
|
ASSERT(args.length() == 2);
|
||||||
|
|
||||||
// First perform a full GC in order to avoid dead objects.
|
// First perform a full GC in order to avoid dead objects.
|
||||||
isolate->heap()->CollectAllGarbage(Heap::kMakeHeapIterableMask);
|
isolate->heap()->CollectAllGarbage(Heap::kMakeHeapIterableMask,
|
||||||
|
"%DebugConstructedBy");
|
||||||
|
|
||||||
// Check parameters.
|
// Check parameters.
|
||||||
CONVERT_CHECKED(JSFunction, constructor, args[0]);
|
CONVERT_CHECKED(JSFunction, constructor, args[0]);
|
||||||
@ -12934,7 +12936,7 @@ RUNTIME_FUNCTION(MaybeObject*, Runtime_SetFlags) {
|
|||||||
// Performs a GC.
|
// Performs a GC.
|
||||||
// Presently, it only does a full GC.
|
// Presently, it only does a full GC.
|
||||||
RUNTIME_FUNCTION(MaybeObject*, Runtime_CollectGarbage) {
|
RUNTIME_FUNCTION(MaybeObject*, Runtime_CollectGarbage) {
|
||||||
isolate->heap()->CollectAllGarbage(true);
|
isolate->heap()->CollectAllGarbage(true, "%CollectGarbage");
|
||||||
return isolate->heap()->undefined_value();
|
return isolate->heap()->undefined_value();
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -13645,12 +13647,14 @@ void Runtime::PerformGC(Object* result) {
|
|||||||
}
|
}
|
||||||
// Try to do a garbage collection; ignore it if it fails. The C
|
// Try to do a garbage collection; ignore it if it fails. The C
|
||||||
// entry stub will throw an out-of-memory exception in that case.
|
// entry stub will throw an out-of-memory exception in that case.
|
||||||
isolate->heap()->CollectGarbage(failure->allocation_space());
|
isolate->heap()->CollectGarbage(failure->allocation_space(),
|
||||||
|
"Runtime::PerformGC");
|
||||||
} else {
|
} else {
|
||||||
// Handle last resort GC and make sure to allow future allocations
|
// Handle last resort GC and make sure to allow future allocations
|
||||||
// to grow the heap without causing GCs (if possible).
|
// to grow the heap without causing GCs (if possible).
|
||||||
isolate->counters()->gc_last_resort_from_js()->Increment();
|
isolate->counters()->gc_last_resort_from_js()->Increment();
|
||||||
isolate->heap()->CollectAllGarbage(Heap::kNoGCFlags);
|
isolate->heap()->CollectAllGarbage(Heap::kNoGCFlags,
|
||||||
|
"Runtime::PerformGC");
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
46
src/spaces.h
46
src/spaces.h
@ -1589,50 +1589,8 @@ class PagedSpace : public Space {
|
|||||||
Page* FirstPage() { return anchor_.next_page(); }
|
Page* FirstPage() { return anchor_.next_page(); }
|
||||||
Page* LastPage() { return anchor_.prev_page(); }
|
Page* LastPage() { return anchor_.prev_page(); }
|
||||||
|
|
||||||
// Returns zero for pages that have so little fragmentation that it is not
|
void CountFreeListItems(Page* p, FreeList::SizeStats* sizes) {
|
||||||
// worth defragmenting them. Otherwise a positive integer that gives an
|
free_list_.CountFreeListItems(p, sizes);
|
||||||
// estimate of fragmentation on an arbitrary scale.
|
|
||||||
int Fragmentation(Page* p) {
|
|
||||||
FreeList::SizeStats sizes;
|
|
||||||
free_list_.CountFreeListItems(p, &sizes);
|
|
||||||
|
|
||||||
intptr_t ratio;
|
|
||||||
intptr_t ratio_threshold;
|
|
||||||
if (identity() == CODE_SPACE) {
|
|
||||||
ratio = (sizes.medium_size_ * 10 + sizes.large_size_ * 2) * 100 /
|
|
||||||
Page::kObjectAreaSize;
|
|
||||||
ratio_threshold = 10;
|
|
||||||
} else {
|
|
||||||
ratio = (sizes.small_size_ * 5 + sizes.medium_size_) * 100 /
|
|
||||||
Page::kObjectAreaSize;
|
|
||||||
ratio_threshold = 15;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (FLAG_trace_fragmentation) {
|
|
||||||
PrintF("%p [%d]: %d (%.2f%%) %d (%.2f%%) %d (%.2f%%) %d (%.2f%%) %s\n",
|
|
||||||
reinterpret_cast<void*>(p),
|
|
||||||
identity(),
|
|
||||||
static_cast<int>(sizes.small_size_),
|
|
||||||
static_cast<double>(sizes.small_size_ * 100) /
|
|
||||||
Page::kObjectAreaSize,
|
|
||||||
static_cast<int>(sizes.medium_size_),
|
|
||||||
static_cast<double>(sizes.medium_size_ * 100) /
|
|
||||||
Page::kObjectAreaSize,
|
|
||||||
static_cast<int>(sizes.large_size_),
|
|
||||||
static_cast<double>(sizes.large_size_ * 100) /
|
|
||||||
Page::kObjectAreaSize,
|
|
||||||
static_cast<int>(sizes.huge_size_),
|
|
||||||
static_cast<double>(sizes.huge_size_ * 100) /
|
|
||||||
Page::kObjectAreaSize,
|
|
||||||
(ratio > ratio_threshold) ? "[fragmented]" : "");
|
|
||||||
}
|
|
||||||
|
|
||||||
if (FLAG_always_compact && sizes.Total() != Page::kObjectAreaSize) {
|
|
||||||
return 1;
|
|
||||||
}
|
|
||||||
if (ratio <= ratio_threshold) return 0; // Not fragmented.
|
|
||||||
|
|
||||||
return static_cast<int>(ratio - ratio_threshold);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
void EvictEvacuationCandidatesFromFreeLists();
|
void EvictEvacuationCandidatesFromFreeLists();
|
||||||
|
Loading…
Reference in New Issue
Block a user